| /linux/lib/crypto/ |
| H A D | gf128mul.c | 57 #define gf128mul_dat(q) { \ argument 58 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\ 59 q( [all...] |
| /linux/Documentation/networking/ |
| H A D | tls-offload-layers.svg | 1 …q-0.609375 -0.359375 -1.203125 -0.359375q-0.546875 0 -0.96875 0.328125q-0.421875 0.328125 -0.60937…
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_rx.c | 50 static void cache_offer(struct funeth_rxq *q, const struct funeth_rxbuf *buf) in cache_offer() argument 52 struct funeth_rx_cache *c = &q->cache; in cache_offer() 58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_offer() 67 static bool cache_get(struct funeth_rxq *q, struct funeth_rxbuf *rb) in cache_get() argument 69 struct funeth_rx_cache *c = &q->cache; in cache_get() 77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, in cache_get() 88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_get() 98 static int funeth_alloc_page(struct funeth_rxq *q, struct funeth_rxbuf *rb, in funeth_alloc_page() argument 103 if (cache_get(q, rb)) in funeth_alloc_page() 110 rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE, in funeth_alloc_page() [all …]
|
| H A D | funeth_tx.c | 56 static void *txq_end(const struct funeth_txq *q) in txq_end() argument 58 return (void *)q->hw_wb; in txq_end() 64 static unsigned int txq_to_end(const struct funeth_txq *q, void *p) in txq_to_end() argument 66 return txq_end(q) - p; in txq_to_end() 78 static struct fun_dataop_gl *fun_write_gl(const struct funeth_txq *q, in fun_write_gl() argument 90 i < ngle && txq_to_end(q, gle); i++, gle++) in fun_write_gl() 93 if (txq_to_end(q, gle) == 0) { in fun_write_gl() 94 gle = (struct fun_dataop_gl *)q->desc; in fun_write_gl() 107 static struct sk_buff *fun_tls_tx(struct sk_buff *skb, struct funeth_txq *q, in fun_tls_tx() argument 132 FUN_QSTAT_INC(q, tx_tls_fallback); in fun_tls_tx() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_guc_submit.c | 53 exec_queue_to_guc(struct xe_exec_queue *q) 55 return &q->gt->uc.guc; in exec_queue_to_guc() 76 static bool exec_queue_registered(struct xe_exec_queue *q) 78 return atomic_read(&q->guc->state) & EXEC_QUEUE_STATE_REGISTERED; in exec_queue_registered() 81 static void set_exec_queue_registered(struct xe_exec_queue *q) 83 atomic_or(EXEC_QUEUE_STATE_REGISTERED, &q->guc->state); in set_exec_queue_registered() 86 static void clear_exec_queue_registered(struct xe_exec_queue *q) 88 atomic_and(~EXEC_QUEUE_STATE_REGISTERED, &q->guc->state); in clear_exec_queue_registered() 91 static bool exec_queue_enabled(struct xe_exec_queue *q) 93 return atomic_read(&q in exec_queue_enabled() 54 exec_queue_to_guc(struct xe_exec_queue * q) exec_queue_to_guc() argument 77 exec_queue_registered(struct xe_exec_queue * q) exec_queue_registered() argument 82 set_exec_queue_registered(struct xe_exec_queue * q) set_exec_queue_registered() argument 87 clear_exec_queue_registered(struct xe_exec_queue * q) clear_exec_queue_registered() argument 92 exec_queue_enabled(struct xe_exec_queue * q) exec_queue_enabled() argument 97 set_exec_queue_enabled(struct xe_exec_queue * q) set_exec_queue_enabled() argument 102 clear_exec_queue_enabled(struct xe_exec_queue * q) clear_exec_queue_enabled() argument 107 exec_queue_pending_enable(struct xe_exec_queue * q) exec_queue_pending_enable() argument 112 set_exec_queue_pending_enable(struct xe_exec_queue * q) set_exec_queue_pending_enable() argument 117 clear_exec_queue_pending_enable(struct xe_exec_queue * q) clear_exec_queue_pending_enable() argument 122 exec_queue_pending_disable(struct xe_exec_queue * q) exec_queue_pending_disable() argument 127 set_exec_queue_pending_disable(struct xe_exec_queue * q) set_exec_queue_pending_disable() argument 132 clear_exec_queue_pending_disable(struct xe_exec_queue * q) clear_exec_queue_pending_disable() argument 137 exec_queue_destroyed(struct xe_exec_queue * q) exec_queue_destroyed() argument 142 set_exec_queue_destroyed(struct xe_exec_queue * q) set_exec_queue_destroyed() argument 147 clear_exec_queue_destroyed(struct xe_exec_queue * q) clear_exec_queue_destroyed() argument 152 exec_queue_banned(struct xe_exec_queue * q) exec_queue_banned() argument 157 set_exec_queue_banned(struct xe_exec_queue * q) set_exec_queue_banned() argument 162 exec_queue_suspended(struct xe_exec_queue * q) exec_queue_suspended() argument 167 set_exec_queue_suspended(struct xe_exec_queue * q) set_exec_queue_suspended() argument 172 clear_exec_queue_suspended(struct xe_exec_queue * q) clear_exec_queue_suspended() argument 177 exec_queue_reset(struct xe_exec_queue * q) exec_queue_reset() argument 182 set_exec_queue_reset(struct xe_exec_queue * q) set_exec_queue_reset() argument 187 exec_queue_killed(struct xe_exec_queue * q) exec_queue_killed() argument 192 set_exec_queue_killed(struct xe_exec_queue * q) set_exec_queue_killed() argument 197 exec_queue_wedged(struct xe_exec_queue * q) exec_queue_wedged() argument 202 set_exec_queue_wedged(struct xe_exec_queue * q) set_exec_queue_wedged() argument 207 exec_queue_pending_resume(struct xe_exec_queue * q) exec_queue_pending_resume() argument 212 set_exec_queue_pending_resume(struct xe_exec_queue * q) set_exec_queue_pending_resume() argument 217 clear_exec_queue_pending_resume(struct xe_exec_queue * q) clear_exec_queue_pending_resume() argument 222 exec_queue_idle_skip_suspend(struct xe_exec_queue * q) exec_queue_idle_skip_suspend() argument 227 set_exec_queue_idle_skip_suspend(struct xe_exec_queue * q) set_exec_queue_idle_skip_suspend() argument 232 clear_exec_queue_idle_skip_suspend(struct xe_exec_queue * q) clear_exec_queue_idle_skip_suspend() argument 237 exec_queue_killed_or_banned_or_wedged(struct xe_exec_queue * q) exec_queue_killed_or_banned_or_wedged() argument 278 struct xe_exec_queue *q; guc_submit_wedged_fini() local 416 __release_guc_id(struct xe_guc * guc,struct xe_exec_queue * q,u32 xa_count) __release_guc_id() argument 432 alloc_guc_id(struct xe_guc * guc,struct xe_exec_queue * q) alloc_guc_id() argument 468 release_guc_id(struct xe_guc * guc,struct xe_exec_queue * q) release_guc_id() argument 524 init_policies(struct xe_guc * guc,struct xe_exec_queue * q) init_policies() argument 549 set_min_preemption_timeout(struct xe_guc * guc,struct xe_exec_queue * q) set_min_preemption_timeout() argument 567 xe_guc_exec_queue_trigger_cleanup(struct xe_exec_queue * q) xe_guc_exec_queue_trigger_cleanup() argument 578 xe_guc_exec_queue_group_trigger_cleanup(struct xe_exec_queue * q) xe_guc_exec_queue_group_trigger_cleanup() argument 597 xe_guc_exec_queue_reset_trigger_cleanup(struct xe_exec_queue * q) xe_guc_exec_queue_reset_trigger_cleanup() argument 625 set_exec_queue_group_banned(struct xe_exec_queue * q) set_exec_queue_group_banned() argument 726 xe_guc_exec_queue_group_cgp_update(struct xe_device * xe,struct xe_exec_queue * q) xe_guc_exec_queue_group_cgp_update() argument 755 xe_guc_exec_queue_group_cgp_sync(struct xe_guc * guc,struct xe_exec_queue * q,const u32 * action,u32 len) xe_guc_exec_queue_group_cgp_sync() argument 791 __register_exec_queue_group(struct xe_guc * guc,struct xe_exec_queue * q,struct guc_ctxt_registration_info * info) __register_exec_queue_group() argument 819 xe_guc_exec_queue_group_add(struct xe_guc * guc,struct xe_exec_queue * q) xe_guc_exec_queue_group_add() argument 842 __register_mlrc_exec_queue(struct xe_guc * guc,struct xe_exec_queue * q,struct guc_ctxt_registration_info * info) __register_mlrc_exec_queue() argument 917 register_exec_queue(struct xe_exec_queue * q,int ctx_type) register_exec_queue() argument 979 wq_space_until_wrap(struct xe_exec_queue * q) wq_space_until_wrap() argument 1002 wq_wait_for_space(struct xe_exec_queue * q,u32 wqi_size) wq_wait_for_space() argument 1032 wq_noop_append(struct xe_exec_queue * q) wq_noop_append() argument 1052 wq_item_append(struct xe_exec_queue * q) wq_item_append() argument 1097 submit_exec_queue(struct xe_exec_queue * q,struct xe_sched_job * job) submit_exec_queue() argument 1161 struct xe_exec_queue *q = job->q; guc_exec_queue_run_job() local 1215 MAKE_SCHED_CONTEXT_ACTION(q,enable_disable) global() argument 1223 disable_scheduling_deregister(struct xe_guc * guc,struct xe_exec_queue * q) disable_scheduling_deregister() argument 1276 struct xe_exec_queue *q; xe_guc_submit_wedge() local 1330 check_timeout(struct xe_exec_queue * q,struct xe_sched_job * job) check_timeout() argument 1385 enable_scheduling(struct xe_exec_queue * q) enable_scheduling() argument 1418 disable_scheduling(struct xe_exec_queue * q,bool immediate) disable_scheduling() argument 1445 struct xe_exec_queue *q = job->q; guc_exec_queue_timedout_job() local 1635 guc_exec_queue_fini(struct xe_exec_queue * q) guc_exec_queue_fini() argument 1655 struct xe_exec_queue *q = ge->q; __guc_exec_queue_destroy_async() local 1675 guc_exec_queue_destroy_async(struct xe_exec_queue * q) guc_exec_queue_destroy_async() argument 1689 __guc_exec_queue_destroy(struct xe_guc * guc,struct xe_exec_queue * q) __guc_exec_queue_destroy() argument 1703 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_cleanup() local 1725 guc_exec_queue_allowed_to_change_state(struct xe_exec_queue * q) guc_exec_queue_allowed_to_change_state() argument 1732 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_set_sched_props() local 1740 __suspend_fence_signal(struct xe_exec_queue * q) __suspend_fence_signal() argument 1762 suspend_fence_signal(struct xe_exec_queue * q) suspend_fence_signal() argument 1775 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_suspend() local 1806 sched_context(struct xe_exec_queue * q) sched_context() argument 1828 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_resume() local 1854 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_set_multi_queue_priority() local 1923 guc_exec_queue_init(struct xe_exec_queue * q) guc_exec_queue_init() argument 2018 guc_exec_queue_kill(struct xe_exec_queue * q) guc_exec_queue_kill() argument 2026 guc_exec_queue_add_msg(struct xe_exec_queue * q,struct xe_sched_msg * msg,u32 opcode) guc_exec_queue_add_msg() argument 2044 guc_exec_queue_try_add_msg_head(struct xe_exec_queue * q,struct xe_sched_msg * msg,u32 opcode) guc_exec_queue_try_add_msg_head() argument 2054 guc_exec_queue_try_add_msg(struct xe_exec_queue * q,struct xe_sched_msg * msg,u32 opcode) guc_exec_queue_try_add_msg() argument 2069 guc_exec_queue_destroy(struct xe_exec_queue * q) guc_exec_queue_destroy() argument 2079 guc_exec_queue_set_priority(struct xe_exec_queue * q,enum xe_exec_queue_priority priority) guc_exec_queue_set_priority() argument 2098 guc_exec_queue_set_timeslice(struct xe_exec_queue * q,u32 timeslice_us) guc_exec_queue_set_timeslice() argument 2116 guc_exec_queue_set_preempt_timeout(struct xe_exec_queue * q,u32 preempt_timeout_us) guc_exec_queue_set_preempt_timeout() argument 2135 guc_exec_queue_set_multi_queue_priority(struct xe_exec_queue * q,enum xe_multi_queue_priority priority) guc_exec_queue_set_multi_queue_priority() argument 2156 guc_exec_queue_suspend(struct xe_exec_queue * q) guc_exec_queue_suspend() argument 2172 guc_exec_queue_suspend_wait(struct xe_exec_queue * q) guc_exec_queue_suspend_wait() argument 2215 guc_exec_queue_resume(struct xe_exec_queue * q) guc_exec_queue_resume() argument 2228 guc_exec_queue_reset_status(struct xe_exec_queue * q) guc_exec_queue_reset_status() argument 2258 guc_exec_queue_stop(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_stop() argument 2347 struct xe_exec_queue *q; xe_guc_submit_stop() local 2372 guc_exec_queue_revert_pending_state_change(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_revert_pending_state_change() argument 2443 guc_exec_queue_pause(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_pause() argument 2490 struct xe_exec_queue *q; xe_guc_submit_pause() local 2505 struct xe_exec_queue *q; xe_guc_submit_pause_vf() local 2522 guc_exec_queue_start(struct xe_exec_queue * q) guc_exec_queue_start() argument 2555 struct xe_exec_queue *q; xe_guc_submit_start() local 2577 guc_exec_queue_unpause_prepare(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_unpause_prepare() argument 2606 struct xe_exec_queue *q; xe_guc_submit_unpause_prepare_vf() local 2623 guc_exec_queue_replay_pending_state_change(struct xe_exec_queue * q) guc_exec_queue_replay_pending_state_change() argument 2661 guc_exec_queue_unpause(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_unpause() argument 2682 struct xe_exec_queue *q; xe_guc_submit_unpause() local 2697 struct xe_exec_queue *q; xe_guc_submit_unpause_vf() local 2723 struct xe_exec_queue *q; xe_guc_submit_pause_abort() local 2744 struct xe_exec_queue *q; g2h_exec_queue_lookup() local 2763 deregister_exec_queue(struct xe_guc * guc,struct xe_exec_queue * q) deregister_exec_queue() argument 2784 handle_sched_done(struct xe_guc * guc,struct xe_exec_queue * q,u32 runnable_state) handle_sched_done() argument 2829 handle_multi_queue_secondary_sched_done(struct xe_guc * guc,struct xe_exec_queue * q,u32 runnable_state) handle_multi_queue_secondary_sched_done() argument 2840 struct xe_exec_queue *q; xe_guc_sched_done_handler() local 2867 handle_deregister_done(struct xe_guc * guc,struct xe_exec_queue * q) handle_deregister_done() argument 2877 struct xe_exec_queue *q; xe_guc_deregister_done_handler() local 2905 struct xe_exec_queue *q; xe_guc_exec_queue_reset_handler() local 2966 struct xe_exec_queue *q; xe_guc_exec_queue_memory_cat_error_handler() local 3040 struct xe_exec_queue *q; xe_guc_exec_queue_cgp_context_error_handler() local 3078 struct xe_exec_queue *q; xe_guc_exec_queue_cgp_sync_done_handler() local 3103 guc_exec_queue_wq_snapshot_capture(struct xe_exec_queue * q,struct xe_guc_submit_exec_queue_snapshot * snapshot) guc_exec_queue_wq_snapshot_capture() argument 3161 xe_guc_exec_queue_snapshot_capture(struct xe_exec_queue * q) xe_guc_exec_queue_snapshot_capture() argument 3294 guc_exec_queue_print(struct xe_exec_queue * q,struct drm_printer * p) guc_exec_queue_print() argument 3315 xe_guc_register_vf_exec_queue(struct xe_exec_queue * q,int ctx_type) xe_guc_register_vf_exec_queue() argument 3340 struct xe_exec_queue *q; xe_guc_submit_print() local 3361 struct xe_exec_queue *q; xe_guc_has_registered_mlrc_queues() local 3383 struct xe_exec_queue *q; xe_guc_contexts_hwsp_rebase() local [all...] |
| H A D | xe_exec_queue.c | 110 static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q, 113 static void xe_exec_queue_group_cleanup(struct xe_exec_queue *q) in xe_exec_queue_group_cleanup() argument 115 struct xe_exec_queue_group *group = q->multi_queue.group; in xe_exec_queue_group_cleanup() 119 if (xe_exec_queue_is_multi_queue_secondary(q)) { in xe_exec_queue_group_cleanup() 124 xe_exec_queue_put(xe_exec_queue_multi_queue_primary(q)); in xe_exec_queue_group_cleanup() 141 static void __xe_exec_queue_free(struct xe_exec_queue *q) in __xe_exec_queue_free() argument 146 if (q->tlb_inval[i].dep_scheduler) in __xe_exec_queue_free() 147 xe_dep_scheduler_fini(q->tlb_inval[i].dep_scheduler); in __xe_exec_queue_free() 149 if (xe_exec_queue_uses_pxp(q)) in __xe_exec_queue_free() 150 xe_pxp_exec_queue_remove(gt_to_xe(q in __xe_exec_queue_free() 165 alloc_dep_schedulers(struct xe_device * xe,struct xe_exec_queue * q) alloc_dep_schedulers() argument 204 struct xe_exec_queue *q; __xe_exec_queue_alloc() local 269 __xe_exec_queue_fini(struct xe_exec_queue * q) __xe_exec_queue_fini() argument 279 __xe_exec_queue_init(struct xe_exec_queue * q,u32 exec_queue_flags) __xe_exec_queue_init() argument 342 struct xe_exec_queue *q; xe_exec_queue_create() local 429 struct xe_exec_queue *q; xe_exec_queue_create_bind() local 473 struct xe_exec_queue *q = container_of(ref, struct xe_exec_queue, refcount); xe_exec_queue_destroy() local 503 xe_exec_queue_fini(struct xe_exec_queue * q) xe_exec_queue_fini() argument 517 xe_exec_queue_assign_name(struct xe_exec_queue * q,u32 instance) xe_exec_queue_assign_name() argument 545 struct xe_exec_queue *q; xe_exec_queue_lookup() local 563 exec_queue_set_priority(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_priority() argument 628 exec_queue_set_timeslice(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_timeslice() argument 645 exec_queue_set_pxp_type(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_pxp_type() argument 661 exec_queue_set_hang_replay_state(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_hang_replay_state() argument 677 xe_exec_queue_group_init(struct xe_device * xe,struct xe_exec_queue * q) xe_exec_queue_group_init() argument 719 xe_exec_queue_supports_multi_queue(struct xe_exec_queue * q) xe_exec_queue_supports_multi_queue() argument 724 xe_exec_queue_group_validate(struct xe_device * xe,struct xe_exec_queue * q,u32 primary_id) xe_exec_queue_group_validate() argument 757 xe_exec_queue_group_add(struct xe_device * xe,struct xe_exec_queue * q) xe_exec_queue_group_add() argument 783 xe_exec_queue_group_delete(struct xe_device * xe,struct xe_exec_queue * q) xe_exec_queue_group_delete() argument 795 exec_queue_set_multi_group(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_multi_group() argument 830 exec_queue_set_multi_queue_priority(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_multi_queue_priority() argument 868 struct xe_exec_queue *q; xe_exec_queue_set_property_ioctl() local 897 exec_queue_user_ext_check(struct xe_exec_queue * q,u64 properties) exec_queue_user_ext_check() argument 913 exec_queue_user_ext_check_final(struct xe_exec_queue * q,u64 properties) exec_queue_user_ext_check_final() argument 924 exec_queue_user_ext_set_property(struct xe_device * xe,struct xe_exec_queue * q,u64 extension,u64 * properties) exec_queue_user_ext_set_property() argument 968 __exec_queue_user_extensions(struct xe_device * xe,struct xe_exec_queue * q,u64 extensions,int ext_number,u64 * properties) __exec_queue_user_extensions() argument 1001 exec_queue_user_extensions(struct xe_device * xe,struct xe_exec_queue * q,u64 extensions) exec_queue_user_extensions() argument 1098 struct xe_exec_queue *q = NULL; xe_exec_queue_create_ioctl() local 1261 struct xe_exec_queue *q; xe_exec_queue_get_property_ioctl() local 1295 xe_exec_queue_lrc(struct xe_exec_queue * q) xe_exec_queue_lrc() argument 1306 xe_exec_queue_is_lr(struct xe_exec_queue * q) xe_exec_queue_is_lr() argument 1326 xe_exec_queue_is_idle(struct xe_exec_queue * q) xe_exec_queue_is_idle() argument 1352 xe_exec_queue_update_run_ticks(struct xe_exec_queue * q) xe_exec_queue_update_run_ticks() argument 1393 xe_exec_queue_kill(struct xe_exec_queue * q) xe_exec_queue_kill() argument 1413 struct xe_exec_queue *q; xe_exec_queue_destroy_ioctl() local 1439 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_lockdep_assert() argument 1457 xe_exec_queue_last_fence_put(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_put() argument 1470 xe_exec_queue_last_fence_put_unlocked(struct xe_exec_queue * q) xe_exec_queue_last_fence_put_unlocked() argument 1487 xe_exec_queue_last_fence_get(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get() argument 1514 xe_exec_queue_last_fence_get_for_resume(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get_for_resume() argument 1539 xe_exec_queue_last_fence_set(struct xe_exec_queue * q,struct xe_vm * vm,struct dma_fence * fence) xe_exec_queue_last_fence_set() argument 1555 xe_exec_queue_tlb_inval_last_fence_put(struct xe_exec_queue * q,struct xe_vm * vm,unsigned int type) xe_exec_queue_tlb_inval_last_fence_put() argument 1574 xe_exec_queue_tlb_inval_last_fence_put_unlocked(struct xe_exec_queue * q,unsigned int type) xe_exec_queue_tlb_inval_last_fence_put_unlocked() argument 1594 xe_exec_queue_tlb_inval_last_fence_get(struct xe_exec_queue * q,struct xe_vm * vm,unsigned int type) xe_exec_queue_tlb_inval_last_fence_get() argument 1627 xe_exec_queue_tlb_inval_last_fence_set(struct xe_exec_queue * q,struct xe_vm * vm,struct dma_fence * fence,unsigned int type) xe_exec_queue_tlb_inval_last_fence_set() argument 1651 xe_exec_queue_contexts_hwsp_rebase(struct xe_exec_queue * q,void * scratch) xe_exec_queue_contexts_hwsp_rebase() argument [all...] |
| /linux/sound/core/seq/ |
| H A D | seq_queue.c | 50 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument 57 queue_list[i] = q; in queue_list_add() 58 q->queue = i; in queue_list_add() 68 struct snd_seq_queue *q; in queue_list_remove() local 71 q = queue_list[id]; in queue_list_remove() 72 if (q) { in queue_list_remove() 73 guard(spinlock)(&q->owner_lock); in queue_list_remove() 74 if (q->owner == client) { in queue_list_remove() 76 q->klocked = 1; in queue_list_remove() 79 return q; in queue_list_remove() [all …]
|
| /linux/sound/core/seq/oss/ |
| H A D | seq_oss_readq.c | 35 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local 37 q = kzalloc_obj(*q); in snd_seq_oss_readq_new() 38 if (!q) in snd_seq_oss_readq_new() 41 q->q = kzalloc_objs(union evrec, maxlen); in snd_seq_oss_readq_new() 42 if (!q->q) { in snd_seq_oss_readq_new() 43 kfree(q); in snd_seq_oss_readq_new() 47 q->maxlen = maxlen; in snd_seq_oss_readq_new() 48 q->qlen = 0; in snd_seq_oss_readq_new() 49 q->head = q->tail = 0; in snd_seq_oss_readq_new() 50 init_waitqueue_head(&q->midi_sleep); in snd_seq_oss_readq_new() [all …]
|
| H A D | seq_oss_writeq.c | 27 struct seq_oss_writeq *q; in snd_seq_oss_writeq_new() local 30 q = kzalloc_obj(*q); in snd_seq_oss_writeq_new() 31 if (!q) in snd_seq_oss_writeq_new() 33 q->dp = dp; in snd_seq_oss_writeq_new() 34 q->maxlen = maxlen; in snd_seq_oss_writeq_new() 35 spin_lock_init(&q->sync_lock); in snd_seq_oss_writeq_new() 36 q->sync_event_put = 0; in snd_seq_oss_writeq_new() 37 q->sync_time = 0; in snd_seq_oss_writeq_new() 38 init_waitqueue_head(&q->sync_sleep); in snd_seq_oss_writeq_new() 47 return q; in snd_seq_oss_writeq_new() [all …]
|
| /linux/net/sched/ |
| H A D | sch_choke.c | 75 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument 77 return (q->tail - q->head) & q->tab_mask; in choke_len() 81 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument 83 return q->flags & TC_RED_ECN; in use_ecn() 87 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument 89 return q->flags & TC_RED_HARDDROP; in use_harddrop() 93 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument 96 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes() 97 if (q->head == q->tail) in choke_zap_head_holes() 99 } while (q->tab[q->head] == NULL); in choke_zap_head_holes() [all …]
|
| H A D | sch_netem.c | 210 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument 212 struct clgstate *clg = &q->clg; in loss_4state() 213 u32 rnd = prandom_u32_state(&q->prng.prng_state); in loss_4state() 275 static bool loss_gilb_ell(struct netem_sched_data *q) in loss_gilb_ell() argument 277 struct clgstate *clg = &q->clg; in loss_gilb_ell() 278 struct rnd_state *s = &q->prng.prng_state; in loss_gilb_ell() 297 static bool loss_event(struct netem_sched_data *q) in loss_event() argument 299 switch (q->loss_model) { in loss_event() 302 return q->loss && q->loss >= get_crandom(&q->loss_cor, &q->prng); in loss_event() 310 return loss_4state(q); in loss_event() [all …]
|
| H A D | sch_sfq.c | 143 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument 146 return &q->slots[val].dep; in sfq_dep_head() 147 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head() 150 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument 153 return skb_get_hash_perturb(skb, &q->perturbation) & (q->divisor - 1); in sfq_hash() 159 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() local 166 TC_H_MIN(skb->priority) <= q->divisor) in sfq_classify() 169 fl = rcu_dereference_bh(q->filter_list); in sfq_classify() 171 return sfq_hash(q, sk in sfq_classify() 196 sfq_link(struct sfq_sched_data * q,sfq_index x) sfq_link() argument 212 sfq_unlink(q,x,n,p) global() argument 221 sfq_dec(struct sfq_sched_data * q,sfq_index x) sfq_dec() argument 234 sfq_inc(struct sfq_sched_data * q,sfq_index x) sfq_inc() argument 288 struct sfq_sched_data *q = qdisc_priv(sch); sfq_drop() local 325 sfq_prob_mark(const struct sfq_sched_data * q) sfq_prob_mark() argument 331 sfq_hard_mark(const struct sfq_sched_data * q) sfq_hard_mark() argument 336 sfq_headdrop(const struct sfq_sched_data * q) sfq_headdrop() argument 344 struct sfq_sched_data *q = qdisc_priv(sch); sfq_enqueue() local 478 struct sfq_sched_data *q = qdisc_priv(sch); sfq_dequeue() local 533 struct sfq_sched_data *q = qdisc_priv(sch); sfq_rehash() local 603 struct sfq_sched_data *q = timer_container_of(q, t, perturb_timer); sfq_perturbation() local 630 struct sfq_sched_data *q = qdisc_priv(sch); sfq_change() local 764 struct sfq_sched_data *q = qdisc_priv(sch); sfq_destroy() local 777 struct sfq_sched_data *q = qdisc_priv(sch); sfq_init() local 832 struct sfq_sched_data *q = qdisc_priv(sch); sfq_dump() local 883 sfq_unbind(struct Qdisc * q,unsigned long cl) sfq_unbind() argument 890 struct sfq_sched_data *q = qdisc_priv(sch); sfq_tcf_block() local 907 struct sfq_sched_data *q = qdisc_priv(sch); sfq_dump_class_stats() local 926 struct sfq_sched_data *q = qdisc_priv(sch); sfq_walk() local [all...] |
| H A D | sch_dualpi2.c | 147 static u64 head_enqueue_time(struct Qdisc *q) in head_enqueue_time() argument 149 struct sk_buff *skb = qdisc_peek_head(q); in head_enqueue_time() 170 static ktime_t next_pi2_timeout(struct dualpi2_sched_data *q) in next_pi2_timeout() argument 172 return ktime_add_ns(ktime_get_ns(), q->pi2_tupdate); in next_pi2_timeout() 185 static bool skb_apply_step(struct sk_buff *skb, struct dualpi2_sched_data *q) in skb_apply_step() argument 187 return skb_is_l4s(skb) && qdisc_qlen(q->l_queue) >= q->min_qlen_step; in skb_apply_step() 190 static bool dualpi2_mark(struct dualpi2_sched_data *q, struct sk_buff *skb) in dualpi2_mark() argument 193 q->ecn_mark++; in dualpi2_mark() 199 static void dualpi2_reset_c_protection(struct dualpi2_sched_data *q) in dualpi2_reset_c_protection() argument 209 dualpi2_calculate_c_protection(struct Qdisc * sch,struct dualpi2_sched_data * q,u32 wc) dualpi2_calculate_c_protection() argument 230 dualpi2_classic_marking(struct dualpi2_sched_data * q,struct sk_buff * skb,u32 prob,bool overload) dualpi2_classic_marking() argument 253 dualpi2_scalable_marking(struct dualpi2_sched_data * q,struct sk_buff * skb,u64 local_l_prob,u32 prob,bool overload) dualpi2_scalable_marking() argument 282 must_drop(struct Qdisc * sch,struct dualpi2_sched_data * q,struct sk_buff * skb) must_drop() argument 341 dualpi2_skb_classify(struct dualpi2_sched_data * q,struct sk_buff * skb) dualpi2_skb_classify() argument 388 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_enqueue_skb() local 444 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_qdisc_enqueue() local 517 dequeue_packet(struct Qdisc * sch,struct dualpi2_sched_data * q,int * credit_change,u64 now) dequeue_packet() argument 551 do_step_aqm(struct dualpi2_sched_data * q,struct sk_buff * skb,u64 now) do_step_aqm() argument 574 drop_and_retry(struct dualpi2_sched_data * q,struct sk_buff * skb,struct Qdisc * sch,enum skb_drop_reason reason) drop_and_retry() argument 585 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_qdisc_dequeue() local 626 get_queue_delays(struct dualpi2_sched_data * q,u64 * qdelay_c,u64 * qdelay_l) get_queue_delays() argument 641 struct dualpi2_sched_data *q = qdisc_priv(sch); calculate_probability() local 711 struct dualpi2_sched_data *q = timer_container_of(q, timer, pi2_timer); dualpi2_timer() local 762 struct dualpi2_sched_data *q; dualpi2_change() local 890 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_reset_default() local 916 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_init() local 947 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_dump() local 1023 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_dump_stats() local 1050 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_reset() local 1069 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_destroy() local 1094 dualpi2_unbind(struct Qdisc * q,unsigned long cl) dualpi2_unbind() argument 1101 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_tcf_block() local [all...] |
| H A D | sch_fq_pie.c | 75 static unsigned int fq_pie_hash(const struct fq_pie_sched_data *q, in fq_pie_hash() argument 78 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_pie_hash() 84 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_classify() local 91 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_pie_classify() 94 filter = rcu_dereference_bh(q->filter_list); in fq_pie_classify() 96 return fq_pie_hash(q, skb) + 1; in fq_pie_classify() 112 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_pie_classify() 134 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_qdisc_enqueue() local 152 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue() 155 memory_limited = q in fq_pie_qdisc_enqueue() 240 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_qdisc_dequeue() local 291 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_change() local 389 struct fq_pie_sched_data *q = timer_container_of(q, t, adapt_timer); fq_pie_timer() local 423 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_init() local 475 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_dump() local 511 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_dump_stats() local 536 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_reset() local 555 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_destroy() local [all...] |
| H A D | sch_red.c | 55 static inline int red_use_ecn(struct red_sched_data *q) in red_use_ecn() argument 57 return q->flags & TC_RED_ECN; in red_use_ecn() 60 static inline int red_use_harddrop(struct red_sched_data *q) in red_use_harddrop() argument 62 return q->flags & TC_RED_HARDDROP; in red_use_harddrop() 65 static int red_use_nodrop(struct red_sched_data *q) in red_use_nodrop() argument 67 return q->flags & TC_RED_NODROP; in red_use_nodrop() 74 struct red_sched_data *q = qdisc_priv(sch); in red_enqueue() local 75 struct Qdisc *child = q->qdisc; in red_enqueue() 79 q->vars.qavg = red_calc_qavg(&q in red_enqueue() 155 struct red_sched_data *q = qdisc_priv(sch); red_dequeue() local 172 struct red_sched_data *q = qdisc_priv(sch); red_peek() local 180 struct red_sched_data *q = qdisc_priv(sch); red_reset() local 188 struct red_sched_data *q = qdisc_priv(sch); red_offload() local 217 struct red_sched_data *q = qdisc_priv(sch); red_destroy() local 240 struct red_sched_data *q = qdisc_priv(sch); __red_change() local 324 struct red_sched_data *q = timer_container_of(q, t, adapt_timer); red_adaptative_timer() local 340 struct red_sched_data *q = qdisc_priv(sch); red_init() local 374 struct red_sched_data *q = qdisc_priv(sch); red_change() local 413 struct red_sched_data *q = qdisc_priv(sch); red_dump() local 450 struct red_sched_data *q = qdisc_priv(sch); red_dump_stats() local 476 struct red_sched_data *q = qdisc_priv(sch); red_dump_class() local 501 struct red_sched_data *q = qdisc_priv(sch); red_graft() local 514 struct red_sched_data *q = qdisc_priv(sch); red_leaf() local [all...] |
| H A D | sch_fq_codel.c | 70 static unsigned int fq_codel_hash(const struct fq_codel_sched_data *q, in fq_codel_hash() argument 73 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_codel_hash() 79 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_classify() local 86 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_codel_classify() 89 filter = rcu_dereference_bh(q->filter_list); in fq_codel_classify() 91 return fq_codel_hash(q, skb) + 1; in fq_codel_classify() 107 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_codel_classify() 140 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_drop() local 154 for (i = 0; i < q->flows_cnt; i++) { in fq_codel_drop() 155 if (q in fq_codel_drop() 188 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_enqueue() local 259 struct fq_codel_sched_data *q = qdisc_priv(sch); dequeue_func() local 284 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dequeue() local 336 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_reset() local 370 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_change() local 462 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_destroy() local 472 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_init() local 535 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dump() local 582 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dump_stats() local 623 fq_codel_unbind(struct Qdisc * q,unsigned long cl) fq_codel_unbind() argument 630 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_tcf_block() local 647 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dump_class_stats() local 693 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_walk() local [all...] |
| H A D | sch_fq.c | 79 /* Following field is only used for q->internal, 80 * because q->internal is not hashed in fq_root[] 93 struct rb_node rate_node; /* anchor in q->delayed tree */ 197 static void fq_flow_add_tail(struct fq_sched_data *q, struct fq_flow *flow, in fq_flow_add_tail() argument 200 struct fq_perband_flows *pband = &q->band_flows[flow->band]; in fq_flow_add_tail() 213 static void fq_flow_unset_throttled(struct fq_sched_data *q, struct fq_flow *f) in fq_flow_unset_throttled() argument 215 rb_erase(&f->rate_node, &q->delayed); in fq_flow_unset_throttled() 216 q->throttled_flows--; in fq_flow_unset_throttled() 217 fq_flow_add_tail(q, f, OLD_FLOW); in fq_flow_unset_throttled() 220 static void fq_flow_set_throttled(struct fq_sched_data *q, struc argument 256 fq_gc(struct fq_sched_data * q,struct rb_root * root,struct sock * sk) fq_gc() argument 313 const struct fq_sched_data *q = qdisc_priv(sch); fq_fastpath_check() local 355 struct fq_sched_data *q = qdisc_priv(sch); fq_classify() local 537 fq_packet_beyond_horizon(const struct sk_buff * skb,const struct fq_sched_data * q,u64 now) fq_packet_beyond_horizon() argument 547 struct fq_sched_data *q = qdisc_priv(sch); fq_enqueue() local 609 fq_check_throttled(struct fq_sched_data * q,u64 now) fq_check_throttled() argument 652 struct fq_sched_data *q = qdisc_priv(sch); fq_dequeue() local 802 struct fq_sched_data *q = qdisc_priv(sch); fq_reset() local 838 fq_rehash(struct fq_sched_data * q,struct rb_root * old_array,u32 old_log,struct rb_root * new_array,u32 new_log) fq_rehash() argument 890 struct fq_sched_data *q = qdisc_priv(sch); fq_resize() local 974 fq_load_weights(struct fq_sched_data * q,const struct nlattr * attr,struct netlink_ext_ack * extack) fq_load_weights() argument 993 fq_load_priomap(struct fq_sched_data * q,const struct nlattr * attr,struct netlink_ext_ack * extack) fq_load_priomap() argument 1019 struct fq_sched_data *q = qdisc_priv(sch); fq_change() local 1159 struct fq_sched_data *q = qdisc_priv(sch); fq_destroy() local 1169 struct fq_sched_data *q = qdisc_priv(sch); fq_init() local 1214 struct fq_sched_data *q = qdisc_priv(sch); fq_dump() local 1287 struct fq_sched_data *q = qdisc_priv(sch); fq_dump_stats() local [all...] |
| H A D | sch_skbprio.c | 40 static u16 calc_new_high_prio(const struct skbprio_sched_data *q) in calc_new_high_prio() argument 44 for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) { in calc_new_high_prio() 45 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_high_prio() 53 static u16 calc_new_low_prio(const struct skbprio_sched_data *q) in calc_new_low_prio() argument 57 for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) { in calc_new_low_prio() 58 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_low_prio() 72 struct skbprio_sched_data *q = qdisc_priv(sch); in skbprio_enqueue() local 81 qdisc = &q->qdiscs[prio]; in skbprio_enqueue() 84 if (sch->q.qlen < READ_ONCE(sch->limit)) { in skbprio_enqueue() 87 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() [all …]
|
| /linux/drivers/media/common/videobuf2/ |
| H A D | videobuf2-core.c | 47 #define dprintk(q, level, fmt, arg...) \ argument 50 pr_info("[%s] %s: " fmt, (q)->name, __func__, \ 103 #define log_qop(q, op) \ argument 104 dprintk(q, 2, "call_qop(%s)%s\n", #op, \ 105 (q)->ops->op ? "" : " (nop)") 107 #define call_qop(q, op, args...) \ argument 111 log_qop(q, op); \ 112 err = (q)->ops->op ? (q)->ops->op(args) : 0; \ 114 (q)->cnt_ ## op++; \ 118 #define call_void_qop(q, op, args...) \ argument [all …]
|
| /linux/net/xdp/ |
| H A D | xsk_queue.h | 125 static inline void __xskq_cons_read_addr_unchecked(struct xsk_queue *q, u32 cached_cons, u64 *addr) in __xskq_cons_read_addr_unchecked() 127 struct xdp_umem_ring *ring = (struct xdp_umem_ring *)q->ring; 128 u32 idx = cached_cons & q->ring_mask; in xskq_cons_read_addr_unchecked() argument 133 static inline bool xskq_cons_read_addr_unchecked(struct xsk_queue *q, u64 *addr) in xskq_cons_read_addr_unchecked() 135 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked() 136 __xskq_cons_read_addr_unchecked(q, q->cached_cons, addr); in xskq_cons_read_addr_unchecked() 222 static inline bool xskq_has_descs(struct xsk_queue *q) in xskq_cons_is_valid_desc() argument 224 return q in xskq_cons_is_valid_desc() 120 __xskq_cons_read_addr_unchecked(struct xsk_queue * q,u32 cached_cons,u64 * addr) __xskq_cons_read_addr_unchecked() argument 217 xskq_has_descs(struct xsk_queue * q) xskq_has_descs() argument 233 xskq_cons_read_desc(struct xsk_queue * q,struct xdp_desc * desc,struct xsk_buff_pool * pool) xskq_cons_read_desc() argument 249 xskq_cons_release_n(struct xsk_queue * q,u32 cnt) xskq_cons_release_n() argument 254 parse_desc(struct xsk_queue * q,struct xsk_buff_pool * pool,struct xdp_desc * desc,struct parsed_desc * parsed) parse_desc() argument 262 xskq_cons_read_desc_batch(struct xsk_queue * q,struct xsk_buff_pool * pool,u32 max) xskq_cons_read_desc_batch() argument 304 __xskq_cons_release(struct xsk_queue * q) __xskq_cons_release() argument 309 __xskq_cons_peek(struct xsk_queue * q) __xskq_cons_peek() argument 315 xskq_cons_get_entries(struct xsk_queue * q) xskq_cons_get_entries() argument 321 xskq_cons_nb_entries(struct xsk_queue * q,u32 max) xskq_cons_nb_entries() argument 334 xskq_cons_peek_addr_unchecked(struct xsk_queue * q,u64 * addr) xskq_cons_peek_addr_unchecked() argument 341 xskq_cons_peek_desc(struct xsk_queue * q,struct xdp_desc * desc,struct xsk_buff_pool * pool) xskq_cons_peek_desc() argument 354 xskq_cons_release(struct xsk_queue * q) xskq_cons_release() argument 359 xskq_cons_cancel_n(struct xsk_queue * q,u32 cnt) xskq_cons_cancel_n() argument 364 xskq_cons_present_entries(struct xsk_queue * q) xskq_cons_present_entries() argument 372 xskq_get_prod(struct xsk_queue * q) xskq_get_prod() argument 377 xskq_prod_nb_free(struct xsk_queue * q,u32 max) xskq_prod_nb_free() argument 391 xskq_prod_is_full(struct xsk_queue * q) xskq_prod_is_full() argument 396 xskq_prod_cancel_n(struct xsk_queue * q,u32 cnt) xskq_prod_cancel_n() argument 401 xskq_prod_reserve(struct xsk_queue * q) xskq_prod_reserve() argument 411 xskq_prod_reserve_addr(struct xsk_queue * q,u64 addr) xskq_prod_reserve_addr() argument 423 xskq_prod_write_addr(struct xsk_queue * q,u32 idx,u64 addr) xskq_prod_write_addr() argument 430 xskq_prod_write_addr_batch(struct xsk_queue * q,struct xdp_desc * descs,u32 nb_entries) xskq_prod_write_addr_batch() argument 443 xskq_prod_reserve_desc(struct xsk_queue * q,u64 addr,u32 len,u32 flags) xskq_prod_reserve_desc() argument 461 __xskq_prod_submit(struct xsk_queue * q,u32 idx) __xskq_prod_submit() argument 466 xskq_prod_submit(struct xsk_queue * q) xskq_prod_submit() argument 471 xskq_prod_submit_n(struct xsk_queue * q,u32 nb_entries) xskq_prod_submit_n() argument 476 xskq_prod_is_empty(struct xsk_queue * q) xskq_prod_is_empty() argument 484 xskq_nb_invalid_descs(struct xsk_queue * q) xskq_nb_invalid_descs() argument 489 xskq_nb_queue_empty_descs(struct xsk_queue * q) xskq_nb_queue_empty_descs() argument [all...] |
| /linux/block/ |
| H A D | blk-pm.c | 29 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument 31 q->dev = dev; in blk_pm_runtime_init() 32 q->rpm_status = RPM_ACTIVE; in blk_pm_runtime_init() 33 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init() 34 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init() 59 int blk_pre_runtime_suspend(struct request_queue *q) in blk_pre_runtime_suspend() argument 63 if (!q->dev) in blk_pre_runtime_suspend() 66 WARN_ON_ONCE(q->rpm_status != RPM_ACTIVE); in blk_pre_runtime_suspend() 68 spin_lock_irq(&q->queue_lock); in blk_pre_runtime_suspend() 69 q->rpm_status = RPM_SUSPENDING; in blk_pre_runtime_suspend() [all …]
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_queue.h | 83 void rxe_queue_reset(struct rxe_queue *q); 88 int rxe_queue_resize(struct rxe_queue *q, unsigned int *num_elem_p, 95 static inline u32 queue_next_index(struct rxe_queue *q, int index) in queue_next_index() argument 97 return (index + 1) & q->index_mask; in queue_next_index() 100 static inline u32 queue_get_producer(const struct rxe_queue *q, in queue_get_producer() argument 108 prod = smp_load_acquire(&q->buf->producer_index); in queue_get_producer() 112 prod = q->index; in queue_get_producer() 116 prod = q->buf->producer_index; in queue_get_producer() 120 prod = smp_load_acquire(&q->buf->producer_index); in queue_get_producer() 127 static inline u32 queue_get_consumer(const struct rxe_queue *q, in queue_get_consumer() argument [all …]
|
| /linux/drivers/net/wireless/broadcom/b43/ |
| H A D | pio.c | 24 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument 37 cookie = (((u16)q->index + 1) << 12); in generate_cookie() 49 struct b43_pio_txqueue *q = NULL; in parse_cookie() local 54 q = pio->tx_queue_AC_BK; in parse_cookie() 57 q = pio->tx_queue_AC_BE; in parse_cookie() 60 q = pio->tx_queue_AC_VI; in parse_cookie() 63 q = pio->tx_queue_AC_VO; in parse_cookie() 66 q = pio->tx_queue_mcast; in parse_cookie() 69 if (B43_WARN_ON(!q)) in parse_cookie() 72 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie() [all …]
|
| /linux/drivers/s390/cio/ |
| H A D | qdio_main.c | 116 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument 119 int tmp_count = count, tmp_start = start, nr = q->nr; in qdio_do_eqbs() 122 qperf_inc(q, eqbs); in qdio_do_eqbs() 124 if (!q->is_input_q) in qdio_do_eqbs() 125 nr += q->irq_ptr->nr_input_qs; in qdio_do_eqbs() 127 ccq = do_eqbs(q->irq_ptr->sch_token, state, nr, &tmp_start, &tmp_count, in qdio_do_eqbs() 137 qperf_inc(q, eqbs_partial); in qdio_do_eqbs() 138 DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "EQBS part:%02x", in qdio_do_eqbs() 143 DBF_DEV_EVENT(DBF_WARN, q->irq_ptr, "EQBS again:%2d", ccq); in qdio_do_eqbs() 146 DBF_ERROR("%4x ccq:%3d", SCH_NO(q), ccq); in qdio_do_eqbs() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | sge.c | 169 static inline struct sge_qset *rspq_to_qset(const struct sge_rspq *q) in rspq_to_qset() argument 171 return container_of(q, struct sge_qset, rspq); in rspq_to_qset() 174 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument 176 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset() 189 const struct sge_rspq *q, unsigned int credits) in refill_rspq() argument 193 V_RSPQ(q->cntxt_id) | V_CREDITS(credits)); in refill_rspq() 233 static inline void unmap_skb(struct sk_buff *skb, struct sge_txq *q, in unmap_skb() argument 237 struct tx_sw_desc *d = &q->sdesc[cidx]; in unmap_skb() 240 sgp = (struct sg_ent *)&q->desc[cidx].flit[d->sflit]; in unmap_skb() 266 d = cidx + 1 == q->size ? q->sdesc : d + 1; in unmap_skb() [all …]
|