Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 1141) sorted by relevance

12345678910>>...46

/linux/lib/crypto/
H A Dgf128mul.c57 #define gf128mul_dat(q) { \ argument
58 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\
59 q(
[all...]
/linux/Documentation/networking/
H A Dtls-offload-layers.svg1q-0.609375 -0.359375 -1.203125 -0.359375q-0.546875 0 -0.96875 0.328125q-0.421875 0.328125 -0.60937…
/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_rx.c50 static void cache_offer(struct funeth_rxq *q, const struct funeth_rxbuf *buf) in cache_offer() argument
52 struct funeth_rx_cache *c = &q->cache; in cache_offer()
58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_offer()
67 static bool cache_get(struct funeth_rxq *q, struct funeth_rxbuf *rb) in cache_get() argument
69 struct funeth_rx_cache *c = &q->cache; in cache_get()
77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, in cache_get()
88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_get()
98 static int funeth_alloc_page(struct funeth_rxq *q, struct funeth_rxbuf *rb, in funeth_alloc_page() argument
103 if (cache_get(q, rb)) in funeth_alloc_page()
110 rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE, in funeth_alloc_page()
[all …]
H A Dfuneth_tx.c56 static void *txq_end(const struct funeth_txq *q) in txq_end() argument
58 return (void *)q->hw_wb; in txq_end()
64 static unsigned int txq_to_end(const struct funeth_txq *q, void *p) in txq_to_end() argument
66 return txq_end(q) - p; in txq_to_end()
78 static struct fun_dataop_gl *fun_write_gl(const struct funeth_txq *q, in fun_write_gl() argument
90 i < ngle && txq_to_end(q, gle); i++, gle++) in fun_write_gl()
93 if (txq_to_end(q, gle) == 0) { in fun_write_gl()
94 gle = (struct fun_dataop_gl *)q->desc; in fun_write_gl()
107 static struct sk_buff *fun_tls_tx(struct sk_buff *skb, struct funeth_txq *q, in fun_tls_tx() argument
132 FUN_QSTAT_INC(q, tx_tls_fallback); in fun_tls_tx()
[all …]
/linux/drivers/gpu/drm/xe/
H A Dxe_guc_submit.c53 exec_queue_to_guc(struct xe_exec_queue *q)
55 return &q->gt->uc.guc; in exec_queue_to_guc()
76 static bool exec_queue_registered(struct xe_exec_queue *q)
78 return atomic_read(&q->guc->state) & EXEC_QUEUE_STATE_REGISTERED; in exec_queue_registered()
81 static void set_exec_queue_registered(struct xe_exec_queue *q)
83 atomic_or(EXEC_QUEUE_STATE_REGISTERED, &q->guc->state); in set_exec_queue_registered()
86 static void clear_exec_queue_registered(struct xe_exec_queue *q)
88 atomic_and(~EXEC_QUEUE_STATE_REGISTERED, &q->guc->state); in clear_exec_queue_registered()
91 static bool exec_queue_enabled(struct xe_exec_queue *q)
93 return atomic_read(&q in exec_queue_enabled()
54 exec_queue_to_guc(struct xe_exec_queue * q) exec_queue_to_guc() argument
77 exec_queue_registered(struct xe_exec_queue * q) exec_queue_registered() argument
82 set_exec_queue_registered(struct xe_exec_queue * q) set_exec_queue_registered() argument
87 clear_exec_queue_registered(struct xe_exec_queue * q) clear_exec_queue_registered() argument
92 exec_queue_enabled(struct xe_exec_queue * q) exec_queue_enabled() argument
97 set_exec_queue_enabled(struct xe_exec_queue * q) set_exec_queue_enabled() argument
102 clear_exec_queue_enabled(struct xe_exec_queue * q) clear_exec_queue_enabled() argument
107 exec_queue_pending_enable(struct xe_exec_queue * q) exec_queue_pending_enable() argument
112 set_exec_queue_pending_enable(struct xe_exec_queue * q) set_exec_queue_pending_enable() argument
117 clear_exec_queue_pending_enable(struct xe_exec_queue * q) clear_exec_queue_pending_enable() argument
122 exec_queue_pending_disable(struct xe_exec_queue * q) exec_queue_pending_disable() argument
127 set_exec_queue_pending_disable(struct xe_exec_queue * q) set_exec_queue_pending_disable() argument
132 clear_exec_queue_pending_disable(struct xe_exec_queue * q) clear_exec_queue_pending_disable() argument
137 exec_queue_destroyed(struct xe_exec_queue * q) exec_queue_destroyed() argument
142 set_exec_queue_destroyed(struct xe_exec_queue * q) set_exec_queue_destroyed() argument
147 clear_exec_queue_destroyed(struct xe_exec_queue * q) clear_exec_queue_destroyed() argument
152 exec_queue_banned(struct xe_exec_queue * q) exec_queue_banned() argument
157 set_exec_queue_banned(struct xe_exec_queue * q) set_exec_queue_banned() argument
162 exec_queue_suspended(struct xe_exec_queue * q) exec_queue_suspended() argument
167 set_exec_queue_suspended(struct xe_exec_queue * q) set_exec_queue_suspended() argument
172 clear_exec_queue_suspended(struct xe_exec_queue * q) clear_exec_queue_suspended() argument
177 exec_queue_reset(struct xe_exec_queue * q) exec_queue_reset() argument
182 set_exec_queue_reset(struct xe_exec_queue * q) set_exec_queue_reset() argument
187 exec_queue_killed(struct xe_exec_queue * q) exec_queue_killed() argument
192 set_exec_queue_killed(struct xe_exec_queue * q) set_exec_queue_killed() argument
197 exec_queue_wedged(struct xe_exec_queue * q) exec_queue_wedged() argument
202 set_exec_queue_wedged(struct xe_exec_queue * q) set_exec_queue_wedged() argument
207 exec_queue_pending_resume(struct xe_exec_queue * q) exec_queue_pending_resume() argument
212 set_exec_queue_pending_resume(struct xe_exec_queue * q) set_exec_queue_pending_resume() argument
217 clear_exec_queue_pending_resume(struct xe_exec_queue * q) clear_exec_queue_pending_resume() argument
222 exec_queue_idle_skip_suspend(struct xe_exec_queue * q) exec_queue_idle_skip_suspend() argument
227 set_exec_queue_idle_skip_suspend(struct xe_exec_queue * q) set_exec_queue_idle_skip_suspend() argument
232 clear_exec_queue_idle_skip_suspend(struct xe_exec_queue * q) clear_exec_queue_idle_skip_suspend() argument
237 exec_queue_killed_or_banned_or_wedged(struct xe_exec_queue * q) exec_queue_killed_or_banned_or_wedged() argument
278 struct xe_exec_queue *q; guc_submit_wedged_fini() local
416 __release_guc_id(struct xe_guc * guc,struct xe_exec_queue * q,u32 xa_count) __release_guc_id() argument
432 alloc_guc_id(struct xe_guc * guc,struct xe_exec_queue * q) alloc_guc_id() argument
468 release_guc_id(struct xe_guc * guc,struct xe_exec_queue * q) release_guc_id() argument
524 init_policies(struct xe_guc * guc,struct xe_exec_queue * q) init_policies() argument
549 set_min_preemption_timeout(struct xe_guc * guc,struct xe_exec_queue * q) set_min_preemption_timeout() argument
567 xe_guc_exec_queue_trigger_cleanup(struct xe_exec_queue * q) xe_guc_exec_queue_trigger_cleanup() argument
578 xe_guc_exec_queue_group_trigger_cleanup(struct xe_exec_queue * q) xe_guc_exec_queue_group_trigger_cleanup() argument
597 xe_guc_exec_queue_reset_trigger_cleanup(struct xe_exec_queue * q) xe_guc_exec_queue_reset_trigger_cleanup() argument
625 set_exec_queue_group_banned(struct xe_exec_queue * q) set_exec_queue_group_banned() argument
726 xe_guc_exec_queue_group_cgp_update(struct xe_device * xe,struct xe_exec_queue * q) xe_guc_exec_queue_group_cgp_update() argument
755 xe_guc_exec_queue_group_cgp_sync(struct xe_guc * guc,struct xe_exec_queue * q,const u32 * action,u32 len) xe_guc_exec_queue_group_cgp_sync() argument
791 __register_exec_queue_group(struct xe_guc * guc,struct xe_exec_queue * q,struct guc_ctxt_registration_info * info) __register_exec_queue_group() argument
819 xe_guc_exec_queue_group_add(struct xe_guc * guc,struct xe_exec_queue * q) xe_guc_exec_queue_group_add() argument
842 __register_mlrc_exec_queue(struct xe_guc * guc,struct xe_exec_queue * q,struct guc_ctxt_registration_info * info) __register_mlrc_exec_queue() argument
917 register_exec_queue(struct xe_exec_queue * q,int ctx_type) register_exec_queue() argument
979 wq_space_until_wrap(struct xe_exec_queue * q) wq_space_until_wrap() argument
1002 wq_wait_for_space(struct xe_exec_queue * q,u32 wqi_size) wq_wait_for_space() argument
1032 wq_noop_append(struct xe_exec_queue * q) wq_noop_append() argument
1052 wq_item_append(struct xe_exec_queue * q) wq_item_append() argument
1097 submit_exec_queue(struct xe_exec_queue * q,struct xe_sched_job * job) submit_exec_queue() argument
1161 struct xe_exec_queue *q = job->q; guc_exec_queue_run_job() local
1215 MAKE_SCHED_CONTEXT_ACTION(q,enable_disable) global() argument
1223 disable_scheduling_deregister(struct xe_guc * guc,struct xe_exec_queue * q) disable_scheduling_deregister() argument
1276 struct xe_exec_queue *q; xe_guc_submit_wedge() local
1330 check_timeout(struct xe_exec_queue * q,struct xe_sched_job * job) check_timeout() argument
1385 enable_scheduling(struct xe_exec_queue * q) enable_scheduling() argument
1418 disable_scheduling(struct xe_exec_queue * q,bool immediate) disable_scheduling() argument
1445 struct xe_exec_queue *q = job->q; guc_exec_queue_timedout_job() local
1635 guc_exec_queue_fini(struct xe_exec_queue * q) guc_exec_queue_fini() argument
1655 struct xe_exec_queue *q = ge->q; __guc_exec_queue_destroy_async() local
1675 guc_exec_queue_destroy_async(struct xe_exec_queue * q) guc_exec_queue_destroy_async() argument
1689 __guc_exec_queue_destroy(struct xe_guc * guc,struct xe_exec_queue * q) __guc_exec_queue_destroy() argument
1703 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_cleanup() local
1725 guc_exec_queue_allowed_to_change_state(struct xe_exec_queue * q) guc_exec_queue_allowed_to_change_state() argument
1732 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_set_sched_props() local
1740 __suspend_fence_signal(struct xe_exec_queue * q) __suspend_fence_signal() argument
1762 suspend_fence_signal(struct xe_exec_queue * q) suspend_fence_signal() argument
1775 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_suspend() local
1806 sched_context(struct xe_exec_queue * q) sched_context() argument
1828 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_resume() local
1854 struct xe_exec_queue *q = msg->private_data; __guc_exec_queue_process_msg_set_multi_queue_priority() local
1923 guc_exec_queue_init(struct xe_exec_queue * q) guc_exec_queue_init() argument
2018 guc_exec_queue_kill(struct xe_exec_queue * q) guc_exec_queue_kill() argument
2026 guc_exec_queue_add_msg(struct xe_exec_queue * q,struct xe_sched_msg * msg,u32 opcode) guc_exec_queue_add_msg() argument
2044 guc_exec_queue_try_add_msg_head(struct xe_exec_queue * q,struct xe_sched_msg * msg,u32 opcode) guc_exec_queue_try_add_msg_head() argument
2054 guc_exec_queue_try_add_msg(struct xe_exec_queue * q,struct xe_sched_msg * msg,u32 opcode) guc_exec_queue_try_add_msg() argument
2069 guc_exec_queue_destroy(struct xe_exec_queue * q) guc_exec_queue_destroy() argument
2079 guc_exec_queue_set_priority(struct xe_exec_queue * q,enum xe_exec_queue_priority priority) guc_exec_queue_set_priority() argument
2098 guc_exec_queue_set_timeslice(struct xe_exec_queue * q,u32 timeslice_us) guc_exec_queue_set_timeslice() argument
2116 guc_exec_queue_set_preempt_timeout(struct xe_exec_queue * q,u32 preempt_timeout_us) guc_exec_queue_set_preempt_timeout() argument
2135 guc_exec_queue_set_multi_queue_priority(struct xe_exec_queue * q,enum xe_multi_queue_priority priority) guc_exec_queue_set_multi_queue_priority() argument
2156 guc_exec_queue_suspend(struct xe_exec_queue * q) guc_exec_queue_suspend() argument
2172 guc_exec_queue_suspend_wait(struct xe_exec_queue * q) guc_exec_queue_suspend_wait() argument
2215 guc_exec_queue_resume(struct xe_exec_queue * q) guc_exec_queue_resume() argument
2228 guc_exec_queue_reset_status(struct xe_exec_queue * q) guc_exec_queue_reset_status() argument
2258 guc_exec_queue_stop(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_stop() argument
2347 struct xe_exec_queue *q; xe_guc_submit_stop() local
2372 guc_exec_queue_revert_pending_state_change(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_revert_pending_state_change() argument
2443 guc_exec_queue_pause(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_pause() argument
2490 struct xe_exec_queue *q; xe_guc_submit_pause() local
2505 struct xe_exec_queue *q; xe_guc_submit_pause_vf() local
2522 guc_exec_queue_start(struct xe_exec_queue * q) guc_exec_queue_start() argument
2555 struct xe_exec_queue *q; xe_guc_submit_start() local
2577 guc_exec_queue_unpause_prepare(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_unpause_prepare() argument
2606 struct xe_exec_queue *q; xe_guc_submit_unpause_prepare_vf() local
2623 guc_exec_queue_replay_pending_state_change(struct xe_exec_queue * q) guc_exec_queue_replay_pending_state_change() argument
2661 guc_exec_queue_unpause(struct xe_guc * guc,struct xe_exec_queue * q) guc_exec_queue_unpause() argument
2682 struct xe_exec_queue *q; xe_guc_submit_unpause() local
2697 struct xe_exec_queue *q; xe_guc_submit_unpause_vf() local
2723 struct xe_exec_queue *q; xe_guc_submit_pause_abort() local
2744 struct xe_exec_queue *q; g2h_exec_queue_lookup() local
2763 deregister_exec_queue(struct xe_guc * guc,struct xe_exec_queue * q) deregister_exec_queue() argument
2784 handle_sched_done(struct xe_guc * guc,struct xe_exec_queue * q,u32 runnable_state) handle_sched_done() argument
2829 handle_multi_queue_secondary_sched_done(struct xe_guc * guc,struct xe_exec_queue * q,u32 runnable_state) handle_multi_queue_secondary_sched_done() argument
2840 struct xe_exec_queue *q; xe_guc_sched_done_handler() local
2867 handle_deregister_done(struct xe_guc * guc,struct xe_exec_queue * q) handle_deregister_done() argument
2877 struct xe_exec_queue *q; xe_guc_deregister_done_handler() local
2905 struct xe_exec_queue *q; xe_guc_exec_queue_reset_handler() local
2966 struct xe_exec_queue *q; xe_guc_exec_queue_memory_cat_error_handler() local
3040 struct xe_exec_queue *q; xe_guc_exec_queue_cgp_context_error_handler() local
3078 struct xe_exec_queue *q; xe_guc_exec_queue_cgp_sync_done_handler() local
3103 guc_exec_queue_wq_snapshot_capture(struct xe_exec_queue * q,struct xe_guc_submit_exec_queue_snapshot * snapshot) guc_exec_queue_wq_snapshot_capture() argument
3161 xe_guc_exec_queue_snapshot_capture(struct xe_exec_queue * q) xe_guc_exec_queue_snapshot_capture() argument
3294 guc_exec_queue_print(struct xe_exec_queue * q,struct drm_printer * p) guc_exec_queue_print() argument
3315 xe_guc_register_vf_exec_queue(struct xe_exec_queue * q,int ctx_type) xe_guc_register_vf_exec_queue() argument
3340 struct xe_exec_queue *q; xe_guc_submit_print() local
3361 struct xe_exec_queue *q; xe_guc_has_registered_mlrc_queues() local
3383 struct xe_exec_queue *q; xe_guc_contexts_hwsp_rebase() local
[all...]
H A Dxe_exec_queue.c110 static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q,
113 static void xe_exec_queue_group_cleanup(struct xe_exec_queue *q) in xe_exec_queue_group_cleanup() argument
115 struct xe_exec_queue_group *group = q->multi_queue.group; in xe_exec_queue_group_cleanup()
119 if (xe_exec_queue_is_multi_queue_secondary(q)) { in xe_exec_queue_group_cleanup()
124 xe_exec_queue_put(xe_exec_queue_multi_queue_primary(q)); in xe_exec_queue_group_cleanup()
141 static void __xe_exec_queue_free(struct xe_exec_queue *q) in __xe_exec_queue_free() argument
146 if (q->tlb_inval[i].dep_scheduler) in __xe_exec_queue_free()
147 xe_dep_scheduler_fini(q->tlb_inval[i].dep_scheduler); in __xe_exec_queue_free()
149 if (xe_exec_queue_uses_pxp(q)) in __xe_exec_queue_free()
150 xe_pxp_exec_queue_remove(gt_to_xe(q in __xe_exec_queue_free()
165 alloc_dep_schedulers(struct xe_device * xe,struct xe_exec_queue * q) alloc_dep_schedulers() argument
204 struct xe_exec_queue *q; __xe_exec_queue_alloc() local
269 __xe_exec_queue_fini(struct xe_exec_queue * q) __xe_exec_queue_fini() argument
279 __xe_exec_queue_init(struct xe_exec_queue * q,u32 exec_queue_flags) __xe_exec_queue_init() argument
342 struct xe_exec_queue *q; xe_exec_queue_create() local
429 struct xe_exec_queue *q; xe_exec_queue_create_bind() local
473 struct xe_exec_queue *q = container_of(ref, struct xe_exec_queue, refcount); xe_exec_queue_destroy() local
503 xe_exec_queue_fini(struct xe_exec_queue * q) xe_exec_queue_fini() argument
517 xe_exec_queue_assign_name(struct xe_exec_queue * q,u32 instance) xe_exec_queue_assign_name() argument
545 struct xe_exec_queue *q; xe_exec_queue_lookup() local
563 exec_queue_set_priority(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_priority() argument
628 exec_queue_set_timeslice(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_timeslice() argument
645 exec_queue_set_pxp_type(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_pxp_type() argument
661 exec_queue_set_hang_replay_state(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_hang_replay_state() argument
677 xe_exec_queue_group_init(struct xe_device * xe,struct xe_exec_queue * q) xe_exec_queue_group_init() argument
719 xe_exec_queue_supports_multi_queue(struct xe_exec_queue * q) xe_exec_queue_supports_multi_queue() argument
724 xe_exec_queue_group_validate(struct xe_device * xe,struct xe_exec_queue * q,u32 primary_id) xe_exec_queue_group_validate() argument
757 xe_exec_queue_group_add(struct xe_device * xe,struct xe_exec_queue * q) xe_exec_queue_group_add() argument
783 xe_exec_queue_group_delete(struct xe_device * xe,struct xe_exec_queue * q) xe_exec_queue_group_delete() argument
795 exec_queue_set_multi_group(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_multi_group() argument
830 exec_queue_set_multi_queue_priority(struct xe_device * xe,struct xe_exec_queue * q,u64 value) exec_queue_set_multi_queue_priority() argument
868 struct xe_exec_queue *q; xe_exec_queue_set_property_ioctl() local
897 exec_queue_user_ext_check(struct xe_exec_queue * q,u64 properties) exec_queue_user_ext_check() argument
913 exec_queue_user_ext_check_final(struct xe_exec_queue * q,u64 properties) exec_queue_user_ext_check_final() argument
924 exec_queue_user_ext_set_property(struct xe_device * xe,struct xe_exec_queue * q,u64 extension,u64 * properties) exec_queue_user_ext_set_property() argument
968 __exec_queue_user_extensions(struct xe_device * xe,struct xe_exec_queue * q,u64 extensions,int ext_number,u64 * properties) __exec_queue_user_extensions() argument
1001 exec_queue_user_extensions(struct xe_device * xe,struct xe_exec_queue * q,u64 extensions) exec_queue_user_extensions() argument
1098 struct xe_exec_queue *q = NULL; xe_exec_queue_create_ioctl() local
1261 struct xe_exec_queue *q; xe_exec_queue_get_property_ioctl() local
1295 xe_exec_queue_lrc(struct xe_exec_queue * q) xe_exec_queue_lrc() argument
1306 xe_exec_queue_is_lr(struct xe_exec_queue * q) xe_exec_queue_is_lr() argument
1326 xe_exec_queue_is_idle(struct xe_exec_queue * q) xe_exec_queue_is_idle() argument
1352 xe_exec_queue_update_run_ticks(struct xe_exec_queue * q) xe_exec_queue_update_run_ticks() argument
1393 xe_exec_queue_kill(struct xe_exec_queue * q) xe_exec_queue_kill() argument
1413 struct xe_exec_queue *q; xe_exec_queue_destroy_ioctl() local
1439 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_lockdep_assert() argument
1457 xe_exec_queue_last_fence_put(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_put() argument
1470 xe_exec_queue_last_fence_put_unlocked(struct xe_exec_queue * q) xe_exec_queue_last_fence_put_unlocked() argument
1487 xe_exec_queue_last_fence_get(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get() argument
1514 xe_exec_queue_last_fence_get_for_resume(struct xe_exec_queue * q,struct xe_vm * vm) xe_exec_queue_last_fence_get_for_resume() argument
1539 xe_exec_queue_last_fence_set(struct xe_exec_queue * q,struct xe_vm * vm,struct dma_fence * fence) xe_exec_queue_last_fence_set() argument
1555 xe_exec_queue_tlb_inval_last_fence_put(struct xe_exec_queue * q,struct xe_vm * vm,unsigned int type) xe_exec_queue_tlb_inval_last_fence_put() argument
1574 xe_exec_queue_tlb_inval_last_fence_put_unlocked(struct xe_exec_queue * q,unsigned int type) xe_exec_queue_tlb_inval_last_fence_put_unlocked() argument
1594 xe_exec_queue_tlb_inval_last_fence_get(struct xe_exec_queue * q,struct xe_vm * vm,unsigned int type) xe_exec_queue_tlb_inval_last_fence_get() argument
1627 xe_exec_queue_tlb_inval_last_fence_set(struct xe_exec_queue * q,struct xe_vm * vm,struct dma_fence * fence,unsigned int type) xe_exec_queue_tlb_inval_last_fence_set() argument
1651 xe_exec_queue_contexts_hwsp_rebase(struct xe_exec_queue * q,void * scratch) xe_exec_queue_contexts_hwsp_rebase() argument
[all...]
/linux/sound/core/seq/
H A Dseq_queue.c50 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument
57 queue_list[i] = q; in queue_list_add()
58 q->queue = i; in queue_list_add()
68 struct snd_seq_queue *q; in queue_list_remove() local
71 q = queue_list[id]; in queue_list_remove()
72 if (q) { in queue_list_remove()
73 guard(spinlock)(&q->owner_lock); in queue_list_remove()
74 if (q->owner == client) { in queue_list_remove()
76 q->klocked = 1; in queue_list_remove()
79 return q; in queue_list_remove()
[all …]
/linux/sound/core/seq/oss/
H A Dseq_oss_readq.c35 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local
37 q = kzalloc_obj(*q); in snd_seq_oss_readq_new()
38 if (!q) in snd_seq_oss_readq_new()
41 q->q = kzalloc_objs(union evrec, maxlen); in snd_seq_oss_readq_new()
42 if (!q->q) { in snd_seq_oss_readq_new()
43 kfree(q); in snd_seq_oss_readq_new()
47 q->maxlen = maxlen; in snd_seq_oss_readq_new()
48 q->qlen = 0; in snd_seq_oss_readq_new()
49 q->head = q->tail = 0; in snd_seq_oss_readq_new()
50 init_waitqueue_head(&q->midi_sleep); in snd_seq_oss_readq_new()
[all …]
H A Dseq_oss_writeq.c27 struct seq_oss_writeq *q; in snd_seq_oss_writeq_new() local
30 q = kzalloc_obj(*q); in snd_seq_oss_writeq_new()
31 if (!q) in snd_seq_oss_writeq_new()
33 q->dp = dp; in snd_seq_oss_writeq_new()
34 q->maxlen = maxlen; in snd_seq_oss_writeq_new()
35 spin_lock_init(&q->sync_lock); in snd_seq_oss_writeq_new()
36 q->sync_event_put = 0; in snd_seq_oss_writeq_new()
37 q->sync_time = 0; in snd_seq_oss_writeq_new()
38 init_waitqueue_head(&q->sync_sleep); in snd_seq_oss_writeq_new()
47 return q; in snd_seq_oss_writeq_new()
[all …]
/linux/net/sched/
H A Dsch_choke.c75 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument
77 return (q->tail - q->head) & q->tab_mask; in choke_len()
81 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument
83 return q->flags & TC_RED_ECN; in use_ecn()
87 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument
89 return q->flags & TC_RED_HARDDROP; in use_harddrop()
93 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument
96 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes()
97 if (q->head == q->tail) in choke_zap_head_holes()
99 } while (q->tab[q->head] == NULL); in choke_zap_head_holes()
[all …]
H A Dsch_netem.c210 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument
212 struct clgstate *clg = &q->clg; in loss_4state()
213 u32 rnd = prandom_u32_state(&q->prng.prng_state); in loss_4state()
275 static bool loss_gilb_ell(struct netem_sched_data *q) in loss_gilb_ell() argument
277 struct clgstate *clg = &q->clg; in loss_gilb_ell()
278 struct rnd_state *s = &q->prng.prng_state; in loss_gilb_ell()
297 static bool loss_event(struct netem_sched_data *q) in loss_event() argument
299 switch (q->loss_model) { in loss_event()
302 return q->loss && q->loss >= get_crandom(&q->loss_cor, &q->prng); in loss_event()
310 return loss_4state(q); in loss_event()
[all …]
H A Dsch_sfq.c143 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument
146 return &q->slots[val].dep; in sfq_dep_head()
147 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head()
150 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument
153 return skb_get_hash_perturb(skb, &q->perturbation) & (q->divisor - 1); in sfq_hash()
159 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() local
166 TC_H_MIN(skb->priority) <= q->divisor) in sfq_classify()
169 fl = rcu_dereference_bh(q->filter_list); in sfq_classify()
171 return sfq_hash(q, sk in sfq_classify()
196 sfq_link(struct sfq_sched_data * q,sfq_index x) sfq_link() argument
212 sfq_unlink(q,x,n,p) global() argument
221 sfq_dec(struct sfq_sched_data * q,sfq_index x) sfq_dec() argument
234 sfq_inc(struct sfq_sched_data * q,sfq_index x) sfq_inc() argument
288 struct sfq_sched_data *q = qdisc_priv(sch); sfq_drop() local
325 sfq_prob_mark(const struct sfq_sched_data * q) sfq_prob_mark() argument
331 sfq_hard_mark(const struct sfq_sched_data * q) sfq_hard_mark() argument
336 sfq_headdrop(const struct sfq_sched_data * q) sfq_headdrop() argument
344 struct sfq_sched_data *q = qdisc_priv(sch); sfq_enqueue() local
478 struct sfq_sched_data *q = qdisc_priv(sch); sfq_dequeue() local
533 struct sfq_sched_data *q = qdisc_priv(sch); sfq_rehash() local
603 struct sfq_sched_data *q = timer_container_of(q, t, perturb_timer); sfq_perturbation() local
630 struct sfq_sched_data *q = qdisc_priv(sch); sfq_change() local
764 struct sfq_sched_data *q = qdisc_priv(sch); sfq_destroy() local
777 struct sfq_sched_data *q = qdisc_priv(sch); sfq_init() local
832 struct sfq_sched_data *q = qdisc_priv(sch); sfq_dump() local
883 sfq_unbind(struct Qdisc * q,unsigned long cl) sfq_unbind() argument
890 struct sfq_sched_data *q = qdisc_priv(sch); sfq_tcf_block() local
907 struct sfq_sched_data *q = qdisc_priv(sch); sfq_dump_class_stats() local
926 struct sfq_sched_data *q = qdisc_priv(sch); sfq_walk() local
[all...]
H A Dsch_dualpi2.c147 static u64 head_enqueue_time(struct Qdisc *q) in head_enqueue_time() argument
149 struct sk_buff *skb = qdisc_peek_head(q); in head_enqueue_time()
170 static ktime_t next_pi2_timeout(struct dualpi2_sched_data *q) in next_pi2_timeout() argument
172 return ktime_add_ns(ktime_get_ns(), q->pi2_tupdate); in next_pi2_timeout()
185 static bool skb_apply_step(struct sk_buff *skb, struct dualpi2_sched_data *q) in skb_apply_step() argument
187 return skb_is_l4s(skb) && qdisc_qlen(q->l_queue) >= q->min_qlen_step; in skb_apply_step()
190 static bool dualpi2_mark(struct dualpi2_sched_data *q, struct sk_buff *skb) in dualpi2_mark() argument
193 q->ecn_mark++; in dualpi2_mark()
199 static void dualpi2_reset_c_protection(struct dualpi2_sched_data *q) in dualpi2_reset_c_protection() argument
209 dualpi2_calculate_c_protection(struct Qdisc * sch,struct dualpi2_sched_data * q,u32 wc) dualpi2_calculate_c_protection() argument
230 dualpi2_classic_marking(struct dualpi2_sched_data * q,struct sk_buff * skb,u32 prob,bool overload) dualpi2_classic_marking() argument
253 dualpi2_scalable_marking(struct dualpi2_sched_data * q,struct sk_buff * skb,u64 local_l_prob,u32 prob,bool overload) dualpi2_scalable_marking() argument
282 must_drop(struct Qdisc * sch,struct dualpi2_sched_data * q,struct sk_buff * skb) must_drop() argument
341 dualpi2_skb_classify(struct dualpi2_sched_data * q,struct sk_buff * skb) dualpi2_skb_classify() argument
388 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_enqueue_skb() local
444 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_qdisc_enqueue() local
517 dequeue_packet(struct Qdisc * sch,struct dualpi2_sched_data * q,int * credit_change,u64 now) dequeue_packet() argument
551 do_step_aqm(struct dualpi2_sched_data * q,struct sk_buff * skb,u64 now) do_step_aqm() argument
574 drop_and_retry(struct dualpi2_sched_data * q,struct sk_buff * skb,struct Qdisc * sch,enum skb_drop_reason reason) drop_and_retry() argument
585 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_qdisc_dequeue() local
626 get_queue_delays(struct dualpi2_sched_data * q,u64 * qdelay_c,u64 * qdelay_l) get_queue_delays() argument
641 struct dualpi2_sched_data *q = qdisc_priv(sch); calculate_probability() local
711 struct dualpi2_sched_data *q = timer_container_of(q, timer, pi2_timer); dualpi2_timer() local
762 struct dualpi2_sched_data *q; dualpi2_change() local
890 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_reset_default() local
916 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_init() local
947 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_dump() local
1023 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_dump_stats() local
1050 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_reset() local
1069 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_destroy() local
1094 dualpi2_unbind(struct Qdisc * q,unsigned long cl) dualpi2_unbind() argument
1101 struct dualpi2_sched_data *q = qdisc_priv(sch); dualpi2_tcf_block() local
[all...]
H A Dsch_fq_pie.c75 static unsigned int fq_pie_hash(const struct fq_pie_sched_data *q, in fq_pie_hash() argument
78 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_pie_hash()
84 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_classify() local
91 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_pie_classify()
94 filter = rcu_dereference_bh(q->filter_list); in fq_pie_classify()
96 return fq_pie_hash(q, skb) + 1; in fq_pie_classify()
112 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_pie_classify()
134 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_qdisc_enqueue() local
152 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue()
155 memory_limited = q in fq_pie_qdisc_enqueue()
240 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_qdisc_dequeue() local
291 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_change() local
389 struct fq_pie_sched_data *q = timer_container_of(q, t, adapt_timer); fq_pie_timer() local
423 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_init() local
475 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_dump() local
511 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_dump_stats() local
536 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_reset() local
555 struct fq_pie_sched_data *q = qdisc_priv(sch); fq_pie_destroy() local
[all...]
H A Dsch_red.c55 static inline int red_use_ecn(struct red_sched_data *q) in red_use_ecn() argument
57 return q->flags & TC_RED_ECN; in red_use_ecn()
60 static inline int red_use_harddrop(struct red_sched_data *q) in red_use_harddrop() argument
62 return q->flags & TC_RED_HARDDROP; in red_use_harddrop()
65 static int red_use_nodrop(struct red_sched_data *q) in red_use_nodrop() argument
67 return q->flags & TC_RED_NODROP; in red_use_nodrop()
74 struct red_sched_data *q = qdisc_priv(sch); in red_enqueue() local
75 struct Qdisc *child = q->qdisc; in red_enqueue()
79 q->vars.qavg = red_calc_qavg(&q in red_enqueue()
155 struct red_sched_data *q = qdisc_priv(sch); red_dequeue() local
172 struct red_sched_data *q = qdisc_priv(sch); red_peek() local
180 struct red_sched_data *q = qdisc_priv(sch); red_reset() local
188 struct red_sched_data *q = qdisc_priv(sch); red_offload() local
217 struct red_sched_data *q = qdisc_priv(sch); red_destroy() local
240 struct red_sched_data *q = qdisc_priv(sch); __red_change() local
324 struct red_sched_data *q = timer_container_of(q, t, adapt_timer); red_adaptative_timer() local
340 struct red_sched_data *q = qdisc_priv(sch); red_init() local
374 struct red_sched_data *q = qdisc_priv(sch); red_change() local
413 struct red_sched_data *q = qdisc_priv(sch); red_dump() local
450 struct red_sched_data *q = qdisc_priv(sch); red_dump_stats() local
476 struct red_sched_data *q = qdisc_priv(sch); red_dump_class() local
501 struct red_sched_data *q = qdisc_priv(sch); red_graft() local
514 struct red_sched_data *q = qdisc_priv(sch); red_leaf() local
[all...]
H A Dsch_fq_codel.c70 static unsigned int fq_codel_hash(const struct fq_codel_sched_data *q, in fq_codel_hash() argument
73 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_codel_hash()
79 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_classify() local
86 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_codel_classify()
89 filter = rcu_dereference_bh(q->filter_list); in fq_codel_classify()
91 return fq_codel_hash(q, skb) + 1; in fq_codel_classify()
107 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_codel_classify()
140 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_drop() local
154 for (i = 0; i < q->flows_cnt; i++) { in fq_codel_drop()
155 if (q in fq_codel_drop()
188 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_enqueue() local
259 struct fq_codel_sched_data *q = qdisc_priv(sch); dequeue_func() local
284 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dequeue() local
336 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_reset() local
370 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_change() local
462 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_destroy() local
472 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_init() local
535 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dump() local
582 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dump_stats() local
623 fq_codel_unbind(struct Qdisc * q,unsigned long cl) fq_codel_unbind() argument
630 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_tcf_block() local
647 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_dump_class_stats() local
693 struct fq_codel_sched_data *q = qdisc_priv(sch); fq_codel_walk() local
[all...]
H A Dsch_fq.c79 /* Following field is only used for q->internal,
80 * because q->internal is not hashed in fq_root[]
93 struct rb_node rate_node; /* anchor in q->delayed tree */
197 static void fq_flow_add_tail(struct fq_sched_data *q, struct fq_flow *flow, in fq_flow_add_tail() argument
200 struct fq_perband_flows *pband = &q->band_flows[flow->band]; in fq_flow_add_tail()
213 static void fq_flow_unset_throttled(struct fq_sched_data *q, struct fq_flow *f) in fq_flow_unset_throttled() argument
215 rb_erase(&f->rate_node, &q->delayed); in fq_flow_unset_throttled()
216 q->throttled_flows--; in fq_flow_unset_throttled()
217 fq_flow_add_tail(q, f, OLD_FLOW); in fq_flow_unset_throttled()
220 static void fq_flow_set_throttled(struct fq_sched_data *q, struc argument
256 fq_gc(struct fq_sched_data * q,struct rb_root * root,struct sock * sk) fq_gc() argument
313 const struct fq_sched_data *q = qdisc_priv(sch); fq_fastpath_check() local
355 struct fq_sched_data *q = qdisc_priv(sch); fq_classify() local
537 fq_packet_beyond_horizon(const struct sk_buff * skb,const struct fq_sched_data * q,u64 now) fq_packet_beyond_horizon() argument
547 struct fq_sched_data *q = qdisc_priv(sch); fq_enqueue() local
609 fq_check_throttled(struct fq_sched_data * q,u64 now) fq_check_throttled() argument
652 struct fq_sched_data *q = qdisc_priv(sch); fq_dequeue() local
802 struct fq_sched_data *q = qdisc_priv(sch); fq_reset() local
838 fq_rehash(struct fq_sched_data * q,struct rb_root * old_array,u32 old_log,struct rb_root * new_array,u32 new_log) fq_rehash() argument
890 struct fq_sched_data *q = qdisc_priv(sch); fq_resize() local
974 fq_load_weights(struct fq_sched_data * q,const struct nlattr * attr,struct netlink_ext_ack * extack) fq_load_weights() argument
993 fq_load_priomap(struct fq_sched_data * q,const struct nlattr * attr,struct netlink_ext_ack * extack) fq_load_priomap() argument
1019 struct fq_sched_data *q = qdisc_priv(sch); fq_change() local
1159 struct fq_sched_data *q = qdisc_priv(sch); fq_destroy() local
1169 struct fq_sched_data *q = qdisc_priv(sch); fq_init() local
1214 struct fq_sched_data *q = qdisc_priv(sch); fq_dump() local
1287 struct fq_sched_data *q = qdisc_priv(sch); fq_dump_stats() local
[all...]
H A Dsch_skbprio.c40 static u16 calc_new_high_prio(const struct skbprio_sched_data *q) in calc_new_high_prio() argument
44 for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) { in calc_new_high_prio()
45 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_high_prio()
53 static u16 calc_new_low_prio(const struct skbprio_sched_data *q) in calc_new_low_prio() argument
57 for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) { in calc_new_low_prio()
58 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_low_prio()
72 struct skbprio_sched_data *q = qdisc_priv(sch); in skbprio_enqueue() local
81 qdisc = &q->qdiscs[prio]; in skbprio_enqueue()
84 if (sch->q.qlen < READ_ONCE(sch->limit)) { in skbprio_enqueue()
87 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue()
[all …]
/linux/drivers/media/common/videobuf2/
H A Dvideobuf2-core.c47 #define dprintk(q, level, fmt, arg...) \ argument
50 pr_info("[%s] %s: " fmt, (q)->name, __func__, \
103 #define log_qop(q, op) \ argument
104 dprintk(q, 2, "call_qop(%s)%s\n", #op, \
105 (q)->ops->op ? "" : " (nop)")
107 #define call_qop(q, op, args...) \ argument
111 log_qop(q, op); \
112 err = (q)->ops->op ? (q)->ops->op(args) : 0; \
114 (q)->cnt_ ## op++; \
118 #define call_void_qop(q, op, args...) \ argument
[all …]
/linux/net/xdp/
H A Dxsk_queue.h125 static inline void __xskq_cons_read_addr_unchecked(struct xsk_queue *q, u32 cached_cons, u64 *addr) in __xskq_cons_read_addr_unchecked()
127 struct xdp_umem_ring *ring = (struct xdp_umem_ring *)q->ring;
128 u32 idx = cached_cons & q->ring_mask; in xskq_cons_read_addr_unchecked() argument
133 static inline bool xskq_cons_read_addr_unchecked(struct xsk_queue *q, u64 *addr) in xskq_cons_read_addr_unchecked()
135 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked()
136 __xskq_cons_read_addr_unchecked(q, q->cached_cons, addr); in xskq_cons_read_addr_unchecked()
222 static inline bool xskq_has_descs(struct xsk_queue *q) in xskq_cons_is_valid_desc() argument
224 return q in xskq_cons_is_valid_desc()
120 __xskq_cons_read_addr_unchecked(struct xsk_queue * q,u32 cached_cons,u64 * addr) __xskq_cons_read_addr_unchecked() argument
217 xskq_has_descs(struct xsk_queue * q) xskq_has_descs() argument
233 xskq_cons_read_desc(struct xsk_queue * q,struct xdp_desc * desc,struct xsk_buff_pool * pool) xskq_cons_read_desc() argument
249 xskq_cons_release_n(struct xsk_queue * q,u32 cnt) xskq_cons_release_n() argument
254 parse_desc(struct xsk_queue * q,struct xsk_buff_pool * pool,struct xdp_desc * desc,struct parsed_desc * parsed) parse_desc() argument
262 xskq_cons_read_desc_batch(struct xsk_queue * q,struct xsk_buff_pool * pool,u32 max) xskq_cons_read_desc_batch() argument
304 __xskq_cons_release(struct xsk_queue * q) __xskq_cons_release() argument
309 __xskq_cons_peek(struct xsk_queue * q) __xskq_cons_peek() argument
315 xskq_cons_get_entries(struct xsk_queue * q) xskq_cons_get_entries() argument
321 xskq_cons_nb_entries(struct xsk_queue * q,u32 max) xskq_cons_nb_entries() argument
334 xskq_cons_peek_addr_unchecked(struct xsk_queue * q,u64 * addr) xskq_cons_peek_addr_unchecked() argument
341 xskq_cons_peek_desc(struct xsk_queue * q,struct xdp_desc * desc,struct xsk_buff_pool * pool) xskq_cons_peek_desc() argument
354 xskq_cons_release(struct xsk_queue * q) xskq_cons_release() argument
359 xskq_cons_cancel_n(struct xsk_queue * q,u32 cnt) xskq_cons_cancel_n() argument
364 xskq_cons_present_entries(struct xsk_queue * q) xskq_cons_present_entries() argument
372 xskq_get_prod(struct xsk_queue * q) xskq_get_prod() argument
377 xskq_prod_nb_free(struct xsk_queue * q,u32 max) xskq_prod_nb_free() argument
391 xskq_prod_is_full(struct xsk_queue * q) xskq_prod_is_full() argument
396 xskq_prod_cancel_n(struct xsk_queue * q,u32 cnt) xskq_prod_cancel_n() argument
401 xskq_prod_reserve(struct xsk_queue * q) xskq_prod_reserve() argument
411 xskq_prod_reserve_addr(struct xsk_queue * q,u64 addr) xskq_prod_reserve_addr() argument
423 xskq_prod_write_addr(struct xsk_queue * q,u32 idx,u64 addr) xskq_prod_write_addr() argument
430 xskq_prod_write_addr_batch(struct xsk_queue * q,struct xdp_desc * descs,u32 nb_entries) xskq_prod_write_addr_batch() argument
443 xskq_prod_reserve_desc(struct xsk_queue * q,u64 addr,u32 len,u32 flags) xskq_prod_reserve_desc() argument
461 __xskq_prod_submit(struct xsk_queue * q,u32 idx) __xskq_prod_submit() argument
466 xskq_prod_submit(struct xsk_queue * q) xskq_prod_submit() argument
471 xskq_prod_submit_n(struct xsk_queue * q,u32 nb_entries) xskq_prod_submit_n() argument
476 xskq_prod_is_empty(struct xsk_queue * q) xskq_prod_is_empty() argument
484 xskq_nb_invalid_descs(struct xsk_queue * q) xskq_nb_invalid_descs() argument
489 xskq_nb_queue_empty_descs(struct xsk_queue * q) xskq_nb_queue_empty_descs() argument
[all...]
/linux/block/
H A Dblk-pm.c29 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument
31 q->dev = dev; in blk_pm_runtime_init()
32 q->rpm_status = RPM_ACTIVE; in blk_pm_runtime_init()
33 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init()
34 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init()
59 int blk_pre_runtime_suspend(struct request_queue *q) in blk_pre_runtime_suspend() argument
63 if (!q->dev) in blk_pre_runtime_suspend()
66 WARN_ON_ONCE(q->rpm_status != RPM_ACTIVE); in blk_pre_runtime_suspend()
68 spin_lock_irq(&q->queue_lock); in blk_pre_runtime_suspend()
69 q->rpm_status = RPM_SUSPENDING; in blk_pre_runtime_suspend()
[all …]
/linux/drivers/infiniband/sw/rxe/
H A Drxe_queue.h83 void rxe_queue_reset(struct rxe_queue *q);
88 int rxe_queue_resize(struct rxe_queue *q, unsigned int *num_elem_p,
95 static inline u32 queue_next_index(struct rxe_queue *q, int index) in queue_next_index() argument
97 return (index + 1) & q->index_mask; in queue_next_index()
100 static inline u32 queue_get_producer(const struct rxe_queue *q, in queue_get_producer() argument
108 prod = smp_load_acquire(&q->buf->producer_index); in queue_get_producer()
112 prod = q->index; in queue_get_producer()
116 prod = q->buf->producer_index; in queue_get_producer()
120 prod = smp_load_acquire(&q->buf->producer_index); in queue_get_producer()
127 static inline u32 queue_get_consumer(const struct rxe_queue *q, in queue_get_consumer() argument
[all …]
/linux/drivers/net/wireless/broadcom/b43/
H A Dpio.c24 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument
37 cookie = (((u16)q->index + 1) << 12); in generate_cookie()
49 struct b43_pio_txqueue *q = NULL; in parse_cookie() local
54 q = pio->tx_queue_AC_BK; in parse_cookie()
57 q = pio->tx_queue_AC_BE; in parse_cookie()
60 q = pio->tx_queue_AC_VI; in parse_cookie()
63 q = pio->tx_queue_AC_VO; in parse_cookie()
66 q = pio->tx_queue_mcast; in parse_cookie()
69 if (B43_WARN_ON(!q)) in parse_cookie()
72 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie()
[all …]
/linux/drivers/s390/cio/
H A Dqdio_main.c116 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument
119 int tmp_count = count, tmp_start = start, nr = q->nr; in qdio_do_eqbs()
122 qperf_inc(q, eqbs); in qdio_do_eqbs()
124 if (!q->is_input_q) in qdio_do_eqbs()
125 nr += q->irq_ptr->nr_input_qs; in qdio_do_eqbs()
127 ccq = do_eqbs(q->irq_ptr->sch_token, state, nr, &tmp_start, &tmp_count, in qdio_do_eqbs()
137 qperf_inc(q, eqbs_partial); in qdio_do_eqbs()
138 DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "EQBS part:%02x", in qdio_do_eqbs()
143 DBF_DEV_EVENT(DBF_WARN, q->irq_ptr, "EQBS again:%2d", ccq); in qdio_do_eqbs()
146 DBF_ERROR("%4x ccq:%3d", SCH_NO(q), ccq); in qdio_do_eqbs()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb3/
H A Dsge.c169 static inline struct sge_qset *rspq_to_qset(const struct sge_rspq *q) in rspq_to_qset() argument
171 return container_of(q, struct sge_qset, rspq); in rspq_to_qset()
174 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument
176 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset()
189 const struct sge_rspq *q, unsigned int credits) in refill_rspq() argument
193 V_RSPQ(q->cntxt_id) | V_CREDITS(credits)); in refill_rspq()
233 static inline void unmap_skb(struct sk_buff *skb, struct sge_txq *q, in unmap_skb() argument
237 struct tx_sw_desc *d = &q->sdesc[cidx]; in unmap_skb()
240 sgp = (struct sg_ent *)&q->desc[cidx].flit[d->sflit]; in unmap_skb()
266 d = cidx + 1 == q->size ? q->sdesc : d + 1; in unmap_skb()
[all …]

12345678910>>...46