Home
last modified time | relevance | path

Searched full:queue (Results 1 – 25 of 4217) sorted by relevance

12345678910>>...169

/linux/drivers/net/wireless/st/cw1200/
H A Dqueue.c3 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers
12 #include "queue.h"
27 static inline void __cw1200_queue_lock(struct cw1200_queue *queue) in __cw1200_queue_lock() argument
29 struct cw1200_queue_stats *stats = queue->stats; in __cw1200_queue_lock()
30 if (queue->tx_locked_cnt++ == 0) { in __cw1200_queue_lock()
31 pr_debug("[TX] Queue %d is locked.\n", in __cw1200_queue_lock()
32 queue->queue_id); in __cw1200_queue_lock()
33 ieee80211_stop_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_lock()
37 static inline void __cw1200_queue_unlock(struct cw1200_queue *queue) in __cw1200_queue_unlock() argument
39 struct cw1200_queue_stats *stats = queue in __cw1200_queue_unlock()
89 __cw1200_queue_gc(struct cw1200_queue * queue,struct list_head * head,bool unlock) __cw1200_queue_gc() argument
135 struct cw1200_queue *queue = cw1200_queue_gc() local
163 cw1200_queue_init(struct cw1200_queue * queue,struct cw1200_queue_stats * stats,u8 queue_id,size_t capacity,unsigned long ttl) cw1200_queue_init() argument
201 cw1200_queue_clear(struct cw1200_queue * queue) cw1200_queue_clear() argument
243 cw1200_queue_deinit(struct cw1200_queue * queue) cw1200_queue_deinit() argument
255 cw1200_queue_get_num_queued(struct cw1200_queue * queue,u32 link_id_map) cw1200_queue_get_num_queued() argument
279 cw1200_queue_put(struct cw1200_queue * queue,struct sk_buff * skb,struct cw1200_txpriv * txpriv) cw1200_queue_put() argument
330 cw1200_queue_get(struct cw1200_queue * queue,u32 link_id_map,struct wsm_tx ** tx,struct ieee80211_tx_info ** tx_info,const struct cw1200_txpriv ** txpriv) cw1200_queue_get() argument
371 cw1200_queue_requeue(struct cw1200_queue * queue,u32 packet_id) cw1200_queue_requeue() argument
413 cw1200_queue_remove(struct cw1200_queue * queue,u32 packet_id) cw1200_queue_remove() argument
464 cw1200_queue_get_skb(struct cw1200_queue * queue,u32 packet_id,struct sk_buff ** skb,const struct cw1200_txpriv ** txpriv) cw1200_queue_get_skb() argument
494 cw1200_queue_lock(struct cw1200_queue * queue) cw1200_queue_lock() argument
501 cw1200_queue_unlock(struct cw1200_queue * queue) cw1200_queue_unlock() argument
508 cw1200_queue_get_xmit_timestamp(struct cw1200_queue * queue,unsigned long * timestamp,u32 pending_frame_id) cw1200_queue_get_xmit_timestamp() argument
[all...]
/linux/drivers/usb/gadget/function/
H A Duvc_queue.c27 * Video buffers queue management.
33 * the videobuf2 queue operations by serializing calls to videobuf2 and a
34 * spinlock to protect the IRQ queue that holds the buffers to be processed by
39 * videobuf2 queue operations
46 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); in uvc_queue_setup() local
47 struct uvc_video *video = container_of(queue, struct uvc_video, queue); in uvc_queue_setup()
63 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); in uvc_buffer_prepare() local
64 struct uvc_video *video = container_of(queue, struct uvc_video, queue); in uvc_buffer_prepare()
74 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED)) in uvc_buffer_prepare()
78 if (queue->use_sg) { in uvc_buffer_prepare()
[all …]
/linux/drivers/nvme/target/
H A Dtcp.c65 * queue before determining it to be idle. This optional module behavior
109 struct nvmet_tcp_queue *queue; member
218 static inline u16 nvmet_tcp_cmd_tag(struct nvmet_tcp_queue *queue, in nvmet_tcp_cmd_tag() argument
221 if (unlikely(!queue->nr_cmds)) { in nvmet_tcp_cmd_tag()
226 return cmd - queue->cmds; in nvmet_tcp_cmd_tag()
254 nvmet_tcp_get_cmd(struct nvmet_tcp_queue *queue) in nvmet_tcp_get_cmd() argument
258 cmd = list_first_entry_or_null(&queue->free_list, in nvmet_tcp_get_cmd()
274 if (unlikely(cmd == &cmd->queue->connect)) in nvmet_tcp_put_cmd()
277 list_add_tail(&cmd->entry, &cmd->queue->free_list); in nvmet_tcp_put_cmd()
280 static inline int queue_cpu(struct nvmet_tcp_queue *queue) in queue_cpu() argument
285 nvmet_tcp_hdgst_len(struct nvmet_tcp_queue * queue) nvmet_tcp_hdgst_len() argument
290 nvmet_tcp_ddgst_len(struct nvmet_tcp_queue * queue) nvmet_tcp_ddgst_len() argument
300 nvmet_tcp_verify_hdgst(struct nvmet_tcp_queue * queue,void * pdu,size_t len) nvmet_tcp_verify_hdgst() argument
326 nvmet_tcp_check_ddgst(struct nvmet_tcp_queue * queue,void * pdu) nvmet_tcp_check_ddgst() argument
399 nvmet_tcp_fatal_error(struct nvmet_tcp_queue * queue) nvmet_tcp_fatal_error() argument
408 nvmet_tcp_socket_error(struct nvmet_tcp_queue * queue,int status) nvmet_tcp_socket_error() argument
476 struct nvmet_tcp_queue *queue = cmd->queue; nvmet_setup_c2h_data_pdu() local
549 nvmet_tcp_process_resp_list(struct nvmet_tcp_queue * queue) nvmet_tcp_process_resp_list() argument
561 nvmet_tcp_fetch_cmd(struct nvmet_tcp_queue * queue) nvmet_tcp_fetch_cmd() argument
591 struct nvmet_tcp_queue *queue = cmd->queue; nvmet_tcp_queue_response() local
657 struct nvmet_tcp_queue *queue = cmd->queue; nvmet_try_send_data() local
769 struct nvmet_tcp_queue *queue = cmd->queue; nvmet_try_send_ddgst() local
802 nvmet_tcp_try_send_one(struct nvmet_tcp_queue * queue,bool last_in_batch) nvmet_tcp_try_send_one() argument
851 nvmet_tcp_try_send(struct nvmet_tcp_queue * queue,int budget,int * sends) nvmet_tcp_try_send() argument
870 nvmet_prepare_receive_pdu(struct nvmet_tcp_queue * queue) nvmet_prepare_receive_pdu() argument
879 nvmet_tcp_handle_icreq(struct nvmet_tcp_queue * queue) nvmet_tcp_handle_icreq() argument
934 nvmet_tcp_handle_req_failure(struct nvmet_tcp_queue * queue,struct nvmet_tcp_cmd * cmd,struct nvmet_req * req) nvmet_tcp_handle_req_failure() argument
965 nvmet_tcp_handle_h2c_data_pdu(struct nvmet_tcp_queue * queue) nvmet_tcp_handle_h2c_data_pdu() argument
1026 nvmet_tcp_done_recv_pdu(struct nvmet_tcp_queue * queue) nvmet_tcp_done_recv_pdu() argument
1137 nvmet_tcp_tls_record_ok(struct nvmet_tcp_queue * queue,struct msghdr * msg,char * cbuf) nvmet_tcp_tls_record_ok() argument
1172 nvmet_tcp_try_recv_pdu(struct nvmet_tcp_queue * queue) nvmet_tcp_try_recv_pdu() argument
1237 struct nvmet_tcp_queue *queue = cmd->queue; nvmet_tcp_prep_recv_ddgst() local
1245 nvmet_tcp_try_recv_data(struct nvmet_tcp_queue * queue) nvmet_tcp_try_recv_data() argument
1278 nvmet_tcp_try_recv_ddgst(struct nvmet_tcp_queue * queue) nvmet_tcp_try_recv_ddgst() argument
1329 nvmet_tcp_try_recv_one(struct nvmet_tcp_queue * queue) nvmet_tcp_try_recv_one() argument
1363 nvmet_tcp_try_recv(struct nvmet_tcp_queue * queue,int budget,int * recvs) nvmet_tcp_try_recv() argument
1384 struct nvmet_tcp_queue *queue = nvmet_tcp_release_queue() local
1391 nvmet_tcp_schedule_release_queue(struct nvmet_tcp_queue * queue) nvmet_tcp_schedule_release_queue() argument
1405 nvmet_tcp_arm_queue_deadline(struct nvmet_tcp_queue * queue) nvmet_tcp_arm_queue_deadline() argument
1410 nvmet_tcp_check_queue_deadline(struct nvmet_tcp_queue * queue,int ops) nvmet_tcp_check_queue_deadline() argument
1424 struct nvmet_tcp_queue *queue = nvmet_tcp_io_work() local
1454 nvmet_tcp_alloc_cmd(struct nvmet_tcp_queue * queue,struct nvmet_tcp_cmd * c) nvmet_tcp_alloc_cmd() argument
1510 nvmet_tcp_alloc_cmds(struct nvmet_tcp_queue * queue) nvmet_tcp_alloc_cmds() argument
1536 nvmet_tcp_free_cmds(struct nvmet_tcp_queue * queue) nvmet_tcp_free_cmds() argument
1548 nvmet_tcp_restore_socket_callbacks(struct nvmet_tcp_queue * queue) nvmet_tcp_restore_socket_callbacks() argument
1563 nvmet_tcp_uninit_data_in_cmds(struct nvmet_tcp_queue * queue) nvmet_tcp_uninit_data_in_cmds() argument
1579 nvmet_tcp_free_cmd_data_in_buffers(struct nvmet_tcp_queue * queue) nvmet_tcp_free_cmd_data_in_buffers() argument
1591 struct nvmet_tcp_queue *queue = nvmet_tcp_release_queue_work() local
1620 struct nvmet_tcp_queue *queue; nvmet_tcp_data_ready() local
1638 struct nvmet_tcp_queue *queue; nvmet_tcp_write_space() local
1660 struct nvmet_tcp_queue *queue; nvmet_tcp_state_change() local
1685 nvmet_tcp_set_queue_sock(struct nvmet_tcp_queue * queue) nvmet_tcp_set_queue_sock() argument
1741 nvmet_tcp_try_peek_pdu(struct nvmet_tcp_queue * queue) nvmet_tcp_try_peek_pdu() argument
1789 nvmet_tcp_tls_key_lookup(struct nvmet_tcp_queue * queue,key_serial_t peerid) nvmet_tcp_tls_key_lookup() argument
1813 struct nvmet_tcp_queue *queue = data; nvmet_tcp_tls_handshake_done() local
1843 struct nvmet_tcp_queue *queue = container_of(to_delayed_work(w), nvmet_tcp_tls_handshake_timeout() local
1863 nvmet_tcp_tls_handshake(struct nvmet_tcp_queue * queue) nvmet_tcp_tls_handshake() argument
1899 struct nvmet_tcp_queue *queue; nvmet_tcp_alloc_queue() local
2113 struct nvmet_tcp_queue *queue; nvmet_tcp_destroy_port_queues() local
2143 struct nvmet_tcp_queue *queue; nvmet_tcp_delete_ctrl() local
2154 struct nvmet_tcp_queue *queue = nvmet_tcp_install_queue() local
2189 struct nvmet_tcp_queue *queue = cmd->queue; nvmet_tcp_disc_port_addr() local
2201 struct nvmet_tcp_queue *queue = nvmet_tcp_host_port_addr() local
2244 struct nvmet_tcp_queue *queue; nvmet_tcp_exit() local
[all...]
H A Drdma.c52 struct nvmet_rdma_queue *queue; member
66 struct nvmet_rdma_queue *queue; member
147 MODULE_PARM_DESC(use_srq, "Use shared receive queue.");
157 MODULE_PARM_DESC(srq_size, "set Shared Receive Queue (SRQ) size, should >= 256 (default: 1024)");
172 static void nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue *queue);
213 nvmet_rdma_get_rsp(struct nvmet_rdma_queue *queue) in nvmet_rdma_get_rsp() argument
218 tag = sbitmap_get(&queue->rsp_tags); in nvmet_rdma_get_rsp()
220 rsp = &queue->rsps[tag]; in nvmet_rdma_get_rsp()
228 ret = nvmet_rdma_alloc_rsp(queue->dev, rsp, in nvmet_rdma_get_rsp()
243 nvmet_rdma_free_rsp(rsp->queue in nvmet_rdma_put_rsp()
448 nvmet_rdma_alloc_rsps(struct nvmet_rdma_queue * queue) nvmet_rdma_alloc_rsps() argument
482 nvmet_rdma_free_rsps(struct nvmet_rdma_queue * queue) nvmet_rdma_free_rsps() argument
513 nvmet_rdma_process_wr_wait_list(struct nvmet_rdma_queue * queue) nvmet_rdma_process_wr_wait_list() argument
662 struct nvmet_rdma_queue *queue = rsp->queue; nvmet_rdma_release_rsp() local
678 nvmet_rdma_error_comp(struct nvmet_rdma_queue * queue) nvmet_rdma_error_comp() argument
696 struct nvmet_rdma_queue *queue = wc->qp->qp_context; nvmet_rdma_send_done() local
749 struct nvmet_rdma_queue *queue = wc->qp->qp_context; nvmet_rdma_read_data_done() local
782 struct nvmet_rdma_queue *queue = wc->qp->qp_context; nvmet_rdma_write_data_done() local
944 struct nvmet_rdma_queue *queue = rsp->queue; nvmet_rdma_execute_command() local
966 nvmet_rdma_handle_command(struct nvmet_rdma_queue * queue,struct nvmet_rdma_rsp * cmd) nvmet_rdma_handle_command() argument
997 nvmet_rdma_recv_not_live(struct nvmet_rdma_queue * queue,struct nvmet_rdma_rsp * rsp) nvmet_rdma_recv_not_live() argument
1022 struct nvmet_rdma_queue *queue = wc->qp->qp_context; nvmet_rdma_recv_done() local
1262 nvmet_rdma_create_queue_ib(struct nvmet_rdma_queue * queue) nvmet_rdma_create_queue_ib() argument
1339 nvmet_rdma_destroy_queue_ib(struct nvmet_rdma_queue * queue) nvmet_rdma_destroy_queue_ib() argument
1349 nvmet_rdma_free_queue(struct nvmet_rdma_queue * queue) nvmet_rdma_free_queue() argument
1369 struct nvmet_rdma_queue *queue = nvmet_rdma_release_queue_work() local
1380 nvmet_rdma_parse_cm_connect_req(struct rdma_conn_param * conn,struct nvmet_rdma_queue * queue) nvmet_rdma_parse_cm_connect_req() argument
1429 struct nvmet_rdma_queue *queue; nvmet_rdma_alloc_queue() local
1529 struct nvmet_rdma_queue *queue = priv; nvmet_rdma_qp_event() local
1547 nvmet_rdma_cm_accept(struct rdma_cm_id * cm_id,struct nvmet_rdma_queue * queue,struct rdma_conn_param * p) nvmet_rdma_cm_accept() argument
1574 struct nvmet_rdma_queue *queue; nvmet_rdma_queue_connect() local
1629 nvmet_rdma_queue_established(struct nvmet_rdma_queue * queue) nvmet_rdma_queue_established() argument
1656 __nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue * queue) __nvmet_rdma_queue_disconnect() argument
1691 nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue * queue) nvmet_rdma_queue_disconnect() argument
1707 nvmet_rdma_queue_connect_fail(struct rdma_cm_id * cm_id,struct nvmet_rdma_queue * queue) nvmet_rdma_queue_connect_fail() argument
1736 nvmet_rdma_device_removal(struct rdma_cm_id * cm_id,struct nvmet_rdma_queue * queue) nvmet_rdma_device_removal() argument
1770 struct nvmet_rdma_queue *queue = NULL; nvmet_rdma_cm_handler() local
1821 struct nvmet_rdma_queue *queue, *n; nvmet_rdma_delete_ctrl() local
1835 struct nvmet_rdma_queue *queue, *tmp; nvmet_rdma_destroy_port_queues() local
2019 struct nvmet_rdma_queue *queue = nvmet_rdma_host_port_addr() local
2057 struct nvmet_rdma_queue *queue, *tmp; nvmet_rdma_remove_one() local
[all...]
/linux/drivers/net/wireless/broadcom/b43legacy/
H A Dpio.c22 static void tx_start(struct b43legacy_pioqueue *queue) in tx_start() argument
24 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, in tx_start()
28 static void tx_octet(struct b43legacy_pioqueue *queue, in tx_octet() argument
31 if (queue->need_workarounds) { in tx_octet()
32 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); in tx_octet()
33 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, in tx_octet()
36 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, in tx_octet()
38 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); in tx_octet()
63 static void tx_data(struct b43legacy_pioqueue *queue, in tx_data() argument
71 if (queue in tx_data()
89 tx_complete(struct b43legacy_pioqueue * queue,struct sk_buff * skb) tx_complete() argument
103 generate_cookie(struct b43legacy_pioqueue * queue,struct b43legacy_pio_txpacket * packet) generate_cookie() argument
141 struct b43legacy_pioqueue *queue = NULL; parse_cookie() local
172 pio_tx_write_fragment(struct b43legacy_pioqueue * queue,struct sk_buff * skb,struct b43legacy_pio_txpacket * packet,size_t txhdr_size) pio_tx_write_fragment() argument
205 struct b43legacy_pioqueue *queue = packet->queue; free_txpacket() local
219 struct b43legacy_pioqueue *queue = packet->queue; pio_tx_packet() local
269 struct b43legacy_pioqueue *queue = from_tasklet(queue, t, txtask); tx_tasklet() local
299 setup_txqueues(struct b43legacy_pioqueue * queue) setup_txqueues() argument
319 struct b43legacy_pioqueue *queue; b43legacy_setup_pioqueue() local
367 cancel_transfers(struct b43legacy_pioqueue * queue) cancel_transfers() argument
379 b43legacy_destroy_pioqueue(struct b43legacy_pioqueue * queue) b43legacy_destroy_pioqueue() argument
409 struct b43legacy_pioqueue *queue; b43legacy_pio_init() local
455 struct b43legacy_pioqueue *queue = dev->pio.queue1; b43legacy_pio_tx() local
477 struct b43legacy_pioqueue *queue; b43legacy_pio_handle_txstatus() local
535 pio_rx_error(struct b43legacy_pioqueue * queue,int clear_buffers,const char * error) pio_rx_error() argument
553 b43legacy_pio_rx(struct b43legacy_pioqueue * queue) b43legacy_pio_rx() argument
634 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue * queue) b43legacy_pio_tx_suspend() argument
642 b43legacy_pio_tx_resume(struct b43legacy_pioqueue * queue) b43legacy_pio_tx_resume() argument
[all...]
/linux/drivers/md/dm-vdo/
H A Dfunnel-workqueue.c15 #include "funnel-queue.h"
28 * DOC: Work queue definition.
36 /* Name of just the work queue (e.g., "cpuQ12") */
73 static inline struct simple_work_queue *as_simple_work_queue(struct vdo_work_queue *queue) in as_simple_work_queue() argument
75 return ((queue == NULL) ? in as_simple_work_queue()
76 NULL : container_of(queue, struct simple_work_queue, common)); in as_simple_work_queue()
79 static inline struct round_robin_work_queue *as_round_robin_work_queue(struct vdo_work_queue *queue) in as_round_robin_work_queue() argument
81 return ((queue == NULL) ? in as_round_robin_work_queue()
83 container_of(queue, struct round_robin_work_queue, common)); in as_round_robin_work_queue()
96 static struct vdo_completion *poll_for_completion(struct simple_work_queue *queue) in poll_for_completion() argument
[all …]
H A Dfunnel-queue.c6 #include "funnel-queue.h"
15 struct funnel_queue *queue; in vdo_make_funnel_queue() local
17 result = vdo_allocate(1, struct funnel_queue, "funnel queue", &queue); in vdo_make_funnel_queue()
22 * Initialize the stub entry and put it in the queue, establishing the invariant that in vdo_make_funnel_queue()
23 * queue->newest and queue->oldest are never null. in vdo_make_funnel_queue()
25 queue->stub.next = NULL; in vdo_make_funnel_queue()
26 queue->newest = &queue->stub; in vdo_make_funnel_queue()
27 queue->oldest = &queue->stub; in vdo_make_funnel_queue()
29 *queue_ptr = queue; in vdo_make_funnel_queue()
33 void vdo_free_funnel_queue(struct funnel_queue *queue) in vdo_free_funnel_queue() argument
[all …]
/linux/drivers/scsi/arm/
H A Dqueue.c3 * linux/drivers/acorn/scsi/queue.c: queue handling primitives
50 #include "queue.h"
55 * Function: void queue_initialise (Queue_t *queue)
56 * Purpose : initialise a queue
57 * Params : queue - queue to initialise
59 int queue_initialise (Queue_t *queue) in queue_initialise() argument
64 spin_lock_init(&queue->queue_lock); in queue_initialise()
65 INIT_LIST_HEAD(&queue in queue_initialise()
91 queue_free(Queue_t * queue) queue_free() argument
107 __queue_add(Queue_t * queue,struct scsi_cmnd * SCpnt,int head) __queue_add() argument
138 __queue_remove(Queue_t * queue,struct list_head * ent) __queue_remove() argument
162 queue_remove_exclude(Queue_t * queue,unsigned long * exclude) queue_remove_exclude() argument
188 queue_remove(Queue_t * queue) queue_remove() argument
210 queue_remove_tgtluntag(Queue_t * queue,int target,int lun,int tag) queue_remove_tgtluntag() argument
238 queue_remove_all_target(Queue_t * queue,int target) queue_remove_all_target() argument
261 queue_probetgtlun(Queue_t * queue,int target,int lun) queue_probetgtlun() argument
287 queue_remove_cmd(Queue_t * queue,struct scsi_cmnd * SCpnt) queue_remove_cmd() argument
[all...]
H A Dqueue.h3 * linux/drivers/acorn/scsi/queue.h: queue handling
18 * Function: void queue_initialise (Queue_t *queue)
19 * Purpose : initialise a queue
20 * Params : queue - queue to initialise
22 extern int queue_initialise (Queue_t *queue);
25 * Function: void queue_free (Queue_t *queue)
26 * Purpose : free a queue
27 * Params : queue - queue to free
29 extern void queue_free (Queue_t *queue);
32 * Function: struct scsi_cmnd *queue_remove (queue)
[all …]
/linux/drivers/nvme/host/
H A Dtcp.c106 struct nvme_tcp_queue *queue; member
207 static int nvme_tcp_try_send(struct nvme_tcp_queue *queue);
214 static inline int nvme_tcp_queue_id(struct nvme_tcp_queue *queue) in nvme_tcp_queue_id() argument
216 return queue - queue->ctrl->queues; in nvme_tcp_queue_id()
233 * Check if the queue is TLS encrypted
235 static inline bool nvme_tcp_queue_tls(struct nvme_tcp_queue *queue) in nvme_tcp_queue_tls() argument
240 return queue->tls_enabled; in nvme_tcp_queue_tls()
254 static inline struct blk_mq_tags *nvme_tcp_tagset(struct nvme_tcp_queue *queue) in nvme_tcp_tagset() argument
256 u32 queue_idx = nvme_tcp_queue_id(queue); in nvme_tcp_tagset()
263 nvme_tcp_hdgst_len(struct nvme_tcp_queue * queue) nvme_tcp_hdgst_len() argument
268 nvme_tcp_ddgst_len(struct nvme_tcp_queue * queue) nvme_tcp_ddgst_len() argument
383 nvme_tcp_send_all(struct nvme_tcp_queue * queue) nvme_tcp_send_all() argument
393 nvme_tcp_queue_has_pending(struct nvme_tcp_queue * queue) nvme_tcp_queue_has_pending() argument
399 nvme_tcp_queue_more(struct nvme_tcp_queue * queue) nvme_tcp_queue_more() argument
408 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_queue_request() local
429 nvme_tcp_process_req_list(struct nvme_tcp_queue * queue) nvme_tcp_process_req_list() argument
441 nvme_tcp_fetch_request(struct nvme_tcp_queue * queue) nvme_tcp_fetch_request() argument
494 nvme_tcp_verify_hdgst(struct nvme_tcp_queue * queue,void * pdu,size_t pdu_len) nvme_tcp_verify_hdgst() argument
520 nvme_tcp_check_ddgst(struct nvme_tcp_queue * queue,void * pdu) nvme_tcp_check_ddgst() argument
556 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx]; nvme_tcp_init_request() local
579 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1]; nvme_tcp_init_hctx() local
589 struct nvme_tcp_queue *queue = &ctrl->queues[0]; nvme_tcp_init_admin_hctx() local
596 nvme_tcp_recv_state(struct nvme_tcp_queue * queue) nvme_tcp_recv_state() argument
603 nvme_tcp_init_recv_ctx(struct nvme_tcp_queue * queue) nvme_tcp_init_recv_ctx() argument
621 nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue * queue,struct nvme_completion * cqe) nvme_tcp_process_nvme_cqe() argument
647 nvme_tcp_handle_c2h_data(struct nvme_tcp_queue * queue,struct nvme_tcp_data_pdu * pdu) nvme_tcp_handle_c2h_data() argument
681 nvme_tcp_handle_comp(struct nvme_tcp_queue * queue,struct nvme_tcp_rsp_pdu * pdu) nvme_tcp_handle_comp() argument
706 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_setup_h2c_data_pdu() local
737 nvme_tcp_handle_r2t(struct nvme_tcp_queue * queue,struct nvme_tcp_r2t_pdu * pdu) nvme_tcp_handle_r2t() argument
796 nvme_tcp_handle_c2h_term(struct nvme_tcp_queue * queue,struct nvme_tcp_term_pdu * pdu) nvme_tcp_handle_c2h_term() argument
830 nvme_tcp_recv_pdu(struct nvme_tcp_queue * queue,struct sk_buff * skb,unsigned int * offset,size_t * len) nvme_tcp_recv_pdu() argument
910 nvme_tcp_recv_data(struct nvme_tcp_queue * queue,struct sk_buff * skb,unsigned int * offset,size_t * len) nvme_tcp_recv_data() argument
981 nvme_tcp_recv_ddgst(struct nvme_tcp_queue * queue,struct sk_buff * skb,unsigned int * offset,size_t * len) nvme_tcp_recv_ddgst() argument
1029 struct nvme_tcp_queue *queue = desc->arg.data; nvme_tcp_recv_skb() local
1064 struct nvme_tcp_queue *queue; nvme_tcp_data_ready() local
1078 struct nvme_tcp_queue *queue; nvme_tcp_write_space() local
1094 struct nvme_tcp_queue *queue; nvme_tcp_state_change() local
1120 nvme_tcp_done_send_req(struct nvme_tcp_queue * queue) nvme_tcp_done_send_req() argument
1140 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_try_send_data() local
1203 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_try_send_cmd_pdu() local
1244 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_try_send_data_pdu() local
1278 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_try_send_ddgst() local
1309 nvme_tcp_try_send(struct nvme_tcp_queue * queue) nvme_tcp_try_send() argument
1359 nvme_tcp_try_recv(struct nvme_tcp_queue * queue) nvme_tcp_try_recv() argument
1377 struct nvme_tcp_queue *queue = nvme_tcp_io_work() local
1422 struct nvme_tcp_queue *queue = &ctrl->queues[0]; nvme_tcp_alloc_async_req() local
1439 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; nvme_tcp_free_queue() local
1458 nvme_tcp_init_connection(struct nvme_tcp_queue * queue) nvme_tcp_init_connection() argument
1589 nvme_tcp_admin_queue(struct nvme_tcp_queue * queue) nvme_tcp_admin_queue() argument
1594 nvme_tcp_default_queue(struct nvme_tcp_queue * queue) nvme_tcp_default_queue() argument
1603 nvme_tcp_read_queue(struct nvme_tcp_queue * queue) nvme_tcp_read_queue() argument
1614 nvme_tcp_poll_queue(struct nvme_tcp_queue * queue) nvme_tcp_poll_queue() argument
1636 nvme_tcp_set_queue_io_cpu(struct nvme_tcp_queue * queue) nvme_tcp_set_queue_io_cpu() argument
1681 struct nvme_tcp_queue *queue = data; nvme_tcp_tls_done() local
1712 nvme_tcp_start_tls(struct nvme_ctrl * nctrl,struct nvme_tcp_queue * queue,key_serial_t pskid) nvme_tcp_start_tls() argument
1768 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; nvme_tcp_alloc_queue() local
1912 nvme_tcp_restore_sock_ops(struct nvme_tcp_queue * queue) nvme_tcp_restore_sock_ops() argument
1924 __nvme_tcp_stop_queue(struct nvme_tcp_queue * queue) __nvme_tcp_stop_queue() argument
1934 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; nvme_tcp_stop_queue_nowait() local
1953 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; nvme_tcp_wait_queue() local
1975 nvme_tcp_setup_sock_ops(struct nvme_tcp_queue * queue) nvme_tcp_setup_sock_ops() argument
1994 struct nvme_tcp_queue *queue = &ctrl->queues[idx]; nvme_tcp_start_queue() local
2579 nvme_tcp_set_sg_inline(struct nvme_tcp_queue * queue,struct nvme_command * c,u32 data_len) nvme_tcp_set_sg_inline() argument
2603 struct nvme_tcp_queue *queue = &ctrl->queues[0]; nvme_tcp_submit_async_event() local
2678 nvme_tcp_map_data(struct nvme_tcp_queue * queue,struct request * rq) nvme_tcp_map_data() argument
2703 struct nvme_tcp_queue *queue = req->queue; nvme_tcp_setup_cmd_pdu() local
2754 struct nvme_tcp_queue *queue = hctx->driver_data; nvme_tcp_commit_rqs() local
2764 struct nvme_tcp_queue *queue = hctx->driver_data; nvme_tcp_queue_rq() local
2793 struct nvme_tcp_queue *queue = hctx->driver_data; nvme_tcp_poll() local
2810 struct nvme_tcp_queue *queue = &to_tcp_ctrl(ctrl)->queues[0]; nvme_tcp_get_address() local
[all...]
/linux/drivers/net/xen-netback/
H A Drx.c42 static void xenvif_update_needed_slots(struct xenvif_queue *queue, in xenvif_update_needed_slots() argument
55 WRITE_ONCE(queue->rx_slots_needed, needed); in xenvif_update_needed_slots()
58 static bool xenvif_rx_ring_slots_available(struct xenvif_queue *queue) in xenvif_rx_ring_slots_available() argument
63 needed = READ_ONCE(queue->rx_slots_needed); in xenvif_rx_ring_slots_available()
68 prod = queue->rx.sring->req_prod; in xenvif_rx_ring_slots_available()
69 cons = queue->rx.req_cons; in xenvif_rx_ring_slots_available()
74 queue->rx.sring->req_event = prod + 1; in xenvif_rx_ring_slots_available()
80 } while (queue->rx.sring->req_prod != prod); in xenvif_rx_ring_slots_available()
85 bool xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb) in xenvif_rx_queue_tail() argument
90 spin_lock_irqsave(&queue->rx_queue.lock, flags); in xenvif_rx_queue_tail()
[all …]
H A Dinterface.c44 /* Number of bytes allowed on the internal guest Rx queue. */
52 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, in xenvif_skb_zerocopy_prepare() argument
56 atomic_inc(&queue->inflight_packets); in xenvif_skb_zerocopy_prepare()
59 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) in xenvif_skb_zerocopy_complete() argument
61 atomic_dec(&queue->inflight_packets); in xenvif_skb_zerocopy_complete()
67 wake_up(&queue->dealloc_wq); in xenvif_skb_zerocopy_complete()
77 static bool xenvif_handle_tx_interrupt(struct xenvif_queue *queue) in xenvif_handle_tx_interrupt() argument
81 rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx); in xenvif_handle_tx_interrupt()
83 napi_schedule(&queue->napi); in xenvif_handle_tx_interrupt()
89 struct xenvif_queue *queue = dev_id; in xenvif_tx_interrupt() local
[all …]
H A Dnetback.c59 /* The time that packets can stay on the guest Rx internal queue
107 static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx,
110 static void make_tx_response(struct xenvif_queue *queue,
115 static void xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx);
117 static inline int tx_work_todo(struct xenvif_queue *queue);
119 static inline unsigned long idx_to_pfn(struct xenvif_queue *queue, in idx_to_pfn() argument
122 return page_to_pfn(queue->mmap_pages[idx]); in idx_to_pfn()
125 static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue, in idx_to_kaddr() argument
128 return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx)); in idx_to_kaddr()
161 void xenvif_kick_thread(struct xenvif_queue *queue) in xenvif_kick_thread() argument
[all …]
/linux/drivers/gpu/drm/imagination/
H A Dpvr_queue.c86 WARN(1, "Invalid queue type"); in get_ctx_state_size()
116 pvr_context_put(fence->queue->ctx); in pvr_queue_fence_release_work()
123 struct pvr_device *pvr_dev = fence->queue->ctx->pvr_dev; in pvr_queue_fence_release()
133 switch (fence->queue->type) { in pvr_queue_job_fence_get_timeline_name()
147 WARN(1, "Invalid queue type"); in pvr_queue_job_fence_get_timeline_name()
156 switch (fence->queue->type) { in pvr_queue_cccb_fence_get_timeline_name()
170 WARN(1, "Invalid queue type"); in pvr_queue_cccb_fence_get_timeline_name()
262 * @queue: The queue this fence belongs to.
267 * pvr_queue_fence::queue fiel
271 pvr_queue_fence_init(struct dma_fence * f,struct pvr_queue * queue,const struct dma_fence_ops * fence_ops,struct pvr_queue_fence_ctx * fence_ctx) pvr_queue_fence_init() argument
296 pvr_queue_cccb_fence_init(struct dma_fence * fence,struct pvr_queue * queue) pvr_queue_cccb_fence_init() argument
314 pvr_queue_job_fence_init(struct dma_fence * fence,struct pvr_queue * queue) pvr_queue_job_fence_init() argument
398 pvr_queue_get_job_cccb_fence(struct pvr_queue * queue,struct pvr_job * job) pvr_queue_get_job_cccb_fence() argument
451 pvr_queue_get_job_kccb_fence(struct pvr_queue * queue,struct pvr_job * job) pvr_queue_get_job_kccb_fence() argument
471 pvr_queue_get_paired_frag_job_dep(struct pvr_queue * queue,struct pvr_job * job) pvr_queue_get_paired_frag_job_dep() argument
509 struct pvr_queue *queue = container_of(s_entity, struct pvr_queue, entity); pvr_queue_prepare_job() local
570 pvr_queue_update_active_state_locked(struct pvr_queue * queue) pvr_queue_update_active_state_locked() argument
602 pvr_queue_update_active_state(struct pvr_queue * queue) pvr_queue_update_active_state() argument
613 struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler); pvr_queue_submit_job_to_cccb() local
755 struct pvr_queue *queue = container_of(job->base.sched, pvr_queue_run_job() local
766 pvr_queue_stop(struct pvr_queue * queue,struct pvr_job * bad_job) pvr_queue_stop() argument
771 pvr_queue_start(struct pvr_queue * queue) pvr_queue_start() argument
812 struct pvr_queue *queue = container_of(sched, struct pvr_queue, scheduler); pvr_queue_timedout_job() local
916 pvr_queue_signal_done_fences(struct pvr_queue * queue) pvr_queue_signal_done_fences() argument
946 pvr_queue_check_job_waiting_for_cccb_space(struct pvr_queue * queue) pvr_queue_check_job_waiting_for_cccb_space() argument
999 pvr_queue_process(struct pvr_queue * queue) pvr_queue_process() argument
1008 get_dm_type(struct pvr_queue * queue) get_dm_type() argument
1032 init_fw_context(struct pvr_queue * queue,void * fw_ctx_map) init_fw_context() argument
1063 pvr_queue_cleanup_fw_context(struct pvr_queue * queue) pvr_queue_cleanup_fw_context() argument
1090 struct pvr_queue *queue; pvr_queue_job_init() local
1166 struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler); pvr_queue_job_push() local
1178 struct pvr_queue *queue = priv; reg_state_init() local
1240 struct pvr_queue *queue; pvr_queue_create() local
1348 struct pvr_queue *queue; pvr_queue_device_pre_reset() local
1360 struct pvr_queue *queue; pvr_queue_device_post_reset() local
1379 pvr_queue_kill(struct pvr_queue * queue) pvr_queue_kill() argument
1393 pvr_queue_destroy(struct pvr_queue * queue) pvr_queue_destroy() argument
[all...]
/linux/fs/fuse/
H A Ddev_uring.c51 static void fuse_uring_flush_bg(struct fuse_ring_queue *queue) in fuse_uring_flush_bg() argument
53 struct fuse_ring *ring = queue->ring; in fuse_uring_flush_bg()
56 lockdep_assert_held(&queue->lock); in fuse_uring_flush_bg()
60 * Allow one bg request per queue, ignoring global fc limits. in fuse_uring_flush_bg()
61 * This prevents a single queue from consuming all resources and in fuse_uring_flush_bg()
62 * eliminates the need for remote queue wake-ups when global in fuse_uring_flush_bg()
63 * limits are met but this queue has no more waiting requests. in fuse_uring_flush_bg()
66 !queue->active_background) && in fuse_uring_flush_bg()
67 (!list_empty(&queue->fuse_req_bg_queue))) { in fuse_uring_flush_bg()
70 req = list_first_entry(&queue in fuse_uring_flush_bg()
82 struct fuse_ring_queue *queue = ent->queue; fuse_uring_req_end() local
107 fuse_uring_abort_end_queue_requests(struct fuse_ring_queue * queue) fuse_uring_abort_end_queue_requests() argument
125 struct fuse_ring_queue *queue; fuse_uring_abort_end_requests() local
163 struct fuse_ring_queue *queue; fuse_uring_request_expired() local
197 struct fuse_ring_queue *queue = ring->queues[qid]; fuse_uring_destruct() local
274 struct fuse_ring_queue *queue; fuse_uring_create_queue() local
333 struct fuse_ring_queue *queue = ent->queue; fuse_uring_entry_teardown() local
363 fuse_uring_stop_list_entries(struct list_head * head,struct fuse_ring_queue * queue,enum fuse_ring_req_state exp_state) fuse_uring_stop_list_entries() argument
392 fuse_uring_teardown_entries(struct fuse_ring_queue * queue) fuse_uring_teardown_entries() argument
409 struct fuse_ring_queue *queue = ring->queues[qid]; fuse_uring_log_ent_state() local
440 struct fuse_ring_queue *queue = READ_ONCE(ring->queues[qid]); fuse_uring_async_stop_queues() local
475 struct fuse_ring_queue *queue = READ_ONCE(ring->queues[qid]); fuse_uring_stop_queues() local
504 struct fuse_ring_queue *queue; fuse_uring_cancel() local
670 struct fuse_ring_queue *queue = ent->queue; fuse_uring_copy_to_ring() local
726 struct fuse_ring_queue *queue = ent->queue; fuse_uring_send_next_to_ring() local
749 fuse_uring_ent_avail(struct fuse_ring_ent * ent,struct fuse_ring_queue * queue) fuse_uring_ent_avail() argument
760 struct fuse_ring_queue *queue = ent->queue; fuse_uring_add_to_pq() local
775 struct fuse_ring_queue *queue = ent->queue; fuse_uring_add_req_to_ring_ent() local
797 struct fuse_ring_queue *queue = ent->queue; fuse_uring_ent_assign_req() local
844 fuse_uring_next_fuse_req(struct fuse_ring_ent * ent,struct fuse_ring_queue * queue,unsigned int issue_flags) fuse_uring_next_fuse_req() argument
865 struct fuse_ring_queue *queue = ent->queue; fuse_ring_ent_set_commit() local
887 struct fuse_ring_queue *queue; fuse_uring_commit_fetch() local
957 struct fuse_ring_queue *queue; is_ring_ready() local
986 struct fuse_ring_queue *queue = ent->queue; fuse_uring_do_register() local
1037 fuse_uring_create_ring_ent(struct io_uring_cmd * cmd,struct fuse_ring_queue * queue) fuse_uring_create_ring_ent() argument
1090 struct fuse_ring_queue *queue; fuse_uring_register() local
1202 struct fuse_ring_queue *queue = ent->queue; fuse_uring_send() local
1223 struct fuse_ring_queue *queue = ent->queue; fuse_uring_send_in_task() local
1242 struct fuse_ring_queue *queue; fuse_uring_task_to_queue() local
1270 struct fuse_ring_queue *queue; fuse_uring_queue_fuse_req() local
1313 struct fuse_ring_queue *queue; fuse_uring_queue_bq_req() local
1360 struct fuse_ring_queue *queue = req->ring_queue; fuse_uring_remove_pending_req() local
[all...]
/linux/drivers/net/
H A Dxen-netfront.c91 /* Queue name is interface name with "-qNNN" appended */
94 /* IRQ name is queue name with "-tx" or "-rx" appended */
108 unsigned int id; /* Queue ID, 0-based */
166 /* Multi-queue support */
218 static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue, in xennet_get_rx_skb() argument
222 struct sk_buff *skb = queue->rx_skbs[i]; in xennet_get_rx_skb()
223 queue->rx_skbs[i] = NULL; in xennet_get_rx_skb()
227 static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue, in xennet_get_rx_ref() argument
231 grant_ref_t ref = queue->grant_rx_ref[i]; in xennet_get_rx_ref()
232 queue in xennet_get_rx_ref()
248 struct netfront_queue *queue = timer_container_of(queue, t, rx_refill_timeout() local
253 netfront_tx_slot_available(struct netfront_queue * queue) netfront_tx_slot_available() argument
259 xennet_maybe_wake_tx(struct netfront_queue * queue) xennet_maybe_wake_tx() argument
271 xennet_alloc_one_rx_buffer(struct netfront_queue * queue) xennet_alloc_one_rx_buffer() argument
299 xennet_alloc_rx_buffers(struct netfront_queue * queue) xennet_alloc_rx_buffers() argument
366 struct netfront_queue *queue = NULL; xennet_open() local
390 xennet_tx_buf_gc(struct netfront_queue * queue) xennet_tx_buf_gc() argument
466 struct netfront_queue *queue; global() member
483 struct netfront_queue *queue = info->queue; xennet_tx_setup_grant() local
604 xennet_mark_tx_pending(struct netfront_queue * queue) xennet_mark_tx_pending() argument
614 xennet_xdp_xmit_one(struct net_device * dev,struct netfront_queue * queue,struct xdp_frame * xdpf) xennet_xdp_xmit_one() argument
649 struct netfront_queue *queue = NULL; xennet_xdp_xmit() local
718 struct netfront_queue *queue = NULL; xennet_start_xmit() local
868 struct netfront_queue *queue; xennet_close() local
885 struct netfront_queue *queue = &info->queues[i]; xennet_destroy_queues() local
902 xennet_set_rx_rsp_cons(struct netfront_queue * queue,RING_IDX val) xennet_set_rx_rsp_cons() argument
912 xennet_move_rx_slot(struct netfront_queue * queue,struct sk_buff * skb,grant_ref_t ref) xennet_move_rx_slot() argument
925 xennet_get_extras(struct netfront_queue * queue,struct xen_netif_extra_info * extras,RING_IDX rp) xennet_get_extras() argument
967 xennet_run_xdp(struct netfront_queue * queue,struct page * pdata,struct xen_netif_rx_response * rx,struct bpf_prog * prog,struct xdp_buff * xdp,bool * need_xdp_flush) xennet_run_xdp() argument
1021 xennet_get_responses(struct netfront_queue * queue,struct netfront_rx_info * rinfo,RING_IDX rp,struct sk_buff_head * list,bool * need_xdp_flush) xennet_get_responses() argument
1164 xennet_fill_frags(struct netfront_queue * queue,struct sk_buff * skb,struct sk_buff_head * list) xennet_fill_frags() argument
1227 handle_incoming_queue(struct netfront_queue * queue,struct sk_buff_head * rxq) handle_incoming_queue() argument
1265 struct netfront_queue *queue = container_of(napi, struct netfront_queue, napi); xennet_poll() local
1423 xennet_release_tx_bufs(struct netfront_queue * queue) xennet_release_tx_bufs() argument
1445 xennet_release_rx_bufs(struct netfront_queue * queue) xennet_release_rx_bufs() argument
1514 xennet_handle_tx(struct netfront_queue * queue,unsigned int * eoi) xennet_handle_tx() argument
1539 xennet_handle_rx(struct netfront_queue * queue,unsigned int * eoi) xennet_handle_rx() argument
1823 struct netfront_queue *queue = &info->queues[i]; xennet_disconnect_backend() local
1905 setup_netfront_single(struct netfront_queue * queue) setup_netfront_single() argument
1931 setup_netfront_split(struct netfront_queue * queue) setup_netfront_split() argument
1976 setup_netfront(struct xenbus_device * dev,struct netfront_queue * queue,unsigned int feature_split_evtchn) setup_netfront() argument
2026 xennet_init_queue(struct netfront_queue * queue) xennet_init_queue() argument
2082 write_queue_xenstore_keys(struct netfront_queue * queue,struct xenbus_transaction * xbt,int write_hierarchical) write_queue_xenstore_keys() argument
2166 xennet_create_page_pool(struct netfront_queue * queue) xennet_create_page_pool() argument
2221 struct netfront_queue *queue = &info->queues[i]; xennet_create_queues() local
2266 struct netfront_queue *queue = NULL; talk_to_netback() local
2430 struct netfront_queue *queue = NULL; xennet_connect() local
[all...]
/linux/drivers/iio/buffer/
H A Dindustrialio-buffer-dma.c38 * means of two queues. The incoming queue and the outgoing queue. Blocks on the
39 * incoming queue are waiting for the DMA controller to pick them up and fill
40 * them with data. Block on the outgoing queue have been filled with data and
56 * incoming or outgoing queue the block will be freed.
102 struct iio_dma_buffer_queue *queue = block->queue; in iio_buffer_block_release() local
107 dma_free_coherent(queue->dev, PAGE_ALIGN(block->size), in iio_buffer_block_release()
110 atomic_dec(&queue->num_dmabufs); in iio_buffer_block_release()
113 iio_buffer_put(&queue in iio_buffer_block_release()
173 iio_dma_buffer_alloc_block(struct iio_dma_buffer_queue * queue,size_t size,bool fileio) iio_dma_buffer_alloc_block() argument
209 iio_dma_buffer_queue_wake(struct iio_dma_buffer_queue * queue) iio_dma_buffer_queue_wake() argument
230 struct iio_dma_buffer_queue *queue = block->queue; iio_dma_buffer_block_done() local
257 iio_dma_buffer_block_list_abort(struct iio_dma_buffer_queue * queue,struct list_head * list) iio_dma_buffer_block_list_abort() argument
302 iio_dma_buffer_can_use_fileio(struct iio_dma_buffer_queue * queue) iio_dma_buffer_can_use_fileio() argument
321 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_request_update() local
414 iio_dma_buffer_fileio_free(struct iio_dma_buffer_queue * queue) iio_dma_buffer_fileio_free() argument
437 iio_dma_buffer_submit_block(struct iio_dma_buffer_queue * queue,struct iio_dma_buffer_block * block) iio_dma_buffer_submit_block() argument
484 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_enable() local
508 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_disable() local
520 iio_dma_buffer_enqueue(struct iio_dma_buffer_queue * queue,struct iio_dma_buffer_block * block) iio_dma_buffer_enqueue() argument
534 iio_dma_buffer_dequeue(struct iio_dma_buffer_queue * queue) iio_dma_buffer_dequeue() argument
556 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_io() local
642 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buf); iio_dma_buffer_usage() local
675 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_attach_dmabuf() local
708 struct iio_dma_buffer_queue *queue = block->queue; iio_dma_can_enqueue_block() local
733 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_enqueue_dmabuf() local
767 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_lock_queue() local
775 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); iio_dma_buffer_unlock_queue() local
827 iio_dma_buffer_init(struct iio_dma_buffer_queue * queue,struct device * dev,const struct iio_dma_buffer_ops * ops) iio_dma_buffer_init() argument
850 iio_dma_buffer_exit(struct iio_dma_buffer_queue * queue) iio_dma_buffer_exit() argument
867 iio_dma_buffer_release(struct iio_dma_buffer_queue * queue) iio_dma_buffer_release() argument
[all...]
/linux/drivers/md/dm-vdo/indexer/
H A Dfunnel-requestqueue.c12 #include "funnel-queue.h"
18 * This queue will attempt to handle requests in reasonably sized batches instead of reacting
22 * If the wait time becomes long enough, the queue will become dormant and must be explicitly
24 * queue via xchg (which is a memory barrier), and later checks "dormant" to decide whether to do a
28 * decide if the funnel queue is idle. In dormant mode, the last examination of "newest" before
31 * queue's "next" field update isn't visible yet to make the entry accessible, its existence will
35 * the queue to awaken immediately.
50 /* Wait queue for synchronizing producers and consumer */
54 /* Queue of new incoming requests */
56 /* Queue of old requests to retry */
[all …]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_userq.c125 * Iterate through all queue types to detect and reset problematic queues in amdgpu_userq_detect_and_reset_queues()
126 * Process each queue type in the defined order in amdgpu_userq_detect_and_reset_queues()
153 struct amdgpu_usermode_queue *queue = container_of(work, in amdgpu_userq_hang_detect_work() local
159 if (!queue || !queue->userq_mgr) in amdgpu_userq_hang_detect_work()
162 uq_mgr = queue->userq_mgr; in amdgpu_userq_hang_detect_work()
163 fence = READ_ONCE(queue->hang_detect_fence); in amdgpu_userq_hang_detect_work()
174 * Start hang detection for a user queue fence. A delayed work will be scheduled
177 void amdgpu_userq_start_hang_detect_work(struct amdgpu_usermode_queue *queue) in amdgpu_userq_start_hang_detect_work() argument
182 if (!queue || !queu in amdgpu_userq_start_hang_detect_work()
208 amdgpu_userq_init_hang_detect_work(struct amdgpu_usermode_queue * queue) amdgpu_userq_init_hang_detect_work() argument
214 amdgpu_userq_buffer_va_list_add(struct amdgpu_usermode_queue * queue,struct amdgpu_bo_va_mapping * va_map,u64 addr) amdgpu_userq_buffer_va_list_add() argument
233 amdgpu_userq_input_va_validate(struct amdgpu_device * adev,struct amdgpu_usermode_queue * queue,u64 addr,u64 expected_size) amdgpu_userq_input_va_validate() argument
286 amdgpu_userq_buffer_vas_mapped(struct amdgpu_usermode_queue * queue) amdgpu_userq_buffer_vas_mapped() argument
313 amdgpu_userq_buffer_vas_list_cleanup(struct amdgpu_device * adev,struct amdgpu_usermode_queue * queue) amdgpu_userq_buffer_vas_list_cleanup() argument
338 amdgpu_userq_preempt_helper(struct amdgpu_usermode_queue * queue) amdgpu_userq_preempt_helper() argument
363 amdgpu_userq_restore_helper(struct amdgpu_usermode_queue * queue) amdgpu_userq_restore_helper() argument
383 amdgpu_userq_unmap_helper(struct amdgpu_usermode_queue * queue) amdgpu_userq_unmap_helper() argument
409 amdgpu_userq_map_helper(struct amdgpu_usermode_queue * queue) amdgpu_userq_map_helper() argument
430 amdgpu_userq_wait_for_last_fence(struct amdgpu_usermode_queue * queue) amdgpu_userq_wait_for_last_fence() argument
449 amdgpu_userq_cleanup(struct amdgpu_usermode_queue * queue,int queue_id) amdgpu_userq_cleanup() argument
633 struct amdgpu_usermode_queue *queue; amdgpu_userq_destroy() local
701 struct amdgpu_usermode_queue *queue = m->private; amdgpu_mqd_info_read() local
745 struct amdgpu_usermode_queue *queue; amdgpu_userq_create() local
1020 struct amdgpu_usermode_queue *queue; amdgpu_userq_restore_all() local
1262 struct amdgpu_usermode_queue *queue; amdgpu_userq_evict_all() local
1299 struct amdgpu_usermode_queue *queue; amdgpu_userq_wait_for_signal() local
1363 struct amdgpu_usermode_queue *queue; amdgpu_userq_mgr_fini() local
1384 struct amdgpu_usermode_queue *queue; amdgpu_userq_suspend() local
1410 struct amdgpu_usermode_queue *queue; amdgpu_userq_resume() local
1436 struct amdgpu_usermode_queue *queue; amdgpu_userq_stop_sched_for_enforce_isolation() local
1470 struct amdgpu_usermode_queue *queue; amdgpu_userq_start_sched_for_enforce_isolation() local
1532 struct amdgpu_usermode_queue *queue; amdgpu_userq_pre_reset() local
1559 struct amdgpu_usermode_queue *queue; amdgpu_userq_post_reset() local
[all...]
/linux/drivers/misc/genwqe/
H A Dcard_ddcb.c14 * Device Driver Control Block (DDCB) queue support. Definition of
15 * interrupt handlers for queue support as well as triggering the
40 * Situation (1): Empty queue
56 * Situation (3): Queue wrapped, A > N
64 * Situation (4a): Queue full N > A
73 * Situation (4a): Queue full A > N
82 static int queue_empty(struct ddcb_queue *queue) in queue_empty() argument
84 return queue->ddcb_next == queue->ddcb_act; in queue_empty()
87 static int queue_enqueued_ddcbs(struct ddcb_queue *queue) in queue_enqueued_ddcbs() argument
95 queue_free_ddcbs(struct ddcb_queue * queue) queue_free_ddcbs() argument
163 print_ddcb_info(struct genwqe_dev * cd,struct ddcb_queue * queue) print_ddcb_info() argument
265 enqueue_ddcb(struct genwqe_dev * cd,struct ddcb_queue * queue,struct ddcb * pddcb,int ddcb_no) enqueue_ddcb() argument
332 struct ddcb_queue *queue = req->queue; copy_ddcb_results() local
366 genwqe_check_ddcb_queue(struct genwqe_dev * cd,struct ddcb_queue * queue) genwqe_check_ddcb_queue() argument
481 struct ddcb_queue *queue; __genwqe_wait_ddcb() local
506 struct ddcb_queue *queue = req->queue; __genwqe_wait_ddcb() local
568 get_next_ddcb(struct genwqe_dev * cd,struct ddcb_queue * queue,int * num) get_next_ddcb() argument
625 struct ddcb_queue *queue = req->queue; __genwqe_purge_ddcb() local
759 struct ddcb_queue *queue; __genwqe_enqueue_ddcb() local
981 struct ddcb_queue *queue = &cd->queue; genwqe_next_ddcb_ready() local
1012 struct ddcb_queue *queue = &cd->queue; genwqe_ddcbs_in_flight() local
1021 setup_ddcb_queue(struct genwqe_dev * cd,struct ddcb_queue * queue) setup_ddcb_queue() argument
1097 ddcb_queue_initialized(struct ddcb_queue * queue) ddcb_queue_initialized() argument
1102 free_ddcb_queue(struct genwqe_dev * cd,struct ddcb_queue * queue) free_ddcb_queue() argument
1230 struct ddcb_queue *queue; genwqe_setup_service_layer() local
1319 struct ddcb_queue *queue = &cd->queue; queue_wake_up_all() local
1346 struct ddcb_queue *queue = &cd->queue; genwqe_finish_queue() local
[all...]
/linux/drivers/net/wireless/ralink/rt2x00/
H A Drt2x00queue.c12 Abstract: rt2x00 queue specific routines.
25 struct data_queue *queue = entry->queue; in rt2x00queue_alloc_rxskb() local
26 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev; in rt2x00queue_alloc_rxskb()
37 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; in rt2x00queue_alloc_rxskb()
95 struct device *dev = entry->queue->rt2x00dev->dev; in rt2x00queue_map_txskb()
112 struct device *dev = entry->queue->rt2x00dev->dev; in rt2x00queue_unmap_skb()
488 struct rt2x00_dev *rt2x00dev = entry->queue in rt2x00queue_write_tx_data()
529 struct data_queue *queue = entry->queue; rt2x00queue_write_tx_descriptor() local
540 rt2x00queue_kick_tx_queue(struct data_queue * queue,struct txentry_desc * txdesc) rt2x00queue_kick_tx_queue() argument
599 rt2x00queue_write_tx_frame(struct data_queue * queue,struct sk_buff * skb,struct ieee80211_sta * sta,bool local) rt2x00queue_write_tx_frame() argument
784 rt2x00queue_for_each_entry(struct data_queue * queue,enum queue_index start,enum queue_index end,void * data,bool (* fn)(struct queue_entry * entry,void * data)) rt2x00queue_for_each_entry() argument
839 rt2x00queue_get_entry(struct data_queue * queue,enum queue_index index) rt2x00queue_get_entry() argument
863 struct data_queue *queue = entry->queue; rt2x00queue_index_inc() local
890 rt2x00queue_pause_queue_nocheck(struct data_queue * queue) rt2x00queue_pause_queue_nocheck() argument
907 rt2x00queue_pause_queue(struct data_queue * queue) rt2x00queue_pause_queue() argument
918 rt2x00queue_unpause_queue(struct data_queue * queue) rt2x00queue_unpause_queue() argument
949 rt2x00queue_start_queue(struct data_queue * queue) rt2x00queue_start_queue() argument
969 rt2x00queue_stop_queue(struct data_queue * queue) rt2x00queue_stop_queue() argument
986 rt2x00queue_flush_queue(struct data_queue * queue,bool drop) rt2x00queue_flush_queue() argument
1025 struct data_queue *queue; rt2x00queue_start_queues() local
1040 struct data_queue *queue; rt2x00queue_stop_queues() local
1059 struct data_queue *queue; rt2x00queue_flush_queues() local
1068 rt2x00queue_reset(struct data_queue * queue) rt2x00queue_reset() argument
1086 struct data_queue *queue; rt2x00queue_init_queues() local
1097 rt2x00queue_alloc_entries(struct data_queue * queue) rt2x00queue_alloc_entries() argument
1134 rt2x00queue_free_skbs(struct data_queue * queue) rt2x00queue_free_skbs() argument
1146 rt2x00queue_alloc_rxskbs(struct data_queue * queue) rt2x00queue_alloc_rxskbs() argument
1163 struct data_queue *queue; rt2x00queue_initialize() local
1202 struct data_queue *queue; rt2x00queue_uninitialize() local
1213 rt2x00queue_init(struct rt2x00_dev * rt2x00dev,struct data_queue * queue,enum data_queue_qid qid) rt2x00queue_init() argument
1233 struct data_queue *queue; rt2x00queue_allocate() local
[all...]
/linux/drivers/net/wireguard/
H A Dqueueing.c25 int wg_packet_queue_init(struct crypt_queue *queue, work_func_t function, in wg_packet_queue_init() argument
30 memset(queue, 0, sizeof(*queue)); in wg_packet_queue_init()
31 queue->last_cpu = -1; in wg_packet_queue_init()
32 ret = ptr_ring_init(&queue->ring, len, GFP_KERNEL); in wg_packet_queue_init()
35 queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue); in wg_packet_queue_init()
36 if (!queue->worker) { in wg_packet_queue_init()
37 ptr_ring_cleanup(&queue->ring, NULL); in wg_packet_queue_init()
43 void wg_packet_queue_free(struct crypt_queue *queue, bool purge) in wg_packet_queue_free() argument
45 free_percpu(queue->worker); in wg_packet_queue_free()
46 WARN_ON(!purge && !__ptr_ring_empty(&queue->ring)); in wg_packet_queue_free()
[all …]
/linux/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h40 * WQE - Work Queue Entry
41 * SWQE - Send Work Queue Entry
42 * RWQE - Receive Work Queue Entry
43 * CQE - Completion Queue Entry
44 * EQE - Event Queue Entry
196 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) in hw_qeit_calc() argument
200 if (q_offset >= queue->queue_length) in hw_qeit_calc()
201 q_offset -= queue->queue_length; in hw_qeit_calc()
202 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT]; in hw_qeit_calc()
206 static inline void *hw_qeit_get(struct hw_queue *queue) in hw_qeit_get() argument
[all …]
/linux/net/sunrpc/
H A Dsched.c91 * queue->lock and bh_disabled in order to avoid races within
95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) in __rpc_disable_timer() argument
101 if (list_empty(&queue->timer_list.list)) in __rpc_disable_timer()
102 cancel_delayed_work(&queue->timer_list.dwork); in __rpc_disable_timer()
106 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) in rpc_set_queue_timer() argument
109 queue->timer_list.expires = expires; in rpc_set_queue_timer()
114 mod_delayed_work(rpciod_workqueue, &queue->timer_list.dwork, expires); in rpc_set_queue_timer()
121 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task, in __rpc_add_timer() argument
125 if (list_empty(&queue->timer_list.list) || time_before(timeout, queue->timer_list.expires)) in __rpc_add_timer()
126 rpc_set_queue_timer(queue, timeout); in __rpc_add_timer()
[all …]
/linux/drivers/net/wireless/ath/ath5k/
H A Dqcu.c20 Queue Control Unit, DCF Control Unit Functions
31 * DOC: Queue Control Unit (QCU)/DCF Control Unit (DCU) functions
35 * basically we have 10 queues to play with. Each queue has a matching
36 * QCU that controls when the queue will get triggered and multiple QCUs
39 * and DCUs allowing us to have different DFS settings for each queue.
41 * When a frame goes into a TX queue, QCU decides when it'll trigger a
43 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue
58 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue
60 * @queue: One of enum ath5k_tx_queue_id
63 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) in ath5k_hw_num_tx_pending() argument
[all …]

12345678910>>...169