Searched refs:bufq (Results 1 – 8 of 8) sorted by relevance
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | xsk.c | 27 struct idpf_buf_queue *bufq) in idpf_xsk_setup_bufq() argument 37 if (&grp->splitq.bufq_sets[j].bufq == bufq) { in idpf_xsk_setup_bufq() 49 bufq->pool = pool; in idpf_xsk_setup_bufq() 51 idpf_queue_set(XSK, bufq); in idpf_xsk_setup_bufq() 127 struct idpf_buf_queue *bufq; in idpf_xsk_clear_queue() local 140 bufq = q; in idpf_xsk_clear_queue() 141 if (!idpf_queue_has_clear(XSK, bufq)) in idpf_xsk_clear_queue() 144 bufq->pool = NULL; in idpf_xsk_clear_queue() 355 static bool idpf_xskfq_refill_thresh(struct idpf_buf_queue *bufq, u32 count) in idpf_xskfq_refill_thresh() argument 358 .pool = bufq->pool, in idpf_xskfq_refill_thresh() [all …]
|
| H A D | idpf_txrx.c | 362 static void idpf_rx_hdr_buf_rel_all(struct idpf_buf_queue *bufq) in idpf_rx_hdr_buf_rel_all() argument 365 .fqes = bufq->hdr_buf, in idpf_rx_hdr_buf_rel_all() 366 .pp = bufq->hdr_pp, in idpf_rx_hdr_buf_rel_all() 369 for (u32 i = 0; i < bufq->desc_count; i++) in idpf_rx_hdr_buf_rel_all() 370 idpf_rx_page_rel(&bufq->hdr_buf[i]); in idpf_rx_hdr_buf_rel_all() 373 bufq->hdr_buf = NULL; in idpf_rx_hdr_buf_rel_all() 374 bufq->hdr_pp = NULL; in idpf_rx_hdr_buf_rel_all() 381 static void idpf_rx_buf_rel_bufq(struct idpf_buf_queue *bufq) in idpf_rx_buf_rel_bufq() argument 384 .fqes = bufq->buf, in idpf_rx_buf_rel_bufq() 385 .pp = bufq->pp, in idpf_rx_buf_rel_bufq() [all …]
|
| H A D | idpf_virtchnl.c | 1449 q = &rx_qgrp->splitq.bufq_sets[j].bufq; in __idpf_queue_reg_init() 1943 q->rx_buf_size = sets[0].bufq.rx_buf_size; in idpf_fill_rxq_config_chunk() 1946 qi->rx_bufq1_id = cpu_to_le16(sets[0].bufq.q_id); in idpf_fill_rxq_config_chunk() 1949 qi->rx_bufq2_id = cpu_to_le16(sets[1].bufq.q_id); in idpf_fill_rxq_config_chunk() 1952 q->rx_hbuf_size = sets[0].bufq.rx_hbuf_size; in idpf_fill_rxq_config_chunk() 2046 idpf_fill_bufq_config_chunk(qs->qv_rsrc, qs->qs[i].bufq, in idpf_send_config_rx_queue_set_msg() 2086 qs->qs[k++].bufq = &rx_qgrp->splitq.bufq_sets[j].bufq; in idpf_send_config_rx_queues_msg() 2183 qid = q->bufq->q_id; in idpf_send_ena_dis_queue_set_msg() 2269 qs->qs[k++].bufq = &rx_qgrp->splitq.bufq_sets[j].bufq; in idpf_send_ena_dis_queues_msg() 3962 q = &rx_qgrp->splitq.bufq_sets[j].bufq; in __idpf_vport_queue_ids_init()
|
| H A D | xdp.c | 55 frag_size = rxq->bufq_sets[0].bufq.truesize; in __idpf_xdp_rxq_info_init() 75 pp = split ? rxq->bufq_sets[0].bufq.pp : rxq->pp; in __idpf_xdp_rxq_info_init()
|
| H A D | idpf_txrx.h | 405 struct idpf_buf_queue **bufq; member 917 struct idpf_buf_queue bufq; member
|
| H A D | idpf_lib.c | 1472 &grp->splitq.bufq_sets[j].bufq; in idpf_rx_init_buf_tail()
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_rx_dqo.c | 24 int buf_count = rx->dqo.bufq.mask + 1; in gve_rx_free_hdr_bufs() 40 rx->dqo.bufq.mask = buffer_queue_slots - 1; in gve_rx_init_ring_state_dqo() 41 rx->dqo.bufq.head = 0; in gve_rx_init_ring_state_dqo() 42 rx->dqo.bufq.tail = 0; in gve_rx_init_ring_state_dqo() 78 if (rx->dqo.bufq.desc_ring) { in gve_rx_reset_ring_dqo() 79 size = sizeof(rx->dqo.bufq.desc_ring[0]) * in gve_rx_reset_ring_dqo() 81 memset(rx->dqo.bufq.desc_ring, 0, size); in gve_rx_reset_ring_dqo() 138 buffer_queue_slots = rx->dqo.bufq.mask + 1; in gve_rx_free_ring_dqo() 165 if (rx->dqo.bufq.desc_ring) { in gve_rx_free_ring_dqo() 166 size = sizeof(rx->dqo.bufq.desc_ring[0]) * buffer_queue_slots; in gve_rx_free_ring_dqo() [all …]
|
| H A D | gve.h | 265 struct gve_rx_buf_queue_dqo bufq; member
|