Home
last modified time | relevance | path

Searched refs:bufq (Results 1 – 8 of 8) sorted by relevance

/linux/drivers/net/ethernet/intel/idpf/
H A Dxsk.c27 struct idpf_buf_queue *bufq) in idpf_xsk_setup_bufq() argument
37 if (&grp->splitq.bufq_sets[j].bufq == bufq) { in idpf_xsk_setup_bufq()
49 bufq->pool = pool; in idpf_xsk_setup_bufq()
51 idpf_queue_set(XSK, bufq); in idpf_xsk_setup_bufq()
127 struct idpf_buf_queue *bufq; in idpf_xsk_clear_queue() local
140 bufq = q; in idpf_xsk_clear_queue()
141 if (!idpf_queue_has_clear(XSK, bufq)) in idpf_xsk_clear_queue()
144 bufq->pool = NULL; in idpf_xsk_clear_queue()
355 static bool idpf_xskfq_refill_thresh(struct idpf_buf_queue *bufq, u32 count) in idpf_xskfq_refill_thresh() argument
358 .pool = bufq->pool, in idpf_xskfq_refill_thresh()
[all …]
H A Didpf_txrx.c362 static void idpf_rx_hdr_buf_rel_all(struct idpf_buf_queue *bufq) in idpf_rx_hdr_buf_rel_all() argument
365 .fqes = bufq->hdr_buf, in idpf_rx_hdr_buf_rel_all()
366 .pp = bufq->hdr_pp, in idpf_rx_hdr_buf_rel_all()
369 for (u32 i = 0; i < bufq->desc_count; i++) in idpf_rx_hdr_buf_rel_all()
370 idpf_rx_page_rel(&bufq->hdr_buf[i]); in idpf_rx_hdr_buf_rel_all()
373 bufq->hdr_buf = NULL; in idpf_rx_hdr_buf_rel_all()
374 bufq->hdr_pp = NULL; in idpf_rx_hdr_buf_rel_all()
381 static void idpf_rx_buf_rel_bufq(struct idpf_buf_queue *bufq) in idpf_rx_buf_rel_bufq() argument
384 .fqes = bufq->buf, in idpf_rx_buf_rel_bufq()
385 .pp = bufq->pp, in idpf_rx_buf_rel_bufq()
[all …]
H A Didpf_virtchnl.c1449 q = &rx_qgrp->splitq.bufq_sets[j].bufq; in __idpf_queue_reg_init()
1943 q->rx_buf_size = sets[0].bufq.rx_buf_size; in idpf_fill_rxq_config_chunk()
1946 qi->rx_bufq1_id = cpu_to_le16(sets[0].bufq.q_id); in idpf_fill_rxq_config_chunk()
1949 qi->rx_bufq2_id = cpu_to_le16(sets[1].bufq.q_id); in idpf_fill_rxq_config_chunk()
1952 q->rx_hbuf_size = sets[0].bufq.rx_hbuf_size; in idpf_fill_rxq_config_chunk()
2046 idpf_fill_bufq_config_chunk(qs->qv_rsrc, qs->qs[i].bufq, in idpf_send_config_rx_queue_set_msg()
2086 qs->qs[k++].bufq = &rx_qgrp->splitq.bufq_sets[j].bufq; in idpf_send_config_rx_queues_msg()
2183 qid = q->bufq->q_id; in idpf_send_ena_dis_queue_set_msg()
2269 qs->qs[k++].bufq = &rx_qgrp->splitq.bufq_sets[j].bufq; in idpf_send_ena_dis_queues_msg()
3962 q = &rx_qgrp->splitq.bufq_sets[j].bufq; in __idpf_vport_queue_ids_init()
H A Dxdp.c55 frag_size = rxq->bufq_sets[0].bufq.truesize; in __idpf_xdp_rxq_info_init()
75 pp = split ? rxq->bufq_sets[0].bufq.pp : rxq->pp; in __idpf_xdp_rxq_info_init()
H A Didpf_txrx.h405 struct idpf_buf_queue **bufq; member
917 struct idpf_buf_queue bufq; member
H A Didpf_lib.c1472 &grp->splitq.bufq_sets[j].bufq; in idpf_rx_init_buf_tail()
/linux/drivers/net/ethernet/google/gve/
H A Dgve_rx_dqo.c24 int buf_count = rx->dqo.bufq.mask + 1; in gve_rx_free_hdr_bufs()
40 rx->dqo.bufq.mask = buffer_queue_slots - 1; in gve_rx_init_ring_state_dqo()
41 rx->dqo.bufq.head = 0; in gve_rx_init_ring_state_dqo()
42 rx->dqo.bufq.tail = 0; in gve_rx_init_ring_state_dqo()
78 if (rx->dqo.bufq.desc_ring) { in gve_rx_reset_ring_dqo()
79 size = sizeof(rx->dqo.bufq.desc_ring[0]) * in gve_rx_reset_ring_dqo()
81 memset(rx->dqo.bufq.desc_ring, 0, size); in gve_rx_reset_ring_dqo()
138 buffer_queue_slots = rx->dqo.bufq.mask + 1; in gve_rx_free_ring_dqo()
165 if (rx->dqo.bufq.desc_ring) { in gve_rx_free_ring_dqo()
166 size = sizeof(rx->dqo.bufq.desc_ring[0]) * buffer_queue_slots; in gve_rx_free_ring_dqo()
[all …]
H A Dgve.h265 struct gve_rx_buf_queue_dqo bufq; member