Home
last modified time | relevance | path

Searched refs:qpl (Results 1 – 7 of 7) sorted by relevance

/linux/drivers/net/ethernet/google/gve/
H A Dgve_main.c767 struct gve_queue_page_list *qpl) in gve_unregister_qpl() argument
771 if (!qpl) in gve_unregister_qpl()
774 err = gve_adminq_unregister_page_list(priv, qpl->id); in gve_unregister_qpl()
778 qpl->id); in gve_unregister_qpl()
782 priv->num_registered_pages -= qpl->num_entries; in gve_unregister_qpl()
787 struct gve_queue_page_list *qpl) in gve_register_qpl() argument
792 if (!qpl) in gve_register_qpl()
795 pages = qpl->num_entries; in gve_register_qpl()
805 err = gve_adminq_register_page_list(priv, qpl); in gve_register_qpl()
809 qpl->id); in gve_register_qpl()
[all …]
H A Dgve_tx.c41 fifo->base = vmap(fifo->qpl->pages, fifo->qpl->num_entries, VM_MAP, in gve_tx_fifo_init()
45 fifo->qpl->id); in gve_tx_fifo_init()
49 fifo->size = fifo->qpl->num_entries * PAGE_SIZE; in gve_tx_fifo_init()
231 if (tx->tx_fifo.qpl) { in gve_tx_free_ring_gqi()
236 gve_free_queue_page_list(priv, tx->tx_fifo.qpl, qpl_id); in gve_tx_free_ring_gqi()
237 tx->tx_fifo.qpl = NULL; in gve_tx_free_ring_gqi()
296 tx->tx_fifo.qpl = gve_alloc_queue_page_list(priv, qpl_id, in gve_tx_alloc_ring_gqi()
298 if (!tx->tx_fifo.qpl) in gve_tx_alloc_ring_gqi()
321 gve_free_queue_page_list(priv, tx->tx_fifo.qpl, qpl_id); in gve_tx_alloc_ring_gqi()
322 tx->tx_fifo.qpl = NULL; in gve_tx_alloc_ring_gqi()
[all …]
H A Dgve_tx_dqo.c23 if (!tx->dqo.qpl) in gve_has_free_tx_qpl_bufs()
199 if (tx->dqo.qpl) in gve_tx_clean_pending_packets()
264 if (tx->dqo.qpl) { in gve_tx_free_ring_dqo()
266 gve_free_queue_page_list(priv, tx->dqo.qpl, qpl_id); in gve_tx_free_ring_dqo()
267 tx->dqo.qpl = NULL; in gve_tx_free_ring_dqo()
276 tx->dqo.qpl->num_entries; in gve_tx_qpl_buf_init()
397 tx->dqo.qpl = gve_alloc_queue_page_list(priv, qpl_id, in gve_tx_alloc_ring_dqo()
399 if (!tx->dqo.qpl) in gve_tx_alloc_ring_dqo()
761 *va = page_address(tx->dqo.qpl->pages[page_id]) + offset; in gve_tx_buf_get_addr()
762 *dma_addr = tx->dqo.qpl->page_buses[page_id] + offset; in gve_tx_buf_get_addr()
[all …]
H A Dgve_rx.c135 if (rx->data.qpl) { in gve_rx_free_ring_gqi()
137 gve_free_queue_page_list(priv, rx->data.qpl, qpl_id); in gve_rx_free_ring_gqi()
138 rx->data.qpl = NULL; in gve_rx_free_ring_gqi()
202 struct page *page = rx->data.qpl->pages[i]; in gve_rx_prefill_pages()
319 rx->data.qpl = gve_alloc_queue_page_list(priv, qpl_id, in gve_rx_alloc_ring_gqi()
321 if (!rx->data.qpl) { in gve_rx_alloc_ring_gqi()
372 gve_free_queue_page_list(priv, rx->data.qpl, qpl_id); in gve_rx_alloc_ring_gqi()
373 rx->data.qpl = NULL; in gve_rx_alloc_ring_gqi()
851 rx->data.qpl->page_buses[idx]; in gve_rx()
H A Dgve.h142 struct gve_queue_page_list *qpl; /* qpl assigned to this queue */ member
297 struct gve_queue_page_list *qpl; member
389 struct gve_queue_page_list *qpl; /* QPL mapped into this FIFO */ member
629 struct gve_queue_page_list *qpl; member
1227 struct gve_queue_page_list *qpl,
H A Dgve_rx_dqo.c159 if (rx->dqo.qpl) { in gve_rx_free_ring_dqo()
161 gve_free_queue_page_list(priv, rx->dqo.qpl, qpl_id); in gve_rx_free_ring_dqo()
162 rx->dqo.qpl = NULL; in gve_rx_free_ring_dqo()
286 rx->dqo.qpl = gve_alloc_queue_page_list(priv, qpl_id, in gve_rx_alloc_ring_dqo()
288 if (!rx->dqo.qpl) in gve_rx_alloc_ring_dqo()
511 if (!rx->dqo.qpl) in gve_rx_should_trigger_copy_ondemand()
H A Dgve_adminq.h639 struct gve_queue_page_list *qpl);