Home
last modified time | relevance | path

Searched refs:tx_queues (Results 1 – 25 of 37) sorted by relevance

12

/freebsd/sys/dev/ice/
H A Dif_ice_iflib.c945 vsi->tx_queues[i].irqv = NULL; in ice_free_irqvs()
1078 if (!(vsi->tx_queues = in ice_if_tx_queues_alloc()
1085 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_if_tx_queues_alloc()
1109 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_if_tx_queues_alloc()
1130 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_if_tx_queues_alloc()
1136 free(vsi->tx_queues, M_ICE); in ice_if_tx_queues_alloc()
1137 vsi->tx_queues = NULL; in ice_if_tx_queues_alloc()
1249 if (vsi->tx_queues != NULL) { in ice_if_queues_free()
1251 for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++) { in ice_if_queues_free()
1257 free(vsi->tx_queues, M_ICE); in ice_if_queues_free()
[all …]
H A Dice_iov.c279 vsi->tx_queues = (struct ice_tx_queue *) in ice_iov_add_vf()
282 if (!vsi->tx_queues) { in ice_iov_add_vf()
288 for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++) { in ice_iov_add_vf()
408 free(vsi->tx_queues, M_ICE); in ice_iov_add_vf()
409 vsi->tx_queues = NULL; in ice_iov_add_vf()
451 if (vsi->tx_queues) { in ice_iov_uninit()
452 free(vsi->tx_queues, M_ICE); in ice_iov_uninit()
453 vsi->tx_queues = NULL; in ice_iov_uninit()
995 txq = &vsi->tx_queues[i]; in ice_vc_cfg_vsi_qs_msg()
1025 txq = &vsi->tx_queues[vqpi->txq.queue_id]; in ice_vc_cfg_vsi_qs_msg()
[all …]
H A Dice_iflib_txrx.c211 struct ice_tx_queue *txq = &sc->pf_vsi.tx_queues[pi->ipi_qsidx]; in ice_ift_txd_credits_update()
229 struct ice_tx_queue *txq = &sc->pf_vsi.tx_queues[txqid]; in ice_ift_txd_credits_update()
308 struct ice_tx_queue *txq = &sc->pf_vsi.tx_queues[txqid]; in ice_ift_rxd_pkt_get()
625 struct ice_tx_queue *txq = &mif->vsi->tx_queues[txqid];
644 struct ice_tx_queue *txq = &mif->vsi->tx_queues[pi->ipi_qsidx];
663 struct ice_tx_queue *txq = &mif->vsi->tx_queues[txqid];
/freebsd/sys/dev/iavf/
H A Dif_iavf_iflib.c687 struct iavf_tx_queue *tx_que = vsi->tx_queues; in iavf_init_queues()
823 struct iavf_tx_queue *tx_que = vsi->tx_queues; in iavf_if_msix_intr_assign()
943 struct iavf_tx_queue *tx_que = &vsi->tx_queues[txqid]; in iavf_if_tx_queue_intr_enable()
978 if (!(vsi->tx_queues = in iavf_if_tx_queues_alloc()
984 for (i = 0, que = vsi->tx_queues; i < ntxqsets; i++, que++) { in iavf_if_tx_queues_alloc()
1093 for (i = 0, que = vsi->tx_queues; i < vsi->shared->isc_ntxqsets; i++, que++) { in iavf_if_queues_free()
1102 if (vsi->tx_queues != NULL) { in iavf_if_queues_free()
1103 free(vsi->tx_queues, M_IAVF); in iavf_if_queues_free()
1104 vsi->tx_queues = NULL; in iavf_if_queues_free()
1844 for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) { in iavf_init_tx_rsqs()
[all …]
H A Diavf_txrx_iflib.c324 struct iavf_tx_queue *que = &vsi->tx_queues[pi->ipi_qsidx]; in iavf_isc_txd_encap()
413 struct tx_ring *txr = &vsi->tx_queues[txqid].txr; in iavf_isc_txd_flush()
478 struct iavf_tx_queue *que = &vsi->tx_queues[qid]; in iavf_isc_txd_credits_update_hwb()
513 struct iavf_tx_queue *tx_que = &vsi->tx_queues[txqid]; in iavf_isc_txd_credits_update_dwb()
H A Diavf_vc_common.c315 vqs.tx_queues = (1 << IAVF_NTXQS(vsi)) - 1; in iavf_enable_queues()
316 vqs.rx_queues = vqs.tx_queues; in iavf_enable_queues()
339 vqs.tx_queues = (1 << IAVF_NTXQS(vsi)) - 1; in iavf_disable_queues()
340 vqs.rx_queues = vqs.tx_queues; in iavf_disable_queues()
H A Diavf_vc_iflib.c58 struct iavf_tx_queue *tx_que = vsi->tx_queues; in iavf_configure_queues()
H A Diavf_iflib.h252 struct iavf_tx_queue *tx_queues; member
/freebsd/sys/dev/ixl/
H A Dixl_txrx.c342 struct ixl_tx_queue *que = &vsi->tx_queues[pi->ipi_qsidx]; in ixl_isc_txd_encap()
417 struct tx_ring *txr = &vsi->tx_queues[txqid].txr; in ixl_isc_txd_flush()
466 struct ixl_tx_queue *que = &vsi->tx_queues[qid]; in ixl_isc_txd_credits_update_hwb()
486 struct ixl_tx_queue *tx_que = &vsi->tx_queues[txqid]; in ixl_isc_txd_credits_update_dwb()
791 for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) { in ixl_init_tx_rsqs()
815 for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) { in ixl_init_tx_cidx()
948 tx_que = &(vsi->tx_queues[q]); in ixl_vsi_add_queues_stats()
H A Dif_ixl.c1078 struct ixl_tx_queue *tx_que = vsi->tx_queues; in ixl_if_msix_intr_assign()
1213 struct ixl_tx_queue *tx_que = &vsi->tx_queues[txqid]; in ixl_if_tx_queue_intr_enable()
1233 if (!(vsi->tx_queues = in ixl_if_tx_queues_alloc()
1239 for (i = 0, que = vsi->tx_queues; i < ntxqsets; i++, que++) { in ixl_if_tx_queues_alloc()
1318 if (vsi->tx_queues != NULL && !vsi->enable_head_writeback) { in ixl_if_queues_free()
1322 for (i = 0, que = vsi->tx_queues; i < vsi->num_tx_queues; i++, que++) { in ixl_if_queues_free()
1331 if (vsi->tx_queues != NULL) { in ixl_if_queues_free()
1332 free(vsi->tx_queues, M_IXL); in ixl_if_queues_free()
1333 vsi->tx_queues = NULL; in ixl_if_queues_free()
H A Dvirtchnl.h397 u32 tx_queues; member
/freebsd/sys/dev/e1000/
H A Dem_txrx.c102 que = &sc->tx_queues[qid]; in em_dump_rs()
139 struct em_tx_queue *que = &sc->tx_queues[pi->ipi_qsidx]; in em_tso_setup()
246 struct em_tx_queue *que = &sc->tx_queues[pi->ipi_qsidx]; in em_transmit_checksum_setup()
346 struct em_tx_queue *que = &sc->tx_queues[pi->ipi_qsidx]; in em_isc_txd_encap()
475 struct em_tx_queue *que = &sc->tx_queues[txqid]; in em_isc_txd_flush()
486 struct em_tx_queue *que = &sc->tx_queues[txqid]; in em_isc_txd_credits_update()
H A Digb_txrx.c237 struct em_tx_queue *que = &sc->tx_queues[pi->ipi_qsidx]; in igb_isc_txd_encap()
303 struct em_tx_queue *que = &sc->tx_queues[txqid]; in igb_isc_txd_flush()
314 struct em_tx_queue *que = &sc->tx_queues[txqid]; in igb_isc_txd_credits_update()
H A Dif_em.c1567 for (i = 0, tx_que = sc->tx_queues; i < sc->tx_num_queues; in em_if_init()
1814 struct tx_ring *txr = &sc->tx_queues[0].txr; in em_intr()
1878 struct em_tx_queue *txq = &sc->tx_queues[txqid]; in em_if_tx_queue_intr_enable()
1898 struct em_tx_queue *txq = &sc->tx_queues[txqid]; in igb_if_tx_queue_intr_enable()
1914 struct tx_ring *txr = &sc->tx_queues[que->msix].txr; in em_msix_que()
2464 struct em_tx_queue *tx_que = sc->tx_queues; in em_if_msix_intr_assign()
2504 tx_que = &sc->tx_queues[i]; in em_if_msix_intr_assign()
2597 tx_que = &sc->tx_queues[i]; in igb_configure_queues()
2637 tx_que = &sc->tx_queues[i]; in igb_configure_queues()
2934 struct tx_ring *txr = &sc->tx_queues in em_flush_tx_ring()
[all...]
H A Dif_em.h486 struct em_tx_queue *tx_queues; member
/freebsd/sys/dev/enetc/
H A Dif_enetc.c488 sc->tx_queues = mallocarray(sc->tx_num_queues, in enetc_tx_queues_alloc()
490 if (sc->tx_queues == NULL) { in enetc_tx_queues_alloc()
497 queue = &sc->tx_queues[i]; in enetc_tx_queues_alloc()
544 if (sc->tx_queues != NULL) { in enetc_queues_free()
545 free(sc->tx_queues, M_DEVBUF); in enetc_queues_free()
546 sc->tx_queues = NULL; in enetc_queues_free()
745 queue = &sc->tx_queues[i]; in enetc_init_tx()
1024 tx_queue = &sc->tx_queues[i]; in enetc_msix_intr_assign()
1100 queue = &sc->tx_queues[ipi->ipi_qsidx]; in enetc_isc_txd_encap()
1164 queue = &sc->tx_queues[qid]; in enetc_isc_txd_credits_update()
H A Denetc.h82 struct enetc_tx_queue *tx_queues; member
/freebsd/sys/dev/igc/
H A Digc_txrx.c91 que = &sc->tx_queues[qid]; in igc_dump_rs()
269 struct igc_tx_queue *que = &sc->tx_queues[pi->ipi_qsidx]; in igc_isc_txd_encap()
331 struct igc_tx_queue *que = &sc->tx_queues[txqid]; in igc_isc_txd_flush()
342 struct igc_tx_queue *que = &sc->tx_queues[txqid]; in igc_isc_txd_credits_update()
H A Dif_igc.c848 for (i = 0, tx_que = sc->tx_queues; i < sc->tx_num_queues; in igc_if_init()
1050 struct tx_ring *txr = &sc->tx_queues[0].txr; in igc_intr()
1108 struct igc_tx_queue *txq = &sc->tx_queues[txqid]; in igc_if_tx_queue_intr_enable()
1124 struct tx_ring *txr = &sc->tx_queues[que->msix].txr; in igc_msix_que()
1527 struct igc_tx_queue *tx_que = sc->tx_queues; in igc_if_msix_intr_assign()
1560 tx_que = &sc->tx_queues[i]; in igc_if_msix_intr_assign()
1628 tx_que = &sc->tx_queues[i]; in igc_configure_queues()
2013 if (!(sc->tx_queues = in igc_if_tx_queues_alloc()
2021 for (i = 0, que = sc->tx_queues; i < sc->tx_num_queues; i++, que++) { in igc_if_tx_queues_alloc()
2100 struct igc_tx_queue *tx_que = sc->tx_queues; in igc_if_queues_free()
[all …]
H A Dif_igc.h311 struct igc_tx_queue *tx_queues; member
/freebsd/sys/dev/ixgbe/
H A Dif_ixv.c259 sc->tx_queues = in ixv_if_tx_queues_alloc()
262 if (!sc->tx_queues) { in ixv_if_tx_queues_alloc()
268 for (i = 0, que = sc->tx_queues; i < ntxqsets; i++, que++) { in ixv_if_tx_queues_alloc()
360 struct ix_tx_queue *que = sc->tx_queues; in ixv_if_queues_free()
374 if (sc->tx_queues != NULL) in ixv_if_queues_free()
375 free(sc->tx_queues, M_DEVBUF); in ixv_if_queues_free()
379 sc->tx_queues = NULL; in ixv_if_queues_free()
1069 tx_que = &sc->tx_queues[i]; in ixv_if_msix_intr_assign()
1246 struct ix_tx_queue *que = sc->tx_queues; in ixv_initialize_transmit_units()
1837 struct ix_tx_queue *tx_que = sc->tx_queues; in ixv_add_stats_sysctls()
H A Dix_txrx.c177 struct ix_tx_queue *que = &sc->tx_queues[pi->ipi_qsidx]; in ixgbe_isc_txd_encap()
257 struct ix_tx_queue *que = &sc->tx_queues[txqid]; in ixgbe_isc_txd_flush()
271 struct ix_tx_queue *que = &sc->tx_queues[txqid]; in ixgbe_isc_txd_credits_update()
H A Dif_ix.c494 sc->tx_queues = in ixgbe_if_tx_queues_alloc()
497 if (!sc->tx_queues) { in ixgbe_if_tx_queues_alloc()
503 for (i = 0, que = sc->tx_queues; i < ntxqsets; i++, que++) { in ixgbe_if_tx_queues_alloc()
599 struct ix_tx_queue *tx_que = sc->tx_queues; in ixgbe_if_queues_free()
613 free(sc->tx_queues, M_IXGBE); in ixgbe_if_queues_free()
614 sc->tx_queues = NULL; in ixgbe_if_queues_free()
863 for (i = 0, que = sc->tx_queues; i < sc->num_tx_queues; in ixgbe_initialize_transmit_units()
2029 for (i = 0, tx_que = sc->tx_queues; i < sc->num_tx_queues; in ixgbe_add_hw_stats()
2560 tx_que = &sc->tx_queues[i]; in ixgbe_if_msix_intr_assign()
3851 for (i = 0, tx_que = sc->tx_queues; i < sc->num_tx_queues; in ixgbe_if_init()
[all …]
H A Dixgbe.h433 struct ix_tx_queue *tx_queues; member
/freebsd/sys/dev/vnic/
H A Dnic.h308 uint8_t tx_queues; member

12