| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_txrx.c | 143 static struct netdev_queue *txring_txq(const struct ice_tx_ring *ring) in txring_txq() function 236 netdev_tx_reset_queue(txring_txq(tx_ring)); in ice_clean_tx_ring() 282 netdev_txq_bql_complete_prefetchw(txring_txq(tx_ring)); in ice_clean_tx_irq() 371 netdev_tx_completed_queue(txring_txq(tx_ring), total_pkts, total_bytes); in ice_clean_tx_irq() 380 if (netif_tx_queue_stopped(txring_txq(tx_ring)) && in ice_clean_tx_irq() 382 netif_tx_wake_queue(txring_txq(tx_ring)); in ice_clean_tx_irq() 1358 netif_tx_stop_queue(txring_txq(tx_ring)); in __ice_maybe_stop_tx() 1367 netif_tx_start_queue(txring_txq(tx_ring)); in __ice_maybe_stop_tx() 1519 kick = __netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount, in ice_tx_map() 2194 netdev_txq_bql_enqueue_prefetchw(txring_txq(tx_ring)); in ice_xmit_frame_ring()
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_tsn.c | 83 netdev_tx_sent_queue(txring_txq(ring), skb->len); in igc_fpe_init_tx_descriptor() 101 nq = txring_txq(ring); in igc_fpe_xmit_smd_frame()
|
| H A D | igc_main.c | 255 netdev_tx_reset_queue(txring_txq(tx_ring)); in igc_clean_tx_ring() 1134 netdev_tx_sent_queue(txring_txq(ring), skb->len); in igc_init_tx_empty_descriptor() 1393 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in igc_tx_map() 1421 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in igc_tx_map() 2449 netdev_tx_sent_queue(txring_txq(ring), head->bytecount); in igc_xdp_init_tx_descriptor() 2503 nq = txring_txq(ring); in igc_xdp_xmit_back() 2580 nq = txring_txq(ring); in igc_finalize_xdp() 3040 struct netdev_queue *nq = txring_txq(ring); in igc_xdp_xmit_zc() 3109 netdev_tx_sent_queue(txring_txq(ring), xdp_desc.len); in igc_xdp_xmit_zc() 3235 netdev_tx_completed_queue(txring_txq(tx_ring), in igc_clean_tx_irq() [all …]
|
| H A D | igc.h | 708 static inline struct netdev_queue *txring_txq(const struct igc_ring *tx_ring) in txring_txq() function
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.h | 388 static inline struct netdev_queue *txring_txq(const struct iavf_ring *ring) in txring_txq() function
|
| H A D | iavf_txrx.c | 105 netdev_tx_reset_queue(txring_txq(tx_ring)); in iavf_clean_tx_ring() 347 netdev_tx_completed_queue(txring_txq(tx_ring), in iavf_clean_tx_irq() 2232 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in iavf_tx_map() 2261 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in iavf_tx_map()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_txrx.h | 562 static inline struct netdev_queue *txring_txq(const struct i40e_ring *ring) in txring_txq() function
|
| H A D | i40e_txrx.c | 814 netdev_tx_reset_queue(txring_txq(tx_ring)); in i40e_clean_tx_ring() 1028 netdev_tx_completed_queue(txring_txq(tx_ring), in i40e_clean_tx_irq() 3655 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in i40e_tx_map() 3696 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in i40e_tx_map()
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| H A D | fm10k_main.c | 1006 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in fm10k_tx_map() 1029 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in fm10k_tx_map() 1306 netdev_tx_completed_queue(txring_txq(tx_ring), in fm10k_clean_tx_irq()
|
| H A D | fm10k.h | 167 static inline struct netdev_queue *txring_txq(const struct fm10k_ring *ring) in txring_txq() function
|
| H A D | fm10k_netdev.c | 187 netdev_tx_reset_queue(txring_txq(tx_ring)); in fm10k_clean_tx_ring()
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 521 netdev_tx_sent_queue(txring_txq(tx_ring), total_bytes); in igb_xmit_zc()
|
| H A D | igb_main.c | 2975 nq = txring_txq(tx_ring); in igb_xdp_xmit_back() 3012 nq = txring_txq(tx_ring); in igb_xdp_xmit() 4994 netdev_tx_reset_queue(txring_txq(tx_ring)); in igb_clean_tx_ring() 6352 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in igb_tx_map() 6380 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in igb_tx_map() 6487 netdev_tx_sent_queue(txring_txq(tx_ring), tx_head->bytecount); in igb_xmit_xdp_ring() 6501 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) in igb_xmit_xdp_ring() 8428 netdev_tx_completed_queue(txring_txq(tx_ring), in igb_clean_tx_irq() 8446 nq = txring_txq(tx_ring); in igb_clean_tx_irq() 8995 nq = txring_txq(tx_ring); in igb_finalize_xdp()
|
| H A D | igb_ethtool.c | 1877 netdev_tx_reset_queue(txring_txq(tx_ring)); in igb_clean_test_rings()
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| H A D | fbnic_txrx.c | 115 static struct netdev_queue *txring_txq(const struct net_device *dev, in txring_txq() function 125 struct netdev_queue *txq = txring_txq(dev, ring); in fbnic_maybe_stop_tx() 141 struct netdev_queue *dev_queue = txring_txq(skb->dev, ring); in fbnic_tx_sent_queue() 606 txq = txring_txq(nv->napi.dev, ring); in fbnic_clean_twq0()
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe.h | 1055 static inline struct netdev_queue *txring_txq(const struct ixgbe_ring *ring) in txring_txq() function
|
| H A D | ixgbe_main.c | 6646 netdev_tx_reset_queue(txring_txq(tx_ring)); in ixgbe_clean_tx_ring() 9038 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in ixgbe_tx_map() 9066 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in ixgbe_tx_map()
|
| H A D | ixgbe_ethtool.c | 2151 netdev_tx_reset_queue(txring_txq(tx_ring)); in ixgbe_clean_test_rings()
|