Lines Matching refs:txq
70 struct netdev_queue *txq) in netpoll_start_xmit() argument
89 status = netdev_start_xmit(skb, dev, txq, false); in netpoll_start_xmit()
102 while ((skb = skb_dequeue(&npinfo->txq))) { in queue_process()
104 struct netdev_queue *txq; in queue_process() local
119 txq = netdev_get_tx_queue(dev, q_index); in queue_process()
120 HARD_TX_LOCK(dev, txq, smp_processor_id()); in queue_process()
121 if (netif_xmit_frozen_or_stopped(txq) || in queue_process()
122 !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) { in queue_process()
123 skb_queue_head(&npinfo->txq, skb); in queue_process()
124 HARD_TX_UNLOCK(dev, txq); in queue_process()
130 HARD_TX_UNLOCK(dev, txq); in queue_process()
140 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_local_xmit_active() local
142 if (READ_ONCE(txq->xmit_lock_owner) == smp_processor_id()) in netif_local_xmit_active()
340 if (skb_queue_len(&npinfo->txq) == 0 && !netpoll_owner_active(dev)) { in __netpoll_send_skb()
341 struct netdev_queue *txq; in __netpoll_send_skb() local
343 txq = netdev_core_pick_tx(dev, skb, NULL); in __netpoll_send_skb()
348 if (HARD_TX_TRYLOCK(dev, txq)) { in __netpoll_send_skb()
349 if (!netif_xmit_stopped(txq)) in __netpoll_send_skb()
350 status = netpoll_start_xmit(skb, dev, txq); in __netpoll_send_skb()
352 HARD_TX_UNLOCK(dev, txq); in __netpoll_send_skb()
372 skb_queue_tail(&npinfo->txq, skb); in __netpoll_send_skb()
653 skb_queue_head_init(&npinfo->txq); in __netpoll_setup()
817 skb_queue_purge(&npinfo->txq); in rcu_cleanup_netpoll_info()
823 __skb_queue_purge(&npinfo->txq); in rcu_cleanup_netpoll_info()