| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | ipoib_tx.c | 29 struct hfi1_ipoib_txq *txq; member 47 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_used() argument 49 return hfi1_ipoib_txreqs(txq->tx_ring.sent_txreqs, in hfi1_ipoib_used() 50 txq->tx_ring.complete_txreqs); in hfi1_ipoib_used() 53 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_stop_txq() argument 55 trace_hfi1_txq_stop(txq); in hfi1_ipoib_stop_txq() 56 if (atomic_inc_return(&txq->tx_ring.stops) == 1) in hfi1_ipoib_stop_txq() 57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_stop_txq() 60 static void hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) in hfi1_ipoib_wake_txq() argument 62 trace_hfi1_txq_wake(txq); in hfi1_ipoib_wake_txq() [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/ |
| H A D | tx.c | 81 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr() argument 84 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr() 86 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr() 109 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr() 118 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr() 119 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr() 121 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr() 130 struct iwl_txq *txq = trans_pcie->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local 135 spin_lock_bh(&txq->lock); in iwl_pcie_txq_check_wrptrs() 136 if (txq->need_update) { in iwl_pcie_txq_check_wrptrs() [all …]
|
| H A D | tx-gen2.c | 297 struct iwl_txq *txq, in iwl_txq_gen2_build_tx_amsdu() argument 304 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_build_tx_amsdu() 305 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx); in iwl_txq_gen2_build_tx_amsdu() 310 tb_phys = iwl_txq_get_first_tb_dma(txq, idx); in iwl_txq_gen2_build_tx_amsdu() 346 memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); in iwl_txq_gen2_build_tx_amsdu() 385 struct iwl_txq *txq, in iwl_txq_gen2_build_tx() argument 393 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_build_tx() 394 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx); in iwl_txq_gen2_build_tx() 400 tb_phys = iwl_txq_get_first_tb_dma(txq, idx); in iwl_txq_gen2_build_tx() 403 memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); in iwl_txq_gen2_build_tx() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_tx.c | 77 static void hinic_txq_clean_stats(struct hinic_txq *txq) in hinic_txq_clean_stats() argument 79 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_clean_stats() 96 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) in hinic_txq_get_stats() argument 98 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in hinic_txq_get_stats() 116 static void txq_stats_init(struct hinic_txq *txq) in txq_stats_init() argument 118 struct hinic_txq_stats *txq_stats = &txq->txq_stats; in txq_stats_init() 121 hinic_txq_clean_stats(txq); in txq_stats_init() 499 struct hinic_txq *txq; in hinic_lb_xmit_frame() local 502 txq = &nic_dev->txqs[q_id]; in hinic_lb_xmit_frame() 503 qp = container_of(txq->sq, struct hinic_qp, sq); in hinic_lb_xmit_frame() [all …]
|
| /linux/drivers/net/wireless/ath/ath9k/ |
| H A D | xmit.c | 56 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, 59 int tx_flags, struct ath_txq *txq, 62 struct ath_txq *txq, struct list_head *bf_q, 65 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, 72 struct ath_txq *txq, 106 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq) in ath_txq_unlock_complete() argument 107 __releases(&txq->axq_lock) in ath_txq_unlock_complete() 114 skb_queue_splice_init(&txq->complete_q, &q); in ath_txq_unlock_complete() 115 spin_unlock_bh(&txq->axq_lock); in ath_txq_unlock_complete() 134 struct ath_txq *txq = tid->txq; in ath9k_wake_tx_queue() local [all …]
|
| H A D | ath9k.h | 179 s8 txq; member 238 struct ath_txq *txq; member 276 struct ath_txq *txq; member 291 struct ath_txq txq[ATH9K_NUM_TX_QUEUES]; member 433 struct ieee80211_txq *txq; in ath_node_to_tid() local 437 txq = sta->txq[tidno % ARRAY_SIZE(sta->txq)]; in ath_node_to_tid() 439 txq = vif->txq; in ath_node_to_tid() 441 return (struct ath_atx_tid *) txq->drv_priv; in ath_node_to_tid() 567 static inline void ath_txq_lock(struct ath_softc *sc, struct ath_txq *txq) in ath_txq_lock() argument 569 spin_lock_bh(&txq->axq_lock); in ath_txq_lock() [all …]
|
| /linux/drivers/net/ethernet/marvell/ |
| H A D | mv643xx_eth.c | 179 #define IS_TSO_HEADER(txq, addr) \ argument 180 ((addr >= txq->tso_hdrs_dma) && \ 181 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 408 struct tx_queue txq[8]; member 446 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) in txq_to_mp() argument 448 return container_of(txq, struct mv643xx_eth_private, txq[txq->index]); in txq_to_mp() 467 static void txq_reset_hw_ptr(struct tx_queue *txq) in txq_reset_hw_ptr() argument 469 struct mv643xx_eth_private *mp = txq_to_mp(txq); in txq_reset_hw_ptr() 472 addr = (u32)txq->tx_desc_dma; in txq_reset_hw_ptr() 473 addr += txq->tx_curr_desc * sizeof(struct tx_desc); in txq_reset_hw_ptr() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede_fp.c | 77 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) in qede_free_tx_pkt() argument 79 u16 idx = txq->sw_tx_cons; in qede_free_tx_pkt() 80 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 85 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 91 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt() 97 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq->tx_pbl); in qede_free_tx_pkt() 105 qed_chain_consume(&txq->tx_pbl); in qede_free_tx_pkt() 115 qed_chain_consume(&txq->tx_pbl); in qede_free_tx_pkt() 121 qed_chain_consume(&txq->tx_pbl); in qede_free_tx_pkt() 125 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_tx_pkt() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | selq.h | 30 static inline u16 mlx5e_txq_to_ch_ix(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix() argument 32 while (unlikely(txq >= num_channels)) in mlx5e_txq_to_ch_ix() 33 txq -= num_channels; in mlx5e_txq_to_ch_ix() 34 return txq; in mlx5e_txq_to_ch_ix() 37 static inline u16 mlx5e_txq_to_ch_ix_htb(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix_htb() argument 39 if (unlikely(txq >= num_channels)) { in mlx5e_txq_to_ch_ix_htb() 40 if (unlikely(txq >= num_channels << 3)) in mlx5e_txq_to_ch_ix_htb() 41 txq %= num_channels; in mlx5e_txq_to_ch_ix_htb() 44 txq -= num_channels; in mlx5e_txq_to_ch_ix_htb() 45 while (txq >= num_channels); in mlx5e_txq_to_ch_ix_htb() [all …]
|
| /linux/include/trace/events/ |
| H A D | qdisc.h | 16 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, 19 TP_ARGS(qdisc, txq, packets, skb), 23 __field(const struct netdev_queue *, txq ) 35 __entry->txq = txq; 38 __entry->ifindex = txq->dev ? txq->dev->ifindex : 0; 41 __entry->txq_state = txq->state; 51 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, struct sk_buff *skb), 53 TP_ARGS(qdisc, txq, sk [all...] |
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| H A D | sge.c | 1133 static void txq_stop(struct sge_eth_txq *txq) in txq_stop() argument 1135 netif_tx_stop_queue(txq->txq); in txq_stop() 1136 txq->q.stops++; in txq_stop() 1164 struct sge_eth_txq *txq; in t4vf_eth_xmit() local 1195 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit() 1205 reclaim_completed_tx(adapter, &txq->q, true); in t4vf_eth_xmit() 1214 credits = txq_avail(&txq->q) - ndesc; in t4vf_eth_xmit() 1223 txq_stop(txq); in t4vf_eth_xmit() 1237 txq->mapping_err++; in t4vf_eth_xmit() 1252 txq_stop(txq); in t4vf_eth_xmit() [all …]
|
| /linux/net/sched/ |
| H A D | sch_generic.c | 41 const struct netdev_queue *txq) in qdisc_maybe_clear_missed() argument 55 if (!netif_xmit_frozen_or_stopped(txq)) in qdisc_maybe_clear_missed() 76 const struct netdev_queue *txq = q->dev_queue; in __skb_dequeue_bad_txq() local 88 txq = skb_get_tx_queue(txq->dev, skb); in __skb_dequeue_bad_txq() 89 if (!netif_xmit_frozen_or_stopped(txq)) { in __skb_dequeue_bad_txq() 100 qdisc_maybe_clear_missed(q, txq); in __skb_dequeue_bad_txq() 182 const struct netdev_queue *txq, in try_bulk_dequeue_skb() argument 185 int bytelimit = qdisc_avail_bulklimit(txq) - skb->len; in try_bulk_dequeue_skb() 236 const struct netdev_queue *txq = q->dev_queue; in dequeue_skb() local 264 txq = skb_get_tx_queue(txq->dev, skb); in dequeue_skb() [all …]
|
| /linux/drivers/net/wireless/ath/ath5k/ |
| H A D | base.c | 769 struct ath5k_txq *txq, int padsize, in ath5k_txbuf_setup() argument 875 spin_lock_bh(&txq->lock); in ath5k_txbuf_setup() 876 list_add_tail(&bf->list, &txq->q); in ath5k_txbuf_setup() 877 txq->txq_len++; in ath5k_txbuf_setup() 878 if (txq->link == NULL) /* is this first packet? */ in ath5k_txbuf_setup() 879 ath5k_hw_set_txdp(ah, txq->qnum, bf->daddr); in ath5k_txbuf_setup() 881 *txq->link = bf->daddr; in ath5k_txbuf_setup() 883 txq->link = &ds->ds_link; in ath5k_txbuf_setup() 884 ath5k_hw_start_tx_dma(ah, txq->qnum); in ath5k_txbuf_setup() 885 spin_unlock_bh(&txq->lock); in ath5k_txbuf_setup() [all …]
|
| /linux/net/core/ |
| H A D | netpoll.c | 63 struct netdev_queue *txq) in netpoll_start_xmit() argument 82 status = netdev_start_xmit(skb, dev, txq, false); in netpoll_start_xmit() 95 while ((skb = skb_dequeue(&npinfo->txq))) { in queue_process() 97 struct netdev_queue *txq; in queue_process() local 112 txq = netdev_get_tx_queue(dev, q_index); in queue_process() 113 HARD_TX_LOCK(dev, txq, smp_processor_id()); in queue_process() 114 if (netif_xmit_frozen_or_stopped(txq) || in queue_process() 115 !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) { in queue_process() 116 skb_queue_head(&npinfo->txq, skb); in queue_process() 117 HARD_TX_UNLOCK(dev, txq); in queue_process() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | sge.c | 176 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset() 654 memset(q->txq, 0, sizeof(struct sge_txq) * SGE_TXQ_PER_SET); in t3_reset_qset() 691 if (q->txq[i].desc) { in t3_free_qset() 693 t3_sge_enable_ecntxt(adapter, q->txq[i].cntxt_id, 0); in t3_free_qset() 695 if (q->txq[i].sdesc) { in t3_free_qset() 696 free_tx_desc(adapter, &q->txq[i], in t3_free_qset() 697 q->txq[i].in_use); in t3_free_qset() 698 kfree(q->txq[i].sdesc); in t3_free_qset() 701 q->txq[i].size * in t3_free_qset() 703 q->txq[i].desc, q->txq[i].phys_addr); in t3_free_qset() [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/mld/ |
| H A D | tx.h | 46 iwl_mld_txq_from_mac80211(struct ieee80211_txq *txq) in iwl_mld_txq_from_mac80211() argument 48 return (void *)txq->drv_priv; in iwl_mld_txq_from_mac80211() 52 void iwl_mld_remove_txq(struct iwl_mld *mld, struct ieee80211_txq *txq); 56 void iwl_mld_tx_from_txq(struct iwl_mld *mld, struct ieee80211_txq *txq); 60 int iwl_mld_ensure_queue(struct iwl_mld *mld, struct ieee80211_txq *txq); 75 struct ieee80211_txq *txq);
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| H A D | common.c | 364 il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB; in il_send_cmd_sync() 2710 il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq) in il_txq_update_write_ptr() argument 2713 int txq_id = txq->q.id; in il_txq_update_write_ptr() 2715 if (txq->need_update == 0) in il_txq_update_write_ptr() 2733 il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); in il_txq_update_write_ptr() 2741 _il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); in il_txq_update_write_ptr() 2742 txq->need_update = 0; in il_txq_update_write_ptr() 2752 struct il_tx_queue *txq = &il->txq[txq_id]; in il_tx_queue_unmap() local 2753 struct il_queue *q = &txq->q; in il_tx_queue_unmap() 2759 il->ops->txq_free_tfd(il, txq); in il_tx_queue_unmap() [all …]
|
| H A D | 3945.c | 274 struct il_tx_queue *txq = &il->txq[txq_id]; in il3945_tx_queue_reclaim() local 275 struct il_queue *q = &txq->q; in il3945_tx_queue_reclaim() 283 skb = txq->skbs[txq->q.read_ptr]; in il3945_tx_queue_reclaim() 285 txq->skbs[txq->q.read_ptr] = NULL; in il3945_tx_queue_reclaim() 286 il->ops->txq_free_tfd(il, txq); in il3945_tx_queue_reclaim() 291 il_wake_queue(il, txq); in il3945_tx_queue_reclaim() 304 struct il_tx_queue *txq = &il->txq[txq_id]; in il3945_hdl_tx() local 311 if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) { in il3945_hdl_tx() 314 txq->q.n_bd, txq->q.write_ptr, txq->q.read_ptr); in il3945_hdl_tx() 331 txq->time_stamp = jiffies; in il3945_hdl_tx() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | sge.c | 1220 netif_tx_stop_queue(q->txq); in eth_txq_stop() 1411 if (!q->in_use || !__netif_tx_trylock(eq->txq)) in t4_sge_eth_txq_egress_update() 1427 if (netif_tx_queue_stopped(eq->txq) && hw_in_use < (q->size / 2)) { in t4_sge_eth_txq_egress_update() 1428 netif_tx_wake_queue(eq->txq); in t4_sge_eth_txq_egress_update() 1432 __netif_tx_unlock(eq->txq); in t4_sge_eth_txq_egress_update() 1819 struct sge_eth_txq *txq; in cxgb4_vf_eth_xmit() local 1845 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in cxgb4_vf_eth_xmit() 1850 reclaim_completed_tx(adapter, &txq->q, -1, true); in cxgb4_vf_eth_xmit() 1858 credits = txq_avail(&txq->q) - ndesc; in cxgb4_vf_eth_xmit() 1866 eth_txq_stop(txq); in cxgb4_vf_eth_xmit() [all …]
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_tx_rx.c | 2870 struct bna_txq *txq; in bna_tx_sm_started_entry() local 2873 list_for_each_entry(txq, &tx->txq_q, qe) { in bna_tx_sm_started_entry() 2874 txq->tcb->priority = txq->priority; in bna_tx_sm_started_entry() 2876 bna_ib_start(tx->bna, &txq->ib, is_regular); in bna_tx_sm_started_entry() 3089 struct bna_txq *txq = NULL; in bna_bfi_tx_enet_start() local 3099 txq = txq ? list_next_entry(txq, qe) in bna_bfi_tx_enet_start() 3101 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt); in bna_bfi_tx_enet_start() 3102 cfg_req->q_cfg[i].q.priority = txq->priority; in bna_bfi_tx_enet_start() 3105 txq->ib.ib_seg_host_addr.lsb; in bna_bfi_tx_enet_start() 3107 txq->ib.ib_seg_host_addr.msb; in bna_bfi_tx_enet_start() [all …]
|
| /linux/drivers/net/ethernet/alacritech/ |
| H A D | slicoss.c | 113 static unsigned int slic_get_free_tx_descs(struct slic_tx_queue *txq) in slic_get_free_tx_descs() argument 117 return slic_get_free_queue_descs(txq->put_idx, txq->done_idx, txq->len); in slic_get_free_tx_descs() 344 struct slic_tx_queue *txq = &sdev->txq; in slic_xmit_complete() local 359 txq->done_idx = idx; in slic_xmit_complete() 360 buff = &txq->txbuffs[idx]; in slic_xmit_complete() 387 (slic_get_free_tx_descs(txq) >= SLIC_MIN_TX_WAKEUP_DESCS)) in slic_xmit_complete() 838 struct slic_tx_queue *txq = &sdev->txq; in slic_init_tx_queue() local 844 txq->len = SLIC_NUM_TX_DESCS; in slic_init_tx_queue() 845 txq->put_idx = 0; in slic_init_tx_queue() 846 txq->done_idx = 0; in slic_init_tx_queue() [all …]
|
| /linux/drivers/net/wwan/t7xx/ |
| H A D | t7xx_hif_dpmaif.c | 236 tx_q = &dpmaif_ctrl->txq[tx_idx]; in t7xx_dpmaif_rxtx_sw_allocs() 261 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_rxtx_sw_allocs() 289 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_sw_release() 304 struct dpmaif_tx_queue *txq; in t7xx_dpmaif_start() local 343 txq = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_start() 344 txq->que_started = true; in t7xx_dpmaif_start() 346 hw_init_para.drb_base_addr[i] = txq->drb_bus_addr; in t7xx_dpmaif_start() 347 hw_init_para.drb_size_cnt[i] = txq->drb_size_cnt; in t7xx_dpmaif_start() 433 struct dpmaif_tx_queue *txq; in t7xx_dpmaif_start_txrx_qs() local 437 txq = &dpmaif_ctrl->txq[que_cnt]; in t7xx_dpmaif_start_txrx_qs() [all …]
|
| /linux/drivers/bluetooth/ |
| H A D | btintel_pcie.c | 241 static void btintel_pcie_prepare_tx(struct txq *txq, u16 tfd_index, in btintel_pcie_prepare_tx() argument 247 tfd = &txq->tfds[tfd_index]; in btintel_pcie_prepare_tx() 250 buf = &txq->bufs[tfd_index]; in btintel_pcie_prepare_tx() 358 struct txq *txq = &data->txq; in btintel_pcie_send_sync() local 362 if (tfd_index > txq->count) in btintel_pcie_send_sync() 388 btintel_pcie_prepare_tx(txq, tfd_index, skb); in btintel_pcie_send_sync() 390 tfd_index = (tfd_index + 1) % txq->count; in btintel_pcie_send_sync() 1000 struct txq *txq; in btintel_pcie_msix_tx_handle() local 1009 txq = &data->txq; in btintel_pcie_msix_tx_handle() 1015 urbd0 = &txq->urbd0s[cr_tia]; in btintel_pcie_msix_tx_handle() [all …]
|
| /linux/drivers/net/wireless/realtek/rtw88/ |
| H A D | tx.c | 575 struct ieee80211_txq *txq = rtwtxq_to_txq(rtwtxq); in rtw_txq_check_agg() local 594 if (!txq->sta) in rtw_txq_check_agg() 597 si = (struct rtw_sta_info *)txq->sta->drv_priv; in rtw_txq_check_agg() 598 set_bit(txq->tid, si->tid_ba); in rtw_txq_check_agg() 607 struct ieee80211_txq *txq = rtwtxq_to_txq(rtwtxq); in rtw_txq_push_skb() local 613 rtw_tx_pkt_info_update(rtwdev, &pkt_info, txq->sta, skb); in rtw_txq_push_skb() 625 struct ieee80211_txq *txq = rtwtxq_to_txq(rtwtxq); in rtw_txq_dequeue() local 628 skb = ieee80211_tx_dequeue(rtwdev->hw, txq); in rtw_txq_dequeue() 667 struct ieee80211_txq *txq = rtwtxq_to_txq(rtwtxq); in __rtw_tx_work() local 670 ieee80211_txq_get_depth(txq, &frame_cnt, NULL); in __rtw_tx_work() [all …]
|
| /linux/drivers/net/wireless/ath/ath6kl/ |
| H A D | htc_pipe.c | 106 if (list_empty(&ep->txq)) in get_htc_packet_credit_based() 110 packet = list_first_entry(&ep->txq, struct htc_packet, list); in get_htc_packet_credit_based() 114 __func__, packet, get_queue_depth(&ep->txq)); in get_htc_packet_credit_based() 159 packet = list_first_entry(&ep->txq, struct htc_packet, list); in get_htc_packet_credit_based() 183 if (list_empty(&ep->txq)) in get_htc_packet() 186 packet = list_first_entry(&ep->txq, struct htc_packet, list); in get_htc_packet() 191 __func__, packet, get_queue_depth(&ep->txq)); in get_htc_packet() 301 struct list_head *txq) in htc_try_send() argument 311 __func__, txq, in htc_try_send() 312 (txq == NULL) ? 0 : get_queue_depth(txq)); in htc_try_send() [all …]
|