Home
last modified time | relevance | path

Searched refs:txqs (Results 1 – 25 of 52) sorted by relevance

123

/linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/
H A Dtx.c130 struct iwl_txq *txq = trans_pcie->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs()
132 if (!test_bit(i, trans_pcie->txqs.queue_used)) in iwl_pcie_txq_check_wrptrs()
170 tfd = (u8 *)txq->tfds + trans_pcie->txqs.tfd.size * txq->write_ptr; in iwl_pcie_txq_build_tfd()
173 memset(tfd, 0, trans_pcie->txqs.tfd.size); in iwl_pcie_txq_build_tfd()
178 if (num_tbs >= trans_pcie->txqs.tfd.max_tbs) { in iwl_pcie_txq_build_tfd()
180 trans_pcie->txqs.tfd.max_tbs); in iwl_pcie_txq_build_tfd()
302 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen1_tfd_unmap()
386 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_pcie_txq_unmap()
441 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_pcie_txq_free()
460 trans_pcie->txqs.tfd.size * in iwl_pcie_txq_free()
[all …]
H A Dtx-gen2.c623 if (le16_to_cpu(tfd->num_tbs) >= trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen2_set_tb()
625 trans_pcie->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb()
647 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen2_tfd_unmap()
722 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_tx()
731 if (WARN_ONCE(!test_bit(txq_id, trans_pcie->txqs.queue_used), in iwl_txq_gen2_tx()
819 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_unmap()
861 trans_pcie->txqs.tfd.size * txq->n_window, in iwl_txq_gen2_free_memory()
870 dma_pool_free(trans_pcie->txqs.bc_pool, in iwl_txq_gen2_free_memory()
893 txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_free()
910 trans_pcie->txqs.txq[txq_id] = NULL; in iwl_txq_gen2_free()
[all …]
H A Dtrans.c2013 if (trans_pcie->txqs.tso_hdr_page) { in iwl_trans_pcie_free()
2016 per_cpu_ptr(trans_pcie->txqs.tso_hdr_page, i); in iwl_trans_pcie_free()
2022 free_percpu(trans_pcie->txqs.tso_hdr_page); in iwl_trans_pcie_free()
2546 if (!test_bit(txq_idx, trans_pcie->txqs.queue_used)) in iwl_trans_pcie_wait_txq_empty()
2550 txq = trans_pcie->txqs.txq[txq_idx]; in iwl_trans_pcie_wait_txq_empty()
2609 if (!test_bit(cnt, trans_pcie->txqs.queue_used)) in iwl_trans_pcie_wait_txqs_empty()
2783 struct iwl_txq *txq = trans_pcie->txqs.txq[state->pos]; in iwl_dbgfs_tx_queue_seq_show()
2787 !!test_bit(state->pos, trans_pcie->txqs.queue_used), in iwl_dbgfs_tx_queue_seq_show()
2788 !!test_bit(state->pos, trans_pcie->txqs.queue_stopped)); in iwl_dbgfs_tx_queue_seq_show()
3269 for (i = 0; i < trans_pcie->txqs.tfd.max_tbs; i++) in iwl_trans_pcie_get_cmdlen()
[all …]
/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_main.c350 static void free_txqs(struct funeth_txq **txqs, unsigned int nqs, in free_txqs() argument
355 for (i = start; i < nqs && txqs[i]; i++) in free_txqs()
356 txqs[i] = funeth_txq_free(txqs[i], state); in free_txqs()
359 static int alloc_txqs(struct net_device *dev, struct funeth_txq **txqs, in alloc_txqs() argument
369 state, &txqs[i]); in alloc_txqs()
371 free_txqs(txqs, nqs, start, FUN_QSTATE_DESTROYED); in alloc_txqs()
453 qset->txqs = fp->txqs; in fun_free_rings()
465 fp->txqs = NULL; in fun_free_rings()
469 free_txqs(qset->txqs, qset->ntxqs, qset->txq_start, qset->state); in fun_free_rings()
481 struct funeth_txq **xdpqs = NULL, **txqs; in fun_alloc_rings() local
[all …]
/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_main.c120 gather_tx_stats(nic_tx_stats, &nic_dev->txqs[i]); in gather_nic_stats()
134 if (nic_dev->txqs) in create_txqs()
137 nic_dev->txqs = devm_kcalloc(&netdev->dev, num_txqs, in create_txqs()
138 sizeof(*nic_dev->txqs), GFP_KERNEL); in create_txqs()
139 if (!nic_dev->txqs) in create_txqs()
147 err = hinic_init_txq(&nic_dev->txqs[i], sq, netdev); in create_txqs()
165 hinic_clean_txq(&nic_dev->txqs[i]); in create_txqs()
168 hinic_sq_debug_rem(nic_dev->txqs[j].sq); in create_txqs()
169 hinic_clean_txq(&nic_dev->txqs[j]); in create_txqs()
174 devm_kfree(&netdev->dev, nic_dev->txqs); in create_txqs()
[all …]
H A Dhinic_dev.h97 struct hinic_txq *txqs; member
/linux/drivers/net/wireless/ath/ath5k/
H A Dmac80211-ops.c69 ath5k_tx_queue(hw, skb, &ah->txqs[qnum], control); in ath5k_tx()
749 *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; in ath5k_get_ringparam()
769 for (qnum = 0; qnum < ARRAY_SIZE(ah->txqs); qnum++) { in ath5k_set_ringparam()
770 if (!ah->txqs[qnum].setup) in ath5k_set_ringparam()
772 if (ah->txqs[qnum].qnum < AR5K_TX_QUEUE_ID_DATA_MIN || in ath5k_set_ringparam()
773 ah->txqs[qnum].qnum > AR5K_TX_QUEUE_ID_DATA_MAX) in ath5k_set_ringparam()
776 ah->txqs[qnum].txq_max = tx; in ath5k_set_ringparam()
777 if (ah->txqs[qnum].txq_len >= ah->txqs[qnum].txq_max) in ath5k_set_ringparam()
778 ieee80211_stop_queue(hw, ah->txqs[qnum].qnum); in ath5k_set_ringparam()
H A Dbase.c1055 txq = &ah->txqs[qnum]; in ath5k_txq_setup()
1067 return &ah->txqs[qnum]; in ath5k_txq_setup()
1161 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_drain_tx_buffs()
1162 if (ah->txqs[i].setup) { in ath5k_drain_tx_buffs()
1163 txq = &ah->txqs[i]; in ath5k_drain_tx_buffs()
1186 struct ath5k_txq *txq = ah->txqs; in ath5k_txq_release()
1189 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) in ath5k_txq_release()
1838 if (ah->txqs[i].setup && (ah->ah_txq_isr_txok_all & BIT(i))) in ath5k_tasklet_tx()
1839 ath5k_tx_processq(ah, &ah->txqs[i]); in ath5k_tasklet_tx()
2054 trace_ath5k_tx(ah, bf->skb, &ah->txqs[ah->bhalq]); in ath5k_beacon_send()
[all …]
/linux/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c593 txp.txq = &priv->txqs[skb_get_queue_mapping(skb)]; in hfi1_ipoib_send()
697 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init()
701 if (!priv->txqs) in hfi1_ipoib_txreq_init()
705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init()
756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init()
765 kfree(priv->txqs); in hfi1_ipoib_txreq_init()
766 priv->txqs = NULL; in hfi1_ipoib_txreq_init()
799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_deinit()
812 kfree(priv->txqs); in hfi1_ipoib_txreq_deinit()
813 priv->txqs = NULL; in hfi1_ipoib_txreq_deinit()
[all …]
H A Dipoib.h125 struct hfi1_ipoib_txq *txqs; member
/linux/net/
H A Ddevres.c22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument
30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
/linux/drivers/net/ethernet/huawei/hinic3/
H A Dhinic3_tx.c51 nic_dev->txqs = kzalloc_objs(*nic_dev->txqs, num_txqs); in hinic3_alloc_txqs()
52 if (!nic_dev->txqs) in hinic3_alloc_txqs()
56 txq = &nic_dev->txqs[q_id]; in hinic3_alloc_txqs()
73 kfree(nic_dev->txqs); in hinic3_free_txqs()
621 return hinic3_send_one_skb(skb, netdev, &nic_dev->txqs[q_id]); in hinic3_xmit_frame()
665 err = hinic3_stop_sq(&nic_dev->txqs[qid]); in hinic3_flush_txqs()
735 txq = &nic_dev->txqs[q_id]; in hinic3_configure_txqs()
H A Dhinic3_netdev_ops.c766 sq = nic_dev->txqs[txqueue].sq; in hinic3_tx_timeout()
794 if (!nic_dev->txqs) in hinic3_get_stats64()
797 txq = &nic_dev->txqs[i]; in hinic3_get_stats64()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c164 idpf_tx_desc_rel(txq_grp->txqs[j]); in idpf_tx_desc_rel_all()
312 struct idpf_tx_queue *txq = rsrc->txq_grps[i].txqs[j]; in idpf_tx_desc_alloc_all()
1163 qs->qs[num++].txq = vport->txqs[idx]; in idpf_vector_to_queue_set()
1166 qs->qs[num++].complq = vport->txqs[idx]->complq; in idpf_vector_to_queue_set()
1317 if (idpf_queue_has(FLOW_SCH_EN, txq_grp->txqs[j])) { in idpf_txq_group_rel()
1318 kfree(txq_grp->txqs[j]->refillq); in idpf_txq_group_rel()
1319 txq_grp->txqs[j]->refillq = NULL; in idpf_txq_group_rel()
1322 kfree(txq_grp->txqs[j]); in idpf_txq_group_rel()
1323 txq_grp->txqs[j] = NULL; in idpf_txq_group_rel()
1418 kfree(vport->txqs); in idpf_vport_queues_rel()
[all …]
H A Dxdp.c78 rxq->xdpsqs = &vport->txqs[rsrc->xdp_txq_offset]; in __idpf_xdp_rxq_info_init()
176 struct idpf_tx_queue *xdpsq = vport->txqs[i]; in idpf_xdpsqs_get()
213 struct idpf_tx_queue *xdpsq = vport->txqs[i]; in idpf_xdpsqs_put()
374 &vport->txqs[xdp_txq_offset], in idpf_xdp_xmit()
/linux/drivers/net/ethernet/netronome/nfp/abm/
H A Dmain.c85 unsigned int txqs; in nfp_abm_spawn_repr() local
90 txqs = 1; in nfp_abm_spawn_repr()
93 txqs = alink->vnic->max_rx_rings; in nfp_abm_spawn_repr()
96 netdev = nfp_repr_alloc_mqs(app, txqs, 1); in nfp_abm_spawn_repr()
/linux/drivers/net/wireless/intel/iwlwifi/
H A Diwl-trans.c712 int iwl_trans_wait_tx_queues_empty(struct iwl_trans *trans, u32 txqs) in iwl_trans_wait_tx_queues_empty() argument
718 return iwl_trans_pcie_wait_txqs_empty(trans, txqs); in iwl_trans_wait_tx_queues_empty()
723 unsigned long txqs, bool freeze) in iwl_trans_freeze_txq_timer() argument
729 iwl_pcie_freeze_txq_timer(trans, txqs, freeze); in iwl_trans_freeze_txq_timer()
/linux/drivers/net/ethernet/netronome/nfp/
H A Dnfp_net_repr.h102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
H A Dnfp_net_repr.c424 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) in nfp_repr_alloc_mqs() argument
429 netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); in nfp_repr_alloc_mqs()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_base.c1052 u8 buf_len = struct_size(qg_buf, txqs, 1); in ice_vsi_cfg_txq()
1073 qg_buf->txqs[0].txq_id = cpu_to_le16(pf_q); in ice_vsi_cfg_txq()
1074 ice_pack_txq_ctx(&tlan_ctx, &qg_buf->txqs[0].txq_ctx); in ice_vsi_cfg_txq()
1110 txq = &qg_buf->txqs[0]; in ice_vsi_cfg_txq()
1145 DEFINE_RAW_FLEX(struct ice_aqc_add_tx_qgrp, qg_buf, txqs, 1); in ice_vsi_cfg_single_txq()
1167 DEFINE_RAW_FLEX(struct ice_aqc_add_tx_qgrp, qg_buf, txqs, 1); in ice_vsi_cfg_txqs()
/linux/drivers/net/wan/
H A Dfarsync.c437 int txqs; /* index to get next buffer to tx */ member
1305 txq_length = port->txqe - port->txqs; in do_bottom_half_tx()
1317 skb = port->txq[port->txqs]; in do_bottom_half_tx()
1318 port->txqs++; in do_bottom_half_tx()
1319 if (port->txqs == FST_TXQ_DEPTH) in do_bottom_half_tx()
1320 port->txqs = 0; in do_bottom_half_tx()
2098 port->txqs = 0; in fst_openport()
2233 txq_length = port->txqe - port->txqs; in fst_start_xmit()
/linux/drivers/net/can/dev/
H A Ddev.c299 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument
326 txqs, rxqs); in alloc_candev_mqs()
/linux/include/linux/can/
H A Ddev.h106 unsigned int txqs, unsigned int rxqs);
/linux/drivers/net/ethernet/marvell/
H A Dmvneta.c511 struct mvneta_tx_queue *txqs; member
1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_port_up()
1864 return &pp->txqs[queue]; in mvneta_tx_done_policy()
2188 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit_back()
2229 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit()
2917 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; in mvneta_tx()
3435 mvneta_txq_done_force(pp, &pp->txqs[queue]); in mvneta_tx_reset()
3656 mvneta_txq_deinit(pp, &pp->txqs[queue]); in mvneta_cleanup_txqs()
3694 int err = mvneta_txq_init(pp, &pp->txqs[queue]); in mvneta_setup_txqs()
4744 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_ethtool_set_coalesce()
[all …]
/linux/drivers/net/ethernet/marvell/mvpp2/
H A Dmvpp2_main.c2355 struct mvpp2_tx_queue *txq = port->txqs[queue]; in mvpp2_egress_enable()
2649 int id = port->txqs[queue]->id; in mvpp2_txq_sent_counter_clear()
2854 return port->txqs[queue]; in mvpp2_get_tx_queue()
3283 txq = port->txqs[queue]; in mvpp2_cleanup_txqs()
3334 txq = port->txqs[queue]; in mvpp2_setup_txqs()
3347 txq = port->txqs[queue]; in mvpp2_setup_txqs()
3645 txq = port->txqs[txq_id]; in mvpp2_xdp_finish_tx()
3681 txq = port->txqs[txq_id]; in mvpp2_xdp_submit_frame()
4373 txq = port->txqs[txq_id]; in mvpp2_tx()
5422 struct mvpp2_tx_queue *txq = port->txqs[queue]; in mvpp2_ethtool_set_coalesce()
[all …]

123