| /linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/ |
| H A D | tx.c | 130 struct iwl_txq *txq = trans_pcie->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() 132 if (!test_bit(i, trans_pcie->txqs.queue_used)) in iwl_pcie_txq_check_wrptrs() 170 tfd = (u8 *)txq->tfds + trans_pcie->txqs.tfd.size * txq->write_ptr; in iwl_pcie_txq_build_tfd() 173 memset(tfd, 0, trans_pcie->txqs.tfd.size); in iwl_pcie_txq_build_tfd() 178 if (num_tbs >= trans_pcie->txqs.tfd.max_tbs) { in iwl_pcie_txq_build_tfd() 180 trans_pcie->txqs.tfd.max_tbs); in iwl_pcie_txq_build_tfd() 302 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen1_tfd_unmap() 386 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_pcie_txq_unmap() 441 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_pcie_txq_free() 460 trans_pcie->txqs.tfd.size * in iwl_pcie_txq_free() [all …]
|
| H A D | tx-gen2.c | 623 if (le16_to_cpu(tfd->num_tbs) >= trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen2_set_tb() 625 trans_pcie->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb() 647 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen2_tfd_unmap() 722 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_tx() 731 if (WARN_ONCE(!test_bit(txq_id, trans_pcie->txqs.queue_used), in iwl_txq_gen2_tx() 819 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_unmap() 861 trans_pcie->txqs.tfd.size * txq->n_window, in iwl_txq_gen2_free_memory() 870 dma_pool_free(trans_pcie->txqs.bc_pool, in iwl_txq_gen2_free_memory() 893 txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_free() 910 trans_pcie->txqs.txq[txq_id] = NULL; in iwl_txq_gen2_free() [all …]
|
| H A D | trans.c | 2013 if (trans_pcie->txqs.tso_hdr_page) { in iwl_trans_pcie_free() 2016 per_cpu_ptr(trans_pcie->txqs.tso_hdr_page, i); in iwl_trans_pcie_free() 2022 free_percpu(trans_pcie->txqs.tso_hdr_page); in iwl_trans_pcie_free() 2546 if (!test_bit(txq_idx, trans_pcie->txqs.queue_used)) in iwl_trans_pcie_wait_txq_empty() 2550 txq = trans_pcie->txqs.txq[txq_idx]; in iwl_trans_pcie_wait_txq_empty() 2609 if (!test_bit(cnt, trans_pcie->txqs.queue_used)) in iwl_trans_pcie_wait_txqs_empty() 2783 struct iwl_txq *txq = trans_pcie->txqs.txq[state->pos]; in iwl_dbgfs_tx_queue_seq_show() 2787 !!test_bit(state->pos, trans_pcie->txqs.queue_used), in iwl_dbgfs_tx_queue_seq_show() 2788 !!test_bit(state->pos, trans_pcie->txqs.queue_stopped)); in iwl_dbgfs_tx_queue_seq_show() 3269 for (i = 0; i < trans_pcie->txqs.tfd.max_tbs; i++) in iwl_trans_pcie_get_cmdlen() [all …]
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_main.c | 350 static void free_txqs(struct funeth_txq **txqs, unsigned int nqs, in free_txqs() argument 355 for (i = start; i < nqs && txqs[i]; i++) in free_txqs() 356 txqs[i] = funeth_txq_free(txqs[i], state); in free_txqs() 359 static int alloc_txqs(struct net_device *dev, struct funeth_txq **txqs, in alloc_txqs() argument 369 state, &txqs[i]); in alloc_txqs() 371 free_txqs(txqs, nqs, start, FUN_QSTATE_DESTROYED); in alloc_txqs() 453 qset->txqs = fp->txqs; in fun_free_rings() 465 fp->txqs = NULL; in fun_free_rings() 469 free_txqs(qset->txqs, qset->ntxqs, qset->txq_start, qset->state); in fun_free_rings() 481 struct funeth_txq **xdpqs = NULL, **txqs; in fun_alloc_rings() local [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_main.c | 120 gather_tx_stats(nic_tx_stats, &nic_dev->txqs[i]); in gather_nic_stats() 134 if (nic_dev->txqs) in create_txqs() 137 nic_dev->txqs = devm_kcalloc(&netdev->dev, num_txqs, in create_txqs() 138 sizeof(*nic_dev->txqs), GFP_KERNEL); in create_txqs() 139 if (!nic_dev->txqs) in create_txqs() 147 err = hinic_init_txq(&nic_dev->txqs[i], sq, netdev); in create_txqs() 165 hinic_clean_txq(&nic_dev->txqs[i]); in create_txqs() 168 hinic_sq_debug_rem(nic_dev->txqs[j].sq); in create_txqs() 169 hinic_clean_txq(&nic_dev->txqs[j]); in create_txqs() 174 devm_kfree(&netdev->dev, nic_dev->txqs); in create_txqs() [all …]
|
| H A D | hinic_dev.h | 97 struct hinic_txq *txqs; member
|
| /linux/drivers/net/wireless/ath/ath5k/ |
| H A D | mac80211-ops.c | 69 ath5k_tx_queue(hw, skb, &ah->txqs[qnum], control); in ath5k_tx() 749 *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; in ath5k_get_ringparam() 769 for (qnum = 0; qnum < ARRAY_SIZE(ah->txqs); qnum++) { in ath5k_set_ringparam() 770 if (!ah->txqs[qnum].setup) in ath5k_set_ringparam() 772 if (ah->txqs[qnum].qnum < AR5K_TX_QUEUE_ID_DATA_MIN || in ath5k_set_ringparam() 773 ah->txqs[qnum].qnum > AR5K_TX_QUEUE_ID_DATA_MAX) in ath5k_set_ringparam() 776 ah->txqs[qnum].txq_max = tx; in ath5k_set_ringparam() 777 if (ah->txqs[qnum].txq_len >= ah->txqs[qnum].txq_max) in ath5k_set_ringparam() 778 ieee80211_stop_queue(hw, ah->txqs[qnum].qnum); in ath5k_set_ringparam()
|
| H A D | base.c | 1055 txq = &ah->txqs[qnum]; in ath5k_txq_setup() 1067 return &ah->txqs[qnum]; in ath5k_txq_setup() 1161 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_drain_tx_buffs() 1162 if (ah->txqs[i].setup) { in ath5k_drain_tx_buffs() 1163 txq = &ah->txqs[i]; in ath5k_drain_tx_buffs() 1186 struct ath5k_txq *txq = ah->txqs; in ath5k_txq_release() 1189 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) in ath5k_txq_release() 1838 if (ah->txqs[i].setup && (ah->ah_txq_isr_txok_all & BIT(i))) in ath5k_tasklet_tx() 1839 ath5k_tx_processq(ah, &ah->txqs[i]); in ath5k_tasklet_tx() 2054 trace_ath5k_tx(ah, bf->skb, &ah->txqs[ah->bhalq]); in ath5k_beacon_send() [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | ipoib_tx.c | 593 txp.txq = &priv->txqs[skb_get_queue_mapping(skb)]; in hfi1_ipoib_send() 697 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init() 701 if (!priv->txqs) in hfi1_ipoib_txreq_init() 705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init() 756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init() 765 kfree(priv->txqs); in hfi1_ipoib_txreq_init() 766 priv->txqs = NULL; in hfi1_ipoib_txreq_init() 799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_deinit() 812 kfree(priv->txqs); in hfi1_ipoib_txreq_deinit() 813 priv->txqs = NULL; in hfi1_ipoib_txreq_deinit() [all …]
|
| H A D | ipoib.h | 125 struct hfi1_ipoib_txq *txqs; member
|
| /linux/net/ |
| H A D | devres.c | 22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument 30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
|
| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_tx.c | 51 nic_dev->txqs = kzalloc_objs(*nic_dev->txqs, num_txqs); in hinic3_alloc_txqs() 52 if (!nic_dev->txqs) in hinic3_alloc_txqs() 56 txq = &nic_dev->txqs[q_id]; in hinic3_alloc_txqs() 73 kfree(nic_dev->txqs); in hinic3_free_txqs() 621 return hinic3_send_one_skb(skb, netdev, &nic_dev->txqs[q_id]); in hinic3_xmit_frame() 665 err = hinic3_stop_sq(&nic_dev->txqs[qid]); in hinic3_flush_txqs() 735 txq = &nic_dev->txqs[q_id]; in hinic3_configure_txqs()
|
| H A D | hinic3_netdev_ops.c | 766 sq = nic_dev->txqs[txqueue].sq; in hinic3_tx_timeout() 794 if (!nic_dev->txqs) in hinic3_get_stats64() 797 txq = &nic_dev->txqs[i]; in hinic3_get_stats64()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 164 idpf_tx_desc_rel(txq_grp->txqs[j]); in idpf_tx_desc_rel_all() 312 struct idpf_tx_queue *txq = rsrc->txq_grps[i].txqs[j]; in idpf_tx_desc_alloc_all() 1163 qs->qs[num++].txq = vport->txqs[idx]; in idpf_vector_to_queue_set() 1166 qs->qs[num++].complq = vport->txqs[idx]->complq; in idpf_vector_to_queue_set() 1317 if (idpf_queue_has(FLOW_SCH_EN, txq_grp->txqs[j])) { in idpf_txq_group_rel() 1318 kfree(txq_grp->txqs[j]->refillq); in idpf_txq_group_rel() 1319 txq_grp->txqs[j]->refillq = NULL; in idpf_txq_group_rel() 1322 kfree(txq_grp->txqs[j]); in idpf_txq_group_rel() 1323 txq_grp->txqs[j] = NULL; in idpf_txq_group_rel() 1418 kfree(vport->txqs); in idpf_vport_queues_rel() [all …]
|
| H A D | xdp.c | 78 rxq->xdpsqs = &vport->txqs[rsrc->xdp_txq_offset]; in __idpf_xdp_rxq_info_init() 176 struct idpf_tx_queue *xdpsq = vport->txqs[i]; in idpf_xdpsqs_get() 213 struct idpf_tx_queue *xdpsq = vport->txqs[i]; in idpf_xdpsqs_put() 374 &vport->txqs[xdp_txq_offset], in idpf_xdp_xmit()
|
| /linux/drivers/net/ethernet/netronome/nfp/abm/ |
| H A D | main.c | 85 unsigned int txqs; in nfp_abm_spawn_repr() local 90 txqs = 1; in nfp_abm_spawn_repr() 93 txqs = alink->vnic->max_rx_rings; in nfp_abm_spawn_repr() 96 netdev = nfp_repr_alloc_mqs(app, txqs, 1); in nfp_abm_spawn_repr()
|
| /linux/drivers/net/wireless/intel/iwlwifi/ |
| H A D | iwl-trans.c | 712 int iwl_trans_wait_tx_queues_empty(struct iwl_trans *trans, u32 txqs) in iwl_trans_wait_tx_queues_empty() argument 718 return iwl_trans_pcie_wait_txqs_empty(trans, txqs); in iwl_trans_wait_tx_queues_empty() 723 unsigned long txqs, bool freeze) in iwl_trans_freeze_txq_timer() argument 729 iwl_pcie_freeze_txq_timer(trans, txqs, freeze); in iwl_trans_freeze_txq_timer()
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | nfp_net_repr.h | 102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
|
| H A D | nfp_net_repr.c | 424 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) in nfp_repr_alloc_mqs() argument 429 netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); in nfp_repr_alloc_mqs()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_base.c | 1052 u8 buf_len = struct_size(qg_buf, txqs, 1); in ice_vsi_cfg_txq() 1073 qg_buf->txqs[0].txq_id = cpu_to_le16(pf_q); in ice_vsi_cfg_txq() 1074 ice_pack_txq_ctx(&tlan_ctx, &qg_buf->txqs[0].txq_ctx); in ice_vsi_cfg_txq() 1110 txq = &qg_buf->txqs[0]; in ice_vsi_cfg_txq() 1145 DEFINE_RAW_FLEX(struct ice_aqc_add_tx_qgrp, qg_buf, txqs, 1); in ice_vsi_cfg_single_txq() 1167 DEFINE_RAW_FLEX(struct ice_aqc_add_tx_qgrp, qg_buf, txqs, 1); in ice_vsi_cfg_txqs()
|
| /linux/drivers/net/wan/ |
| H A D | farsync.c | 437 int txqs; /* index to get next buffer to tx */ member 1305 txq_length = port->txqe - port->txqs; in do_bottom_half_tx() 1317 skb = port->txq[port->txqs]; in do_bottom_half_tx() 1318 port->txqs++; in do_bottom_half_tx() 1319 if (port->txqs == FST_TXQ_DEPTH) in do_bottom_half_tx() 1320 port->txqs = 0; in do_bottom_half_tx() 2098 port->txqs = 0; in fst_openport() 2233 txq_length = port->txqe - port->txqs; in fst_start_xmit()
|
| /linux/drivers/net/can/dev/ |
| H A D | dev.c | 299 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument 326 txqs, rxqs); in alloc_candev_mqs()
|
| /linux/include/linux/can/ |
| H A D | dev.h | 106 unsigned int txqs, unsigned int rxqs);
|
| /linux/drivers/net/ethernet/marvell/ |
| H A D | mvneta.c | 511 struct mvneta_tx_queue *txqs; member 1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_port_up() 1864 return &pp->txqs[queue]; in mvneta_tx_done_policy() 2188 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit_back() 2229 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit() 2917 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; in mvneta_tx() 3435 mvneta_txq_done_force(pp, &pp->txqs[queue]); in mvneta_tx_reset() 3656 mvneta_txq_deinit(pp, &pp->txqs[queue]); in mvneta_cleanup_txqs() 3694 int err = mvneta_txq_init(pp, &pp->txqs[queue]); in mvneta_setup_txqs() 4744 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_ethtool_set_coalesce() [all …]
|
| /linux/drivers/net/ethernet/marvell/mvpp2/ |
| H A D | mvpp2_main.c | 2355 struct mvpp2_tx_queue *txq = port->txqs[queue]; in mvpp2_egress_enable() 2649 int id = port->txqs[queue]->id; in mvpp2_txq_sent_counter_clear() 2854 return port->txqs[queue]; in mvpp2_get_tx_queue() 3283 txq = port->txqs[queue]; in mvpp2_cleanup_txqs() 3334 txq = port->txqs[queue]; in mvpp2_setup_txqs() 3347 txq = port->txqs[queue]; in mvpp2_setup_txqs() 3645 txq = port->txqs[txq_id]; in mvpp2_xdp_finish_tx() 3681 txq = port->txqs[txq_id]; in mvpp2_xdp_submit_frame() 4373 txq = port->txqs[txq_id]; in mvpp2_tx() 5422 struct mvpp2_tx_queue *txq = port->txqs[queue]; in mvpp2_ethtool_set_coalesce() [all …]
|