/linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | tx.c | 94 txq_id != trans_pcie->txqs.cmd.q_id && in iwl_pcie_txq_inc_wr_ptr() 129 struct iwl_txq *txq = trans_pcie->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() 131 if (!test_bit(i, trans_pcie->txqs.queue_used)) in iwl_pcie_txq_check_wrptrs() 169 tfd = (u8 *)txq->tfds + trans_pcie->txqs.tfd.size * txq->write_ptr; in iwl_pcie_txq_build_tfd() 172 memset(tfd, 0, trans_pcie->txqs.tfd.size); in iwl_pcie_txq_build_tfd() 177 if (num_tbs >= trans_pcie->txqs.tfd.max_tbs) { in iwl_pcie_txq_build_tfd() 179 trans_pcie->txqs.tfd.max_tbs); in iwl_pcie_txq_build_tfd() 233 page_ptr = (void *)((u8 *)skb->cb + trans_pcie->txqs.page_offs); in iwl_pcie_free_tso_pages() 300 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen1_tfd_unmap() 384 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_pcie_txq_unmap() [all …]
|
H A D | tx-gen2.c | 27 page_ptr = (void *)((u8 *)skb->cb + trans_pcie->txqs.page_offs); in get_workaround_page() 589 WARN_ON(trans_pcie->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl() 597 WARN_ON(!trans_pcie->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl() 632 if (le16_to_cpu(tfd->num_tbs) >= trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen2_set_tb() 634 trans_pcie->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb() 656 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) { in iwl_txq_gen2_tfd_unmap() 731 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_tx() 740 if (WARN_ONCE(!test_bit(txq_id, trans_pcie->txqs.queue_used), in iwl_txq_gen2_tx() 759 trans_pcie->txqs.dev_cmd_offs); in iwl_txq_gen2_tx() 827 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_unmap() [all …]
|
H A D | internal.h | 520 struct iwl_pcie_txqs txqs; member 602 #define IWL_TRANS_PCIE_MAX_FRAGS(trans_pcie) ((trans_pcie)->txqs.tfd.max_tbs - 3) 685 return (u8 *)txq->tfds + trans_pcie->txqs.tfd.size * idx; in iwl_txq_get_tfd() 704 if (!test_and_set_bit(txq->id, trans_pcie->txqs.queue_stopped)) { in iwl_txq_stop() 742 if (test_and_clear_bit(txq->id, trans_pcie->txqs.queue_stopped)) { in iwl_trans_pcie_wake_queue() 802 unsigned long txqs, bool freeze);
|
H A D | ctxt-info.c | 221 cpu_to_le64(trans_pcie->txqs.txq[trans_pcie->txqs.cmd.q_id]->dma_addr); in iwl_pcie_ctxt_info_init()
|
/linux/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_main.c | 120 gather_tx_stats(nic_tx_stats, &nic_dev->txqs[i]); in gather_nic_stats() 134 if (nic_dev->txqs) in create_txqs() 137 nic_dev->txqs = devm_kcalloc(&netdev->dev, num_txqs, in create_txqs() 138 sizeof(*nic_dev->txqs), GFP_KERNEL); in create_txqs() 139 if (!nic_dev->txqs) in create_txqs() 147 err = hinic_init_txq(&nic_dev->txqs[i], sq, netdev); in create_txqs() 165 hinic_clean_txq(&nic_dev->txqs[i]); in create_txqs() 168 hinic_sq_debug_rem(nic_dev->txqs[j].sq); in create_txqs() 169 hinic_clean_txq(&nic_dev->txqs[j]); in create_txqs() 174 devm_kfree(&netdev->dev, nic_dev->txqs); in create_txqs() [all …]
|
H A D | hinic_dev.h | 97 struct hinic_txq *txqs; member
|
/linux/drivers/net/wireless/ath/ath5k/ |
H A D | mac80211-ops.c | 69 ath5k_tx_queue(hw, skb, &ah->txqs[qnum], control); in ath5k_tx() 745 *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; in ath5k_get_ringparam() 765 for (qnum = 0; qnum < ARRAY_SIZE(ah->txqs); qnum++) { in ath5k_set_ringparam() 766 if (!ah->txqs[qnum].setup) in ath5k_set_ringparam() 768 if (ah->txqs[qnum].qnum < AR5K_TX_QUEUE_ID_DATA_MIN || in ath5k_set_ringparam() 769 ah->txqs[qnum].qnum > AR5K_TX_QUEUE_ID_DATA_MAX) in ath5k_set_ringparam() 772 ah->txqs[qnum].txq_max = tx; in ath5k_set_ringparam() 773 if (ah->txqs[qnum].txq_len >= ah->txqs[qnum].txq_max) in ath5k_set_ringparam() 774 ieee80211_stop_queue(hw, ah->txqs[qnum].qnum); in ath5k_set_ringparam()
|
H A D | base.c | 1055 txq = &ah->txqs[qnum]; in ath5k_txq_setup() 1067 return &ah->txqs[qnum]; in ath5k_txq_setup() 1161 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_drain_tx_buffs() 1162 if (ah->txqs[i].setup) { in ath5k_drain_tx_buffs() 1163 txq = &ah->txqs[i]; in ath5k_drain_tx_buffs() 1186 struct ath5k_txq *txq = ah->txqs; in ath5k_txq_release() 1189 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) in ath5k_txq_release() 1838 if (ah->txqs[i].setup && (ah->ah_txq_isr_txok_all & BIT(i))) in ath5k_tasklet_tx() 1839 ath5k_tx_processq(ah, &ah->txqs[i]); in ath5k_tasklet_tx() 2054 trace_ath5k_tx(ah, bf->skb, &ah->txqs[ah->bhalq]); in ath5k_beacon_send() [all …]
|
/linux/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_main.c | 350 static void free_txqs(struct funeth_txq **txqs, unsigned int nqs, in free_txqs() argument 355 for (i = start; i < nqs && txqs[i]; i++) in free_txqs() 356 txqs[i] = funeth_txq_free(txqs[i], state); in free_txqs() 359 static int alloc_txqs(struct net_device *dev, struct funeth_txq **txqs, in alloc_txqs() argument 369 state, &txqs[i]); in alloc_txqs() 371 free_txqs(txqs, nqs, start, FUN_QSTATE_DESTROYED); in alloc_txqs() 453 qset->txqs = fp->txqs; in fun_free_rings() 465 fp->txqs = NULL; in fun_free_rings() 469 free_txqs(qset->txqs, qset->ntxqs, qset->txq_start, qset->state); in fun_free_rings() 481 struct funeth_txq **xdpqs = NULL, **txqs; in fun_alloc_rings() local [all …]
|
H A D | funeth.h | 67 struct funeth_txq **txqs; member 88 struct funeth_txq **txqs; member
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 593 txp.txq = &priv->txqs[skb_get_queue_mapping(skb)]; in hfi1_ipoib_send() 697 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init() 701 if (!priv->txqs) in hfi1_ipoib_txreq_init() 705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init() 756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init() 765 kfree(priv->txqs); in hfi1_ipoib_txreq_init() 766 priv->txqs = NULL; in hfi1_ipoib_txreq_init() 799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_deinit() 812 kfree(priv->txqs); in hfi1_ipoib_txreq_deinit() 813 priv->txqs = NULL; in hfi1_ipoib_txreq_deinit() [all …]
|
H A D | ipoib.h | 125 struct hfi1_ipoib_txq *txqs; member
|
/linux/net/ |
H A D | devres.c | 22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument 30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
|
/linux/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_txrx.c | 182 idpf_tx_desc_rel(txq_grp->txqs[j]); in idpf_tx_desc_rel_all() 318 struct idpf_tx_queue *txq = vport->txq_grps[i].txqs[j]; in idpf_tx_desc_alloc_all() 997 kfree(txq_grp->txqs[j]); in idpf_txq_group_rel() 998 txq_grp->txqs[j] = NULL; in idpf_txq_group_rel() 1094 kfree(vport->txqs); in idpf_vport_queues_rel() 1095 vport->txqs = NULL; in idpf_vport_queues_rel() 1112 vport->txqs = kcalloc(vport->num_txq, sizeof(*vport->txqs), in idpf_vport_init_fast_path_txqs() 1115 if (!vport->txqs) in idpf_vport_init_fast_path_txqs() 1122 vport->txqs[k] = tx_grp->txqs[j]; in idpf_vport_init_fast_path_txqs() 1123 vport->txqs[k]->idx = k; in idpf_vport_init_fast_path_txqs() [all …]
|
H A D | idpf_virtchnl.c | 759 idpf_queue_set(SW_MARKER, vport->txqs[i]); in idpf_wait_for_marker_event() 767 idpf_queue_clear(POLL_MODE, vport->txqs[i]); in idpf_wait_for_marker_event() 1109 tx_qgrp->txqs[j]->tail = in __idpf_queue_reg_init() 1460 cpu_to_le32(tx_qgrp->txqs[j]->q_id); in idpf_send_config_tx_queues_msg() 1466 cpu_to_le16(tx_qgrp->txqs[j]->desc_count); in idpf_send_config_tx_queues_msg() 1468 cpu_to_le64(tx_qgrp->txqs[j]->dma); in idpf_send_config_tx_queues_msg() 1470 struct idpf_tx_queue *q = tx_qgrp->txqs[j]; in idpf_send_config_tx_queues_msg() 1735 qc[k].start_queue_id = cpu_to_le32(tx_qgrp->txqs[j]->q_id); in idpf_send_ena_dis_queues_msg() 1880 vqv[k].queue_id = cpu_to_le32(tx_qgrp->txqs[j]->q_id); in idpf_send_map_unmap_queue_vector_msg() 1889 cpu_to_le16(tx_qgrp->txqs[j]->q_vector->v_idx); in idpf_send_map_unmap_queue_vector_msg() [all …]
|
H A D | idpf_ethtool.c | 832 struct idpf_tx_queue *txq = txq_grp->txqs[j]; in idpf_collect_queue_stats() 899 struct idpf_tx_queue *txq = txq_grp->txqs[j]; in idpf_get_ethtool_stats() 987 return vport->txqs[q_num]->q_vector; in idpf_find_txq_vec()
|
/linux/drivers/net/ethernet/netronome/nfp/abm/ |
H A D | main.c | 85 unsigned int txqs; in nfp_abm_spawn_repr() local 90 txqs = 1; in nfp_abm_spawn_repr() 93 txqs = alink->vnic->max_rx_rings; in nfp_abm_spawn_repr() 96 netdev = nfp_repr_alloc_mqs(app, txqs, 1); in nfp_abm_spawn_repr()
|
/linux/drivers/net/wireless/intel/iwlwifi/ |
H A D | iwl-trans.c | 684 int iwl_trans_wait_tx_queues_empty(struct iwl_trans *trans, u32 txqs) in iwl_trans_wait_tx_queues_empty() argument 690 return iwl_trans_pcie_wait_txqs_empty(trans, txqs); in iwl_trans_wait_tx_queues_empty() 695 unsigned long txqs, bool freeze) in iwl_trans_freeze_txq_timer() argument 701 iwl_pcie_freeze_txq_timer(trans, txqs, freeze); in iwl_trans_freeze_txq_timer()
|
/linux/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_repr.h | 102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
|
H A D | nfp_net_repr.c | 424 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) in nfp_repr_alloc_mqs() argument 429 netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); in nfp_repr_alloc_mqs()
|
/linux/net/ethernet/ |
H A D | eth.c | 368 * @txqs: The number of TX queues this device has. 379 struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs, 383 ether_setup, txqs, rxqs); in alloc_etherdev_mqs() 381 alloc_etherdev_mqs(int sizeof_priv,unsigned int txqs,unsigned int rxqs) alloc_etherdev_mqs() argument
|
/linux/drivers/net/can/dev/ |
H A D | dev.c | 248 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument 275 txqs, rxqs); in alloc_candev_mqs()
|
/linux/drivers/net/wan/ |
H A D | farsync.c | 437 int txqs; /* index to get next buffer to tx */ member 1305 txq_length = port->txqe - port->txqs; in do_bottom_half_tx() 1317 skb = port->txq[port->txqs]; in do_bottom_half_tx() 1318 port->txqs++; in do_bottom_half_tx() 1319 if (port->txqs == FST_TXQ_DEPTH) in do_bottom_half_tx() 1320 port->txqs = 0; in do_bottom_half_tx() 2098 port->txqs = 0; in fst_openport() 2233 txq_length = port->txqe - port->txqs; in fst_start_xmit()
|
/linux/include/linux/can/ |
H A D | dev.h | 174 unsigned int txqs, unsigned int rxqs);
|
/linux/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 511 struct mvneta_tx_queue *txqs; member 1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_port_up() 1864 return &pp->txqs[queue]; in mvneta_tx_done_policy() 2188 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit_back() 2229 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit() 2915 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; in mvneta_tx() 3426 mvneta_txq_done_force(pp, &pp->txqs[queue]); in mvneta_tx_reset() 3647 mvneta_txq_deinit(pp, &pp->txqs[queue]); in mvneta_cleanup_txqs() 3685 int err = mvneta_txq_init(pp, &pp->txqs[queue]); in mvneta_setup_txqs() 4735 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_ethtool_set_coalesce() [all …]
|