Home
last modified time | relevance | path

Searched refs:tx_rings (Results 1 – 25 of 36) sorted by relevance

12

/linux/drivers/net/ethernet/aquantia/atlantic/
H A Daq_vec.c18 unsigned int tx_rings; member
41 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_poll()
119 self->tx_rings = 0; in aq_vec_alloc()
144 ++self->tx_rings; in aq_vec_ring_alloc()
190 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_init()
232 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_start()
256 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_stop()
276 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_deinit()
305 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_ring_free()
314 self->tx_rings = 0; in aq_vec_ring_free()
[all …]
H A Daq_hw.h66 u8 tx_rings; member
/linux/drivers/net/ethernet/netronome/nfp/
H A Dnfp_net_dp.c187 dp->tx_rings = kzalloc_objs(*dp->tx_rings, dp->num_tx_rings); in nfp_net_tx_rings_prepare()
188 if (!dp->tx_rings) in nfp_net_tx_rings_prepare()
205 nfp_net_tx_ring_init(&dp->tx_rings[r], dp, in nfp_net_tx_rings_prepare()
208 if (nfp_net_tx_ring_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare()
211 if (nfp_net_tx_ring_bufs_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare()
219 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare()
221 nfp_net_tx_ring_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare()
227 kfree(dp->tx_rings); in nfp_net_tx_rings_prepare()
236 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free()
237 nfp_net_tx_ring_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free()
[all …]
H A Dnfp_net_common.c797 idx < dp->num_stack_tx_rings ? &dp->tx_rings[idx] : NULL; in nfp_net_vector_assign_rings()
800 &dp->tx_rings[dp->num_stack_tx_rings + idx] : NULL; in nfp_net_vector_assign_rings()
968 nfp_net_tx_ring_reset(&nn->dp, &nn->dp.tx_rings[r]); in nfp_net_clear_config_and_disable()
1001 nfp_net_tx_ring_hw_cfg_write(nn, &nn->dp.tx_rings[r], r); in nfp_net_set_config_and_enable()
1574 new->tx_rings = NULL; in nfp_net_clone_dp()
H A Dnfp_net.h520 struct nfp_net_tx_ring *tx_rings; member
/linux/drivers/net/wireless/realtek/rtw88/
H A Dpci.c168 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_free_trx_ring()
329 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_init_trx_ring()
351 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_init_trx_ring()
405 dma = rtwpci->tx_rings[RTW_TX_QUEUE_BCN].r.dma; in rtw_pci_reset_buf_desc()
409 len = rtwpci->tx_rings[RTW_TX_QUEUE_H2C].r.len; in rtw_pci_reset_buf_desc()
410 dma = rtwpci->tx_rings[RTW_TX_QUEUE_H2C].r.dma; in rtw_pci_reset_buf_desc()
411 rtwpci->tx_rings[RTW_TX_QUEUE_H2C].r.rp = 0; in rtw_pci_reset_buf_desc()
412 rtwpci->tx_rings[RTW_TX_QUEUE_H2C].r.wp = 0; in rtw_pci_reset_buf_desc()
417 len = rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.len; in rtw_pci_reset_buf_desc()
418 dma = rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.dma; in rtw_pci_reset_buf_desc()
[all …]
H A Dpci.h223 struct rtw_pci_tx_ring tx_rings[RTK_MAX_TX_QUEUE_NUM]; member
/linux/drivers/net/ethernet/intel/ice/
H A Dice_ethtool.c1290 tx_ring = test_vsi->tx_rings[0]; in ice_loopback_test()
1951 tx_ring = READ_ONCE(vsi->tx_rings[j]); in __ice_get_ethtool_stats()
3165 if (vsi->tx_rings && vsi->rx_rings) { in ice_get_ringparam()
3167 ring->tx_pending = vsi->tx_rings[0]->count; in ice_get_ringparam()
3191 struct ice_tx_ring *tx_rings = NULL; in ice_set_ringparam() local
3212 if (!vsi->tx_rings || !vsi->rx_rings) in ice_set_ringparam()
3227 if (new_tx_cnt == vsi->tx_rings[0]->count && in ice_set_ringparam()
3251 vsi->tx_rings[i]->count = new_tx_cnt; in ice_set_ringparam()
3265 if (new_tx_cnt == vsi->tx_rings[0]->count) in ice_set_ringparam()
3270 vsi->tx_rings[0]->count, new_tx_cnt); in ice_set_ringparam()
[all …]
H A Dice_base.c952 struct ice_tx_ring *tx_ring = vsi->tx_rings[q_id]; in ice_vsi_map_rings_to_vectors()
1132 int ice_vsi_cfg_single_txq(struct ice_vsi *vsi, struct ice_tx_ring **tx_rings, in ice_vsi_cfg_single_txq() argument
1137 if (q_idx >= vsi->alloc_txq || !tx_rings || !tx_rings[q_idx]) in ice_vsi_cfg_single_txq()
1142 return ice_vsi_cfg_txq(vsi, tx_rings[q_idx], qg_buf); in ice_vsi_cfg_single_txq()
1181 return ice_vsi_cfg_txqs(vsi, vsi->tx_rings, vsi->num_txq); in ice_vsi_cfg_lan_txqs()
1423 ice_clean_tx_ring(vsi->tx_rings[q_idx]); in ice_qp_clean_rings()
1448 tx_ring = vsi->tx_rings[q_idx]; in ice_qp_dis()
1495 err = ice_vsi_cfg_single_txq(vsi, vsi->tx_rings, q_idx); in ice_qp_ena()
H A Dice_lib.c83 vsi->tx_rings = devm_kcalloc(dev, vsi->alloc_txq, in ice_vsi_alloc_arrays()
84 sizeof(*vsi->tx_rings), GFP_KERNEL); in ice_vsi_alloc_arrays()
85 if (!vsi->tx_rings) in ice_vsi_alloc_arrays()
125 devm_kfree(dev, vsi->tx_rings); in ice_vsi_alloc_arrays()
323 devm_kfree(dev, vsi->tx_rings); in ice_vsi_free_arrays()
324 vsi->tx_rings = NULL; in ice_vsi_free_arrays()
393 ring = vsi->tx_rings[i]; in ice_vsi_alloc_ring_stats()
1367 if (vsi->tx_rings) { in ice_vsi_clear_rings()
1369 if (vsi->tx_rings[i]) { in ice_vsi_clear_rings()
1370 kfree_rcu(vsi->tx_rings[i], rcu); in ice_vsi_clear_rings()
[all …]
H A Dice_dcb_lib.c209 return vsi->tx_rings[queue_index]->dcb_tc; in ice_dcb_get_tc()
226 tx_ring = vsi->tx_rings[i]; in ice_vsi_cfg_dcb_rings()
243 vsi->tx_rings[i]->dcb_tc = n; in ice_vsi_cfg_dcb_rings()
265 vsi->tx_rings[i]->dcb_tc = first_droptc; in ice_vsi_cfg_dcb_rings()
H A Dice_main.c142 struct ice_tx_ring *tx_ring = vsi->tx_rings[i]; in ice_check_for_hang_subtask()
6125 q_handle = vsi->tx_rings[queue_index]->q_handle; in ice_set_tx_maxrate()
6975 ice_update_vsi_tx_ring_stats(vsi, &tx_stats, vsi->tx_rings, in ice_update_vsi_ring_stats()
7357 ice_clean_tx_ring(vsi->tx_rings[i]); in ice_down()
7417 struct ice_tx_ring *ring = vsi->tx_rings[i]; in ice_vsi_setup_tx_rings()
8299 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) in ice_tx_timeout()
8300 if (txqueue == vsi->tx_rings[i]->q_index) { in ice_tx_timeout()
8301 tx_ring = vsi->tx_rings[i]; in ice_tx_timeout()
8691 tx_ring = vsi->tx_rings[ch->base_q + i]; in ice_chnl_cfg_res()
8984 tx_ring = vsi->tx_rings[ch->base_q + i]; in ice_remove_q_channels()
[all …]
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_ethtool.c2030 ring->tx_pending = vsi->tx_rings[0]->count; in i40e_get_ringparam()
2052 struct i40e_ring *tx_rings = NULL, *rx_rings = NULL; in i40e_set_ringparam() local
2080 if ((new_tx_count == vsi->tx_rings[0]->count) && in i40e_set_ringparam()
2101 vsi->tx_rings[i]->count = new_tx_count; in i40e_set_ringparam()
2119 if (new_tx_count != vsi->tx_rings[0]->count) { in i40e_set_ringparam()
2122 vsi->tx_rings[0]->count, new_tx_count); in i40e_set_ringparam()
2123 tx_rings = kzalloc_objs(struct i40e_ring, tx_alloc_queue_pairs); in i40e_set_ringparam()
2124 if (!tx_rings) { in i40e_set_ringparam()
2133 tx_rings[i] = *vsi->tx_rings[i]; in i40e_set_ringparam()
2134 tx_rings[i].count = new_tx_count; in i40e_set_ringparam()
[all …]
H A Di40e_main.c367 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) { in i40e_tx_timeout()
369 vsi->tx_rings[i]->queue_index) { in i40e_tx_timeout()
370 tx_ring = vsi->tx_rings[i]; in i40e_tx_timeout()
480 if (!vsi->tx_rings) in i40e_get_netdev_stats_struct()
488 ring = READ_ONCE(vsi->tx_rings[i]); in i40e_get_netdev_stats_struct()
549 memset(&vsi->tx_rings[i]->stats, 0, in i40e_vsi_reset_stats()
550 sizeof(vsi->tx_rings[i]->stats)); in i40e_vsi_reset_stats()
551 memset(&vsi->tx_rings[i]->tx_stats, 0, in i40e_vsi_reset_stats()
552 sizeof(vsi->tx_rings[i]->tx_stats)); in i40e_vsi_reset_stats()
913 p = READ_ONCE(vsi->tx_rings[q]); in i40e_update_vsi_stats()
[all …]
H A Di40e_debugfs.c232 struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid()
514 if (!vsi->tx_rings || !vsi->tx_rings[0]->desc) { in i40e_dbg_dump_desc()
526 ring = kmemdup(vsi->tx_rings[ring_id], sizeof(*ring), GFP_KERNEL); in i40e_dbg_dump_desc()
/linux/drivers/thunderbolt/
H A Dnhi.c532 if (!nhi->tx_rings[i]) { in nhi_alloc_hop()
555 if (ring->is_tx && nhi->tx_rings[ring->hop]) { in nhi_alloc_hop()
569 nhi->tx_rings[ring->hop] = ring; in nhi_alloc_hop()
824 ring->nhi->tx_rings[ring->hop] = NULL; in tb_ring_free()
949 ring = nhi->tx_rings[hop]; in nhi_interrupt_work()
1147 if (nhi->tx_rings[i]) in nhi_shutdown()
1368 nhi->tx_rings = devm_kcalloc(&pdev->dev, nhi->hop_count, in nhi_probe()
1369 sizeof(*nhi->tx_rings), GFP_KERNEL); in nhi_probe()
1372 if (!nhi->tx_rings || !nhi->rx_rings) in nhi_probe()
/linux/drivers/net/ethernet/broadcom/
H A Dbcmsysport.c435 ring = &priv->tx_rings[q]; in bcm_sysport_update_tx_stats()
501 ring = &priv->tx_rings[i]; in bcm_sysport_get_stats()
633 bcm_sysport_set_tx_coalesce(&priv->tx_rings[i], ec); in bcm_sysport_set_coalesce()
989 bcm_sysport_tx_reclaim(priv, &priv->tx_rings[q]); in bcm_sysport_tx_reclaim_all()
1143 txr = &priv->tx_rings[ring]; in bcm_sysport_rx_isr()
1175 txr = &priv->tx_rings[ring]; in bcm_sysport_tx_isr()
1300 ring = &priv->tx_rings[queue]; in bcm_sysport_xmit()
1482 struct bcm_sysport_tx_ring *ring = &priv->tx_rings[index]; in bcm_sysport_init_tx_ring()
1573 struct bcm_sysport_tx_ring *ring = &priv->tx_rings[index]; in bcm_sysport_fini_tx_ring()
2338 ring = &priv->tx_rings[q]; in bcm_sysport_map_queues()
[all …]
/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_main.c458 struct iavf_ring *tx_ring = &adapter->tx_rings[t_idx]; in iavf_map_vector_to_txq()
647 adapter->tx_rings[i].tail = hw->hw_addr + IAVF_QTX_TAIL1(i); in iavf_configure_tx()
1452 kfree(adapter->tx_rings); in iavf_free_queues()
1453 adapter->tx_rings = NULL; in iavf_free_queues()
1472 struct iavf_ring *tx_ring = &adapter->tx_rings[i]; in iavf_set_queue_vlan_tag_loc()
1566 adapter->tx_rings = kzalloc_objs(struct iavf_ring, num_active_queues); in iavf_alloc_queues()
1567 if (!adapter->tx_rings) in iavf_alloc_queues()
1577 tx_ring = &adapter->tx_rings[i]; in iavf_alloc_queues()
3083 if (adapter->tx_rings[i].q_shaper.bw_min || in iavf_reconfig_qs_bw()
3084 adapter->tx_rings[i].q_shaper.bw_max) { in iavf_reconfig_qs_bw()
[all …]
H A Diavf_virtchnl.c410 vqpi->txq.ring_len = adapter->tx_rings[i].count; in iavf_configure_queues()
411 vqpi->txq.dma_ring_addr = adapter->tx_rings[i].dma; in iavf_configure_queues()
1778 if (adapter->tx_rings[i].q_shaper_update) in iavf_cfg_queues_bw()
1790 struct iavf_ring *tx_ring = &adapter->tx_rings[i]; in iavf_cfg_queues_bw()
2948 adapter->tx_rings[i].q_shaper_update = false; in iavf_virtchnl_completion()
H A Diavf_ethtool.c355 ring = &adapter->tx_rings[i]; in iavf_get_ethtool_stats()
571 tx_ring = &adapter->tx_rings[queue]; in __iavf_get_coalesce()
631 struct iavf_ring *tx_ring = &adapter->tx_rings[queue]; in iavf_set_itr_per_queue()
H A Diavf.h276 struct iavf_ring *tx_rings; member
/linux/drivers/net/ethernet/broadcom/genet/
H A Dbcmgenet.c1028 tx_rings[num].stats64, packets), \
1030 tx_rings[num].stats64, bytes), \
1032 tx_rings[num].stats64, errors), \
1034 tx_rings[num].stats64, dropped), \
2041 bcmgenet_tx_reclaim(dev, &priv->tx_rings[i++], true); in bcmgenet_tx_reclaim_all()
2138 ring = &priv->tx_rings[index]; in bcmgenet_xmit()
2705 struct bcmgenet_tx_ring *ring = &priv->tx_rings[index]; in bcmgenet_init_tx_ring()
2807 ring = &priv->tx_rings[i]; in bcmgenet_enable_tx_napi()
2819 ring = &priv->tx_rings[i]; in bcmgenet_disable_tx_napi()
2830 ring = &priv->tx_rings[i]; in bcmgenet_fini_tx_napi()
[all …]
H A Dbcmgenet.h606 struct bcmgenet_tx_ring tx_rings[GENET_MAX_MQ_CNT + 1]; member
/linux/drivers/net/ethernet/sun/
H A Dniu.c3610 index = (rp - np->tx_rings); in niu_tx_work()
3757 struct tx_ring_info *rp = &np->tx_rings[i]; in niu_poll_core()
4109 struct tx_ring_info *rp = &np->tx_rings[i]; in niu_slowpath_interrupt()
4187 struct tx_ring_info *rp = &np->tx_rings[i]; in __niu_fastpath_interrupt()
4327 if (np->tx_rings) { in niu_free_channels()
4329 struct tx_ring_info *rp = &np->tx_rings[i]; in niu_free_channels()
4333 kfree(np->tx_rings); in niu_free_channels()
4334 np->tx_rings = NULL; in niu_free_channels()
4474 struct tx_ring_info *tx_rings; in niu_alloc_channels() local
4526 tx_rings = kzalloc_objs(struct tx_ring_info, num_tx_rings); in niu_alloc_channels()
[all …]
/linux/Documentation/networking/
H A Ddriver.rst66 dr = dp->tx_rings[idx];

12