| /linux/drivers/net/ethernet/aquantia/atlantic/ |
| H A D | aq_vec.c | 18 unsigned int tx_rings; member 41 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_poll() 119 self->tx_rings = 0; in aq_vec_alloc() 144 ++self->tx_rings; in aq_vec_ring_alloc() 190 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_init() 232 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_start() 256 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_stop() 276 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_deinit() 305 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_ring_free() 314 self->tx_rings = 0; in aq_vec_ring_free() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | nfp_net_dp.c | 187 dp->tx_rings = kzalloc_objs(*dp->tx_rings, dp->num_tx_rings); in nfp_net_tx_rings_prepare() 188 if (!dp->tx_rings) in nfp_net_tx_rings_prepare() 205 nfp_net_tx_ring_init(&dp->tx_rings[r], dp, in nfp_net_tx_rings_prepare() 208 if (nfp_net_tx_ring_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare() 211 if (nfp_net_tx_ring_bufs_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare() 219 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare() 221 nfp_net_tx_ring_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare() 227 kfree(dp->tx_rings); in nfp_net_tx_rings_prepare() 236 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free() 237 nfp_net_tx_ring_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free() [all …]
|
| H A D | nfp_net_common.c | 797 idx < dp->num_stack_tx_rings ? &dp->tx_rings[idx] : NULL; in nfp_net_vector_assign_rings() 800 &dp->tx_rings[dp->num_stack_tx_rings + idx] : NULL; in nfp_net_vector_assign_rings() 968 nfp_net_tx_ring_reset(&nn->dp, &nn->dp.tx_rings[r]); in nfp_net_clear_config_and_disable() 1001 nfp_net_tx_ring_hw_cfg_write(nn, &nn->dp.tx_rings[r], r); in nfp_net_set_config_and_enable() 1574 new->tx_rings = NULL; in nfp_net_clone_dp()
|
| H A D | nfp_net.h | 520 struct nfp_net_tx_ring *tx_rings; member
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_ethtool.c | 1290 tx_ring = test_vsi->tx_rings[0]; in ice_loopback_test() 1959 tx_ring = READ_ONCE(vsi->tx_rings[j]); in __ice_get_ethtool_stats() 3173 if (vsi->tx_rings && vsi->rx_rings) { in ice_get_ringparam() 3175 ring->tx_pending = vsi->tx_rings[0]->count; in ice_get_ringparam() 3199 struct ice_tx_ring *tx_rings = NULL; in ice_set_ringparam() local 3220 if (!vsi->tx_rings || !vsi->rx_rings) in ice_set_ringparam() 3235 if (new_tx_cnt == vsi->tx_rings[0]->count && in ice_set_ringparam() 3259 vsi->tx_rings[i]->count = new_tx_cnt; in ice_set_ringparam() 3273 if (new_tx_cnt == vsi->tx_rings[0]->count) in ice_set_ringparam() 3278 vsi->tx_rings[0]->count, new_tx_cnt); in ice_set_ringparam() [all …]
|
| H A D | ice_base.c | 952 struct ice_tx_ring *tx_ring = vsi->tx_rings[q_id]; in ice_vsi_map_rings_to_vectors() 1132 int ice_vsi_cfg_single_txq(struct ice_vsi *vsi, struct ice_tx_ring **tx_rings, in ice_vsi_cfg_single_txq() argument 1137 if (q_idx >= vsi->alloc_txq || !tx_rings || !tx_rings[q_idx]) in ice_vsi_cfg_single_txq() 1142 return ice_vsi_cfg_txq(vsi, tx_rings[q_idx], qg_buf); in ice_vsi_cfg_single_txq() 1181 return ice_vsi_cfg_txqs(vsi, vsi->tx_rings, vsi->num_txq); in ice_vsi_cfg_lan_txqs() 1423 ice_clean_tx_ring(vsi->tx_rings[q_idx]); in ice_qp_clean_rings() 1448 tx_ring = vsi->tx_rings[q_idx]; in ice_qp_dis() 1495 err = ice_vsi_cfg_single_txq(vsi, vsi->tx_rings, q_idx); in ice_qp_ena()
|
| H A D | ice_lib.c | 83 vsi->tx_rings = devm_kcalloc(dev, vsi->alloc_txq, in ice_vsi_alloc_arrays() 84 sizeof(*vsi->tx_rings), GFP_KERNEL); in ice_vsi_alloc_arrays() 85 if (!vsi->tx_rings) in ice_vsi_alloc_arrays() 125 devm_kfree(dev, vsi->tx_rings); in ice_vsi_alloc_arrays() 323 devm_kfree(dev, vsi->tx_rings); in ice_vsi_free_arrays() 324 vsi->tx_rings = NULL; in ice_vsi_free_arrays() 393 ring = vsi->tx_rings[i]; in ice_vsi_alloc_ring_stats() 1367 if (vsi->tx_rings) { in ice_vsi_clear_rings() 1369 if (vsi->tx_rings[i]) { in ice_vsi_clear_rings() 1370 kfree_rcu(vsi->tx_rings[i], rcu); in ice_vsi_clear_rings() [all …]
|
| H A D | ice_main.c | 142 struct ice_tx_ring *tx_ring = vsi->tx_rings[i]; in ice_check_for_hang_subtask() 6060 q_handle = vsi->tx_rings[queue_index]->q_handle; in ice_set_tx_maxrate() 6910 ice_update_vsi_tx_ring_stats(vsi, &tx_stats, vsi->tx_rings, in ice_update_vsi_ring_stats() 7292 ice_clean_tx_ring(vsi->tx_rings[i]); in ice_down() 7352 struct ice_tx_ring *ring = vsi->tx_rings[i]; in ice_vsi_setup_tx_rings() 8234 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) in ice_tx_timeout() 8235 if (txqueue == vsi->tx_rings[i]->q_index) { in ice_tx_timeout() 8236 tx_ring = vsi->tx_rings[i]; in ice_tx_timeout() 8626 tx_ring = vsi->tx_rings[ch->base_q + i]; in ice_chnl_cfg_res() 8919 tx_ring = vsi->tx_rings[ch->base_q + i]; in ice_remove_q_channels() [all …]
|
| H A D | ice_txrx.c | 49 tx_ring = vsi->tx_rings[0]; in ice_prgm_fdir_fltr() 2279 tx_ring = vsi->tx_rings[skb->queue_mapping]; in ice_start_xmit()
|
| /linux/drivers/thunderbolt/ |
| H A D | nhi.c | 532 if (!nhi->tx_rings[i]) { in nhi_alloc_hop() 555 if (ring->is_tx && nhi->tx_rings[ring->hop]) { in nhi_alloc_hop() 569 nhi->tx_rings[ring->hop] = ring; in nhi_alloc_hop() 824 ring->nhi->tx_rings[ring->hop] = NULL; in tb_ring_free() 949 ring = nhi->tx_rings[hop]; in nhi_interrupt_work() 1147 if (nhi->tx_rings[i]) in nhi_shutdown() 1368 nhi->tx_rings = devm_kcalloc(&pdev->dev, nhi->hop_count, in nhi_probe() 1369 sizeof(*nhi->tx_rings), GFP_KERNEL); in nhi_probe() 1372 if (!nhi->tx_rings || !nhi->rx_rings) in nhi_probe()
|
| /linux/drivers/net/ethernet/broadcom/ |
| H A D | bcmsysport.c | 435 ring = &priv->tx_rings[q]; in bcm_sysport_update_tx_stats() 501 ring = &priv->tx_rings[i]; in bcm_sysport_get_stats() 633 bcm_sysport_set_tx_coalesce(&priv->tx_rings[i], ec); in bcm_sysport_set_coalesce() 989 bcm_sysport_tx_reclaim(priv, &priv->tx_rings[q]); in bcm_sysport_tx_reclaim_all() 1143 txr = &priv->tx_rings[ring]; in bcm_sysport_rx_isr() 1175 txr = &priv->tx_rings[ring]; in bcm_sysport_tx_isr() 1300 ring = &priv->tx_rings[queue]; in bcm_sysport_xmit() 1482 struct bcm_sysport_tx_ring *ring = &priv->tx_rings[index]; in bcm_sysport_init_tx_ring() 1573 struct bcm_sysport_tx_ring *ring = &priv->tx_rings[index]; in bcm_sysport_fini_tx_ring() 2338 ring = &priv->tx_rings[q]; in bcm_sysport_map_queues() [all …]
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_main.c | 458 struct iavf_ring *tx_ring = &adapter->tx_rings[t_idx]; in iavf_map_vector_to_txq() 647 adapter->tx_rings[i].tail = hw->hw_addr + IAVF_QTX_TAIL1(i); in iavf_configure_tx() 1432 kfree(adapter->tx_rings); in iavf_free_queues() 1433 adapter->tx_rings = NULL; in iavf_free_queues() 1452 struct iavf_ring *tx_ring = &adapter->tx_rings[i]; in iavf_set_queue_vlan_tag_loc() 1546 adapter->tx_rings = kzalloc_objs(struct iavf_ring, num_active_queues); in iavf_alloc_queues() 1547 if (!adapter->tx_rings) in iavf_alloc_queues() 1557 tx_ring = &adapter->tx_rings[i]; in iavf_alloc_queues() 3063 if (adapter->tx_rings[i].q_shaper.bw_min || in iavf_reconfig_qs_bw() 3064 adapter->tx_rings[i].q_shaper.bw_max) { in iavf_reconfig_qs_bw() [all …]
|
| H A D | iavf_virtchnl.c | 410 vqpi->txq.ring_len = adapter->tx_rings[i].count; in iavf_configure_queues() 411 vqpi->txq.dma_ring_addr = adapter->tx_rings[i].dma; in iavf_configure_queues() 1753 if (adapter->tx_rings[i].q_shaper_update) in iavf_cfg_queues_bw() 1765 struct iavf_ring *tx_ring = &adapter->tx_rings[i]; in iavf_cfg_queues_bw() 2944 adapter->tx_rings[i].q_shaper_update = false; in iavf_virtchnl_completion()
|
| H A D | iavf_ethtool.c | 354 tx_ring = &adapter->tx_rings[i]; in iavf_get_ethtool_stats() 570 tx_ring = &adapter->tx_rings[queue]; in __iavf_get_coalesce() 630 struct iavf_ring *tx_ring = &adapter->tx_rings[queue]; in iavf_set_itr_per_queue()
|
| H A D | iavf.h | 275 struct iavf_ring *tx_rings; member
|
| H A D | iavf_txrx.c | 198 tx_ring = &vsi->back->tx_rings[i]; in iavf_detect_recover_hung() 2394 struct iavf_ring *tx_ring = &adapter->tx_rings[skb->queue_mapping]; in iavf_xmit_frame()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_main.c | 367 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) { in i40e_tx_timeout() 369 vsi->tx_rings[i]->queue_index) { in i40e_tx_timeout() 370 tx_ring = vsi->tx_rings[i]; in i40e_tx_timeout() 480 if (!vsi->tx_rings) in i40e_get_netdev_stats_struct() 488 ring = READ_ONCE(vsi->tx_rings[i]); in i40e_get_netdev_stats_struct() 549 memset(&vsi->tx_rings[i]->stats, 0, in i40e_vsi_reset_stats() 550 sizeof(vsi->tx_rings[i]->stats)); in i40e_vsi_reset_stats() 551 memset(&vsi->tx_rings[i]->tx_stats, 0, in i40e_vsi_reset_stats() 552 sizeof(vsi->tx_rings[i]->tx_stats)); in i40e_vsi_reset_stats() 913 p = READ_ONCE(vsi->tx_rings[q]); in i40e_update_vsi_stats() [all …]
|
| H A D | i40e_debugfs.c | 232 struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid() 514 if (!vsi->tx_rings || !vsi->tx_rings[0]->desc) { in i40e_dbg_dump_desc() 526 ring = kmemdup(vsi->tx_rings[ring_id], sizeof(*ring), GFP_KERNEL); in i40e_dbg_dump_desc()
|
| /linux/drivers/net/ethernet/broadcom/genet/ |
| H A D | bcmgenet.c | 1028 tx_rings[num].stats64, packets), \ 1030 tx_rings[num].stats64, bytes), \ 1032 tx_rings[num].stats64, errors), \ 1034 tx_rings[num].stats64, dropped), \ 2043 bcmgenet_tx_reclaim(dev, &priv->tx_rings[i++], true); in bcmgenet_tx_reclaim_all() 2140 ring = &priv->tx_rings[index]; in bcmgenet_xmit() 2707 struct bcmgenet_tx_ring *ring = &priv->tx_rings[index]; in bcmgenet_init_tx_ring() 2809 ring = &priv->tx_rings[i]; in bcmgenet_enable_tx_napi() 2821 ring = &priv->tx_rings[i]; in bcmgenet_disable_tx_napi() 2832 ring = &priv->tx_rings[i]; in bcmgenet_fini_tx_napi() [all …]
|
| H A D | bcmgenet.h | 606 struct bcmgenet_tx_ring tx_rings[GENET_MAX_MQ_CNT + 1]; member
|
| /linux/drivers/net/wireless/realtek/rtw88/ |
| H A D | pci.h | 223 struct rtw_pci_tx_ring tx_rings[RTK_MAX_TX_QUEUE_NUM]; member
|
| /linux/drivers/net/ethernet/sun/ |
| H A D | niu.c | 3610 index = (rp - np->tx_rings); in niu_tx_work() 3757 struct tx_ring_info *rp = &np->tx_rings[i]; in niu_poll_core() 4109 struct tx_ring_info *rp = &np->tx_rings[i]; in niu_slowpath_interrupt() 4187 struct tx_ring_info *rp = &np->tx_rings[i]; in __niu_fastpath_interrupt() 4327 if (np->tx_rings) { in niu_free_channels() 4329 struct tx_ring_info *rp = &np->tx_rings[i]; in niu_free_channels() 4333 kfree(np->tx_rings); in niu_free_channels() 4334 np->tx_rings = NULL; in niu_free_channels() 4474 struct tx_ring_info *tx_rings; in niu_alloc_channels() local 4526 tx_rings = kzalloc_objs(struct tx_ring_info, num_tx_rings); in niu_alloc_channels() [all …]
|
| /linux/Documentation/networking/ |
| H A D | driver.rst | 66 dr = dp->tx_rings[idx];
|
| /linux/drivers/net/ethernet/aquantia/atlantic/hw_atl2/ |
| H A D | hw_atl2.c | 38 .tx_rings = HW_ATL2_TX_RINGS, \
|
| /linux/drivers/net/ethernet/qlogic/qlcnic/ |
| H A D | qlcnic_main.c | 3977 u8 tx_rings, rx_rings; in qlcnic_setup_rings() local 3983 tx_rings = adapter->drv_tss_rings; in qlcnic_setup_rings() 3988 err = qlcnic_set_real_num_queues(adapter, tx_rings, rx_rings); in qlcnic_setup_rings() 4014 if ((tx_rings != adapter->drv_tx_rings) || in qlcnic_setup_rings()
|