Home
last modified time | relevance | path

Searched refs:num_tx_queues (Results 1 – 25 of 97) sorted by relevance

1234

/linux/drivers/net/ethernet/google/gve/
H A Dgve_ethtool.c96 int num_tx_queues; in gve_get_strings() local
99 num_tx_queues = gve_num_tx_queues(priv); in gve_get_strings()
110 for (i = 0; i < num_tx_queues; i++) in gve_get_strings()
133 int num_tx_queues; in gve_get_sset_count() local
135 num_tx_queues = gve_num_tx_queues(priv); in gve_get_sset_count()
140 (num_tx_queues * NUM_GVE_TX_CNTS); in gve_get_sset_count()
170 int num_tx_queues; in gve_get_ethtool_stats() local
177 num_tx_queues = gve_num_tx_queues(priv); in gve_get_ethtool_stats()
187 tx_qid_to_stats_idx = kmalloc_objs(int, num_tx_queues); in gve_get_ethtool_stats()
192 for (ring = 0; ring < num_tx_queues; ring++) { in gve_get_ethtool_stats()
[all …]
/linux/net/sched/
H A Dsch_mq.c56 for (ntx = 0; ntx < dev->num_tx_queues && priv->qdiscs[ntx]; ntx++) in mq_destroy_common()
85 priv->qdiscs = kzalloc_objs(priv->qdiscs[0], dev->num_tx_queues); in mq_init_common()
89 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mq_init_common()
127 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mq_attach()
158 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mq_dump_common()
184 if (ntx >= dev->num_tx_queues) in mq_queue_get()
274 for (ntx = arg->skip; ntx < dev->num_tx_queues; ntx++) { in mq_walk()
H A Dsch_mqprio.c103 ntx < dev->num_tx_queues && priv->qdiscs[ntx]; in mqprio_destroy()
367 if (dev->num_tx_queues >= TC_H_MIN_PRIORITY) in mqprio_init()
391 priv->qdiscs = kzalloc_objs(priv->qdiscs[0], dev->num_tx_queues); in mqprio_init()
395 for (i = 0; i < dev->num_tx_queues; i++) { in mqprio_init()
439 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mqprio_attach()
457 if (ntx >= dev->num_tx_queues) in mqprio_queue_get()
570 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mqprio_dump()
628 * num_tx_queues. All of these are backed by actual Qdiscs. in mqprio_find()
631 return (ntx <= dev->num_tx_queues) ? ntx : 0; in mqprio_find()
738 for (ntx -= TC_MAX_QUEUE; ntx < dev->num_tx_queues; nt in mqprio_walk()
[all...]
H A Dsch_taprio.c798 if (q->cur_txq[tc] >= dev->num_tx_queues) in taprio_dequeue_tc_priority()
820 for (i = 0; i < dev->num_tx_queues; i++) { in taprio_dequeue_txq_priority()
1195 if (qopt->num_tc > dev->num_tx_queues) { in taprio_parse_mqprio_opt()
2018 for (i = 0; i < dev->num_tx_queues; i++) in taprio_reset()
2042 for (i = 0; i < dev->num_tx_queues; i++) in taprio_destroy()
2094 q->qdiscs = kzalloc_objs(q->qdiscs[0], dev->num_tx_queues); in taprio_init()
2101 for (i = 0; i < dev->num_tx_queues; i++) { in taprio_init()
2135 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in taprio_attach()
2170 if (ntx >= dev->num_tx_queues) in taprio_queue_get()
2457 if (ntx >= dev->num_tx_queues) in taprio_leaf()
[all …]
H A Dsch_generic.c464 for (i = 1; i < dev->num_tx_queues; i++) { in dev_trans_start()
480 for (i = 0; i < dev->num_tx_queues; i++) { in netif_freeze_queues()
506 for (i = 0; i < dev->num_tx_queues; i++) { in netif_unfreeze_queues()
540 for (i = 0; i < dev->num_tx_queues; i++) { in dev_watchdog()
1320 for (i = 0; i < dev->num_tx_queues; i++) { in some_qdisc_is_busy()
1455 for (i = 0; i < dev->num_tx_queues; i++) { in dev_qdisc_change_tx_queue_len()
/linux/drivers/net/ethernet/broadcom/
H A Dbcmsysport.c338 return j + dev->num_tx_queues * NUM_SYSPORT_TXQ_STAT; in bcm_sysport_get_sset_count()
362 for (i = 0; i < dev->num_tx_queues; i++) { in bcm_sysport_get_strings()
434 for (q = 0; q < priv->netdev->num_tx_queues; q++) { in bcm_sysport_update_tx_stats()
498 dev->num_tx_queues * NUM_SYSPORT_TXQ_STAT; in bcm_sysport_get_stats()
500 for (i = 0; i < dev->num_tx_queues; i++) { in bcm_sysport_get_stats()
632 for (i = 0; i < dev->num_tx_queues; i++) in bcm_sysport_set_coalesce()
988 for (q = 0; q < priv->netdev->num_tx_queues; q++) in bcm_sysport_tx_reclaim_all()
1138 for (ring = 0; ring < dev->num_tx_queues; ring++) { in bcm_sysport_rx_isr()
1171 for (ring = 0; ring < dev->num_tx_queues; ring++) { in bcm_sysport_tx_isr()
2003 for (i = 0; i < dev->num_tx_queues; i++) { in bcm_sysport_open()
[all …]
/linux/drivers/net/ethernet/engleder/
H A Dtsnep_xdp.c28 queue_id >= adapter->num_tx_queues) in tsnep_xdp_enable_pool()
64 queue_id >= adapter->num_tx_queues) in tsnep_xdp_disable_pool()
/linux/drivers/net/ethernet/intel/igc/
H A Digc_tsn.c165 for (int i = 0; i < adapter->num_tx_queues; i++) { in igc_fpe_clear_preempt_queue()
198 for (int i = 0; i < adapter->num_tx_queues; i++) { in igc_fpe_save_preempt_queue()
209 for (i = 0; i < adapter->num_tx_queues; i++) { in is_any_launchtime()
223 for (i = 0; i < adapter->num_tx_queues; i++) { in is_cbs_enabled()
379 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_tsn_disable_offload()
468 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_tsn_enable_offload()
H A Digc_xdp.c68 queue_id >= adapter->num_tx_queues) in igc_xdp_enable_pool()
126 queue_id >= adapter->num_tx_queues) in igc_xdp_disable_pool()
H A Digc_dump.c135 for (n = 0; n < adapter->num_tx_queues; n++) { in igc_rings_dump()
166 for (n = 0; n < adapter->num_tx_queues; n++) { in igc_rings_dump()
H A Digc_main.c309 for (i = 0; i < adapter->num_tx_queues; i++) in igc_free_all_tx_resources()
321 for (i = 0; i < adapter->num_tx_queues; i++) in igc_clean_all_tx_rings()
346 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_disable_all_tx_rings_hw()
402 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_setup_all_tx_resources()
775 for (i = 0; i < adapter->num_tx_queues; i++) in igc_configure_tx()
1726 if (r_idx >= adapter->num_tx_queues) in igc_tx_queue_mapping()
1727 r_idx = r_idx % adapter->num_tx_queues; in igc_tx_queue_mapping()
2485 while (index >= adapter->num_tx_queues) in igc_get_tx_ring()
2486 index -= adapter->num_tx_queues; in igc_get_tx_ring()
4447 adapter->num_tx_queues = 0; in igc_free_q_vectors()
[all …]
H A Digc_ethtool.c120 (((struct igc_adapter *)netdev_priv(netdev))->num_tx_queues * \
620 for (i = 0; i < adapter->num_tx_queues; i++) in igc_ethtool_set_ringparam()
629 if (adapter->num_tx_queues > adapter->num_rx_queues) in igc_ethtool_set_ringparam()
630 temp_ring = vmalloc_array(adapter->num_tx_queues, in igc_ethtool_set_ringparam()
648 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_ethtool_set_ringparam()
663 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_ethtool_set_ringparam()
782 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_ethtool_get_strings()
840 for (j = 0; j < adapter->num_tx_queues; j++) { in igc_ethtool_get_stats()
/linux/drivers/net/ethernet/intel/fm10k/
H A Dfm10k_netdev.c59 for (i = 0; i < interface->num_tx_queues; i++) { in fm10k_setup_all_tx_resources()
226 for (i = 0; i < interface->num_tx_queues; i++) in fm10k_clean_all_tx_rings()
238 int i = interface->num_tx_queues; in fm10k_free_all_tx_resources()
462 interface->num_tx_queues); in fm10k_open()
513 int num_tx_queues = READ_ONCE(interface->num_tx_queues); in fm10k_xmit_frame() local
517 if (!num_tx_queues) in fm10k_xmit_frame()
572 if (r_idx >= num_tx_queues) in fm10k_xmit_frame()
573 r_idx %= num_tx_queues; in fm10k_xmit_frame()
591 if (txqueue >= interface->num_tx_queues) { in fm10k_tx_timeout()
1241 for (i = 0; i < interface->num_tx_queues; i++) { in fm10k_get_stats64()
H A Dfm10k_main.c1510 interface->num_tx_queues = rss_i * pcs; in fm10k_set_qos_queues()
1536 interface->num_tx_queues = rss_i; in fm10k_set_rss_queues()
1571 interface->num_tx_queues = 0; in fm10k_reset_num_queues()
1718 unsigned int txr_remaining = interface->num_tx_queues; in fm10k_alloc_q_vectors()
1813 v_budget = max(interface->num_rx_queues, interface->num_tx_queues); in fm10k_init_msix_capability()
1899 for (i = 0; i < interface->num_tx_queues; i++) in fm10k_cache_ring_rss()
/linux/drivers/crypto/caam/
H A Ddpseci.h53 u8 num_tx_queues; member
77 u8 num_tx_queues; member
H A Ddpseci_cmd.h71 u8 num_tx_queues; member
H A Ddpseci.c178 attr->num_tx_queues = rsp_params->num_tx_queues; in dpseci_get_attributes()
/linux/drivers/net/
H A Difb.c163 for (i = 0; i < dev->num_tx_queues; i++,txp++) { in ifb_stats64()
190 txp = kzalloc_objs(*txp, dev->num_tx_queues); in ifb_dev_init()
194 for (i = 0; i < dev->num_tx_queues; i++,txp++) { in ifb_dev_init()
302 for (i = 0; i < dev->num_tx_queues; i++,txp++) { in ifb_dev_free()
/linux/drivers/net/dsa/microchip/
H A Dksz_dcb.c339 ipm = ieee8021q_tt_to_tc(tt, dev->info->num_tx_queues); in ksz_init_global_dscp_map()
391 dev->info->num_tx_queues); in ksz_port_del_dscp_prio()
584 dev->info->num_tx_queues); in ksz_dcb_init_port()
/linux/drivers/net/ethernet/wangxun/libwx/
H A Dwx_lib.c1676 if (r_idx >= wx->num_tx_queues) in wx_xmit_frame()
1677 r_idx = r_idx % wx->num_tx_queues; in wx_xmit_frame()
1828 wx->num_tx_queues = vmdq_i * rss_i; in wx_set_vmdq_queues()
1874 wx->num_tx_queues = f->indices; in wx_set_rss_queues()
1881 wx->num_tx_queues = 1; in wx_set_num_queues()
1904 nvecs = max(wx->num_rx_queues, wx->num_tx_queues); in wx_acquire_msix_vectors()
2024 for (i = 0; i < wx->num_tx_queues; i++, reg_idx++) { in wx_cache_ring_vmdq()
2038 for (i = 0; i < wx->num_tx_queues; i++) in wx_cache_ring_vmdq()
2063 for (i = 0; i < wx->num_tx_queues; i++) in wx_cache_ring_rss()
2229 unsigned int txr_remaining = wx->num_tx_queues; in wx_alloc_q_vectors()
[all …]
/linux/drivers/net/ethernet/intel/ixgbevf/
H A Dixgbevf_main.c1760 for (i = 0; i < adapter->num_tx_queues; i++) in ixgbevf_configure_tx()
2186 unsigned int num_tx_queues = adapter->num_tx_queues; in ixgbevf_configure_dcb() local
2201 num_tx_queues = 1; in ixgbevf_configure_dcb()
2212 (adapter->num_tx_queues != num_tx_queues)) { in ixgbevf_configure_dcb()
2513 for (i = 0; i < adapter->num_tx_queues; i++) in ixgbevf_clean_all_tx_rings()
2548 for (i = 0; i < adapter->num_tx_queues; i++) { in ixgbevf_down()
2657 adapter->num_tx_queues = 1; in ixgbevf_set_num_queues()
2689 adapter->num_tx_queues = rss; in ixgbevf_set_num_queues()
2715 v_budget = max(adapter->num_rx_queues, adapter->num_tx_queues); in ixgbevf_set_interrupt_capability()
2914 int txr_remaining = adapter->num_tx_queues; in ixgbevf_alloc_q_vectors()
[all …]
/linux/drivers/net/ethernet/freescale/
H A Dgianfar.c133 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_init_tx_rx_base()
247 for_each_set_bit(i, &tx_mask, priv->num_tx_queues) { in gfar_configure_coalescing()
289 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_get_stats64()
415 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_alloc_tx_queues()
447 for (i = 0; i < priv->num_tx_queues; i++) in gfar_free_tx_queues()
558 for_each_set_bit(i, &grp->tx_bit_map, priv->num_tx_queues) { in gfar_parse_group()
561 grp->num_tx_queues++; in gfar_parse_group()
682 priv->num_tx_queues = num_tx_qs; in gfar_of_init()
1121 for (i = 0; i < priv->num_tx_queues; i++) { in free_skb_resources()
1282 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_init_bds()
[all …]
/linux/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c697 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init()
704 for (i = 0; i < dev->num_tx_queues; i++) { in hfi1_ipoib_txreq_init()
798 for (i = 0; i < priv->netdev->num_tx_queues; i++) { in hfi1_ipoib_txreq_deinit()
821 for (i = 0; i < dev->num_tx_queues; i++) { in hfi1_ipoib_napi_tx_enable()
833 for (i = 0; i < dev->num_tx_queues; i++) { in hfi1_ipoib_napi_tx_disable()
/linux/include/net/
H A Dnetdev_lock.h127 for (i = 0; i < (dev)->num_tx_queues; i++) \
/linux/net/core/
H A Ddev.h262 netdev->num_tx_queues); in netdev_set_defer_hard_irqs()
308 netdev->num_tx_queues); in netdev_set_gro_flush_timeout()

1234