| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_lib.c | 48 for (i = 0; i < adapter->num_tx_queues; i++, reg_idx++) { in ixgbe_cache_ring_dcb_sriov() 78 for (i = fcoe->offset; i < adapter->num_tx_queues; i++) { in ixgbe_cache_ring_dcb_sriov() 225 for (i = 0; i < adapter->num_tx_queues; i++, reg_idx++) { in ixgbe_cache_ring_sriov() 239 for (; i < adapter->num_tx_queues; i++, reg_idx++) in ixgbe_cache_ring_sriov() 262 for (i = 0, reg_idx = 0; i < adapter->num_tx_queues; i++, reg_idx++) in ixgbe_cache_ring_rss() 385 adapter->num_tx_queues = vmdq_i * tcs; in ixgbe_set_dcb_sriov_queues() 404 adapter->num_tx_queues += fcoe_i; in ixgbe_set_dcb_sriov_queues() 441 rss_i = dev->num_tx_queues / tcs; in ixgbe_set_dcb_queues() 483 adapter->num_tx_queues = rss_i * tcs; in ixgbe_set_dcb_queues() 556 adapter->num_tx_queues = vmdq_i * rss_i; in ixgbe_set_sriov_queues() [all …]
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_ethtool.c | 96 int num_tx_queues; in gve_get_strings() local 99 num_tx_queues = gve_num_tx_queues(priv); in gve_get_strings() 110 for (i = 0; i < num_tx_queues; i++) in gve_get_strings() 133 int num_tx_queues; in gve_get_sset_count() local 135 num_tx_queues = gve_num_tx_queues(priv); in gve_get_sset_count() 140 (num_tx_queues * NUM_GVE_TX_CNTS); in gve_get_sset_count() 170 int num_tx_queues; in gve_get_ethtool_stats() local 177 num_tx_queues = gve_num_tx_queues(priv); in gve_get_ethtool_stats() 187 tx_qid_to_stats_idx = kmalloc_objs(int, num_tx_queues); in gve_get_ethtool_stats() 192 for (ring = 0; ring < num_tx_queues; ring++) { in gve_get_ethtool_stats() [all …]
|
| /linux/net/sched/ |
| H A D | sch_mq.c | 56 for (ntx = 0; ntx < dev->num_tx_queues && priv->qdiscs[ntx]; ntx++) in mq_destroy_common() 85 priv->qdiscs = kzalloc_objs(priv->qdiscs[0], dev->num_tx_queues); in mq_init_common() 89 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mq_init_common() 127 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mq_attach() 158 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mq_dump_common() 184 if (ntx >= dev->num_tx_queues) in mq_queue_get() 274 for (ntx = arg->skip; ntx < dev->num_tx_queues; ntx++) { in mq_walk()
|
| H A D | sch_mqprio.c | 103 ntx < dev->num_tx_queues && priv->qdiscs[ntx]; in mqprio_destroy() 367 if (dev->num_tx_queues >= TC_H_MIN_PRIORITY) in mqprio_init() 391 priv->qdiscs = kzalloc_objs(priv->qdiscs[0], dev->num_tx_queues); in mqprio_init() 395 for (i = 0; i < dev->num_tx_queues; i++) { in mqprio_init() 439 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mqprio_attach() 457 if (ntx >= dev->num_tx_queues) in mqprio_queue_get() 570 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in mqprio_dump() 631 return (ntx <= dev->num_tx_queues) ? ntx : 0; in mqprio_find() 738 for (ntx -= TC_MAX_QUEUE; ntx < dev->num_tx_queues; ntx++) { in mqprio_walk()
|
| H A D | sch_generic.c | 439 for (i = 1; i < dev->num_tx_queues; i++) { in dev_trans_start() 455 for (i = 0; i < dev->num_tx_queues; i++) { in netif_freeze_queues() 481 for (i = 0; i < dev->num_tx_queues; i++) { in netif_unfreeze_queues() 515 for (i = 0; i < dev->num_tx_queues; i++) { in dev_watchdog() 1295 for (i = 0; i < dev->num_tx_queues; i++) { in some_qdisc_is_busy() 1426 for (i = 0; i < dev->num_tx_queues; i++) { in dev_qdisc_change_tx_queue_len()
|
| H A D | sch_taprio.c | 798 if (q->cur_txq[tc] >= dev->num_tx_queues) in taprio_dequeue_tc_priority() 820 for (i = 0; i < dev->num_tx_queues; i++) { in taprio_dequeue_txq_priority() 1194 if (qopt->num_tc > dev->num_tx_queues) { in taprio_parse_mqprio_opt() 2017 for (i = 0; i < dev->num_tx_queues; i++) in taprio_reset() 2041 for (i = 0; i < dev->num_tx_queues; i++) in taprio_destroy() 2093 q->qdiscs = kzalloc_objs(q->qdiscs[0], dev->num_tx_queues); in taprio_init() 2100 for (i = 0; i < dev->num_tx_queues; i++) { in taprio_init() 2134 for (ntx = 0; ntx < dev->num_tx_queues; ntx++) { in taprio_attach() 2169 if (ntx >= dev->num_tx_queues) in taprio_queue_get() 2453 if (ntx >= dev->num_tx_queues) in taprio_leaf() [all …]
|
| /linux/drivers/net/ethernet/broadcom/ |
| H A D | bcmsysport.c | 338 return j + dev->num_tx_queues * NUM_SYSPORT_TXQ_STAT; in bcm_sysport_get_sset_count() 362 for (i = 0; i < dev->num_tx_queues; i++) { in bcm_sysport_get_strings() 434 for (q = 0; q < priv->netdev->num_tx_queues; q++) { in bcm_sysport_update_tx_stats() 498 dev->num_tx_queues * NUM_SYSPORT_TXQ_STAT; in bcm_sysport_get_stats() 500 for (i = 0; i < dev->num_tx_queues; i++) { in bcm_sysport_get_stats() 632 for (i = 0; i < dev->num_tx_queues; i++) in bcm_sysport_set_coalesce() 988 for (q = 0; q < priv->netdev->num_tx_queues; q++) in bcm_sysport_tx_reclaim_all() 1138 for (ring = 0; ring < dev->num_tx_queues; ring++) { in bcm_sysport_rx_isr() 1171 for (ring = 0; ring < dev->num_tx_queues; ring++) { in bcm_sysport_tx_isr() 2003 for (i = 0; i < dev->num_tx_queues; i++) { in bcm_sysport_open() [all …]
|
| /linux/drivers/net/ethernet/engleder/ |
| H A D | tsnep_xdp.c | 28 queue_id >= adapter->num_tx_queues) in tsnep_xdp_enable_pool() 64 queue_id >= adapter->num_tx_queues) in tsnep_xdp_disable_pool()
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_tsn.c | 165 for (int i = 0; i < adapter->num_tx_queues; i++) { in igc_fpe_clear_preempt_queue() 198 for (int i = 0; i < adapter->num_tx_queues; i++) { in igc_fpe_save_preempt_queue() 209 for (i = 0; i < adapter->num_tx_queues; i++) { in is_any_launchtime() 223 for (i = 0; i < adapter->num_tx_queues; i++) { in is_cbs_enabled() 379 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_tsn_disable_offload() 468 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_tsn_enable_offload()
|
| H A D | igc_xdp.c | 68 queue_id >= adapter->num_tx_queues) in igc_xdp_enable_pool() 126 queue_id >= adapter->num_tx_queues) in igc_xdp_disable_pool()
|
| H A D | igc_dump.c | 135 for (n = 0; n < adapter->num_tx_queues; n++) { in igc_rings_dump() 166 for (n = 0; n < adapter->num_tx_queues; n++) { in igc_rings_dump()
|
| H A D | igc_main.c | 309 for (i = 0; i < adapter->num_tx_queues; i++) in igc_free_all_tx_resources() 321 for (i = 0; i < adapter->num_tx_queues; i++) in igc_clean_all_tx_rings() 346 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_disable_all_tx_rings_hw() 402 for (i = 0; i < adapter->num_tx_queues; i++) { in igc_setup_all_tx_resources() 775 for (i = 0; i < adapter->num_tx_queues; i++) in igc_configure_tx() 1726 if (r_idx >= adapter->num_tx_queues) in igc_tx_queue_mapping() 1727 r_idx = r_idx % adapter->num_tx_queues; in igc_tx_queue_mapping() 2485 while (index >= adapter->num_tx_queues) in igc_get_tx_ring() 2486 index -= adapter->num_tx_queues; in igc_get_tx_ring() 4447 adapter->num_tx_queues = 0; in igc_free_q_vectors() [all …]
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| H A D | fm10k_netdev.c | 59 for (i = 0; i < interface->num_tx_queues; i++) { in fm10k_setup_all_tx_resources() 226 for (i = 0; i < interface->num_tx_queues; i++) in fm10k_clean_all_tx_rings() 238 int i = interface->num_tx_queues; in fm10k_free_all_tx_resources() 462 interface->num_tx_queues); in fm10k_open() 513 int num_tx_queues = READ_ONCE(interface->num_tx_queues); in fm10k_xmit_frame() local 517 if (!num_tx_queues) in fm10k_xmit_frame() 572 if (r_idx >= num_tx_queues) in fm10k_xmit_frame() 573 r_idx %= num_tx_queues; in fm10k_xmit_frame() 591 if (txqueue >= interface->num_tx_queues) { in fm10k_tx_timeout() 1241 for (i = 0; i < interface->num_tx_queues; i++) { in fm10k_get_stats64()
|
| /linux/drivers/net/vmxnet3/ |
| H A D | vmxnet3_drv.c | 221 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_check_link() 230 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_check_link() 258 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_process_events() 562 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_tq_destroy_all() 662 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_tq_cleanup_all() 940 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_tq_init_all() 1439 BUG_ON(skb->queue_mapping > adapter->num_tx_queues); in vmxnet3_xmit_frame() 2358 for (i = 0; i < adapter->num_tx_queues; i++) in vmxnet3_do_poll() 2435 for (i = 0; i < adapter->num_tx_queues; i++) { in vmxnet3_msix_tx() 2560 for (i = 0; i < adapter->num_tx_queues; i++) { in vmxnet3_request_irqs() [all …]
|
| /linux/drivers/net/dsa/microchip/ |
| H A D | ksz_common.c | 1502 .num_tx_queues = 4, 1526 .num_tx_queues = 4, 1557 .num_tx_queues = 4, 1598 .num_tx_queues = 4, 1625 .num_tx_queues = 4, 1652 .num_tx_queues = 4, 1686 .num_tx_queues = 4, 1709 .num_tx_queues = 4, 1733 .num_tx_queues = 4, 1771 .num_tx_queues = 4, [all …]
|
| H A D | ksz_dcb.c | 339 ipm = ieee8021q_tt_to_tc(tt, dev->info->num_tx_queues); in ksz_init_global_dscp_map() 391 dev->info->num_tx_queues); in ksz_port_del_dscp_prio() 584 dev->info->num_tx_queues); in ksz_dcb_init_port()
|
| /linux/drivers/net/ |
| H A D | ifb.c | 163 for (i = 0; i < dev->num_tx_queues; i++,txp++) { in ifb_stats64() 190 txp = kzalloc_objs(*txp, dev->num_tx_queues); in ifb_dev_init() 194 for (i = 0; i < dev->num_tx_queues; i++,txp++) { in ifb_dev_init() 302 for (i = 0; i < dev->num_tx_queues; i++,txp++) { in ifb_dev_free()
|
| /linux/drivers/crypto/caam/ |
| H A D | dpseci.h | 53 u8 num_tx_queues; member 77 u8 num_tx_queues; member
|
| /linux/drivers/net/ethernet/wangxun/ngbe/ |
| H A D | ngbe_ethtool.c | 72 for (i = 0; i < wx->num_tx_queues; i++) in ngbe_set_ringparam() 83 i = max_t(int, wx->num_tx_queues, wx->num_rx_queues); in ngbe_set_ringparam()
|
| /linux/drivers/net/ethernet/wangxun/libwx/ |
| H A D | wx_lib.c | 1676 if (r_idx >= wx->num_tx_queues) in wx_xmit_frame() 1677 r_idx = r_idx % wx->num_tx_queues; in wx_xmit_frame() 1828 wx->num_tx_queues = vmdq_i * rss_i; in wx_set_vmdq_queues() 1874 wx->num_tx_queues = f->indices; in wx_set_rss_queues() 1881 wx->num_tx_queues = 1; in wx_set_num_queues() 1904 nvecs = max(wx->num_rx_queues, wx->num_tx_queues); in wx_acquire_msix_vectors() 2024 for (i = 0; i < wx->num_tx_queues; i++, reg_idx++) { in wx_cache_ring_vmdq() 2038 for (i = 0; i < wx->num_tx_queues; i++) in wx_cache_ring_vmdq() 2063 for (i = 0; i < wx->num_tx_queues; i++) in wx_cache_ring_rss() 2229 unsigned int txr_remaining = wx->num_tx_queues; in wx_alloc_q_vectors() [all …]
|
| /linux/drivers/net/ethernet/intel/ixgbevf/ |
| H A D | ixgbevf_main.c | 1760 for (i = 0; i < adapter->num_tx_queues; i++) in ixgbevf_configure_tx() 2186 unsigned int num_tx_queues = adapter->num_tx_queues; in ixgbevf_configure_dcb() local 2201 num_tx_queues = 1; in ixgbevf_configure_dcb() 2212 (adapter->num_tx_queues != num_tx_queues)) { in ixgbevf_configure_dcb() 2513 for (i = 0; i < adapter->num_tx_queues; i++) in ixgbevf_clean_all_tx_rings() 2548 for (i = 0; i < adapter->num_tx_queues; i++) { in ixgbevf_down() 2657 adapter->num_tx_queues = 1; in ixgbevf_set_num_queues() 2689 adapter->num_tx_queues = rss; in ixgbevf_set_num_queues() 2715 v_budget = max(adapter->num_rx_queues, adapter->num_tx_queues); in ixgbevf_set_interrupt_capability() 2914 int txr_remaining = adapter->num_tx_queues; in ixgbevf_alloc_q_vectors() [all …]
|
| /linux/drivers/net/ethernet/freescale/ |
| H A D | gianfar.c | 133 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_init_tx_rx_base() 247 for_each_set_bit(i, &tx_mask, priv->num_tx_queues) { in gfar_configure_coalescing() 289 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_get_stats64() 415 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_alloc_tx_queues() 447 for (i = 0; i < priv->num_tx_queues; i++) in gfar_free_tx_queues() 558 for_each_set_bit(i, &grp->tx_bit_map, priv->num_tx_queues) { in gfar_parse_group() 561 grp->num_tx_queues++; in gfar_parse_group() 682 priv->num_tx_queues = num_tx_qs; in gfar_of_init() 1121 for (i = 0; i < priv->num_tx_queues; i++) { in free_skb_resources() 1282 for (i = 0; i < priv->num_tx_queues; i++) { in gfar_init_bds() [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | ipoib_tx.c | 697 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init() 704 for (i = 0; i < dev->num_tx_queues; i++) { in hfi1_ipoib_txreq_init() 798 for (i = 0; i < priv->netdev->num_tx_queues; i++) { in hfi1_ipoib_txreq_deinit() 821 for (i = 0; i < dev->num_tx_queues; i++) { in hfi1_ipoib_napi_tx_enable() 833 for (i = 0; i < dev->num_tx_queues; i++) { in hfi1_ipoib_napi_tx_disable()
|
| /linux/net/core/ |
| H A D | dev.h | 239 netdev->num_tx_queues); in netdev_set_defer_hard_irqs() 285 netdev->num_tx_queues); in netdev_set_gro_flush_timeout()
|
| /linux/include/net/ |
| H A D | netdev_lock.h | 127 for (i = 0; i < (dev)->num_tx_queues; i++) \
|