/linux/net/sched/ |
H A D | sch_mqprio_lib.c | 38 if (qopt->offset[i] >= dev->real_num_tx_queues || in mqprio_validate_queue_counts() 39 last > dev->real_num_tx_queues) { in mqprio_validate_queue_counts() 43 i, dev->real_num_tx_queues); in mqprio_validate_queue_counts()
|
H A D | sch_mq.c | 119 if (ntx < dev->real_num_tx_queues) in mq_attach()
|
H A D | sch_multiq.c | 186 qopt->bands = qdisc_dev(sch)->real_num_tx_queues; in multiq_tune()
|
H A D | act_skbedit.c | 76 skb->dev->real_num_tx_queues > params->queue_mapping) { in tcf_skbedit_act()
|
H A D | sch_generic.c | 1431 for (i = new_real_tx; i < dev->real_num_tx_queues; i++) { in mq_change_real_num_tx() 1439 for (i = dev->real_num_tx_queues; i < new_real_tx; i++) { in mq_change_real_num_tx()
|
/linux/drivers/net/ethernet/cavium/liquidio/ |
H A D | octeon_network.h | 563 for (i = 0; i < netdev->real_num_tx_queues; i++) in stop_txqs() 576 for (i = 0; i < netdev->real_num_tx_queues; i++) { in wake_txqs() 597 for (i = 0; i < netdev->real_num_tx_queues; i++) in start_txqs()
|
/linux/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_ethtool.c | 518 for (i = 0; i < netdev->real_num_tx_queues; i++) in fun_set_coalesce() 531 chan->tx_count = netdev->real_num_tx_queues; in fun_get_channels() 540 if (chan->tx_count == netdev->real_num_tx_queues && in fun_set_channels() 622 n = (dev->real_num_tx_queues + 1) * ARRAY_SIZE(txq_stat_names) + in fun_get_sset_count() 652 for (i = 0; i < netdev->real_num_tx_queues; i++) { in fun_get_strings() 770 totals = data + netdev->real_num_tx_queues * ARRAY_SIZE(txq_stat_names); in fun_get_ethtool_stats() 772 for (i = 0; i < netdev->real_num_tx_queues; i++) { in fun_get_ethtool_stats()
|
H A D | funeth_main.c | 455 qset->ntxqs = netdev->real_num_tx_queues; in fun_free_rings() 839 .ntxqs = netdev->real_num_tx_queues, in funeth_open() 890 for (i = 0; i < netdev->real_num_tx_queues; i++) { in fun_get_stats64() 1592 newqs->ntxqs = dev->real_num_tx_queues; in fun_replace_queues() 1625 unsigned int keep_tx = min(dev->real_num_tx_queues, ntx); in fun_change_num_queues() 1632 .ntxqs = dev->real_num_tx_queues, in fun_change_num_queues() 1673 if (ntx < dev->real_num_tx_queues) in fun_change_num_queues() 1680 if (ntx > dev->real_num_tx_queues) in fun_change_num_queues()
|
/linux/drivers/net/ |
H A D | xen-netfront.c | 364 unsigned int num_queues = dev->real_num_tx_queues; in xennet_open() 589 unsigned int num_queues = dev->real_num_tx_queues; in xennet_select_queue() 649 unsigned int num_queues = dev->real_num_tx_queues; in xennet_xdp_xmit() 722 unsigned int num_queues = dev->real_num_tx_queues; in xennet_start_xmit() 871 unsigned int num_queues = np->queues ? dev->real_num_tx_queues : 0; in xennet_close() 889 for (i = 0; i < info->netdev->real_num_tx_queues; i++) { in xennet_destroy_queues() 1600 unsigned int num_queues = dev->real_num_tx_queues; in xennet_poll_controller() 1661 bpf_prog_add(prog, dev->real_num_tx_queues); in xennet_xdp_set() 1663 for (i = 0; i < dev->real_num_tx_queues; ++i) in xennet_xdp_set() 1667 for (i = 0; i < dev->real_num_tx_queues; ++i) in xennet_xdp_set() [all …]
|
H A D | veth.c | 158 for (i = 0; i < dev->real_num_tx_queues; i++) in veth_get_strings() 174 VETH_TQ_STATS_LEN * dev->real_num_tx_queues + in veth_get_sset_count() 233 tx_idx += (i % dev->real_num_tx_queues) * VETH_TQ_STATS_LEN; in veth_get_ethtool_stats() 242 pp_idx = idx + dev->real_num_tx_queues * VETH_TQ_STATS_LEN; in veth_get_ethtool_stats() 251 channels->tx_count = dev->real_num_tx_queues; in veth_get_channels() 1281 if (peer && peer->real_num_tx_queues <= dev->real_num_rx_queues) { in veth_set_xdp_features() 1312 if (priv->_xdp_prog && peer && ch->rx_count < peer->real_num_tx_queues) in veth_set_channels() 1593 if (dev->real_num_rx_queues < peer->real_num_tx_queues) { in veth_xdp_set()
|
H A D | net_failover.c | 140 if (unlikely(txq >= dev->real_num_tx_queues)) { in net_failover_select_queue() 142 txq -= dev->real_num_tx_queues; in net_failover_select_queue() 143 } while (txq >= dev->real_num_tx_queues); in net_failover_select_queue()
|
/linux/drivers/s390/net/ |
H A D | qeth_ethtool.c | 209 channels->tx_count = dev->real_num_tx_queues; in qeth_get_channels() 238 channels->tx_count < dev->real_num_tx_queues) in qeth_set_channels()
|
/linux/net/core/ |
H A D | netdev-genl.c | 440 if (q_id >= netdev->real_num_tx_queues) in netdev_nl_queue_validate() 517 for (; ctx->txq_idx < netdev->real_num_tx_queues; ctx->txq_idx++) { in netdev_nl_queue_dump_one() 692 while (ops->get_queue_stats_tx && i < netdev->real_num_tx_queues) { in netdev_nl_stats_by_queue() 778 0, netdev->real_num_tx_queues, &tx_sum); in netdev_nl_stats_by_netdev()
|
H A D | netpoll.c | 115 if (unlikely(q_index >= dev->real_num_tx_queues)) { in queue_process() 116 q_index = q_index % dev->real_num_tx_queues; in queue_process()
|
H A D | net-sysfs.c | 2091 real_tx = dev->real_num_tx_queues; in register_queue_kobjects() 2126 real_tx = ndev->real_num_tx_queues; in queue_change_owner() 2146 real_tx = dev->real_num_tx_queues; in remove_queue_kobjects() 2153 dev->real_num_tx_queues = 0; in remove_queue_kobjects()
|
H A D | dev.c | 3129 if ((offset + count) > dev->real_num_tx_queues) in netdev_bind_sb_channel_queue() 3175 disabling = txq < dev->real_num_tx_queues; in netif_set_real_num_tx_queues() 3185 rc = netdev_queue_update_kobjects(dev, dev->real_num_tx_queues, in netif_set_real_num_tx_queues() 3197 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues() 3207 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues() 3273 if (txq > dev->real_num_tx_queues) { in netif_set_real_num_queues() 3280 if (txq < dev->real_num_tx_queues) in netif_set_real_num_queues() 3491 u16 qcount = dev->real_num_tx_queues; in skb_tx_hash() 3502 qcount = dev->real_num_tx_queues; in skb_tx_hash() 4488 if (unlikely(queue_index >= dev->real_num_tx_queues)) in __get_xps_queue_idx() [all …]
|
/linux/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_bpf.c | 62 q_idx = smp_processor_id() % ndev->real_num_tx_queues; in mana_xdp_xmit()
|
/linux/drivers/net/hyperv/ |
H A D | netvsc_bpf.c | 276 q_idx = smp_processor_id() % ndev->real_num_tx_queues; in netvsc_ndoxdp_xmit()
|
H A D | netvsc_drv.c | 283 if (q_idx < 0 || skb->ooo_okay || q_idx >= ndev->real_num_tx_queues) { in netvsc_pick_tx() 323 while (txq >= ndev->real_num_tx_queues) in netvsc_select_queue() 324 txq -= ndev->real_num_tx_queues; in netvsc_select_queue() 520 if (hash != 0 && net->real_num_tx_queues > 1) { in netvsc_xmit()
|
/linux/net/xdp/ |
H A D | xsk.c | 104 if (queue_id < dev->real_num_tx_queues) in xsk_get_pool_from_qid() 128 dev->real_num_tx_queues)) in xsk_reg_pool_at_qid() 133 if (queue_id < dev->real_num_tx_queues) in xsk_reg_pool_at_qid() 796 if (xs->queue_id >= xs->dev->real_num_tx_queues) in __xsk_generic_xmit()
|
/linux/drivers/net/xen-netback/ |
H A D | interface.c | 193 dev->real_num_tx_queues; in xenvif_select_queue() 198 return skb_get_hash_raw(skb) % dev->real_num_tx_queues; in xenvif_select_queue()
|
/linux/drivers/net/ethernet/myricom/myri10ge/ |
H A D | myri10ge.c | 977 if (mgp->dev->real_num_tx_queues > 1) in myri10ge_reset() 985 if (status != 0 && mgp->dev->real_num_tx_queues > 1) { in myri10ge_reset() 1426 if ((ss->mgp->dev->real_num_tx_queues > 1) && in myri10ge_tx_done() 1551 if ((mgp->dev->real_num_tx_queues == 1) && (ss != mgp->ss)) { in myri10ge_intr() 2239 if (slice == 0 || (mgp->dev->real_num_tx_queues > 1)) { in myri10ge_get_txrx() 2315 if (mgp->dev->real_num_tx_queues > 1) in myri10ge_open() 2404 if (slice == 0 || mgp->dev->real_num_tx_queues > 1) in myri10ge_open() 2856 if ((mgp->dev->real_num_tx_queues > 1) && tx->queue_active == 0) { in myri10ge_xmit()
|
/linux/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_stats.c | 171 #define MAX_Q(lif) ((lif)->netdev->real_num_tx_queues)
|
/linux/drivers/net/team/ |
H A D | team_core.c | 1554 if (new_queue_id >= team->dev->real_num_tx_queues) in team_queue_id_option_set() 1767 if (unlikely(txq >= dev->real_num_tx_queues)) { in team_select_queue() 1769 txq -= dev->real_num_tx_queues; in team_select_queue() 1770 } while (txq >= dev->real_num_tx_queues); in team_select_queue()
|
/linux/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_xsk.c | 97 qid >= netdev->real_num_tx_queues) in i40e_xsk_pool_enable()
|