| /linux/drivers/net/xen-netback/ |
| H A D | interface.c | 183 unsigned int num_queues; in xenvif_select_queue() local 187 num_queues = READ_ONCE(vif->num_queues); in xenvif_select_queue() 188 if (num_queues < 1) in xenvif_select_queue() 209 unsigned int num_queues; in xenvif_start_xmit() local 219 num_queues = READ_ONCE(vif->num_queues); in xenvif_start_xmit() 220 if (num_queues < 1) in xenvif_start_xmit() 225 if (index >= num_queues) { in xenvif_start_xmit() 228 index %= num_queues; in xenvif_start_xmit() 275 unsigned int num_queues; in xenvif_get_stats() local 283 num_queues = READ_ONCE(vif->num_queues); in xenvif_get_stats() [all …]
|
| H A D | xenbus.c | 177 for (i = 0; i < vif->num_queues; ++i) { in xenvif_debugfs_addif() 255 unsigned int num_queues = vif->num_queues; in backend_disconnect() local 267 vif->num_queues = 0; in backend_disconnect() 270 for (queue_index = 0; queue_index < num_queues; ++queue_index) in backend_disconnect() 533 for (queue_index = 0; queue_index < vif->num_queues; queue_index++) { in xen_net_rate_changed() 773 be->vif->num_queues = requested_num_queues; in connect() 791 be->vif->num_queues = queue_index; in connect() 807 be->vif->num_queues = queue_index; in connect() 838 if (be->vif->num_queues > 0) in connect() 840 for (queue_index = 0; queue_index < be->vif->num_queues; ++queue_index) in connect() [all …]
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf_main.c | 38 if (cptvf->num_queues) { in init_worker_threads() 40 cptvf->num_queues); in init_worker_threads() 43 for (i = 0; i < cptvf->num_queues; i++) { in init_worker_threads() 63 if (cptvf->num_queues) { in cleanup_worker_threads() 65 cptvf->num_queues); in cleanup_worker_threads() 68 for (i = 0; i < cptvf->num_queues; i++) in cleanup_worker_threads() 90 pqinfo->num_queues = 0; in free_pending_queues() 94 u32 num_queues) in alloc_pending_queues() argument 100 pqinfo->num_queues = num_queues; in alloc_pending_queues() 126 u32 num_queues) in init_pending_queues() argument [all …]
|
| H A D | otx_cptvf.h | 52 u32 num_queues; /* Number of queues supported */ member 57 for (i = 0, q = &qinfo->queue[i]; i < qinfo->num_queues; i++, \ 87 u32 num_queues; member
|
| /linux/arch/mips/cavium-octeon/executive/ |
| H A D | cvmx-pko.c | 71 const int num_queues = 1; in __cvmx_pko_iport_config() local 76 for (queue = 0; queue < num_queues; queue++) { in __cvmx_pko_iport_config() 85 config.s.tail = (queue == (num_queues - 1)); in __cvmx_pko_iport_config() 101 num_queues, queue); in __cvmx_pko_iport_config() 327 uint64_t num_queues, in cvmx_pko_config_port() argument 347 if (base_queue + num_queues > CVMX_PKO_MAX_OUTPUT_QUEUES) { in cvmx_pko_config_port() 350 (unsigned long long)(base_queue + num_queues)); in cvmx_pko_config_port() 360 for (queue = 0; queue < num_queues; queue++) { in cvmx_pko_config_port() 374 && queue == num_queues - 1) in cvmx_pko_config_port() 417 cvmx_dprintf("num queues: %d (%lld,%lld)\n", num_queues, in cvmx_pko_config_port() [all …]
|
| /linux/tools/testing/selftests/drivers/net/hw/ |
| H A D | ncdevmem.c | 95 static int num_queues = -1; variable 820 queues = netdev_queue_id_alloc(num_queues); in create_queues() 821 for (i = 0; i < num_queues; i++) { in create_queues() 877 if (bind_rx_queue(ifindex, mem->fd, create_queues(), num_queues, &ys)) { in do_server() 1096 queues = netdev_queue_id_alloc(num_queues); in run_devmem_tests() 1102 if (!bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) { in run_devmem_tests() 1118 if (!bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) { in run_devmem_tests() 1134 if (bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) { in run_devmem_tests() 1140 if (!check_changing_channels(num_queues, num_queues)) { in run_devmem_tests() 1428 num_queues = atoi(optarg); in main() [all …]
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_main.c | 118 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_stats() 364 if (block->tx->q_num < priv->tx_cfg.num_queues) in gve_napi_poll() 419 if (block->tx->q_num < priv->tx_cfg.num_queues) in gve_napi_poll_dqo() 517 if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues) in gve_alloc_notify_blocks() 518 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues; in gve_alloc_notify_blocks() 519 if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues) in gve_alloc_notify_blocks() 520 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; in gve_alloc_notify_blocks() 690 err = gve_init_rss_config(priv, priv->rx_cfg.num_queues); in gve_setup_device_resources() 904 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues); in gve_create_rings() 907 priv->rx_cfg.num_queues); in gve_create_rings() [all …]
|
| H A D | gve_ethtool.c | 105 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_get_strings() 139 (priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS) + in gve_get_sset_count() 179 rx_qid_to_stats_idx = kmalloc_objs(int, priv->rx_cfg.num_queues); in gve_get_ethtool_stats() 182 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 203 ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 279 nic_stats_len = (NIC_RX_STATS_REPORT_NUM * priv->rx_cfg.num_queues + in gve_get_ethtool_stats() 297 (priv->rx_cfg.num_queues - num_stopped_rxqs) + in gve_get_ethtool_stats() 315 if (queue_id < 0 || queue_id >= priv->rx_cfg.num_queues) { in gve_get_ethtool_stats() 323 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 382 i += priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS; in gve_get_ethtool_stats() [all …]
|
| H A D | gve.h | 681 u16 num_queues; member 688 u16 num_queues; /* number of TX queues, excluding XDP queues */ member 1115 return tx_cfg->num_queues + tx_cfg->num_xdp_queues; in gve_num_tx_qpls() 1124 return rx_cfg->num_queues; in gve_num_rx_qpls() 1185 return priv->tx_cfg.num_queues + priv->tx_cfg.num_xdp_queues; in gve_num_tx_queues() 1190 return priv->tx_cfg.num_queues + queue_id; in gve_xdp_tx_queue_id() 1325 int gve_init_rss_config(struct gve_priv *priv, u16 num_queues);
|
| /linux/drivers/net/wireless/silabs/wfx/ |
| H A D | queue.c | 234 int i, j, num_queues = 0; in wfx_tx_queues_get_skb() local 243 WARN_ON(num_queues >= ARRAY_SIZE(queues)); in wfx_tx_queues_get_skb() 244 queues[num_queues] = &wvif->tx_queue[i]; in wfx_tx_queues_get_skb() 245 for (j = num_queues; j > 0; j--) in wfx_tx_queues_get_skb() 249 num_queues++; in wfx_tx_queues_get_skb() 255 for (i = 0; i < num_queues; i++) { in wfx_tx_queues_get_skb() 277 for (i = 0; i < num_queues; i++) { in wfx_tx_queues_get_skb() 296 for (i = 0; i < num_queues; i++) { in wfx_tx_queues_get_skb()
|
| /linux/drivers/net/netdevsim/ |
| H A D | bus.c | 156 nsim_bus_dev_new(unsigned int id, unsigned int port_count, unsigned int num_queues); 161 unsigned int id, port_count, num_queues; in new_device_store() local 165 err = sscanf(buf, "%u %u %u", &id, &port_count, &num_queues); in new_device_store() 171 num_queues = 1; in new_device_store() 191 nsim_bus_dev = nsim_bus_dev_new(id, port_count, num_queues); in new_device_store() 449 nsim_bus_dev_new(unsigned int id, unsigned int port_count, unsigned int num_queues) in nsim_bus_dev_new() argument 465 nsim_bus_dev->num_queues = num_queues; in nsim_bus_dev_new()
|
| /linux/drivers/net/ |
| H A D | xen-netfront.c | 364 unsigned int num_queues = dev->real_num_tx_queues; in xennet_open() local 371 for (i = 0; i < num_queues; ++i) { in xennet_open() 589 unsigned int num_queues = dev->real_num_tx_queues; in xennet_select_queue() local 594 if (num_queues == 1) { in xennet_select_queue() 598 queue_idx = hash % num_queues; in xennet_select_queue() 647 unsigned int num_queues = dev->real_num_tx_queues; in xennet_xdp_xmit() local 659 queue = &np->queues[smp_processor_id() % num_queues]; in xennet_xdp_xmit() 720 unsigned int num_queues = dev->real_num_tx_queues; in xennet_start_xmit() local 725 if (num_queues < 1) in xennet_start_xmit() 866 unsigned int num_queues = np->queues ? dev->real_num_tx_queues : 0; in xennet_close() local [all …]
|
| /linux/drivers/net/ethernet/freescale/dpaa2/ |
| H A D | dpaa2-eth-debugfs.c | 125 int i, j, num_queues, buf_cnt; in dpaa2_dbg_bp_show() local 133 num_queues = dpaa2_eth_queue_count(priv); in dpaa2_dbg_bp_show() 134 for (i = 0; i < num_queues; i++) { in dpaa2_dbg_bp_show() 153 for (j = 0; j < num_queues; j++) { in dpaa2_dbg_bp_show()
|
| /linux/net/core/ |
| H A D | ieee8021q_helpers.c | 99 int ieee8021q_tt_to_tc(enum ieee8021q_traffic_type tt, unsigned int num_queues) in ieee8021q_tt_to_tc() argument 107 switch (num_queues) { in ieee8021q_tt_to_tc() 134 pr_err("Invalid number of queues %d\n", num_queues); in ieee8021q_tt_to_tc()
|
| /linux/drivers/scsi/ |
| H A D | virtio_scsi.c | 83 u32 num_queues; member 214 num_vqs = vscsi->num_queues; in virtscsi_poll_requests() 854 num_req_vqs = vscsi->num_queues; in virtscsi_init() 918 u32 num_queues; in virtscsi_probe() local 927 num_queues = virtscsi_config_get(vdev, num_queues) ? : 1; in virtscsi_probe() 928 num_queues = min_t(unsigned int, nr_cpu_ids, num_queues); in virtscsi_probe() 929 num_queues = blk_mq_num_possible_queues(num_queues); in virtscsi_probe() 934 struct_size(vscsi, req_vqs, num_queues)); in virtscsi_probe() 943 vscsi->num_queues = num_queues; in virtscsi_probe() 966 shost->nr_hw_queues = num_queues; in virtscsi_probe()
|
| /linux/include/net/ |
| H A D | ieee8021q.h | 41 int ieee8021q_tt_to_tc(enum ieee8021q_traffic_type tt, unsigned int num_queues); 51 unsigned int num_queues) in ieee8021q_tt_to_tc() argument
|
| /linux/drivers/soc/ti/ |
| H A D | knav_qmss_acc.c | 33 for (queue = 0; queue < range->num_queues; queue++) { in __knav_acc_notify() 150 queue >= range_base + range->num_queues) { in knav_acc_int_handler() 154 range_base + range->num_queues); in knav_acc_int_handler() 313 queue_mask = BIT(range->num_queues) - 1; in knav_acc_setup_cmd() 379 for (queue = 0; queue < range->num_queues; queue++) { in knav_acc_init_range() 439 channels = range->num_queues; in knav_acc_free_range() 523 channels = range->num_queues; in knav_init_acc_range() 533 if (range->num_queues > 32) { in knav_init_acc_range()
|
| H A D | knav_qmss.h | 140 unsigned num_queues; member 297 unsigned num_queues; member 332 unsigned num_queues; member
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede.h | 184 #define QEDE_MAX_RSS_CNT(edev) ((edev)->dev_info.num_queues) 185 #define QEDE_MAX_TSS_CNT(edev) ((edev)->dev_info.num_queues) 198 u16 num_queues; member 201 #define QEDE_QUEUE_CNT(edev) ((edev)->num_queues) 202 #define QEDE_RSS_COUNT(edev) ((edev)->num_queues - (edev)->fp_num_tx) 204 #define QEDE_TSS_COUNT(edev) ((edev)->num_queues - (edev)->fp_num_rx) 605 #define for_each_queue(i) for (i = 0; i < edev->num_queues; i++)
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | cn23xx_vf_device.c | 50 static int cn23xx_vf_reset_io_queues(struct octeon_device *oct, u32 num_queues) in cn23xx_vf_reset_io_queues() argument 57 for (q_no = 0; q_no < num_queues; q_no++) { in cn23xx_vf_reset_io_queues() 67 for (q_no = 0; q_no < num_queues; q_no++) { in cn23xx_vf_reset_io_queues() 359 u32 num_queues = oct->num_iqs; in cn23xx_disable_vf_io_queues() local 364 if (num_queues < oct->num_oqs) in cn23xx_disable_vf_io_queues() 365 num_queues = oct->num_oqs; in cn23xx_disable_vf_io_queues() 367 cn23xx_vf_reset_io_queues(oct, num_queues); in cn23xx_disable_vf_io_queues()
|
| /linux/drivers/gpu/drm/nouveau/include/nvfw/ |
| H A D | sec2.h | 21 u8 num_queues; member 42 u8 num_queues; member
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_pm4_headers.h | 89 uint32_t num_queues:10; member 142 uint32_t num_queues:10; member
|
| H A D | kfd_pm4_headers_vi.h | 192 uint32_t num_queues:10; member 241 uint32_t num_queues:3; member 362 uint32_t num_queues:3; member
|
| /linux/drivers/net/ethernet/cadence/ |
| H A D | macb_main.c | 469 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { in macb_init_buffers() 660 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) in macb_mac_link_down() 708 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { in macb_mac_link_up() 1762 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { in macb_hresp_error_task() 1780 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) in macb_hresp_error_task() 1984 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) in macb_poll_controller() 2394 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { in gem_free_rx_buffers() 2455 size = bp->num_queues * macb_tx_ring_size_per_queue(bp); in macb_free_consistent() 2458 size = bp->num_queues * macb_rx_ring_size_per_queue(bp); in macb_free_consistent() 2461 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { in macb_free_consistent() [all …]
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | mana_en.c | 313 err = mana_pre_alloc_rxbufs(apc, ndev->mtu, apc->num_queues); in mana_per_port_queue_reset_work_handler() 586 unsigned int num_queues = apc->num_queues; in mana_get_stats64() local 603 for (q = 0; q < num_queues; q++) { in mana_get_stats64() 616 for (q = 0; q < num_queues; q++) { in mana_get_stats64() 784 int mana_pre_alloc_rxbufs(struct mana_port_context *mpc, int new_mtu, int num_queues) in mana_pre_alloc_rxbufs() argument 799 num_rxb = num_queues * mpc->rx_queue_size; in mana_pre_alloc_rxbufs() 834 netdev_err(mpc->ndev, "Failed to pre-allocate RX buffers for %d queues\n", num_queues); in mana_pre_alloc_rxbufs() 846 err = mana_pre_alloc_rxbufs(mpc, new_mtu, mpc->num_queues); in mana_change_mtu() 998 apc->rxqs = kzalloc_objs(struct mana_rxq *, apc->num_queues); in mana_init_port_context() 1709 for (rxq_idx = 0; rxq_idx < apc->num_queues; rxq_idx++) { in mana_fence_rqs() [all …]
|