Lines Matching refs:num_queues

117 		for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) {  in gve_get_stats()
287 priv->rx_cfg.num_queues; in gve_alloc_stats_report()
363 if (block->tx->q_num < priv->tx_cfg.num_queues) in gve_napi_poll()
497 if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues) in gve_alloc_notify_blocks()
498 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues; in gve_alloc_notify_blocks()
499 if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues) in gve_alloc_notify_blocks()
500 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; in gve_alloc_notify_blocks()
655 err = gve_init_rss_config(priv, priv->rx_cfg.num_queues); in gve_setup_device_resources()
866 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues); in gve_create_rings()
869 priv->rx_cfg.num_queues); in gve_create_rings()
876 priv->rx_cfg.num_queues); in gve_create_rings()
885 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_create_rings()
888 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_create_rings()
915 for (i = 0; i < priv->tx_cfg.num_queues; i++) in gve_init_sync_stats()
921 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_init_sync_stats()
1005 err = gve_adminq_destroy_rx_queues(priv, priv->rx_cfg.num_queues); in gve_destroy_rings()
1145 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_reg_xdp_info()
1202 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_unreg_xdp_info()
1222 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_drain_page_cache()
1315 gve_rx_start_rings(priv, rx_alloc_cfg->qcfg_rx->num_queues); in gve_queues_start()
1318 err = netif_set_real_num_tx_queues(dev, priv->tx_cfg.num_queues); in gve_queues_start()
1321 err = netif_set_real_num_rx_queues(dev, priv->rx_cfg.num_queues); in gve_queues_start()
1330 err = gve_init_rss_config(priv, priv->rx_cfg.num_queues); in gve_queues_start()
1368 gve_rx_stop_rings(priv, priv->rx_cfg.num_queues); in gve_queues_start()
1417 gve_rx_stop_rings(priv, priv->rx_cfg.num_queues); in gve_queues_stop()
1491 priv->rx_cfg.num_queues : 0; in gve_set_xdp()
1496 err = gve_configure_rings_xdp(priv, priv->rx_cfg.num_queues); in gve_set_xdp()
1523 if (qid >= priv->rx_cfg.num_queues) { in gve_xsk_pool_enable()
1581 if (qid >= priv->rx_cfg.num_queues) in gve_xsk_pool_disable()
1624 if (queue_id >= priv->rx_cfg.num_queues || !priv->xdp_prog) in gve_xsk_wakeup()
1664 if (priv->rx_cfg.num_queues != priv->tx_cfg.num_queues || in verify_xdp_configuration()
1665 (2 * priv->tx_cfg.num_queues > priv->tx_cfg.max_queues)) { in verify_xdp_configuration()
1667 priv->rx_cfg.num_queues, in verify_xdp_configuration()
1668 priv->tx_cfg.num_queues, in verify_xdp_configuration()
1696 int gve_init_rss_config(struct gve_priv *priv, u16 num_queues) in gve_init_rss_config() argument
1707 ethtool_rxfh_indir_default(i, num_queues); in gve_init_rss_config()
1785 err = gve_init_rss_config(priv, new_rx_config.num_queues); in gve_adjust_queues()
1813 if (idx < priv->tx_cfg.num_queues) in gve_turndown()
1819 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_turndown()
1860 if (idx < priv->tx_cfg.num_queues) in gve_turnup()
1879 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_turnup()
1931 if (txqueue > priv->tx_cfg.num_queues) in gve_tx_timeout()
2149 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_handle_report_stats()
2255 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues; in gve_init_priv()
2256 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; in gve_init_priv()
2258 priv->tx_cfg.num_queues = min_t(int, priv->default_num_queues, in gve_init_priv()
2259 priv->tx_cfg.num_queues); in gve_init_priv()
2260 priv->rx_cfg.num_queues = min_t(int, priv->default_num_queues, in gve_init_priv()
2261 priv->rx_cfg.num_queues); in gve_init_priv()
2266 priv->tx_cfg.num_queues, priv->rx_cfg.num_queues); in gve_init_priv()