| /linux/net/core/ |
| H A D | netdev_rx_queue.c | 180 struct netdev_queue_config qcfg; in netdev_rx_queue_restart() local 182 netdev_queue_config(dev, rxq_idx, &qcfg); in netdev_rx_queue_restart() 183 return netdev_rx_queue_reconfig(dev, rxq_idx, &qcfg, &qcfg); in netdev_rx_queue_restart() 192 struct netdev_queue_config qcfg[2]; in __netif_mp_open_rxq() local 228 netdev_queue_config(dev, rxq_idx, &qcfg[0]); in __netif_mp_open_rxq() 230 ret = netdev_queue_config_validate(dev, rxq_idx, &qcfg[1], extack); in __netif_mp_open_rxq() 234 ret = netdev_rx_queue_reconfig(dev, rxq_idx, &qcfg[0], &qcfg[1]); in __netif_mp_open_rxq() 281 struct netdev_queue_config qcfg[2]; in __netif_mp_close_rxq() local 298 netdev_queue_config(dev, rxq_idx, &qcfg[0]); in __netif_mp_close_rxq() 300 netdev_queue_config(dev, rxq_idx, &qcfg[1]); in __netif_mp_close_rxq() [all …]
|
| H A D | dev.h | 106 struct netdev_queue_config *qcfg,
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_tx_dqo.c | 361 if (idx >= cfg->qcfg->num_queues && cfg->num_xdp_rings) { in gve_tx_alloc_ring_dqo() 419 total_queues = cfg->qcfg->num_queues + cfg->num_xdp_rings; in gve_tx_alloc_rings_dqo() 420 if (total_queues > cfg->qcfg->max_queues) { in gve_tx_alloc_rings_dqo() 426 tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues); in gve_tx_alloc_rings_dqo() 459 for (i = 0; i < cfg->qcfg->num_queues + cfg->qcfg->num_xdp_queues; i++) in gve_tx_free_rings_dqo()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_main.c | 2828 struct netdev_queue_config *qcfg, in mlx5e_open_channel() argument 2862 err = mlx5e_build_channel_param(mdev, params, qcfg, cparam); in mlx5e_open_channel() 5619 struct netdev_queue_config *qcfg) in mlx5e_queue_default_qcfg() argument 5621 qcfg->rx_page_size = PAGE_SIZE; in mlx5e_queue_default_qcfg() 5625 struct netdev_queue_config *qcfg, in mlx5e_queue_validate_qcfg() argument 5632 if (!is_power_of_2(qcfg->rx_page_size)) { in mlx5e_queue_validate_qcfg() 5634 qcfg->rx_page_size); in mlx5e_queue_validate_qcfg() 5639 if (qcfg->rx_page_size < PAGE_SIZE || qcfg->rx_page_size > max) { in mlx5e_queue_validate_qcfg() 5642 qcfg->rx_page_size, PAGE_SIZE, max); in mlx5e_queue_validate_qcfg() 5650 struct netdev_queue_config *qcfg, in mlx5e_queue_validate_page_size() argument [all …]
|
| /linux/drivers/net/netdevsim/ |
| H A D | netdev.c | 750 struct netdev_queue_config *qcfg, in nsim_queue_mem_alloc() argument 800 nsim_queue_start(struct net_device *dev, struct netdev_queue_config *qcfg, in nsim_queue_start() argument
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_tx_rx.c | 1264 #define BNA_GET_RXQS(qcfg) (((qcfg)->rxp_type == BNA_RXP_SINGLE) ? \ argument 1265 (qcfg)->num_paths : ((qcfg)->num_paths * 2))
|
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt.c | 4418 struct netdev_queue_config qcfg; in bnxt_init_ring_struct() local 4441 netdev_queue_config(bp->dev, i, &qcfg); in bnxt_init_ring_struct() 4442 rxr->rx_page_size = qcfg.rx_page_size; in bnxt_init_ring_struct() 16108 struct netdev_queue_config *qcfg) in bnxt_queue_default_qcfg() argument 16110 qcfg->rx_page_size = BNXT_RX_PAGE_SIZE; in bnxt_queue_default_qcfg() 16114 struct netdev_queue_config *qcfg, in bnxt_validate_qcfg() argument 16121 qcfg->rx_page_size != BNXT_RX_PAGE_SIZE) in bnxt_validate_qcfg() 16124 if (!is_power_of_2(qcfg->rx_page_size)) in bnxt_validate_qcfg() 16127 if (qcfg->rx_page_size < BNXT_RX_PAGE_SIZE || in bnxt_validate_qcfg() 16128 qcfg->rx_page_size > BNXT_MAX_RX_PAGE_SIZE) in bnxt_validate_qcfg() [all …]
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| H A D | fbnic_txrx.c | 2832 struct netdev_queue_config *qcfg, in fbnic_queue_mem_alloc() argument 2887 struct netdev_queue_config *qcfg, in fbnic_queue_start() argument
|