| /linux/block/ |
| H A D | blk-mq-cpumap.c | 21 unsigned int max_queues) in blk_mq_num_queues() argument 26 return min_not_zero(num, max_queues); in blk_mq_num_queues() 38 unsigned int blk_mq_num_possible_queues(unsigned int max_queues) in blk_mq_num_possible_queues() argument 40 return blk_mq_num_queues(cpu_possible_mask, max_queues); in blk_mq_num_possible_queues() 53 unsigned int blk_mq_num_online_queues(unsigned int max_queues) in blk_mq_num_online_queues() argument 55 return blk_mq_num_queues(cpu_online_mask, max_queues); in blk_mq_num_online_queues()
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | hw_queue.c | 53 if (!hdev->asic_prop.max_queues || q->queue_type == QUEUE_TYPE_HW) in hl_hw_queue_update_ci() 61 for (i = 0 ; i < hdev->asic_prop.max_queues ; i++, q++) { in hl_hw_queue_update_ci() 646 u32 max_queues; in hl_hw_queue_schedule_cs() local 661 max_queues = hdev->asic_prop.max_queues; in hl_hw_queue_schedule_cs() 664 for (i = 0, cq_cnt = 0 ; i < max_queues ; i++, q++) { in hl_hw_queue_schedule_cs() 795 for (i = 0 ; (i < max_queues) && (cq_cnt > 0) ; i++, q++) { in hl_hw_queue_schedule_cs() 1086 asic->max_queues); in hl_hw_queues_create() 1095 i < asic->max_queues ; i++, q_ready_cnt++, q++) { in hl_hw_queues_create() 1129 u32 max_queues = hdev->asic_prop.max_queues; in hl_hw_queues_destroy() local 1132 for (i = 0, q = hdev->kernel_queues ; i < max_queues ; i++, q++) in hl_hw_queues_destroy() [all …]
|
| H A D | command_submission.c | 950 cs->jobs_in_queue_cnt = kcalloc(hdev->asic_prop.max_queues, in allocate_cs() 953 cs->jobs_in_queue_cnt = kcalloc(hdev->asic_prop.max_queues, in allocate_cs() 1214 if (chunk->queue_index >= asic->max_queues) { in validate_queue_index() 2022 if (q_idx >= hdev->asic_prop.max_queues) { in cs_ioctl_reserve_signals() 2230 if (chunk->queue_index >= hdev->asic_prop.max_queues) { in cs_ioctl_signal_wait()
|
| /linux/arch/mips/cavium-octeon/executive/ |
| H A D | cvmx-pko.c | 215 int max_queues = in cvmx_pko_initialize_global() local 219 if (max_queues <= 32) in cvmx_pko_initialize_global() 221 else if (max_queues <= 64) in cvmx_pko_initialize_global() 224 if (max_queues <= 64) in cvmx_pko_initialize_global() 226 else if (max_queues <= 128) in cvmx_pko_initialize_global()
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_main.c | 286 priv->tx_cfg.max_queues; in gve_alloc_stats_report() 288 priv->rx_cfg.max_queues; in gve_alloc_stats_report() 509 priv->tx_cfg.max_queues = min_t(int, priv->tx_cfg.max_queues, in gve_alloc_notify_blocks() 511 priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues, in gve_alloc_notify_blocks() 515 vecs_enabled, priv->tx_cfg.max_queues, in gve_alloc_notify_blocks() 516 priv->rx_cfg.max_queues); in gve_alloc_notify_blocks() 517 if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues) in gve_alloc_notify_blocks() 518 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues; in gve_alloc_notify_blocks() 519 if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues) in gve_alloc_notify_blocks() 520 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; in gve_alloc_notify_blocks() [all …]
|
| H A D | gve.h | 680 u16 max_queues; member 687 u16 max_queues; member 1134 return priv->tx_cfg.max_queues + rx_qid; in gve_rx_qpl_id() 1140 return tx_cfg->max_queues + rx_qid; in gve_get_rx_qpl_id()
|
| H A D | gve_ethtool.c | 490 cmd->max_rx = priv->rx_cfg.max_queues; in gve_get_channels() 491 cmd->max_tx = priv->tx_cfg.max_queues; in gve_get_channels() 522 (2 * new_tx > priv->tx_cfg.max_queues)) { in gve_set_channels()
|
| H A D | gve_tx.c | 342 if (total_queues > cfg->qcfg->max_queues) { in gve_tx_alloc_rings_gqi() 348 tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues); in gve_tx_alloc_rings_gqi()
|
| H A D | gve_tx_dqo.c | 422 if (total_queues > cfg->qcfg->max_queues) { in gve_tx_alloc_rings_dqo() 428 tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues); in gve_tx_alloc_rings_dqo()
|
| /linux/drivers/net/ethernet/samsung/sxgbe/ |
| H A D | sxgbe_common.h | 194 #define SXGBE_FOR_EACH_QUEUE(max_queues, queue_num) \ argument 195 for (queue_num = 0; queue_num < max_queues; queue_num++)
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_process_queue_manager.c | 324 unsigned int max_queues = 127; /* HWS limit */ in pqm_create_queue() local 333 max_queues = 255; in pqm_create_queue() 351 max_queues = dev->kfd->device_info.max_no_of_hqd/2; in pqm_create_queue() 353 if (pdd->qpd.queue_count >= max_queues) in pqm_create_queue()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_qp.c | 74 u16 num_queues, u16 max_queues) in hinic_qp_prepare_header() argument 76 u16 max_sqs = max_queues; in hinic_qp_prepare_header() 77 u16 max_rqs = max_queues; in hinic_qp_prepare_header()
|
| H A D | hinic_hw_qp.h | 130 u16 num_queues, u16 max_queues);
|
| /linux/drivers/net/ethernet/amazon/ena/ |
| H A D | ena_netdev.c | 2925 struct ena_admin_queue_feature_desc *max_queues = in ena_calc_io_queue_size() local 2926 &get_feat_ctx->max_queues; in ena_calc_io_queue_size() 2927 max_rx_queue_size = min_t(u32, max_queues->max_cq_depth, in ena_calc_io_queue_size() 2928 max_queues->max_sq_depth); in ena_calc_io_queue_size() 2929 max_tx_queue_size = max_queues->max_cq_depth; in ena_calc_io_queue_size() 2936 max_queues->max_sq_depth); in ena_calc_io_queue_size() 2939 max_queues->max_packet_tx_descs); in ena_calc_io_queue_size() 2941 max_queues->max_packet_rx_descs); in ena_calc_io_queue_size() 3731 struct ena_admin_queue_feature_desc *max_queues = in ena_calc_max_io_queue_num() local 3732 &get_feat_ctx->max_queues; in ena_calc_max_io_queue_num() [all …]
|
| H A D | ena_com.h | 394 struct ena_admin_queue_feature_desc max_queues; member
|
| /linux/drivers/net/ethernet/cisco/enic/ |
| H A D | enic_main.c | 2260 unsigned int max_queues; in enic_adjust_resources() local 2302 max_queues = min(enic->cq_avail, in enic_adjust_resources() 2304 if (wq_avail + rq_avail <= max_queues) { in enic_adjust_resources() 2310 enic->rq_count = min(rq_avail, max_queues / 2); in enic_adjust_resources() 2311 enic->wq_count = max_queues - enic->rq_count; in enic_adjust_resources() 2313 enic->wq_count = min(wq_avail, max_queues / 2); in enic_adjust_resources() 2314 enic->rq_count = max_queues - enic->wq_count; in enic_adjust_resources()
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | mana_en.c | 3083 u32 max_txq, max_rxq, max_queues; in mana_init_port() local 3103 max_queues = min_t(u32, max_txq, max_rxq); in mana_init_port() 3104 if (apc->max_queues > max_queues) in mana_init_port() 3105 apc->max_queues = max_queues; in mana_init_port() 3107 if (apc->num_queues > apc->max_queues) in mana_init_port() 3108 apc->num_queues = apc->max_queues; in mana_init_port() 3328 apc->max_queues = gc->max_num_queues; in mana_probe_port()
|
| /linux/drivers/net/ethernet/cavium/thunder/ |
| H A D | nicvf_main.c | 1850 if ((nic->rx_queues + nic->tx_queues) > nic->max_queues) { in nicvf_xdp_setup() 1853 nic->max_queues); in nicvf_xdp_setup() 2150 nic->max_queues = qcount; in nicvf_probe() 2155 nic->max_queues *= 2; in nicvf_probe() 2221 nic->rx_queues + nic->tx_queues <= nic->max_queues) in nicvf_probe()
|
| H A D | nic.h | 310 u8 max_queues; member
|
| /linux/drivers/net/ |
| H A D | xen-netfront.c | 65 module_param_named(max_queues, xennet_max_queues, uint, 0644); 66 MODULE_PARM_DESC(max_queues, 2264 unsigned int max_queues = 0; in talk_to_netback() local 2276 max_queues = xenbus_read_unsigned(info->xbdev->otherend, in talk_to_netback() 2278 num_queues = min(max_queues, xennet_max_queues); in talk_to_netback()
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| H A D | fm10k_type.h | 546 u16 max_queues; member
|
| H A D | fm10k_main.c | 1496 rss_i = interface->hw.mac.max_queues / pcs; in fm10k_set_qos_queues() 1529 rss_i = min_t(u16, interface->hw.mac.max_queues, f->limit); in fm10k_set_rss_queues()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_ethtool.c | 266 channel->max_rx = pfvf->hw.max_queues; in otx2_get_channels() 267 channel->max_tx = pfvf->hw.max_queues; in otx2_get_channels()
|
| /linux/drivers/net/xen-netback/ |
| H A D | netback.c | 73 module_param_named(max_queues, xenvif_max_queues, uint, 0644); 74 MODULE_PARM_DESC(max_queues,
|
| /linux/drivers/block/xen-blkback/ |
| H A D | blkback.c | 103 module_param_named(max_queues, xenblk_max_queues, uint, 0644); 104 MODULE_PARM_DESC(max_queues,
|