| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_main.c | 42 if (cptvf->nr_queues) { in init_worker_threads() 44 cptvf->nr_queues); in init_worker_threads() 47 for (i = 0; i < cptvf->nr_queues; i++) { in init_worker_threads() 69 if (cptvf->nr_queues) { in cleanup_worker_threads() 71 cptvf->nr_queues); in cleanup_worker_threads() 74 for (i = 0; i < cptvf->nr_queues; i++) in cleanup_worker_threads() 100 pqinfo->nr_queues = 0; in free_pending_queues() 104 u32 nr_queues) in alloc_pending_queues() argument 110 pqinfo->nr_queues = nr_queues; in alloc_pending_queues() 136 static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) in init_pending_queues() argument [all …]
|
| H A D | cptvf.h | 85 u32 nr_queues; /* Number of queues supported */ member 91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \ 110 u32 nr_queues; member
|
| H A D | cptvf_reqmanager.c | 233 if (unlikely(qno >= cptvf->nr_queues)) { in send_cpt_command() 235 qno, cptvf->nr_queues); in send_cpt_command() 545 if (unlikely(qno > cptvf->nr_queues)) { in vq_post_process()
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_sriov.c | 58 int nr_queues = 0; in vf_mode_to_nr_queues() local 62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues() 65 nr_queues = 8; in vf_mode_to_nr_queues() 68 nr_queues = 4; in vf_mode_to_nr_queues() 71 nr_queues = 2; in vf_mode_to_nr_queues() 74 nr_queues = 1; in vf_mode_to_nr_queues() 78 return nr_queues; in vf_mode_to_nr_queues()
|
| H A D | nitrox_lib.c | 91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues() 102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues() 142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues() 155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues() 161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
|
| H A D | nitrox_dev.h | 160 int nr_queues; member 250 u16 nr_queues; member
|
| H A D | nitrox_mbx.c | 68 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response() 76 vfdev->nr_queues = 0; in pf2vf_send_response()
|
| H A D | nitrox_hal.c | 124 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_config_pkt_input_rings() 240 for (i = 0; i < ndev->nr_queues; i++) in nitrox_config_pkt_solicit_ports() 356 for (ring = 0; ring < ndev->nr_queues; ring++) { in nitrox_config_aqm_rings()
|
| /linux/block/ |
| H A D | blk-mq-cpumap.c | 64 masks = group_cpus_evenly(qmap->nr_queues, &nr_masks); in blk_mq_map_queues() 71 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_queues() 118 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_hw_queues()
|
| /linux/fs/fuse/ |
| H A D | dev_uring.c | 128 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_abort_end_requests() 169 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_request_expired() 196 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_destruct() 230 size_t nr_queues = num_possible_cpus(); in fuse_uring_create() local 238 ring->queues = kzalloc_objs(struct fuse_ring_queue *, nr_queues, in fuse_uring_create() 256 ring->nr_queues = nr_queues; in fuse_uring_create() 408 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_log_ent_state() 439 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_async_stop_queues() 474 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_stop_queues() 897 if (qid >= ring->nr_queues) in fuse_uring_commit_fetch() [all …]
|
| /linux/tools/perf/util/ |
| H A D | auxtrace.c | 214 static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues) in auxtrace_alloc_queue_array() argument 220 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array() 223 queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue)); in auxtrace_alloc_queue_array() 227 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array() 235 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues) in auxtrace_queues__init_nr() argument 237 queues->nr_queues = nr_queues; in auxtrace_queues__init_nr() 238 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init_nr() 252 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local 256 if (!nr_queues) in auxtrace_queues__grow() 257 nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__grow() [all …]
|
| H A D | auxtrace.h | 292 unsigned int nr_queues; member 518 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues);
|
| H A D | cs-etm.c | 309 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__map_trace_id_v0() 367 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__process_trace_id_v0_1() 995 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__free_events() 2590 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__process_timeless_queues() 2628 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues() 2723 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues() 2872 for (i = 0; i < etm->queues.nr_queues; ++i) in dump_queued_data() 3330 for (unsigned int i = 0; i < queues->nr_queues; i++) { in cs_etm__create_decoders() 3442 for (unsigned int j = 0; j < etm->queues.nr_queues; ++j) { in cs_etm__process_auxtrace_info_full()
|
| H A D | s390-cpumsf.c | 203 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) in s390_cpumcf_dumpctr() 827 for (i = 0; i < sf->queues.nr_queues; i++) { in s390_cpumsf_setup_queues() 1016 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
|
| H A D | intel-bts.c | 211 for (i = 0; i < bts->queues.nr_queues; i++) { in intel_bts_setup_queues() 544 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit() 715 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
|
| H A D | powerpc-vpadtl.c | 491 for (i = 0; i < vpa->queues.nr_queues; i++) { in powerpc_vpadtl__setup_queues() 587 for (unsigned int i = 0; i < queues->nr_queues; i++) in powerpc_vpadtl_free_events()
|
| /linux/drivers/block/rnbd/ |
| H A D | rnbd-clt.c | 1173 set->map[HCTX_TYPE_DEFAULT].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1175 set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1182 set->map[HCTX_TYPE_POLL].nr_queues = sess->nr_poll_queues; in rnbd_rdma_map_queues() 1184 set->map[HCTX_TYPE_READ].nr_queues; in rnbd_rdma_map_queues() 1188 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1189 set->map[HCTX_TYPE_READ].nr_queues, in rnbd_rdma_map_queues() 1190 set->map[HCTX_TYPE_POLL].nr_queues); in rnbd_rdma_map_queues() 1194 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1195 set->map[HCTX_TYPE_READ].nr_queues); in rnbd_rdma_map_queues()
|
| /linux/drivers/nvme/target/ |
| H A D | pci-epf.c | 164 unsigned int nr_queues; member 506 ctrl->nr_queues); in nvmet_pci_epf_alloc_irq_vectors() 532 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_find_irq_vector() 555 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_add_irq_vector() 1565 ctrl->sq = kzalloc_objs(struct nvmet_pci_epf_queue, ctrl->nr_queues); in nvmet_pci_epf_alloc_queues() 1569 ctrl->cq = kzalloc_objs(struct nvmet_pci_epf_queue, ctrl->nr_queues); in nvmet_pci_epf_alloc_queues() 1576 for (qid = 0; qid < ctrl->nr_queues; qid++) { in nvmet_pci_epf_alloc_queues() 1716 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_poll_sqs_work() 1930 for (qid = 1; qid < ctrl->nr_queues; qid++) in nvmet_pci_epf_disable_ctrl() 1933 for (qid = 1; qid < ctrl->nr_queues; qid++) in nvmet_pci_epf_disable_ctrl() [all …]
|
| /linux/drivers/virtio/ |
| H A D | virtio_rtc_driver.c | 1221 int nr_queues, ret; in viortc_init_vqs() local 1227 nr_queues = VIORTC_ALARMQ + 1; in viortc_init_vqs() 1229 nr_queues = VIORTC_REQUESTQ + 1; in viortc_init_vqs() 1231 ret = virtio_find_vqs(vdev, nr_queues, vqs, vqs_info, NULL); in viortc_init_vqs()
|
| /linux/drivers/nvme/host/ |
| H A D | fabrics.c | 1163 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues() 1166 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues() 1172 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues() 1175 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues() 1184 set->map[HCTX_TYPE_POLL].nr_queues = io_queues[HCTX_TYPE_POLL]; in nvmf_map_queues()
|
| H A D | pci.c | 690 map->nr_queues = dev->io_queues[i]; in nvme_pci_map_queues() 691 if (!map->nr_queues) { in nvme_pci_map_queues() 705 qoff += map->nr_queues; in nvme_pci_map_queues() 706 offset += map->nr_queues; in nvme_pci_map_queues() 3092 int nr_queues = dev->online_queues - 1, sent = 0; in __nvme_delete_io_queues() local 3097 while (nr_queues > 0) { in __nvme_delete_io_queues() 3098 if (nvme_delete_queue(&dev->queues[nr_queues], opcode)) in __nvme_delete_io_queues() 3100 nr_queues--; in __nvme_delete_io_queues() 3104 struct nvme_queue *nvmeq = &dev->queues[nr_queues + sent]; in __nvme_delete_io_queues() 3112 if (nr_queues) in __nvme_delete_io_queues()
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | kublk.c | 1302 unsigned nr_queues = ctx->nr_hw_queues; in __cmd_dev_add() local 1320 if (nr_queues > UBLK_MAX_QUEUES || depth > UBLK_QUEUE_DEPTH) { in __cmd_dev_add() 1322 __func__, nr_queues, depth); in __cmd_dev_add() 1329 nthreads = nr_queues; in __cmd_dev_add() 1338 if (nthreads != nr_queues && (!ctx->per_io_tasks && in __cmd_dev_add() 1342 __func__, nthreads, nr_queues); in __cmd_dev_add() 1369 info->nr_hw_queues = nr_queues; in __cmd_dev_add()
|
| /linux/drivers/block/ |
| H A D | zloop.c | 77 unsigned int nr_queues; member 1051 opts->nr_queues * opts->queue_depth, zlo->id); in zloop_ctl_add() 1093 zlo->tag_set.nr_hw_queues = opts->nr_queues; in zloop_ctl_add() 1227 opts->nr_queues = ZLOOP_DEF_NR_QUEUES; in zloop_parse_options() 1324 opts->nr_queues = min(token, num_online_cpus()); in zloop_parse_options()
|
| /linux/drivers/char/ |
| H A D | virtio_console.c | 1806 u32 i, j, nr_ports, nr_queues; in init_vqs() local 1810 nr_queues = use_multiport(portdev) ? (nr_ports + 1) * 2 : 2; in init_vqs() 1812 vqs = kmalloc_objs(struct virtqueue *, nr_queues); in init_vqs() 1813 vqs_info = kzalloc_objs(*vqs_info, nr_queues); in init_vqs() 1847 err = virtio_find_vqs(portdev->vdev, nr_queues, vqs, vqs_info, NULL); in init_vqs()
|
| /linux/Documentation/admin-guide/blockdev/ |
| H A D | zoned_loop.rst | 65 …ity_mb=%u,zone_size_mb=%u,zone_capacity_mb=%u,conv_zones=%u,base_dir=%s,nr_queues=%u,queue_depth=%… 89 nr_queues Number of I/O queues of the zoned block device. This
|