Home
last modified time | relevance | path

Searched refs:nr_queues (Results 1 – 25 of 49) sorted by relevance

12

/linux/drivers/crypto/cavium/cpt/
H A Dcptvf_main.c42 if (cptvf->nr_queues) { in init_worker_threads()
44 cptvf->nr_queues); in init_worker_threads()
47 for (i = 0; i < cptvf->nr_queues; i++) { in init_worker_threads()
69 if (cptvf->nr_queues) { in cleanup_worker_threads()
71 cptvf->nr_queues); in cleanup_worker_threads()
74 for (i = 0; i < cptvf->nr_queues; i++) in cleanup_worker_threads()
100 pqinfo->nr_queues = 0; in free_pending_queues()
104 u32 nr_queues) in alloc_pending_queues() argument
110 pqinfo->nr_queues = nr_queues; in alloc_pending_queues()
136 static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) in init_pending_queues() argument
[all …]
H A Dcptvf.h85 u32 nr_queues; /* Number of queues supported */ member
91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \
110 u32 nr_queues; member
H A Dcptvf_reqmanager.c233 if (unlikely(qno >= cptvf->nr_queues)) { in send_cpt_command()
235 qno, cptvf->nr_queues); in send_cpt_command()
545 if (unlikely(qno > cptvf->nr_queues)) { in vq_post_process()
/linux/drivers/crypto/cavium/nitrox/
H A Dnitrox_sriov.c58 int nr_queues = 0; in vf_mode_to_nr_queues() local
62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues()
65 nr_queues = 8; in vf_mode_to_nr_queues()
68 nr_queues = 4; in vf_mode_to_nr_queues()
71 nr_queues = 2; in vf_mode_to_nr_queues()
74 nr_queues = 1; in vf_mode_to_nr_queues()
78 return nr_queues; in vf_mode_to_nr_queues()
H A Dnitrox_lib.c91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues()
102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues()
142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues()
155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues()
161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
H A Dnitrox_dev.h160 int nr_queues; member
250 u16 nr_queues; member
H A Dnitrox_mbx.c68 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response()
76 vfdev->nr_queues = 0; in pf2vf_send_response()
H A Dnitrox_hal.c124 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_config_pkt_input_rings()
240 for (i = 0; i < ndev->nr_queues; i++) in nitrox_config_pkt_solicit_ports()
356 for (ring = 0; ring < ndev->nr_queues; ring++) { in nitrox_config_aqm_rings()
/linux/block/
H A Dblk-mq-cpumap.c64 masks = group_cpus_evenly(qmap->nr_queues, &nr_masks); in blk_mq_map_queues()
71 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_queues()
118 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_hw_queues()
/linux/fs/fuse/
H A Ddev_uring.c128 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_abort_end_requests()
169 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_request_expired()
196 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_destruct()
230 size_t nr_queues = num_possible_cpus(); in fuse_uring_create() local
238 ring->queues = kzalloc_objs(struct fuse_ring_queue *, nr_queues, in fuse_uring_create()
256 ring->nr_queues = nr_queues; in fuse_uring_create()
408 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_log_ent_state()
439 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_async_stop_queues()
474 for (qid = 0; qid < ring->nr_queues; qid++) { in fuse_uring_stop_queues()
897 if (qid >= ring->nr_queues) in fuse_uring_commit_fetch()
[all …]
/linux/tools/perf/util/
H A Dauxtrace.c214 static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues) in auxtrace_alloc_queue_array() argument
220 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array()
223 queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue)); in auxtrace_alloc_queue_array()
227 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array()
235 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues) in auxtrace_queues__init_nr() argument
237 queues->nr_queues = nr_queues; in auxtrace_queues__init_nr()
238 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init_nr()
252 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local
256 if (!nr_queues) in auxtrace_queues__grow()
257 nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__grow()
[all …]
H A Dauxtrace.h292 unsigned int nr_queues; member
518 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues);
H A Dcs-etm.c309 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__map_trace_id_v0()
367 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__process_trace_id_v0_1()
995 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__free_events()
2590 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__process_timeless_queues()
2628 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues()
2723 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues()
2872 for (i = 0; i < etm->queues.nr_queues; ++i) in dump_queued_data()
3330 for (unsigned int i = 0; i < queues->nr_queues; i++) { in cs_etm__create_decoders()
3442 for (unsigned int j = 0; j < etm->queues.nr_queues; ++j) { in cs_etm__process_auxtrace_info_full()
H A Ds390-cpumsf.c203 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) in s390_cpumcf_dumpctr()
827 for (i = 0; i < sf->queues.nr_queues; i++) { in s390_cpumsf_setup_queues()
1016 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
H A Dintel-bts.c211 for (i = 0; i < bts->queues.nr_queues; i++) { in intel_bts_setup_queues()
544 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit()
715 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
H A Dpowerpc-vpadtl.c491 for (i = 0; i < vpa->queues.nr_queues; i++) { in powerpc_vpadtl__setup_queues()
587 for (unsigned int i = 0; i < queues->nr_queues; i++) in powerpc_vpadtl_free_events()
/linux/drivers/block/rnbd/
H A Drnbd-clt.c1173 set->map[HCTX_TYPE_DEFAULT].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues()
1175 set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues()
1182 set->map[HCTX_TYPE_POLL].nr_queues = sess->nr_poll_queues; in rnbd_rdma_map_queues()
1184 set->map[HCTX_TYPE_READ].nr_queues; in rnbd_rdma_map_queues()
1188 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues()
1189 set->map[HCTX_TYPE_READ].nr_queues, in rnbd_rdma_map_queues()
1190 set->map[HCTX_TYPE_POLL].nr_queues); in rnbd_rdma_map_queues()
1194 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues()
1195 set->map[HCTX_TYPE_READ].nr_queues); in rnbd_rdma_map_queues()
/linux/drivers/nvme/target/
H A Dpci-epf.c164 unsigned int nr_queues; member
506 ctrl->nr_queues); in nvmet_pci_epf_alloc_irq_vectors()
532 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_find_irq_vector()
555 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_add_irq_vector()
1565 ctrl->sq = kzalloc_objs(struct nvmet_pci_epf_queue, ctrl->nr_queues); in nvmet_pci_epf_alloc_queues()
1569 ctrl->cq = kzalloc_objs(struct nvmet_pci_epf_queue, ctrl->nr_queues); in nvmet_pci_epf_alloc_queues()
1576 for (qid = 0; qid < ctrl->nr_queues; qid++) { in nvmet_pci_epf_alloc_queues()
1716 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_poll_sqs_work()
1930 for (qid = 1; qid < ctrl->nr_queues; qid++) in nvmet_pci_epf_disable_ctrl()
1933 for (qid = 1; qid < ctrl->nr_queues; qid++) in nvmet_pci_epf_disable_ctrl()
[all …]
/linux/drivers/virtio/
H A Dvirtio_rtc_driver.c1221 int nr_queues, ret; in viortc_init_vqs() local
1227 nr_queues = VIORTC_ALARMQ + 1; in viortc_init_vqs()
1229 nr_queues = VIORTC_REQUESTQ + 1; in viortc_init_vqs()
1231 ret = virtio_find_vqs(vdev, nr_queues, vqs, vqs_info, NULL); in viortc_init_vqs()
/linux/drivers/nvme/host/
H A Dfabrics.c1163 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues()
1166 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues()
1172 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues()
1175 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues()
1184 set->map[HCTX_TYPE_POLL].nr_queues = io_queues[HCTX_TYPE_POLL]; in nvmf_map_queues()
H A Dpci.c690 map->nr_queues = dev->io_queues[i]; in nvme_pci_map_queues()
691 if (!map->nr_queues) { in nvme_pci_map_queues()
705 qoff += map->nr_queues; in nvme_pci_map_queues()
706 offset += map->nr_queues; in nvme_pci_map_queues()
3092 int nr_queues = dev->online_queues - 1, sent = 0; in __nvme_delete_io_queues() local
3097 while (nr_queues > 0) { in __nvme_delete_io_queues()
3098 if (nvme_delete_queue(&dev->queues[nr_queues], opcode)) in __nvme_delete_io_queues()
3100 nr_queues--; in __nvme_delete_io_queues()
3104 struct nvme_queue *nvmeq = &dev->queues[nr_queues + sent]; in __nvme_delete_io_queues()
3112 if (nr_queues) in __nvme_delete_io_queues()
/linux/tools/testing/selftests/ublk/
H A Dkublk.c1302 unsigned nr_queues = ctx->nr_hw_queues; in __cmd_dev_add() local
1320 if (nr_queues > UBLK_MAX_QUEUES || depth > UBLK_QUEUE_DEPTH) { in __cmd_dev_add()
1322 __func__, nr_queues, depth); in __cmd_dev_add()
1329 nthreads = nr_queues; in __cmd_dev_add()
1338 if (nthreads != nr_queues && (!ctx->per_io_tasks && in __cmd_dev_add()
1342 __func__, nthreads, nr_queues); in __cmd_dev_add()
1369 info->nr_hw_queues = nr_queues; in __cmd_dev_add()
/linux/drivers/block/
H A Dzloop.c77 unsigned int nr_queues; member
1051 opts->nr_queues * opts->queue_depth, zlo->id); in zloop_ctl_add()
1093 zlo->tag_set.nr_hw_queues = opts->nr_queues; in zloop_ctl_add()
1227 opts->nr_queues = ZLOOP_DEF_NR_QUEUES; in zloop_parse_options()
1324 opts->nr_queues = min(token, num_online_cpus()); in zloop_parse_options()
/linux/drivers/char/
H A Dvirtio_console.c1806 u32 i, j, nr_ports, nr_queues; in init_vqs() local
1810 nr_queues = use_multiport(portdev) ? (nr_ports + 1) * 2 : 2; in init_vqs()
1812 vqs = kmalloc_objs(struct virtqueue *, nr_queues); in init_vqs()
1813 vqs_info = kzalloc_objs(*vqs_info, nr_queues); in init_vqs()
1847 err = virtio_find_vqs(portdev->vdev, nr_queues, vqs, vqs_info, NULL); in init_vqs()
/linux/Documentation/admin-guide/blockdev/
H A Dzoned_loop.rst65 …ity_mb=%u,zone_size_mb=%u,zone_capacity_mb=%u,conv_zones=%u,base_dir=%s,nr_queues=%u,queue_depth=%…
89 nr_queues Number of I/O queues of the zoned block device. This

12