| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_main.c | 42 if (cptvf->nr_queues) { in init_worker_threads() 44 cptvf->nr_queues); in init_worker_threads() 47 for (i = 0; i < cptvf->nr_queues; i++) { in init_worker_threads() 69 if (cptvf->nr_queues) { in cleanup_worker_threads() 71 cptvf->nr_queues); in cleanup_worker_threads() 74 for (i = 0; i < cptvf->nr_queues; i++) in cleanup_worker_threads() 100 pqinfo->nr_queues = 0; in free_pending_queues() 104 u32 nr_queues) in alloc_pending_queues() argument 110 pqinfo->nr_queues = nr_queues; in alloc_pending_queues() 136 static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) in init_pending_queues() argument [all …]
|
| H A D | cptvf.h | 85 u32 nr_queues; /* Number of queues supported */ member 91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \ 110 u32 nr_queues; member
|
| H A D | cptvf_reqmanager.c | 233 if (unlikely(qno >= cptvf->nr_queues)) { in send_cpt_command() 235 qno, cptvf->nr_queues); in send_cpt_command() 545 if (unlikely(qno > cptvf->nr_queues)) { in vq_post_process()
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_sriov.c | 58 int nr_queues = 0; in vf_mode_to_nr_queues() local 62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues() 65 nr_queues = 8; in vf_mode_to_nr_queues() 68 nr_queues = 4; in vf_mode_to_nr_queues() 71 nr_queues = 2; in vf_mode_to_nr_queues() 74 nr_queues = 1; in vf_mode_to_nr_queues() 78 return nr_queues; in vf_mode_to_nr_queues()
|
| H A D | nitrox_lib.c | 91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues() 102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues() 142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues() 155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues() 161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
|
| H A D | nitrox_dev.h | 160 int nr_queues; member 250 u16 nr_queues; member
|
| H A D | nitrox_mbx.c | 68 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response() 76 vfdev->nr_queues = 0; in pf2vf_send_response()
|
| H A D | nitrox_hal.c | 124 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_config_pkt_input_rings() 240 for (i = 0; i < ndev->nr_queues; i++) in nitrox_config_pkt_solicit_ports() 356 for (ring = 0; ring < ndev->nr_queues; ring++) { in nitrox_config_aqm_rings()
|
| /linux/block/ |
| H A D | blk-mq-cpumap.c | 64 masks = group_cpus_evenly(qmap->nr_queues, &nr_masks); in blk_mq_map_queues() 71 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_queues() 118 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_hw_queues()
|
| H A D | blk-mq.h | 462 q->tag_set->map[HCTX_TYPE_POLL].nr_queues; in blk_mq_can_poll()
|
| /linux/tools/perf/util/ |
| H A D | auxtrace.c | 214 static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues) in auxtrace_alloc_queue_array() argument 220 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array() 223 queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue)); in auxtrace_alloc_queue_array() 227 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array() 235 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues) in auxtrace_queues__init_nr() argument 237 queues->nr_queues = nr_queues; in auxtrace_queues__init_nr() 238 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init_nr() 252 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local 256 if (!nr_queues) in auxtrace_queues__grow() 257 nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__grow() [all …]
|
| H A D | auxtrace.h | 292 unsigned int nr_queues; member 518 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues);
|
| H A D | cs-etm.c | 309 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__map_trace_id_v0() 367 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__process_trace_id_v0_1() 995 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__free_events() 2590 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__process_timeless_queues() 2628 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues() 2723 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues() 2872 for (i = 0; i < etm->queues.nr_queues; ++i) in dump_queued_data() 3330 for (unsigned int i = 0; i < queues->nr_queues; i++) { in cs_etm__create_decoders() 3442 for (unsigned int j = 0; j < etm->queues.nr_queues; ++j) { in cs_etm__process_auxtrace_info_full()
|
| H A D | s390-cpumsf.c | 203 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) in s390_cpumcf_dumpctr() 827 for (i = 0; i < sf->queues.nr_queues; i++) { in s390_cpumsf_setup_queues() 1016 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
|
| H A D | intel-bts.c | 211 for (i = 0; i < bts->queues.nr_queues; i++) { in intel_bts_setup_queues() 544 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit() 715 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
|
| H A D | powerpc-vpadtl.c | 494 for (i = 0; i < vpa->queues.nr_queues; i++) { in powerpc_vpadtl__setup_queues() 590 for (unsigned int i = 0; i < queues->nr_queues; i++) in powerpc_vpadtl_free_events()
|
| H A D | intel-pt.c | 1422 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_first_timestamp() 1662 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_setup_queues() 3022 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_enable_sync_switch() 3037 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_disable_sync_switch() 3275 for (i = 0; i < queues->nr_queues; i++) { in intel_pt_process_timeless_queues() 3349 if (cpu < 0 || !pt->queues.nr_queues) in intel_pt_cpu_to_ptq() 3352 if ((unsigned)cpu >= pt->queues.nr_queues) in intel_pt_cpu_to_ptq() 3353 i = pt->queues.nr_queues - 1; in intel_pt_cpu_to_ptq() 3365 for (; j < pt->queues.nr_queues; j++) { in intel_pt_cpu_to_ptq() 3816 for (i = 0; i < queues->nr_queues; i++) { in intel_pt_free_events()
|
| /linux/drivers/block/rnbd/ |
| H A D | rnbd-clt.c | 1173 set->map[HCTX_TYPE_DEFAULT].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1175 set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1182 set->map[HCTX_TYPE_POLL].nr_queues = sess->nr_poll_queues; in rnbd_rdma_map_queues() 1184 set->map[HCTX_TYPE_READ].nr_queues; in rnbd_rdma_map_queues() 1188 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1189 set->map[HCTX_TYPE_READ].nr_queues, in rnbd_rdma_map_queues() 1190 set->map[HCTX_TYPE_POLL].nr_queues); in rnbd_rdma_map_queues() 1194 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1195 set->map[HCTX_TYPE_READ].nr_queues); in rnbd_rdma_map_queues()
|
| /linux/drivers/nvme/target/ |
| H A D | pci-epf.c | 164 unsigned int nr_queues; member 506 ctrl->nr_queues); in nvmet_pci_epf_alloc_irq_vectors() 532 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_find_irq_vector() 555 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_add_irq_vector() 1565 ctrl->sq = kzalloc_objs(struct nvmet_pci_epf_queue, ctrl->nr_queues); in nvmet_pci_epf_alloc_queues() 1569 ctrl->cq = kzalloc_objs(struct nvmet_pci_epf_queue, ctrl->nr_queues); in nvmet_pci_epf_alloc_queues() 1576 for (qid = 0; qid < ctrl->nr_queues; qid++) { in nvmet_pci_epf_alloc_queues() 1716 for (i = 0; i < ctrl->nr_queues; i++) { in nvmet_pci_epf_poll_sqs_work() 1930 for (qid = 1; qid < ctrl->nr_queues; qid++) in nvmet_pci_epf_disable_ctrl() 1933 for (qid = 1; qid < ctrl->nr_queues; qid++) in nvmet_pci_epf_disable_ctrl() [all …]
|
| /linux/drivers/virtio/ |
| H A D | virtio_rtc_driver.c | 1221 int nr_queues, ret; in viortc_init_vqs() local 1227 nr_queues = VIORTC_ALARMQ + 1; in viortc_init_vqs() 1229 nr_queues = VIORTC_REQUESTQ + 1; in viortc_init_vqs() 1231 ret = virtio_find_vqs(vdev, nr_queues, vqs, vqs_info, NULL); in viortc_init_vqs()
|
| /linux/drivers/nvme/host/ |
| H A D | fabrics.c | 1163 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues() 1166 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues() 1172 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues() 1175 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues() 1184 set->map[HCTX_TYPE_POLL].nr_queues = io_queues[HCTX_TYPE_POLL]; in nvmf_map_queues()
|
| H A D | pci.c | 690 map->nr_queues = dev->io_queues[i]; in nvme_pci_map_queues() 691 if (!map->nr_queues) { in nvme_pci_map_queues() 705 qoff += map->nr_queues; in nvme_pci_map_queues() 706 offset += map->nr_queues; in nvme_pci_map_queues() 3093 int nr_queues = dev->online_queues - 1, sent = 0; in __nvme_delete_io_queues() local 3098 while (nr_queues > 0) { in __nvme_delete_io_queues() 3099 if (nvme_delete_queue(&dev->queues[nr_queues], opcode)) in __nvme_delete_io_queues() 3101 nr_queues--; in __nvme_delete_io_queues() 3105 struct nvme_queue *nvmeq = &dev->queues[nr_queues + sent]; in __nvme_delete_io_queues() 3113 if (nr_queues) in __nvme_delete_io_queues()
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | kublk.c | 1642 unsigned nr_queues = ctx->nr_hw_queues; in cmd_dev_get_features() 1660 if (nr_queues > UBLK_MAX_QUEUES || depth > UBLK_QUEUE_DEPTH) { in cmd_dev_update_size() 1661 ublk_err("%s: invalid nr_queues or depth queues %u depth %u\n", in cmd_dev_update_size() 1662 __func__, nr_queues, depth); in cmd_dev_update_size() 1669 nthreads = nr_queues; in cmd_dev_update_size() 1678 if (nthreads != nr_queues && (!ctx->per_io_tasks && in cmd_dev_update_size() 1682 __func__, nthreads, nr_queues); in cmd_dev_update_size() 1709 info->nr_hw_queues = nr_queues; in __cmd_create_help() 2052 printf("%s %s -t [null|loop|stripe|fault_inject] [-q nr_queues] [-d depth] [-n dev_id]\n", 2061 printf("\tdefault: nr_queues 1302 unsigned nr_queues = ctx->nr_hw_queues; __cmd_dev_add() local [all...] |
| /linux/drivers/char/ |
| H A D | virtio_console.c | 1806 u32 i, j, nr_ports, nr_queues; in init_vqs() local 1810 nr_queues = use_multiport(portdev) ? (nr_ports + 1) * 2 : 2; in init_vqs() 1812 vqs = kmalloc_objs(struct virtqueue *, nr_queues); in init_vqs() 1813 vqs_info = kzalloc_objs(*vqs_info, nr_queues); in init_vqs() 1847 err = virtio_find_vqs(portdev->vdev, nr_queues, vqs, vqs_info, NULL); in init_vqs()
|
| /linux/drivers/block/ |
| H A D | zloop.c | 84 unsigned int nr_queues; member 1244 opts->nr_queues * opts->queue_depth, zlo->id); in zloop_ctl_add() 1287 zlo->tag_set.nr_hw_queues = opts->nr_queues; in zloop_ctl_add() 1465 opts->nr_queues = ZLOOP_DEF_NR_QUEUES; in zloop_parse_options() 1569 opts->nr_queues = min(token, num_online_cpus()); in zloop_parse_options()
|