Home
last modified time | relevance | path

Searched refs:queue_depth (Results 1 – 25 of 89) sorted by relevance

1234

/linux/drivers/target/
H A Dtarget_core_tpg.c158 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() argument
160 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth()
162 if (!acl->queue_depth) { in target_set_nacl_queue_depth()
166 acl->queue_depth = 1; in target_set_nacl_queue_depth()
174 u32 queue_depth; in target_alloc_node_acl() local
191 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl()
193 queue_depth = 1; in target_alloc_node_acl()
194 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl()
218 acl->queue_depth, in target_add_node_acl()
369 tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth, in core_tpg_del_initiator_node_acl()
[all …]
H A Dtarget_core_pscsi.c291 if (!sd->queue_depth) { in pscsi_add_device_to_list()
292 sd->queue_depth = PSCSI_DEFAULT_QUEUEDEPTH; in pscsi_add_device_to_list()
296 sd->lun, sd->queue_depth); in pscsi_add_device_to_list()
303 dev->dev_attrib.hw_queue_depth = sd->queue_depth; in pscsi_add_device_to_list()
/linux/arch/um/drivers/
H A Dvector_kern.c267 atomic_sub(advance, &qi->queue_depth); in vector_advancehead()
268 return atomic_read(&qi->queue_depth); in vector_advancehead()
281 atomic_add(advance, &qi->queue_depth); in vector_advancetail()
282 return atomic_read(&qi->queue_depth); in vector_advancetail()
329 int queue_depth; in vector_enqueue() local
335 queue_depth = atomic_read(&qi->queue_depth); in vector_enqueue()
340 if (queue_depth < qi->max_depth) { in vector_enqueue()
355 queue_depth = vector_advancetail(qi, 1); in vector_enqueue()
359 return queue_depth; in vector_enqueue()
368 return queue_depth; in vector_enqueue()
[all …]
H A Dvector_kern.h48 atomic_t queue_depth; member
/linux/drivers/infiniband/ulp/rtrs/
H A Drtrs-clt.c71 size_t max_depth = clt->queue_depth; in __rtrs_get_permit()
497 if (WARN_ON(msg_id >= clt_path->queue_depth)) in process_io_rsp()
551 if (WARN_ON(buf_id >= clt_path->queue_depth)) in rtrs_clt_rkey_rsp_done()
724 q_size = clt_path->queue_depth; in post_recv_path()
1323 for (i = 0; i < clt_path->queue_depth; ++i) { in fail_all_outstanding_reqs()
1349 for (i = 0; i < clt_path->queue_depth; ++i) { in free_path_reqs()
1367 clt_path->reqs = kzalloc_objs(*clt_path->reqs, clt_path->queue_depth); in alloc_path_reqs()
1376 for (i = 0; i < clt_path->queue_depth; ++i) { in alloc_path_reqs()
1415 clt->permits_map = bitmap_zalloc(clt->queue_depth, GFP_KERNEL); in alloc_permits()
1420 clt->permits = kcalloc(clt->queue_depth, permit_size(clt), GFP_KERNEL); in alloc_permits()
[all …]
H A Drtrs-srv.c112 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids()
141 srv_path->ops_ids = kzalloc_objs(*srv_path->ops_ids, srv->queue_depth); in rtrs_srv_alloc_ops_ids()
145 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids()
582 mrs_num = srv->queue_depth; in map_cont_bufs()
586 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs()
587 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs()
604 srv->queue_depth - chunks); in map_cont_bufs()
672 chunk_bits = ilog2(srv->queue_depth - 1) + 1; in map_cont_bufs()
1015 q_size = srv->queue_depth; in post_recv_path()
1284 if (msg_id >= srv->queue_depth || off >= max_chunk_size) { in rtrs_srv_rdma_done()
[all …]
H A Drtrs.h117 u32 queue_depth; member
/linux/drivers/ata/
H A Dlibata-sata.c1281 int queue_depth) in ata_change_queue_depth() argument
1290 if (!dev || queue_depth < 1 || queue_depth == sdev->queue_depth) { in ata_change_queue_depth()
1292 return sdev->queue_depth; in ata_change_queue_depth()
1301 if (queue_depth > max_queue_depth) { in ata_change_queue_depth()
1310 if (queue_depth == 1 || !ata_ncq_supported(dev)) { in ata_change_queue_depth()
1312 queue_depth = 1; in ata_change_queue_depth()
1319 if (queue_depth == sdev->queue_depth) in ata_change_queue_depth()
1320 return sdev->queue_depth; in ata_change_queue_depth()
1322 return scsi_change_queue_depth(sdev, queue_depth); in ata_change_queue_depth()
1341 int ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) in ata_scsi_change_queue_depth() argument
[all …]
/linux/drivers/target/iscsi/
H A Discsi_target_device.c41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn()
42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
H A Discsi_target_nego.c1118 u32 payload_length, queue_depth = 0; in iscsi_target_locate_portal() local
1299 queue_depth = se_nacl->queue_depth; in iscsi_target_locate_portal()
1309 tag_num = max_t(u32, ISCSIT_MIN_TAGS, queue_depth); in iscsi_target_locate_portal()
/linux/include/net/mana/
H A Dhw_channel.h150 u16 queue_depth;
164 u16 queue_depth;
148 u16 queue_depth; global() member
162 u16 queue_depth; global() member
/linux/block/
H A Dblk-mq.c3928 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument
3938 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
3962 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
4147 set->queue_depth); in __blk_mq_alloc_map_and_rqs()
4665 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
4666 q->async_depth = set->queue_depth; in blk_mq_init_allocated_queue()
4699 set->queue_depth); in __blk_mq_alloc_rq_maps()
4734 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs()
4740 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs()
4741 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_set_map_and_rqs()
[all …]
/linux/tools/testing/selftests/ublk/
H A Dkublk.c349 info->dev_id, info->nr_hw_queues, info->queue_depth, in ublk_ctrl_dump()
453 int depth = dev->dev_info.queue_depth; in ublk_queue_init()
525 cq_depth += dev->dev_info.queue_depth * 2; in ublk_thread_init()
538 unsigned nr_ios = dev->dev_info.queue_depth * dev->dev_info.nr_hw_queues; in ublk_thread_init()
792 int nr_ios = dinfo->nr_hw_queues * dinfo->queue_depth; in ublk_submit_fetch_commands()
794 int q_id = i / dinfo->queue_depth; in ublk_submit_fetch_commands()
795 int tag = i % dinfo->queue_depth; in ublk_submit_fetch_commands()
1304 unsigned depth = ctx->queue_depth; in __cmd_dev_add()
1370 info->queue_depth = depth; in __cmd_dev_add()
1800 .queue_depth = 128, in main()
[all …]
H A Dstripe.c350 dev->tgt.sq_depth = mul * dev->dev_info.queue_depth * conf->nr_files; in ublk_stripe_tgt_init()
351 dev->tgt.cq_depth = mul * dev->dev_info.queue_depth * conf->nr_files; in ublk_stripe_tgt_init()
/linux/drivers/block/rnbd/
H A Drnbd-srv.h25 int queue_depth; member
H A Drnbd-clt.h86 size_t queue_depth; member
/linux/drivers/scsi/snic/
H A Dsnic_main.c91 if (qsz < sdev->queue_depth) in snic_change_queue_depth()
93 else if (qsz > sdev->queue_depth) in snic_change_queue_depth()
96 atomic64_set(&snic->s_stats.misc.last_qsz, sdev->queue_depth); in snic_change_queue_depth()
100 return sdev->queue_depth; in snic_change_queue_depth()
/linux/drivers/block/
H A Dvirtio_blk.c1242 module_param_named(queue_depth, virtblk_queue_depth, uint, 0444);
1444 unsigned int queue_depth; in virtblk_probe() local
1476 queue_depth = vblk->vqs[0].vq->num_free; in virtblk_probe()
1479 queue_depth /= 2; in virtblk_probe()
1481 queue_depth = virtblk_queue_depth; in virtblk_probe()
1486 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe()
H A Dzloop.c78 unsigned int queue_depth; member
1051 opts->nr_queues * opts->queue_depth, zlo->id); in zloop_ctl_add()
1094 zlo->tag_set.queue_depth = opts->queue_depth; in zloop_ctl_add()
1228 opts->queue_depth = ZLOOP_DEF_QUEUE_DEPTH; in zloop_parse_options()
1336 opts->queue_depth = token; in zloop_parse_options()
/linux/rust/kernel/block/mq/
H A Dtag_set.rs50 queue_depth: num_tags, in new()
/linux/drivers/scsi/
H A Dscsi.c225 sdev->queue_depth = depth; in scsi_change_queue_depth()
232 sbitmap_resize(&sdev->budget_map, sdev->queue_depth); in scsi_change_queue_depth()
234 return sdev->queue_depth; in scsi_change_queue_depth()
/linux/drivers/net/ethernet/microsoft/mana/
H A Dhw_channel.c357 comp_read = mana_gd_poll_cq(q_self, completions, hwc_cq->queue_depth); in mana_hwc_comp_event()
358 WARN_ON_ONCE(comp_read <= 0 || comp_read > hwc_cq->queue_depth); in mana_hwc_comp_event()
437 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq()
552 hwc_wq->queue_depth = q_depth; in mana_hwc_create_wq()
/linux/drivers/scsi/bfa/
H A Dbfad_im.c91 (bfa_lun_queue_depth > cmnd->device->queue_depth)) { in bfa_cb_ioim_done()
119 if (bfa_lun_queue_depth > cmnd->device->queue_depth) { in bfa_cb_ioim_good_comp()
879 if (bfa_lun_queue_depth > tmp_sdev->queue_depth) { in bfad_ramp_up_qdepth()
883 tmp_sdev->queue_depth + 1); in bfad_ramp_up_qdepth()
901 scsi_track_queue_full(tmp_sdev, tmp_sdev->queue_depth - 1); in bfad_handle_qfull()
/linux/include/target/
H A Dtarget_core_base.h585 u32 queue_depth;
733 u32 queue_depth;
836 u32 queue_depth;
581 u32 queue_depth; global() member
729 u32 queue_depth; global() member
832 u32 queue_depth; global() member
/linux/drivers/mmc/core/
H A Dqueue.c439 mq->tag_set.queue_depth = in mmc_init_queue()
442 mq->tag_set.queue_depth = MMC_QUEUE_DEPTH; in mmc_init_queue()

1234