Home
last modified time | relevance | path

Searched refs:tag_set (Results 1 – 25 of 61) sorted by relevance

123

/linux/rust/kernel/block/mq/
H A Dtag_set.rs43 let tag_set: bindings::blk_mq_tag_set = unsafe { core::mem::zeroed() }; in new() localVariable
44 let tag_set: Result<_> = core::mem::size_of::<RequestDataWrapper>() in new() localVariable
57 ..tag_set in new()
64 inner <- tag_set.pin_chain(|tag_set| { in new()
66 let tag_set: &mut Opaque<_> = unsafe { Pin::get_unchecked_mut(tag_set) }; in new() localVariable
68 error::to_result( unsafe { bindings::blk_mq_alloc_tag_set(tag_set.get())}) in new()
/linux/drivers/md/
H A Ddm-rq.c543 md->tag_set = kzalloc_node(sizeof(struct blk_mq_tag_set), GFP_KERNEL, md->numa_node_id); in dm_mq_init_request_queue()
544 if (!md->tag_set) in dm_mq_init_request_queue()
547 md->tag_set->ops = &dm_mq_ops; in dm_mq_init_request_queue()
548 md->tag_set->queue_depth = dm_get_blk_mq_queue_depth(); in dm_mq_init_request_queue()
549 md->tag_set->numa_node = md->numa_node_id; in dm_mq_init_request_queue()
550 md->tag_set->flags = BLK_MQ_F_STACKING; in dm_mq_init_request_queue()
551 md->tag_set->nr_hw_queues = dm_get_blk_mq_nr_hw_queues(); in dm_mq_init_request_queue()
552 md->tag_set->driver_data = md; in dm_mq_init_request_queue()
554 md->tag_set->cmd_size = sizeof(struct dm_rq_target_io); in dm_mq_init_request_queue()
558 md->tag_set->cmd_size += immutable_tgt->per_io_data_size; in dm_mq_init_request_queue()
[all …]
/linux/drivers/mmc/core/
H A Dqueue.c384 disk = blk_mq_alloc_disk(&mq->tag_set, &lim, mq); in mmc_alloc_disk()
432 memset(&mq->tag_set, 0, sizeof(mq->tag_set)); in mmc_init_queue()
433 mq->tag_set.ops = &mmc_mq_ops; in mmc_init_queue()
439 mq->tag_set.queue_depth = in mmc_init_queue()
442 mq->tag_set.queue_depth = MMC_QUEUE_DEPTH; in mmc_init_queue()
443 mq->tag_set.numa_node = NUMA_NO_NODE; in mmc_init_queue()
444 mq->tag_set.flags = BLK_MQ_F_BLOCKING; in mmc_init_queue()
445 mq->tag_set.nr_hw_queues = 1; in mmc_init_queue()
446 mq->tag_set.cmd_size = sizeof(struct mmc_queue_req); in mmc_init_queue()
447 mq->tag_set.driver_data = mq; in mmc_init_queue()
[all …]
/linux/drivers/mtd/ubi/
H A Dblock.c85 struct blk_mq_tag_set tag_set; member
383 dev->tag_set.ops = &ubiblock_mq_ops; in ubiblock_create()
384 dev->tag_set.queue_depth = 64; in ubiblock_create()
385 dev->tag_set.numa_node = NUMA_NO_NODE; in ubiblock_create()
386 dev->tag_set.flags = BLK_MQ_F_BLOCKING; in ubiblock_create()
387 dev->tag_set.cmd_size = sizeof(struct ubiblock_pdu); in ubiblock_create()
388 dev->tag_set.driver_data = dev; in ubiblock_create()
389 dev->tag_set.nr_hw_queues = 1; in ubiblock_create()
391 ret = blk_mq_alloc_tag_set(&dev->tag_set); in ubiblock_create()
400 gd = blk_mq_alloc_disk(&dev->tag_set, &lim, dev); in ubiblock_create()
[all …]
/linux/block/
H A Dbsg-lib.c22 struct blk_mq_tag_set tag_set; member
279 container_of(q->tag_set, struct bsg_set, tag_set); in bsg_queue_rq()
324 container_of(q->tag_set, struct bsg_set, tag_set); in bsg_remove_queue()
329 blk_mq_free_tag_set(&bset->tag_set); in bsg_remove_queue()
338 container_of(rq->q->tag_set, struct bsg_set, tag_set); in bsg_timeout()
378 set = &bset->tag_set; in bsg_setup_queue()
H A Dblk-mq-sysfs.c232 mutex_lock(&q->tag_set->tag_list_lock); in blk_mq_sysfs_register()
238 mutex_unlock(&q->tag_set->tag_list_lock); in blk_mq_sysfs_register()
246 mutex_unlock(&q->tag_set->tag_list_lock); in blk_mq_sysfs_register()
259 mutex_lock(&q->tag_set->tag_list_lock); in blk_mq_sysfs_unregister()
262 mutex_unlock(&q->tag_set->tag_list_lock); in blk_mq_sysfs_unregister()
H A Delevator.c603 q->nr_requests = q->tag_set->queue_depth; in elevator_switch()
632 blk_mq_free_sched_tags(e->et, q->tag_set); in elv_exit_and_release()
647 blk_mq_free_sched_tags(ctx->old->et, q->tag_set); in elevator_change_done()
666 struct blk_mq_tag_set *set = q->tag_set; in elevator_change()
711 struct blk_mq_tag_set *set = q->tag_set; in elv_update_nr_hw_queues()
752 if (q->tag_set->flags & BLK_MQ_F_NO_SCHED_BY_DEFAULT) in elevator_set_default()
765 blk_mq_is_shared_tags(q->tag_set->flags))) { in elevator_set_default()
805 struct blk_mq_tag_set *set = q->tag_set; in elv_iosched_store()
H A Dblk-mq-tag.c271 struct blk_mq_tag_set *set = q->tag_set; in bt_iter()
507 if (blk_mq_is_shared_tags(q->tag_set->flags)) { in blk_mq_queue_tag_busy_iter()
508 struct blk_mq_tags *tags = q->tag_set->shared_tags; in blk_mq_queue_tag_busy_iter()
600 struct blk_mq_tag_set *set = hctx->queue->tag_set; in blk_mq_tag_update_depth()
648 q->nr_requests - q->tag_set->reserved_tags); in blk_mq_tag_update_sched_shared_tags()
H A Dblk-mq-sched.c549 unsigned int flags = q->tag_set->flags; in blk_mq_init_sched()
606 if (blk_mq_is_shared_tags(q->tag_set->flags)) { in blk_mq_sched_free_rqs()
607 blk_mq_free_rqs(q->tag_set, q->sched_shared_tags, in blk_mq_sched_free_rqs()
612 blk_mq_free_rqs(q->tag_set, in blk_mq_sched_free_rqs()
/linux/drivers/mtd/
H A Dmtd_blkdevs.c33 blk_mq_free_tag_set(dev->tag_set); in blktrans_dev_release()
34 kfree(dev->tag_set); in blktrans_dev_release()
327 new->tag_set = kzalloc(sizeof(*new->tag_set), GFP_KERNEL); in add_mtd_blktrans_dev()
328 if (!new->tag_set) in add_mtd_blktrans_dev()
331 ret = blk_mq_alloc_sq_tag_set(new->tag_set, &mtd_mq_ops, 2, in add_mtd_blktrans_dev()
343 gd = blk_mq_alloc_disk(new->tag_set, &lim, new); in add_mtd_blktrans_dev()
396 blk_mq_free_tag_set(new->tag_set); in add_mtd_blktrans_dev()
398 kfree(new->tag_set); in add_mtd_blktrans_dev()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Ddefiner.c133 (fc)->tag_set = &hws_definer_generic_set; \
557 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
562 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
567 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
572 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
577 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
582 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
587 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
592 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_steering_ok_bits_handler()
612 fc->tag_set = &hws_definer_generic_set; in hws_definer_flex_parser_handler()
[all …]
/linux/rust/kernel/block/
H A Dmq.rs94 mod tag_set; module
98 pub use tag_set::TagSet;
/linux/include/scsi/
H A Dscsi_tcq.h32 if (hwq < shost->tag_set.nr_hw_queues) { in scsi_host_find_tag()
33 req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq], in scsi_host_find_tag()
/linux/drivers/block/
H A Dnbd.c117 struct blk_mq_tag_set tag_set; member
269 blk_mq_free_tag_set(&nbd->tag_set); in nbd_dev_remove()
480 (config->num_connections == 1 && nbd->tag_set.timeout)) { in nbd_xmit_timeout()
513 if (!nbd->tag_set.timeout) { in nbd_xmit_timeout()
883 if (hwq < nbd->tag_set.nr_hw_queues) in nbd_handle_reply()
884 req = blk_mq_tag_to_rq(nbd->tag_set.tags[hwq], in nbd_handle_reply()
1050 blk_mq_tagset_busy_iter(&nbd->tag_set, nbd_clear_req, NULL); in nbd_clear_que()
1332 if (nbd->tag_set.timeout) in nbd_reconnect_socket()
1333 sock->sk->sk_sndtimeo = nbd->tag_set.timeout; in nbd_reconnect_socket()
1452 nbd->tag_set.timeout = 0; in nbd_config_put()
[all …]
H A Dvirtio_blk.c71 struct blk_mq_tag_set tag_set; member
1478 memset(&vblk->tag_set, 0, sizeof(vblk->tag_set)); in virtblk_probe()
1479 vblk->tag_set.ops = &virtio_mq_ops; in virtblk_probe()
1480 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe()
1481 vblk->tag_set.numa_node = NUMA_NO_NODE; in virtblk_probe()
1482 vblk->tag_set.cmd_size = in virtblk_probe()
1485 vblk->tag_set.driver_data = vblk; in virtblk_probe()
1486 vblk->tag_set.nr_hw_queues = vblk->num_vqs; in virtblk_probe()
1487 vblk->tag_set.nr_maps = 1; in virtblk_probe()
1489 vblk->tag_set.nr_maps = 3; in virtblk_probe()
[all …]
H A Dataflop.c305 struct blk_mq_tag_set tag_set; member
2001 disk = blk_mq_alloc_disk(&unit[drive].tag_set, &lim, NULL); in ataflop_alloc_disk()
2055 blk_mq_free_tag_set(&unit[i].tag_set); in atari_floppy_cleanup()
2073 blk_mq_free_tag_set(&fs->tag_set); in atari_cleanup_floppy_disk()
2086 memset(&unit[i].tag_set, 0, sizeof(unit[i].tag_set)); in atari_floppy_init()
2087 unit[i].tag_set.ops = &ataflop_mq_ops; in atari_floppy_init()
2088 unit[i].tag_set.nr_hw_queues = 1; in atari_floppy_init()
2089 unit[i].tag_set.nr_maps = 1; in atari_floppy_init()
2090 unit[i].tag_set.queue_depth = 2; in atari_floppy_init()
2091 unit[i].tag_set.numa_node = NUMA_NO_NODE; in atari_floppy_init()
[all …]
H A Dzloop.c106 struct blk_mq_tag_set tag_set; member
703 blk_mq_free_tag_set(&zlo->tag_set); in zloop_free_disk()
980 zlo->tag_set.ops = &zloop_mq_ops; in zloop_ctl_add()
981 zlo->tag_set.nr_hw_queues = opts->nr_queues; in zloop_ctl_add()
982 zlo->tag_set.queue_depth = opts->queue_depth; in zloop_ctl_add()
983 zlo->tag_set.numa_node = NUMA_NO_NODE; in zloop_ctl_add()
984 zlo->tag_set.cmd_size = sizeof(struct zloop_cmd); in zloop_ctl_add()
985 zlo->tag_set.driver_data = zlo; in zloop_ctl_add()
987 ret = blk_mq_alloc_tag_set(&zlo->tag_set); in zloop_ctl_add()
993 zlo->disk = blk_mq_alloc_disk(&zlo->tag_set, &lim, zlo); in zloop_ctl_add()
[all …]
H A Dxen-blkfront.c226 struct blk_mq_tag_set tag_set; member
1120 memset(&info->tag_set, 0, sizeof(info->tag_set)); in xlvbd_alloc_gendisk()
1121 info->tag_set.ops = &blkfront_mq_ops; in xlvbd_alloc_gendisk()
1122 info->tag_set.nr_hw_queues = info->nr_rings; in xlvbd_alloc_gendisk()
1130 info->tag_set.queue_depth = BLK_RING_SIZE(info) / 2; in xlvbd_alloc_gendisk()
1132 info->tag_set.queue_depth = BLK_RING_SIZE(info); in xlvbd_alloc_gendisk()
1133 info->tag_set.numa_node = NUMA_NO_NODE; in xlvbd_alloc_gendisk()
1134 info->tag_set.cmd_size = sizeof(struct blkif_req); in xlvbd_alloc_gendisk()
1135 info->tag_set.driver_data = info; in xlvbd_alloc_gendisk()
1137 err = blk_mq_alloc_tag_set(&info->tag_set); in xlvbd_alloc_gendisk()
[all …]
H A Damiflop.c205 struct blk_mq_tag_set tag_set; member
1786 disk = blk_mq_alloc_disk(&unit[drive].tag_set, &lim, NULL); in fd_alloc_disk()
1816 memset(&unit[drive].tag_set, 0, sizeof(unit[drive].tag_set)); in fd_alloc_drive()
1817 unit[drive].tag_set.ops = &amiflop_mq_ops; in fd_alloc_drive()
1818 unit[drive].tag_set.nr_hw_queues = 1; in fd_alloc_drive()
1819 unit[drive].tag_set.nr_maps = 1; in fd_alloc_drive()
1820 unit[drive].tag_set.queue_depth = 2; in fd_alloc_drive()
1821 unit[drive].tag_set.numa_node = NUMA_NO_NODE; in fd_alloc_drive()
1822 if (blk_mq_alloc_tag_set(&unit[drive].tag_set)) in fd_alloc_drive()
1832 blk_mq_free_tag_set(&unit[drive].tag_set); in fd_alloc_drive()
H A Dloop.c73 struct blk_mq_tag_set tag_set; member
2031 lo->tag_set.ops = &loop_mq_ops; in loop_add()
2032 lo->tag_set.nr_hw_queues = 1; in loop_add()
2033 lo->tag_set.queue_depth = hw_queue_depth; in loop_add()
2034 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add()
2035 lo->tag_set.cmd_size = sizeof(struct loop_cmd); in loop_add()
2036 lo->tag_set.flags = BLK_MQ_F_STACKING | BLK_MQ_F_NO_SCHED_BY_DEFAULT; in loop_add()
2037 lo->tag_set.driver_data = lo; in loop_add()
2039 err = blk_mq_alloc_tag_set(&lo->tag_set); in loop_add()
2043 disk = lo->lo_disk = blk_mq_alloc_disk(&lo->tag_set, &lim, lo); in loop_add()
[all …]
H A Dswim.c191 struct blk_mq_tag_set tag_set; member
785 blk_mq_free_tag_set(&fs->tag_set); in swim_cleanup_floppy_disk()
820 err = blk_mq_alloc_sq_tag_set(&swd->unit[drive].tag_set, in swim_floppy_init()
826 blk_mq_alloc_disk(&swd->unit[drive].tag_set, &lim, in swim_floppy_init()
829 blk_mq_free_tag_set(&swd->unit[drive].tag_set); in swim_floppy_init()
/linux/drivers/block/rnbd/
H A Drnbd-clt.c737 if (sess->tag_set.tags) in destroy_mq_tags()
738 blk_mq_free_tag_set(&sess->tag_set); in destroy_mq_tags()
1206 struct blk_mq_tag_set *tag_set = &sess->tag_set; in setup_mq_tags() local
1208 memset(tag_set, 0, sizeof(*tag_set)); in setup_mq_tags()
1209 tag_set->ops = &rnbd_mq_ops; in setup_mq_tags()
1210 tag_set->queue_depth = sess->queue_depth; in setup_mq_tags()
1211 tag_set->numa_node = NUMA_NO_NODE; in setup_mq_tags()
1212 tag_set->flags = BLK_MQ_F_TAG_QUEUE_SHARED; in setup_mq_tags()
1213 tag_set->cmd_size = sizeof(struct rnbd_iu) + RNBD_RDMA_SGL_SIZE; in setup_mq_tags()
1216 tag_set->nr_maps = sess->nr_poll_queues ? HCTX_MAX_TYPES : 2; in setup_mq_tags()
[all …]
/linux/drivers/cdrom/
H A Dgdrom.c104 struct blk_mq_tag_set tag_set; member
779 err = blk_mq_alloc_sq_tag_set(&gd.tag_set, &gdrom_mq_ops, 1, in probe_gdrom()
784 gd.disk = blk_mq_alloc_disk(&gd.tag_set, &lim, NULL); in probe_gdrom()
825 blk_mq_free_tag_set(&gd.tag_set); in probe_gdrom()
837 blk_mq_free_tag_set(&gd.tag_set); in remove_gdrom()
/linux/drivers/s390/block/
H A Dscm_blk.h20 struct blk_mq_tag_set tag_set; member
/linux/include/linux/mtd/
H A Dblktrans.h34 struct blk_mq_tag_set *tag_set; member

123