/linux/block/ |
H A D | elevator.c | 63 struct elevator_queue *e = q->elevator; in elv_iosched_allow_bio_merge() 153 struct elevator_queue *e = q->elevator; in elevator_exit() 180 struct elevator_queue *e = q->elevator; in elv_rqhash_add() 196 struct elevator_queue *e = q->elevator; in elv_rqhash_find() 271 struct elevator_queue *e = q->elevator; in elv_merge() 365 struct elevator_queue *e = q->elevator; in elv_merged_request() 379 struct elevator_queue *e = q->elevator; in elv_merge_requests() 390 struct elevator_queue *e = q->elevator; in elv_latter_request() 400 struct elevator_queue *e = q->elevator; in elv_former_request() 457 struct elevator_queue *e = q->elevator; in elv_register_queue() [all …]
|
H A D | mq-deadline.c | 192 struct deadline_data *dd = q->elevator->elevator_data; in dd_request_merged() 212 struct deadline_data *dd = q->elevator->elevator_data; in dd_merged_requests() 464 struct deadline_data *dd = hctx->queue->elevator->elevator_data; in dd_dispatch_request() 510 struct deadline_data *dd = data->q->elevator->elevator_data; in dd_limit_depth() 527 struct deadline_data *dd = q->elevator->elevator_data; in dd_depth_updated() 609 q->elevator = eq; in dd_init_sched() 624 struct deadline_data *dd = q->elevator->elevator_data; in dd_request_merge() 656 struct deadline_data *dd = q->elevator->elevator_data; in dd_bio_merge() 677 struct deadline_data *dd = q->elevator->elevator_data; in dd_insert_request() 729 struct deadline_data *dd = q->elevator->elevator_data; in dd_insert_requests() [all …]
|
H A D | blk-mq-sched.h | 41 struct elevator_queue *e = q->elevator; in blk_mq_sched_allow_merge() 52 struct elevator_queue *e = rq->q->elevator; in blk_mq_sched_completed_request() 63 struct elevator_queue *e = q->elevator; in blk_mq_sched_requeue_request() 72 struct elevator_queue *e = hctx->queue->elevator; in blk_mq_sched_has_work()
|
H A D | blk-mq-sched.c | 90 struct elevator_queue *e = q->elevator; in __blk_mq_do_dispatch_sched() 308 if (hctx->queue->elevator) in __blk_mq_sched_dispatch_requests() 340 struct elevator_queue *e = q->elevator; in blk_mq_sched_bio_merge() 481 eq = q->elevator; in blk_mq_init_sched() 499 q->elevator = NULL; in blk_mq_init_sched() 549 q->elevator = NULL; in blk_mq_exit_sched()
|
H A D | kyber-iosched.c | 425 q->elevator = eq; in kyber_init_sched() 455 struct kyber_queue_data *kqd = hctx->queue->elevator->elevator_data; in kyber_depth_updated() 561 struct kyber_queue_data *kqd = data->q->elevator->elevator_data; in kyber_limit_depth() 616 struct kyber_queue_data *kqd = rq->q->elevator->elevator_data; in kyber_finish_request() 641 struct kyber_queue_data *kqd = rq->q->elevator->elevator_data; in kyber_completed_request() 805 struct kyber_queue_data *kqd = hctx->queue->elevator->elevator_data; in kyber_dispatch_request() 904 struct kyber_queue_data *kqd = q->elevator->elevator_data; \ 963 struct kyber_queue_data *kqd = q->elevator->elevator_data; in KYBER_DEBUGFS_DOMAIN_ATTRS()
|
H A D | blk-flush.c | 231 if (!q->elevator) { in flush_end_io() 312 if (!q->elevator) in blk_kick_flush() 346 if (q->elevator) { in mq_flush_data_end_io()
|
H A D | blk-ioc.c | 48 struct elevator_type *et = icq->q->elevator->type; in ioc_exit_icq() 77 struct elevator_type *et = q->elevator->type; in ioc_destroy_icq() 366 struct elevator_type *et = q->elevator->type; in ioc_create_icq()
|
H A D | blk-mq-debugfs.c | 650 if (q->elevator && !q->sched_debugfs_dir) in blk_mq_debugfs_register() 657 if (q->elevator && !hctx->sched_debugfs_dir) in blk_mq_debugfs_register() 731 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched() 802 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched_hctx()
|
H A D | Makefile | 6 obj-y := bdev.o fops.o bio.o elevator.o blk-core.o blk-sysfs.o \
|
H A D | blk-mq.c | 390 struct elevator_queue *e = data->q->elevator; in blk_mq_rq_ctx_init() 453 if (q->elevator) { in __blk_mq_alloc_requests() 466 struct elevator_mq_ops *ops = &q->elevator->type->ops; in __blk_mq_alloc_requests() 661 if (q->elevator) in blk_mq_alloc_request_hctx() 695 q->elevator->type->ops.finish_request(rq); in blk_mq_finish_request() 2515 } else if (q->elevator) { in blk_mq_insert_request() 2521 q->elevator->type->ops.insert_requests(hctx, &list, flags); in blk_mq_insert_request() 2744 } else if (this_hctx->queue->elevator) { in blk_mq_dispatch_plug_list() 2745 this_hctx->queue->elevator->type->ops.insert_requests(this_hctx, in blk_mq_dispatch_plug_list() 4689 if (q->elevator && q->elevator->type->ops.depth_updated) in blk_mq_update_nr_requests() [all …]
|
H A D | blk-sysfs.c | 768 if (q->elevator) { in blk_register_queue() 783 if (q->elevator) in blk_register_queue() 784 kobject_uevent(&q->elevator->kobj, KOBJ_ADD); in blk_register_queue()
|
H A D | bfq-iosched.c | 438 return bic->icq.q->elevator->elevator_data; in bic_to_bfqd() 692 struct bfq_data *bfqd = data->q->elevator->elevator_data; in bfq_limit_depth() 2448 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_bio_merge() 2488 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_request_merge() 3238 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_allow_bio_merge() 3672 bfq_update_peak_rate(q->elevator->elevator_data, rq); in bfq_dispatch_remove() 5151 struct bfq_data *bfqd = hctx->queue->elevator->elevator_data; in bfq_has_work() 5163 struct bfq_data *bfqd = hctx->queue->elevator->elevator_data; in __bfq_dispatch_request() 5304 struct bfq_data *bfqd = hctx->queue->elevator->elevator_data; in bfq_dispatch_request() 6246 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_insert_request() [all …]
|
H A D | blk-wbt.c | 710 if (q->elevator && in wbt_enable_default() 711 test_bit(ELEVATOR_FLAG_DISABLE_WBT, &q->elevator->flags)) in wbt_enable_default()
|
H A D | blk-mq-tag.c | 109 if (!data->q->elevator && !(data->flags & BLK_MQ_REQ_RESERVED) && in __blk_mq_get_tag()
|
H A D | genhd.c | 552 if (disk->queue->elevator) in device_add_disk() 709 if (q->elevator) { in del_gendisk()
|
H A D | bfq-cgroup.c | 533 struct bfq_data *bfqd = blkg->q->elevator->elevator_data; in bfq_pd_init()
|
/linux/tools/cgroup/ |
H A D | iocost_coef_gen.py | 107 global elevator_path, nomerges_path, elevator, nomerges 111 f.write(elevator) 143 elevator = re.sub(r'.*\[(.*)\].*', r'\1', f.read().strip()) variable
|
/linux/Documentation/filesystems/ |
H A D | zonefs.rst | 148 implemented by the block layer elevator. An elevator implementing the sequential 149 write feature for zoned block device (ELEVATOR_F_ZBD_SEQ_WRITE elevator feature) 150 must be used. This type of elevator (e.g. mq-deadline) is set by default
|
/linux/include/linux/ |
H A D | blkdev.h | 451 struct elevator_queue *elevator; member
|
/linux/Documentation/ABI/stable/ |
H A D | sysfs-block | 528 [RW] Standard I/O elevator operations include attempts to merge
|
/linux/Documentation/admin-guide/LSM/ |
H A D | Smack.rst | 6 "Good for you, you've decided to clean the elevator!"
|
/linux/ |
H A D | CREDITS | 159 D: elevator + block layer rewrites
|