Searched refs:nr_requests (Results 1 – 11 of 11) sorted by relevance
| /linux/block/ |
| H A D | blk-mq-sched.c | 98 max_dispatch = hctx->queue->nr_requests; in __blk_mq_do_dispatch_sched() 505 unsigned int nr_hw_queues, unsigned int nr_requests) in blk_mq_alloc_sched_tags() argument 521 et->nr_requests = nr_requests; in blk_mq_alloc_sched_tags() 533 et->nr_requests); in blk_mq_alloc_sched_tags() 628 q->nr_requests = et->nr_requests; in blk_mq_init_sched() 633 blk_mq_tag_update_sched_shared_tags(q, et->nr_requests); in blk_mq_init_sched()
|
| H A D | blk-sysfs.c | 59 ret = queue_var_show(disk->queue->nr_requests, page); in queue_requests_show() 90 if (nr == q->nr_requests) in queue_requests_store() 110 nr > q->elevator->et->nr_requests) { in queue_requests_store() 164 q->async_depth = min(q->nr_requests, nr); in queue_async_depth_store()
|
| H A D | elevator.h | 30 unsigned int nr_requests; member
|
| H A D | blk-iolatency.c | 333 unsigned long qd = blkiolat->rqos.disk->queue->nr_requests; in scale_cookie_change() 375 unsigned long qd = iolat->blkiolat->rqos.disk->queue->nr_requests; in scale_change()
|
| H A D | elevator.c | 591 q->nr_requests = q->tag_set->queue_depth; in elevator_switch()
|
| H A D | blk-mq.c | 4665 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue() 5024 } else if (nr <= q->elevator->et->nr_requests) { in blk_mq_update_nr_requests() 5044 q->async_depth = max(q->async_depth * nr / q->nr_requests, 1); in blk_mq_update_nr_requests() 5045 q->nr_requests = nr; in blk_mq_update_nr_requests()
|
| H A D | bfq-iosched.c | 701 limit = data->q->nr_requests; in bfq_limit_depth() 725 if (limit < data->q->nr_requests) in bfq_limit_depth() 7358 q->async_depth = (q->nr_requests * 3) >> 2; in bfq_init_queue()
|
| /linux/drivers/s390/block/ |
| H A D | scm_blk.c | 28 static unsigned int nr_requests = 64; variable 31 module_param(nr_requests, uint, S_IRUGO); 32 MODULE_PARM_DESC(nr_requests, "Number of parallel requests."); 460 bdev->tag_set.nr_hw_queues = nr_requests; in scm_blk_dev_setup() 461 bdev->tag_set.queue_depth = nr_requests_per_io * nr_requests; in scm_blk_dev_setup() 545 ret = scm_alloc_rqs(nr_requests); in scm_blk_init()
|
| /linux/drivers/dma/stm32/ |
| H A D | stm32-mdma.c | 255 u32 nr_requests; member 1562 if (config.request >= dmadev->nr_requests) { in stm32_mdma_of_xlate() 1597 u32 nr_channels, nr_requests; in stm32_mdma_probe() local 1613 &nr_requests); in stm32_mdma_probe() 1615 nr_requests = STM32_MDMA_MAX_REQUESTS; in stm32_mdma_probe() 1617 nr_requests); in stm32_mdma_probe() 1632 dmadev->nr_requests = nr_requests; in stm32_mdma_probe()
|
| /linux/drivers/dma/ |
| H A D | owl-dma.c | 1095 int ret, i, nr_channels, nr_requests; in owl_dma_probe() local 1111 ret = of_property_read_u32(np, "dma-requests", &nr_requests); in owl_dma_probe() 1118 nr_channels, nr_requests); in owl_dma_probe() 1123 od->nr_vchans = nr_requests; in owl_dma_probe()
|
| /linux/drivers/target/ |
| H A D | target_core_iblock.c | 139 dev->dev_attrib.hw_queue_depth = q->nr_requests; in iblock_configure_device()
|