Home
last modified time | relevance | path

Searched refs:blk_rq_sectors (Results 1 – 25 of 27) sorted by relevance

12

/linux/block/
H A Dblk-merge.c516 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_back_merge_fn()
535 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_front_merge_fn()
551 if (blk_rq_sectors(req) + bio_sectors(next->bio) > in req_attempt_discard_merge()
573 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > in ll_merge_requests_fn()
669 else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next)) in blk_try_req_merge()
839 else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector) in blk_try_merge()
927 if (blk_rq_sectors(req) + bio_sectors(bio) > in bio_attempt_discard_merge()
H A Dbfq-iosched.c235 blk_rq_sectors(rq) < BFQQ_SECT_THR_NONROT))
1050 return blk_rq_sectors(rq); in bfq_serv_to_charge()
1052 return blk_rq_sectors(rq) * bfq_async_charge_factor; in bfq_serv_to_charge()
3437 blk_rq_sectors(rq); in bfq_reset_rate_computation()
3623 bfqd->tot_sectors_dispatched += blk_rq_sectors(rq); in bfq_update_peak_rate()
3628 max_t(u32, blk_rq_sectors(rq), bfqd->last_rq_max_size); in bfq_update_peak_rate()
3630 bfqd->last_rq_max_size = blk_rq_sectors(rq); in bfq_update_peak_rate()
3641 bfqd->last_position = blk_rq_pos(rq) + blk_rq_sectors(rq); in bfq_update_peak_rate()
4114 next_rq ? blk_rq_sectors(next_rq) : 0, in __bfq_bfqq_recalc_budget()
4726 blk_rq_sectors(bfqq->next_rq) >= in bfq_choose_bfqq_for_injection()
[all …]
H A Dblk-flush.c392 if (blk_rq_sectors(rq)) in blk_insert_flush()
H A Delevator.c67 #define rq_hash_key(rq) (blk_rq_pos(rq) + blk_rq_sectors(rq))
H A Dblk-mq.c823 blk_rq_sectors(rq), blk_rq_cur_sectors(rq)); in blk_dump_rq_flags()
1350 rq->stats_sectors = blk_rq_sectors(rq); in blk_mq_start_request()
3247 if (blk_rq_sectors(rq) > max_sectors) { in blk_insert_cloned_request()
3262 __func__, blk_rq_sectors(rq), max_sectors); in blk_insert_cloned_request()
H A Dblk-zoned.c877 sector_t req_back_sector = blk_rq_pos(req) + blk_rq_sectors(req); in blk_zone_write_plug_init_request()
H A Dblk-iocost.c2767 blk_rq_pos(rq) + blk_rq_sectors(rq) == iocg->cursor) in ioc_rqos_merge()
/linux/include/linux/
H A Dblktrace_api.h128 return blk_rq_is_passthrough(rq) ? 0 : blk_rq_sectors(rq); in blk_rq_trace_nr_sectors()
H A Dblk-mq.h1097 static inline unsigned int blk_rq_sectors(const struct request *rq) in blk_rq_sectors() function
/linux/drivers/s390/block/
H A Ddasd_fba.c336 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; in dasd_fba_build_cp_discard()
453 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; in dasd_fba_build_cp_regular()
489 block->bp_block, blk_rq_pos(req), blk_rq_sectors(req)); in dasd_fba_build_cp_regular()
H A Ddasd.c785 sectors = blk_rq_sectors(req); in dasd_profile_end()
/linux/drivers/scsi/
H A Dsd.c914 u32 nr_blocks = sectors_to_logical(sdp, blk_rq_sectors(rq)); in sd_setup_unmap_cmnd()
1004 u32 nr_blocks = sectors_to_logical(sdp, blk_rq_sectors(rq)); in sd_setup_write_same16_cmnd()
1031 u32 nr_blocks = sectors_to_logical(sdp, blk_rq_sectors(rq)); in sd_setup_write_same10_cmnd()
1057 u32 nr_blocks = sectors_to_logical(sdp, blk_rq_sectors(rq)); in sd_setup_write_zeroes_cmnd()
1327 unsigned int nr_blocks = sectors_to_logical(sdp, blk_rq_sectors(rq)); in sd_setup_read_write_cmnd()
1346 if (blk_rq_pos(rq) + blk_rq_sectors(rq) > get_capacity(rq->q->disk)) { in sd_setup_read_write_cmnd()
1351 if ((blk_rq_pos(rq) & mask) || (blk_rq_sectors(rq) & mask)) { in sd_setup_read_write_cmnd()
1418 blk_rq_sectors(rq))); in sd_setup_read_write_cmnd()
1423 blk_rq_sectors(rq))); in sd_setup_read_write_cmnd()
H A Dsr.c376 "Finishing %u sectors\n", blk_rq_sectors(rq))); in sr_init_command()
443 this_count, blk_rq_sectors(rq))); in sr_init_command()
H A Dvirtio_scsi.c547 blk_rq_sectors(rq))); in virtio_scsi_init_hdr_pi()
551 blk_rq_sectors(rq))); in virtio_scsi_init_hdr_pi()
/linux/drivers/block/
H A Dfloppy.c2440 blk_rq_sectors(current_req)); in copy_buffer()
2443 buffer_max > fsector_t + blk_rq_sectors(current_req)) in copy_buffer()
2445 blk_rq_sectors(current_req)); in copy_buffer()
2453 blk_rq_sectors(current_req)); in copy_buffer()
2592 blk_rq_sectors(current_req)); in make_raw_rw_request()
2644 max_size = blk_rq_sectors(current_req); in make_raw_rw_request()
2653 } else if (in_sector_offset || blk_rq_sectors(current_req) < ssize) { in make_raw_rw_request()
2657 sectors = fsector_t + blk_rq_sectors(current_req); in make_raw_rw_request()
2677 (!in_sector_offset && blk_rq_sectors(current_req) >= ssize)) && in make_raw_rw_request()
H A Dzloop.c364 sector_t nr_sectors = blk_rq_sectors(rq); in zloop_rw()
H A Dataflop.c1513 blk_rq_sectors(bd->rq), bd->last)); in ataflop_queue_rq()
H A Dvirtio_blk.c183 range[0].num_sectors = cpu_to_le32(blk_rq_sectors(req)); in virtblk_setup_discard_write_zeroes_erase()
H A Dublk_drv.c502 iod->nr_sectors = blk_rq_sectors(req); in ublk_setup_iod_zoned()
1106 iod->nr_sectors = blk_rq_sectors(req); in ublk_setup_iod()
H A Dxen-blkfront.c563 ring_req->u.discard.nr_sectors = blk_rq_sectors(req); in blkif_queue_discard_req()
/linux/drivers/md/
H A Ddm-rq.c441 tio->n_sectors = blk_rq_sectors(orig); in dm_start_request()
/linux/drivers/mmc/core/
H A Dblock.c1229 nr = blk_rq_sectors(req); in mmc_blk_issue_erase_rq()
1285 nr = blk_rq_sectors(req); in mmc_blk_issue_secdiscard_rq()
1464 brq->data.blocks = blk_rq_sectors(req); in mmc_blk_data_prep()
1489 (blk_rq_pos(req) + blk_rq_sectors(req) == in mmc_blk_data_prep()
1539 if (brq->data.blocks != blk_rq_sectors(req)) { in mmc_blk_data_prep()
/linux/include/trace/events/
H A Dblock.h587 __entry->nr_sector = blk_rq_sectors(rq);
/linux/drivers/cdrom/
H A Dgdrom.c584 block_cnt = blk_rq_sectors(req)/GD_TO_BLK; in gdrom_readdisk_dma()
/linux/drivers/block/mtip32xx/
H A Dmtip32xx.c2056 unsigned int nsect = blk_rq_sectors(rq); in mtip_hw_submit_io()
3257 if (blk_rq_sectors(rq) <= 64) { in mtip_check_unal_depth()
3258 if ((blk_rq_pos(rq) & 7) || (blk_rq_sectors(rq) & 7)) in mtip_check_unal_depth()

12