Lines Matching refs:sectors
348 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos()
532 r10_bio->sectors) && in raid10_end_write_request()
729 int sectors = r10_bio->sectors; in read_balance() local
749 if (raid1_should_read_first(conf->mddev, this_sector, sectors)) in read_balance()
764 r10_bio->devs[slot].addr + sectors > in read_balance()
771 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
775 if (is_badblock(rdev, dev_sector, sectors, in read_balance()
786 if (!do_balance && sectors > bad_sectors) in read_balance()
787 sectors = bad_sectors; in read_balance()
788 if (best_good_sectors > sectors) in read_balance()
789 best_good_sectors = sectors; in read_balance()
804 best_good_sectors = sectors; in read_balance()
1123 struct bio *bio, sector_t sectors) in regular_request_wait() argument
1132 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1142 sectors); in regular_request_wait()
1186 if (!regular_request_wait(mddev, conf, bio, r10_bio->sectors)) { in raid10_read_request()
1217 r10_bio->sectors = max_sectors; in raid10_read_request()
1308 r10_bio->sectors && in wait_blocked_dev()
1310 r10_bio->sectors) < 0) in wait_blocked_dev()
1349 sector_t sectors; in raid10_write_request() local
1373 sectors = r10_bio->sectors; in raid10_write_request()
1374 if (!regular_request_wait(mddev, conf, bio, sectors)) { in raid10_write_request()
1382 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
1383 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe && in raid10_write_request()
1418 max_sectors = r10_bio->sectors; in raid10_write_request()
1482 if (max_sectors < r10_bio->sectors) in raid10_write_request()
1483 r10_bio->sectors = max_sectors; in raid10_write_request()
1485 if (r10_bio->sectors < bio_sectors(bio)) { in raid10_write_request()
1487 bio = bio_submit_split_bioset(bio, r10_bio->sectors, in raid10_write_request()
1529 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors) in __make_request() argument
1537 r10_bio->sectors = sectors; in __make_request()
1729 r10_bio->sectors = 0; in raid10_handle_discard()
1870 int sectors = bio_sectors(bio); in raid10_make_request() local
1887 sectors > chunk_sects in raid10_make_request()
1891 sectors = chunk_sects - in raid10_make_request()
1894 __make_request(mddev, bio, sectors); in raid10_make_request()
2229 atomic_add(r10_bio->sectors, in __end_sync_read()
2269 sector_t s = r10_bio->sectors; in end_sync_request()
2316 r10_bio->sectors)) { in end_sync_write()
2361 fbio->bi_iter.bi_size = r10_bio->sectors << 9; in sync_request_write()
2365 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9); in sync_request_write()
2387 int sectors = r10_bio->sectors; in sync_request_write() local
2390 if (sectors < (len / 512)) in sync_request_write()
2391 len = sectors * 512; in sync_request_write()
2396 sectors -= len/512; in sync_request_write()
2400 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches); in sync_request_write()
2451 md_done_sync(mddev, r10_bio->sectors); in sync_request_write()
2479 int sectors = r10_bio->sectors; in fix_recovery_read_error() local
2485 while (sectors) { in fix_recovery_read_error()
2486 int s = sectors; in fix_recovery_read_error()
2541 sectors -= s; in fix_recovery_read_error()
2586 int sectors, struct page *page, enum req_op op) in r10_sync_page_io() argument
2588 if (rdev_has_badblock(rdev, sector, sectors) && in r10_sync_page_io()
2591 if (sync_page_io(rdev, sector, sectors << 9, page, op, false)) in r10_sync_page_io()
2601 rdev_set_badblocks(rdev, sector, sectors, 0); in r10_sync_page_io()
2616 int sectors = r10_bio->sectors, slot = r10_bio->read_slot; in fix_read_error() local
2635 while(sectors) { in fix_read_error()
2636 int s = sectors; in fix_read_error()
2763 sectors -= s; in fix_read_error()
2787 int sectors; in narrow_write_error() local
2788 int sect_to_write = r10_bio->sectors; in narrow_write_error()
2796 sectors = ((r10_bio->sector + block_sectors) in narrow_write_error()
2803 if (sectors > sect_to_write) in narrow_write_error()
2804 sectors = sect_to_write; in narrow_write_error()
2808 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors); in narrow_write_error()
2815 !rdev_set_badblocks(rdev, wsector, sectors, 0)) { in narrow_write_error()
2825 sect_to_write -= sectors; in narrow_write_error()
2826 sector += sectors; in narrow_write_error()
2827 sectors = block_sectors; in narrow_write_error()
2892 r10_bio->sectors, 0); in handle_write_completed()
2896 r10_bio->sectors, 0); in handle_write_completed()
2906 r10_bio->sectors, 0); in handle_write_completed()
2910 r10_bio->sectors, 0); in handle_write_completed()
2923 r10_bio->sectors, 0); in handle_write_completed()
2936 r10_bio->sectors, 0); in handle_write_completed()
3543 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
3647 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3701 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3713 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid10_size() argument
3721 if (!sectors) in raid10_size()
3722 sectors = conf->dev_sectors; in raid10_size()
3724 size = sectors >> conf->geo.chunk_shift; in raid10_size()
4137 static int raid10_resize(struct mddev *mddev, sector_t sectors) in raid10_resize() argument
4161 size = raid10_size(mddev, sectors, 0); in raid10_resize()
4174 if (sectors > mddev->dev_sectors && in raid10_resize()
4179 calc_sectors(conf, sectors); in raid10_resize()
4213 rdev->sectors = size; in raid10_takeover_raid0()
4711 r10_bio->sectors = last - sector_nr + 1; in reshape_request()
4808 r10_bio->sectors = nr_sectors; in reshape_request()
4847 md_done_sync(mddev, r10_bio->sectors); in reshape_request_write()
4912 int sectors = r10_bio->sectors; in handle_reshape_read_error() local
4931 while (sectors) { in handle_reshape_read_error()
4932 int s = sectors; in handle_reshape_read_error()
4972 sectors -= s; in handle_reshape_read_error()
5006 md_done_sync(r10_bio->mddev, r10_bio->sectors); in end_reshape_request()