Lines Matching refs:max_sectors

593 			     int *max_sectors)  in choose_first_rdev()  argument
614 *max_sectors = read_len; in choose_first_rdev()
629 int *max_sectors) in choose_bb_rdev() argument
660 *max_sectors = best_len; in choose_bb_rdev()
668 int *max_sectors) in choose_slow_rdev() argument
693 *max_sectors = read_len; in choose_slow_rdev()
709 *max_sectors = bb_read_len; in choose_slow_rdev()
866 int *max_sectors) in read_balance() argument
874 return choose_first_rdev(conf, r1_bio, max_sectors); in read_balance()
878 *max_sectors = r1_bio->sectors; in read_balance()
889 disk = choose_bb_rdev(conf, r1_bio, max_sectors); in read_balance()
893 return choose_slow_rdev(conf, r1_bio, max_sectors); in read_balance()
1324 int max_sectors; in raid1_read_request() local
1355 rdisk = read_balance(conf, r1_bio, &max_sectors); in raid1_read_request()
1383 if (max_sectors < bio_sectors(bio)) { in raid1_read_request()
1384 struct bio *split = bio_split(bio, max_sectors, in raid1_read_request()
1390 r1_bio->sectors = max_sectors; in raid1_read_request()
1424 int max_sectors; in raid1_write_request() local
1477 max_sectors = r1_bio->sectors; in raid1_write_request()
1507 is_bad = is_badblock(rdev, r1_bio->sector, max_sectors, in raid1_write_request()
1519 if (bad_sectors < max_sectors) in raid1_write_request()
1523 max_sectors = bad_sectors; in raid1_write_request()
1539 if (good_sectors < max_sectors) in raid1_write_request()
1540 max_sectors = good_sectors; in raid1_write_request()
1574 max_sectors = min_t(int, max_sectors, in raid1_write_request()
1576 if (max_sectors < bio_sectors(bio)) { in raid1_write_request()
1577 struct bio *split = bio_split(bio, max_sectors, in raid1_write_request()
1583 r1_bio->sectors = max_sectors; in raid1_write_request()