Lines Matching refs:max_sectors

591 			     int *max_sectors)  in choose_first_rdev()  argument
612 *max_sectors = read_len; in choose_first_rdev()
627 int *max_sectors) in choose_bb_rdev() argument
658 *max_sectors = best_len; in choose_bb_rdev()
666 int *max_sectors) in choose_slow_rdev() argument
691 *max_sectors = read_len; in choose_slow_rdev()
707 *max_sectors = bb_read_len; in choose_slow_rdev()
864 int *max_sectors) in read_balance() argument
872 return choose_first_rdev(conf, r1_bio, max_sectors); in read_balance()
876 *max_sectors = r1_bio->sectors; in read_balance()
887 disk = choose_bb_rdev(conf, r1_bio, max_sectors); in read_balance()
891 return choose_slow_rdev(conf, r1_bio, max_sectors); in read_balance()
1319 int max_sectors; in raid1_read_request() local
1350 rdisk = read_balance(conf, r1_bio, &max_sectors); in raid1_read_request()
1378 if (max_sectors < bio_sectors(bio)) { in raid1_read_request()
1379 struct bio *split = bio_split(bio, max_sectors, in raid1_read_request()
1390 r1_bio->sectors = max_sectors; in raid1_read_request()
1463 int max_sectors; in raid1_write_request() local
1519 max_sectors = r1_bio->sectors; in raid1_write_request()
1541 is_bad = is_badblock(rdev, r1_bio->sector, max_sectors, in raid1_write_request()
1546 if (bad_sectors < max_sectors) in raid1_write_request()
1550 max_sectors = bad_sectors; in raid1_write_request()
1570 if (good_sectors < max_sectors) in raid1_write_request()
1571 max_sectors = good_sectors; in raid1_write_request()
1584 max_sectors = min_t(int, max_sectors, in raid1_write_request()
1586 if (max_sectors < bio_sectors(bio)) { in raid1_write_request()
1587 struct bio *split = bio_split(bio, max_sectors, in raid1_write_request()
1598 r1_bio->sectors = max_sectors; in raid1_write_request()