Lines Matching refs:sectors
346 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos()
432 r10_bio->sectors, in close_write()
538 r10_bio->sectors, in raid10_end_write_request()
735 int sectors = r10_bio->sectors; in read_balance() local
762 && (this_sector + sectors >= conf->next_resync)) || in read_balance()
765 this_sector + sectors))) in read_balance()
780 r10_bio->devs[slot].addr + sectors > in read_balance()
794 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
798 if (is_badblock(rdev, dev_sector, sectors, in read_balance()
809 if (!do_balance && sectors > bad_sectors) in read_balance()
810 sectors = bad_sectors; in read_balance()
811 if (best_good_sectors > sectors) in read_balance()
812 best_good_sectors = sectors; in read_balance()
827 best_good_sectors = sectors; in read_balance()
1147 struct bio *bio, sector_t sectors) in regular_request_wait() argument
1156 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1166 sectors); in regular_request_wait()
1214 if (!regular_request_wait(mddev, conf, bio, r10_bio->sectors)) in raid10_read_request()
1240 r10_bio->sectors = max_sectors; in raid10_read_request()
1377 if (!r10_bio->sectors) in wait_blocked_dev()
1380 is_bad = is_badblock(rdev, dev_sector, r10_bio->sectors, in wait_blocked_dev()
1412 sector_t sectors; in raid10_write_request() local
1436 sectors = r10_bio->sectors; in raid10_write_request()
1437 if (!regular_request_wait(mddev, conf, bio, sectors)) in raid10_write_request()
1442 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
1443 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe && in raid10_write_request()
1478 max_sectors = r10_bio->sectors; in raid10_write_request()
1540 if (max_sectors < r10_bio->sectors) in raid10_write_request()
1541 r10_bio->sectors = max_sectors; in raid10_write_request()
1543 if (r10_bio->sectors < bio_sectors(bio)) { in raid10_write_request()
1544 struct bio *split = bio_split(bio, r10_bio->sectors, in raid10_write_request()
1557 md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0); in raid10_write_request()
1568 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors) in __make_request() argument
1576 r10_bio->sectors = sectors; in __make_request()
1764 r10_bio->sectors = 0; in raid10_handle_discard()
1905 int sectors = bio_sectors(bio); in raid10_make_request() local
1923 sectors > chunk_sects in raid10_make_request()
1927 sectors = chunk_sects - in raid10_make_request()
1930 __make_request(mddev, bio, sectors); in raid10_make_request()
2288 atomic_add(r10_bio->sectors, in __end_sync_read()
2328 sector_t s = r10_bio->sectors; in end_sync_request()
2378 r10_bio->sectors, in end_sync_write()
2423 fbio->bi_iter.bi_size = r10_bio->sectors << 9; in sync_request_write()
2427 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9); in sync_request_write()
2449 int sectors = r10_bio->sectors; in sync_request_write() local
2452 if (sectors < (len / 512)) in sync_request_write()
2453 len = sectors * 512; in sync_request_write()
2458 sectors -= len/512; in sync_request_write()
2462 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches); in sync_request_write()
2519 md_done_sync(mddev, r10_bio->sectors, 1); in sync_request_write()
2547 int sectors = r10_bio->sectors; in fix_recovery_read_error() local
2553 while (sectors) { in fix_recovery_read_error()
2554 int s = sectors; in fix_recovery_read_error()
2611 sectors -= s; in fix_recovery_read_error()
2695 int sectors, struct page *page, enum req_op op) in r10_sync_page_io() argument
2700 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) in r10_sync_page_io()
2703 if (sync_page_io(rdev, sector, sectors << 9, page, op, false)) in r10_sync_page_io()
2713 if (!rdev_set_badblocks(rdev, sector, sectors, 0)) in r10_sync_page_io()
2729 int sectors = r10_bio->sectors, slot = r10_bio->read_slot; in fix_read_error() local
2757 while(sectors) { in fix_read_error()
2758 int s = sectors; in fix_read_error()
2898 sectors -= s; in fix_read_error()
2922 int sectors; in narrow_write_error() local
2923 int sect_to_write = r10_bio->sectors; in narrow_write_error()
2932 sectors = ((r10_bio->sector + block_sectors) in narrow_write_error()
2939 if (sectors > sect_to_write) in narrow_write_error()
2940 sectors = sect_to_write; in narrow_write_error()
2944 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors); in narrow_write_error()
2953 sectors, 0) in narrow_write_error()
2957 sect_to_write -= sectors; in narrow_write_error()
2958 sector += sectors; in narrow_write_error()
2959 sectors = block_sectors; in narrow_write_error()
3025 r10_bio->sectors, 0); in handle_write_completed()
3030 r10_bio->sectors, 0)) in handle_write_completed()
3042 r10_bio->sectors, 0); in handle_write_completed()
3047 r10_bio->sectors, 0)) in handle_write_completed()
3062 r10_bio->sectors, 0); in handle_write_completed()
3079 r10_bio->sectors, 0); in handle_write_completed()
3737 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
3845 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3899 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3930 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid10_size() argument
3938 if (!sectors) in raid10_size()
3939 sectors = conf->dev_sectors; in raid10_size()
3941 size = sectors >> conf->geo.chunk_shift; in raid10_size()
4345 static int raid10_resize(struct mddev *mddev, sector_t sectors) in raid10_resize() argument
4369 size = raid10_size(mddev, sectors, 0); in raid10_resize()
4379 if (sectors > mddev->dev_sectors && in raid10_resize()
4384 calc_sectors(conf, sectors); in raid10_resize()
4418 rdev->sectors = size; in raid10_takeover_raid0()
4925 r10_bio->sectors = last - sector_nr + 1; in reshape_request()
5024 r10_bio->sectors = nr_sectors; in reshape_request()
5027 md_sync_acct_bio(read_bio, r10_bio->sectors); in reshape_request()
5064 md_done_sync(mddev, r10_bio->sectors, 0); in reshape_request_write()
5090 md_sync_acct_bio(b, r10_bio->sectors); in reshape_request_write()
5133 int sectors = r10_bio->sectors; in handle_reshape_read_error() local
5152 while (sectors) { in handle_reshape_read_error()
5153 int s = sectors; in handle_reshape_read_error()
5197 sectors -= s; in handle_reshape_read_error()
5235 md_done_sync(r10_bio->mddev, r10_bio->sectors, 1); in end_reshape_request()