Lines Matching refs:sectors
318 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos()
406 r10_bio->sectors, in close_write()
512 r10_bio->sectors, in raid10_end_write_request()
709 int sectors = r10_bio->sectors; in read_balance() local
736 && (this_sector + sectors >= conf->next_resync)) || in read_balance()
739 this_sector + sectors))) in read_balance()
754 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
760 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
764 if (is_badblock(rdev, dev_sector, sectors, in read_balance()
775 if (!do_balance && sectors > bad_sectors) in read_balance()
776 sectors = bad_sectors; in read_balance()
777 if (best_good_sectors > sectors) in read_balance()
778 best_good_sectors = sectors; in read_balance()
793 best_good_sectors = sectors; in read_balance()
1101 struct bio *bio, sector_t sectors) in regular_request_wait() argument
1106 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1112 sectors); in regular_request_wait()
1159 regular_request_wait(mddev, conf, bio, r10_bio->sectors); in raid10_read_request()
1184 r10_bio->sectors = max_sectors; in raid10_read_request()
1285 sector_t sectors; in raid10_write_request() local
1304 sectors = r10_bio->sectors; in raid10_write_request()
1305 regular_request_wait(mddev, conf, bio, sectors); in raid10_write_request()
1309 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
1310 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe && in raid10_write_request()
1345 max_sectors = r10_bio->sectors; in raid10_write_request()
1457 if (max_sectors < r10_bio->sectors) in raid10_write_request()
1458 r10_bio->sectors = max_sectors; in raid10_write_request()
1460 if (r10_bio->sectors < bio_sectors(bio)) { in raid10_write_request()
1461 struct bio *split = bio_split(bio, r10_bio->sectors, in raid10_write_request()
1472 md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0); in raid10_write_request()
1483 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors) in __make_request() argument
1491 r10_bio->sectors = sectors; in __make_request()
1509 int sectors = bio_sectors(bio); in raid10_make_request() local
1523 sectors > chunk_sects in raid10_make_request()
1527 sectors = chunk_sects - in raid10_make_request()
1530 __make_request(mddev, bio, sectors); in raid10_make_request()
1876 atomic_add(r10_bio->sectors, in __end_sync_read()
1916 sector_t s = r10_bio->sectors; in end_sync_request()
1966 r10_bio->sectors, in end_sync_write()
2011 fbio->bi_iter.bi_size = r10_bio->sectors << 9; in sync_request_write()
2015 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9); in sync_request_write()
2037 int sectors = r10_bio->sectors; in sync_request_write() local
2040 if (sectors < (len / 512)) in sync_request_write()
2041 len = sectors * 512; in sync_request_write()
2046 sectors -= len/512; in sync_request_write()
2050 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches); in sync_request_write()
2109 md_done_sync(mddev, r10_bio->sectors, 1); in sync_request_write()
2137 int sectors = r10_bio->sectors; in fix_recovery_read_error() local
2143 while (sectors) { in fix_recovery_read_error()
2144 int s = sectors; in fix_recovery_read_error()
2201 sectors -= s; in fix_recovery_read_error()
2282 int sectors, struct page *page, int rw) in r10_sync_page_io() argument
2287 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) in r10_sync_page_io()
2290 if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) in r10_sync_page_io()
2300 if (!rdev_set_badblocks(rdev, sector, sectors, 0)) in r10_sync_page_io()
2316 int sectors = r10_bio->sectors; in fix_read_error() local
2347 while(sectors) { in fix_read_error()
2348 int s = sectors; in fix_read_error()
2493 sectors -= s; in fix_read_error()
2517 int sectors; in narrow_write_error() local
2518 int sect_to_write = r10_bio->sectors; in narrow_write_error()
2527 sectors = ((r10_bio->sector + block_sectors) in narrow_write_error()
2534 if (sectors > sect_to_write) in narrow_write_error()
2535 sectors = sect_to_write; in narrow_write_error()
2538 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors); in narrow_write_error()
2548 sectors, 0) in narrow_write_error()
2552 sect_to_write -= sectors; in narrow_write_error()
2553 sector += sectors; in narrow_write_error()
2554 sectors = block_sectors; in narrow_write_error()
2616 r10_bio->sectors, 0); in handle_write_completed()
2621 r10_bio->sectors, 0)) in handle_write_completed()
2633 r10_bio->sectors, 0); in handle_write_completed()
2638 r10_bio->sectors, 0)) in handle_write_completed()
2653 r10_bio->sectors, 0); in handle_write_completed()
2670 r10_bio->sectors, 0); in handle_write_completed()
3315 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
3423 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3477 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3508 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid10_size() argument
3516 if (!sectors) in raid10_size()
3517 sectors = conf->dev_sectors; in raid10_size()
3519 size = sectors >> conf->geo.chunk_shift; in raid10_size()
3940 static int raid10_resize(struct mddev *mddev, sector_t sectors) in raid10_resize() argument
3964 size = raid10_size(mddev, sectors, 0); in raid10_resize()
3974 if (sectors > mddev->dev_sectors && in raid10_resize()
3979 calc_sectors(conf, sectors); in raid10_resize()
4013 rdev->sectors = size; in raid10_takeover_raid0()
4521 r10_bio->sectors = last - sector_nr + 1; in reshape_request()
4626 r10_bio->sectors = nr_sectors; in reshape_request()
4629 md_sync_acct_bio(read_bio, r10_bio->sectors); in reshape_request()
4666 md_done_sync(mddev, r10_bio->sectors, 0); in reshape_request_write()
4692 md_sync_acct_bio(b, r10_bio->sectors); in reshape_request_write()
4735 int sectors = r10_bio->sectors; in handle_reshape_read_error() local
4754 while (sectors) { in handle_reshape_read_error()
4755 int s = sectors; in handle_reshape_read_error()
4799 sectors -= s; in handle_reshape_read_error()
4837 md_done_sync(r10_bio->mddev, r10_bio->sectors, 1); in end_reshape_request()