Lines Matching refs:sectors
318 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos()
406 r10_bio->sectors, in close_write()
512 r10_bio->sectors, in raid10_end_write_request()
709 int sectors = r10_bio->sectors; in read_balance() local
736 && (this_sector + sectors >= conf->next_resync)) || in read_balance()
739 this_sector + sectors))) in read_balance()
754 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
760 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
764 if (is_badblock(rdev, dev_sector, sectors, in read_balance()
775 if (!do_balance && sectors > bad_sectors) in read_balance()
776 sectors = bad_sectors; in read_balance()
777 if (best_good_sectors > sectors) in read_balance()
778 best_good_sectors = sectors; in read_balance()
793 best_good_sectors = sectors; in read_balance()
1118 struct bio *bio, sector_t sectors) in regular_request_wait() argument
1123 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1129 sectors); in regular_request_wait()
1176 regular_request_wait(mddev, conf, bio, r10_bio->sectors); in raid10_read_request()
1201 r10_bio->sectors = max_sectors; in raid10_read_request()
1302 sector_t sectors; in raid10_write_request() local
1321 sectors = r10_bio->sectors; in raid10_write_request()
1322 regular_request_wait(mddev, conf, bio, sectors); in raid10_write_request()
1326 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
1327 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe && in raid10_write_request()
1362 max_sectors = r10_bio->sectors; in raid10_write_request()
1474 if (max_sectors < r10_bio->sectors) in raid10_write_request()
1475 r10_bio->sectors = max_sectors; in raid10_write_request()
1477 if (r10_bio->sectors < bio_sectors(bio)) { in raid10_write_request()
1478 struct bio *split = bio_split(bio, r10_bio->sectors, in raid10_write_request()
1489 md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0); in raid10_write_request()
1500 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors) in __make_request() argument
1508 r10_bio->sectors = sectors; in __make_request()
1526 int sectors = bio_sectors(bio); in raid10_make_request() local
1541 sectors > chunk_sects in raid10_make_request()
1545 sectors = chunk_sects - in raid10_make_request()
1548 __make_request(mddev, bio, sectors); in raid10_make_request()
1894 atomic_add(r10_bio->sectors, in __end_sync_read()
1934 sector_t s = r10_bio->sectors; in end_sync_request()
1984 r10_bio->sectors, in end_sync_write()
2029 fbio->bi_iter.bi_size = r10_bio->sectors << 9; in sync_request_write()
2033 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9); in sync_request_write()
2055 int sectors = r10_bio->sectors; in sync_request_write() local
2058 if (sectors < (len / 512)) in sync_request_write()
2059 len = sectors * 512; in sync_request_write()
2064 sectors -= len/512; in sync_request_write()
2068 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches); in sync_request_write()
2127 md_done_sync(mddev, r10_bio->sectors, 1); in sync_request_write()
2155 int sectors = r10_bio->sectors; in fix_recovery_read_error() local
2161 while (sectors) { in fix_recovery_read_error()
2162 int s = sectors; in fix_recovery_read_error()
2219 sectors -= s; in fix_recovery_read_error()
2300 int sectors, struct page *page, int rw) in r10_sync_page_io() argument
2305 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) in r10_sync_page_io()
2308 if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) in r10_sync_page_io()
2318 if (!rdev_set_badblocks(rdev, sector, sectors, 0)) in r10_sync_page_io()
2334 int sectors = r10_bio->sectors; in fix_read_error() local
2365 while(sectors) { in fix_read_error()
2366 int s = sectors; in fix_read_error()
2511 sectors -= s; in fix_read_error()
2535 int sectors; in narrow_write_error() local
2536 int sect_to_write = r10_bio->sectors; in narrow_write_error()
2545 sectors = ((r10_bio->sector + block_sectors) in narrow_write_error()
2552 if (sectors > sect_to_write) in narrow_write_error()
2553 sectors = sect_to_write; in narrow_write_error()
2556 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors); in narrow_write_error()
2566 sectors, 0) in narrow_write_error()
2570 sect_to_write -= sectors; in narrow_write_error()
2571 sector += sectors; in narrow_write_error()
2572 sectors = block_sectors; in narrow_write_error()
2634 r10_bio->sectors, 0); in handle_write_completed()
2639 r10_bio->sectors, 0)) in handle_write_completed()
2651 r10_bio->sectors, 0); in handle_write_completed()
2656 r10_bio->sectors, 0)) in handle_write_completed()
2671 r10_bio->sectors, 0); in handle_write_completed()
2688 r10_bio->sectors, 0); in handle_write_completed()
3333 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
3441 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3495 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3526 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid10_size() argument
3534 if (!sectors) in raid10_size()
3535 sectors = conf->dev_sectors; in raid10_size()
3537 size = sectors >> conf->geo.chunk_shift; in raid10_size()
3966 static int raid10_resize(struct mddev *mddev, sector_t sectors) in raid10_resize() argument
3990 size = raid10_size(mddev, sectors, 0); in raid10_resize()
4000 if (sectors > mddev->dev_sectors && in raid10_resize()
4005 calc_sectors(conf, sectors); in raid10_resize()
4039 rdev->sectors = size; in raid10_takeover_raid0()
4547 r10_bio->sectors = last - sector_nr + 1; in reshape_request()
4652 r10_bio->sectors = nr_sectors; in reshape_request()
4655 md_sync_acct_bio(read_bio, r10_bio->sectors); in reshape_request()
4692 md_done_sync(mddev, r10_bio->sectors, 0); in reshape_request_write()
4718 md_sync_acct_bio(b, r10_bio->sectors); in reshape_request_write()
4769 int sectors = r10_bio->sectors; in handle_reshape_read_error() local
4788 while (sectors) { in handle_reshape_read_error()
4789 int s = sectors; in handle_reshape_read_error()
4833 sectors -= s; in handle_reshape_read_error()
4871 md_done_sync(r10_bio->mddev, r10_bio->sectors, 1); in end_reshape_request()