Lines Matching refs:sector_nr

46 static void allow_barrier(struct r1conf *conf, sector_t sector_nr);
47 static void lower_barrier(struct r1conf *conf, sector_t sector_nr);
869 static int raise_barrier(struct r1conf *conf, sector_t sector_nr) in raise_barrier() argument
871 int idx = sector_to_idx(sector_nr); in raise_barrier()
919 static void lower_barrier(struct r1conf *conf, sector_t sector_nr) in lower_barrier() argument
921 int idx = sector_to_idx(sector_nr); in lower_barrier()
999 static bool wait_read_barrier(struct r1conf *conf, sector_t sector_nr, bool nowait) in wait_read_barrier() argument
1001 int idx = sector_to_idx(sector_nr); in wait_read_barrier()
1042 static bool wait_barrier(struct r1conf *conf, sector_t sector_nr, bool nowait) in wait_barrier() argument
1044 int idx = sector_to_idx(sector_nr); in wait_barrier()
1055 static void allow_barrier(struct r1conf *conf, sector_t sector_nr) in allow_barrier() argument
1057 int idx = sector_to_idx(sector_nr); in allow_barrier()
2661 static sector_t raid1_sync_request(struct mddev *mddev, sector_t sector_nr, in raid1_sync_request() argument
2676 int idx = sector_to_idx(sector_nr); in raid1_sync_request()
2684 if (sector_nr >= max_sector) { in raid1_sync_request()
2711 return max_sector - sector_nr; in raid1_sync_request()
2716 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, 1) && in raid1_sync_request()
2734 md_bitmap_cond_end_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2735 mddev_is_clustered(mddev) && (sector_nr + 2 * RESYNC_SECTORS > conf->cluster_sync_high)); in raid1_sync_request()
2738 if (raise_barrier(conf, sector_nr)) in raid1_sync_request()
2754 r1_bio->sector = sector_nr; in raid1_sync_request()
2758 good_sectors = align_to_barrier_unit_end(sector_nr, good_sectors); in raid1_sync_request()
2778 if (is_badblock(rdev, sector_nr, good_sectors, in raid1_sync_request()
2780 if (first_bad > sector_nr) in raid1_sync_request()
2781 good_sectors = first_bad - sector_nr; in raid1_sync_request()
2783 bad_sectors -= (sector_nr - first_bad); in raid1_sync_request()
2789 if (sector_nr < first_bad) { in raid1_sync_request()
2816 bio->bi_iter.bi_sector = sector_nr + rdev->data_offset; in raid1_sync_request()
2835 ok = rdev_set_badblocks(rdev, sector_nr, in raid1_sync_request()
2872 max_sector = sector_nr + min_bad; in raid1_sync_request()
2873 rv = max_sector - sector_nr; in raid1_sync_request()
2881 if (max_sector > sector_nr + good_sectors) in raid1_sync_request()
2882 max_sector = sector_nr + good_sectors; in raid1_sync_request()
2888 if (sector_nr + (len>>9) > max_sector) in raid1_sync_request()
2889 len = (max_sector - sector_nr) << 9; in raid1_sync_request()
2893 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2918 sector_nr += len>>9; in raid1_sync_request()
2925 conf->cluster_sync_high < sector_nr + nr_sectors) { in raid1_sync_request()