Lines Matching refs:mddev
84 struct mddev *mddev = rdev->mddev; in wait_for_serialization() local
89 if (WARN_ON(!mddev->serial_info_pool)) in wait_for_serialization()
91 si = mempool_alloc(mddev->serial_info_pool, GFP_NOIO); in wait_for_serialization()
101 struct mddev *mddev = rdev->mddev; in remove_serial() local
110 mempool_free(si, mddev->serial_info_pool); in remove_serial()
180 if (test_bit(MD_RECOVERY_REQUESTED, &pi->mddev->recovery)) in r1buf_pool_alloc()
255 struct r1conf *conf = r1_bio->mddev->private; in free_r1bio()
263 struct r1conf *conf = r1_bio->mddev->private; in put_buf()
270 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev); in put_buf()
281 struct mddev *mddev = r1_bio->mddev; in reschedule_retry() local
282 struct r1conf *conf = mddev->private; in reschedule_retry()
292 md_wakeup_thread(mddev->thread); in reschedule_retry()
313 struct r1conf *conf = r1_bio->mddev->private; in raid_end_bio_io()
339 struct r1conf *conf = r1_bio->mddev->private; in update_head_pos()
351 struct r1conf *conf = r1_bio->mddev->private; in find_bio_disk()
368 struct r1conf *conf = r1_bio->mddev->private; in raid1_end_read_request()
390 if (r1_bio->mddev->degraded == conf->raid_disks || in raid1_end_read_request()
391 (r1_bio->mddev->degraded == conf->raid_disks-1 && in raid1_end_read_request()
399 rdev_dec_pending(rdev, conf->mddev); in raid1_end_read_request()
405 mdname(conf->mddev), in raid1_end_read_request()
423 md_bitmap_endwrite(r1_bio->mddev->bitmap, r1_bio->sector, in close_write()
427 md_write_end(r1_bio->mddev); in close_write()
450 struct r1conf *conf = r1_bio->mddev->private; in raid1_end_write_request()
467 conf->mddev->recovery); in raid1_end_write_request()
473 md_error(r1_bio->mddev, rdev); in raid1_end_write_request()
550 } else if (rdev->mddev->serialize_policy) in raid1_end_write_request()
553 rdev_dec_pending(rdev, conf->mddev); in raid1_end_write_request()
630 if ((conf->mddev->recovery_cp < this_sector + sectors) || in read_balance()
631 (mddev_is_clustered(conf->mddev) && in read_balance()
632 md_cluster_ops->area_resyncing(conf->mddev, READ, this_sector, in read_balance()
802 raid1_prepare_flush_writes(conf->mddev->bitmap); in flush_bio_list()
903 test_bit(MD_RECOVERY_INTR, &conf->mddev->recovery), in raise_barrier()
906 if (test_bit(MD_RECOVERY_INTR, &conf->mddev->recovery)) { in raise_barrier()
1102 raid1_log(conf->mddev, "wait freeze"); in freeze_array()
1128 &r1_bio->mddev->bio_set); in alloc_behind_master_bio()
1173 struct mddev *mddev = plug->cb.data; in raid1_unplug() local
1174 struct r1conf *conf = mddev->private; in raid1_unplug()
1182 md_wakeup_thread(mddev->thread); in raid1_unplug()
1193 static void init_r1bio(struct r1bio *r1_bio, struct mddev *mddev, struct bio *bio) in init_r1bio() argument
1198 r1_bio->mddev = mddev; in init_r1bio()
1203 alloc_r1bio(struct mddev *mddev, struct bio *bio) in alloc_r1bio() argument
1205 struct r1conf *conf = mddev->private; in alloc_r1bio()
1211 init_r1bio(r1_bio, mddev, bio); in alloc_r1bio()
1215 static void raid1_read_request(struct mddev *mddev, struct bio *bio, in raid1_read_request() argument
1218 struct r1conf *conf = mddev->private; in raid1_read_request()
1221 struct bitmap *bitmap = mddev->bitmap; in raid1_read_request()
1259 r1_bio = alloc_r1bio(mddev, bio); in raid1_read_request()
1261 init_r1bio(r1_bio, mddev, bio); in raid1_read_request()
1274 mdname(mddev), in raid1_read_request()
1285 mdname(mddev), in raid1_read_request()
1295 raid1_log(mddev, "wait behind writes"); in raid1_read_request()
1312 md_account_bio(mddev, &bio); in raid1_read_request()
1316 &mddev->bio_set); in raid1_read_request()
1329 if (mddev->gendisk) in raid1_read_request()
1330 trace_block_bio_remap(read_bio, disk_devt(mddev->gendisk), in raid1_read_request()
1336 static void raid1_write_request(struct mddev *mddev, struct bio *bio, in raid1_write_request() argument
1339 struct r1conf *conf = mddev->private; in raid1_write_request()
1342 struct bitmap *bitmap = mddev->bitmap; in raid1_write_request()
1349 if (mddev_is_clustered(mddev) && in raid1_write_request()
1350 md_cluster_ops->area_resyncing(mddev, WRITE, in raid1_write_request()
1361 if (!md_cluster_ops->area_resyncing(mddev, WRITE, in raid1_write_request()
1382 r1_bio = alloc_r1bio(mddev, bio); in raid1_write_request()
1446 rdev_dec_pending(rdev, mddev); in raid1_write_request()
1475 rdev_dec_pending(conf->mirrors[j].rdev, mddev); in raid1_write_request()
1483 raid1_log(mddev, "wait rdev %d blocked", blocked_rdev->raid_disk); in raid1_write_request()
1484 md_wait_for_blocked_rdev(blocked_rdev, mddev); in raid1_write_request()
1508 md_account_bio(mddev, &bio); in raid1_write_request()
1528 < mddev->bitmap_info.max_write_behind) && in raid1_write_request()
1541 GFP_NOIO, &mddev->bio_set); in raid1_write_request()
1548 &mddev->bio_set); in raid1_write_request()
1550 if (mddev->serialize_policy) in raid1_write_request()
1561 conf->raid_disks - mddev->degraded > 1) in raid1_write_request()
1567 if (mddev->gendisk) in raid1_write_request()
1568 trace_block_bio_remap(mbio, disk_devt(mddev->gendisk), in raid1_write_request()
1572 if (!raid1_add_bio_to_plug(mddev, mbio, raid1_unplug, disks)) { in raid1_write_request()
1576 md_wakeup_thread(mddev->thread); in raid1_write_request()
1586 static bool raid1_make_request(struct mddev *mddev, struct bio *bio) in raid1_make_request() argument
1591 && md_flush_request(mddev, bio)) in raid1_make_request()
1605 raid1_read_request(mddev, bio, sectors, NULL); in raid1_make_request()
1607 if (!md_write_start(mddev,bio)) in raid1_make_request()
1609 raid1_write_request(mddev, bio, sectors); in raid1_make_request()
1614 static void raid1_status(struct seq_file *seq, struct mddev *mddev) in raid1_status() argument
1616 struct r1conf *conf = mddev->private; in raid1_status()
1620 conf->raid_disks - mddev->degraded); in raid1_status()
1647 static void raid1_error(struct mddev *mddev, struct md_rdev *rdev) in raid1_error() argument
1649 struct r1conf *conf = mddev->private; in raid1_error()
1655 (conf->raid_disks - mddev->degraded) == 1) { in raid1_error()
1656 set_bit(MD_BROKEN, &mddev->flags); in raid1_error()
1658 if (!mddev->fail_last_dev) { in raid1_error()
1659 conf->recovery_disabled = mddev->recovery_disabled; in raid1_error()
1666 mddev->degraded++; in raid1_error()
1672 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in raid1_error()
1673 set_mask_bits(&mddev->sb_flags, 0, in raid1_error()
1677 mdname(mddev), rdev->bdev, in raid1_error()
1678 mdname(mddev), conf->raid_disks - mddev->degraded); in raid1_error()
1690 pr_debug(" --- wd:%d rd:%d\n", conf->raid_disks - conf->mddev->degraded, in print_conf()
1717 static int raid1_spare_active(struct mddev *mddev) in raid1_spare_active() argument
1720 struct r1conf *conf = mddev->private; in raid1_spare_active()
1762 mddev->degraded -= count; in raid1_spare_active()
1769 static int raid1_add_disk(struct mddev *mddev, struct md_rdev *rdev) in raid1_add_disk() argument
1771 struct r1conf *conf = mddev->private; in raid1_add_disk()
1778 if (mddev->recovery_disabled == conf->recovery_disabled) in raid1_add_disk()
1781 if (md_integrity_add_rdev(rdev, mddev)) in raid1_add_disk()
1800 if (mddev->gendisk) in raid1_add_disk()
1801 disk_stack_limits(mddev->gendisk, rdev->bdev, in raid1_add_disk()
1835 static int raid1_remove_disk(struct mddev *mddev, struct md_rdev *rdev) in raid1_remove_disk() argument
1837 struct r1conf *conf = mddev->private; in raid1_remove_disk()
1859 mddev->recovery_disabled != conf->recovery_disabled && in raid1_remove_disk()
1860 mddev->degraded < conf->raid_disks) { in raid1_remove_disk()
1900 err = md_integrity_register(mddev); in raid1_remove_disk()
1926 static void abort_sync_write(struct mddev *mddev, struct r1bio *r1_bio) in abort_sync_write() argument
1934 md_bitmap_end_sync(mddev->bitmap, s, &sync_blocks, 1); in abort_sync_write()
1943 struct mddev *mddev = r1_bio->mddev; in put_sync_write_buf() local
1951 md_done_sync(mddev, s, uptodate); in put_sync_write_buf()
1960 struct mddev *mddev = r1_bio->mddev; in end_sync_write() local
1961 struct r1conf *conf = mddev->private; in end_sync_write()
1967 abort_sync_write(mddev, r1_bio); in end_sync_write()
1971 mddev->recovery); in end_sync_write()
1996 rdev->mddev->recovery); in r1_sync_page_io()
2000 md_error(rdev->mddev, rdev); in r1_sync_page_io()
2017 struct mddev *mddev = r1_bio->mddev; in fix_sync_read_error() local
2018 struct r1conf *conf = mddev->private; in fix_sync_read_error()
2030 md_error(mddev, rdev); in fix_sync_read_error()
2073 mdname(mddev), bio->bi_bdev, in fix_sync_read_error()
2084 mddev->recovery_disabled; in fix_sync_read_error()
2085 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in fix_sync_read_error()
2086 md_done_sync(mddev, r1_bio->sectors, 0); in fix_sync_read_error()
2110 rdev_dec_pending(rdev, mddev); in fix_sync_read_error()
2144 struct mddev *mddev = r1_bio->mddev; in process_checks() local
2145 struct r1conf *conf = mddev->private; in process_checks()
2175 rdev_dec_pending(conf->mirrors[primary].rdev, mddev); in process_checks()
2208 atomic64_add(r1_bio->sectors, &mddev->resync_mismatches); in process_checks()
2209 if (j < 0 || (test_bit(MD_RECOVERY_CHECK, &mddev->recovery) in process_checks()
2213 rdev_dec_pending(conf->mirrors[i].rdev, mddev); in process_checks()
2221 static void sync_request_write(struct mddev *mddev, struct r1bio *r1_bio) in sync_request_write() argument
2223 struct r1conf *conf = mddev->private; in sync_request_write()
2233 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) in sync_request_write()
2245 !test_bit(MD_RECOVERY_SYNC, &mddev->recovery)))) in sync_request_write()
2248 abort_sync_write(mddev, r1_bio); in sync_request_write()
2277 struct mddev *mddev = conf->mddev; in fix_read_error() local
2305 rdev_dec_pending(rdev, mddev); in fix_read_error()
2319 md_error(mddev, rdev); in fix_read_error()
2336 rdev_dec_pending(rdev, mddev); in fix_read_error()
2355 mdname(mddev), s, in fix_read_error()
2360 rdev_dec_pending(rdev, mddev); in fix_read_error()
2371 struct mddev *mddev = r1_bio->mddev; in narrow_write_error() local
2372 struct r1conf *conf = mddev->private; in narrow_write_error()
2411 GFP_NOIO, &mddev->bio_set); in narrow_write_error()
2414 GFP_NOIO, &mddev->bio_set); in narrow_write_error()
2454 md_error(conf->mddev, rdev); in handle_sync_write_finished()
2458 md_done_sync(conf->mddev, s, 1); in handle_sync_write_finished()
2472 rdev_dec_pending(rdev, conf->mddev); in handle_write_finished()
2480 md_error(conf->mddev, in handle_write_finished()
2486 conf->mddev); in handle_write_finished()
2499 md_wakeup_thread(conf->mddev->thread); in handle_write_finished()
2509 struct mddev *mddev = conf->mddev; in handle_read_error() local
2529 if (mddev->ro == 0 in handle_read_error()
2535 } else if (mddev->ro == 0 && test_bit(FailFast, &rdev->flags)) { in handle_read_error()
2536 md_error(mddev, rdev); in handle_read_error()
2541 rdev_dec_pending(rdev, conf->mddev); in handle_read_error()
2547 raid1_read_request(mddev, bio, r1_bio->sectors, r1_bio); in handle_read_error()
2553 struct mddev *mddev = thread->mddev; in raid1d() local
2556 struct r1conf *conf = mddev->private; in raid1d()
2561 md_check_recovery(mddev); in raid1d()
2564 !test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) { in raid1d()
2567 if (!test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) in raid1d()
2576 if (mddev->degraded) in raid1d()
2600 mddev = r1_bio->mddev; in raid1d()
2601 conf = mddev->private; in raid1d()
2607 sync_request_write(mddev, r1_bio); in raid1d()
2617 if (mddev->sb_flags & ~(1<<MD_SB_CHANGE_PENDING)) in raid1d()
2618 md_check_recovery(mddev); in raid1d()
2661 static sector_t raid1_sync_request(struct mddev *mddev, sector_t sector_nr, in raid1_sync_request() argument
2664 struct r1conf *conf = mddev->private; in raid1_sync_request()
2683 max_sector = mddev->dev_sectors; in raid1_sync_request()
2690 if (mddev->curr_resync < max_sector) /* aborted */ in raid1_sync_request()
2691 md_bitmap_end_sync(mddev->bitmap, mddev->curr_resync, in raid1_sync_request()
2696 md_bitmap_close_sync(mddev->bitmap); in raid1_sync_request()
2699 if (mddev_is_clustered(mddev)) { in raid1_sync_request()
2706 if (mddev->bitmap == NULL && in raid1_sync_request()
2707 mddev->recovery_cp == MaxSector && in raid1_sync_request()
2708 !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery) && in raid1_sync_request()
2716 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, 1) && in raid1_sync_request()
2717 !conf->fullsync && !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) { in raid1_sync_request()
2734 md_bitmap_cond_end_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2735 mddev_is_clustered(mddev) && (sector_nr + 2 * RESYNC_SECTORS > conf->cluster_sync_high)); in raid1_sync_request()
2753 r1_bio->mddev = mddev; in raid1_sync_request()
2801 test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && in raid1_sync_request()
2802 !test_bit(MD_RECOVERY_CHECK, &mddev->recovery)) { in raid1_sync_request()
2839 set_bit(MD_SB_CHANGE_DEVS, &mddev->sb_flags); in raid1_sync_request()
2849 conf->recovery_disabled = mddev->recovery_disabled; in raid1_sync_request()
2850 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in raid1_sync_request()
2862 if (test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && read_targets > 0) in raid1_sync_request()
2879 if (max_sector > mddev->resync_max) in raid1_sync_request()
2880 max_sector = mddev->resync_max; /* Don't do IO beyond here */ in raid1_sync_request()
2893 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2896 !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) in raid1_sync_request()
2924 if (mddev_is_clustered(mddev) && in raid1_sync_request()
2926 conf->cluster_sync_low = mddev->curr_resync_completed; in raid1_sync_request()
2929 md_cluster_ops->resync_info_update(mddev, in raid1_sync_request()
2937 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) { in raid1_sync_request()
2960 static sector_t raid1_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid1_size() argument
2965 return mddev->dev_sectors; in raid1_size()
2968 static struct r1conf *setup_conf(struct mddev *mddev) in setup_conf() argument
3001 mddev->raid_disks, 2), in setup_conf()
3013 conf->poolinfo->raid_disks = mddev->raid_disks * 2; in setup_conf()
3023 conf->poolinfo->mddev = mddev; in setup_conf()
3027 rdev_for_each(rdev, mddev) { in setup_conf()
3029 if (disk_idx >= mddev->raid_disks in setup_conf()
3033 disk = conf->mirrors + mddev->raid_disks + disk_idx; in setup_conf()
3043 conf->raid_disks = mddev->raid_disks; in setup_conf()
3044 conf->mddev = mddev; in setup_conf()
3052 conf->recovery_disabled = mddev->recovery_disabled - 1; in setup_conf()
3085 md_register_thread(raid1d, mddev, "raid1")); in setup_conf()
3107 static void raid1_free(struct mddev *mddev, void *priv);
3108 static int raid1_run(struct mddev *mddev) in raid1_run() argument
3115 if (mddev->level != 1) { in raid1_run()
3117 mdname(mddev), mddev->level); in raid1_run()
3120 if (mddev->reshape_position != MaxSector) { in raid1_run()
3122 mdname(mddev)); in raid1_run()
3125 if (mddev_init_writes_pending(mddev) < 0) in raid1_run()
3132 if (mddev->private == NULL) in raid1_run()
3133 conf = setup_conf(mddev); in raid1_run()
3135 conf = mddev->private; in raid1_run()
3140 if (mddev->queue) in raid1_run()
3141 blk_queue_max_write_zeroes_sectors(mddev->queue, 0); in raid1_run()
3143 rdev_for_each(rdev, mddev) { in raid1_run()
3144 if (!mddev->gendisk) in raid1_run()
3146 disk_stack_limits(mddev->gendisk, rdev->bdev, in raid1_run()
3150 mddev->degraded = 0; in raid1_run()
3155 mddev->degraded++; in raid1_run()
3159 if (conf->raid_disks - mddev->degraded < 1) { in raid1_run()
3160 md_unregister_thread(mddev, &conf->thread); in raid1_run()
3165 if (conf->raid_disks - mddev->degraded == 1) in raid1_run()
3166 mddev->recovery_cp = MaxSector; in raid1_run()
3168 if (mddev->recovery_cp != MaxSector) in raid1_run()
3170 mdname(mddev)); in raid1_run()
3172 mdname(mddev), mddev->raid_disks - mddev->degraded, in raid1_run()
3173 mddev->raid_disks); in raid1_run()
3178 rcu_assign_pointer(mddev->thread, conf->thread); in raid1_run()
3180 mddev->private = conf; in raid1_run()
3181 set_bit(MD_FAILFAST_SUPPORTED, &mddev->flags); in raid1_run()
3183 md_set_array_sectors(mddev, raid1_size(mddev, 0, 0)); in raid1_run()
3185 ret = md_integrity_register(mddev); in raid1_run()
3187 md_unregister_thread(mddev, &mddev->thread); in raid1_run()
3193 raid1_free(mddev, conf); in raid1_run()
3197 static void raid1_free(struct mddev *mddev, void *priv) in raid1_free() argument
3213 static int raid1_resize(struct mddev *mddev, sector_t sectors) in raid1_resize() argument
3222 sector_t newsize = raid1_size(mddev, sectors, 0); in raid1_resize()
3223 if (mddev->external_size && in raid1_resize()
3224 mddev->array_sectors > newsize) in raid1_resize()
3226 if (mddev->bitmap) { in raid1_resize()
3227 int ret = md_bitmap_resize(mddev->bitmap, newsize, 0, 0); in raid1_resize()
3231 md_set_array_sectors(mddev, newsize); in raid1_resize()
3232 if (sectors > mddev->dev_sectors && in raid1_resize()
3233 mddev->recovery_cp > mddev->dev_sectors) { in raid1_resize()
3234 mddev->recovery_cp = mddev->dev_sectors; in raid1_resize()
3235 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in raid1_resize()
3237 mddev->dev_sectors = sectors; in raid1_resize()
3238 mddev->resync_max_sectors = sectors; in raid1_resize()
3242 static int raid1_reshape(struct mddev *mddev) in raid1_reshape() argument
3258 struct r1conf *conf = mddev->private; in raid1_reshape()
3268 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape()
3269 mddev->layout != mddev->new_layout || in raid1_reshape()
3270 mddev->level != mddev->new_level) { in raid1_reshape()
3271 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
3272 mddev->new_layout = mddev->layout; in raid1_reshape()
3273 mddev->new_level = mddev->level; in raid1_reshape()
3277 if (!mddev_is_clustered(mddev)) in raid1_reshape()
3278 md_allow_write(mddev); in raid1_reshape()
3280 raid_disks = mddev->raid_disks + mddev->delta_disks; in raid1_reshape()
3294 newpoolinfo->mddev = mddev; in raid1_reshape()
3321 sysfs_unlink_rdev(mddev, rdev); in raid1_reshape()
3323 sysfs_unlink_rdev(mddev, rdev); in raid1_reshape()
3324 if (sysfs_link_rdev(mddev, rdev)) in raid1_reshape()
3326 mdname(mddev), rdev->raid_disk); in raid1_reshape()
3337 mddev->degraded += (raid_disks - conf->raid_disks); in raid1_reshape()
3339 conf->raid_disks = mddev->raid_disks = raid_disks; in raid1_reshape()
3340 mddev->delta_disks = 0; in raid1_reshape()
3344 set_bit(MD_RECOVERY_RECOVER, &mddev->recovery); in raid1_reshape()
3345 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in raid1_reshape()
3346 md_wakeup_thread(mddev->thread); in raid1_reshape()
3352 static void raid1_quiesce(struct mddev *mddev, int quiesce) in raid1_quiesce() argument
3354 struct r1conf *conf = mddev->private; in raid1_quiesce()
3362 static void *raid1_takeover(struct mddev *mddev) in raid1_takeover() argument
3367 if (mddev->level == 5 && mddev->raid_disks == 2) { in raid1_takeover()
3369 mddev->new_level = 1; in raid1_takeover()
3370 mddev->new_layout = 0; in raid1_takeover()
3371 mddev->new_chunk_sectors = 0; in raid1_takeover()
3372 conf = setup_conf(mddev); in raid1_takeover()
3376 mddev_clear_unsupported_flags(mddev, in raid1_takeover()