Lines Matching refs:log

221 	struct r5l_log *log;  member
261 bool r5c_is_writeback(struct r5l_log *log) in r5c_is_writeback() argument
263 return (log != NULL && in r5c_is_writeback()
264 log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK); in r5c_is_writeback()
267 static sector_t r5l_ring_add(struct r5l_log *log, sector_t start, sector_t inc) in r5l_ring_add() argument
270 if (start >= log->device_size) in r5l_ring_add()
271 start = start - log->device_size; in r5l_ring_add()
275 static sector_t r5l_ring_distance(struct r5l_log *log, sector_t start, in r5l_ring_distance() argument
281 return end + log->device_size - start; in r5l_ring_distance()
284 static bool r5l_has_free_space(struct r5l_log *log, sector_t size) in r5l_has_free_space() argument
288 used_size = r5l_ring_distance(log, log->last_checkpoint, in r5l_has_free_space()
289 log->log_start); in r5l_has_free_space()
291 return log->device_size > used_size + size; in r5l_has_free_space()
335 void r5l_wake_reclaim(struct r5l_log *log, sector_t space);
342 if (!r5c_is_writeback(conf->log)) in r5c_check_stripe_cache_usage()
358 r5l_wake_reclaim(conf->log, 0); in r5c_check_stripe_cache_usage()
367 if (!r5c_is_writeback(conf->log)) in r5c_check_cached_full_stripe()
377 r5l_wake_reclaim(conf->log, 0); in r5c_check_cached_full_stripe()
410 struct r5l_log *log = conf->log; in r5c_log_required_to_flush_cache() local
412 if (!r5c_is_writeback(log)) in r5c_log_required_to_flush_cache()
416 ((conf->max_degraded + 1) * atomic_read(&log->stripe_in_journal_count) + in r5c_log_required_to_flush_cache()
427 static inline void r5c_update_log_state(struct r5l_log *log) in r5c_update_log_state() argument
429 struct r5conf *conf = log->rdev->mddev->private; in r5c_update_log_state()
434 if (!r5c_is_writeback(log)) in r5c_update_log_state()
437 free_space = r5l_ring_distance(log, log->log_start, in r5c_update_log_state()
438 log->last_checkpoint); in r5c_update_log_state()
453 r5l_wake_reclaim(log, 0); in r5c_update_log_state()
463 struct r5l_log *log = conf->log; in r5c_make_stripe_write_out() local
465 BUG_ON(!r5c_is_writeback(log)); in r5c_make_stripe_write_out()
505 struct r5l_log *log = sh->raid_conf->log; in r5c_finish_cache_stripe() local
507 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) { in r5c_finish_cache_stripe()
538 static void r5l_log_run_stripes(struct r5l_log *log) in r5l_log_run_stripes() argument
542 lockdep_assert_held(&log->io_list_lock); in r5l_log_run_stripes()
544 list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) { in r5l_log_run_stripes()
549 list_move_tail(&io->log_sibling, &log->finished_ios); in r5l_log_run_stripes()
554 static void r5l_move_to_end_ios(struct r5l_log *log) in r5l_move_to_end_ios() argument
558 lockdep_assert_held(&log->io_list_lock); in r5l_move_to_end_ios()
560 list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) { in r5l_move_to_end_ios()
564 list_move_tail(&io->log_sibling, &log->io_end_ios); in r5l_move_to_end_ios()
573 struct r5l_log *log = io->log; in r5l_log_endio() local
579 md_error(log->rdev->mddev, log->rdev); in r5l_log_endio()
582 mempool_free(io->meta_page, &log->meta_pool); in r5l_log_endio()
584 spin_lock_irqsave(&log->io_list_lock, flags); in r5l_log_endio()
596 if (log->need_cache_flush && !list_empty(&io->stripe_list)) in r5l_log_endio()
597 r5l_move_to_end_ios(log); in r5l_log_endio()
599 r5l_log_run_stripes(log); in r5l_log_endio()
600 if (!list_empty(&log->running_ios)) { in r5l_log_endio()
605 io_deferred = list_first_entry(&log->running_ios, in r5l_log_endio()
608 schedule_work(&log->deferred_io_work); in r5l_log_endio()
611 spin_unlock_irqrestore(&log->io_list_lock, flags); in r5l_log_endio()
613 if (log->need_cache_flush) in r5l_log_endio()
614 md_wakeup_thread(log->rdev->mddev->thread); in r5l_log_endio()
635 static void r5l_do_submit_io(struct r5l_log *log, struct r5l_io_unit *io) in r5l_do_submit_io() argument
639 spin_lock_irqsave(&log->io_list_lock, flags); in r5l_do_submit_io()
641 spin_unlock_irqrestore(&log->io_list_lock, flags); in r5l_do_submit_io()
672 struct r5l_log *log = container_of(work, struct r5l_log, in r5l_submit_io_async() local
677 spin_lock_irqsave(&log->io_list_lock, flags); in r5l_submit_io_async()
678 if (!list_empty(&log->running_ios)) { in r5l_submit_io_async()
679 io = list_first_entry(&log->running_ios, struct r5l_io_unit, in r5l_submit_io_async()
686 spin_unlock_irqrestore(&log->io_list_lock, flags); in r5l_submit_io_async()
688 r5l_do_submit_io(log, io); in r5l_submit_io_async()
693 struct r5l_log *log = container_of(work, struct r5l_log, in r5c_disable_writeback_async() local
695 struct mddev *mddev = log->rdev->mddev; in r5c_disable_writeback_async()
699 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) in r5c_disable_writeback_async()
706 conf->log == NULL || in r5c_disable_writeback_async()
711 log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; in r5c_disable_writeback_async()
717 static void r5l_submit_current_io(struct r5l_log *log) in r5l_submit_current_io() argument
719 struct r5l_io_unit *io = log->current_io; in r5l_submit_current_io()
730 crc = crc32c_le(log->uuid_checksum, block, PAGE_SIZE); in r5l_submit_current_io()
733 log->current_io = NULL; in r5l_submit_current_io()
734 spin_lock_irqsave(&log->io_list_lock, flags); in r5l_submit_current_io()
736 if (io != list_first_entry(&log->running_ios, in r5l_submit_current_io()
742 spin_unlock_irqrestore(&log->io_list_lock, flags); in r5l_submit_current_io()
744 r5l_do_submit_io(log, io); in r5l_submit_current_io()
747 static struct bio *r5l_bio_alloc(struct r5l_log *log) in r5l_bio_alloc() argument
749 struct bio *bio = bio_alloc_bioset(GFP_NOIO, BIO_MAX_PAGES, &log->bs); in r5l_bio_alloc()
752 bio_set_dev(bio, log->rdev->bdev); in r5l_bio_alloc()
753 bio->bi_iter.bi_sector = log->rdev->data_offset + log->log_start; in r5l_bio_alloc()
758 static void r5_reserve_log_entry(struct r5l_log *log, struct r5l_io_unit *io) in r5_reserve_log_entry() argument
760 log->log_start = r5l_ring_add(log, log->log_start, BLOCK_SECTORS); in r5_reserve_log_entry()
762 r5c_update_log_state(log); in r5_reserve_log_entry()
770 if (log->log_start == 0) in r5_reserve_log_entry()
773 io->log_end = log->log_start; in r5_reserve_log_entry()
776 static struct r5l_io_unit *r5l_new_meta(struct r5l_log *log) in r5l_new_meta() argument
781 io = mempool_alloc(&log->io_pool, GFP_ATOMIC); in r5l_new_meta()
786 io->log = log; in r5l_new_meta()
792 io->meta_page = mempool_alloc(&log->meta_pool, GFP_NOIO); in r5l_new_meta()
797 block->seq = cpu_to_le64(log->seq); in r5l_new_meta()
798 block->position = cpu_to_le64(log->log_start); in r5l_new_meta()
800 io->log_start = log->log_start; in r5l_new_meta()
802 io->seq = log->seq++; in r5l_new_meta()
804 io->current_bio = r5l_bio_alloc(log); in r5l_new_meta()
809 r5_reserve_log_entry(log, io); in r5l_new_meta()
811 spin_lock_irq(&log->io_list_lock); in r5l_new_meta()
812 list_add_tail(&io->log_sibling, &log->running_ios); in r5l_new_meta()
813 spin_unlock_irq(&log->io_list_lock); in r5l_new_meta()
818 static int r5l_get_meta(struct r5l_log *log, unsigned int payload_size) in r5l_get_meta() argument
820 if (log->current_io && in r5l_get_meta()
821 log->current_io->meta_offset + payload_size > PAGE_SIZE) in r5l_get_meta()
822 r5l_submit_current_io(log); in r5l_get_meta()
824 if (!log->current_io) { in r5l_get_meta()
825 log->current_io = r5l_new_meta(log); in r5l_get_meta()
826 if (!log->current_io) in r5l_get_meta()
833 static void r5l_append_payload_meta(struct r5l_log *log, u16 type, in r5l_append_payload_meta() argument
838 struct r5l_io_unit *io = log->current_io; in r5l_append_payload_meta()
855 static void r5l_append_payload_page(struct r5l_log *log, struct page *page) in r5l_append_payload_page() argument
857 struct r5l_io_unit *io = log->current_io; in r5l_append_payload_page()
862 io->current_bio = r5l_bio_alloc(log); in r5l_append_payload_page()
870 r5_reserve_log_entry(log, io); in r5l_append_payload_page()
873 static void r5l_append_flush_payload(struct r5l_log *log, sector_t sect) in r5l_append_flush_payload() argument
875 struct mddev *mddev = log->rdev->mddev; in r5l_append_flush_payload()
889 mutex_lock(&log->io_mutex); in r5l_append_flush_payload()
892 if (r5l_get_meta(log, meta_size)) { in r5l_append_flush_payload()
893 mutex_unlock(&log->io_mutex); in r5l_append_flush_payload()
898 io = log->current_io; in r5l_append_flush_payload()
910 mutex_unlock(&log->io_mutex); in r5l_append_flush_payload()
913 static int r5l_log_stripe(struct r5l_log *log, struct stripe_head *sh, in r5l_log_stripe() argument
927 ret = r5l_get_meta(log, meta_size); in r5l_log_stripe()
931 io = log->current_io; in r5l_log_stripe()
943 log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) { in r5l_log_stripe()
951 r5l_append_payload_meta(log, R5LOG_PAYLOAD_DATA, in r5l_log_stripe()
954 r5l_append_payload_page(log, sh->dev[i].page); in r5l_log_stripe()
958 r5l_append_payload_meta(log, R5LOG_PAYLOAD_PARITY, in r5l_log_stripe()
961 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); in r5l_log_stripe()
962 r5l_append_payload_page(log, sh->dev[sh->qd_idx].page); in r5l_log_stripe()
964 r5l_append_payload_meta(log, R5LOG_PAYLOAD_PARITY, in r5l_log_stripe()
967 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); in r5l_log_stripe()
975 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) in r5l_log_stripe()
981 spin_lock_irq(&log->stripe_in_journal_lock); in r5l_log_stripe()
983 &log->stripe_in_journal_list); in r5l_log_stripe()
984 spin_unlock_irq(&log->stripe_in_journal_lock); in r5l_log_stripe()
985 atomic_inc(&log->stripe_in_journal_count); in r5l_log_stripe()
991 static inline void r5l_add_no_space_stripe(struct r5l_log *log, in r5l_add_no_space_stripe() argument
994 spin_lock(&log->no_space_stripes_lock); in r5l_add_no_space_stripe()
995 list_add_tail(&sh->log_list, &log->no_space_stripes); in r5l_add_no_space_stripe()
996 spin_unlock(&log->no_space_stripes_lock); in r5l_add_no_space_stripe()
1003 int r5l_write_stripe(struct r5l_log *log, struct stripe_head *sh) in r5l_write_stripe() argument
1013 if (!log) in r5l_write_stripe()
1037 sh->dev[i].log_checksum = crc32c_le(log->uuid_checksum, in r5l_write_stripe()
1052 mutex_lock(&log->io_mutex); in r5l_write_stripe()
1056 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) { in r5l_write_stripe()
1057 if (!r5l_has_free_space(log, reserve)) { in r5l_write_stripe()
1058 r5l_add_no_space_stripe(log, sh); in r5l_write_stripe()
1061 ret = r5l_log_stripe(log, sh, data_pages, parity_pages); in r5l_write_stripe()
1063 spin_lock_irq(&log->io_list_lock); in r5l_write_stripe()
1065 &log->no_mem_stripes); in r5l_write_stripe()
1066 spin_unlock_irq(&log->io_list_lock); in r5l_write_stripe()
1076 r5l_add_no_space_stripe(log, sh); in r5l_write_stripe()
1079 } else if (!r5l_has_free_space(log, reserve)) { in r5l_write_stripe()
1080 if (sh->log_start == log->last_checkpoint) in r5l_write_stripe()
1083 r5l_add_no_space_stripe(log, sh); in r5l_write_stripe()
1085 ret = r5l_log_stripe(log, sh, data_pages, parity_pages); in r5l_write_stripe()
1087 spin_lock_irq(&log->io_list_lock); in r5l_write_stripe()
1089 &log->no_mem_stripes); in r5l_write_stripe()
1090 spin_unlock_irq(&log->io_list_lock); in r5l_write_stripe()
1095 mutex_unlock(&log->io_mutex); in r5l_write_stripe()
1097 r5l_wake_reclaim(log, reserve); in r5l_write_stripe()
1101 void r5l_write_stripe_run(struct r5l_log *log) in r5l_write_stripe_run() argument
1103 if (!log) in r5l_write_stripe_run()
1105 mutex_lock(&log->io_mutex); in r5l_write_stripe_run()
1106 r5l_submit_current_io(log); in r5l_write_stripe_run()
1107 mutex_unlock(&log->io_mutex); in r5l_write_stripe_run()
1110 int r5l_handle_flush_request(struct r5l_log *log, struct bio *bio) in r5l_handle_flush_request() argument
1112 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) { in r5l_handle_flush_request()
1128 mutex_lock(&log->io_mutex); in r5l_handle_flush_request()
1129 r5l_get_meta(log, 0); in r5l_handle_flush_request()
1130 bio_list_add(&log->current_io->flush_barriers, bio); in r5l_handle_flush_request()
1131 log->current_io->has_flush = 1; in r5l_handle_flush_request()
1132 log->current_io->has_null_flush = 1; in r5l_handle_flush_request()
1133 atomic_inc(&log->current_io->pending_stripe); in r5l_handle_flush_request()
1134 r5l_submit_current_io(log); in r5l_handle_flush_request()
1135 mutex_unlock(&log->io_mutex); in r5l_handle_flush_request()
1143 static void r5l_run_no_space_stripes(struct r5l_log *log) in r5l_run_no_space_stripes() argument
1147 spin_lock(&log->no_space_stripes_lock); in r5l_run_no_space_stripes()
1148 while (!list_empty(&log->no_space_stripes)) { in r5l_run_no_space_stripes()
1149 sh = list_first_entry(&log->no_space_stripes, in r5l_run_no_space_stripes()
1155 spin_unlock(&log->no_space_stripes_lock); in r5l_run_no_space_stripes()
1166 struct r5l_log *log = conf->log; in r5c_calculate_new_cp() local
1170 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) in r5c_calculate_new_cp()
1171 return log->next_checkpoint; in r5c_calculate_new_cp()
1173 spin_lock_irqsave(&log->stripe_in_journal_lock, flags); in r5c_calculate_new_cp()
1174 if (list_empty(&conf->log->stripe_in_journal_list)) { in r5c_calculate_new_cp()
1176 spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags); in r5c_calculate_new_cp()
1177 return log->next_checkpoint; in r5c_calculate_new_cp()
1179 sh = list_first_entry(&conf->log->stripe_in_journal_list, in r5c_calculate_new_cp()
1182 spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags); in r5c_calculate_new_cp()
1186 static sector_t r5l_reclaimable_space(struct r5l_log *log) in r5l_reclaimable_space() argument
1188 struct r5conf *conf = log->rdev->mddev->private; in r5l_reclaimable_space()
1190 return r5l_ring_distance(log, log->last_checkpoint, in r5l_reclaimable_space()
1194 static void r5l_run_no_mem_stripe(struct r5l_log *log) in r5l_run_no_mem_stripe() argument
1198 lockdep_assert_held(&log->io_list_lock); in r5l_run_no_mem_stripe()
1200 if (!list_empty(&log->no_mem_stripes)) { in r5l_run_no_mem_stripe()
1201 sh = list_first_entry(&log->no_mem_stripes, in r5l_run_no_mem_stripe()
1209 static bool r5l_complete_finished_ios(struct r5l_log *log) in r5l_complete_finished_ios() argument
1214 lockdep_assert_held(&log->io_list_lock); in r5l_complete_finished_ios()
1216 list_for_each_entry_safe(io, next, &log->finished_ios, log_sibling) { in r5l_complete_finished_ios()
1221 log->next_checkpoint = io->log_start; in r5l_complete_finished_ios()
1224 mempool_free(io, &log->io_pool); in r5l_complete_finished_ios()
1225 r5l_run_no_mem_stripe(log); in r5l_complete_finished_ios()
1235 struct r5l_log *log = io->log; in __r5l_stripe_write_finished() local
1236 struct r5conf *conf = log->rdev->mddev->private; in __r5l_stripe_write_finished()
1239 spin_lock_irqsave(&log->io_list_lock, flags); in __r5l_stripe_write_finished()
1242 if (!r5l_complete_finished_ios(log)) { in __r5l_stripe_write_finished()
1243 spin_unlock_irqrestore(&log->io_list_lock, flags); in __r5l_stripe_write_finished()
1247 if (r5l_reclaimable_space(log) > log->max_free_space || in __r5l_stripe_write_finished()
1249 r5l_wake_reclaim(log, 0); in __r5l_stripe_write_finished()
1251 spin_unlock_irqrestore(&log->io_list_lock, flags); in __r5l_stripe_write_finished()
1252 wake_up(&log->iounit_wait); in __r5l_stripe_write_finished()
1268 struct r5l_log *log = container_of(bio, struct r5l_log, in r5l_log_flush_endio() local
1274 md_error(log->rdev->mddev, log->rdev); in r5l_log_flush_endio()
1276 spin_lock_irqsave(&log->io_list_lock, flags); in r5l_log_flush_endio()
1277 list_for_each_entry(io, &log->flushing_ios, log_sibling) in r5l_log_flush_endio()
1279 list_splice_tail_init(&log->flushing_ios, &log->finished_ios); in r5l_log_flush_endio()
1280 spin_unlock_irqrestore(&log->io_list_lock, flags); in r5l_log_flush_endio()
1297 void r5l_flush_stripe_to_raid(struct r5l_log *log) in r5l_flush_stripe_to_raid() argument
1301 if (!log || !log->need_cache_flush) in r5l_flush_stripe_to_raid()
1304 spin_lock_irq(&log->io_list_lock); in r5l_flush_stripe_to_raid()
1306 if (!list_empty(&log->flushing_ios)) { in r5l_flush_stripe_to_raid()
1307 spin_unlock_irq(&log->io_list_lock); in r5l_flush_stripe_to_raid()
1310 list_splice_tail_init(&log->io_end_ios, &log->flushing_ios); in r5l_flush_stripe_to_raid()
1311 do_flush = !list_empty(&log->flushing_ios); in r5l_flush_stripe_to_raid()
1312 spin_unlock_irq(&log->io_list_lock); in r5l_flush_stripe_to_raid()
1316 bio_reset(&log->flush_bio); in r5l_flush_stripe_to_raid()
1317 bio_set_dev(&log->flush_bio, log->rdev->bdev); in r5l_flush_stripe_to_raid()
1318 log->flush_bio.bi_end_io = r5l_log_flush_endio; in r5l_flush_stripe_to_raid()
1319 log->flush_bio.bi_opf = REQ_OP_WRITE | REQ_PREFLUSH; in r5l_flush_stripe_to_raid()
1320 submit_bio(&log->flush_bio); in r5l_flush_stripe_to_raid()
1323 static void r5l_write_super(struct r5l_log *log, sector_t cp);
1324 static void r5l_write_super_and_discard_space(struct r5l_log *log, in r5l_write_super_and_discard_space() argument
1327 struct block_device *bdev = log->rdev->bdev; in r5l_write_super_and_discard_space()
1330 r5l_write_super(log, end); in r5l_write_super_and_discard_space()
1335 mddev = log->rdev->mddev; in r5l_write_super_and_discard_space()
1355 if (log->last_checkpoint < end) { in r5l_write_super_and_discard_space()
1357 log->last_checkpoint + log->rdev->data_offset, in r5l_write_super_and_discard_space()
1358 end - log->last_checkpoint, GFP_NOIO, 0); in r5l_write_super_and_discard_space()
1361 log->last_checkpoint + log->rdev->data_offset, in r5l_write_super_and_discard_space()
1362 log->device_size - log->last_checkpoint, in r5l_write_super_and_discard_space()
1364 blkdev_issue_discard(bdev, log->rdev->data_offset, end, in r5l_write_super_and_discard_space()
1414 if (!conf->log) in r5c_flush_cache()
1435 struct r5l_log *log = conf->log; in r5c_do_reclaim() local
1443 if (!r5c_is_writeback(log)) in r5c_do_reclaim()
1479 spin_lock_irqsave(&log->stripe_in_journal_lock, flags); in r5c_do_reclaim()
1481 list_for_each_entry(sh, &log->stripe_in_journal_list, r5c) { in r5c_do_reclaim()
1499 spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags); in r5c_do_reclaim()
1503 r5l_run_no_space_stripes(log); in r5c_do_reclaim()
1508 static void r5l_do_reclaim(struct r5l_log *log) in r5l_do_reclaim() argument
1510 struct r5conf *conf = log->rdev->mddev->private; in r5l_do_reclaim()
1511 sector_t reclaim_target = xchg(&log->reclaim_target, 0); in r5l_do_reclaim()
1516 spin_lock_irq(&log->io_list_lock); in r5l_do_reclaim()
1517 write_super = r5l_reclaimable_space(log) > log->max_free_space || in r5l_do_reclaim()
1518 reclaim_target != 0 || !list_empty(&log->no_space_stripes); in r5l_do_reclaim()
1525 reclaimable = r5l_reclaimable_space(log); in r5l_do_reclaim()
1527 (list_empty(&log->running_ios) && in r5l_do_reclaim()
1528 list_empty(&log->io_end_ios) && in r5l_do_reclaim()
1529 list_empty(&log->flushing_ios) && in r5l_do_reclaim()
1530 list_empty(&log->finished_ios))) in r5l_do_reclaim()
1533 md_wakeup_thread(log->rdev->mddev->thread); in r5l_do_reclaim()
1534 wait_event_lock_irq(log->iounit_wait, in r5l_do_reclaim()
1535 r5l_reclaimable_space(log) > reclaimable, in r5l_do_reclaim()
1536 log->io_list_lock); in r5l_do_reclaim()
1540 spin_unlock_irq(&log->io_list_lock); in r5l_do_reclaim()
1550 r5l_write_super_and_discard_space(log, next_checkpoint); in r5l_do_reclaim()
1552 mutex_lock(&log->io_mutex); in r5l_do_reclaim()
1553 log->last_checkpoint = next_checkpoint; in r5l_do_reclaim()
1554 r5c_update_log_state(log); in r5l_do_reclaim()
1555 mutex_unlock(&log->io_mutex); in r5l_do_reclaim()
1557 r5l_run_no_space_stripes(log); in r5l_do_reclaim()
1564 struct r5l_log *log = conf->log; in r5l_reclaim_thread() local
1566 if (!log) in r5l_reclaim_thread()
1569 r5l_do_reclaim(log); in r5l_reclaim_thread()
1572 void r5l_wake_reclaim(struct r5l_log *log, sector_t space) in r5l_wake_reclaim() argument
1577 if (!log) in r5l_wake_reclaim()
1580 target = log->reclaim_target; in r5l_wake_reclaim()
1583 } while (cmpxchg(&log->reclaim_target, target, new) != target); in r5l_wake_reclaim()
1584 md_wakeup_thread(log->reclaim_thread); in r5l_wake_reclaim()
1587 void r5l_quiesce(struct r5l_log *log, int quiesce) in r5l_quiesce() argument
1593 mddev = log->rdev->mddev; in r5l_quiesce()
1595 kthread_park(log->reclaim_thread->tsk); in r5l_quiesce()
1596 r5l_wake_reclaim(log, MaxSector); in r5l_quiesce()
1597 r5l_do_reclaim(log); in r5l_quiesce()
1599 kthread_unpark(log->reclaim_thread->tsk); in r5l_quiesce()
1604 struct r5l_log *log; in r5l_log_disk_error() local
1608 log = rcu_dereference(conf->log); in r5l_log_disk_error()
1610 if (!log) in r5l_log_disk_error()
1613 ret = test_bit(Faulty, &log->rdev->flags); in r5l_log_disk_error()
1643 static int r5l_recovery_allocate_ra_pool(struct r5l_log *log, in r5l_recovery_allocate_ra_pool() argument
1648 ctx->ra_bio = bio_alloc_bioset(GFP_KERNEL, BIO_MAX_PAGES, &log->bs); in r5l_recovery_allocate_ra_pool()
1672 static void r5l_recovery_free_ra_pool(struct r5l_log *log, in r5l_recovery_free_ra_pool() argument
1688 static int r5l_recovery_fetch_ra_pool(struct r5l_log *log, in r5l_recovery_fetch_ra_pool() argument
1693 bio_set_dev(ctx->ra_bio, log->rdev->bdev); in r5l_recovery_fetch_ra_pool()
1695 ctx->ra_bio->bi_iter.bi_sector = log->rdev->data_offset + offset; in r5l_recovery_fetch_ra_pool()
1705 offset = r5l_ring_add(log, offset, BLOCK_SECTORS); in r5l_recovery_fetch_ra_pool()
1718 static int r5l_recovery_read_page(struct r5l_log *log, in r5l_recovery_read_page() argument
1727 ret = r5l_recovery_fetch_ra_pool(log, ctx, offset); in r5l_recovery_read_page()
1742 static int r5l_recovery_read_meta_block(struct r5l_log *log, in r5l_recovery_read_meta_block() argument
1750 ret = r5l_recovery_read_page(log, ctx, page, ctx->pos); in r5l_recovery_read_meta_block()
1764 crc = crc32c_le(log->uuid_checksum, mb, PAGE_SIZE); in r5l_recovery_read_meta_block()
1777 r5l_recovery_create_empty_meta_block(struct r5l_log *log, in r5l_recovery_create_empty_meta_block() argument
1792 static int r5l_log_write_empty_meta_block(struct r5l_log *log, sector_t pos, in r5l_log_write_empty_meta_block() argument
1801 r5l_recovery_create_empty_meta_block(log, page, pos, seq); in r5l_log_write_empty_meta_block()
1803 mb->checksum = cpu_to_le32(crc32c_le(log->uuid_checksum, in r5l_log_write_empty_meta_block()
1805 if (!sync_page_io(log->rdev, pos, PAGE_SIZE, page, REQ_OP_WRITE, in r5l_log_write_empty_meta_block()
1821 static void r5l_recovery_load_data(struct r5l_log *log, in r5l_recovery_load_data() argument
1827 struct mddev *mddev = log->rdev->mddev; in r5l_recovery_load_data()
1834 r5l_recovery_read_page(log, ctx, sh->dev[dd_idx].page, log_offset); in r5l_recovery_load_data()
1843 static void r5l_recovery_load_parity(struct r5l_log *log, in r5l_recovery_load_parity() argument
1849 struct mddev *mddev = log->rdev->mddev; in r5l_recovery_load_parity()
1853 r5l_recovery_read_page(log, ctx, sh->dev[sh->pd_idx].page, log_offset); in r5l_recovery_load_parity()
1860 log, ctx, sh->dev[sh->qd_idx].page, in r5l_recovery_load_parity()
1861 r5l_ring_add(log, log_offset, BLOCK_SECTORS)); in r5l_recovery_load_parity()
1992 r5l_recovery_verify_data_checksum(struct r5l_log *log, in r5l_recovery_verify_data_checksum() argument
2000 r5l_recovery_read_page(log, ctx, page, log_offset); in r5l_recovery_verify_data_checksum()
2002 checksum = crc32c_le(log->uuid_checksum, addr, PAGE_SIZE); in r5l_recovery_verify_data_checksum()
2012 r5l_recovery_verify_data_checksum_for_mb(struct r5l_log *log, in r5l_recovery_verify_data_checksum_for_mb() argument
2015 struct mddev *mddev = log->rdev->mddev; in r5l_recovery_verify_data_checksum_for_mb()
2019 sector_t log_offset = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); in r5l_recovery_verify_data_checksum_for_mb()
2034 log, ctx, page, log_offset, in r5l_recovery_verify_data_checksum_for_mb()
2039 log, ctx, page, log_offset, in r5l_recovery_verify_data_checksum_for_mb()
2044 log, ctx, page, in r5l_recovery_verify_data_checksum_for_mb()
2045 r5l_ring_add(log, log_offset, in r5l_recovery_verify_data_checksum_for_mb()
2059 log_offset = r5l_ring_add(log, log_offset, in r5l_recovery_verify_data_checksum_for_mb()
2085 r5c_recovery_analyze_meta_block(struct r5l_log *log, in r5c_recovery_analyze_meta_block() argument
2089 struct mddev *mddev = log->rdev->mddev; in r5c_recovery_analyze_meta_block()
2105 ret = r5l_recovery_verify_data_checksum_for_mb(log, ctx); in r5c_recovery_analyze_meta_block()
2113 log_offset = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); in r5c_recovery_analyze_meta_block()
2187 r5l_recovery_load_data(log, sh, ctx, payload, in r5c_recovery_analyze_meta_block()
2190 r5l_recovery_load_parity(log, sh, ctx, payload, in r5c_recovery_analyze_meta_block()
2195 log_offset = r5l_ring_add(log, log_offset, in r5c_recovery_analyze_meta_block()
2210 static void r5c_recovery_load_one_stripe(struct r5l_log *log, in r5c_recovery_load_one_stripe() argument
2241 static int r5c_recovery_flush_log(struct r5l_log *log, in r5c_recovery_flush_log() argument
2249 if (r5l_recovery_read_meta_block(log, ctx)) in r5c_recovery_flush_log()
2252 ret = r5c_recovery_analyze_meta_block(log, ctx, in r5c_recovery_flush_log()
2261 ctx->pos = r5l_ring_add(log, ctx->pos, ctx->meta_total_blocks); in r5c_recovery_flush_log()
2275 r5c_recovery_load_one_stripe(log, sh); in r5c_recovery_flush_log()
2352 r5c_recovery_rewrite_data_only_stripes(struct r5l_log *log, in r5c_recovery_rewrite_data_only_stripes() argument
2356 struct mddev *mddev = log->rdev->mddev; in r5c_recovery_rewrite_data_only_stripes()
2376 r5l_recovery_create_empty_meta_block(log, page, in r5c_recovery_rewrite_data_only_stripes()
2380 write_pos = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); in r5c_recovery_rewrite_data_only_stripes()
2396 crc32c_le(log->uuid_checksum, addr, in r5c_recovery_rewrite_data_only_stripes()
2399 sync_page_io(log->rdev, write_pos, PAGE_SIZE, in r5c_recovery_rewrite_data_only_stripes()
2401 write_pos = r5l_ring_add(log, write_pos, in r5c_recovery_rewrite_data_only_stripes()
2409 mb->checksum = cpu_to_le32(crc32c_le(log->uuid_checksum, in r5c_recovery_rewrite_data_only_stripes()
2411 sync_page_io(log->rdev, ctx->pos, PAGE_SIZE, page, in r5c_recovery_rewrite_data_only_stripes()
2414 list_add_tail(&sh->r5c, &log->stripe_in_journal_list); in r5c_recovery_rewrite_data_only_stripes()
2415 atomic_inc(&log->stripe_in_journal_count); in r5c_recovery_rewrite_data_only_stripes()
2420 log->next_checkpoint = next_checkpoint; in r5c_recovery_rewrite_data_only_stripes()
2425 static void r5c_recovery_flush_data_only_stripes(struct r5l_log *log, in r5c_recovery_flush_data_only_stripes() argument
2428 struct mddev *mddev = log->rdev->mddev; in r5c_recovery_flush_data_only_stripes()
2435 log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_BACK; in r5c_recovery_flush_data_only_stripes()
2448 log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; in r5c_recovery_flush_data_only_stripes()
2451 static int r5l_recovery_log(struct r5l_log *log) in r5l_recovery_log() argument
2453 struct mddev *mddev = log->rdev->mddev; in r5l_recovery_log()
2462 ctx->pos = log->last_checkpoint; in r5l_recovery_log()
2463 ctx->seq = log->last_cp_seq; in r5l_recovery_log()
2472 if (r5l_recovery_allocate_ra_pool(log, ctx) != 0) { in r5l_recovery_log()
2477 ret = r5c_recovery_flush_log(log, ctx); in r5l_recovery_log()
2494 log->next_checkpoint = ctx->pos; in r5l_recovery_log()
2495 r5l_log_write_empty_meta_block(log, ctx->pos, ctx->seq++); in r5l_recovery_log()
2496 ctx->pos = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); in r5l_recovery_log()
2497 } else if (r5c_recovery_rewrite_data_only_stripes(log, ctx)) { in r5l_recovery_log()
2504 log->log_start = ctx->pos; in r5l_recovery_log()
2505 log->seq = ctx->seq; in r5l_recovery_log()
2506 log->last_checkpoint = pos; in r5l_recovery_log()
2507 r5l_write_super(log, pos); in r5l_recovery_log()
2509 r5c_recovery_flush_data_only_stripes(log, ctx); in r5l_recovery_log()
2512 r5l_recovery_free_ra_pool(log, ctx); in r5l_recovery_log()
2520 static void r5l_write_super(struct r5l_log *log, sector_t cp) in r5l_write_super() argument
2522 struct mddev *mddev = log->rdev->mddev; in r5l_write_super()
2524 log->rdev->journal_tail = cp; in r5l_write_super()
2538 if (!conf || !conf->log) { in r5c_journal_mode_show()
2543 switch (conf->log->r5c_journal_mode) { in r5c_journal_mode_show()
2578 if (!conf || !conf->log) in r5c_journal_mode_set()
2586 conf->log->r5c_journal_mode = mode; in r5c_journal_mode_set()
2637 struct r5l_log *log = conf->log; in r5c_try_caching_write() local
2646 BUG_ON(!r5c_is_writeback(log)); in r5c_try_caching_write()
2696 spin_lock(&log->tree_lock); in r5c_try_caching_write()
2697 pslot = radix_tree_lookup_slot(&log->big_stripe_tree, in r5c_try_caching_write()
2701 pslot, &log->tree_lock) >> in r5c_try_caching_write()
2704 &log->big_stripe_tree, pslot, in r5c_try_caching_write()
2712 &log->big_stripe_tree, tree_index, in r5c_try_caching_write()
2715 spin_unlock(&log->tree_lock); in r5c_try_caching_write()
2720 spin_unlock(&log->tree_lock); in r5c_try_caching_write()
2804 struct r5l_log *log = conf->log; in r5c_finish_stripe_write_out() local
2811 if (!log || !test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags)) in r5c_finish_stripe_write_out()
2817 if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) in r5c_finish_stripe_write_out()
2839 spin_lock_irq(&log->stripe_in_journal_lock); in r5c_finish_stripe_write_out()
2841 spin_unlock_irq(&log->stripe_in_journal_lock); in r5c_finish_stripe_write_out()
2844 atomic_dec(&log->stripe_in_journal_count); in r5c_finish_stripe_write_out()
2845 r5c_update_log_state(log); in r5c_finish_stripe_write_out()
2851 spin_lock(&log->tree_lock); in r5c_finish_stripe_write_out()
2852 pslot = radix_tree_lookup_slot(&log->big_stripe_tree, in r5c_finish_stripe_write_out()
2856 pslot, &log->tree_lock) >> in r5c_finish_stripe_write_out()
2859 radix_tree_delete(&log->big_stripe_tree, tree_index); in r5c_finish_stripe_write_out()
2862 &log->big_stripe_tree, pslot, in r5c_finish_stripe_write_out()
2864 spin_unlock(&log->tree_lock); in r5c_finish_stripe_write_out()
2879 r5l_append_flush_payload(log, sh->sector); in r5c_finish_stripe_write_out()
2885 int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh) in r5c_cache_data() argument
2893 BUG_ON(!log); in r5c_cache_data()
2901 sh->dev[i].log_checksum = crc32c_le(log->uuid_checksum, in r5c_cache_data()
2915 mutex_lock(&log->io_mutex); in r5c_cache_data()
2921 r5l_add_no_space_stripe(log, sh); in r5c_cache_data()
2922 else if (!r5l_has_free_space(log, reserve)) { in r5c_cache_data()
2923 if (sh->log_start == log->last_checkpoint) in r5c_cache_data()
2926 r5l_add_no_space_stripe(log, sh); in r5c_cache_data()
2928 ret = r5l_log_stripe(log, sh, pages, 0); in r5c_cache_data()
2930 spin_lock_irq(&log->io_list_lock); in r5c_cache_data()
2931 list_add_tail(&sh->log_list, &log->no_mem_stripes); in r5c_cache_data()
2932 spin_unlock_irq(&log->io_list_lock); in r5c_cache_data()
2936 mutex_unlock(&log->io_mutex); in r5c_cache_data()
2943 struct r5l_log *log = conf->log; in r5c_big_stripe_cached() local
2947 if (!log) in r5c_big_stripe_cached()
2952 slot = radix_tree_lookup(&log->big_stripe_tree, tree_index); in r5c_big_stripe_cached()
2956 static int r5l_load_log(struct r5l_log *log) in r5l_load_log() argument
2958 struct md_rdev *rdev = log->rdev; in r5l_load_log()
2961 sector_t cp = log->rdev->journal_tail; in r5l_load_log()
2986 expected_crc = crc32c_le(log->uuid_checksum, mb, PAGE_SIZE); in r5l_load_log()
2997 log->last_cp_seq = prandom_u32(); in r5l_load_log()
2999 r5l_log_write_empty_meta_block(log, cp, log->last_cp_seq); in r5l_load_log()
3005 r5l_write_super(log, cp); in r5l_load_log()
3007 log->last_cp_seq = le64_to_cpu(mb->seq); in r5l_load_log()
3009 log->device_size = round_down(rdev->sectors, BLOCK_SECTORS); in r5l_load_log()
3010 log->max_free_space = log->device_size >> RECLAIM_MAX_FREE_SPACE_SHIFT; in r5l_load_log()
3011 if (log->max_free_space > RECLAIM_MAX_FREE_SPACE) in r5l_load_log()
3012 log->max_free_space = RECLAIM_MAX_FREE_SPACE; in r5l_load_log()
3013 log->last_checkpoint = cp; in r5l_load_log()
3018 log->log_start = r5l_ring_add(log, cp, BLOCK_SECTORS); in r5l_load_log()
3019 log->seq = log->last_cp_seq + 1; in r5l_load_log()
3020 log->next_checkpoint = cp; in r5l_load_log()
3022 ret = r5l_recovery_log(log); in r5l_load_log()
3024 r5c_update_log_state(log); in r5l_load_log()
3031 int r5l_start(struct r5l_log *log) in r5l_start() argument
3035 if (!log) in r5l_start()
3038 ret = r5l_load_log(log); in r5l_start()
3040 struct mddev *mddev = log->rdev->mddev; in r5l_start()
3051 struct r5l_log *log = conf->log; in r5c_update_on_rdev_error() local
3053 if (!log) in r5c_update_on_rdev_error()
3058 conf->log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) in r5c_update_on_rdev_error()
3059 schedule_work(&log->disable_writeback_work); in r5c_update_on_rdev_error()
3065 struct r5l_log *log; in r5l_init_log() local
3090 log = kzalloc(sizeof(*log), GFP_KERNEL); in r5l_init_log()
3091 if (!log) in r5l_init_log()
3093 log->rdev = rdev; in r5l_init_log()
3095 log->need_cache_flush = test_bit(QUEUE_FLAG_WC, &q->queue_flags) != 0; in r5l_init_log()
3097 log->uuid_checksum = crc32c_le(~0, rdev->mddev->uuid, in r5l_init_log()
3100 mutex_init(&log->io_mutex); in r5l_init_log()
3102 spin_lock_init(&log->io_list_lock); in r5l_init_log()
3103 INIT_LIST_HEAD(&log->running_ios); in r5l_init_log()
3104 INIT_LIST_HEAD(&log->io_end_ios); in r5l_init_log()
3105 INIT_LIST_HEAD(&log->flushing_ios); in r5l_init_log()
3106 INIT_LIST_HEAD(&log->finished_ios); in r5l_init_log()
3107 bio_init(&log->flush_bio, NULL, 0); in r5l_init_log()
3109 log->io_kc = KMEM_CACHE(r5l_io_unit, 0); in r5l_init_log()
3110 if (!log->io_kc) in r5l_init_log()
3113 ret = mempool_init_slab_pool(&log->io_pool, R5L_POOL_SIZE, log->io_kc); in r5l_init_log()
3117 ret = bioset_init(&log->bs, R5L_POOL_SIZE, 0, BIOSET_NEED_BVECS); in r5l_init_log()
3121 ret = mempool_init_page_pool(&log->meta_pool, R5L_POOL_SIZE, 0); in r5l_init_log()
3125 spin_lock_init(&log->tree_lock); in r5l_init_log()
3126 INIT_RADIX_TREE(&log->big_stripe_tree, GFP_NOWAIT | __GFP_NOWARN); in r5l_init_log()
3128 log->reclaim_thread = md_register_thread(r5l_reclaim_thread, in r5l_init_log()
3129 log->rdev->mddev, "reclaim"); in r5l_init_log()
3130 if (!log->reclaim_thread) in r5l_init_log()
3132 log->reclaim_thread->timeout = R5C_RECLAIM_WAKEUP_INTERVAL; in r5l_init_log()
3134 init_waitqueue_head(&log->iounit_wait); in r5l_init_log()
3136 INIT_LIST_HEAD(&log->no_mem_stripes); in r5l_init_log()
3138 INIT_LIST_HEAD(&log->no_space_stripes); in r5l_init_log()
3139 spin_lock_init(&log->no_space_stripes_lock); in r5l_init_log()
3141 INIT_WORK(&log->deferred_io_work, r5l_submit_io_async); in r5l_init_log()
3142 INIT_WORK(&log->disable_writeback_work, r5c_disable_writeback_async); in r5l_init_log()
3144 log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; in r5l_init_log()
3145 INIT_LIST_HEAD(&log->stripe_in_journal_list); in r5l_init_log()
3146 spin_lock_init(&log->stripe_in_journal_lock); in r5l_init_log()
3147 atomic_set(&log->stripe_in_journal_count, 0); in r5l_init_log()
3149 rcu_assign_pointer(conf->log, log); in r5l_init_log()
3154 rcu_assign_pointer(conf->log, NULL); in r5l_init_log()
3155 md_unregister_thread(&log->reclaim_thread); in r5l_init_log()
3157 mempool_exit(&log->meta_pool); in r5l_init_log()
3159 bioset_exit(&log->bs); in r5l_init_log()
3161 mempool_exit(&log->io_pool); in r5l_init_log()
3163 kmem_cache_destroy(log->io_kc); in r5l_init_log()
3165 kfree(log); in r5l_init_log()
3171 struct r5l_log *log = conf->log; in r5l_exit_log() local
3173 conf->log = NULL; in r5l_exit_log()
3178 flush_work(&log->disable_writeback_work); in r5l_exit_log()
3179 md_unregister_thread(&log->reclaim_thread); in r5l_exit_log()
3180 mempool_exit(&log->meta_pool); in r5l_exit_log()
3181 bioset_exit(&log->bs); in r5l_exit_log()
3182 mempool_exit(&log->io_pool); in r5l_exit_log()
3183 kmem_cache_destroy(log->io_kc); in r5l_exit_log()
3184 kfree(log); in r5l_exit_log()