Lines Matching refs:mddev
49 static void raid1_free(struct mddev *mddev, void *priv);
84 struct mddev *mddev = rdev->mddev; in wait_for_serialization() local
89 if (WARN_ON(!mddev->serial_info_pool)) in wait_for_serialization()
91 si = mempool_alloc(mddev->serial_info_pool, GFP_NOIO); in wait_for_serialization()
101 struct mddev *mddev = rdev->mddev; in remove_serial() local
110 mempool_free(si, mddev->serial_info_pool); in remove_serial()
178 if (test_bit(MD_RECOVERY_REQUESTED, &conf->mddev->recovery)) in r1buf_pool_alloc()
253 struct r1conf *conf = r1_bio->mddev->private; in free_r1bio()
261 struct r1conf *conf = r1_bio->mddev->private; in put_buf()
268 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev); in put_buf()
279 struct mddev *mddev = r1_bio->mddev; in reschedule_retry() local
280 struct r1conf *conf = mddev->private; in reschedule_retry()
290 md_wakeup_thread(mddev->thread); in reschedule_retry()
311 struct r1conf *conf = r1_bio->mddev->private; in raid_end_bio_io()
337 struct r1conf *conf = r1_bio->mddev->private; in update_head_pos()
349 struct r1conf *conf = r1_bio->mddev->private; in find_bio_disk()
366 struct r1conf *conf = r1_bio->mddev->private; in raid1_end_read_request()
390 if (r1_bio->mddev->degraded == conf->raid_disks || in raid1_end_read_request()
391 (r1_bio->mddev->degraded == conf->raid_disks-1 && in raid1_end_read_request()
399 rdev_dec_pending(rdev, conf->mddev); in raid1_end_read_request()
405 mdname(conf->mddev), in raid1_end_read_request()
416 struct mddev *mddev = r1_bio->mddev; in close_write() local
426 mddev->bitmap_ops->end_behind_write(mddev); in close_write()
427 md_write_end(mddev); in close_write()
450 struct r1conf *conf = r1_bio->mddev->private; in raid1_end_write_request()
466 conf->mddev->recovery); in raid1_end_write_request()
472 md_error(r1_bio->mddev, rdev); in raid1_end_write_request()
544 } else if (test_bit(MD_SERIALIZE_POLICY, &rdev->mddev->flags)) in raid1_end_write_request()
547 rdev_dec_pending(rdev, conf->mddev); in raid1_end_write_request()
869 if (raid1_should_read_first(conf->mddev, r1_bio->sector, in read_balance()
902 raid1_prepare_flush_writes(conf->mddev); in flush_bio_list()
1003 test_bit(MD_RECOVERY_INTR, &conf->mddev->recovery), in raise_barrier()
1006 if (test_bit(MD_RECOVERY_INTR, &conf->mddev->recovery)) { in raise_barrier()
1202 mddev_add_trace_msg(conf->mddev, "raid1 wait freeze"); in freeze_array()
1228 &r1_bio->mddev->bio_set); in alloc_behind_master_bio()
1271 struct mddev *mddev = plug->cb.data; in raid1_unplug() local
1272 struct r1conf *conf = mddev->private; in raid1_unplug()
1280 md_wakeup_thread(mddev->thread); in raid1_unplug()
1291 static void init_r1bio(struct r1bio *r1_bio, struct mddev *mddev, struct bio *bio) in init_r1bio() argument
1296 r1_bio->mddev = mddev; in init_r1bio()
1301 alloc_r1bio(struct mddev *mddev, struct bio *bio) in alloc_r1bio() argument
1303 struct r1conf *conf = mddev->private; in alloc_r1bio()
1308 init_r1bio(r1_bio, mddev, bio); in alloc_r1bio()
1312 static void raid1_read_request(struct mddev *mddev, struct bio *bio, in raid1_read_request() argument
1315 struct r1conf *conf = mddev->private; in raid1_read_request()
1340 r1_bio = alloc_r1bio(mddev, bio); in raid1_read_request()
1342 init_r1bio(r1_bio, mddev, bio); in raid1_read_request()
1354 mdname(mddev), in raid1_read_request()
1364 mdname(mddev), in raid1_read_request()
1369 md_bitmap_enabled(mddev, false)) { in raid1_read_request()
1374 mddev_add_trace_msg(mddev, "raid1 wait behind writes"); in raid1_read_request()
1375 mddev->bitmap_ops->wait_behind_writes(mddev); in raid1_read_request()
1392 md_account_bio(mddev, &bio); in raid1_read_request()
1396 &mddev->bio_set); in raid1_read_request()
1407 mddev_trace_remap(mddev, read_bio, r1_bio->sector); in raid1_read_request()
1416 static bool wait_blocked_rdev(struct mddev *mddev, struct bio *bio) in wait_blocked_rdev() argument
1418 struct r1conf *conf = mddev->private; in wait_blocked_rdev()
1439 mddev_add_trace_msg(rdev->mddev, "raid1 wait rdev %d blocked", in wait_blocked_rdev()
1442 md_wait_for_blocked_rdev(rdev, rdev->mddev); in wait_blocked_rdev()
1450 static void raid1_start_write_behind(struct mddev *mddev, struct r1bio *r1_bio, in raid1_start_write_behind() argument
1453 unsigned long max_write_behind = mddev->bitmap_info.max_write_behind; in raid1_start_write_behind()
1458 if (!md_bitmap_enabled(mddev, false)) in raid1_start_write_behind()
1461 err = mddev->bitmap_ops->get_stats(mddev->bitmap, &stats); in raid1_start_write_behind()
1470 mddev->bitmap_ops->start_behind_write(mddev); in raid1_start_write_behind()
1474 static void raid1_write_request(struct mddev *mddev, struct bio *bio, in raid1_write_request() argument
1477 struct r1conf *conf = mddev->private; in raid1_write_request()
1486 if (mddev_is_clustered(mddev) && in raid1_write_request()
1487 mddev->cluster_ops->area_resyncing(mddev, WRITE, in raid1_write_request()
1498 if (!mddev->cluster_ops->area_resyncing(mddev, WRITE, in raid1_write_request()
1518 if (!wait_blocked_rdev(mddev, bio)) { in raid1_write_request()
1523 r1_bio = alloc_r1bio(mddev, bio); in raid1_write_request()
1570 rdev_dec_pending(rdev, mddev); in raid1_write_request()
1600 if (write_behind && mddev->bitmap) in raid1_write_request()
1615 md_account_bio(mddev, &bio); in raid1_write_request()
1630 raid1_start_write_behind(mddev, r1_bio, bio); in raid1_write_request()
1637 GFP_NOIO, &mddev->bio_set); in raid1_write_request()
1644 &mddev->bio_set); in raid1_write_request()
1646 if (test_bit(MD_SERIALIZE_POLICY, &mddev->flags)) in raid1_write_request()
1657 conf->raid_disks - mddev->degraded > 1) in raid1_write_request()
1662 mddev_trace_remap(mddev, mbio, r1_bio->sector); in raid1_write_request()
1665 if (!raid1_add_bio_to_plug(mddev, mbio, raid1_unplug, disks)) { in raid1_write_request()
1669 md_wakeup_thread(mddev->thread); in raid1_write_request()
1681 rdev_dec_pending(conf->mirrors[k].rdev, mddev); in raid1_write_request()
1689 static bool raid1_make_request(struct mddev *mddev, struct bio *bio) in raid1_make_request() argument
1694 && md_flush_request(mddev, bio)) in raid1_make_request()
1708 raid1_read_request(mddev, bio, sectors, NULL); in raid1_make_request()
1710 md_write_start(mddev,bio); in raid1_make_request()
1711 raid1_write_request(mddev, bio, sectors); in raid1_make_request()
1716 static void raid1_status(struct seq_file *seq, struct mddev *mddev) in raid1_status() argument
1718 struct r1conf *conf = mddev->private; in raid1_status()
1721 lockdep_assert_held(&mddev->lock); in raid1_status()
1724 conf->raid_disks - mddev->degraded); in raid1_status()
1750 static void raid1_error(struct mddev *mddev, struct md_rdev *rdev) in raid1_error() argument
1752 struct r1conf *conf = mddev->private; in raid1_error()
1758 (conf->raid_disks - mddev->degraded) == 1) { in raid1_error()
1759 set_bit(MD_BROKEN, &mddev->flags); in raid1_error()
1761 if (!test_bit(MD_FAILLAST_DEV, &mddev->flags)) { in raid1_error()
1768 mddev->degraded++; in raid1_error()
1774 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in raid1_error()
1775 set_mask_bits(&mddev->sb_flags, 0, in raid1_error()
1779 mdname(mddev), rdev->bdev, in raid1_error()
1780 mdname(mddev), conf->raid_disks - mddev->degraded); in raid1_error()
1792 pr_debug(" --- wd:%d rd:%d\n", conf->raid_disks - conf->mddev->degraded, in print_conf()
1795 lockdep_assert_held(&conf->mddev->reconfig_mutex); in print_conf()
1818 static int raid1_spare_active(struct mddev *mddev) in raid1_spare_active() argument
1821 struct r1conf *conf = mddev->private; in raid1_spare_active()
1863 mddev->degraded -= count; in raid1_spare_active()
1905 rdev->mddev->degraded < conf->raid_disks) in raid1_remove_conf()
1915 static int raid1_add_disk(struct mddev *mddev, struct md_rdev *rdev) in raid1_add_disk() argument
1917 struct r1conf *conf = mddev->private; in raid1_add_disk()
1940 err = mddev_stack_new_rdev(mddev, rdev); in raid1_add_disk()
1970 static int raid1_remove_disk(struct mddev *mddev, struct md_rdev *rdev) in raid1_remove_disk() argument
1972 struct r1conf *conf = mddev->private; in raid1_remove_disk()
2019 err = md_integrity_register(mddev); in raid1_remove_disk()
2045 static void abort_sync_write(struct mddev *mddev, struct r1bio *r1_bio) in abort_sync_write() argument
2053 md_bitmap_end_sync(mddev, s, &sync_blocks); in abort_sync_write()
2062 struct mddev *mddev = r1_bio->mddev; in put_sync_write_buf() local
2070 md_done_sync(mddev, s); in put_sync_write_buf()
2078 struct mddev *mddev = r1_bio->mddev; in end_sync_write() local
2079 struct r1conf *conf = mddev->private; in end_sync_write()
2083 abort_sync_write(mddev, r1_bio); in end_sync_write()
2087 mddev->recovery); in end_sync_write()
2109 rdev->mddev->recovery); in r1_sync_page_io()
2129 struct mddev *mddev = r1_bio->mddev; in fix_sync_read_error() local
2130 struct r1conf *conf = mddev->private; in fix_sync_read_error()
2142 md_error(mddev, rdev); in fix_sync_read_error()
2185 mdname(mddev), bio->bi_bdev, in fix_sync_read_error()
2217 rdev_dec_pending(rdev, mddev); in fix_sync_read_error()
2251 struct mddev *mddev = r1_bio->mddev; in process_checks() local
2252 struct r1conf *conf = mddev->private; in process_checks()
2282 rdev_dec_pending(conf->mirrors[primary].rdev, mddev); in process_checks()
2315 atomic64_add(r1_bio->sectors, &mddev->resync_mismatches); in process_checks()
2316 if (j < 0 || (test_bit(MD_RECOVERY_CHECK, &mddev->recovery) in process_checks()
2320 rdev_dec_pending(conf->mirrors[i].rdev, mddev); in process_checks()
2328 static void sync_request_write(struct mddev *mddev, struct r1bio *r1_bio) in sync_request_write() argument
2330 struct r1conf *conf = mddev->private; in sync_request_write()
2341 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery) || in sync_request_write()
2343 md_done_sync(mddev, r1_bio->sectors); in sync_request_write()
2344 md_sync_error(mddev); in sync_request_write()
2350 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) in sync_request_write()
2362 !test_bit(MD_RECOVERY_SYNC, &mddev->recovery)))) in sync_request_write()
2365 abort_sync_write(mddev, r1_bio); in sync_request_write()
2395 struct mddev *mddev = conf->mddev; in fix_read_error() local
2398 if (exceed_read_errors(mddev, rdev)) { in fix_read_error()
2423 rdev_dec_pending(rdev, mddev); in fix_read_error()
2451 rdev_dec_pending(rdev, mddev); in fix_read_error()
2467 mdname(mddev), s, in fix_read_error()
2472 rdev_dec_pending(rdev, mddev); in fix_read_error()
2482 struct mddev *mddev = r1_bio->mddev; in narrow_write_error() local
2483 struct r1conf *conf = mddev->private; in narrow_write_error()
2521 GFP_NOIO, &mddev->bio_set); in narrow_write_error()
2524 GFP_NOIO, &mddev->bio_set); in narrow_write_error()
2568 md_done_sync(conf->mddev, s); in handle_sync_write_finished()
2582 rdev_dec_pending(rdev, conf->mddev); in handle_write_finished()
2591 conf->mddev); in handle_write_finished()
2604 md_wakeup_thread(conf->mddev->thread); in handle_write_finished()
2614 struct mddev *mddev = conf->mddev; in handle_read_error() local
2634 if (mddev->ro == 0 in handle_read_error()
2639 } else if (mddev->ro == 0 && test_bit(FailFast, &rdev->flags)) { in handle_read_error()
2640 md_error(mddev, rdev); in handle_read_error()
2645 rdev_dec_pending(rdev, conf->mddev); in handle_read_error()
2651 raid1_read_request(mddev, bio, r1_bio->sectors, r1_bio); in handle_read_error()
2657 struct mddev *mddev = thread->mddev; in raid1d() local
2660 struct r1conf *conf = mddev->private; in raid1d()
2665 md_check_recovery(mddev); in raid1d()
2668 !test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) { in raid1d()
2671 if (!test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) in raid1d()
2702 mddev = r1_bio->mddev; in raid1d()
2703 conf = mddev->private; in raid1d()
2709 sync_request_write(mddev, r1_bio); in raid1d()
2719 if (mddev->sb_flags & ~(1<<MD_SB_CHANGE_PENDING)) in raid1d()
2720 md_check_recovery(mddev); in raid1d()
2763 static sector_t raid1_sync_request(struct mddev *mddev, sector_t sector_nr, in raid1_sync_request() argument
2766 struct r1conf *conf = mddev->private; in raid1_sync_request()
2791 if (mddev->curr_resync < max_sector) /* aborted */ in raid1_sync_request()
2792 md_bitmap_end_sync(mddev, mddev->curr_resync, in raid1_sync_request()
2797 if (md_bitmap_enabled(mddev, false)) in raid1_sync_request()
2798 mddev->bitmap_ops->close_sync(mddev); in raid1_sync_request()
2801 if (mddev_is_clustered(mddev)) { in raid1_sync_request()
2808 if (mddev->bitmap == NULL && in raid1_sync_request()
2809 mddev->resync_offset == MaxSector && in raid1_sync_request()
2810 !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery) && in raid1_sync_request()
2818 if (!md_bitmap_start_sync(mddev, sector_nr, &sync_blocks, true) && in raid1_sync_request()
2819 !conf->fullsync && !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) { in raid1_sync_request()
2835 if (md_bitmap_enabled(mddev, false)) in raid1_sync_request()
2836 mddev->bitmap_ops->cond_end_sync(mddev, sector_nr, in raid1_sync_request()
2837 mddev_is_clustered(mddev) && in raid1_sync_request()
2855 r1_bio->mddev = mddev; in raid1_sync_request()
2903 test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && in raid1_sync_request()
2904 !test_bit(MD_RECOVERY_CHECK, &mddev->recovery)) { in raid1_sync_request()
2940 set_bit(MD_SB_CHANGE_DEVS, &mddev->sb_flags); in raid1_sync_request()
2959 if (test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && read_targets > 0) in raid1_sync_request()
2976 if (max_sector > mddev->resync_max) in raid1_sync_request()
2977 max_sector = mddev->resync_max; /* Don't do IO beyond here */ in raid1_sync_request()
2990 if (!md_bitmap_start_sync(mddev, sector_nr, in raid1_sync_request()
2993 !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) in raid1_sync_request()
3021 if (mddev_is_clustered(mddev) && in raid1_sync_request()
3023 conf->cluster_sync_low = mddev->curr_resync_completed; in raid1_sync_request()
3026 mddev->cluster_ops->resync_info_update(mddev, in raid1_sync_request()
3034 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) { in raid1_sync_request()
3055 static sector_t raid1_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid1_size() argument
3060 return mddev->dev_sectors; in raid1_size()
3063 static struct r1conf *setup_conf(struct mddev *mddev) in setup_conf() argument
3093 mddev->raid_disks, 2), in setup_conf()
3102 r1bio_size = offsetof(struct r1bio, bios[mddev->raid_disks * 2]); in setup_conf()
3113 conf->raid_disks = mddev->raid_disks; in setup_conf()
3114 rdev_for_each(rdev, mddev) { in setup_conf()
3124 conf->mddev = mddev; in setup_conf()
3164 md_register_thread(raid1d, mddev, "raid1")); in setup_conf()
3185 static int raid1_set_limits(struct mddev *mddev) in raid1_set_limits() argument
3193 lim.logical_block_size = mddev->logical_block_size; in raid1_set_limits()
3195 err = mddev_stack_rdev_limits(mddev, &lim, MDDEV_STACK_INTEGRITY); in raid1_set_limits()
3198 return queue_limits_set(mddev->gendisk->queue, &lim); in raid1_set_limits()
3201 static int raid1_run(struct mddev *mddev) in raid1_run() argument
3207 if (mddev->level != 1) { in raid1_run()
3209 mdname(mddev), mddev->level); in raid1_run()
3212 if (mddev->reshape_position != MaxSector) { in raid1_run()
3214 mdname(mddev)); in raid1_run()
3223 if (mddev->private == NULL) in raid1_run()
3224 conf = setup_conf(mddev); in raid1_run()
3226 conf = mddev->private; in raid1_run()
3231 if (!mddev_is_dm(mddev)) { in raid1_run()
3232 ret = raid1_set_limits(mddev); in raid1_run()
3234 md_unregister_thread(mddev, &conf->thread); in raid1_run()
3235 if (!mddev->private) in raid1_run()
3236 raid1_free(mddev, conf); in raid1_run()
3241 mddev->degraded = 0; in raid1_run()
3246 mddev->degraded++; in raid1_run()
3250 if (conf->raid_disks - mddev->degraded < 1) { in raid1_run()
3251 md_unregister_thread(mddev, &conf->thread); in raid1_run()
3252 if (!mddev->private) in raid1_run()
3253 raid1_free(mddev, conf); in raid1_run()
3257 if (conf->raid_disks - mddev->degraded == 1) in raid1_run()
3258 mddev->resync_offset = MaxSector; in raid1_run()
3260 if (mddev->resync_offset != MaxSector) in raid1_run()
3262 mdname(mddev)); in raid1_run()
3264 mdname(mddev), mddev->raid_disks - mddev->degraded, in raid1_run()
3265 mddev->raid_disks); in raid1_run()
3270 rcu_assign_pointer(mddev->thread, conf->thread); in raid1_run()
3272 mddev->private = conf; in raid1_run()
3273 set_bit(MD_FAILFAST_SUPPORTED, &mddev->flags); in raid1_run()
3275 md_set_array_sectors(mddev, raid1_size(mddev, 0, 0)); in raid1_run()
3277 ret = md_integrity_register(mddev); in raid1_run()
3279 md_unregister_thread(mddev, &mddev->thread); in raid1_run()
3283 static void raid1_free(struct mddev *mddev, void *priv) in raid1_free() argument
3298 static int raid1_resize(struct mddev *mddev, sector_t sectors) in raid1_resize() argument
3307 sector_t newsize = raid1_size(mddev, sectors, 0); in raid1_resize()
3309 if (mddev->external_size && in raid1_resize()
3310 mddev->array_sectors > newsize) in raid1_resize()
3313 if (md_bitmap_enabled(mddev, false)) { in raid1_resize()
3314 int ret = mddev->bitmap_ops->resize(mddev, newsize, 0); in raid1_resize()
3320 md_set_array_sectors(mddev, newsize); in raid1_resize()
3321 if (sectors > mddev->dev_sectors && in raid1_resize()
3322 mddev->resync_offset > mddev->dev_sectors) { in raid1_resize()
3323 mddev->resync_offset = mddev->dev_sectors; in raid1_resize()
3324 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in raid1_resize()
3326 mddev->dev_sectors = sectors; in raid1_resize()
3327 mddev->resync_max_sectors = sectors; in raid1_resize()
3331 static int raid1_reshape(struct mddev *mddev) in raid1_reshape() argument
3347 struct r1conf *conf = mddev->private; in raid1_reshape()
3353 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape()
3354 mddev->layout != mddev->new_layout || in raid1_reshape()
3355 mddev->level != mddev->new_level) { in raid1_reshape()
3356 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
3357 mddev->new_layout = mddev->layout; in raid1_reshape()
3358 mddev->new_level = mddev->level; in raid1_reshape()
3362 if (!mddev_is_clustered(mddev)) in raid1_reshape()
3363 md_allow_write(mddev); in raid1_reshape()
3365 raid_disks = mddev->raid_disks + mddev->delta_disks; in raid1_reshape()
3398 sysfs_unlink_rdev(mddev, rdev); in raid1_reshape()
3400 sysfs_unlink_rdev(mddev, rdev); in raid1_reshape()
3401 if (sysfs_link_rdev(mddev, rdev)) in raid1_reshape()
3403 mdname(mddev), rdev->raid_disk); in raid1_reshape()
3412 mddev->degraded += (raid_disks - conf->raid_disks); in raid1_reshape()
3414 conf->raid_disks = mddev->raid_disks = raid_disks; in raid1_reshape()
3415 mddev->delta_disks = 0; in raid1_reshape()
3419 set_bit(MD_RECOVERY_RECOVER, &mddev->recovery); in raid1_reshape()
3420 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in raid1_reshape()
3421 md_wakeup_thread(mddev->thread); in raid1_reshape()
3427 static void raid1_quiesce(struct mddev *mddev, int quiesce) in raid1_quiesce() argument
3429 struct r1conf *conf = mddev->private; in raid1_quiesce()
3437 static void *raid1_takeover(struct mddev *mddev) in raid1_takeover() argument
3442 if (mddev->level == 5 && mddev->raid_disks == 2) { in raid1_takeover()
3444 mddev->new_level = 1; in raid1_takeover()
3445 mddev->new_layout = 0; in raid1_takeover()
3446 mddev->new_chunk_sectors = 0; in raid1_takeover()
3447 conf = setup_conf(mddev); in raid1_takeover()
3451 mddev_clear_unsupported_flags(mddev, in raid1_takeover()