• Home
  • Raw
  • Download

Lines Matching refs:mddev

58 	struct mddev *mddev = rdev->mddev;  in check_and_add_wb()  local
60 wi = mempool_alloc(mddev->wb_info_pool, GFP_NOIO); in check_and_add_wb()
76 mempool_free(wi, mddev->wb_info_pool); in check_and_add_wb()
87 struct mddev *mddev = rdev->mddev; in remove_wb() local
93 mempool_free(wi, mddev->wb_info_pool); in remove_wb()
162 if (test_bit(MD_RECOVERY_REQUESTED, &pi->mddev->recovery)) in r1buf_pool_alloc()
234 struct r1conf *conf = r1_bio->mddev->private; in free_r1bio()
242 struct r1conf *conf = r1_bio->mddev->private; in put_buf()
249 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev); in put_buf()
260 struct mddev *mddev = r1_bio->mddev; in reschedule_retry() local
261 struct r1conf *conf = mddev->private; in reschedule_retry()
271 md_wakeup_thread(mddev->thread); in reschedule_retry()
282 struct r1conf *conf = r1_bio->mddev->private; in call_bio_endio()
316 struct r1conf *conf = r1_bio->mddev->private; in update_head_pos()
328 struct r1conf *conf = r1_bio->mddev->private; in find_bio_disk()
345 struct r1conf *conf = r1_bio->mddev->private; in raid1_end_read_request()
367 if (r1_bio->mddev->degraded == conf->raid_disks || in raid1_end_read_request()
368 (r1_bio->mddev->degraded == conf->raid_disks-1 && in raid1_end_read_request()
376 rdev_dec_pending(rdev, conf->mddev); in raid1_end_read_request()
383 mdname(conf->mddev), in raid1_end_read_request()
401 md_bitmap_endwrite(r1_bio->mddev->bitmap, r1_bio->sector, in close_write()
405 md_write_end(r1_bio->mddev); in close_write()
428 struct r1conf *conf = r1_bio->mddev->private; in raid1_end_write_request()
443 conf->mddev->recovery); in raid1_end_write_request()
449 md_error(r1_bio->mddev, rdev); in raid1_end_write_request()
532 rdev_dec_pending(rdev, conf->mddev); in raid1_end_write_request()
609 if ((conf->mddev->recovery_cp < this_sector + sectors) || in read_balance()
610 (mddev_is_clustered(conf->mddev) && in read_balance()
611 md_cluster_ops->area_resyncing(conf->mddev, READ, this_sector, in read_balance()
772 static int raid1_congested(struct mddev *mddev, int bits) in raid1_congested() argument
774 struct r1conf *conf = mddev->private; in raid1_congested()
805 md_bitmap_unplug(conf->mddev->bitmap); in flush_bio_list()
915 test_bit(MD_RECOVERY_INTR, &conf->mddev->recovery), in raise_barrier()
918 if (test_bit(MD_RECOVERY_INTR, &conf->mddev->recovery)) { in raise_barrier()
1094 raid1_log(conf->mddev, "wait freeze"); in freeze_array()
1119 behind_bio = bio_alloc_mddev(GFP_NOIO, vcnt, r1_bio->mddev); in alloc_behind_master_bio()
1169 struct mddev *mddev = plug->cb.data; in raid1_unplug() local
1170 struct r1conf *conf = mddev->private; in raid1_unplug()
1179 md_wakeup_thread(mddev->thread); in raid1_unplug()
1190 static void init_r1bio(struct r1bio *r1_bio, struct mddev *mddev, struct bio *bio) in init_r1bio() argument
1195 r1_bio->mddev = mddev; in init_r1bio()
1200 alloc_r1bio(struct mddev *mddev, struct bio *bio) in alloc_r1bio() argument
1202 struct r1conf *conf = mddev->private; in alloc_r1bio()
1208 init_r1bio(r1_bio, mddev, bio); in alloc_r1bio()
1212 static void raid1_read_request(struct mddev *mddev, struct bio *bio, in raid1_read_request() argument
1215 struct r1conf *conf = mddev->private; in raid1_read_request()
1218 struct bitmap *bitmap = mddev->bitmap; in raid1_read_request()
1252 r1_bio = alloc_r1bio(mddev, bio); in raid1_read_request()
1254 init_r1bio(r1_bio, mddev, bio); in raid1_read_request()
1267 mdname(mddev), in raid1_read_request()
1278 mdname(mddev), in raid1_read_request()
1288 raid1_log(mddev, "wait behind writes"); in raid1_read_request()
1305 read_bio = bio_clone_fast(bio, gfp, &mddev->bio_set); in raid1_read_request()
1319 if (mddev->gendisk) in raid1_read_request()
1321 disk_devt(mddev->gendisk), r1_bio->sector); in raid1_read_request()
1326 static void raid1_write_request(struct mddev *mddev, struct bio *bio, in raid1_write_request() argument
1329 struct r1conf *conf = mddev->private; in raid1_write_request()
1332 struct bitmap *bitmap = mddev->bitmap; in raid1_write_request()
1340 if (mddev_is_clustered(mddev) && in raid1_write_request()
1341 md_cluster_ops->area_resyncing(mddev, WRITE, in raid1_write_request()
1348 if (!md_cluster_ops->area_resyncing(mddev, WRITE, in raid1_write_request()
1364 r1_bio = alloc_r1bio(mddev, bio); in raid1_write_request()
1368 md_wakeup_thread(mddev->thread); in raid1_write_request()
1369 raid1_log(mddev, "wait queued"); in raid1_write_request()
1426 rdev_dec_pending(rdev, mddev); in raid1_write_request()
1455 rdev_dec_pending(conf->mirrors[j].rdev, mddev); in raid1_write_request()
1458 raid1_log(mddev, "wait rdev %d blocked", blocked_rdev->raid_disk); in raid1_write_request()
1459 md_wait_for_blocked_rdev(blocked_rdev, mddev); in raid1_write_request()
1491 < mddev->bitmap_info.max_write_behind) && in raid1_write_request()
1503 GFP_NOIO, &mddev->bio_set); in raid1_write_request()
1505 mbio = bio_clone_fast(bio, GFP_NOIO, &mddev->bio_set); in raid1_write_request()
1530 conf->raid_disks - mddev->degraded > 1) in raid1_write_request()
1536 if (mddev->gendisk) in raid1_write_request()
1538 mbio, disk_devt(mddev->gendisk), in raid1_write_request()
1543 cb = blk_check_plugged(raid1_unplug, mddev, sizeof(*plug)); in raid1_write_request()
1556 md_wakeup_thread(mddev->thread); in raid1_write_request()
1566 static bool raid1_make_request(struct mddev *mddev, struct bio *bio) in raid1_make_request() argument
1571 && md_flush_request(mddev, bio)) in raid1_make_request()
1585 raid1_read_request(mddev, bio, sectors, NULL); in raid1_make_request()
1587 if (!md_write_start(mddev,bio)) in raid1_make_request()
1589 raid1_write_request(mddev, bio, sectors); in raid1_make_request()
1594 static void raid1_status(struct seq_file *seq, struct mddev *mddev) in raid1_status() argument
1596 struct r1conf *conf = mddev->private; in raid1_status()
1600 conf->raid_disks - mddev->degraded); in raid1_status()
1611 static void raid1_error(struct mddev *mddev, struct md_rdev *rdev) in raid1_error() argument
1614 struct r1conf *conf = mddev->private; in raid1_error()
1624 if (test_bit(In_sync, &rdev->flags) && !mddev->fail_last_dev in raid1_error()
1625 && (conf->raid_disks - mddev->degraded) == 1) { in raid1_error()
1632 conf->recovery_disabled = mddev->recovery_disabled; in raid1_error()
1638 mddev->degraded++; in raid1_error()
1644 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in raid1_error()
1645 set_mask_bits(&mddev->sb_flags, 0, in raid1_error()
1649 mdname(mddev), bdevname(rdev->bdev, b), in raid1_error()
1650 mdname(mddev), conf->raid_disks - mddev->degraded); in raid1_error()
1662 pr_debug(" --- wd:%d rd:%d\n", conf->raid_disks - conf->mddev->degraded, in print_conf()
1690 static int raid1_spare_active(struct mddev *mddev) in raid1_spare_active() argument
1693 struct r1conf *conf = mddev->private; in raid1_spare_active()
1735 mddev->degraded -= count; in raid1_spare_active()
1742 static int raid1_add_disk(struct mddev *mddev, struct md_rdev *rdev) in raid1_add_disk() argument
1744 struct r1conf *conf = mddev->private; in raid1_add_disk()
1751 if (mddev->recovery_disabled == conf->recovery_disabled) in raid1_add_disk()
1754 if (md_integrity_add_rdev(rdev, mddev)) in raid1_add_disk()
1773 if (mddev->gendisk) in raid1_add_disk()
1774 disk_stack_limits(mddev->gendisk, rdev->bdev, in raid1_add_disk()
1800 if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid1_add_disk()
1801 blk_queue_flag_set(QUEUE_FLAG_DISCARD, mddev->queue); in raid1_add_disk()
1806 static int raid1_remove_disk(struct mddev *mddev, struct md_rdev *rdev) in raid1_remove_disk() argument
1808 struct r1conf *conf = mddev->private; in raid1_remove_disk()
1827 mddev->recovery_disabled != conf->recovery_disabled && in raid1_remove_disk()
1828 mddev->degraded < conf->raid_disks) { in raid1_remove_disk()
1868 err = md_integrity_register(mddev); in raid1_remove_disk()
1894 static void abort_sync_write(struct mddev *mddev, struct r1bio *r1_bio) in abort_sync_write() argument
1902 md_bitmap_end_sync(mddev->bitmap, s, &sync_blocks, 1); in abort_sync_write()
1911 struct mddev *mddev = r1_bio->mddev; in put_sync_write_buf() local
1919 md_done_sync(mddev, s, uptodate); in put_sync_write_buf()
1928 struct mddev *mddev = r1_bio->mddev; in end_sync_write() local
1929 struct r1conf *conf = mddev->private; in end_sync_write()
1935 abort_sync_write(mddev, r1_bio); in end_sync_write()
1939 mddev->recovery); in end_sync_write()
1964 rdev->mddev->recovery); in r1_sync_page_io()
1968 md_error(rdev->mddev, rdev); in r1_sync_page_io()
1985 struct mddev *mddev = r1_bio->mddev; in fix_sync_read_error() local
1986 struct r1conf *conf = mddev->private; in fix_sync_read_error()
1998 md_error(mddev, rdev); in fix_sync_read_error()
2042 mdname(mddev), bio_devname(bio, b), in fix_sync_read_error()
2053 mddev->recovery_disabled; in fix_sync_read_error()
2054 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in fix_sync_read_error()
2055 md_done_sync(mddev, r1_bio->sectors, 0); in fix_sync_read_error()
2079 rdev_dec_pending(rdev, mddev); in fix_sync_read_error()
2113 struct mddev *mddev = r1_bio->mddev; in process_checks() local
2114 struct r1conf *conf = mddev->private; in process_checks()
2145 rdev_dec_pending(conf->mirrors[primary].rdev, mddev); in process_checks()
2178 atomic64_add(r1_bio->sectors, &mddev->resync_mismatches); in process_checks()
2179 if (j < 0 || (test_bit(MD_RECOVERY_CHECK, &mddev->recovery) in process_checks()
2183 rdev_dec_pending(conf->mirrors[i].rdev, mddev); in process_checks()
2191 static void sync_request_write(struct mddev *mddev, struct r1bio *r1_bio) in sync_request_write() argument
2193 struct r1conf *conf = mddev->private; in sync_request_write()
2203 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) in sync_request_write()
2215 !test_bit(MD_RECOVERY_SYNC, &mddev->recovery)))) in sync_request_write()
2218 abort_sync_write(mddev, r1_bio); in sync_request_write()
2247 struct mddev *mddev = conf->mddev; in fix_read_error() local
2275 rdev_dec_pending(rdev, mddev); in fix_read_error()
2289 md_error(mddev, rdev); in fix_read_error()
2306 rdev_dec_pending(rdev, mddev); in fix_read_error()
2326 mdname(mddev), s, in fix_read_error()
2331 rdev_dec_pending(rdev, mddev); in fix_read_error()
2342 struct mddev *mddev = r1_bio->mddev; in narrow_write_error() local
2343 struct r1conf *conf = mddev->private; in narrow_write_error()
2382 &mddev->bio_set); in narrow_write_error()
2385 &mddev->bio_set); in narrow_write_error()
2426 md_error(conf->mddev, rdev); in handle_sync_write_finished()
2430 md_done_sync(conf->mddev, s, 1); in handle_sync_write_finished()
2444 rdev_dec_pending(rdev, conf->mddev); in handle_write_finished()
2452 md_error(conf->mddev, in handle_write_finished()
2458 conf->mddev); in handle_write_finished()
2471 md_wakeup_thread(conf->mddev->thread); in handle_write_finished()
2481 struct mddev *mddev = conf->mddev; in handle_read_error() local
2500 if (mddev->ro == 0 in handle_read_error()
2506 } else if (mddev->ro == 0 && test_bit(FailFast, &rdev->flags)) { in handle_read_error()
2507 md_error(mddev, rdev); in handle_read_error()
2512 rdev_dec_pending(rdev, conf->mddev); in handle_read_error()
2518 raid1_read_request(mddev, bio, r1_bio->sectors, r1_bio); in handle_read_error()
2523 struct mddev *mddev = thread->mddev; in raid1d() local
2526 struct r1conf *conf = mddev->private; in raid1d()
2531 md_check_recovery(mddev); in raid1d()
2534 !test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) { in raid1d()
2537 if (!test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) in raid1d()
2546 if (mddev->degraded) in raid1d()
2570 mddev = r1_bio->mddev; in raid1d()
2571 conf = mddev->private; in raid1d()
2577 sync_request_write(mddev, r1_bio); in raid1d()
2587 if (mddev->sb_flags & ~(1<<MD_SB_CHANGE_PENDING)) in raid1d()
2588 md_check_recovery(mddev); in raid1d()
2631 static sector_t raid1_sync_request(struct mddev *mddev, sector_t sector_nr, in raid1_sync_request() argument
2634 struct r1conf *conf = mddev->private; in raid1_sync_request()
2653 max_sector = mddev->dev_sectors; in raid1_sync_request()
2660 if (mddev->curr_resync < max_sector) /* aborted */ in raid1_sync_request()
2661 md_bitmap_end_sync(mddev->bitmap, mddev->curr_resync, in raid1_sync_request()
2666 md_bitmap_close_sync(mddev->bitmap); in raid1_sync_request()
2669 if (mddev_is_clustered(mddev)) { in raid1_sync_request()
2676 if (mddev->bitmap == NULL && in raid1_sync_request()
2677 mddev->recovery_cp == MaxSector && in raid1_sync_request()
2678 !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery) && in raid1_sync_request()
2686 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, 1) && in raid1_sync_request()
2687 !conf->fullsync && !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) { in raid1_sync_request()
2704 md_bitmap_cond_end_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2705 mddev_is_clustered(mddev) && (sector_nr + 2 * RESYNC_SECTORS > conf->cluster_sync_high)); in raid1_sync_request()
2723 r1_bio->mddev = mddev; in raid1_sync_request()
2771 test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && in raid1_sync_request()
2772 !test_bit(MD_RECOVERY_CHECK, &mddev->recovery)) { in raid1_sync_request()
2809 set_bit(MD_SB_CHANGE_DEVS, &mddev->sb_flags); in raid1_sync_request()
2819 conf->recovery_disabled = mddev->recovery_disabled; in raid1_sync_request()
2820 set_bit(MD_RECOVERY_INTR, &mddev->recovery); in raid1_sync_request()
2832 if (test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && read_targets > 0) in raid1_sync_request()
2849 if (max_sector > mddev->resync_max) in raid1_sync_request()
2850 max_sector = mddev->resync_max; /* Don't do IO beyond here */ in raid1_sync_request()
2863 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2866 !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) in raid1_sync_request()
2894 if (mddev_is_clustered(mddev) && in raid1_sync_request()
2896 conf->cluster_sync_low = mddev->curr_resync_completed; in raid1_sync_request()
2899 md_cluster_ops->resync_info_update(mddev, in raid1_sync_request()
2907 if (test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) { in raid1_sync_request()
2930 static sector_t raid1_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid1_size() argument
2935 return mddev->dev_sectors; in raid1_size()
2938 static struct r1conf *setup_conf(struct mddev *mddev) in setup_conf() argument
2971 mddev->raid_disks, 2), in setup_conf()
2983 conf->poolinfo->raid_disks = mddev->raid_disks * 2; in setup_conf()
2993 conf->poolinfo->mddev = mddev; in setup_conf()
2997 rdev_for_each(rdev, mddev) { in setup_conf()
2999 if (disk_idx >= mddev->raid_disks in setup_conf()
3003 disk = conf->mirrors + mddev->raid_disks + disk_idx; in setup_conf()
3013 conf->raid_disks = mddev->raid_disks; in setup_conf()
3014 conf->mddev = mddev; in setup_conf()
3023 conf->recovery_disabled = mddev->recovery_disabled - 1; in setup_conf()
3055 conf->thread = md_register_thread(raid1d, mddev, "raid1"); in setup_conf()
3077 static void raid1_free(struct mddev *mddev, void *priv);
3078 static int raid1_run(struct mddev *mddev) in raid1_run() argument
3086 if (mddev->level != 1) { in raid1_run()
3088 mdname(mddev), mddev->level); in raid1_run()
3091 if (mddev->reshape_position != MaxSector) { in raid1_run()
3093 mdname(mddev)); in raid1_run()
3096 if (mddev_init_writes_pending(mddev) < 0) in raid1_run()
3103 if (mddev->private == NULL) in raid1_run()
3104 conf = setup_conf(mddev); in raid1_run()
3106 conf = mddev->private; in raid1_run()
3111 if (mddev->queue) { in raid1_run()
3112 blk_queue_max_write_same_sectors(mddev->queue, 0); in raid1_run()
3113 blk_queue_max_write_zeroes_sectors(mddev->queue, 0); in raid1_run()
3116 rdev_for_each(rdev, mddev) { in raid1_run()
3117 if (!mddev->gendisk) in raid1_run()
3119 disk_stack_limits(mddev->gendisk, rdev->bdev, in raid1_run()
3125 mddev->degraded = 0; in raid1_run()
3130 mddev->degraded++; in raid1_run()
3134 if (conf->raid_disks - mddev->degraded < 1) { in raid1_run()
3139 if (conf->raid_disks - mddev->degraded == 1) in raid1_run()
3140 mddev->recovery_cp = MaxSector; in raid1_run()
3142 if (mddev->recovery_cp != MaxSector) in raid1_run()
3144 mdname(mddev)); in raid1_run()
3146 mdname(mddev), mddev->raid_disks - mddev->degraded, in raid1_run()
3147 mddev->raid_disks); in raid1_run()
3152 mddev->thread = conf->thread; in raid1_run()
3154 mddev->private = conf; in raid1_run()
3155 set_bit(MD_FAILFAST_SUPPORTED, &mddev->flags); in raid1_run()
3157 md_set_array_sectors(mddev, raid1_size(mddev, 0, 0)); in raid1_run()
3159 if (mddev->queue) { in raid1_run()
3162 mddev->queue); in raid1_run()
3165 mddev->queue); in raid1_run()
3168 ret = md_integrity_register(mddev); in raid1_run()
3170 md_unregister_thread(&mddev->thread); in raid1_run()
3176 raid1_free(mddev, conf); in raid1_run()
3180 static void raid1_free(struct mddev *mddev, void *priv) in raid1_free() argument
3196 static int raid1_resize(struct mddev *mddev, sector_t sectors) in raid1_resize() argument
3205 sector_t newsize = raid1_size(mddev, sectors, 0); in raid1_resize()
3206 if (mddev->external_size && in raid1_resize()
3207 mddev->array_sectors > newsize) in raid1_resize()
3209 if (mddev->bitmap) { in raid1_resize()
3210 int ret = md_bitmap_resize(mddev->bitmap, newsize, 0, 0); in raid1_resize()
3214 md_set_array_sectors(mddev, newsize); in raid1_resize()
3215 if (sectors > mddev->dev_sectors && in raid1_resize()
3216 mddev->recovery_cp > mddev->dev_sectors) { in raid1_resize()
3217 mddev->recovery_cp = mddev->dev_sectors; in raid1_resize()
3218 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in raid1_resize()
3220 mddev->dev_sectors = sectors; in raid1_resize()
3221 mddev->resync_max_sectors = sectors; in raid1_resize()
3225 static int raid1_reshape(struct mddev *mddev) in raid1_reshape() argument
3241 struct r1conf *conf = mddev->private; in raid1_reshape()
3251 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape()
3252 mddev->layout != mddev->new_layout || in raid1_reshape()
3253 mddev->level != mddev->new_level) { in raid1_reshape()
3254 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
3255 mddev->new_layout = mddev->layout; in raid1_reshape()
3256 mddev->new_level = mddev->level; in raid1_reshape()
3260 if (!mddev_is_clustered(mddev)) in raid1_reshape()
3261 md_allow_write(mddev); in raid1_reshape()
3263 raid_disks = mddev->raid_disks + mddev->delta_disks; in raid1_reshape()
3277 newpoolinfo->mddev = mddev; in raid1_reshape()
3304 sysfs_unlink_rdev(mddev, rdev); in raid1_reshape()
3306 sysfs_unlink_rdev(mddev, rdev); in raid1_reshape()
3307 if (sysfs_link_rdev(mddev, rdev)) in raid1_reshape()
3309 mdname(mddev), rdev->raid_disk); in raid1_reshape()
3320 mddev->degraded += (raid_disks - conf->raid_disks); in raid1_reshape()
3322 conf->raid_disks = mddev->raid_disks = raid_disks; in raid1_reshape()
3323 mddev->delta_disks = 0; in raid1_reshape()
3327 set_bit(MD_RECOVERY_RECOVER, &mddev->recovery); in raid1_reshape()
3328 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in raid1_reshape()
3329 md_wakeup_thread(mddev->thread); in raid1_reshape()
3335 static void raid1_quiesce(struct mddev *mddev, int quiesce) in raid1_quiesce() argument
3337 struct r1conf *conf = mddev->private; in raid1_quiesce()
3345 static void *raid1_takeover(struct mddev *mddev) in raid1_takeover() argument
3350 if (mddev->level == 5 && mddev->raid_disks == 2) { in raid1_takeover()
3352 mddev->new_level = 1; in raid1_takeover()
3353 mddev->new_layout = 0; in raid1_takeover()
3354 mddev->new_chunk_sectors = 0; in raid1_takeover()
3355 conf = setup_conf(mddev); in raid1_takeover()
3359 mddev_clear_unsupported_flags(mddev, in raid1_takeover()