Home
last modified time | relevance | path

Searched refs:mddev (Results 1 – 19 of 19) sorted by relevance

/drivers/md/
Dmd.c80 static int remove_and_add_spares(struct mddev *mddev,
82 static void mddev_detach(struct mddev *mddev);
105 static inline int speed_min(struct mddev *mddev) in speed_min() argument
107 return mddev->sync_speed_min ? in speed_min()
108 mddev->sync_speed_min : sysctl_speed_limit_min; in speed_min()
111 static inline int speed_max(struct mddev *mddev) in speed_max() argument
113 return mddev->sync_speed_max ? in speed_max()
114 mddev->sync_speed_max : sysctl_speed_limit_max; in speed_max()
166 struct mddev *mddev) in bio_alloc_mddev() argument
170 if (!mddev || !mddev->bio_set) in bio_alloc_mddev()
[all …]
Dmd.h39 struct mddev *mddev; /* RAID array if running */ member
192 struct mddev { struct
440 void (*sync_super)(struct mddev *mddev, struct md_rdev *rdev); argument
445 static inline int __must_check mddev_lock(struct mddev *mddev) in mddev_lock() argument
447 return mutex_lock_interruptible(&mddev->reconfig_mutex); in mddev_lock()
453 static inline void mddev_lock_nointr(struct mddev *mddev) in mddev_lock_nointr() argument
455 mutex_lock(&mddev->reconfig_mutex); in mddev_lock_nointr()
458 static inline int mddev_is_locked(struct mddev *mddev) in mddev_is_locked() argument
460 return mutex_is_locked(&mddev->reconfig_mutex); in mddev_is_locked()
463 static inline int mddev_trylock(struct mddev *mddev) in mddev_trylock() argument
[all …]
Draid0.c28 static int raid0_congested(struct mddev *mddev, int bits) in raid0_congested() argument
30 struct r0conf *conf = mddev->private; in raid0_congested()
46 static void dump_zones(struct mddev *mddev) in dump_zones() argument
52 struct r0conf *conf = mddev->private; in dump_zones()
55 mdname(mddev), in dump_zones()
75 static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) in create_strip_zones() argument
90 rdev_for_each(rdev1, mddev) { in create_strip_zones()
92 mdname(mddev), in create_strip_zones()
98 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones()
99 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones()
[all …]
Dmd-cluster.h8 struct mddev;
12 int (*join)(struct mddev *mddev, int nodes);
13 int (*leave)(struct mddev *mddev);
14 int (*slot_number)(struct mddev *mddev);
15 int (*resync_info_update)(struct mddev *mddev, sector_t lo, sector_t hi);
16 int (*metadata_update_start)(struct mddev *mddev);
17 int (*metadata_update_finish)(struct mddev *mddev);
18 void (*metadata_update_cancel)(struct mddev *mddev);
19 int (*resync_start)(struct mddev *mddev);
20 int (*resync_finish)(struct mddev *mddev);
[all …]
Dlinear.c30 static inline struct dev_info *which_dev(struct mddev *mddev, sector_t sector) in which_dev() argument
36 hi = mddev->raid_disks - 1; in which_dev()
37 conf = mddev->private; in which_dev()
61 static int linear_congested(struct mddev *mddev, int bits) in linear_congested() argument
67 conf = rcu_dereference(mddev->private); in linear_congested()
78 static sector_t linear_size(struct mddev *mddev, sector_t sectors, int raid_disks) in linear_size() argument
83 conf = mddev->private; in linear_size()
91 static struct linear_conf *linear_conf(struct mddev *mddev, int raid_disks) in linear_conf() argument
106 rdev_for_each(rdev, mddev) { in linear_conf()
113 mdname(mddev)); in linear_conf()
[all …]
Dmd-cluster.c32 struct mddev *mddev; /* pointing back to mddev. */ member
153 struct mddev *mddev) in dlm_lock_sync_interruptible() argument
165 || test_bit(MD_CLOSING, &mddev->flags)); in dlm_lock_sync_interruptible()
186 static struct dlm_lock_resource *lockres_init(struct mddev *mddev, in lockres_init() argument
191 struct md_cluster_info *cinfo = mddev->cluster_info; in lockres_init()
199 res->mddev = mddev; in lockres_init()
271 static struct suspend_info *read_resync_info(struct mddev *mddev, struct dlm_lock_resource *lockres) in read_resync_info() argument
294 struct mddev *mddev = thread->mddev; in recover_bitmaps() local
295 struct md_cluster_info *cinfo = mddev->cluster_info; in recover_bitmaps()
315 bm_lockres = lockres_init(mddev, str, NULL, 1); in recover_bitmaps()
[all …]
Dmultipath.c62 struct mddev *mddev = mp_bh->mddev; in multipath_reschedule_retry() local
63 struct mpconf *conf = mddev->private; in multipath_reschedule_retry()
68 md_wakeup_thread(mddev->thread); in multipath_reschedule_retry()
79 struct mpconf *conf = mp_bh->mddev->private; in multipath_end_bh_io()
89 struct mpconf *conf = mp_bh->mddev->private; in multipath_end_request()
99 md_error (mp_bh->mddev, rdev); in multipath_end_request()
106 rdev_dec_pending(rdev, conf->mddev); in multipath_end_request()
109 static void multipath_make_request(struct mddev *mddev, struct bio * bio) in multipath_make_request() argument
111 struct mpconf *conf = mddev->private; in multipath_make_request()
116 md_flush_request(mddev, bio); in multipath_make_request()
[all …]
Ddm-raid.c223 struct mddev md;
232 struct mddev *mddev = &rs->md; in rs_config_backup() local
234 l->new_level = mddev->new_level; in rs_config_backup()
235 l->new_layout = mddev->new_layout; in rs_config_backup()
236 l->new_chunk_sectors = mddev->new_chunk_sectors; in rs_config_backup()
241 struct mddev *mddev = &rs->md; in rs_config_restore() local
243 mddev->new_level = l->new_level; in rs_config_restore()
244 mddev->new_layout = l->new_layout; in rs_config_restore()
245 mddev->new_chunk_sectors = l->new_chunk_sectors; in rs_config_restore()
623 struct mddev *mddev = &rs->md; in rs_set_capacity() local
[all …]
Dbitmap.c35 return bitmap->mddev ? mdname(bitmap->mddev) : "mdX"; in bmname()
147 static int read_sb_page(struct mddev *mddev, loff_t offset, in read_sb_page() argument
156 rdev_for_each(rdev, mddev) { in read_sb_page()
173 static struct md_rdev *next_active_rdev(struct md_rdev *rdev, struct mddev *mddev) in next_active_rdev() argument
191 rdev = list_entry(&mddev->disks, struct md_rdev, same_set); in next_active_rdev()
194 rdev_dec_pending(rdev, mddev); in next_active_rdev()
196 list_for_each_entry_continue_rcu(rdev, &mddev->disks, same_set) { in next_active_rdev()
213 struct mddev *mddev = bitmap->mddev; in write_sb_page() local
216 while ((rdev = next_active_rdev(rdev, mddev)) != NULL) { in write_sb_page()
218 loff_t offset = mddev->bitmap_info.offset; in write_sb_page()
[all …]
Draid1.c124 if (test_bit(MD_RECOVERY_REQUESTED, &pi->mddev->recovery)) in r1buf_pool_alloc()
136 if (!test_bit(MD_RECOVERY_REQUESTED, &pi->mddev->recovery)) { in r1buf_pool_alloc()
191 struct r1conf *conf = r1_bio->mddev->private; in free_r1bio()
199 struct r1conf *conf = r1_bio->mddev->private; in put_buf()
205 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev); in put_buf()
216 struct mddev *mddev = r1_bio->mddev; in reschedule_retry() local
217 struct r1conf *conf = mddev->private; in reschedule_retry()
225 md_wakeup_thread(mddev->thread); in reschedule_retry()
237 struct r1conf *conf = r1_bio->mddev->private; in call_bio_endio()
289 struct r1conf *conf = r1_bio->mddev->private; in update_head_pos()
[all …]
Draid10.c102 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr,
104 static void reshape_request_write(struct mddev *mddev, struct r10bio *r10_bio);
151 if (test_bit(MD_RECOVERY_SYNC, &conf->mddev->recovery) || in r10buf_pool_alloc()
152 test_bit(MD_RECOVERY_RESHAPE, &conf->mddev->recovery)) in r10buf_pool_alloc()
181 &conf->mddev->recovery)) { in r10buf_pool_alloc()
259 struct r10conf *conf = r10_bio->mddev->private; in free_r10bio()
267 struct r10conf *conf = r10_bio->mddev->private; in put_buf()
277 struct mddev *mddev = r10_bio->mddev; in reschedule_retry() local
278 struct r10conf *conf = mddev->private; in reschedule_retry()
288 md_wakeup_thread(mddev->thread); in reschedule_retry()
[all …]
Draid5.c266 md_wakeup_thread(conf->mddev->thread); in raid5_wakeup_stripe_thread()
310 md_wakeup_thread(conf->mddev->thread); in do_release_stripe()
316 md_wakeup_thread(conf->mddev->thread); in do_release_stripe()
375 md_wakeup_thread(conf->mddev->thread); in release_inactive_stripe_list()
423 if (unlikely(!conf->mddev->thread) || in raid5_release_stripe()
428 md_wakeup_thread(conf->mddev->thread); in raid5_release_stripe()
646 if (conf->mddev->reshape_position == MaxSector) in has_failed()
647 return conf->mddev->degraded > conf->max_degraded; in has_failed()
855 md_wakeup_thread(conf->mddev->thread); in stripe_add_to_batch_list()
982 if (!conf->mddev->external && in ops_run_io()
[all …]
Dfaulty.c173 static void faulty_make_request(struct mddev *mddev, struct bio *bio) in faulty_make_request() argument
175 struct faulty_conf *conf = mddev->private; in faulty_make_request()
217 struct bio *b = bio_clone_mddev(bio, GFP_NOIO, mddev); in faulty_make_request()
229 static void faulty_status(struct seq_file *seq, struct mddev *mddev) in faulty_status() argument
231 struct faulty_conf *conf = mddev->private; in faulty_status()
262 static int faulty_reshape(struct mddev *mddev) in faulty_reshape() argument
264 int mode = mddev->new_layout & ModeMask; in faulty_reshape()
265 int count = mddev->new_layout >> ModeShift; in faulty_reshape()
266 struct faulty_conf *conf = mddev->private; in faulty_reshape()
268 if (mddev->new_layout < 0) in faulty_reshape()
[all …]
Dbitmap.h194 struct mddev *mddev; /* the md device that the bitmap is for */ member
238 struct bitmap *bitmap_create(struct mddev *mddev, int slot);
239 int bitmap_load(struct mddev *mddev);
240 void bitmap_flush(struct mddev *mddev);
241 void bitmap_destroy(struct mddev *mddev);
261 void bitmap_sync_with_cluster(struct mddev *mddev,
266 void bitmap_daemon_work(struct mddev *mddev);
270 int bitmap_copy_from_slot(struct mddev *mddev, int slot,
Draid1.h27 struct mddev *mddev; member
32 struct mddev *mddev; member
141 struct mddev *mddev; member
Draid5-cache.c219 md_error(log->rdev->mddev, log->rdev); in r5l_log_endio()
233 md_wakeup_thread(log->rdev->mddev->thread); in r5l_log_endio()
643 md_error(log->rdev->mddev, log->rdev); in r5l_log_flush_endio()
697 struct mddev *mddev; in r5l_write_super_and_discard_space() local
704 mddev = log->rdev->mddev; in r5l_write_super_and_discard_space()
716 set_mask_bits(&mddev->flags, 0, in r5l_write_super_and_discard_space()
718 if (!mddev_trylock(mddev)) in r5l_write_super_and_discard_space()
720 md_update_sb(mddev, 1); in r5l_write_super_and_discard_space()
721 mddev_unlock(mddev); in r5l_write_super_and_discard_space()
761 md_wakeup_thread(log->rdev->mddev->thread); in r5l_do_reclaim()
[all …]
Dmultipath.h9 struct mddev *mddev; member
25 struct mddev *mddev; member
Draid10.h15 struct mddev *mddev; member
106 struct mddev *mddev; member
Draid5.h439 struct mddev *mddev; member
618 extern int raid5_set_cache_size(struct mddev *mddev, int size);