/linux/drivers/md/ |
H A D | raid0.c | 35 static void dump_zones(struct mddev *mddev) in dump_zones() argument 40 struct r0conf *conf = mddev->private; in dump_zones() 43 mdname(mddev), in dump_zones() 63 static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) in create_strip_zones() argument 76 rdev_for_each(rdev1, mddev) { in create_strip_zones() 78 mdname(mddev), in create_strip_zones() 84 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 85 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones() 90 rdev_for_each(rdev2, mddev) { in create_strip_zones() 93 mdname(mddev), in create_strip_zones() [all …]
|
H A D | md-cluster.h | 9 struct mddev; 13 int (*join)(struct mddev *mddev, int nodes); 14 int (*leave)(struct mddev *mddev); 15 int (*slot_number)(struct mddev *mddev); 16 int (*resync_info_update)(struct mddev *mddev, sector_t lo, sector_t hi); 17 int (*resync_start_notify)(struct mddev *mddev); 18 int (*resync_status_get)(struct mddev *mddev); 19 void (*resync_info_get)(struct mddev *mddev, sector_t *lo, sector_t *hi); 20 int (*metadata_update_start)(struct mddev *mddev); 21 int (*metadata_update_finish)(struct mddev *mddev); [all …]
|
H A D | md-cluster.c | 28 struct mddev *mddev; /* pointing back to mddev. */ member 63 struct mddev *mddev; /* the md device which md_cluster_info belongs to */ member 159 struct mddev *mddev) in dlm_lock_sync_interruptible() argument 171 || test_bit(MD_CLOSING, &mddev->flags)); in dlm_lock_sync_interruptible() 192 static struct dlm_lock_resource *lockres_init(struct mddev *mddev, in lockres_init() argument 197 struct md_cluster_info *cinfo = mddev->cluster_info; in lockres_init() 205 res->mddev = mddev; in lockres_init() 277 static int read_resync_info(struct mddev *mddev, in read_resync_info() argument 281 struct md_cluster_info *cinfo = mddev->cluster_info; in read_resync_info() 297 struct mddev *mddev = thread->mddev; in recover_bitmaps() local [all …]
|
H A D | md-linear.c | 31 static inline struct dev_info *which_dev(struct mddev *mddev, sector_t sector) in which_dev() argument 37 hi = mddev->raid_disks - 1; in which_dev() 38 conf = mddev->private; in which_dev() 56 static sector_t linear_size(struct mddev *mddev, sector_t sectors, int raid_disks) in linear_size() argument 61 conf = mddev->private; in linear_size() 69 static int linear_set_limits(struct mddev *mddev) in linear_set_limits() argument 75 lim.max_hw_sectors = mddev->chunk_sectors; in linear_set_limits() 76 lim.max_write_zeroes_sectors = mddev->chunk_sectors; in linear_set_limits() 77 lim.io_min = mddev->chunk_sectors << 9; in linear_set_limits() 78 err = mddev_stack_rdev_limits(mddev, &lim, MDDEV_STACK_INTEGRITY); in linear_set_limits() [all …]
|
H A D | raid10.c | 73 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr, 75 static void reshape_request_write(struct mddev *mddev, struct r10bio *r10_bio); 143 if (test_bit(MD_RECOVERY_SYNC, &conf->mddev->recovery) || in r10buf_pool_alloc() 144 test_bit(MD_RECOVERY_RESHAPE, &conf->mddev->recovery)) in r10buf_pool_alloc() 190 &conf->mddev->recovery)) { in r10buf_pool_alloc() 276 struct r10conf *conf = r10_bio->mddev->private; in free_r10bio() 284 struct r10conf *conf = r10_bio->mddev->private; in put_buf() 300 struct mddev *mddev = r10_bio->mddev; in reschedule_retry() local 301 struct r10conf *conf = mddev->private; in reschedule_retry() 311 md_wakeup_thread(mddev->thread); in reschedule_retry() [all …]
|
H A D | raid1.c | 82 struct mddev *mddev = rdev->mddev; in wait_for_serialization() local 87 if (WARN_ON(!mddev->serial_info_pool)) in wait_for_serialization() 89 si = mempool_alloc(mddev->serial_info_pool, GFP_NOIO); in wait_for_serialization() 99 struct mddev *mddev = rdev->mddev; in remove_serial() local 108 mempool_free(si, mddev->serial_info_pool); in remove_serial() 178 if (test_bit(MD_RECOVERY_REQUESTED, &pi->mddev->recovery)) in r1buf_pool_alloc() 253 struct r1conf *conf = r1_bio->mddev->private; in free_r1bio() 261 struct r1conf *conf = r1_bio->mddev->private; in put_buf() 268 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev); in put_buf() 279 struct mddev *mddev = r1_bio->mddev; in reschedule_retry() local [all …]
|
H A D | dm-raid.c | 245 struct mddev md; 263 struct mddev *mddev = &rs->md; in rs_config_backup() local 265 l->new_level = mddev->new_level; in rs_config_backup() 266 l->new_layout = mddev->new_layout; in rs_config_backup() 267 l->new_chunk_sectors = mddev->new_chunk_sectors; in rs_config_backup() 272 struct mddev *mddev = &rs->md; in rs_config_restore() local 274 mddev->new_level = l->new_level; in rs_config_restore() 275 mddev in rs_config_restore() 686 struct mddev *mddev = &rs->md; rs_set_rdev_sectors() local 714 struct mddev *mddev = &rs->md; rs_set_cur() local 727 struct mddev *mddev = &rs->md; rs_set_new() local 1545 struct mddev *mddev = &rs->md; rs_set_raid456_stripe_cache() local 1652 struct mddev *mddev = &rs->md; rs_set_dev_and_array_sectors() local 1749 struct mddev *mddev = &rs->md; rs_check_takeover() local 1919 struct mddev *mddev = &rs->md; rs_reshape_requested() local 2047 struct mddev *mddev = &rs->md; rs_check_reshape() local 2114 super_sync(struct mddev * mddev,struct md_rdev * rdev) super_sync() argument 2244 struct mddev *mddev = &rs->md; super_init_validation() local 2463 struct mddev *mddev = &rs->md; super_validate() local 2536 struct mddev *mddev = &rs->md; analyse_superblocks() local 2739 struct mddev *mddev = &rs->md; rs_setup_takeover() local 2785 struct mddev *mddev = &rs->md; rs_prepare_reshape() local 2852 struct mddev *mddev = &rs->md; rs_setup_reshape() local 3340 struct mddev *mddev = &rs->md; raid_map() local 3378 decipher_sync_action(struct mddev * mddev,unsigned long recovery) decipher_sync_action() argument 3439 struct mddev *mddev = &rs->md; rs_get_progress() local 3526 struct mddev *mddev = &rs->md; raid_status() local 3732 struct mddev *mddev = &rs->md; raid_message() local 3821 struct mddev *mddev = &rs->md; raid_presuspend() local 3869 struct mddev *mddev = &rs->md; attempt_restore_of_faulty_devices() local 3952 struct mddev *mddev = &rs->md; __load_dirty_region_bitmap() local 3965 struct mddev *mddev = &rs->md; rs_update_sbs() local 3984 struct mddev *mddev = &rs->md; rs_start_reshape() local 4031 struct mddev *mddev = &rs->md; raid_preresume() local 4103 struct mddev *mddev = &rs->md; raid_resume() local [all...] |
H A D | raid1-10.c | 133 static inline bool raid1_add_bio_to_plug(struct mddev *mddev, struct bio *bio, in raid1_add_bio_to_plug() argument 143 if (!mddev->bitmap_ops->enabled(mddev)) { in raid1_add_bio_to_plug() 148 cb = blk_check_plugged(unplug, mddev, sizeof(*plug)); in raid1_add_bio_to_plug() 169 static inline void raid1_prepare_flush_writes(struct mddev *mddev) in raid1_prepare_flush_writes() argument 171 mddev->bitmap_ops->unplug(mddev, current->bio_list == NULL); in raid1_prepare_flush_writes() 179 static inline void check_decay_read_errors(struct mddev *mddev, struct md_rdev *rdev) in check_decay_read_errors() argument 209 static inline bool exceed_read_errors(struct mddev *mddev, struct md_rdev *rdev) in exceed_read_errors() argument 211 int max_read_errors = atomic_read(&mddev->max_corr_read_errors); in exceed_read_errors() 214 check_decay_read_errors(mddev, rdev); in exceed_read_errors() 218 mdname(mddev), rdev->bdev, read_errors, max_read_errors); in exceed_read_errors() [all …]
|
H A D | raid5-log.h | 30 void r5c_update_on_rdev_error(struct mddev *mddev, struct md_rdev *rdev); 49 return test_bit(MD_HAS_JOURNAL, &conf->mddev->flags); in raid5_has_log() 54 return test_bit(MD_HAS_PPL, &conf->mddev->flags); in raid5_has_ppl()
|