| /linux/drivers/md/ |
| H A D | md.c | 98 static int remove_and_add_spares(struct mddev *mddev, 100 static void mddev_detach(struct mddev *mddev); 101 static void export_rdev(struct md_rdev *rdev, struct mddev *mddev); 137 static int speed_min(struct mddev *mddev) in speed_min() argument 139 return mddev->sync_speed_min ? in speed_min() 140 mddev->sync_speed_min : sysctl_speed_limit_min; in speed_min() 143 static int speed_max(struct mddev *mddev) in speed_max() argument 145 return mddev->sync_speed_max ? in speed_max() 146 mddev->sync_speed_max : sysctl_speed_limit_max; in speed_max() 149 static int sync_io_depth(struct mddev *mddev) in sync_io_depth() argument [all …]
|
| H A D | md-cluster.h | 9 struct mddev; 15 int (*join)(struct mddev *mddev, int nodes); 16 int (*leave)(struct mddev *mddev); 17 int (*slot_number)(struct mddev *mddev); 18 int (*resync_info_update)(struct mddev *mddev, sector_t lo, sector_t hi); 19 int (*resync_start_notify)(struct mddev *mddev); 20 int (*resync_status_get)(struct mddev *mddev); 21 void (*resync_info_get)(struct mddev *mddev, sector_t *lo, sector_t *hi); 22 int (*metadata_update_start)(struct mddev *mddev); 23 int (*metadata_update_finish)(struct mddev *mddev); [all …]
|
| H A D | raid0.c | 37 static void dump_zones(struct mddev *mddev) in dump_zones() argument 42 struct r0conf *conf = mddev->private; in dump_zones() 45 mdname(mddev), in dump_zones() 65 static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) in create_strip_zones() argument 75 if (!mddev_is_dm(mddev)) in create_strip_zones() 76 blksize = queue_logical_block_size(mddev->gendisk->queue); in create_strip_zones() 81 rdev_for_each(rdev1, mddev) { in create_strip_zones() 83 mdname(mddev), in create_strip_zones() 89 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 90 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones() [all …]
|
| H A D | md-cluster.c | 28 struct mddev *mddev; /* pointing back to mddev. */ member 63 struct mddev *mddev; /* the md device which md_cluster_info belongs to */ member 159 struct mddev *mddev) in dlm_lock_sync_interruptible() argument 171 || test_bit(MD_CLOSING, &mddev->flags)); in dlm_lock_sync_interruptible() 192 static struct dlm_lock_resource *lockres_init(struct mddev *mddev, in lockres_init() argument 197 struct md_cluster_info *cinfo = mddev->cluster_info; in lockres_init() 205 res->mddev = mddev; in lockres_init() 277 static int read_resync_info(struct mddev *mddev, in read_resync_info() argument 281 struct md_cluster_info *cinfo = mddev->cluster_info; in read_resync_info() 297 struct mddev *mddev = thread->mddev; in recover_bitmaps() local [all …]
|
| H A D | md-linear.c | 30 static inline struct dev_info *which_dev(struct mddev *mddev, sector_t sector) in which_dev() argument 36 hi = mddev->raid_disks - 1; in which_dev() 37 conf = mddev->private; in which_dev() 55 static sector_t linear_size(struct mddev *mddev, sector_t sectors, int raid_disks) in linear_size() argument 60 conf = mddev->private; in linear_size() 68 static int linear_set_limits(struct mddev *mddev) in linear_set_limits() argument 74 lim.max_hw_sectors = mddev->chunk_sectors; in linear_set_limits() 75 lim.logical_block_size = mddev->logical_block_size; in linear_set_limits() 76 lim.max_write_zeroes_sectors = mddev->chunk_sectors; in linear_set_limits() 77 lim.max_hw_wzeroes_unmap_sectors = mddev->chunk_sectors; in linear_set_limits() [all …]
|
| H A D | md-llbitmap.c | 257 struct mddev *mddev; member 350 static void __llbitmap_flush(struct mddev *mddev); 369 bool level_456 = raid_is_456(llbitmap->mddev); in llbitmap_infect_dirty_bits() 401 if (llbitmap->mddev->degraded) { in llbitmap_set_page_dirty() 446 struct mddev *mddev = llbitmap->mddev; in llbitmap_read_page() local 459 rdev_for_each(rdev, mddev) { in llbitmap_read_page() 465 sector = mddev->bitmap_info.offset + in llbitmap_read_page() 472 md_error(mddev, rdev); in llbitmap_read_page() 482 struct mddev *mddev = llbitmap->mddev; in llbitmap_write_page() local 492 rdev_for_each(rdev, mddev) { in llbitmap_write_page() [all …]
|
| H A D | md-bitmap.c | 167 struct mddev *mddev; /* the md device that the bitmap is for */ member 225 return bitmap->mddev ? mdname(bitmap->mddev) : "mdX"; in bmname() 347 static int read_sb_page(struct mddev *mddev, loff_t offset, in read_sb_page() argument 351 sector_t sector = mddev->bitmap_info.offset + offset + in read_sb_page() 355 rdev_for_each(rdev, mddev) { in read_sb_page() 369 static struct md_rdev *next_active_rdev(struct md_rdev *rdev, struct mddev *mddev) in next_active_rdev() argument 387 rdev = list_entry(&mddev->disks, struct md_rdev, same_set); in next_active_rdev() 390 rdev_dec_pending(rdev, mddev); in next_active_rdev() 392 list_for_each_entry_continue_rcu(rdev, &mddev->disks, same_set) { in next_active_rdev() 431 struct mddev *mddev = bitmap->mddev; in __write_sb_page() local [all …]
|
| H A D | raid10.c | 74 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr, 76 static void reshape_request_write(struct mddev *mddev, struct r10bio *r10_bio); 144 if (test_bit(MD_RECOVERY_SYNC, &conf->mddev->recovery) || in r10buf_pool_alloc() 145 test_bit(MD_RECOVERY_RESHAPE, &conf->mddev->recovery)) in r10buf_pool_alloc() 191 &conf->mddev->recovery)) { in r10buf_pool_alloc() 277 struct r10conf *conf = r10_bio->mddev->private; in free_r10bio() 285 struct r10conf *conf = r10_bio->mddev->private; in put_buf() 301 struct mddev *mddev = r10_bio->mddev; in reschedule_retry() local 302 struct r10conf *conf = mddev->private; in reschedule_retry() 312 md_wakeup_thread(mddev->thread); in reschedule_retry() [all …]
|
| H A D | raid1.c | 49 static void raid1_free(struct mddev *mddev, void *priv); 84 struct mddev *mddev = rdev->mddev; in wait_for_serialization() local 89 if (WARN_ON(!mddev->serial_info_pool)) in wait_for_serialization() 91 si = mempool_alloc(mddev->serial_info_pool, GFP_NOIO); in wait_for_serialization() 101 struct mddev *mddev = rdev->mddev; in remove_serial() local 110 mempool_free(si, mddev->serial_info_pool); in remove_serial() 178 if (test_bit(MD_RECOVERY_REQUESTED, &conf->mddev->recovery)) in r1buf_pool_alloc() 253 struct r1conf *conf = r1_bio->mddev->private; in free_r1bio() 261 struct r1conf *conf = r1_bio->mddev->private; in put_buf() 268 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev); in put_buf() [all …]
|
| H A D | dm-raid.c | 245 struct mddev md; 263 struct mddev *mddev = &rs->md; in rs_config_backup() local 265 l->new_level = mddev->new_level; in rs_config_backup() 266 l->new_layout = mddev->new_layout; in rs_config_backup() 267 l->new_chunk_sectors = mddev->new_chunk_sectors; in rs_config_backup() 272 struct mddev *mddev = &rs->md; in rs_config_restore() local 274 mddev->new_level = l->new_level; in rs_config_restore() 275 mddev->new_layout = l->new_layout; in rs_config_restore() 276 mddev->new_chunk_sectors = l->new_chunk_sectors; in rs_config_restore() 686 struct mddev *mddev = &rs->md; in rs_set_rdev_sectors() local [all …]
|
| H A D | raid5.c | 76 static void raid5_quiesce(struct mddev *mddev, int quiesce); 203 md_wakeup_thread(conf->mddev->thread); in raid5_wakeup_stripe_thread() 276 md_wakeup_thread(conf->mddev->thread); in do_release_stripe() 282 md_wakeup_thread(conf->mddev->thread); in do_release_stripe() 364 md_wakeup_thread(conf->mddev->thread); in release_inactive_stripe_list() 410 if (unlikely(!conf->mddev->thread) || in raid5_release_stripe() 415 md_wakeup_thread(conf->mddev->thread); in raid5_release_stripe() 751 int degraded = conf->mddev->degraded; in has_failed() 753 if (test_bit(MD_BROKEN, &conf->mddev->flags)) in has_failed() 756 if (conf->mddev->reshape_position != MaxSector) in has_failed() [all …]
|
| H A D | raid5-ppl.c | 88 struct mddev *mddev; member 408 md_error(ppl_conf->mddev, log->rdev); in ppl_log_endio() 551 struct r5conf *conf = ppl_conf->mddev->private; in ppl_io_unit_finished() 586 struct r5conf *conf = ppl_conf->mddev->private; in ppl_flush_endio() 594 rdev = md_find_rdev_rcu(conf->mddev, bio_dev(bio)); in ppl_flush_endio() 596 md_error(rdev->mddev, rdev); in ppl_flush_endio() 604 md_wakeup_thread(conf->mddev->thread); in ppl_flush_endio() 612 struct r5conf *conf = ppl_conf->mddev->private; in ppl_do_flush() 788 struct mddev *mddev = ppl_conf->mddev; in ppl_recover_entry() local 789 struct r5conf *conf = mddev->private; in ppl_recover_entry() [all …]
|
| H A D | raid5-cache.c | 301 md_write_end(conf->mddev); in r5c_return_dev_pending_writes() 417 struct r5conf *conf = log->rdev->mddev->private; in r5c_update_log_state() 567 md_error(log->rdev->mddev, log->rdev); in r5l_log_endio() 602 md_wakeup_thread(log->rdev->mddev->thread); in r5l_log_endio() 683 struct mddev *mddev = log->rdev->mddev; in r5c_disable_writeback_async() local 684 struct r5conf *conf = mddev->private; in r5c_disable_writeback_async() 689 mdname(mddev)); in r5c_disable_writeback_async() 692 wait_event(mddev->sb_wait, in r5c_disable_writeback_async() 694 !test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)); in r5c_disable_writeback_async() 698 mddev_suspend(mddev, false); in r5c_disable_writeback_async() [all …]
|
| H A D | raid5-log.h | 30 void r5c_update_on_rdev_error(struct mddev *mddev, struct md_rdev *rdev); 49 return test_bit(MD_HAS_JOURNAL, &conf->mddev->flags); in raid5_has_log() 54 return test_bit(MD_HAS_PPL, &conf->mddev->flags); in raid5_has_ppl()
|