1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2001, 2002 Sistina Software (UK) Limited. 4 * Copyright (C) 2004-2008 Red Hat, Inc. All rights reserved. 5 * 6 * This file is released under the GPL. 7 */ 8 9 #include "dm-core.h" 10 #include "dm-rq.h" 11 #include "dm-uevent.h" 12 #include "dm-ima.h" 13 14 #include <linux/bio-integrity.h> 15 #include <linux/init.h> 16 #include <linux/module.h> 17 #include <linux/mutex.h> 18 #include <linux/sched/mm.h> 19 #include <linux/sched/signal.h> 20 #include <linux/blkpg.h> 21 #include <linux/bio.h> 22 #include <linux/mempool.h> 23 #include <linux/dax.h> 24 #include <linux/slab.h> 25 #include <linux/idr.h> 26 #include <linux/uio.h> 27 #include <linux/hdreg.h> 28 #include <linux/delay.h> 29 #include <linux/wait.h> 30 #include <linux/pr.h> 31 #include <linux/refcount.h> 32 #include <linux/part_stat.h> 33 #include <linux/blk-crypto.h> 34 #include <linux/blk-crypto-profile.h> 35 36 #define DM_MSG_PREFIX "core" 37 38 /* 39 * Cookies are numeric values sent with CHANGE and REMOVE 40 * uevents while resuming, removing or renaming the device. 41 */ 42 #define DM_COOKIE_ENV_VAR_NAME "DM_COOKIE" 43 #define DM_COOKIE_LENGTH 24 44 45 /* 46 * For REQ_POLLED fs bio, this flag is set if we link mapped underlying 47 * dm_io into one list, and reuse bio->bi_private as the list head. Before 48 * ending this fs bio, we will recover its ->bi_private. 49 */ 50 #define REQ_DM_POLL_LIST REQ_DRV 51 52 static const char *_name = DM_NAME; 53 54 static unsigned int major; 55 static unsigned int _major; 56 57 static DEFINE_IDR(_minor_idr); 58 59 static DEFINE_SPINLOCK(_minor_lock); 60 61 static void do_deferred_remove(struct work_struct *w); 62 63 static DECLARE_WORK(deferred_remove_work, do_deferred_remove); 64 65 static struct workqueue_struct *deferred_remove_workqueue; 66 67 atomic_t dm_global_event_nr = ATOMIC_INIT(0); 68 DECLARE_WAIT_QUEUE_HEAD(dm_global_eventq); 69 70 void dm_issue_global_event(void) 71 { 72 atomic_inc(&dm_global_event_nr); 73 wake_up(&dm_global_eventq); 74 } 75 76 DEFINE_STATIC_KEY_FALSE(stats_enabled); 77 DEFINE_STATIC_KEY_FALSE(swap_bios_enabled); 78 DEFINE_STATIC_KEY_FALSE(zoned_enabled); 79 80 /* 81 * One of these is allocated (on-stack) per original bio. 82 */ 83 struct clone_info { 84 struct dm_table *map; 85 struct bio *bio; 86 struct dm_io *io; 87 sector_t sector; 88 unsigned int sector_count; 89 bool is_abnormal_io:1; 90 bool submit_as_polled:1; 91 }; 92 93 static inline struct dm_target_io *clone_to_tio(struct bio *clone) 94 { 95 return container_of(clone, struct dm_target_io, clone); 96 } 97 98 void *dm_per_bio_data(struct bio *bio, size_t data_size) 99 { 100 if (!dm_tio_flagged(clone_to_tio(bio), DM_TIO_INSIDE_DM_IO)) 101 return (char *)bio - DM_TARGET_IO_BIO_OFFSET - data_size; 102 return (char *)bio - DM_IO_BIO_OFFSET - data_size; 103 } 104 EXPORT_SYMBOL_GPL(dm_per_bio_data); 105 106 struct bio *dm_bio_from_per_bio_data(void *data, size_t data_size) 107 { 108 struct dm_io *io = (struct dm_io *)((char *)data + data_size); 109 110 if (io->magic == DM_IO_MAGIC) 111 return (struct bio *)((char *)io + DM_IO_BIO_OFFSET); 112 BUG_ON(io->magic != DM_TIO_MAGIC); 113 return (struct bio *)((char *)io + DM_TARGET_IO_BIO_OFFSET); 114 } 115 EXPORT_SYMBOL_GPL(dm_bio_from_per_bio_data); 116 117 unsigned int dm_bio_get_target_bio_nr(const struct bio *bio) 118 { 119 return container_of(bio, struct dm_target_io, clone)->target_bio_nr; 120 } 121 EXPORT_SYMBOL_GPL(dm_bio_get_target_bio_nr); 122 123 #define MINOR_ALLOCED ((void *)-1) 124 125 #define DM_NUMA_NODE NUMA_NO_NODE 126 static int dm_numa_node = DM_NUMA_NODE; 127 128 #define DEFAULT_SWAP_BIOS (8 * 1048576 / PAGE_SIZE) 129 static int swap_bios = DEFAULT_SWAP_BIOS; 130 static int get_swap_bios(void) 131 { 132 int latch = READ_ONCE(swap_bios); 133 134 if (unlikely(latch <= 0)) 135 latch = DEFAULT_SWAP_BIOS; 136 return latch; 137 } 138 139 struct table_device { 140 struct list_head list; 141 refcount_t count; 142 struct dm_dev dm_dev; 143 }; 144 145 /* 146 * Bio-based DM's mempools' reserved IOs set by the user. 147 */ 148 #define RESERVED_BIO_BASED_IOS 16 149 static unsigned int reserved_bio_based_ios = RESERVED_BIO_BASED_IOS; 150 151 static int __dm_get_module_param_int(int *module_param, int min, int max) 152 { 153 int param = READ_ONCE(*module_param); 154 int modified_param = 0; 155 bool modified = true; 156 157 if (param < min) 158 modified_param = min; 159 else if (param > max) 160 modified_param = max; 161 else 162 modified = false; 163 164 if (modified) { 165 (void)cmpxchg(module_param, param, modified_param); 166 param = modified_param; 167 } 168 169 return param; 170 } 171 172 unsigned int __dm_get_module_param(unsigned int *module_param, unsigned int def, unsigned int max) 173 { 174 unsigned int param = READ_ONCE(*module_param); 175 unsigned int modified_param = 0; 176 177 if (!param) 178 modified_param = def; 179 else if (param > max) 180 modified_param = max; 181 182 if (modified_param) { 183 (void)cmpxchg(module_param, param, modified_param); 184 param = modified_param; 185 } 186 187 return param; 188 } 189 190 unsigned int dm_get_reserved_bio_based_ios(void) 191 { 192 return __dm_get_module_param(&reserved_bio_based_ios, 193 RESERVED_BIO_BASED_IOS, DM_RESERVED_MAX_IOS); 194 } 195 EXPORT_SYMBOL_GPL(dm_get_reserved_bio_based_ios); 196 197 static unsigned int dm_get_numa_node(void) 198 { 199 return __dm_get_module_param_int(&dm_numa_node, 200 DM_NUMA_NODE, num_online_nodes() - 1); 201 } 202 203 static int __init local_init(void) 204 { 205 int r; 206 207 r = dm_uevent_init(); 208 if (r) 209 return r; 210 211 deferred_remove_workqueue = alloc_ordered_workqueue("kdmremove", 0); 212 if (!deferred_remove_workqueue) { 213 r = -ENOMEM; 214 goto out_uevent_exit; 215 } 216 217 _major = major; 218 r = register_blkdev(_major, _name); 219 if (r < 0) 220 goto out_free_workqueue; 221 222 if (!_major) 223 _major = r; 224 225 return 0; 226 227 out_free_workqueue: 228 destroy_workqueue(deferred_remove_workqueue); 229 out_uevent_exit: 230 dm_uevent_exit(); 231 232 return r; 233 } 234 235 static void local_exit(void) 236 { 237 destroy_workqueue(deferred_remove_workqueue); 238 239 unregister_blkdev(_major, _name); 240 dm_uevent_exit(); 241 242 _major = 0; 243 244 DMINFO("cleaned up"); 245 } 246 247 static int (*_inits[])(void) __initdata = { 248 local_init, 249 dm_target_init, 250 dm_linear_init, 251 dm_stripe_init, 252 dm_io_init, 253 dm_kcopyd_init, 254 dm_interface_init, 255 dm_statistics_init, 256 }; 257 258 static void (*_exits[])(void) = { 259 local_exit, 260 dm_target_exit, 261 dm_linear_exit, 262 dm_stripe_exit, 263 dm_io_exit, 264 dm_kcopyd_exit, 265 dm_interface_exit, 266 dm_statistics_exit, 267 }; 268 269 static int __init dm_init(void) 270 { 271 const int count = ARRAY_SIZE(_inits); 272 int r, i; 273 274 #if (IS_ENABLED(CONFIG_IMA) && !IS_ENABLED(CONFIG_IMA_DISABLE_HTABLE)) 275 DMINFO("CONFIG_IMA_DISABLE_HTABLE is disabled." 276 " Duplicate IMA measurements will not be recorded in the IMA log."); 277 #endif 278 279 for (i = 0; i < count; i++) { 280 r = _inits[i](); 281 if (r) 282 goto bad; 283 } 284 285 return 0; 286 bad: 287 while (i--) 288 _exits[i](); 289 290 return r; 291 } 292 293 static void __exit dm_exit(void) 294 { 295 int i = ARRAY_SIZE(_exits); 296 297 while (i--) 298 _exits[i](); 299 300 /* 301 * Should be empty by this point. 302 */ 303 idr_destroy(&_minor_idr); 304 } 305 306 /* 307 * Block device functions 308 */ 309 int dm_deleting_md(struct mapped_device *md) 310 { 311 return test_bit(DMF_DELETING, &md->flags); 312 } 313 314 static int dm_blk_open(struct gendisk *disk, blk_mode_t mode) 315 { 316 struct mapped_device *md; 317 318 spin_lock(&_minor_lock); 319 320 md = disk->private_data; 321 if (!md) 322 goto out; 323 324 if (test_bit(DMF_FREEING, &md->flags) || 325 dm_deleting_md(md)) { 326 md = NULL; 327 goto out; 328 } 329 330 dm_get(md); 331 atomic_inc(&md->open_count); 332 out: 333 spin_unlock(&_minor_lock); 334 335 return md ? 0 : -ENXIO; 336 } 337 338 static void dm_blk_close(struct gendisk *disk) 339 { 340 struct mapped_device *md; 341 342 spin_lock(&_minor_lock); 343 344 md = disk->private_data; 345 if (WARN_ON(!md)) 346 goto out; 347 348 if (atomic_dec_and_test(&md->open_count) && 349 (test_bit(DMF_DEFERRED_REMOVE, &md->flags))) 350 queue_work(deferred_remove_workqueue, &deferred_remove_work); 351 352 dm_put(md); 353 out: 354 spin_unlock(&_minor_lock); 355 } 356 357 int dm_open_count(struct mapped_device *md) 358 { 359 return atomic_read(&md->open_count); 360 } 361 362 /* 363 * Guarantees nothing is using the device before it's deleted. 364 */ 365 int dm_lock_for_deletion(struct mapped_device *md, bool mark_deferred, bool only_deferred) 366 { 367 int r = 0; 368 369 spin_lock(&_minor_lock); 370 371 if (dm_open_count(md)) { 372 r = -EBUSY; 373 if (mark_deferred) 374 set_bit(DMF_DEFERRED_REMOVE, &md->flags); 375 } else if (only_deferred && !test_bit(DMF_DEFERRED_REMOVE, &md->flags)) 376 r = -EEXIST; 377 else 378 set_bit(DMF_DELETING, &md->flags); 379 380 spin_unlock(&_minor_lock); 381 382 return r; 383 } 384 385 int dm_cancel_deferred_remove(struct mapped_device *md) 386 { 387 int r = 0; 388 389 spin_lock(&_minor_lock); 390 391 if (test_bit(DMF_DELETING, &md->flags)) 392 r = -EBUSY; 393 else 394 clear_bit(DMF_DEFERRED_REMOVE, &md->flags); 395 396 spin_unlock(&_minor_lock); 397 398 return r; 399 } 400 401 static void do_deferred_remove(struct work_struct *w) 402 { 403 dm_deferred_remove(); 404 } 405 406 static int dm_blk_getgeo(struct gendisk *disk, struct hd_geometry *geo) 407 { 408 struct mapped_device *md = disk->private_data; 409 410 return dm_get_geometry(md, geo); 411 } 412 413 static int dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, 414 struct block_device **bdev, unsigned int cmd, 415 unsigned long arg, bool *forward) 416 { 417 struct dm_target *ti; 418 struct dm_table *map; 419 int r; 420 421 retry: 422 r = -ENOTTY; 423 map = dm_get_live_table(md, srcu_idx); 424 if (!map || !dm_table_get_size(map)) 425 return r; 426 427 /* We only support devices that have a single target */ 428 if (map->num_targets != 1) 429 return r; 430 431 ti = dm_table_get_target(map, 0); 432 if (!ti->type->prepare_ioctl) 433 return r; 434 435 if (dm_suspended_md(md)) 436 return -EAGAIN; 437 438 r = ti->type->prepare_ioctl(ti, bdev, cmd, arg, forward); 439 if (r == -ENOTCONN && *forward && !fatal_signal_pending(current)) { 440 dm_put_live_table(md, *srcu_idx); 441 fsleep(10000); 442 goto retry; 443 } 444 445 return r; 446 } 447 448 static void dm_unprepare_ioctl(struct mapped_device *md, int srcu_idx) 449 { 450 dm_put_live_table(md, srcu_idx); 451 } 452 453 static int dm_blk_ioctl(struct block_device *bdev, blk_mode_t mode, 454 unsigned int cmd, unsigned long arg) 455 { 456 struct mapped_device *md = bdev->bd_disk->private_data; 457 int r, srcu_idx; 458 bool forward = true; 459 460 r = dm_prepare_ioctl(md, &srcu_idx, &bdev, cmd, arg, &forward); 461 if (!forward || r < 0) 462 goto out; 463 464 if (r > 0) { 465 /* 466 * Target determined this ioctl is being issued against a 467 * subset of the parent bdev; require extra privileges. 468 */ 469 if (!capable(CAP_SYS_RAWIO)) { 470 DMDEBUG_LIMIT( 471 "%s: sending ioctl %x to DM device without required privilege.", 472 current->comm, cmd); 473 r = -ENOIOCTLCMD; 474 goto out; 475 } 476 } 477 478 if (!bdev->bd_disk->fops->ioctl) 479 r = -ENOTTY; 480 else 481 r = bdev->bd_disk->fops->ioctl(bdev, mode, cmd, arg); 482 out: 483 dm_unprepare_ioctl(md, srcu_idx); 484 return r; 485 } 486 487 u64 dm_start_time_ns_from_clone(struct bio *bio) 488 { 489 return jiffies_to_nsecs(clone_to_tio(bio)->io->start_time); 490 } 491 EXPORT_SYMBOL_GPL(dm_start_time_ns_from_clone); 492 493 static inline unsigned int dm_io_sectors(struct dm_io *io, struct bio *bio) 494 { 495 /* 496 * If REQ_PREFLUSH set, don't account payload, it will be 497 * submitted (and accounted) after this flush completes. 498 */ 499 if (io->requeue_flush_with_data) 500 return 0; 501 if (unlikely(dm_io_flagged(io, DM_IO_WAS_SPLIT))) 502 return io->sectors; 503 return bio_sectors(bio); 504 } 505 506 static void dm_io_acct(struct dm_io *io, bool end) 507 { 508 struct bio *bio = io->orig_bio; 509 510 if (dm_io_flagged(io, DM_IO_BLK_STAT)) { 511 if (!end) 512 bdev_start_io_acct(bio->bi_bdev, bio_op(bio), 513 io->start_time); 514 else 515 bdev_end_io_acct(bio->bi_bdev, bio_op(bio), 516 dm_io_sectors(io, bio), 517 io->start_time); 518 } 519 520 if (static_branch_unlikely(&stats_enabled) && 521 unlikely(dm_stats_used(&io->md->stats))) { 522 sector_t sector; 523 524 if (unlikely(dm_io_flagged(io, DM_IO_WAS_SPLIT))) 525 sector = bio_end_sector(bio) - io->sector_offset; 526 else 527 sector = bio->bi_iter.bi_sector; 528 529 dm_stats_account_io(&io->md->stats, bio_data_dir(bio), 530 sector, dm_io_sectors(io, bio), 531 end, io->start_time, &io->stats_aux); 532 } 533 } 534 535 static void __dm_start_io_acct(struct dm_io *io) 536 { 537 dm_io_acct(io, false); 538 } 539 540 static void dm_start_io_acct(struct dm_io *io, struct bio *clone) 541 { 542 /* 543 * Ensure IO accounting is only ever started once. 544 */ 545 if (dm_io_flagged(io, DM_IO_ACCOUNTED)) 546 return; 547 548 /* Expect no possibility for race unless DM_TIO_IS_DUPLICATE_BIO. */ 549 if (!clone || likely(dm_tio_is_normal(clone_to_tio(clone)))) { 550 dm_io_set_flag(io, DM_IO_ACCOUNTED); 551 } else { 552 unsigned long flags; 553 /* Can afford locking given DM_TIO_IS_DUPLICATE_BIO */ 554 spin_lock_irqsave(&io->lock, flags); 555 if (dm_io_flagged(io, DM_IO_ACCOUNTED)) { 556 spin_unlock_irqrestore(&io->lock, flags); 557 return; 558 } 559 dm_io_set_flag(io, DM_IO_ACCOUNTED); 560 spin_unlock_irqrestore(&io->lock, flags); 561 } 562 563 __dm_start_io_acct(io); 564 } 565 566 static void dm_end_io_acct(struct dm_io *io) 567 { 568 dm_io_acct(io, true); 569 } 570 571 static struct dm_io *alloc_io(struct mapped_device *md, struct bio *bio, gfp_t gfp_mask) 572 { 573 struct dm_io *io; 574 struct dm_target_io *tio; 575 struct bio *clone; 576 577 clone = bio_alloc_clone(NULL, bio, gfp_mask, &md->mempools->io_bs); 578 if (unlikely(!clone)) 579 return NULL; 580 tio = clone_to_tio(clone); 581 tio->flags = 0; 582 dm_tio_set_flag(tio, DM_TIO_INSIDE_DM_IO); 583 tio->io = NULL; 584 585 io = container_of(tio, struct dm_io, tio); 586 io->magic = DM_IO_MAGIC; 587 io->status = BLK_STS_OK; 588 io->requeue_flush_with_data = false; 589 590 /* one ref is for submission, the other is for completion */ 591 atomic_set(&io->io_count, 2); 592 this_cpu_inc(*md->pending_io); 593 io->orig_bio = bio; 594 io->md = md; 595 spin_lock_init(&io->lock); 596 io->start_time = jiffies; 597 io->flags = 0; 598 if (blk_queue_io_stat(md->queue)) 599 dm_io_set_flag(io, DM_IO_BLK_STAT); 600 601 if (static_branch_unlikely(&stats_enabled) && 602 unlikely(dm_stats_used(&md->stats))) 603 dm_stats_record_start(&md->stats, &io->stats_aux); 604 605 return io; 606 } 607 608 static void free_io(struct dm_io *io) 609 { 610 bio_put(&io->tio.clone); 611 } 612 613 static struct bio *alloc_tio(struct clone_info *ci, struct dm_target *ti, 614 unsigned int target_bio_nr, unsigned int *len, gfp_t gfp_mask) 615 { 616 struct mapped_device *md = ci->io->md; 617 struct dm_target_io *tio; 618 struct bio *clone; 619 620 if (!ci->io->tio.io) { 621 /* the dm_target_io embedded in ci->io is available */ 622 tio = &ci->io->tio; 623 /* alloc_io() already initialized embedded clone */ 624 clone = &tio->clone; 625 } else { 626 clone = bio_alloc_clone(NULL, ci->bio, gfp_mask, 627 &md->mempools->bs); 628 if (!clone) 629 return NULL; 630 631 /* REQ_DM_POLL_LIST shouldn't be inherited */ 632 clone->bi_opf &= ~REQ_DM_POLL_LIST; 633 634 tio = clone_to_tio(clone); 635 tio->flags = 0; /* also clears DM_TIO_INSIDE_DM_IO */ 636 } 637 638 tio->magic = DM_TIO_MAGIC; 639 tio->io = ci->io; 640 tio->ti = ti; 641 tio->target_bio_nr = target_bio_nr; 642 tio->len_ptr = len; 643 tio->old_sector = 0; 644 645 /* Set default bdev, but target must bio_set_dev() before issuing IO */ 646 clone->bi_bdev = md->disk->part0; 647 if (likely(ti != NULL) && unlikely(ti->needs_bio_set_dev)) 648 bio_set_dev(clone, md->disk->part0); 649 650 if (len) { 651 clone->bi_iter.bi_size = to_bytes(*len); 652 if (bio_integrity(clone)) 653 bio_integrity_trim(clone); 654 } 655 656 return clone; 657 } 658 659 static void free_tio(struct bio *clone) 660 { 661 if (dm_tio_flagged(clone_to_tio(clone), DM_TIO_INSIDE_DM_IO)) 662 return; 663 bio_put(clone); 664 } 665 666 /* 667 * Add the bio to the list of deferred io. 668 */ 669 static void queue_io(struct mapped_device *md, struct bio *bio) 670 { 671 unsigned long flags; 672 673 spin_lock_irqsave(&md->deferred_lock, flags); 674 bio_list_add(&md->deferred, bio); 675 spin_unlock_irqrestore(&md->deferred_lock, flags); 676 queue_work(md->wq, &md->work); 677 } 678 679 /* 680 * Everyone (including functions in this file), should use this 681 * function to access the md->map field, and make sure they call 682 * dm_put_live_table() when finished. 683 */ 684 struct dm_table *dm_get_live_table(struct mapped_device *md, 685 int *srcu_idx) __acquires(md->io_barrier) 686 { 687 *srcu_idx = srcu_read_lock(&md->io_barrier); 688 689 return srcu_dereference(md->map, &md->io_barrier); 690 } 691 692 void dm_put_live_table(struct mapped_device *md, 693 int srcu_idx) __releases(md->io_barrier) 694 { 695 srcu_read_unlock(&md->io_barrier, srcu_idx); 696 } 697 698 void dm_sync_table(struct mapped_device *md) 699 { 700 synchronize_srcu(&md->io_barrier); 701 synchronize_rcu_expedited(); 702 } 703 704 /* 705 * A fast alternative to dm_get_live_table/dm_put_live_table. 706 * The caller must not block between these two functions. 707 */ 708 static struct dm_table *dm_get_live_table_fast(struct mapped_device *md) __acquires(RCU) 709 { 710 rcu_read_lock(); 711 return rcu_dereference(md->map); 712 } 713 714 static void dm_put_live_table_fast(struct mapped_device *md) __releases(RCU) 715 { 716 rcu_read_unlock(); 717 } 718 719 static char *_dm_claim_ptr = "I belong to device-mapper"; 720 721 /* 722 * Open a table device so we can use it as a map destination. 723 */ 724 static struct table_device *open_table_device(struct mapped_device *md, 725 dev_t dev, blk_mode_t mode) 726 { 727 struct table_device *td; 728 struct file *bdev_file; 729 struct block_device *bdev; 730 u64 part_off; 731 int r; 732 733 td = kmalloc_node(sizeof(*td), GFP_KERNEL, md->numa_node_id); 734 if (!td) 735 return ERR_PTR(-ENOMEM); 736 refcount_set(&td->count, 1); 737 738 bdev_file = bdev_file_open_by_dev(dev, mode, _dm_claim_ptr, NULL); 739 if (IS_ERR(bdev_file)) { 740 r = PTR_ERR(bdev_file); 741 goto out_free_td; 742 } 743 744 bdev = file_bdev(bdev_file); 745 746 /* 747 * We can be called before the dm disk is added. In that case we can't 748 * register the holder relation here. It will be done once add_disk was 749 * called. 750 */ 751 if (md->disk->slave_dir) { 752 r = bd_link_disk_holder(bdev, md->disk); 753 if (r) 754 goto out_blkdev_put; 755 } 756 757 td->dm_dev.mode = mode; 758 td->dm_dev.bdev = bdev; 759 td->dm_dev.bdev_file = bdev_file; 760 td->dm_dev.dax_dev = fs_dax_get_by_bdev(bdev, &part_off, 761 NULL, NULL); 762 format_dev_t(td->dm_dev.name, dev); 763 list_add(&td->list, &md->table_devices); 764 return td; 765 766 out_blkdev_put: 767 __fput_sync(bdev_file); 768 out_free_td: 769 kfree(td); 770 return ERR_PTR(r); 771 } 772 773 /* 774 * Close a table device that we've been using. 775 */ 776 static void close_table_device(struct table_device *td, struct mapped_device *md) 777 { 778 if (md->disk->slave_dir) 779 bd_unlink_disk_holder(td->dm_dev.bdev, md->disk); 780 781 /* Leverage async fput() if DMF_DEFERRED_REMOVE set */ 782 if (unlikely(test_bit(DMF_DEFERRED_REMOVE, &md->flags))) 783 fput(td->dm_dev.bdev_file); 784 else 785 __fput_sync(td->dm_dev.bdev_file); 786 787 put_dax(td->dm_dev.dax_dev); 788 list_del(&td->list); 789 kfree(td); 790 } 791 792 static struct table_device *find_table_device(struct list_head *l, dev_t dev, 793 blk_mode_t mode) 794 { 795 struct table_device *td; 796 797 list_for_each_entry(td, l, list) 798 if (td->dm_dev.bdev->bd_dev == dev && td->dm_dev.mode == mode) 799 return td; 800 801 return NULL; 802 } 803 804 int dm_get_table_device(struct mapped_device *md, dev_t dev, blk_mode_t mode, 805 struct dm_dev **result) 806 { 807 struct table_device *td; 808 809 mutex_lock(&md->table_devices_lock); 810 td = find_table_device(&md->table_devices, dev, mode); 811 if (!td) { 812 td = open_table_device(md, dev, mode); 813 if (IS_ERR(td)) { 814 mutex_unlock(&md->table_devices_lock); 815 return PTR_ERR(td); 816 } 817 } else { 818 refcount_inc(&td->count); 819 } 820 mutex_unlock(&md->table_devices_lock); 821 822 *result = &td->dm_dev; 823 return 0; 824 } 825 826 void dm_put_table_device(struct mapped_device *md, struct dm_dev *d) 827 { 828 struct table_device *td = container_of(d, struct table_device, dm_dev); 829 830 mutex_lock(&md->table_devices_lock); 831 if (refcount_dec_and_test(&td->count)) 832 close_table_device(td, md); 833 mutex_unlock(&md->table_devices_lock); 834 } 835 836 /* 837 * Get the geometry associated with a dm device 838 */ 839 int dm_get_geometry(struct mapped_device *md, struct hd_geometry *geo) 840 { 841 *geo = md->geometry; 842 843 return 0; 844 } 845 846 /* 847 * Set the geometry of a device. 848 */ 849 int dm_set_geometry(struct mapped_device *md, struct hd_geometry *geo) 850 { 851 sector_t sz = (sector_t)geo->cylinders * geo->heads * geo->sectors; 852 853 if (geo->start > sz) { 854 DMERR("Start sector is beyond the geometry limits."); 855 return -EINVAL; 856 } 857 858 md->geometry = *geo; 859 860 return 0; 861 } 862 863 static int __noflush_suspending(struct mapped_device *md) 864 { 865 return test_bit(DMF_NOFLUSH_SUSPENDING, &md->flags); 866 } 867 868 static void dm_requeue_add_io(struct dm_io *io, bool first_stage) 869 { 870 struct mapped_device *md = io->md; 871 872 if (first_stage) { 873 struct dm_io *next = md->requeue_list; 874 875 md->requeue_list = io; 876 io->next = next; 877 } else { 878 bio_list_add_head(&md->deferred, io->orig_bio); 879 } 880 } 881 882 static void dm_kick_requeue(struct mapped_device *md, bool first_stage) 883 { 884 if (first_stage) 885 queue_work(md->wq, &md->requeue_work); 886 else 887 queue_work(md->wq, &md->work); 888 } 889 890 /* 891 * Return true if the dm_io's original bio is requeued. 892 * io->status is updated with error if requeue disallowed. 893 */ 894 static bool dm_handle_requeue(struct dm_io *io, bool first_stage) 895 { 896 struct bio *bio = io->orig_bio; 897 bool handle_requeue = (io->status == BLK_STS_DM_REQUEUE); 898 bool handle_polled_eagain = ((io->status == BLK_STS_AGAIN) && 899 (bio->bi_opf & REQ_POLLED)); 900 struct mapped_device *md = io->md; 901 bool requeued = false; 902 903 if (handle_requeue || handle_polled_eagain) { 904 unsigned long flags; 905 906 if (bio->bi_opf & REQ_POLLED) { 907 /* 908 * Upper layer won't help us poll split bio 909 * (io->orig_bio may only reflect a subset of the 910 * pre-split original) so clear REQ_POLLED. 911 */ 912 bio_clear_polled(bio); 913 } 914 915 /* 916 * Target requested pushing back the I/O or 917 * polled IO hit BLK_STS_AGAIN. 918 */ 919 spin_lock_irqsave(&md->deferred_lock, flags); 920 if ((__noflush_suspending(md) && 921 !WARN_ON_ONCE(dm_is_zone_write(md, bio))) || 922 handle_polled_eagain || first_stage) { 923 dm_requeue_add_io(io, first_stage); 924 requeued = true; 925 } else { 926 /* 927 * noflush suspend was interrupted or this is 928 * a write to a zoned target. 929 */ 930 io->status = BLK_STS_IOERR; 931 } 932 spin_unlock_irqrestore(&md->deferred_lock, flags); 933 } 934 935 if (requeued) 936 dm_kick_requeue(md, first_stage); 937 938 return requeued; 939 } 940 941 static void __dm_io_complete(struct dm_io *io, bool first_stage) 942 { 943 struct bio *bio = io->orig_bio; 944 struct mapped_device *md = io->md; 945 blk_status_t io_error; 946 bool requeued; 947 bool requeue_flush_with_data; 948 949 requeued = dm_handle_requeue(io, first_stage); 950 if (requeued && first_stage) 951 return; 952 953 io_error = io->status; 954 if (dm_io_flagged(io, DM_IO_ACCOUNTED)) 955 dm_end_io_acct(io); 956 else if (!io_error) { 957 /* 958 * Must handle target that DM_MAPIO_SUBMITTED only to 959 * then bio_endio() rather than dm_submit_bio_remap() 960 */ 961 __dm_start_io_acct(io); 962 dm_end_io_acct(io); 963 } 964 requeue_flush_with_data = io->requeue_flush_with_data; 965 free_io(io); 966 smp_wmb(); 967 this_cpu_dec(*md->pending_io); 968 969 /* nudge anyone waiting on suspend queue */ 970 if (unlikely(wq_has_sleeper(&md->wait))) 971 wake_up(&md->wait); 972 973 /* Return early if the original bio was requeued */ 974 if (requeued) 975 return; 976 977 if (unlikely(requeue_flush_with_data)) { 978 /* 979 * Preflush done for flush with data, reissue 980 * without REQ_PREFLUSH. 981 */ 982 bio->bi_opf &= ~REQ_PREFLUSH; 983 queue_io(md, bio); 984 } else { 985 /* done with normal IO or empty flush */ 986 if (io_error) 987 bio->bi_status = io_error; 988 bio_endio(bio); 989 } 990 } 991 992 static void dm_wq_requeue_work(struct work_struct *work) 993 { 994 struct mapped_device *md = container_of(work, struct mapped_device, 995 requeue_work); 996 unsigned long flags; 997 struct dm_io *io; 998 999 /* reuse deferred lock to simplify dm_handle_requeue */ 1000 spin_lock_irqsave(&md->deferred_lock, flags); 1001 io = md->requeue_list; 1002 md->requeue_list = NULL; 1003 spin_unlock_irqrestore(&md->deferred_lock, flags); 1004 1005 while (io) { 1006 struct dm_io *next = io->next; 1007 1008 dm_io_rewind(io, &md->disk->bio_split); 1009 1010 io->next = NULL; 1011 __dm_io_complete(io, false); 1012 io = next; 1013 cond_resched(); 1014 } 1015 } 1016 1017 /* 1018 * Two staged requeue: 1019 * 1020 * 1) io->orig_bio points to the real original bio, and the part mapped to 1021 * this io must be requeued, instead of other parts of the original bio. 1022 * 1023 * 2) io->orig_bio points to new cloned bio which matches the requeued dm_io. 1024 */ 1025 static inline void dm_io_complete(struct dm_io *io) 1026 { 1027 /* 1028 * Only dm_io that has been split needs two stage requeue, otherwise 1029 * we may run into long bio clone chain during suspend and OOM could 1030 * be triggered. 1031 * 1032 * Also flush data dm_io won't be marked as DM_IO_WAS_SPLIT, so they 1033 * also aren't handled via the first stage requeue. 1034 */ 1035 __dm_io_complete(io, dm_io_flagged(io, DM_IO_WAS_SPLIT)); 1036 } 1037 1038 /* 1039 * Decrements the number of outstanding ios that a bio has been 1040 * cloned into, completing the original io if necc. 1041 */ 1042 static inline void __dm_io_dec_pending(struct dm_io *io) 1043 { 1044 if (atomic_dec_and_test(&io->io_count)) 1045 dm_io_complete(io); 1046 } 1047 1048 static void dm_io_set_error(struct dm_io *io, blk_status_t error) 1049 { 1050 unsigned long flags; 1051 1052 /* Push-back supersedes any I/O errors */ 1053 spin_lock_irqsave(&io->lock, flags); 1054 if (!(io->status == BLK_STS_DM_REQUEUE && 1055 __noflush_suspending(io->md))) { 1056 io->status = error; 1057 } 1058 spin_unlock_irqrestore(&io->lock, flags); 1059 } 1060 1061 static void dm_io_dec_pending(struct dm_io *io, blk_status_t error) 1062 { 1063 if (unlikely(error)) 1064 dm_io_set_error(io, error); 1065 1066 __dm_io_dec_pending(io); 1067 } 1068 1069 /* 1070 * The queue_limits are only valid as long as you have a reference 1071 * count on 'md'. But _not_ imposing verification to avoid atomic_read(), 1072 */ 1073 static inline struct queue_limits *dm_get_queue_limits(struct mapped_device *md) 1074 { 1075 return &md->queue->limits; 1076 } 1077 1078 static bool swap_bios_limit(struct dm_target *ti, struct bio *bio) 1079 { 1080 return unlikely((bio->bi_opf & REQ_SWAP) != 0) && unlikely(ti->limit_swap_bios); 1081 } 1082 1083 static void clone_endio(struct bio *bio) 1084 { 1085 blk_status_t error = bio->bi_status; 1086 struct dm_target_io *tio = clone_to_tio(bio); 1087 struct dm_target *ti = tio->ti; 1088 dm_endio_fn endio = likely(ti != NULL) ? ti->type->end_io : NULL; 1089 struct dm_io *io = tio->io; 1090 struct mapped_device *md = io->md; 1091 1092 if (unlikely(error == BLK_STS_TARGET)) { 1093 if (bio_op(bio) == REQ_OP_DISCARD && 1094 !bdev_max_discard_sectors(bio->bi_bdev)) 1095 blk_queue_disable_discard(md->queue); 1096 else if (bio_op(bio) == REQ_OP_WRITE_ZEROES && 1097 !bdev_write_zeroes_sectors(bio->bi_bdev)) 1098 blk_queue_disable_write_zeroes(md->queue); 1099 } 1100 1101 if (static_branch_unlikely(&zoned_enabled) && 1102 unlikely(bdev_is_zoned(bio->bi_bdev))) 1103 dm_zone_endio(io, bio); 1104 1105 if (endio) { 1106 int r = endio(ti, bio, &error); 1107 1108 switch (r) { 1109 case DM_ENDIO_REQUEUE: 1110 if (static_branch_unlikely(&zoned_enabled)) { 1111 /* 1112 * Requeuing writes to a sequential zone of a zoned 1113 * target will break the sequential write pattern: 1114 * fail such IO. 1115 */ 1116 if (WARN_ON_ONCE(dm_is_zone_write(md, bio))) 1117 error = BLK_STS_IOERR; 1118 else 1119 error = BLK_STS_DM_REQUEUE; 1120 } else 1121 error = BLK_STS_DM_REQUEUE; 1122 fallthrough; 1123 case DM_ENDIO_DONE: 1124 break; 1125 case DM_ENDIO_INCOMPLETE: 1126 /* The target will handle the io */ 1127 return; 1128 default: 1129 DMCRIT("unimplemented target endio return value: %d", r); 1130 BUG(); 1131 } 1132 } 1133 1134 if (static_branch_unlikely(&swap_bios_enabled) && 1135 likely(ti != NULL) && unlikely(swap_bios_limit(ti, bio))) 1136 up(&md->swap_bios_semaphore); 1137 1138 free_tio(bio); 1139 dm_io_dec_pending(io, error); 1140 } 1141 1142 /* 1143 * Return maximum size of I/O possible at the supplied sector up to the current 1144 * target boundary. 1145 */ 1146 static inline sector_t max_io_len_target_boundary(struct dm_target *ti, 1147 sector_t target_offset) 1148 { 1149 return ti->len - target_offset; 1150 } 1151 1152 static sector_t __max_io_len(struct dm_target *ti, sector_t sector, 1153 unsigned int max_granularity, 1154 unsigned int max_sectors) 1155 { 1156 sector_t target_offset = dm_target_offset(ti, sector); 1157 sector_t len = max_io_len_target_boundary(ti, target_offset); 1158 1159 /* 1160 * Does the target need to split IO even further? 1161 * - varied (per target) IO splitting is a tenet of DM; this 1162 * explains why stacked chunk_sectors based splitting via 1163 * bio_split_to_limits() isn't possible here. 1164 */ 1165 if (!max_granularity) 1166 return len; 1167 return min_t(sector_t, len, 1168 min(max_sectors ? : queue_max_sectors(ti->table->md->queue), 1169 blk_boundary_sectors_left(target_offset, max_granularity))); 1170 } 1171 1172 static inline sector_t max_io_len(struct dm_target *ti, sector_t sector) 1173 { 1174 return __max_io_len(ti, sector, ti->max_io_len, 0); 1175 } 1176 1177 int dm_set_target_max_io_len(struct dm_target *ti, sector_t len) 1178 { 1179 if (len > UINT_MAX) { 1180 DMERR("Specified maximum size of target IO (%llu) exceeds limit (%u)", 1181 (unsigned long long)len, UINT_MAX); 1182 ti->error = "Maximum size of target IO is too large"; 1183 return -EINVAL; 1184 } 1185 1186 ti->max_io_len = (uint32_t) len; 1187 1188 return 0; 1189 } 1190 EXPORT_SYMBOL_GPL(dm_set_target_max_io_len); 1191 1192 static struct dm_target *dm_dax_get_live_target(struct mapped_device *md, 1193 sector_t sector, int *srcu_idx) 1194 __acquires(md->io_barrier) 1195 { 1196 struct dm_table *map; 1197 struct dm_target *ti; 1198 1199 map = dm_get_live_table(md, srcu_idx); 1200 if (!map) 1201 return NULL; 1202 1203 ti = dm_table_find_target(map, sector); 1204 if (!ti) 1205 return NULL; 1206 1207 return ti; 1208 } 1209 1210 static long dm_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, 1211 long nr_pages, enum dax_access_mode mode, void **kaddr, 1212 unsigned long *pfn) 1213 { 1214 struct mapped_device *md = dax_get_private(dax_dev); 1215 sector_t sector = pgoff * PAGE_SECTORS; 1216 struct dm_target *ti; 1217 long len, ret = -EIO; 1218 int srcu_idx; 1219 1220 ti = dm_dax_get_live_target(md, sector, &srcu_idx); 1221 1222 if (!ti) 1223 goto out; 1224 if (!ti->type->direct_access) 1225 goto out; 1226 len = max_io_len(ti, sector) / PAGE_SECTORS; 1227 if (len < 1) 1228 goto out; 1229 nr_pages = min(len, nr_pages); 1230 ret = ti->type->direct_access(ti, pgoff, nr_pages, mode, kaddr, pfn); 1231 1232 out: 1233 dm_put_live_table(md, srcu_idx); 1234 1235 return ret; 1236 } 1237 1238 static int dm_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, 1239 size_t nr_pages) 1240 { 1241 struct mapped_device *md = dax_get_private(dax_dev); 1242 sector_t sector = pgoff * PAGE_SECTORS; 1243 struct dm_target *ti; 1244 int ret = -EIO; 1245 int srcu_idx; 1246 1247 ti = dm_dax_get_live_target(md, sector, &srcu_idx); 1248 1249 if (!ti) 1250 goto out; 1251 if (WARN_ON(!ti->type->dax_zero_page_range)) { 1252 /* 1253 * ->zero_page_range() is mandatory dax operation. If we are 1254 * here, something is wrong. 1255 */ 1256 goto out; 1257 } 1258 ret = ti->type->dax_zero_page_range(ti, pgoff, nr_pages); 1259 out: 1260 dm_put_live_table(md, srcu_idx); 1261 1262 return ret; 1263 } 1264 1265 static size_t dm_dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, 1266 void *addr, size_t bytes, struct iov_iter *i) 1267 { 1268 struct mapped_device *md = dax_get_private(dax_dev); 1269 sector_t sector = pgoff * PAGE_SECTORS; 1270 struct dm_target *ti; 1271 int srcu_idx; 1272 long ret = 0; 1273 1274 ti = dm_dax_get_live_target(md, sector, &srcu_idx); 1275 if (!ti || !ti->type->dax_recovery_write) 1276 goto out; 1277 1278 ret = ti->type->dax_recovery_write(ti, pgoff, addr, bytes, i); 1279 out: 1280 dm_put_live_table(md, srcu_idx); 1281 return ret; 1282 } 1283 1284 /* 1285 * A target may call dm_accept_partial_bio only from the map routine. It is 1286 * allowed for all bio types except REQ_PREFLUSH, REQ_OP_ZONE_* zone management 1287 * operations, zone append writes (native with REQ_OP_ZONE_APPEND or emulated 1288 * with write BIOs flagged with BIO_EMULATES_ZONE_APPEND) and any bio serviced 1289 * by __send_duplicate_bios(). 1290 * 1291 * dm_accept_partial_bio informs the dm that the target only wants to process 1292 * additional n_sectors sectors of the bio and the rest of the data should be 1293 * sent in a next bio. 1294 * 1295 * A diagram that explains the arithmetics: 1296 * +--------------------+---------------+-------+ 1297 * | 1 | 2 | 3 | 1298 * +--------------------+---------------+-------+ 1299 * 1300 * <-------------- *tio->len_ptr ---------------> 1301 * <----- bio_sectors -----> 1302 * <-- n_sectors --> 1303 * 1304 * Region 1 was already iterated over with bio_advance or similar function. 1305 * (it may be empty if the target doesn't use bio_advance) 1306 * Region 2 is the remaining bio size that the target wants to process. 1307 * (it may be empty if region 1 is non-empty, although there is no reason 1308 * to make it empty) 1309 * The target requires that region 3 is to be sent in the next bio. 1310 * 1311 * If the target wants to receive multiple copies of the bio (via num_*bios, etc), 1312 * the partially processed part (the sum of regions 1+2) must be the same for all 1313 * copies of the bio. 1314 */ 1315 void dm_accept_partial_bio(struct bio *bio, unsigned int n_sectors) 1316 { 1317 struct dm_target_io *tio = clone_to_tio(bio); 1318 struct dm_io *io = tio->io; 1319 unsigned int bio_sectors = bio_sectors(bio); 1320 1321 BUG_ON(dm_tio_flagged(tio, DM_TIO_IS_DUPLICATE_BIO)); 1322 BUG_ON(bio_sectors > *tio->len_ptr); 1323 BUG_ON(n_sectors > bio_sectors); 1324 1325 if (static_branch_unlikely(&zoned_enabled) && 1326 unlikely(bdev_is_zoned(bio->bi_bdev))) { 1327 enum req_op op = bio_op(bio); 1328 1329 BUG_ON(op_is_zone_mgmt(op)); 1330 BUG_ON(op == REQ_OP_WRITE); 1331 BUG_ON(op == REQ_OP_WRITE_ZEROES); 1332 BUG_ON(op == REQ_OP_ZONE_APPEND); 1333 } 1334 1335 *tio->len_ptr -= bio_sectors - n_sectors; 1336 bio->bi_iter.bi_size = n_sectors << SECTOR_SHIFT; 1337 1338 /* 1339 * __split_and_process_bio() may have already saved mapped part 1340 * for accounting but it is being reduced so update accordingly. 1341 */ 1342 dm_io_set_flag(io, DM_IO_WAS_SPLIT); 1343 io->sectors = n_sectors; 1344 io->sector_offset = bio_sectors(io->orig_bio); 1345 } 1346 EXPORT_SYMBOL_GPL(dm_accept_partial_bio); 1347 1348 /* 1349 * @clone: clone bio that DM core passed to target's .map function 1350 * @tgt_clone: clone of @clone bio that target needs submitted 1351 * 1352 * Targets should use this interface to submit bios they take 1353 * ownership of when returning DM_MAPIO_SUBMITTED. 1354 * 1355 * Target should also enable ti->accounts_remapped_io 1356 */ 1357 void dm_submit_bio_remap(struct bio *clone, struct bio *tgt_clone) 1358 { 1359 struct dm_target_io *tio = clone_to_tio(clone); 1360 struct dm_io *io = tio->io; 1361 1362 /* establish bio that will get submitted */ 1363 if (!tgt_clone) 1364 tgt_clone = clone; 1365 1366 /* 1367 * Account io->origin_bio to DM dev on behalf of target 1368 * that took ownership of IO with DM_MAPIO_SUBMITTED. 1369 */ 1370 dm_start_io_acct(io, clone); 1371 1372 trace_block_bio_remap(tgt_clone, disk_devt(io->md->disk), 1373 tio->old_sector); 1374 submit_bio_noacct(tgt_clone); 1375 } 1376 EXPORT_SYMBOL_GPL(dm_submit_bio_remap); 1377 1378 static noinline void __set_swap_bios_limit(struct mapped_device *md, int latch) 1379 { 1380 mutex_lock(&md->swap_bios_lock); 1381 while (latch < md->swap_bios) { 1382 cond_resched(); 1383 down(&md->swap_bios_semaphore); 1384 md->swap_bios--; 1385 } 1386 while (latch > md->swap_bios) { 1387 cond_resched(); 1388 up(&md->swap_bios_semaphore); 1389 md->swap_bios++; 1390 } 1391 mutex_unlock(&md->swap_bios_lock); 1392 } 1393 1394 static void __map_bio(struct bio *clone) 1395 { 1396 struct dm_target_io *tio = clone_to_tio(clone); 1397 struct dm_target *ti = tio->ti; 1398 struct dm_io *io = tio->io; 1399 struct mapped_device *md = io->md; 1400 int r; 1401 1402 clone->bi_end_io = clone_endio; 1403 1404 /* 1405 * Map the clone. 1406 */ 1407 tio->old_sector = clone->bi_iter.bi_sector; 1408 1409 if (static_branch_unlikely(&swap_bios_enabled) && 1410 unlikely(swap_bios_limit(ti, clone))) { 1411 int latch = get_swap_bios(); 1412 1413 if (unlikely(latch != md->swap_bios)) 1414 __set_swap_bios_limit(md, latch); 1415 down(&md->swap_bios_semaphore); 1416 } 1417 1418 if (likely(ti->type->map == linear_map)) 1419 r = linear_map(ti, clone); 1420 else if (ti->type->map == stripe_map) 1421 r = stripe_map(ti, clone); 1422 else 1423 r = ti->type->map(ti, clone); 1424 1425 switch (r) { 1426 case DM_MAPIO_SUBMITTED: 1427 /* target has assumed ownership of this io */ 1428 if (!ti->accounts_remapped_io) 1429 dm_start_io_acct(io, clone); 1430 break; 1431 case DM_MAPIO_REMAPPED: 1432 dm_submit_bio_remap(clone, NULL); 1433 break; 1434 case DM_MAPIO_KILL: 1435 case DM_MAPIO_REQUEUE: 1436 if (static_branch_unlikely(&swap_bios_enabled) && 1437 unlikely(swap_bios_limit(ti, clone))) 1438 up(&md->swap_bios_semaphore); 1439 free_tio(clone); 1440 if (r == DM_MAPIO_KILL) 1441 dm_io_dec_pending(io, BLK_STS_IOERR); 1442 else 1443 dm_io_dec_pending(io, BLK_STS_DM_REQUEUE); 1444 break; 1445 default: 1446 DMCRIT("unimplemented target map return value: %d", r); 1447 BUG(); 1448 } 1449 } 1450 1451 static void setup_split_accounting(struct clone_info *ci, unsigned int len) 1452 { 1453 struct dm_io *io = ci->io; 1454 1455 if (ci->sector_count > len) { 1456 /* 1457 * Split needed, save the mapped part for accounting. 1458 * NOTE: dm_accept_partial_bio() will update accordingly. 1459 */ 1460 dm_io_set_flag(io, DM_IO_WAS_SPLIT); 1461 io->sectors = len; 1462 io->sector_offset = bio_sectors(ci->bio); 1463 } 1464 } 1465 1466 static void alloc_multiple_bios(struct bio_list *blist, struct clone_info *ci, 1467 struct dm_target *ti, unsigned int num_bios, 1468 unsigned *len) 1469 { 1470 struct bio *bio; 1471 int try; 1472 1473 for (try = 0; try < 2; try++) { 1474 int bio_nr; 1475 1476 if (try && num_bios > 1) 1477 mutex_lock(&ci->io->md->table_devices_lock); 1478 for (bio_nr = 0; bio_nr < num_bios; bio_nr++) { 1479 bio = alloc_tio(ci, ti, bio_nr, len, 1480 try ? GFP_NOIO : GFP_NOWAIT); 1481 if (!bio) 1482 break; 1483 1484 bio_list_add(blist, bio); 1485 } 1486 if (try && num_bios > 1) 1487 mutex_unlock(&ci->io->md->table_devices_lock); 1488 if (bio_nr == num_bios) 1489 return; 1490 1491 while ((bio = bio_list_pop(blist))) 1492 free_tio(bio); 1493 } 1494 } 1495 1496 static unsigned int __send_duplicate_bios(struct clone_info *ci, struct dm_target *ti, 1497 unsigned int num_bios, unsigned int *len) 1498 { 1499 struct bio_list blist = BIO_EMPTY_LIST; 1500 struct bio *clone; 1501 unsigned int ret = 0; 1502 1503 if (WARN_ON_ONCE(num_bios == 0)) /* num_bios = 0 is a bug in caller */ 1504 return 0; 1505 1506 /* dm_accept_partial_bio() is not supported with shared tio->len_ptr */ 1507 if (len) 1508 setup_split_accounting(ci, *len); 1509 1510 /* 1511 * Using alloc_multiple_bios(), even if num_bios is 1, to consistently 1512 * support allocating using GFP_NOWAIT with GFP_NOIO fallback. 1513 */ 1514 alloc_multiple_bios(&blist, ci, ti, num_bios, len); 1515 while ((clone = bio_list_pop(&blist))) { 1516 if (num_bios > 1) 1517 dm_tio_set_flag(clone_to_tio(clone), DM_TIO_IS_DUPLICATE_BIO); 1518 __map_bio(clone); 1519 ret += 1; 1520 } 1521 1522 return ret; 1523 } 1524 1525 static void __send_empty_flush(struct clone_info *ci) 1526 { 1527 struct dm_table *t = ci->map; 1528 struct bio flush_bio; 1529 blk_opf_t opf = REQ_OP_WRITE | REQ_PREFLUSH | REQ_SYNC; 1530 1531 if ((ci->io->orig_bio->bi_opf & (REQ_IDLE | REQ_SYNC)) == 1532 (REQ_IDLE | REQ_SYNC)) 1533 opf |= REQ_IDLE; 1534 1535 /* 1536 * Use an on-stack bio for this, it's safe since we don't 1537 * need to reference it after submit. It's just used as 1538 * the basis for the clone(s). 1539 */ 1540 bio_init(&flush_bio, ci->io->md->disk->part0, NULL, 0, opf); 1541 1542 ci->bio = &flush_bio; 1543 ci->sector_count = 0; 1544 ci->io->tio.clone.bi_iter.bi_size = 0; 1545 1546 if (!t->flush_bypasses_map) { 1547 for (unsigned int i = 0; i < t->num_targets; i++) { 1548 unsigned int bios; 1549 struct dm_target *ti = dm_table_get_target(t, i); 1550 1551 if (unlikely(ti->num_flush_bios == 0)) 1552 continue; 1553 1554 atomic_add(ti->num_flush_bios, &ci->io->io_count); 1555 bios = __send_duplicate_bios(ci, ti, ti->num_flush_bios, 1556 NULL); 1557 atomic_sub(ti->num_flush_bios - bios, &ci->io->io_count); 1558 } 1559 } else { 1560 /* 1561 * Note that there's no need to grab t->devices_lock here 1562 * because the targets that support flush optimization don't 1563 * modify the list of devices. 1564 */ 1565 struct list_head *devices = dm_table_get_devices(t); 1566 unsigned int len = 0; 1567 struct dm_dev_internal *dd; 1568 list_for_each_entry(dd, devices, list) { 1569 struct bio *clone; 1570 /* 1571 * Note that the structure dm_target_io is not 1572 * associated with any target (because the device may be 1573 * used by multiple targets), so we set tio->ti = NULL. 1574 * We must check for NULL in the I/O processing path, to 1575 * avoid NULL pointer dereference. 1576 */ 1577 clone = alloc_tio(ci, NULL, 0, &len, GFP_NOIO); 1578 atomic_add(1, &ci->io->io_count); 1579 bio_set_dev(clone, dd->dm_dev->bdev); 1580 clone->bi_end_io = clone_endio; 1581 dm_submit_bio_remap(clone, NULL); 1582 } 1583 } 1584 1585 /* 1586 * alloc_io() takes one extra reference for submission, so the 1587 * reference won't reach 0 without the following subtraction 1588 */ 1589 atomic_sub(1, &ci->io->io_count); 1590 1591 bio_uninit(ci->bio); 1592 } 1593 1594 static void __send_abnormal_io(struct clone_info *ci, struct dm_target *ti, 1595 unsigned int num_bios, unsigned int max_granularity, 1596 unsigned int max_sectors) 1597 { 1598 unsigned int len, bios; 1599 1600 len = min_t(sector_t, ci->sector_count, 1601 __max_io_len(ti, ci->sector, max_granularity, max_sectors)); 1602 1603 atomic_add(num_bios, &ci->io->io_count); 1604 bios = __send_duplicate_bios(ci, ti, num_bios, &len); 1605 /* 1606 * alloc_io() takes one extra reference for submission, so the 1607 * reference won't reach 0 without the following (+1) subtraction 1608 */ 1609 atomic_sub(num_bios - bios + 1, &ci->io->io_count); 1610 1611 ci->sector += len; 1612 ci->sector_count -= len; 1613 } 1614 1615 static bool is_abnormal_io(struct bio *bio) 1616 { 1617 switch (bio_op(bio)) { 1618 case REQ_OP_READ: 1619 case REQ_OP_WRITE: 1620 case REQ_OP_FLUSH: 1621 return false; 1622 case REQ_OP_DISCARD: 1623 case REQ_OP_SECURE_ERASE: 1624 case REQ_OP_WRITE_ZEROES: 1625 case REQ_OP_ZONE_RESET_ALL: 1626 return true; 1627 default: 1628 return false; 1629 } 1630 } 1631 1632 static blk_status_t __process_abnormal_io(struct clone_info *ci, 1633 struct dm_target *ti) 1634 { 1635 unsigned int num_bios = 0; 1636 unsigned int max_granularity = 0; 1637 unsigned int max_sectors = 0; 1638 struct queue_limits *limits = dm_get_queue_limits(ti->table->md); 1639 1640 switch (bio_op(ci->bio)) { 1641 case REQ_OP_DISCARD: 1642 num_bios = ti->num_discard_bios; 1643 max_sectors = limits->max_discard_sectors; 1644 if (ti->max_discard_granularity) 1645 max_granularity = max_sectors; 1646 break; 1647 case REQ_OP_SECURE_ERASE: 1648 num_bios = ti->num_secure_erase_bios; 1649 max_sectors = limits->max_secure_erase_sectors; 1650 break; 1651 case REQ_OP_WRITE_ZEROES: 1652 num_bios = ti->num_write_zeroes_bios; 1653 max_sectors = limits->max_write_zeroes_sectors; 1654 break; 1655 default: 1656 break; 1657 } 1658 1659 /* 1660 * Even though the device advertised support for this type of 1661 * request, that does not mean every target supports it, and 1662 * reconfiguration might also have changed that since the 1663 * check was performed. 1664 */ 1665 if (unlikely(!num_bios)) 1666 return BLK_STS_NOTSUPP; 1667 1668 __send_abnormal_io(ci, ti, num_bios, max_granularity, max_sectors); 1669 1670 return BLK_STS_OK; 1671 } 1672 1673 /* 1674 * Reuse ->bi_private as dm_io list head for storing all dm_io instances 1675 * associated with this bio, and this bio's bi_private needs to be 1676 * stored in dm_io->data before the reuse. 1677 * 1678 * bio->bi_private is owned by fs or upper layer, so block layer won't 1679 * touch it after splitting. Meantime it won't be changed by anyone after 1680 * bio is submitted. So this reuse is safe. 1681 */ 1682 static inline struct dm_io **dm_poll_list_head(struct bio *bio) 1683 { 1684 return (struct dm_io **)&bio->bi_private; 1685 } 1686 1687 static void dm_queue_poll_io(struct bio *bio, struct dm_io *io) 1688 { 1689 struct dm_io **head = dm_poll_list_head(bio); 1690 1691 if (!(bio->bi_opf & REQ_DM_POLL_LIST)) { 1692 bio->bi_opf |= REQ_DM_POLL_LIST; 1693 /* 1694 * Save .bi_private into dm_io, so that we can reuse 1695 * .bi_private as dm_io list head for storing dm_io list 1696 */ 1697 io->data = bio->bi_private; 1698 1699 /* tell block layer to poll for completion */ 1700 bio->bi_cookie = ~BLK_QC_T_NONE; 1701 1702 io->next = NULL; 1703 } else { 1704 /* 1705 * bio recursed due to split, reuse original poll list, 1706 * and save bio->bi_private too. 1707 */ 1708 io->data = (*head)->data; 1709 io->next = *head; 1710 } 1711 1712 *head = io; 1713 } 1714 1715 /* 1716 * Select the correct strategy for processing a non-flush bio. 1717 */ 1718 static blk_status_t __split_and_process_bio(struct clone_info *ci) 1719 { 1720 struct bio *clone; 1721 struct dm_target *ti; 1722 unsigned int len; 1723 1724 ti = dm_table_find_target(ci->map, ci->sector); 1725 if (unlikely(!ti)) 1726 return BLK_STS_IOERR; 1727 1728 if (unlikely(ci->is_abnormal_io)) 1729 return __process_abnormal_io(ci, ti); 1730 1731 /* 1732 * Only support bio polling for normal IO, and the target io is 1733 * exactly inside the dm_io instance (verified in dm_poll_dm_io) 1734 */ 1735 ci->submit_as_polled = !!(ci->bio->bi_opf & REQ_POLLED); 1736 1737 len = min_t(sector_t, max_io_len(ti, ci->sector), ci->sector_count); 1738 if (ci->bio->bi_opf & REQ_ATOMIC && len != ci->sector_count) 1739 return BLK_STS_IOERR; 1740 1741 setup_split_accounting(ci, len); 1742 1743 if (unlikely(ci->bio->bi_opf & REQ_NOWAIT)) { 1744 if (unlikely(!dm_target_supports_nowait(ti->type))) 1745 return BLK_STS_NOTSUPP; 1746 1747 clone = alloc_tio(ci, ti, 0, &len, GFP_NOWAIT); 1748 if (unlikely(!clone)) 1749 return BLK_STS_AGAIN; 1750 } else { 1751 clone = alloc_tio(ci, ti, 0, &len, GFP_NOIO); 1752 } 1753 __map_bio(clone); 1754 1755 ci->sector += len; 1756 ci->sector_count -= len; 1757 1758 return BLK_STS_OK; 1759 } 1760 1761 static void init_clone_info(struct clone_info *ci, struct dm_io *io, 1762 struct dm_table *map, struct bio *bio, bool is_abnormal) 1763 { 1764 ci->map = map; 1765 ci->io = io; 1766 ci->bio = bio; 1767 ci->is_abnormal_io = is_abnormal; 1768 ci->submit_as_polled = false; 1769 ci->sector = bio->bi_iter.bi_sector; 1770 ci->sector_count = bio_sectors(bio); 1771 1772 /* Shouldn't happen but sector_count was being set to 0 so... */ 1773 if (static_branch_unlikely(&zoned_enabled) && 1774 WARN_ON_ONCE(op_is_zone_mgmt(bio_op(bio)) && ci->sector_count)) 1775 ci->sector_count = 0; 1776 } 1777 1778 #ifdef CONFIG_BLK_DEV_ZONED 1779 static inline bool dm_zone_bio_needs_split(struct bio *bio) 1780 { 1781 /* 1782 * Special case the zone operations that cannot or should not be split. 1783 */ 1784 switch (bio_op(bio)) { 1785 case REQ_OP_ZONE_APPEND: 1786 case REQ_OP_ZONE_FINISH: 1787 case REQ_OP_ZONE_RESET: 1788 case REQ_OP_ZONE_RESET_ALL: 1789 return false; 1790 default: 1791 break; 1792 } 1793 1794 /* 1795 * When mapped devices use the block layer zone write plugging, we must 1796 * split any large BIO to the mapped device limits to not submit BIOs 1797 * that span zone boundaries and to avoid potential deadlocks with 1798 * queue freeze operations. 1799 */ 1800 return bio_needs_zone_write_plugging(bio) || bio_straddles_zones(bio); 1801 } 1802 1803 static inline bool dm_zone_plug_bio(struct mapped_device *md, struct bio *bio) 1804 { 1805 if (!bio_needs_zone_write_plugging(bio)) 1806 return false; 1807 return blk_zone_plug_bio(bio, 0); 1808 } 1809 1810 static blk_status_t __send_zone_reset_all_emulated(struct clone_info *ci, 1811 struct dm_target *ti) 1812 { 1813 struct bio_list blist = BIO_EMPTY_LIST; 1814 struct mapped_device *md = ci->io->md; 1815 unsigned int zone_sectors = md->disk->queue->limits.chunk_sectors; 1816 unsigned long *need_reset; 1817 unsigned int i, nr_zones, nr_reset; 1818 unsigned int num_bios = 0; 1819 blk_status_t sts = BLK_STS_OK; 1820 sector_t sector = ti->begin; 1821 struct bio *clone; 1822 int ret; 1823 1824 nr_zones = ti->len >> ilog2(zone_sectors); 1825 need_reset = bitmap_zalloc(nr_zones, GFP_NOIO); 1826 if (!need_reset) 1827 return BLK_STS_RESOURCE; 1828 1829 ret = dm_zone_get_reset_bitmap(md, ci->map, ti->begin, 1830 nr_zones, need_reset); 1831 if (ret) { 1832 sts = BLK_STS_IOERR; 1833 goto free_bitmap; 1834 } 1835 1836 /* If we have no zone to reset, we are done. */ 1837 nr_reset = bitmap_weight(need_reset, nr_zones); 1838 if (!nr_reset) 1839 goto free_bitmap; 1840 1841 atomic_add(nr_zones, &ci->io->io_count); 1842 1843 for (i = 0; i < nr_zones; i++) { 1844 1845 if (!test_bit(i, need_reset)) { 1846 sector += zone_sectors; 1847 continue; 1848 } 1849 1850 if (bio_list_empty(&blist)) { 1851 /* This may take a while, so be nice to others */ 1852 if (num_bios) 1853 cond_resched(); 1854 1855 /* 1856 * We may need to reset thousands of zones, so let's 1857 * not go crazy with the clone allocation. 1858 */ 1859 alloc_multiple_bios(&blist, ci, ti, min(nr_reset, 32), 1860 NULL); 1861 } 1862 1863 /* Get a clone and change it to a regular reset operation. */ 1864 clone = bio_list_pop(&blist); 1865 clone->bi_opf &= ~REQ_OP_MASK; 1866 clone->bi_opf |= REQ_OP_ZONE_RESET | REQ_SYNC; 1867 clone->bi_iter.bi_sector = sector; 1868 clone->bi_iter.bi_size = 0; 1869 __map_bio(clone); 1870 1871 sector += zone_sectors; 1872 num_bios++; 1873 nr_reset--; 1874 } 1875 1876 WARN_ON_ONCE(!bio_list_empty(&blist)); 1877 atomic_sub(nr_zones - num_bios, &ci->io->io_count); 1878 ci->sector_count = 0; 1879 1880 free_bitmap: 1881 bitmap_free(need_reset); 1882 1883 return sts; 1884 } 1885 1886 static void __send_zone_reset_all_native(struct clone_info *ci, 1887 struct dm_target *ti) 1888 { 1889 unsigned int bios; 1890 1891 atomic_add(1, &ci->io->io_count); 1892 bios = __send_duplicate_bios(ci, ti, 1, NULL); 1893 atomic_sub(1 - bios, &ci->io->io_count); 1894 1895 ci->sector_count = 0; 1896 } 1897 1898 static blk_status_t __send_zone_reset_all(struct clone_info *ci) 1899 { 1900 struct dm_table *t = ci->map; 1901 blk_status_t sts = BLK_STS_OK; 1902 1903 for (unsigned int i = 0; i < t->num_targets; i++) { 1904 struct dm_target *ti = dm_table_get_target(t, i); 1905 1906 if (ti->zone_reset_all_supported) { 1907 __send_zone_reset_all_native(ci, ti); 1908 continue; 1909 } 1910 1911 sts = __send_zone_reset_all_emulated(ci, ti); 1912 if (sts != BLK_STS_OK) 1913 break; 1914 } 1915 1916 /* Release the reference that alloc_io() took for submission. */ 1917 atomic_sub(1, &ci->io->io_count); 1918 1919 return sts; 1920 } 1921 1922 #else 1923 static inline bool dm_zone_bio_needs_split(struct bio *bio) 1924 { 1925 return false; 1926 } 1927 static inline bool dm_zone_plug_bio(struct mapped_device *md, struct bio *bio) 1928 { 1929 return false; 1930 } 1931 static blk_status_t __send_zone_reset_all(struct clone_info *ci) 1932 { 1933 return BLK_STS_NOTSUPP; 1934 } 1935 #endif 1936 1937 /* 1938 * Entry point to split a bio into clones and submit them to the targets. 1939 */ 1940 static void dm_split_and_process_bio(struct mapped_device *md, 1941 struct dm_table *map, struct bio *bio) 1942 { 1943 struct clone_info ci; 1944 struct dm_io *io; 1945 blk_status_t error = BLK_STS_OK; 1946 bool is_abnormal, need_split; 1947 1948 is_abnormal = is_abnormal_io(bio); 1949 if (static_branch_unlikely(&zoned_enabled)) { 1950 need_split = is_abnormal || dm_zone_bio_needs_split(bio); 1951 } else { 1952 need_split = is_abnormal; 1953 } 1954 1955 if (unlikely(need_split)) { 1956 /* 1957 * Use bio_split_to_limits() for abnormal IO (e.g. discard, etc) 1958 * otherwise associated queue_limits won't be imposed. 1959 * Also split the BIO for mapped devices needing zone append 1960 * emulation to ensure that the BIO does not cross zone 1961 * boundaries. 1962 */ 1963 bio = bio_split_to_limits(bio); 1964 if (!bio) 1965 return; 1966 } 1967 1968 /* 1969 * Use the block layer zone write plugging for mapped devices that 1970 * need zone append emulation (e.g. dm-crypt). 1971 */ 1972 if (static_branch_unlikely(&zoned_enabled) && dm_zone_plug_bio(md, bio)) 1973 return; 1974 1975 /* Only support nowait for normal IO */ 1976 if (unlikely(bio->bi_opf & REQ_NOWAIT) && !is_abnormal) { 1977 /* 1978 * Don't support NOWAIT for FLUSH because it may allocate 1979 * multiple bios and there's no easy way how to undo the 1980 * allocations. 1981 */ 1982 if (bio->bi_opf & REQ_PREFLUSH) { 1983 bio_wouldblock_error(bio); 1984 return; 1985 } 1986 io = alloc_io(md, bio, GFP_NOWAIT); 1987 if (unlikely(!io)) { 1988 /* Unable to do anything without dm_io. */ 1989 bio_wouldblock_error(bio); 1990 return; 1991 } 1992 } else { 1993 io = alloc_io(md, bio, GFP_NOIO); 1994 } 1995 init_clone_info(&ci, io, map, bio, is_abnormal); 1996 1997 if (unlikely((bio->bi_opf & REQ_PREFLUSH) != 0)) { 1998 /* 1999 * The "flush_bypasses_map" is set on targets where it is safe 2000 * to skip the map function and submit bios directly to the 2001 * underlying block devices - currently, it is set for dm-linear 2002 * and dm-stripe. 2003 * 2004 * If we have just one underlying device (i.e. there is one 2005 * linear target or multiple linear targets pointing to the same 2006 * device), we can send the flush with data directly to it. 2007 */ 2008 if (map->flush_bypasses_map) { 2009 struct list_head *devices = dm_table_get_devices(map); 2010 if (devices->next == devices->prev) 2011 goto send_preflush_with_data; 2012 } 2013 if (bio->bi_iter.bi_size) 2014 io->requeue_flush_with_data = true; 2015 __send_empty_flush(&ci); 2016 /* dm_io_complete submits any data associated with flush */ 2017 goto out; 2018 } 2019 2020 send_preflush_with_data: 2021 if (static_branch_unlikely(&zoned_enabled) && 2022 (bio_op(bio) == REQ_OP_ZONE_RESET_ALL)) { 2023 error = __send_zone_reset_all(&ci); 2024 goto out; 2025 } 2026 2027 error = __split_and_process_bio(&ci); 2028 if (error || !ci.sector_count) 2029 goto out; 2030 /* 2031 * Remainder must be passed to submit_bio_noacct() so it gets handled 2032 * *after* bios already submitted have been completely processed. 2033 */ 2034 bio_trim(bio, io->sectors, ci.sector_count); 2035 trace_block_split(bio, bio->bi_iter.bi_sector); 2036 bio_inc_remaining(bio); 2037 submit_bio_noacct(bio); 2038 out: 2039 /* 2040 * Drop the extra reference count for non-POLLED bio, and hold one 2041 * reference for POLLED bio, which will be released in dm_poll_bio 2042 * 2043 * Add every dm_io instance into the dm_io list head which is stored 2044 * in bio->bi_private, so that dm_poll_bio can poll them all. 2045 */ 2046 if (error || !ci.submit_as_polled) { 2047 /* 2048 * In case of submission failure, the extra reference for 2049 * submitting io isn't consumed yet 2050 */ 2051 if (error) 2052 atomic_dec(&io->io_count); 2053 dm_io_dec_pending(io, error); 2054 } else 2055 dm_queue_poll_io(bio, io); 2056 } 2057 2058 static void dm_submit_bio(struct bio *bio) 2059 { 2060 struct mapped_device *md = bio->bi_bdev->bd_disk->private_data; 2061 int srcu_idx; 2062 struct dm_table *map; 2063 2064 map = dm_get_live_table(md, &srcu_idx); 2065 if (unlikely(!map)) { 2066 DMERR_LIMIT("%s: mapping table unavailable, erroring io", 2067 dm_device_name(md)); 2068 bio_io_error(bio); 2069 goto out; 2070 } 2071 2072 /* If suspended, queue this IO for later */ 2073 if (unlikely(test_bit(DMF_BLOCK_IO_FOR_SUSPEND, &md->flags))) { 2074 if (bio->bi_opf & REQ_NOWAIT) 2075 bio_wouldblock_error(bio); 2076 else if (bio->bi_opf & REQ_RAHEAD) 2077 bio_io_error(bio); 2078 else 2079 queue_io(md, bio); 2080 goto out; 2081 } 2082 2083 dm_split_and_process_bio(md, map, bio); 2084 out: 2085 dm_put_live_table(md, srcu_idx); 2086 } 2087 2088 static bool dm_poll_dm_io(struct dm_io *io, struct io_comp_batch *iob, 2089 unsigned int flags) 2090 { 2091 WARN_ON_ONCE(!dm_tio_is_normal(&io->tio)); 2092 2093 /* don't poll if the mapped io is done */ 2094 if (atomic_read(&io->io_count) > 1) 2095 bio_poll(&io->tio.clone, iob, flags); 2096 2097 /* bio_poll holds the last reference */ 2098 return atomic_read(&io->io_count) == 1; 2099 } 2100 2101 static int dm_poll_bio(struct bio *bio, struct io_comp_batch *iob, 2102 unsigned int flags) 2103 { 2104 struct dm_io **head = dm_poll_list_head(bio); 2105 struct dm_io *list = *head; 2106 struct dm_io *tmp = NULL; 2107 struct dm_io *curr, *next; 2108 2109 /* Only poll normal bio which was marked as REQ_DM_POLL_LIST */ 2110 if (!(bio->bi_opf & REQ_DM_POLL_LIST)) 2111 return 0; 2112 2113 WARN_ON_ONCE(!list); 2114 2115 /* 2116 * Restore .bi_private before possibly completing dm_io. 2117 * 2118 * bio_poll() is only possible once @bio has been completely 2119 * submitted via submit_bio_noacct()'s depth-first submission. 2120 * So there is no dm_queue_poll_io() race associated with 2121 * clearing REQ_DM_POLL_LIST here. 2122 */ 2123 bio->bi_opf &= ~REQ_DM_POLL_LIST; 2124 bio->bi_private = list->data; 2125 2126 for (curr = list, next = curr->next; curr; curr = next, next = 2127 curr ? curr->next : NULL) { 2128 if (dm_poll_dm_io(curr, iob, flags)) { 2129 /* 2130 * clone_endio() has already occurred, so no 2131 * error handling is needed here. 2132 */ 2133 __dm_io_dec_pending(curr); 2134 } else { 2135 curr->next = tmp; 2136 tmp = curr; 2137 } 2138 } 2139 2140 /* Not done? */ 2141 if (tmp) { 2142 bio->bi_opf |= REQ_DM_POLL_LIST; 2143 /* Reset bio->bi_private to dm_io list head */ 2144 *head = tmp; 2145 return 0; 2146 } 2147 return 1; 2148 } 2149 2150 /* 2151 *--------------------------------------------------------------- 2152 * An IDR is used to keep track of allocated minor numbers. 2153 *--------------------------------------------------------------- 2154 */ 2155 static void free_minor(int minor) 2156 { 2157 spin_lock(&_minor_lock); 2158 idr_remove(&_minor_idr, minor); 2159 spin_unlock(&_minor_lock); 2160 } 2161 2162 /* 2163 * See if the device with a specific minor # is free. 2164 */ 2165 static int specific_minor(int minor) 2166 { 2167 int r; 2168 2169 if (minor >= (1 << MINORBITS)) 2170 return -EINVAL; 2171 2172 idr_preload(GFP_KERNEL); 2173 spin_lock(&_minor_lock); 2174 2175 r = idr_alloc(&_minor_idr, MINOR_ALLOCED, minor, minor + 1, GFP_NOWAIT); 2176 2177 spin_unlock(&_minor_lock); 2178 idr_preload_end(); 2179 if (r < 0) 2180 return r == -ENOSPC ? -EBUSY : r; 2181 return 0; 2182 } 2183 2184 static int next_free_minor(int *minor) 2185 { 2186 int r; 2187 2188 idr_preload(GFP_KERNEL); 2189 spin_lock(&_minor_lock); 2190 2191 r = idr_alloc(&_minor_idr, MINOR_ALLOCED, 0, 1 << MINORBITS, GFP_NOWAIT); 2192 2193 spin_unlock(&_minor_lock); 2194 idr_preload_end(); 2195 if (r < 0) 2196 return r; 2197 *minor = r; 2198 return 0; 2199 } 2200 2201 static const struct block_device_operations dm_blk_dops; 2202 static const struct block_device_operations dm_rq_blk_dops; 2203 static const struct dax_operations dm_dax_ops; 2204 2205 static void dm_wq_work(struct work_struct *work); 2206 2207 #ifdef CONFIG_BLK_INLINE_ENCRYPTION 2208 static void dm_queue_destroy_crypto_profile(struct request_queue *q) 2209 { 2210 dm_destroy_crypto_profile(q->crypto_profile); 2211 } 2212 2213 #else /* CONFIG_BLK_INLINE_ENCRYPTION */ 2214 2215 static inline void dm_queue_destroy_crypto_profile(struct request_queue *q) 2216 { 2217 } 2218 #endif /* !CONFIG_BLK_INLINE_ENCRYPTION */ 2219 2220 static void cleanup_mapped_device(struct mapped_device *md) 2221 { 2222 if (md->wq) 2223 destroy_workqueue(md->wq); 2224 dm_free_md_mempools(md->mempools); 2225 2226 if (md->dax_dev) { 2227 dax_remove_host(md->disk); 2228 kill_dax(md->dax_dev); 2229 put_dax(md->dax_dev); 2230 md->dax_dev = NULL; 2231 } 2232 2233 if (md->disk) { 2234 spin_lock(&_minor_lock); 2235 md->disk->private_data = NULL; 2236 spin_unlock(&_minor_lock); 2237 if (dm_get_md_type(md) != DM_TYPE_NONE) { 2238 struct table_device *td; 2239 2240 dm_sysfs_exit(md); 2241 list_for_each_entry(td, &md->table_devices, list) { 2242 bd_unlink_disk_holder(td->dm_dev.bdev, 2243 md->disk); 2244 } 2245 2246 /* 2247 * Hold lock to make sure del_gendisk() won't concurrent 2248 * with open/close_table_device(). 2249 */ 2250 mutex_lock(&md->table_devices_lock); 2251 del_gendisk(md->disk); 2252 mutex_unlock(&md->table_devices_lock); 2253 } 2254 dm_queue_destroy_crypto_profile(md->queue); 2255 put_disk(md->disk); 2256 } 2257 2258 if (md->pending_io) { 2259 free_percpu(md->pending_io); 2260 md->pending_io = NULL; 2261 } 2262 2263 cleanup_srcu_struct(&md->io_barrier); 2264 2265 mutex_destroy(&md->suspend_lock); 2266 mutex_destroy(&md->type_lock); 2267 mutex_destroy(&md->table_devices_lock); 2268 mutex_destroy(&md->swap_bios_lock); 2269 2270 dm_mq_cleanup_mapped_device(md); 2271 } 2272 2273 /* 2274 * Allocate and initialise a blank device with a given minor. 2275 */ 2276 static struct mapped_device *alloc_dev(int minor) 2277 { 2278 int r, numa_node_id = dm_get_numa_node(); 2279 struct dax_device *dax_dev; 2280 struct mapped_device *md; 2281 void *old_md; 2282 2283 md = kvzalloc_node(sizeof(*md), GFP_KERNEL, numa_node_id); 2284 if (!md) { 2285 DMERR("unable to allocate device, out of memory."); 2286 return NULL; 2287 } 2288 2289 if (!try_module_get(THIS_MODULE)) 2290 goto bad_module_get; 2291 2292 /* get a minor number for the dev */ 2293 if (minor == DM_ANY_MINOR) 2294 r = next_free_minor(&minor); 2295 else 2296 r = specific_minor(minor); 2297 if (r < 0) 2298 goto bad_minor; 2299 2300 r = init_srcu_struct(&md->io_barrier); 2301 if (r < 0) 2302 goto bad_io_barrier; 2303 2304 md->numa_node_id = numa_node_id; 2305 md->init_tio_pdu = false; 2306 md->type = DM_TYPE_NONE; 2307 mutex_init(&md->suspend_lock); 2308 mutex_init(&md->type_lock); 2309 mutex_init(&md->table_devices_lock); 2310 spin_lock_init(&md->deferred_lock); 2311 atomic_set(&md->holders, 1); 2312 atomic_set(&md->open_count, 0); 2313 atomic_set(&md->event_nr, 0); 2314 atomic_set(&md->uevent_seq, 0); 2315 INIT_LIST_HEAD(&md->uevent_list); 2316 INIT_LIST_HEAD(&md->table_devices); 2317 spin_lock_init(&md->uevent_lock); 2318 2319 /* 2320 * default to bio-based until DM table is loaded and md->type 2321 * established. If request-based table is loaded: blk-mq will 2322 * override accordingly. 2323 */ 2324 md->disk = blk_alloc_disk(NULL, md->numa_node_id); 2325 if (IS_ERR(md->disk)) { 2326 md->disk = NULL; 2327 goto bad; 2328 } 2329 md->queue = md->disk->queue; 2330 2331 init_waitqueue_head(&md->wait); 2332 INIT_WORK(&md->work, dm_wq_work); 2333 INIT_WORK(&md->requeue_work, dm_wq_requeue_work); 2334 init_waitqueue_head(&md->eventq); 2335 init_completion(&md->kobj_holder.completion); 2336 2337 md->requeue_list = NULL; 2338 md->swap_bios = get_swap_bios(); 2339 sema_init(&md->swap_bios_semaphore, md->swap_bios); 2340 mutex_init(&md->swap_bios_lock); 2341 2342 md->disk->major = _major; 2343 md->disk->first_minor = minor; 2344 md->disk->minors = 1; 2345 md->disk->flags |= GENHD_FL_NO_PART; 2346 md->disk->fops = &dm_blk_dops; 2347 md->disk->private_data = md; 2348 sprintf(md->disk->disk_name, "dm-%d", minor); 2349 2350 dax_dev = alloc_dax(md, &dm_dax_ops); 2351 if (IS_ERR(dax_dev)) { 2352 if (PTR_ERR(dax_dev) != -EOPNOTSUPP) 2353 goto bad; 2354 } else { 2355 set_dax_nocache(dax_dev); 2356 set_dax_nomc(dax_dev); 2357 md->dax_dev = dax_dev; 2358 if (dax_add_host(dax_dev, md->disk)) 2359 goto bad; 2360 } 2361 2362 format_dev_t(md->name, MKDEV(_major, minor)); 2363 2364 md->wq = alloc_workqueue("kdmflush/%s", WQ_MEM_RECLAIM, 0, md->name); 2365 if (!md->wq) 2366 goto bad; 2367 2368 md->pending_io = alloc_percpu(unsigned long); 2369 if (!md->pending_io) 2370 goto bad; 2371 2372 r = dm_stats_init(&md->stats); 2373 if (r < 0) 2374 goto bad; 2375 2376 /* Populate the mapping, nobody knows we exist yet */ 2377 spin_lock(&_minor_lock); 2378 old_md = idr_replace(&_minor_idr, md, minor); 2379 spin_unlock(&_minor_lock); 2380 2381 BUG_ON(old_md != MINOR_ALLOCED); 2382 2383 return md; 2384 2385 bad: 2386 cleanup_mapped_device(md); 2387 bad_io_barrier: 2388 free_minor(minor); 2389 bad_minor: 2390 module_put(THIS_MODULE); 2391 bad_module_get: 2392 kvfree(md); 2393 return NULL; 2394 } 2395 2396 static void unlock_fs(struct mapped_device *md); 2397 2398 static void free_dev(struct mapped_device *md) 2399 { 2400 int minor = MINOR(disk_devt(md->disk)); 2401 2402 unlock_fs(md); 2403 2404 cleanup_mapped_device(md); 2405 2406 WARN_ON_ONCE(!list_empty(&md->table_devices)); 2407 dm_stats_cleanup(&md->stats); 2408 free_minor(minor); 2409 2410 module_put(THIS_MODULE); 2411 kvfree(md); 2412 } 2413 2414 /* 2415 * Bind a table to the device. 2416 */ 2417 static void event_callback(void *context) 2418 { 2419 unsigned long flags; 2420 LIST_HEAD(uevents); 2421 struct mapped_device *md = context; 2422 2423 spin_lock_irqsave(&md->uevent_lock, flags); 2424 list_splice_init(&md->uevent_list, &uevents); 2425 spin_unlock_irqrestore(&md->uevent_lock, flags); 2426 2427 dm_send_uevents(&uevents, &disk_to_dev(md->disk)->kobj); 2428 2429 atomic_inc(&md->event_nr); 2430 wake_up(&md->eventq); 2431 dm_issue_global_event(); 2432 } 2433 2434 /* 2435 * Returns old map, which caller must destroy. 2436 */ 2437 static struct dm_table *__bind(struct mapped_device *md, struct dm_table *t, 2438 struct queue_limits *limits) 2439 { 2440 struct dm_table *old_map; 2441 sector_t size, old_size; 2442 2443 lockdep_assert_held(&md->suspend_lock); 2444 2445 size = dm_table_get_size(t); 2446 2447 old_size = dm_get_size(md); 2448 2449 if (!dm_table_supports_size_change(t, old_size, size)) { 2450 old_map = ERR_PTR(-EINVAL); 2451 goto out; 2452 } 2453 2454 set_capacity(md->disk, size); 2455 2456 if (limits) { 2457 int ret = dm_table_set_restrictions(t, md->queue, limits); 2458 if (ret) { 2459 set_capacity(md->disk, old_size); 2460 old_map = ERR_PTR(ret); 2461 goto out; 2462 } 2463 } 2464 2465 /* 2466 * Wipe any geometry if the size of the table changed. 2467 */ 2468 if (size != old_size) 2469 memset(&md->geometry, 0, sizeof(md->geometry)); 2470 2471 dm_table_event_callback(t, event_callback, md); 2472 2473 if (dm_table_request_based(t)) { 2474 /* 2475 * Leverage the fact that request-based DM targets are 2476 * immutable singletons - used to optimize dm_mq_queue_rq. 2477 */ 2478 md->immutable_target = dm_table_get_immutable_target(t); 2479 2480 /* 2481 * There is no need to reload with request-based dm because the 2482 * size of front_pad doesn't change. 2483 * 2484 * Note for future: If you are to reload bioset, prep-ed 2485 * requests in the queue may refer to bio from the old bioset, 2486 * so you must walk through the queue to unprep. 2487 */ 2488 if (!md->mempools) 2489 md->mempools = t->mempools; 2490 else 2491 dm_free_md_mempools(t->mempools); 2492 } else { 2493 /* 2494 * The md may already have mempools that need changing. 2495 * If so, reload bioset because front_pad may have changed 2496 * because a different table was loaded. 2497 */ 2498 dm_free_md_mempools(md->mempools); 2499 md->mempools = t->mempools; 2500 } 2501 t->mempools = NULL; 2502 2503 old_map = rcu_dereference_protected(md->map, lockdep_is_held(&md->suspend_lock)); 2504 rcu_assign_pointer(md->map, (void *)t); 2505 md->immutable_target_type = dm_table_get_immutable_target_type(t); 2506 2507 if (old_map) 2508 dm_sync_table(md); 2509 out: 2510 return old_map; 2511 } 2512 2513 /* 2514 * Returns unbound table for the caller to free. 2515 */ 2516 static struct dm_table *__unbind(struct mapped_device *md) 2517 { 2518 struct dm_table *map = rcu_dereference_protected(md->map, 1); 2519 2520 if (!map) 2521 return NULL; 2522 2523 dm_table_event_callback(map, NULL, NULL); 2524 RCU_INIT_POINTER(md->map, NULL); 2525 dm_sync_table(md); 2526 2527 return map; 2528 } 2529 2530 /* 2531 * Constructor for a new device. 2532 */ 2533 int dm_create(int minor, struct mapped_device **result) 2534 { 2535 struct mapped_device *md; 2536 2537 md = alloc_dev(minor); 2538 if (!md) 2539 return -ENXIO; 2540 2541 dm_ima_reset_data(md); 2542 2543 *result = md; 2544 return 0; 2545 } 2546 2547 /* 2548 * Functions to manage md->type. 2549 * All are required to hold md->type_lock. 2550 */ 2551 void dm_lock_md_type(struct mapped_device *md) 2552 { 2553 mutex_lock(&md->type_lock); 2554 } 2555 2556 void dm_unlock_md_type(struct mapped_device *md) 2557 { 2558 mutex_unlock(&md->type_lock); 2559 } 2560 2561 enum dm_queue_mode dm_get_md_type(struct mapped_device *md) 2562 { 2563 return md->type; 2564 } 2565 2566 struct target_type *dm_get_immutable_target_type(struct mapped_device *md) 2567 { 2568 return md->immutable_target_type; 2569 } 2570 2571 /* 2572 * Setup the DM device's queue based on md's type 2573 */ 2574 int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t) 2575 { 2576 enum dm_queue_mode type = dm_table_get_type(t); 2577 struct queue_limits limits; 2578 struct table_device *td; 2579 int r; 2580 2581 WARN_ON_ONCE(type == DM_TYPE_NONE); 2582 2583 if (type == DM_TYPE_REQUEST_BASED) { 2584 md->disk->fops = &dm_rq_blk_dops; 2585 r = dm_mq_init_request_queue(md, t); 2586 if (r) { 2587 DMERR("Cannot initialize queue for request-based dm mapped device"); 2588 return r; 2589 } 2590 } 2591 2592 r = dm_calculate_queue_limits(t, &limits); 2593 if (r) { 2594 DMERR("Cannot calculate initial queue limits"); 2595 return r; 2596 } 2597 r = dm_table_set_restrictions(t, md->queue, &limits); 2598 if (r) 2599 return r; 2600 2601 /* 2602 * Hold lock to make sure add_disk() and del_gendisk() won't concurrent 2603 * with open_table_device() and close_table_device(). 2604 */ 2605 mutex_lock(&md->table_devices_lock); 2606 r = add_disk(md->disk); 2607 mutex_unlock(&md->table_devices_lock); 2608 if (r) 2609 return r; 2610 2611 /* 2612 * Register the holder relationship for devices added before the disk 2613 * was live. 2614 */ 2615 list_for_each_entry(td, &md->table_devices, list) { 2616 r = bd_link_disk_holder(td->dm_dev.bdev, md->disk); 2617 if (r) 2618 goto out_undo_holders; 2619 } 2620 2621 r = dm_sysfs_init(md); 2622 if (r) 2623 goto out_undo_holders; 2624 2625 md->type = type; 2626 return 0; 2627 2628 out_undo_holders: 2629 list_for_each_entry_continue_reverse(td, &md->table_devices, list) 2630 bd_unlink_disk_holder(td->dm_dev.bdev, md->disk); 2631 mutex_lock(&md->table_devices_lock); 2632 del_gendisk(md->disk); 2633 mutex_unlock(&md->table_devices_lock); 2634 return r; 2635 } 2636 2637 struct mapped_device *dm_get_md(dev_t dev) 2638 { 2639 struct mapped_device *md; 2640 unsigned int minor = MINOR(dev); 2641 2642 if (MAJOR(dev) != _major || minor >= (1 << MINORBITS)) 2643 return NULL; 2644 2645 spin_lock(&_minor_lock); 2646 2647 md = idr_find(&_minor_idr, minor); 2648 if (!md || md == MINOR_ALLOCED || (MINOR(disk_devt(dm_disk(md))) != minor) || 2649 test_bit(DMF_FREEING, &md->flags) || dm_deleting_md(md)) { 2650 md = NULL; 2651 goto out; 2652 } 2653 dm_get(md); 2654 out: 2655 spin_unlock(&_minor_lock); 2656 2657 return md; 2658 } 2659 EXPORT_SYMBOL_GPL(dm_get_md); 2660 2661 void *dm_get_mdptr(struct mapped_device *md) 2662 { 2663 return md->interface_ptr; 2664 } 2665 2666 void dm_set_mdptr(struct mapped_device *md, void *ptr) 2667 { 2668 md->interface_ptr = ptr; 2669 } 2670 2671 void dm_get(struct mapped_device *md) 2672 { 2673 atomic_inc(&md->holders); 2674 BUG_ON(test_bit(DMF_FREEING, &md->flags)); 2675 } 2676 2677 int dm_hold(struct mapped_device *md) 2678 { 2679 spin_lock(&_minor_lock); 2680 if (test_bit(DMF_FREEING, &md->flags)) { 2681 spin_unlock(&_minor_lock); 2682 return -EBUSY; 2683 } 2684 dm_get(md); 2685 spin_unlock(&_minor_lock); 2686 return 0; 2687 } 2688 EXPORT_SYMBOL_GPL(dm_hold); 2689 2690 const char *dm_device_name(struct mapped_device *md) 2691 { 2692 return md->name; 2693 } 2694 EXPORT_SYMBOL_GPL(dm_device_name); 2695 2696 static void __dm_destroy(struct mapped_device *md, bool wait) 2697 { 2698 struct dm_table *map; 2699 int srcu_idx; 2700 2701 might_sleep(); 2702 2703 spin_lock(&_minor_lock); 2704 idr_replace(&_minor_idr, MINOR_ALLOCED, MINOR(disk_devt(dm_disk(md)))); 2705 set_bit(DMF_FREEING, &md->flags); 2706 spin_unlock(&_minor_lock); 2707 2708 blk_mark_disk_dead(md->disk); 2709 2710 /* 2711 * Take suspend_lock so that presuspend and postsuspend methods 2712 * do not race with internal suspend. 2713 */ 2714 mutex_lock(&md->suspend_lock); 2715 map = dm_get_live_table(md, &srcu_idx); 2716 if (!dm_suspended_md(md)) { 2717 dm_table_presuspend_targets(map); 2718 set_bit(DMF_SUSPENDED, &md->flags); 2719 set_bit(DMF_POST_SUSPENDING, &md->flags); 2720 dm_table_postsuspend_targets(map); 2721 } 2722 /* dm_put_live_table must be before fsleep, otherwise deadlock is possible */ 2723 dm_put_live_table(md, srcu_idx); 2724 mutex_unlock(&md->suspend_lock); 2725 2726 /* 2727 * Rare, but there may be I/O requests still going to complete, 2728 * for example. Wait for all references to disappear. 2729 * No one should increment the reference count of the mapped_device, 2730 * after the mapped_device state becomes DMF_FREEING. 2731 */ 2732 if (wait) 2733 while (atomic_read(&md->holders)) 2734 fsleep(1000); 2735 else if (atomic_read(&md->holders)) 2736 DMWARN("%s: Forcibly removing mapped_device still in use! (%d users)", 2737 dm_device_name(md), atomic_read(&md->holders)); 2738 2739 dm_table_destroy(__unbind(md)); 2740 free_dev(md); 2741 } 2742 2743 void dm_destroy(struct mapped_device *md) 2744 { 2745 __dm_destroy(md, true); 2746 } 2747 2748 void dm_destroy_immediate(struct mapped_device *md) 2749 { 2750 __dm_destroy(md, false); 2751 } 2752 2753 void dm_put(struct mapped_device *md) 2754 { 2755 atomic_dec(&md->holders); 2756 } 2757 EXPORT_SYMBOL_GPL(dm_put); 2758 2759 static bool dm_in_flight_bios(struct mapped_device *md) 2760 { 2761 int cpu; 2762 unsigned long sum = 0; 2763 2764 for_each_possible_cpu(cpu) 2765 sum += *per_cpu_ptr(md->pending_io, cpu); 2766 2767 return sum != 0; 2768 } 2769 2770 static int dm_wait_for_bios_completion(struct mapped_device *md, unsigned int task_state) 2771 { 2772 int r = 0; 2773 DEFINE_WAIT(wait); 2774 2775 while (true) { 2776 prepare_to_wait(&md->wait, &wait, task_state); 2777 2778 if (!dm_in_flight_bios(md)) 2779 break; 2780 2781 if (signal_pending_state(task_state, current)) { 2782 r = -ERESTARTSYS; 2783 break; 2784 } 2785 2786 io_schedule(); 2787 } 2788 finish_wait(&md->wait, &wait); 2789 2790 smp_rmb(); 2791 2792 return r; 2793 } 2794 2795 static int dm_wait_for_completion(struct mapped_device *md, unsigned int task_state) 2796 { 2797 int r = 0; 2798 2799 if (!queue_is_mq(md->queue)) 2800 return dm_wait_for_bios_completion(md, task_state); 2801 2802 while (true) { 2803 if (!blk_mq_queue_inflight(md->queue)) 2804 break; 2805 2806 if (signal_pending_state(task_state, current)) { 2807 r = -ERESTARTSYS; 2808 break; 2809 } 2810 2811 fsleep(5000); 2812 } 2813 2814 return r; 2815 } 2816 2817 /* 2818 * Process the deferred bios 2819 */ 2820 static void dm_wq_work(struct work_struct *work) 2821 { 2822 struct mapped_device *md = container_of(work, struct mapped_device, work); 2823 struct bio *bio; 2824 2825 while (!test_bit(DMF_BLOCK_IO_FOR_SUSPEND, &md->flags)) { 2826 spin_lock_irq(&md->deferred_lock); 2827 bio = bio_list_pop(&md->deferred); 2828 spin_unlock_irq(&md->deferred_lock); 2829 2830 if (!bio) 2831 break; 2832 2833 submit_bio_noacct(bio); 2834 cond_resched(); 2835 } 2836 } 2837 2838 static void dm_queue_flush(struct mapped_device *md) 2839 { 2840 clear_bit(DMF_NOFLUSH_SUSPENDING, &md->flags); 2841 clear_bit(DMF_BLOCK_IO_FOR_SUSPEND, &md->flags); 2842 smp_mb__after_atomic(); 2843 queue_work(md->wq, &md->work); 2844 } 2845 2846 /* 2847 * Swap in a new table, returning the old one for the caller to destroy. 2848 */ 2849 struct dm_table *dm_swap_table(struct mapped_device *md, struct dm_table *table) 2850 { 2851 struct dm_table *live_map = NULL, *map = ERR_PTR(-EINVAL); 2852 struct queue_limits limits; 2853 bool update_limits = true; 2854 int r; 2855 2856 mutex_lock(&md->suspend_lock); 2857 2858 /* device must be suspended */ 2859 if (!dm_suspended_md(md)) 2860 goto out; 2861 2862 /* 2863 * To avoid a potential deadlock locking the queue limits, disallow 2864 * updating the queue limits during a table swap, when updating an 2865 * immutable request-based dm device (dm-multipath) during a noflush 2866 * suspend. It is userspace's responsibility to make sure that the new 2867 * table uses the same limits as the existing table, if it asks for a 2868 * noflush suspend. 2869 */ 2870 if (dm_request_based(md) && md->immutable_target && 2871 __noflush_suspending(md)) 2872 update_limits = false; 2873 /* 2874 * If the new table has no data devices, retain the existing limits. 2875 * This helps multipath with queue_if_no_path if all paths disappear, 2876 * then new I/O is queued based on these limits, and then some paths 2877 * reappear. 2878 */ 2879 else if (dm_table_has_no_data_devices(table)) { 2880 live_map = dm_get_live_table_fast(md); 2881 if (live_map) 2882 limits = md->queue->limits; 2883 dm_put_live_table_fast(md); 2884 } 2885 2886 if (update_limits && !live_map) { 2887 r = dm_calculate_queue_limits(table, &limits); 2888 if (r) { 2889 map = ERR_PTR(r); 2890 goto out; 2891 } 2892 } 2893 2894 map = __bind(md, table, update_limits ? &limits : NULL); 2895 dm_issue_global_event(); 2896 2897 out: 2898 mutex_unlock(&md->suspend_lock); 2899 return map; 2900 } 2901 2902 /* 2903 * Functions to lock and unlock any filesystem running on the 2904 * device. 2905 */ 2906 static int lock_fs(struct mapped_device *md) 2907 { 2908 int r; 2909 2910 WARN_ON(test_bit(DMF_FROZEN, &md->flags)); 2911 2912 r = bdev_freeze(md->disk->part0); 2913 if (!r) 2914 set_bit(DMF_FROZEN, &md->flags); 2915 return r; 2916 } 2917 2918 static void unlock_fs(struct mapped_device *md) 2919 { 2920 if (!test_bit(DMF_FROZEN, &md->flags)) 2921 return; 2922 bdev_thaw(md->disk->part0); 2923 clear_bit(DMF_FROZEN, &md->flags); 2924 } 2925 2926 /* 2927 * @suspend_flags: DM_SUSPEND_LOCKFS_FLAG and/or DM_SUSPEND_NOFLUSH_FLAG 2928 * @task_state: e.g. TASK_INTERRUPTIBLE or TASK_UNINTERRUPTIBLE 2929 * @dmf_suspended_flag: DMF_SUSPENDED or DMF_SUSPENDED_INTERNALLY 2930 * 2931 * If __dm_suspend returns 0, the device is completely quiescent 2932 * now. There is no request-processing activity. All new requests 2933 * are being added to md->deferred list. 2934 */ 2935 static int __dm_suspend(struct mapped_device *md, struct dm_table *map, 2936 unsigned int suspend_flags, unsigned int task_state, 2937 int dmf_suspended_flag) 2938 { 2939 bool do_lockfs = suspend_flags & DM_SUSPEND_LOCKFS_FLAG; 2940 bool noflush = suspend_flags & DM_SUSPEND_NOFLUSH_FLAG; 2941 int r = 0; 2942 2943 lockdep_assert_held(&md->suspend_lock); 2944 2945 /* 2946 * DMF_NOFLUSH_SUSPENDING must be set before presuspend. 2947 */ 2948 if (noflush) 2949 set_bit(DMF_NOFLUSH_SUSPENDING, &md->flags); 2950 else 2951 DMDEBUG("%s: suspending with flush", dm_device_name(md)); 2952 2953 /* 2954 * This gets reverted if there's an error later and the targets 2955 * provide the .presuspend_undo hook. 2956 */ 2957 dm_table_presuspend_targets(map); 2958 2959 /* 2960 * Flush I/O to the device. 2961 * Any I/O submitted after lock_fs() may not be flushed. 2962 * noflush takes precedence over do_lockfs. 2963 * (lock_fs() flushes I/Os and waits for them to complete.) 2964 */ 2965 if (!noflush && do_lockfs) { 2966 r = lock_fs(md); 2967 if (r) { 2968 dm_table_presuspend_undo_targets(map); 2969 return r; 2970 } 2971 } 2972 2973 /* 2974 * Here we must make sure that no processes are submitting requests 2975 * to target drivers i.e. no one may be executing 2976 * dm_split_and_process_bio from dm_submit_bio. 2977 * 2978 * To get all processes out of dm_split_and_process_bio in dm_submit_bio, 2979 * we take the write lock. To prevent any process from reentering 2980 * dm_split_and_process_bio from dm_submit_bio and quiesce the thread 2981 * (dm_wq_work), we set DMF_BLOCK_IO_FOR_SUSPEND and call 2982 * flush_workqueue(md->wq). 2983 */ 2984 set_bit(DMF_BLOCK_IO_FOR_SUSPEND, &md->flags); 2985 if (map) 2986 synchronize_srcu(&md->io_barrier); 2987 2988 /* 2989 * Stop md->queue before flushing md->wq in case request-based 2990 * dm defers requests to md->wq from md->queue. 2991 */ 2992 if (map && dm_request_based(md)) { 2993 dm_stop_queue(md->queue); 2994 set_bit(DMF_QUEUE_STOPPED, &md->flags); 2995 } 2996 2997 flush_workqueue(md->wq); 2998 2999 /* 3000 * At this point no more requests are entering target request routines. 3001 * We call dm_wait_for_completion to wait for all existing requests 3002 * to finish. 3003 */ 3004 if (map) 3005 r = dm_wait_for_completion(md, task_state); 3006 if (!r) 3007 set_bit(dmf_suspended_flag, &md->flags); 3008 3009 if (map) 3010 synchronize_srcu(&md->io_barrier); 3011 3012 /* were we interrupted ? */ 3013 if (r < 0) { 3014 dm_queue_flush(md); 3015 3016 if (test_and_clear_bit(DMF_QUEUE_STOPPED, &md->flags)) 3017 dm_start_queue(md->queue); 3018 3019 unlock_fs(md); 3020 dm_table_presuspend_undo_targets(map); 3021 /* pushback list is already flushed, so skip flush */ 3022 } 3023 3024 return r; 3025 } 3026 3027 /* 3028 * We need to be able to change a mapping table under a mounted 3029 * filesystem. For example we might want to move some data in 3030 * the background. Before the table can be swapped with 3031 * dm_bind_table, dm_suspend must be called to flush any in 3032 * flight bios and ensure that any further io gets deferred. 3033 */ 3034 /* 3035 * Suspend mechanism in request-based dm. 3036 * 3037 * 1. Flush all I/Os by lock_fs() if needed. 3038 * 2. Stop dispatching any I/O by stopping the request_queue. 3039 * 3. Wait for all in-flight I/Os to be completed or requeued. 3040 * 3041 * To abort suspend, start the request_queue. 3042 */ 3043 int dm_suspend(struct mapped_device *md, unsigned int suspend_flags) 3044 { 3045 struct dm_table *map = NULL; 3046 int r = 0; 3047 3048 retry: 3049 mutex_lock_nested(&md->suspend_lock, SINGLE_DEPTH_NESTING); 3050 3051 if (dm_suspended_md(md)) { 3052 r = -EINVAL; 3053 goto out_unlock; 3054 } 3055 3056 if (dm_suspended_internally_md(md)) { 3057 /* already internally suspended, wait for internal resume */ 3058 mutex_unlock(&md->suspend_lock); 3059 r = wait_on_bit(&md->flags, DMF_SUSPENDED_INTERNALLY, TASK_INTERRUPTIBLE); 3060 if (r) 3061 return r; 3062 goto retry; 3063 } 3064 3065 map = rcu_dereference_protected(md->map, lockdep_is_held(&md->suspend_lock)); 3066 if (!map) { 3067 /* avoid deadlock with fs/namespace.c:do_mount() */ 3068 suspend_flags &= ~DM_SUSPEND_LOCKFS_FLAG; 3069 } 3070 3071 r = __dm_suspend(md, map, suspend_flags, TASK_INTERRUPTIBLE, DMF_SUSPENDED); 3072 if (r) 3073 goto out_unlock; 3074 3075 set_bit(DMF_POST_SUSPENDING, &md->flags); 3076 dm_table_postsuspend_targets(map); 3077 clear_bit(DMF_POST_SUSPENDING, &md->flags); 3078 3079 out_unlock: 3080 mutex_unlock(&md->suspend_lock); 3081 return r; 3082 } 3083 3084 static int __dm_resume(struct mapped_device *md, struct dm_table *map) 3085 { 3086 if (map) { 3087 int r = dm_table_resume_targets(map); 3088 3089 if (r) 3090 return r; 3091 } 3092 3093 dm_queue_flush(md); 3094 3095 /* 3096 * Flushing deferred I/Os must be done after targets are resumed 3097 * so that mapping of targets can work correctly. 3098 * Request-based dm is queueing the deferred I/Os in its request_queue. 3099 */ 3100 if (test_and_clear_bit(DMF_QUEUE_STOPPED, &md->flags)) 3101 dm_start_queue(md->queue); 3102 3103 unlock_fs(md); 3104 3105 return 0; 3106 } 3107 3108 int dm_resume(struct mapped_device *md) 3109 { 3110 int r; 3111 struct dm_table *map = NULL; 3112 3113 retry: 3114 r = -EINVAL; 3115 mutex_lock_nested(&md->suspend_lock, SINGLE_DEPTH_NESTING); 3116 3117 if (!dm_suspended_md(md)) 3118 goto out; 3119 3120 if (dm_suspended_internally_md(md)) { 3121 /* already internally suspended, wait for internal resume */ 3122 mutex_unlock(&md->suspend_lock); 3123 r = wait_on_bit(&md->flags, DMF_SUSPENDED_INTERNALLY, TASK_INTERRUPTIBLE); 3124 if (r) 3125 return r; 3126 goto retry; 3127 } 3128 3129 map = rcu_dereference_protected(md->map, lockdep_is_held(&md->suspend_lock)); 3130 if (!map || !dm_table_get_size(map)) 3131 goto out; 3132 3133 r = __dm_resume(md, map); 3134 if (r) 3135 goto out; 3136 3137 clear_bit(DMF_SUSPENDED, &md->flags); 3138 out: 3139 mutex_unlock(&md->suspend_lock); 3140 3141 return r; 3142 } 3143 3144 /* 3145 * Internal suspend/resume works like userspace-driven suspend. It waits 3146 * until all bios finish and prevents issuing new bios to the target drivers. 3147 * It may be used only from the kernel. 3148 */ 3149 3150 static void __dm_internal_suspend(struct mapped_device *md, unsigned int suspend_flags) 3151 { 3152 struct dm_table *map = NULL; 3153 3154 lockdep_assert_held(&md->suspend_lock); 3155 3156 if (md->internal_suspend_count++) 3157 return; /* nested internal suspend */ 3158 3159 if (dm_suspended_md(md)) { 3160 set_bit(DMF_SUSPENDED_INTERNALLY, &md->flags); 3161 return; /* nest suspend */ 3162 } 3163 3164 map = rcu_dereference_protected(md->map, lockdep_is_held(&md->suspend_lock)); 3165 3166 /* 3167 * Using TASK_UNINTERRUPTIBLE because only NOFLUSH internal suspend is 3168 * supported. Properly supporting a TASK_INTERRUPTIBLE internal suspend 3169 * would require changing .presuspend to return an error -- avoid this 3170 * until there is a need for more elaborate variants of internal suspend. 3171 */ 3172 (void) __dm_suspend(md, map, suspend_flags, TASK_UNINTERRUPTIBLE, 3173 DMF_SUSPENDED_INTERNALLY); 3174 3175 set_bit(DMF_POST_SUSPENDING, &md->flags); 3176 dm_table_postsuspend_targets(map); 3177 clear_bit(DMF_POST_SUSPENDING, &md->flags); 3178 } 3179 3180 static void __dm_internal_resume(struct mapped_device *md) 3181 { 3182 int r; 3183 struct dm_table *map; 3184 3185 BUG_ON(!md->internal_suspend_count); 3186 3187 if (--md->internal_suspend_count) 3188 return; /* resume from nested internal suspend */ 3189 3190 if (dm_suspended_md(md)) 3191 goto done; /* resume from nested suspend */ 3192 3193 map = rcu_dereference_protected(md->map, lockdep_is_held(&md->suspend_lock)); 3194 r = __dm_resume(md, map); 3195 if (r) { 3196 /* 3197 * If a preresume method of some target failed, we are in a 3198 * tricky situation. We can't return an error to the caller. We 3199 * can't fake success because then the "resume" and 3200 * "postsuspend" methods would not be paired correctly, and it 3201 * would break various targets, for example it would cause list 3202 * corruption in the "origin" target. 3203 * 3204 * So, we fake normal suspend here, to make sure that the 3205 * "resume" and "postsuspend" methods will be paired correctly. 3206 */ 3207 DMERR("Preresume method failed: %d", r); 3208 set_bit(DMF_SUSPENDED, &md->flags); 3209 } 3210 done: 3211 clear_bit(DMF_SUSPENDED_INTERNALLY, &md->flags); 3212 smp_mb__after_atomic(); 3213 wake_up_bit(&md->flags, DMF_SUSPENDED_INTERNALLY); 3214 } 3215 3216 void dm_internal_suspend_noflush(struct mapped_device *md) 3217 { 3218 mutex_lock(&md->suspend_lock); 3219 __dm_internal_suspend(md, DM_SUSPEND_NOFLUSH_FLAG); 3220 mutex_unlock(&md->suspend_lock); 3221 } 3222 EXPORT_SYMBOL_GPL(dm_internal_suspend_noflush); 3223 3224 void dm_internal_resume(struct mapped_device *md) 3225 { 3226 mutex_lock(&md->suspend_lock); 3227 __dm_internal_resume(md); 3228 mutex_unlock(&md->suspend_lock); 3229 } 3230 EXPORT_SYMBOL_GPL(dm_internal_resume); 3231 3232 /* 3233 * Fast variants of internal suspend/resume hold md->suspend_lock, 3234 * which prevents interaction with userspace-driven suspend. 3235 */ 3236 3237 void dm_internal_suspend_fast(struct mapped_device *md) 3238 { 3239 mutex_lock(&md->suspend_lock); 3240 if (dm_suspended_md(md) || dm_suspended_internally_md(md)) 3241 return; 3242 3243 set_bit(DMF_BLOCK_IO_FOR_SUSPEND, &md->flags); 3244 synchronize_srcu(&md->io_barrier); 3245 flush_workqueue(md->wq); 3246 dm_wait_for_completion(md, TASK_UNINTERRUPTIBLE); 3247 } 3248 EXPORT_SYMBOL_GPL(dm_internal_suspend_fast); 3249 3250 void dm_internal_resume_fast(struct mapped_device *md) 3251 { 3252 if (dm_suspended_md(md) || dm_suspended_internally_md(md)) 3253 goto done; 3254 3255 dm_queue_flush(md); 3256 3257 done: 3258 mutex_unlock(&md->suspend_lock); 3259 } 3260 EXPORT_SYMBOL_GPL(dm_internal_resume_fast); 3261 3262 /* 3263 *--------------------------------------------------------------- 3264 * Event notification. 3265 *--------------------------------------------------------------- 3266 */ 3267 int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action, 3268 unsigned int cookie, bool need_resize_uevent) 3269 { 3270 int r; 3271 unsigned int noio_flag; 3272 char udev_cookie[DM_COOKIE_LENGTH]; 3273 char *envp[3] = { NULL, NULL, NULL }; 3274 char **envpp = envp; 3275 if (cookie) { 3276 snprintf(udev_cookie, DM_COOKIE_LENGTH, "%s=%u", 3277 DM_COOKIE_ENV_VAR_NAME, cookie); 3278 *envpp++ = udev_cookie; 3279 } 3280 if (need_resize_uevent) { 3281 *envpp++ = "RESIZE=1"; 3282 } 3283 3284 noio_flag = memalloc_noio_save(); 3285 3286 r = kobject_uevent_env(&disk_to_dev(md->disk)->kobj, action, envp); 3287 3288 memalloc_noio_restore(noio_flag); 3289 3290 return r; 3291 } 3292 3293 uint32_t dm_next_uevent_seq(struct mapped_device *md) 3294 { 3295 return atomic_add_return(1, &md->uevent_seq); 3296 } 3297 3298 uint32_t dm_get_event_nr(struct mapped_device *md) 3299 { 3300 return atomic_read(&md->event_nr); 3301 } 3302 3303 int dm_wait_event(struct mapped_device *md, int event_nr) 3304 { 3305 return wait_event_interruptible(md->eventq, 3306 (event_nr != atomic_read(&md->event_nr))); 3307 } 3308 3309 void dm_uevent_add(struct mapped_device *md, struct list_head *elist) 3310 { 3311 unsigned long flags; 3312 3313 spin_lock_irqsave(&md->uevent_lock, flags); 3314 list_add(elist, &md->uevent_list); 3315 spin_unlock_irqrestore(&md->uevent_lock, flags); 3316 } 3317 3318 /* 3319 * The gendisk is only valid as long as you have a reference 3320 * count on 'md'. 3321 */ 3322 struct gendisk *dm_disk(struct mapped_device *md) 3323 { 3324 return md->disk; 3325 } 3326 EXPORT_SYMBOL_GPL(dm_disk); 3327 3328 struct kobject *dm_kobject(struct mapped_device *md) 3329 { 3330 return &md->kobj_holder.kobj; 3331 } 3332 3333 struct mapped_device *dm_get_from_kobject(struct kobject *kobj) 3334 { 3335 struct mapped_device *md; 3336 3337 md = container_of(kobj, struct mapped_device, kobj_holder.kobj); 3338 3339 spin_lock(&_minor_lock); 3340 if (test_bit(DMF_FREEING, &md->flags) || dm_deleting_md(md)) { 3341 md = NULL; 3342 goto out; 3343 } 3344 dm_get(md); 3345 out: 3346 spin_unlock(&_minor_lock); 3347 3348 return md; 3349 } 3350 3351 int dm_suspended_md(struct mapped_device *md) 3352 { 3353 return test_bit(DMF_SUSPENDED, &md->flags); 3354 } 3355 3356 static int dm_post_suspending_md(struct mapped_device *md) 3357 { 3358 return test_bit(DMF_POST_SUSPENDING, &md->flags); 3359 } 3360 3361 int dm_suspended_internally_md(struct mapped_device *md) 3362 { 3363 return test_bit(DMF_SUSPENDED_INTERNALLY, &md->flags); 3364 } 3365 3366 int dm_test_deferred_remove_flag(struct mapped_device *md) 3367 { 3368 return test_bit(DMF_DEFERRED_REMOVE, &md->flags); 3369 } 3370 3371 int dm_suspended(struct dm_target *ti) 3372 { 3373 return dm_suspended_md(ti->table->md); 3374 } 3375 EXPORT_SYMBOL_GPL(dm_suspended); 3376 3377 int dm_post_suspending(struct dm_target *ti) 3378 { 3379 return dm_post_suspending_md(ti->table->md); 3380 } 3381 EXPORT_SYMBOL_GPL(dm_post_suspending); 3382 3383 int dm_noflush_suspending(struct dm_target *ti) 3384 { 3385 return __noflush_suspending(ti->table->md); 3386 } 3387 EXPORT_SYMBOL_GPL(dm_noflush_suspending); 3388 3389 void dm_free_md_mempools(struct dm_md_mempools *pools) 3390 { 3391 if (!pools) 3392 return; 3393 3394 bioset_exit(&pools->bs); 3395 bioset_exit(&pools->io_bs); 3396 3397 kfree(pools); 3398 } 3399 3400 struct dm_blkdev_id { 3401 u8 *id; 3402 enum blk_unique_id type; 3403 }; 3404 3405 static int __dm_get_unique_id(struct dm_target *ti, struct dm_dev *dev, 3406 sector_t start, sector_t len, void *data) 3407 { 3408 struct dm_blkdev_id *dm_id = data; 3409 const struct block_device_operations *fops = dev->bdev->bd_disk->fops; 3410 3411 if (!fops->get_unique_id) 3412 return 0; 3413 3414 return fops->get_unique_id(dev->bdev->bd_disk, dm_id->id, dm_id->type); 3415 } 3416 3417 /* 3418 * Allow access to get_unique_id() for the first device returning a 3419 * non-zero result. Reasonable use expects all devices to have the 3420 * same unique id. 3421 */ 3422 static int dm_blk_get_unique_id(struct gendisk *disk, u8 *id, 3423 enum blk_unique_id type) 3424 { 3425 struct mapped_device *md = disk->private_data; 3426 struct dm_table *table; 3427 struct dm_target *ti; 3428 int ret = 0, srcu_idx; 3429 3430 struct dm_blkdev_id dm_id = { 3431 .id = id, 3432 .type = type, 3433 }; 3434 3435 table = dm_get_live_table(md, &srcu_idx); 3436 if (!table || !dm_table_get_size(table)) 3437 goto out; 3438 3439 /* We only support devices that have a single target */ 3440 if (table->num_targets != 1) 3441 goto out; 3442 ti = dm_table_get_target(table, 0); 3443 3444 if (!ti->type->iterate_devices) 3445 goto out; 3446 3447 ret = ti->type->iterate_devices(ti, __dm_get_unique_id, &dm_id); 3448 out: 3449 dm_put_live_table(md, srcu_idx); 3450 return ret; 3451 } 3452 3453 struct dm_pr { 3454 u64 old_key; 3455 u64 new_key; 3456 u32 flags; 3457 bool abort; 3458 bool fail_early; 3459 int ret; 3460 enum pr_type type; 3461 struct pr_keys *read_keys; 3462 struct pr_held_reservation *rsv; 3463 }; 3464 3465 static int dm_call_pr(struct block_device *bdev, iterate_devices_callout_fn fn, 3466 struct dm_pr *pr) 3467 { 3468 struct mapped_device *md = bdev->bd_disk->private_data; 3469 struct dm_table *table; 3470 struct dm_target *ti; 3471 int ret = -ENOTTY, srcu_idx; 3472 3473 table = dm_get_live_table(md, &srcu_idx); 3474 if (!table || !dm_table_get_size(table)) 3475 goto out; 3476 3477 /* We only support devices that have a single target */ 3478 if (table->num_targets != 1) 3479 goto out; 3480 ti = dm_table_get_target(table, 0); 3481 3482 if (dm_suspended_md(md)) { 3483 ret = -EAGAIN; 3484 goto out; 3485 } 3486 3487 ret = -EINVAL; 3488 if (!ti->type->iterate_devices) 3489 goto out; 3490 3491 ti->type->iterate_devices(ti, fn, pr); 3492 ret = 0; 3493 out: 3494 dm_put_live_table(md, srcu_idx); 3495 return ret; 3496 } 3497 3498 /* 3499 * For register / unregister we need to manually call out to every path. 3500 */ 3501 static int __dm_pr_register(struct dm_target *ti, struct dm_dev *dev, 3502 sector_t start, sector_t len, void *data) 3503 { 3504 struct dm_pr *pr = data; 3505 const struct pr_ops *ops = dev->bdev->bd_disk->fops->pr_ops; 3506 int ret; 3507 3508 if (!ops || !ops->pr_register) { 3509 pr->ret = -EOPNOTSUPP; 3510 return -1; 3511 } 3512 3513 ret = ops->pr_register(dev->bdev, pr->old_key, pr->new_key, pr->flags); 3514 if (!ret) 3515 return 0; 3516 3517 if (!pr->ret) 3518 pr->ret = ret; 3519 3520 if (pr->fail_early) 3521 return -1; 3522 3523 return 0; 3524 } 3525 3526 static int dm_pr_register(struct block_device *bdev, u64 old_key, u64 new_key, 3527 u32 flags) 3528 { 3529 struct dm_pr pr = { 3530 .old_key = old_key, 3531 .new_key = new_key, 3532 .flags = flags, 3533 .fail_early = true, 3534 .ret = 0, 3535 }; 3536 int ret; 3537 3538 ret = dm_call_pr(bdev, __dm_pr_register, &pr); 3539 if (ret) { 3540 /* Didn't even get to register a path */ 3541 return ret; 3542 } 3543 3544 if (!pr.ret) 3545 return 0; 3546 ret = pr.ret; 3547 3548 if (!new_key) 3549 return ret; 3550 3551 /* unregister all paths if we failed to register any path */ 3552 pr.old_key = new_key; 3553 pr.new_key = 0; 3554 pr.flags = 0; 3555 pr.fail_early = false; 3556 (void) dm_call_pr(bdev, __dm_pr_register, &pr); 3557 return ret; 3558 } 3559 3560 3561 static int __dm_pr_reserve(struct dm_target *ti, struct dm_dev *dev, 3562 sector_t start, sector_t len, void *data) 3563 { 3564 struct dm_pr *pr = data; 3565 const struct pr_ops *ops = dev->bdev->bd_disk->fops->pr_ops; 3566 3567 if (!ops || !ops->pr_reserve) { 3568 pr->ret = -EOPNOTSUPP; 3569 return -1; 3570 } 3571 3572 pr->ret = ops->pr_reserve(dev->bdev, pr->old_key, pr->type, pr->flags); 3573 if (!pr->ret) 3574 return -1; 3575 3576 return 0; 3577 } 3578 3579 static int dm_pr_reserve(struct block_device *bdev, u64 key, enum pr_type type, 3580 u32 flags) 3581 { 3582 struct dm_pr pr = { 3583 .old_key = key, 3584 .flags = flags, 3585 .type = type, 3586 .fail_early = false, 3587 .ret = 0, 3588 }; 3589 int ret; 3590 3591 ret = dm_call_pr(bdev, __dm_pr_reserve, &pr); 3592 if (ret) 3593 return ret; 3594 3595 return pr.ret; 3596 } 3597 3598 /* 3599 * If there is a non-All Registrants type of reservation, the release must be 3600 * sent down the holding path. For the cases where there is no reservation or 3601 * the path is not the holder the device will also return success, so we must 3602 * try each path to make sure we got the correct path. 3603 */ 3604 static int __dm_pr_release(struct dm_target *ti, struct dm_dev *dev, 3605 sector_t start, sector_t len, void *data) 3606 { 3607 struct dm_pr *pr = data; 3608 const struct pr_ops *ops = dev->bdev->bd_disk->fops->pr_ops; 3609 3610 if (!ops || !ops->pr_release) { 3611 pr->ret = -EOPNOTSUPP; 3612 return -1; 3613 } 3614 3615 pr->ret = ops->pr_release(dev->bdev, pr->old_key, pr->type); 3616 if (pr->ret) 3617 return -1; 3618 3619 return 0; 3620 } 3621 3622 static int dm_pr_release(struct block_device *bdev, u64 key, enum pr_type type) 3623 { 3624 struct dm_pr pr = { 3625 .old_key = key, 3626 .type = type, 3627 .fail_early = false, 3628 }; 3629 int ret; 3630 3631 ret = dm_call_pr(bdev, __dm_pr_release, &pr); 3632 if (ret) 3633 return ret; 3634 3635 return pr.ret; 3636 } 3637 3638 static int __dm_pr_preempt(struct dm_target *ti, struct dm_dev *dev, 3639 sector_t start, sector_t len, void *data) 3640 { 3641 struct dm_pr *pr = data; 3642 const struct pr_ops *ops = dev->bdev->bd_disk->fops->pr_ops; 3643 3644 if (!ops || !ops->pr_preempt) { 3645 pr->ret = -EOPNOTSUPP; 3646 return -1; 3647 } 3648 3649 pr->ret = ops->pr_preempt(dev->bdev, pr->old_key, pr->new_key, pr->type, 3650 pr->abort); 3651 if (!pr->ret) 3652 return -1; 3653 3654 return 0; 3655 } 3656 3657 static int dm_pr_preempt(struct block_device *bdev, u64 old_key, u64 new_key, 3658 enum pr_type type, bool abort) 3659 { 3660 struct dm_pr pr = { 3661 .new_key = new_key, 3662 .old_key = old_key, 3663 .type = type, 3664 .fail_early = false, 3665 }; 3666 int ret; 3667 3668 ret = dm_call_pr(bdev, __dm_pr_preempt, &pr); 3669 if (ret) 3670 return ret; 3671 3672 return pr.ret; 3673 } 3674 3675 static int dm_pr_clear(struct block_device *bdev, u64 key) 3676 { 3677 struct mapped_device *md = bdev->bd_disk->private_data; 3678 const struct pr_ops *ops; 3679 int r, srcu_idx; 3680 bool forward = true; 3681 3682 /* Not a real ioctl, but targets must not interpret non-DM ioctls */ 3683 r = dm_prepare_ioctl(md, &srcu_idx, &bdev, 0, 0, &forward); 3684 if (r < 0) 3685 goto out; 3686 WARN_ON_ONCE(!forward); 3687 3688 ops = bdev->bd_disk->fops->pr_ops; 3689 if (ops && ops->pr_clear) 3690 r = ops->pr_clear(bdev, key); 3691 else 3692 r = -EOPNOTSUPP; 3693 out: 3694 dm_unprepare_ioctl(md, srcu_idx); 3695 return r; 3696 } 3697 3698 static int __dm_pr_read_keys(struct dm_target *ti, struct dm_dev *dev, 3699 sector_t start, sector_t len, void *data) 3700 { 3701 struct dm_pr *pr = data; 3702 const struct pr_ops *ops = dev->bdev->bd_disk->fops->pr_ops; 3703 3704 if (!ops || !ops->pr_read_keys) { 3705 pr->ret = -EOPNOTSUPP; 3706 return -1; 3707 } 3708 3709 pr->ret = ops->pr_read_keys(dev->bdev, pr->read_keys); 3710 if (!pr->ret) 3711 return -1; 3712 3713 return 0; 3714 } 3715 3716 static int dm_pr_read_keys(struct block_device *bdev, struct pr_keys *keys) 3717 { 3718 struct dm_pr pr = { 3719 .read_keys = keys, 3720 }; 3721 int ret; 3722 3723 ret = dm_call_pr(bdev, __dm_pr_read_keys, &pr); 3724 if (ret) 3725 return ret; 3726 3727 return pr.ret; 3728 } 3729 3730 static int __dm_pr_read_reservation(struct dm_target *ti, struct dm_dev *dev, 3731 sector_t start, sector_t len, void *data) 3732 { 3733 struct dm_pr *pr = data; 3734 const struct pr_ops *ops = dev->bdev->bd_disk->fops->pr_ops; 3735 3736 if (!ops || !ops->pr_read_reservation) { 3737 pr->ret = -EOPNOTSUPP; 3738 return -1; 3739 } 3740 3741 pr->ret = ops->pr_read_reservation(dev->bdev, pr->rsv); 3742 if (!pr->ret) 3743 return -1; 3744 3745 return 0; 3746 } 3747 3748 static int dm_pr_read_reservation(struct block_device *bdev, 3749 struct pr_held_reservation *rsv) 3750 { 3751 struct dm_pr pr = { 3752 .rsv = rsv, 3753 }; 3754 int ret; 3755 3756 ret = dm_call_pr(bdev, __dm_pr_read_reservation, &pr); 3757 if (ret) 3758 return ret; 3759 3760 return pr.ret; 3761 } 3762 3763 static const struct pr_ops dm_pr_ops = { 3764 .pr_register = dm_pr_register, 3765 .pr_reserve = dm_pr_reserve, 3766 .pr_release = dm_pr_release, 3767 .pr_preempt = dm_pr_preempt, 3768 .pr_clear = dm_pr_clear, 3769 .pr_read_keys = dm_pr_read_keys, 3770 .pr_read_reservation = dm_pr_read_reservation, 3771 }; 3772 3773 static const struct block_device_operations dm_blk_dops = { 3774 .submit_bio = dm_submit_bio, 3775 .poll_bio = dm_poll_bio, 3776 .open = dm_blk_open, 3777 .release = dm_blk_close, 3778 .ioctl = dm_blk_ioctl, 3779 .getgeo = dm_blk_getgeo, 3780 .report_zones = dm_blk_report_zones, 3781 .get_unique_id = dm_blk_get_unique_id, 3782 .pr_ops = &dm_pr_ops, 3783 .owner = THIS_MODULE 3784 }; 3785 3786 static const struct block_device_operations dm_rq_blk_dops = { 3787 .open = dm_blk_open, 3788 .release = dm_blk_close, 3789 .ioctl = dm_blk_ioctl, 3790 .getgeo = dm_blk_getgeo, 3791 .get_unique_id = dm_blk_get_unique_id, 3792 .pr_ops = &dm_pr_ops, 3793 .owner = THIS_MODULE 3794 }; 3795 3796 static const struct dax_operations dm_dax_ops = { 3797 .direct_access = dm_dax_direct_access, 3798 .zero_page_range = dm_dax_zero_page_range, 3799 .recovery_write = dm_dax_recovery_write, 3800 }; 3801 3802 /* 3803 * module hooks 3804 */ 3805 module_init(dm_init); 3806 module_exit(dm_exit); 3807 3808 module_param(major, uint, 0); 3809 MODULE_PARM_DESC(major, "The major number of the device mapper"); 3810 3811 module_param(reserved_bio_based_ios, uint, 0644); 3812 MODULE_PARM_DESC(reserved_bio_based_ios, "Reserved IOs in bio-based mempools"); 3813 3814 module_param(dm_numa_node, int, 0644); 3815 MODULE_PARM_DESC(dm_numa_node, "NUMA node for DM device memory allocations"); 3816 3817 module_param(swap_bios, int, 0644); 3818 MODULE_PARM_DESC(swap_bios, "Maximum allowed inflight swap IOs"); 3819 3820 MODULE_DESCRIPTION(DM_NAME " driver"); 3821 MODULE_AUTHOR("Joe Thornber <dm-devel@lists.linux.dev>"); 3822 MODULE_LICENSE("GPL"); 3823