Lines Matching refs:lo
109 static int loop_global_lock_killable(struct loop_device *lo, bool global) in loop_global_lock_killable() argument
118 err = mutex_lock_killable(&lo->lo_mutex); in loop_global_lock_killable()
130 static void loop_global_unlock(struct loop_device *lo, bool global) in loop_global_unlock() argument
132 mutex_unlock(&lo->lo_mutex); in loop_global_unlock()
140 static loff_t lo_calculate_size(struct loop_device *lo, struct file *file) in lo_calculate_size() argument
162 if (lo->lo_offset > 0) in lo_calculate_size()
163 loopsize -= lo->lo_offset; in lo_calculate_size()
167 if (lo->lo_sizelimit > 0 && lo->lo_sizelimit < loopsize) in lo_calculate_size()
168 loopsize = lo->lo_sizelimit; in lo_calculate_size()
181 static bool lo_can_use_dio(struct loop_device *lo) in lo_can_use_dio() argument
183 if (!(lo->lo_backing_file->f_mode & FMODE_CAN_ODIRECT)) in lo_can_use_dio()
185 if (queue_logical_block_size(lo->lo_queue) < lo->lo_min_dio_size) in lo_can_use_dio()
187 if (lo->lo_offset & (lo->lo_min_dio_size - 1)) in lo_can_use_dio()
200 static inline void loop_update_dio(struct loop_device *lo) in loop_update_dio() argument
202 lockdep_assert_held(&lo->lo_mutex); in loop_update_dio()
203 WARN_ON_ONCE(lo->lo_state == Lo_bound && in loop_update_dio()
204 lo->lo_queue->mq_freeze_depth == 0); in loop_update_dio()
206 if ((lo->lo_flags & LO_FLAGS_DIRECT_IO) && !lo_can_use_dio(lo)) in loop_update_dio()
207 lo->lo_flags &= ~LO_FLAGS_DIRECT_IO; in loop_update_dio()
218 static void loop_set_size(struct loop_device *lo, loff_t size) in loop_set_size() argument
220 if (!set_capacity_and_notify(lo->lo_disk, size)) in loop_set_size()
221 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_set_size()
224 static void loop_clear_limits(struct loop_device *lo, int mode) in loop_clear_limits() argument
226 struct queue_limits lim = queue_limits_start_update(lo->lo_queue); in loop_clear_limits()
243 queue_limits_commit_update(lo->lo_queue, &lim); in loop_clear_limits()
246 static int lo_fallocate(struct loop_device *lo, struct request *rq, loff_t pos, in lo_fallocate() argument
253 struct file *file = lo->lo_backing_file; in lo_fallocate()
258 if (!bdev_max_discard_sectors(lo->lo_device)) in lo_fallocate()
270 loop_clear_limits(lo, mode); in lo_fallocate()
275 static int lo_req_flush(struct loop_device *lo, struct request *rq) in lo_req_flush() argument
277 int ret = vfs_fsync(lo->lo_backing_file, 0); in lo_req_flush()
340 static int lo_rw_aio(struct loop_device *lo, struct loop_cmd *cmd, in lo_rw_aio() argument
348 struct file *file = lo->lo_backing_file; in lo_rw_aio()
413 static int do_req_filebacked(struct loop_device *lo, struct request *rq) in do_req_filebacked() argument
416 loff_t pos = ((loff_t) blk_rq_pos(rq) << 9) + lo->lo_offset; in do_req_filebacked()
420 return lo_req_flush(lo, rq); in do_req_filebacked()
426 return lo_fallocate(lo, rq, pos, in do_req_filebacked()
431 return lo_fallocate(lo, rq, pos, FALLOC_FL_PUNCH_HOLE); in do_req_filebacked()
433 return lo_rw_aio(lo, cmd, pos, ITER_SOURCE); in do_req_filebacked()
435 return lo_rw_aio(lo, cmd, pos, ITER_DEST); in do_req_filebacked()
442 static void loop_reread_partitions(struct loop_device *lo) in loop_reread_partitions() argument
446 mutex_lock(&lo->lo_disk->open_mutex); in loop_reread_partitions()
447 rc = bdev_disk_changed(lo->lo_disk, false); in loop_reread_partitions()
448 mutex_unlock(&lo->lo_disk->open_mutex); in loop_reread_partitions()
451 __func__, lo->lo_number, lo->lo_file_name, rc); in loop_reread_partitions()
454 static unsigned int loop_query_min_dio_size(struct loop_device *lo) in loop_query_min_dio_size() argument
456 struct file *file = lo->lo_backing_file; in loop_query_min_dio_size()
508 static void loop_assign_backing_file(struct loop_device *lo, struct file *file) in loop_assign_backing_file() argument
510 lo->lo_backing_file = file; in loop_assign_backing_file()
511 lo->old_gfp_mask = mapping_gfp_mask(file->f_mapping); in loop_assign_backing_file()
513 lo->old_gfp_mask & ~(__GFP_IO | __GFP_FS)); in loop_assign_backing_file()
514 if (lo->lo_backing_file->f_flags & O_DIRECT) in loop_assign_backing_file()
515 lo->lo_flags |= LO_FLAGS_DIRECT_IO; in loop_assign_backing_file()
516 lo->lo_min_dio_size = loop_query_min_dio_size(lo); in loop_assign_backing_file()
538 static int loop_change_fd(struct loop_device *lo, struct block_device *bdev, in loop_change_fd() argument
558 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 1); in loop_change_fd()
561 error = loop_global_lock_killable(lo, is_loop); in loop_change_fd()
565 if (lo->lo_state != Lo_bound) in loop_change_fd()
570 if (!(lo->lo_flags & LO_FLAGS_READ_ONLY)) in loop_change_fd()
577 old_file = lo->lo_backing_file; in loop_change_fd()
582 if (lo_calculate_size(lo, file) != lo_calculate_size(lo, old_file)) in loop_change_fd()
593 disk_force_media_change(lo->lo_disk); in loop_change_fd()
594 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_change_fd()
595 mapping_set_gfp_mask(old_file->f_mapping, lo->old_gfp_mask); in loop_change_fd()
596 loop_assign_backing_file(lo, file); in loop_change_fd()
597 loop_update_dio(lo); in loop_change_fd()
598 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_change_fd()
599 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in loop_change_fd()
600 loop_global_unlock(lo, is_loop); in loop_change_fd()
616 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0); in loop_change_fd()
618 loop_reread_partitions(lo); in loop_change_fd()
622 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_change_fd()
626 loop_global_unlock(lo, is_loop); in loop_change_fd()
629 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0); in loop_change_fd()
639 struct loop_device *lo = disk->private_data; in loop_attr_show() local
641 return callback(lo, page); in loop_attr_show()
654 static ssize_t loop_attr_backing_file_show(struct loop_device *lo, char *buf) in loop_attr_backing_file_show() argument
659 spin_lock_irq(&lo->lo_lock); in loop_attr_backing_file_show()
660 if (lo->lo_backing_file) in loop_attr_backing_file_show()
661 p = file_path(lo->lo_backing_file, buf, PAGE_SIZE - 1); in loop_attr_backing_file_show()
662 spin_unlock_irq(&lo->lo_lock); in loop_attr_backing_file_show()
676 static ssize_t loop_attr_offset_show(struct loop_device *lo, char *buf) in loop_attr_offset_show() argument
678 return sysfs_emit(buf, "%llu\n", (unsigned long long)lo->lo_offset); in loop_attr_offset_show()
681 static ssize_t loop_attr_sizelimit_show(struct loop_device *lo, char *buf) in loop_attr_sizelimit_show() argument
683 return sysfs_emit(buf, "%llu\n", (unsigned long long)lo->lo_sizelimit); in loop_attr_sizelimit_show()
686 static ssize_t loop_attr_autoclear_show(struct loop_device *lo, char *buf) in loop_attr_autoclear_show() argument
688 int autoclear = (lo->lo_flags & LO_FLAGS_AUTOCLEAR); in loop_attr_autoclear_show()
693 static ssize_t loop_attr_partscan_show(struct loop_device *lo, char *buf) in loop_attr_partscan_show() argument
695 int partscan = (lo->lo_flags & LO_FLAGS_PARTSCAN); in loop_attr_partscan_show()
700 static ssize_t loop_attr_dio_show(struct loop_device *lo, char *buf) in loop_attr_dio_show() argument
702 int dio = (lo->lo_flags & LO_FLAGS_DIRECT_IO); in loop_attr_dio_show()
729 static void loop_sysfs_init(struct loop_device *lo) in loop_sysfs_init() argument
731 lo->sysfs_inited = !sysfs_create_group(&disk_to_dev(lo->lo_disk)->kobj, in loop_sysfs_init()
735 static void loop_sysfs_exit(struct loop_device *lo) in loop_sysfs_exit() argument
737 if (lo->sysfs_inited) in loop_sysfs_exit()
738 sysfs_remove_group(&disk_to_dev(lo->lo_disk)->kobj, in loop_sysfs_exit()
742 static void loop_get_discard_config(struct loop_device *lo, in loop_get_discard_config() argument
745 struct file *file = lo->lo_backing_file; in loop_get_discard_config()
777 struct loop_device *lo; member
796 static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd) in loop_queue_work() argument
803 spin_lock_irq(&lo->lo_work_lock); in loop_queue_work()
808 node = &lo->worker_tree.rb_node; in loop_queue_work()
843 worker->lo = lo; in loop_queue_work()
845 rb_insert_color(&worker->rb_node, &lo->worker_tree); in loop_queue_work()
858 work = &lo->rootcg_work; in loop_queue_work()
859 cmd_list = &lo->rootcg_cmd_list; in loop_queue_work()
862 queue_work(lo->workqueue, work); in loop_queue_work()
863 spin_unlock_irq(&lo->lo_work_lock); in loop_queue_work()
866 static void loop_set_timer(struct loop_device *lo) in loop_set_timer() argument
868 timer_reduce(&lo->timer, jiffies + LOOP_IDLE_WORKER_TIMEOUT); in loop_set_timer()
871 static void loop_free_idle_workers(struct loop_device *lo, bool delete_all) in loop_free_idle_workers() argument
875 spin_lock_irq(&lo->lo_work_lock); in loop_free_idle_workers()
876 list_for_each_entry_safe(worker, pos, &lo->idle_worker_list, in loop_free_idle_workers()
883 rb_erase(&worker->rb_node, &lo->worker_tree); in loop_free_idle_workers()
887 if (!list_empty(&lo->idle_worker_list)) in loop_free_idle_workers()
888 loop_set_timer(lo); in loop_free_idle_workers()
889 spin_unlock_irq(&lo->lo_work_lock); in loop_free_idle_workers()
894 struct loop_device *lo = container_of(timer, struct loop_device, timer); in loop_free_idle_workers_timer() local
896 return loop_free_idle_workers(lo, false); in loop_free_idle_workers_timer()
908 loop_set_status_from_info(struct loop_device *lo, in loop_set_status_from_info() argument
931 lo->lo_offset = info->lo_offset; in loop_set_status_from_info()
932 lo->lo_sizelimit = info->lo_sizelimit; in loop_set_status_from_info()
934 memcpy(lo->lo_file_name, info->lo_file_name, LO_NAME_SIZE); in loop_set_status_from_info()
935 lo->lo_file_name[LO_NAME_SIZE-1] = 0; in loop_set_status_from_info()
939 static unsigned int loop_default_blocksize(struct loop_device *lo) in loop_default_blocksize() argument
942 if (lo->lo_flags & LO_FLAGS_DIRECT_IO) in loop_default_blocksize()
943 return lo->lo_min_dio_size; in loop_default_blocksize()
947 static void loop_update_limits(struct loop_device *lo, struct queue_limits *lim, in loop_update_limits() argument
950 struct file *file = lo->lo_backing_file; in loop_update_limits()
961 bsize = loop_default_blocksize(lo); in loop_update_limits()
963 loop_get_discard_config(lo, &granularity, &max_discard_sectors); in loop_update_limits()
969 if (file->f_op->fsync && !(lo->lo_flags & LO_FLAGS_READ_ONLY)) in loop_update_limits()
981 static int loop_configure(struct loop_device *lo, blk_mode_t mode, in loop_configure() argument
1016 error = loop_global_lock_killable(lo, is_loop); in loop_configure()
1021 if (lo->lo_state != Lo_unbound) in loop_configure()
1033 error = loop_set_status_from_info(lo, &config->info); in loop_configure()
1036 lo->lo_flags = config->info.lo_flags; in loop_configure()
1040 lo->lo_flags |= LO_FLAGS_READ_ONLY; in loop_configure()
1042 if (!lo->workqueue) { in loop_configure()
1043 lo->workqueue = alloc_workqueue("loop%d", in loop_configure()
1045 0, lo->lo_number); in loop_configure()
1046 if (!lo->workqueue) { in loop_configure()
1053 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 1); in loop_configure()
1055 disk_force_media_change(lo->lo_disk); in loop_configure()
1056 set_disk_ro(lo->lo_disk, (lo->lo_flags & LO_FLAGS_READ_ONLY) != 0); in loop_configure()
1058 lo->lo_device = bdev; in loop_configure()
1059 loop_assign_backing_file(lo, file); in loop_configure()
1061 lim = queue_limits_start_update(lo->lo_queue); in loop_configure()
1062 loop_update_limits(lo, &lim, config->block_size); in loop_configure()
1064 error = queue_limits_commit_update(lo->lo_queue, &lim); in loop_configure()
1075 loop_update_dio(lo); in loop_configure()
1076 loop_sysfs_init(lo); in loop_configure()
1078 size = lo_calculate_size(lo, file); in loop_configure()
1079 loop_set_size(lo, size); in loop_configure()
1084 WRITE_ONCE(lo->lo_state, Lo_bound); in loop_configure()
1086 lo->lo_flags |= LO_FLAGS_PARTSCAN; in loop_configure()
1087 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in loop_configure()
1089 clear_bit(GD_SUPPRESS_PART_SCAN, &lo->lo_disk->state); in loop_configure()
1091 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0); in loop_configure()
1092 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_configure()
1094 loop_global_unlock(lo, is_loop); in loop_configure()
1096 loop_reread_partitions(lo); in loop_configure()
1104 loop_global_unlock(lo, is_loop); in loop_configure()
1115 static void __loop_clr_fd(struct loop_device *lo) in __loop_clr_fd() argument
1119 gfp_t gfp = lo->old_gfp_mask; in __loop_clr_fd()
1121 spin_lock_irq(&lo->lo_lock); in __loop_clr_fd()
1122 filp = lo->lo_backing_file; in __loop_clr_fd()
1123 lo->lo_backing_file = NULL; in __loop_clr_fd()
1124 spin_unlock_irq(&lo->lo_lock); in __loop_clr_fd()
1126 lo->lo_device = NULL; in __loop_clr_fd()
1127 lo->lo_offset = 0; in __loop_clr_fd()
1128 lo->lo_sizelimit = 0; in __loop_clr_fd()
1129 memset(lo->lo_file_name, 0, LO_NAME_SIZE); in __loop_clr_fd()
1137 lim = queue_limits_start_update(lo->lo_queue); in __loop_clr_fd()
1141 queue_limits_commit_update(lo->lo_queue, &lim); in __loop_clr_fd()
1143 invalidate_disk(lo->lo_disk); in __loop_clr_fd()
1144 loop_sysfs_exit(lo); in __loop_clr_fd()
1146 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in __loop_clr_fd()
1151 disk_force_media_change(lo->lo_disk); in __loop_clr_fd()
1153 if (lo->lo_flags & LO_FLAGS_PARTSCAN) { in __loop_clr_fd()
1164 err = bdev_disk_changed(lo->lo_disk, false); in __loop_clr_fd()
1167 __func__, lo->lo_number, err); in __loop_clr_fd()
1177 lo->lo_flags = 0; in __loop_clr_fd()
1179 set_bit(GD_SUPPRESS_PART_SCAN, &lo->lo_disk->state); in __loop_clr_fd()
1180 mutex_lock(&lo->lo_mutex); in __loop_clr_fd()
1181 WRITE_ONCE(lo->lo_state, Lo_unbound); in __loop_clr_fd()
1182 mutex_unlock(&lo->lo_mutex); in __loop_clr_fd()
1192 static int loop_clr_fd(struct loop_device *lo) in loop_clr_fd() argument
1205 err = loop_global_lock_killable(lo, true); in loop_clr_fd()
1208 if (lo->lo_state != Lo_bound) { in loop_clr_fd()
1209 loop_global_unlock(lo, true); in loop_clr_fd()
1218 lo->lo_flags |= LO_FLAGS_AUTOCLEAR; in loop_clr_fd()
1219 if (disk_openers(lo->lo_disk) == 1) in loop_clr_fd()
1220 WRITE_ONCE(lo->lo_state, Lo_rundown); in loop_clr_fd()
1221 loop_global_unlock(lo, true); in loop_clr_fd()
1227 loop_set_status(struct loop_device *lo, const struct loop_info64 *info) in loop_set_status() argument
1234 err = mutex_lock_killable(&lo->lo_mutex); in loop_set_status()
1237 if (lo->lo_state != Lo_bound) { in loop_set_status()
1242 if (lo->lo_offset != info->lo_offset || in loop_set_status()
1243 lo->lo_sizelimit != info->lo_sizelimit) { in loop_set_status()
1245 sync_blockdev(lo->lo_device); in loop_set_status()
1246 invalidate_bdev(lo->lo_device); in loop_set_status()
1250 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_set_status()
1252 err = loop_set_status_from_info(lo, info); in loop_set_status()
1256 partscan = !(lo->lo_flags & LO_FLAGS_PARTSCAN) && in loop_set_status()
1259 lo->lo_flags &= ~LOOP_SET_STATUS_CLEARABLE_FLAGS; in loop_set_status()
1260 lo->lo_flags |= (info->lo_flags & LOOP_SET_STATUS_SETTABLE_FLAGS); in loop_set_status()
1263 loop_update_dio(lo); in loop_set_status()
1266 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_set_status()
1268 clear_bit(GD_SUPPRESS_PART_SCAN, &lo->lo_disk->state); in loop_set_status()
1270 loff_t new_size = lo_calculate_size(lo, lo->lo_backing_file); in loop_set_status()
1271 loop_set_size(lo, new_size); in loop_set_status()
1274 mutex_unlock(&lo->lo_mutex); in loop_set_status()
1276 loop_reread_partitions(lo); in loop_set_status()
1282 loop_get_status(struct loop_device *lo, struct loop_info64 *info) in loop_get_status() argument
1288 ret = mutex_lock_killable(&lo->lo_mutex); in loop_get_status()
1291 if (lo->lo_state != Lo_bound) { in loop_get_status()
1292 mutex_unlock(&lo->lo_mutex); in loop_get_status()
1297 info->lo_number = lo->lo_number; in loop_get_status()
1298 info->lo_offset = lo->lo_offset; in loop_get_status()
1299 info->lo_sizelimit = lo->lo_sizelimit; in loop_get_status()
1300 info->lo_flags = lo->lo_flags; in loop_get_status()
1301 memcpy(info->lo_file_name, lo->lo_file_name, LO_NAME_SIZE); in loop_get_status()
1304 path = lo->lo_backing_file->f_path; in loop_get_status()
1306 mutex_unlock(&lo->lo_mutex); in loop_get_status()
1354 loop_set_status_old(struct loop_device *lo, const struct loop_info __user *arg) in loop_set_status_old() argument
1362 return loop_set_status(lo, &info64); in loop_set_status_old()
1366 loop_set_status64(struct loop_device *lo, const struct loop_info64 __user *arg) in loop_set_status64() argument
1372 return loop_set_status(lo, &info64); in loop_set_status64()
1376 loop_get_status_old(struct loop_device *lo, struct loop_info __user *arg) { in loop_get_status_old() argument
1383 err = loop_get_status(lo, &info64); in loop_get_status_old()
1393 loop_get_status64(struct loop_device *lo, struct loop_info64 __user *arg) { in loop_get_status64() argument
1399 err = loop_get_status(lo, &info64); in loop_get_status64()
1406 static int loop_set_capacity(struct loop_device *lo) in loop_set_capacity() argument
1410 if (unlikely(lo->lo_state != Lo_bound)) in loop_set_capacity()
1413 size = lo_calculate_size(lo, lo->lo_backing_file); in loop_set_capacity()
1414 loop_set_size(lo, size); in loop_set_capacity()
1419 static int loop_set_dio(struct loop_device *lo, unsigned long arg) in loop_set_dio() argument
1424 if (lo->lo_state != Lo_bound) in loop_set_dio()
1426 if (use_dio == !!(lo->lo_flags & LO_FLAGS_DIRECT_IO)) in loop_set_dio()
1430 if (!lo_can_use_dio(lo)) in loop_set_dio()
1433 vfs_fsync(lo->lo_backing_file, 0); in loop_set_dio()
1436 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_set_dio()
1438 lo->lo_flags |= LO_FLAGS_DIRECT_IO; in loop_set_dio()
1440 lo->lo_flags &= ~LO_FLAGS_DIRECT_IO; in loop_set_dio()
1441 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_set_dio()
1445 static int loop_set_block_size(struct loop_device *lo, blk_mode_t mode, in loop_set_block_size() argument
1462 err = mutex_lock_killable(&lo->lo_mutex); in loop_set_block_size()
1466 if (lo->lo_state != Lo_bound) { in loop_set_block_size()
1471 if (lo->lo_queue->limits.logical_block_size == arg) in loop_set_block_size()
1474 sync_blockdev(lo->lo_device); in loop_set_block_size()
1475 invalidate_bdev(lo->lo_device); in loop_set_block_size()
1477 lim = queue_limits_start_update(lo->lo_queue); in loop_set_block_size()
1478 loop_update_limits(lo, &lim, arg); in loop_set_block_size()
1480 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_set_block_size()
1481 err = queue_limits_commit_update(lo->lo_queue, &lim); in loop_set_block_size()
1482 loop_update_dio(lo); in loop_set_block_size()
1483 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_set_block_size()
1486 mutex_unlock(&lo->lo_mutex); in loop_set_block_size()
1493 static int lo_simple_ioctl(struct loop_device *lo, unsigned int cmd, in lo_simple_ioctl() argument
1498 err = mutex_lock_killable(&lo->lo_mutex); in lo_simple_ioctl()
1503 err = loop_set_capacity(lo); in lo_simple_ioctl()
1506 err = loop_set_dio(lo, arg); in lo_simple_ioctl()
1511 mutex_unlock(&lo->lo_mutex); in lo_simple_ioctl()
1518 struct loop_device *lo = bdev->bd_disk->private_data; in lo_ioctl() local
1534 return loop_configure(lo, mode, bdev, &config); in lo_ioctl()
1542 return loop_configure(lo, mode, bdev, &config); in lo_ioctl()
1545 return loop_change_fd(lo, bdev, arg); in lo_ioctl()
1547 return loop_clr_fd(lo); in lo_ioctl()
1551 err = loop_set_status_old(lo, argp); in lo_ioctl()
1554 return loop_get_status_old(lo, argp); in lo_ioctl()
1558 err = loop_set_status64(lo, argp); in lo_ioctl()
1561 return loop_get_status64(lo, argp); in lo_ioctl()
1565 return loop_set_block_size(lo, mode, bdev, arg); in lo_ioctl()
1572 err = lo_simple_ioctl(lo, cmd, arg); in lo_ioctl()
1652 loop_set_status_compat(struct loop_device *lo, in loop_set_status_compat() argument
1661 return loop_set_status(lo, &info64); in loop_set_status_compat()
1665 loop_get_status_compat(struct loop_device *lo, in loop_get_status_compat() argument
1673 err = loop_get_status(lo, &info64); in loop_get_status_compat()
1682 struct loop_device *lo = bdev->bd_disk->private_data; in lo_compat_ioctl() local
1687 err = loop_set_status_compat(lo, in lo_compat_ioctl()
1691 err = loop_get_status_compat(lo, in lo_compat_ioctl()
1717 struct loop_device *lo = disk->private_data; in lo_open() local
1720 err = mutex_lock_killable(&lo->lo_mutex); in lo_open()
1724 if (lo->lo_state == Lo_deleting || lo->lo_state == Lo_rundown) in lo_open()
1726 mutex_unlock(&lo->lo_mutex); in lo_open()
1732 struct loop_device *lo = disk->private_data; in lo_release() local
1743 mutex_lock(&lo->lo_mutex); in lo_release()
1744 if (lo->lo_state == Lo_bound && (lo->lo_flags & LO_FLAGS_AUTOCLEAR)) in lo_release()
1745 WRITE_ONCE(lo->lo_state, Lo_rundown); in lo_release()
1747 need_clear = (lo->lo_state == Lo_rundown); in lo_release()
1748 mutex_unlock(&lo->lo_mutex); in lo_release()
1751 __loop_clr_fd(lo); in lo_release()
1756 struct loop_device *lo = disk->private_data; in lo_free_disk() local
1758 if (lo->workqueue) in lo_free_disk()
1759 destroy_workqueue(lo->workqueue); in lo_free_disk()
1760 loop_free_idle_workers(lo, true); in lo_free_disk()
1761 timer_shutdown_sync(&lo->timer); in lo_free_disk()
1762 mutex_destroy(&lo->lo_mutex); in lo_free_disk()
1763 kfree(lo); in lo_free_disk()
1856 struct loop_device *lo = rq->q->queuedata; in loop_queue_rq() local
1860 if (data_race(READ_ONCE(lo->lo_state)) != Lo_bound) in loop_queue_rq()
1870 cmd->use_aio = lo->lo_flags & LO_FLAGS_DIRECT_IO; in loop_queue_rq()
1889 loop_queue_work(lo, cmd); in loop_queue_rq()
1900 struct loop_device *lo = rq->q->queuedata; in loop_handle_cmd() local
1904 if (write && (lo->lo_flags & LO_FLAGS_READ_ONLY)) { in loop_handle_cmd()
1925 ret = do_req_filebacked(lo, rq); in loop_handle_cmd()
1947 struct list_head *cmd_list, struct loop_device *lo) in loop_process_work() argument
1953 spin_lock_irq(&lo->lo_work_lock); in loop_process_work()
1958 spin_unlock_irq(&lo->lo_work_lock); in loop_process_work()
1963 spin_lock_irq(&lo->lo_work_lock); in loop_process_work()
1973 list_add_tail(&worker->idle_list, &lo->idle_worker_list); in loop_process_work()
1974 loop_set_timer(lo); in loop_process_work()
1976 spin_unlock_irq(&lo->lo_work_lock); in loop_process_work()
1984 loop_process_work(worker, &worker->cmd_list, worker->lo); in loop_workfn()
1989 struct loop_device *lo = in loop_rootcg_workfn() local
1991 loop_process_work(NULL, &lo->rootcg_cmd_list, lo); in loop_rootcg_workfn()
2007 struct loop_device *lo; in loop_add() local
2012 lo = kzalloc_obj(*lo); in loop_add()
2013 if (!lo) in loop_add()
2015 lo->worker_tree = RB_ROOT; in loop_add()
2016 INIT_LIST_HEAD(&lo->idle_worker_list); in loop_add()
2017 timer_setup(&lo->timer, loop_free_idle_workers_timer, TIMER_DEFERRABLE); in loop_add()
2018 WRITE_ONCE(lo->lo_state, Lo_unbound); in loop_add()
2026 err = idr_alloc(&loop_index_idr, lo, i, i + 1, GFP_KERNEL); in loop_add()
2030 err = idr_alloc(&loop_index_idr, lo, 0, 0, GFP_KERNEL); in loop_add()
2037 lo->tag_set.ops = &loop_mq_ops; in loop_add()
2038 lo->tag_set.nr_hw_queues = 1; in loop_add()
2039 lo->tag_set.queue_depth = hw_queue_depth; in loop_add()
2040 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add()
2041 lo->tag_set.cmd_size = sizeof(struct loop_cmd); in loop_add()
2042 lo->tag_set.flags = BLK_MQ_F_STACKING | BLK_MQ_F_NO_SCHED_BY_DEFAULT; in loop_add()
2043 lo->tag_set.driver_data = lo; in loop_add()
2045 err = blk_mq_alloc_tag_set(&lo->tag_set); in loop_add()
2049 disk = lo->lo_disk = blk_mq_alloc_disk(&lo->tag_set, &lim, lo); in loop_add()
2054 lo->lo_queue = lo->lo_disk->queue; in loop_add()
2076 mutex_init(&lo->lo_mutex); in loop_add()
2077 lo->lo_number = i; in loop_add()
2078 spin_lock_init(&lo->lo_lock); in loop_add()
2079 spin_lock_init(&lo->lo_work_lock); in loop_add()
2080 INIT_WORK(&lo->rootcg_work, loop_rootcg_workfn); in loop_add()
2081 INIT_LIST_HEAD(&lo->rootcg_cmd_list); in loop_add()
2086 disk->private_data = lo; in loop_add()
2087 disk->queue = lo->lo_queue; in loop_add()
2098 lo->idr_visible = true; in loop_add()
2106 blk_mq_free_tag_set(&lo->tag_set); in loop_add()
2112 kfree(lo); in loop_add()
2117 static void loop_remove(struct loop_device *lo) in loop_remove() argument
2120 del_gendisk(lo->lo_disk); in loop_remove()
2121 blk_mq_free_tag_set(&lo->tag_set); in loop_remove()
2124 idr_remove(&loop_index_idr, lo->lo_number); in loop_remove()
2127 put_disk(lo->lo_disk); in loop_remove()
2145 struct loop_device *lo; in loop_control_remove() local
2157 lo = idr_find(&loop_index_idr, idx); in loop_control_remove()
2158 if (!lo || !lo->idr_visible) in loop_control_remove()
2161 lo->idr_visible = false; in loop_control_remove()
2167 ret = mutex_lock_killable(&lo->lo_mutex); in loop_control_remove()
2170 if (lo->lo_state != Lo_unbound || disk_openers(lo->lo_disk) > 0) { in loop_control_remove()
2171 mutex_unlock(&lo->lo_mutex); in loop_control_remove()
2176 WRITE_ONCE(lo->lo_state, Lo_deleting); in loop_control_remove()
2177 mutex_unlock(&lo->lo_mutex); in loop_control_remove()
2179 loop_remove(lo); in loop_control_remove()
2185 lo->idr_visible = true; in loop_control_remove()
2192 struct loop_device *lo; in loop_control_get_free() local
2198 idr_for_each_entry(&loop_index_idr, lo, id) { in loop_control_get_free()
2203 if (lo->idr_visible && in loop_control_get_free()
2204 data_race(READ_ONCE(lo->lo_state)) == Lo_unbound) in loop_control_get_free()
2301 struct loop_device *lo; in loop_exit() local
2313 idr_for_each_entry(&loop_index_idr, lo, id) in loop_exit()
2314 loop_remove(lo); in loop_exit()