Lines Matching full:lo
101 * @lo: struct loop_device
102 * @global: true if @lo is about to bind another "struct loop_device", false otherwise
110 static int loop_global_lock_killable(struct loop_device *lo, bool global) in loop_global_lock_killable() argument
119 err = mutex_lock_killable(&lo->lo_mutex); in loop_global_lock_killable()
128 * @lo: struct loop_device
129 * @global: true if @lo was about to bind another "struct loop_device", false otherwise
131 static void loop_global_unlock(struct loop_device *lo, bool global) in loop_global_unlock() argument
133 mutex_unlock(&lo->lo_mutex); in loop_global_unlock()
162 static loff_t get_loop_size(struct loop_device *lo, struct file *file) in get_loop_size() argument
164 return get_size(lo->lo_offset, lo->lo_sizelimit, file); in get_loop_size()
172 static bool lo_bdev_can_use_dio(struct loop_device *lo, in lo_bdev_can_use_dio() argument
177 if (queue_logical_block_size(lo->lo_queue) < sb_bsize) in lo_bdev_can_use_dio()
179 if (lo->lo_offset & (sb_bsize - 1)) in lo_bdev_can_use_dio()
184 static bool lo_can_use_dio(struct loop_device *lo) in lo_can_use_dio() argument
186 struct inode *inode = lo->lo_backing_file->f_mapping->host; in lo_can_use_dio()
188 if (!(lo->lo_backing_file->f_mode & FMODE_CAN_ODIRECT)) in lo_can_use_dio()
192 return lo_bdev_can_use_dio(lo, I_BDEV(inode)); in lo_can_use_dio()
194 return lo_bdev_can_use_dio(lo, inode->i_sb->s_bdev); in lo_can_use_dio()
206 static inline void loop_update_dio(struct loop_device *lo) in loop_update_dio() argument
208 bool dio_in_use = lo->lo_flags & LO_FLAGS_DIRECT_IO; in loop_update_dio()
210 lockdep_assert_held(&lo->lo_mutex); in loop_update_dio()
211 WARN_ON_ONCE(lo->lo_state == Lo_bound && in loop_update_dio()
212 lo->lo_queue->mq_freeze_depth == 0); in loop_update_dio()
214 if (lo->lo_backing_file->f_flags & O_DIRECT) in loop_update_dio()
215 lo->lo_flags |= LO_FLAGS_DIRECT_IO; in loop_update_dio()
216 if ((lo->lo_flags & LO_FLAGS_DIRECT_IO) && !lo_can_use_dio(lo)) in loop_update_dio()
217 lo->lo_flags &= ~LO_FLAGS_DIRECT_IO; in loop_update_dio()
220 if ((lo->lo_flags & LO_FLAGS_DIRECT_IO) && !dio_in_use) in loop_update_dio()
221 vfs_fsync(lo->lo_backing_file, 0); in loop_update_dio()
226 * @lo: struct loop_device to set the size for
232 static void loop_set_size(struct loop_device *lo, loff_t size) in loop_set_size() argument
234 if (!set_capacity_and_notify(lo->lo_disk, size)) in loop_set_size()
235 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_set_size()
238 static void loop_clear_limits(struct loop_device *lo, int mode) in loop_clear_limits() argument
240 struct queue_limits lim = queue_limits_start_update(lo->lo_queue); in loop_clear_limits()
257 queue_limits_commit_update(lo->lo_queue, &lim); in loop_clear_limits()
260 static int lo_fallocate(struct loop_device *lo, struct request *rq, loff_t pos, in lo_fallocate() argument
267 struct file *file = lo->lo_backing_file; in lo_fallocate()
272 if (!bdev_max_discard_sectors(lo->lo_device)) in lo_fallocate()
284 loop_clear_limits(lo, mode); in lo_fallocate()
289 static int lo_req_flush(struct loop_device *lo, struct request *rq) in lo_req_flush() argument
291 int ret = vfs_fsync(lo->lo_backing_file, 0); in lo_req_flush()
352 static int lo_rw_aio(struct loop_device *lo, struct loop_cmd *cmd, in lo_rw_aio() argument
360 struct file *file = lo->lo_backing_file; in lo_rw_aio()
426 static int do_req_filebacked(struct loop_device *lo, struct request *rq) in do_req_filebacked() argument
429 loff_t pos = ((loff_t) blk_rq_pos(rq) << 9) + lo->lo_offset; in do_req_filebacked()
433 return lo_req_flush(lo, rq); in do_req_filebacked()
439 return lo_fallocate(lo, rq, pos, in do_req_filebacked()
444 return lo_fallocate(lo, rq, pos, FALLOC_FL_PUNCH_HOLE); in do_req_filebacked()
446 return lo_rw_aio(lo, cmd, pos, ITER_SOURCE); in do_req_filebacked()
448 return lo_rw_aio(lo, cmd, pos, ITER_DEST); in do_req_filebacked()
455 static void loop_reread_partitions(struct loop_device *lo) in loop_reread_partitions() argument
459 mutex_lock(&lo->lo_disk->open_mutex); in loop_reread_partitions()
460 rc = bdev_disk_changed(lo->lo_disk, false); in loop_reread_partitions()
461 mutex_unlock(&lo->lo_disk->open_mutex); in loop_reread_partitions()
464 __func__, lo->lo_number, lo->lo_file_name, rc); in loop_reread_partitions()
490 /* Order wrt setting lo->lo_backing_file in loop_configure(). */ in loop_validate_file()
507 static int loop_change_fd(struct loop_device *lo, struct block_device *bdev, in loop_change_fd() argument
521 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 1); in loop_change_fd()
524 error = loop_global_lock_killable(lo, is_loop); in loop_change_fd()
528 if (lo->lo_state != Lo_bound) in loop_change_fd()
533 if (!(lo->lo_flags & LO_FLAGS_READ_ONLY)) in loop_change_fd()
540 old_file = lo->lo_backing_file; in loop_change_fd()
545 if (get_loop_size(lo, file) != get_loop_size(lo, old_file)) in loop_change_fd()
549 disk_force_media_change(lo->lo_disk); in loop_change_fd()
550 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_change_fd()
551 mapping_set_gfp_mask(old_file->f_mapping, lo->old_gfp_mask); in loop_change_fd()
552 lo->lo_backing_file = file; in loop_change_fd()
553 lo->old_gfp_mask = mapping_gfp_mask(file->f_mapping); in loop_change_fd()
555 lo->old_gfp_mask & ~(__GFP_IO|__GFP_FS)); in loop_change_fd()
556 loop_update_dio(lo); in loop_change_fd()
557 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_change_fd()
558 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in loop_change_fd()
559 loop_global_unlock(lo, is_loop); in loop_change_fd()
575 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0); in loop_change_fd()
577 loop_reread_partitions(lo); in loop_change_fd()
581 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_change_fd()
585 loop_global_unlock(lo, is_loop); in loop_change_fd()
588 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0); in loop_change_fd()
598 struct loop_device *lo = disk->private_data; in loop_attr_show() local
600 return callback(lo, page); in loop_attr_show()
613 static ssize_t loop_attr_backing_file_show(struct loop_device *lo, char *buf) in loop_attr_backing_file_show() argument
618 spin_lock_irq(&lo->lo_lock); in loop_attr_backing_file_show()
619 if (lo->lo_backing_file) in loop_attr_backing_file_show()
620 p = file_path(lo->lo_backing_file, buf, PAGE_SIZE - 1); in loop_attr_backing_file_show()
621 spin_unlock_irq(&lo->lo_lock); in loop_attr_backing_file_show()
635 static ssize_t loop_attr_offset_show(struct loop_device *lo, char *buf) in loop_attr_offset_show() argument
637 return sysfs_emit(buf, "%llu\n", (unsigned long long)lo->lo_offset); in loop_attr_offset_show()
640 static ssize_t loop_attr_sizelimit_show(struct loop_device *lo, char *buf) in loop_attr_sizelimit_show() argument
642 return sysfs_emit(buf, "%llu\n", (unsigned long long)lo->lo_sizelimit); in loop_attr_sizelimit_show()
645 static ssize_t loop_attr_autoclear_show(struct loop_device *lo, char *buf) in loop_attr_autoclear_show() argument
647 int autoclear = (lo->lo_flags & LO_FLAGS_AUTOCLEAR); in loop_attr_autoclear_show()
652 static ssize_t loop_attr_partscan_show(struct loop_device *lo, char *buf) in loop_attr_partscan_show() argument
654 int partscan = (lo->lo_flags & LO_FLAGS_PARTSCAN); in loop_attr_partscan_show()
659 static ssize_t loop_attr_dio_show(struct loop_device *lo, char *buf) in loop_attr_dio_show() argument
661 int dio = (lo->lo_flags & LO_FLAGS_DIRECT_IO); in loop_attr_dio_show()
688 static void loop_sysfs_init(struct loop_device *lo) in loop_sysfs_init() argument
690 lo->sysfs_inited = !sysfs_create_group(&disk_to_dev(lo->lo_disk)->kobj, in loop_sysfs_init()
694 static void loop_sysfs_exit(struct loop_device *lo) in loop_sysfs_exit() argument
696 if (lo->sysfs_inited) in loop_sysfs_exit()
697 sysfs_remove_group(&disk_to_dev(lo->lo_disk)->kobj, in loop_sysfs_exit()
701 static void loop_get_discard_config(struct loop_device *lo, in loop_get_discard_config() argument
704 struct file *file = lo->lo_backing_file; in loop_get_discard_config()
736 struct loop_device *lo; member
755 static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd) in loop_queue_work() argument
762 spin_lock_irq(&lo->lo_work_lock); in loop_queue_work()
767 node = &lo->worker_tree.rb_node; in loop_queue_work()
802 worker->lo = lo; in loop_queue_work()
804 rb_insert_color(&worker->rb_node, &lo->worker_tree); in loop_queue_work()
817 work = &lo->rootcg_work; in loop_queue_work()
818 cmd_list = &lo->rootcg_cmd_list; in loop_queue_work()
821 queue_work(lo->workqueue, work); in loop_queue_work()
822 spin_unlock_irq(&lo->lo_work_lock); in loop_queue_work()
825 static void loop_set_timer(struct loop_device *lo) in loop_set_timer() argument
827 timer_reduce(&lo->timer, jiffies + LOOP_IDLE_WORKER_TIMEOUT); in loop_set_timer()
830 static void loop_free_idle_workers(struct loop_device *lo, bool delete_all) in loop_free_idle_workers() argument
834 spin_lock_irq(&lo->lo_work_lock); in loop_free_idle_workers()
835 list_for_each_entry_safe(worker, pos, &lo->idle_worker_list, in loop_free_idle_workers()
842 rb_erase(&worker->rb_node, &lo->worker_tree); in loop_free_idle_workers()
846 if (!list_empty(&lo->idle_worker_list)) in loop_free_idle_workers()
847 loop_set_timer(lo); in loop_free_idle_workers()
848 spin_unlock_irq(&lo->lo_work_lock); in loop_free_idle_workers()
853 struct loop_device *lo = container_of(timer, struct loop_device, timer); in loop_free_idle_workers_timer() local
855 return loop_free_idle_workers(lo, false); in loop_free_idle_workers_timer()
860 * @lo: struct loop_device to configure
867 loop_set_status_from_info(struct loop_device *lo, in loop_set_status_from_info() argument
890 lo->lo_offset = info->lo_offset; in loop_set_status_from_info()
891 lo->lo_sizelimit = info->lo_sizelimit; in loop_set_status_from_info()
893 memcpy(lo->lo_file_name, info->lo_file_name, LO_NAME_SIZE); in loop_set_status_from_info()
894 lo->lo_file_name[LO_NAME_SIZE-1] = 0; in loop_set_status_from_info()
898 static unsigned int loop_default_blocksize(struct loop_device *lo, in loop_default_blocksize() argument
902 if ((lo->lo_backing_file->f_flags & O_DIRECT) && backing_bdev) in loop_default_blocksize()
907 static void loop_update_limits(struct loop_device *lo, struct queue_limits *lim, in loop_update_limits() argument
910 struct file *file = lo->lo_backing_file; in loop_update_limits()
921 bsize = loop_default_blocksize(lo, backing_bdev); in loop_update_limits()
923 loop_get_discard_config(lo, &granularity, &max_discard_sectors); in loop_update_limits()
929 if (file->f_op->fsync && !(lo->lo_flags & LO_FLAGS_READ_ONLY)) in loop_update_limits()
941 static int loop_configure(struct loop_device *lo, blk_mode_t mode, in loop_configure() argument
970 error = loop_global_lock_killable(lo, is_loop); in loop_configure()
975 if (lo->lo_state != Lo_unbound) in loop_configure()
989 error = loop_set_status_from_info(lo, &config->info); in loop_configure()
992 lo->lo_flags = config->info.lo_flags; in loop_configure()
996 lo->lo_flags |= LO_FLAGS_READ_ONLY; in loop_configure()
998 if (!lo->workqueue) { in loop_configure()
999 lo->workqueue = alloc_workqueue("loop%d", in loop_configure()
1001 0, lo->lo_number); in loop_configure()
1002 if (!lo->workqueue) { in loop_configure()
1009 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 1); in loop_configure()
1011 disk_force_media_change(lo->lo_disk); in loop_configure()
1012 set_disk_ro(lo->lo_disk, (lo->lo_flags & LO_FLAGS_READ_ONLY) != 0); in loop_configure()
1014 lo->lo_device = bdev; in loop_configure()
1015 lo->lo_backing_file = file; in loop_configure()
1016 lo->old_gfp_mask = mapping_gfp_mask(mapping); in loop_configure()
1017 mapping_set_gfp_mask(mapping, lo->old_gfp_mask & ~(__GFP_IO|__GFP_FS)); in loop_configure()
1019 lim = queue_limits_start_update(lo->lo_queue); in loop_configure()
1020 loop_update_limits(lo, &lim, config->block_size); in loop_configure()
1022 error = queue_limits_commit_update(lo->lo_queue, &lim); in loop_configure()
1026 loop_update_dio(lo); in loop_configure()
1027 loop_sysfs_init(lo); in loop_configure()
1029 size = get_loop_size(lo, file); in loop_configure()
1030 loop_set_size(lo, size); in loop_configure()
1035 lo->lo_state = Lo_bound; in loop_configure()
1037 lo->lo_flags |= LO_FLAGS_PARTSCAN; in loop_configure()
1038 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in loop_configure()
1040 clear_bit(GD_SUPPRESS_PART_SCAN, &lo->lo_disk->state); in loop_configure()
1042 dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0); in loop_configure()
1043 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_configure()
1045 loop_global_unlock(lo, is_loop); in loop_configure()
1047 loop_reread_partitions(lo); in loop_configure()
1055 loop_global_unlock(lo, is_loop); in loop_configure()
1066 static void __loop_clr_fd(struct loop_device *lo) in __loop_clr_fd() argument
1070 gfp_t gfp = lo->old_gfp_mask; in __loop_clr_fd()
1072 spin_lock_irq(&lo->lo_lock); in __loop_clr_fd()
1073 filp = lo->lo_backing_file; in __loop_clr_fd()
1074 lo->lo_backing_file = NULL; in __loop_clr_fd()
1075 spin_unlock_irq(&lo->lo_lock); in __loop_clr_fd()
1077 lo->lo_device = NULL; in __loop_clr_fd()
1078 lo->lo_offset = 0; in __loop_clr_fd()
1079 lo->lo_sizelimit = 0; in __loop_clr_fd()
1080 memset(lo->lo_file_name, 0, LO_NAME_SIZE); in __loop_clr_fd()
1088 lim = queue_limits_start_update(lo->lo_queue); in __loop_clr_fd()
1092 queue_limits_commit_update(lo->lo_queue, &lim); in __loop_clr_fd()
1094 invalidate_disk(lo->lo_disk); in __loop_clr_fd()
1095 loop_sysfs_exit(lo); in __loop_clr_fd()
1097 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in __loop_clr_fd()
1102 disk_force_media_change(lo->lo_disk); in __loop_clr_fd()
1104 if (lo->lo_flags & LO_FLAGS_PARTSCAN) { in __loop_clr_fd()
1115 err = bdev_disk_changed(lo->lo_disk, false); in __loop_clr_fd()
1118 __func__, lo->lo_number, err); in __loop_clr_fd()
1123 * lo->lo_state is set to Lo_unbound here after above partscan has in __loop_clr_fd()
1126 * change the 'lo' device. in __loop_clr_fd()
1128 lo->lo_flags = 0; in __loop_clr_fd()
1130 set_bit(GD_SUPPRESS_PART_SCAN, &lo->lo_disk->state); in __loop_clr_fd()
1131 mutex_lock(&lo->lo_mutex); in __loop_clr_fd()
1132 lo->lo_state = Lo_unbound; in __loop_clr_fd()
1133 mutex_unlock(&lo->lo_mutex); in __loop_clr_fd()
1143 static int loop_clr_fd(struct loop_device *lo) in loop_clr_fd() argument
1156 err = loop_global_lock_killable(lo, true); in loop_clr_fd()
1159 if (lo->lo_state != Lo_bound) { in loop_clr_fd()
1160 loop_global_unlock(lo, true); in loop_clr_fd()
1169 lo->lo_flags |= LO_FLAGS_AUTOCLEAR; in loop_clr_fd()
1170 if (disk_openers(lo->lo_disk) == 1) in loop_clr_fd()
1171 lo->lo_state = Lo_rundown; in loop_clr_fd()
1172 loop_global_unlock(lo, true); in loop_clr_fd()
1178 loop_set_status(struct loop_device *lo, const struct loop_info64 *info) in loop_set_status() argument
1185 err = mutex_lock_killable(&lo->lo_mutex); in loop_set_status()
1188 if (lo->lo_state != Lo_bound) { in loop_set_status()
1193 if (lo->lo_offset != info->lo_offset || in loop_set_status()
1194 lo->lo_sizelimit != info->lo_sizelimit) { in loop_set_status()
1196 sync_blockdev(lo->lo_device); in loop_set_status()
1197 invalidate_bdev(lo->lo_device); in loop_set_status()
1201 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_set_status()
1203 err = loop_set_status_from_info(lo, info); in loop_set_status()
1207 partscan = !(lo->lo_flags & LO_FLAGS_PARTSCAN) && in loop_set_status()
1210 lo->lo_flags &= ~LOOP_SET_STATUS_CLEARABLE_FLAGS; in loop_set_status()
1211 lo->lo_flags |= (info->lo_flags & LOOP_SET_STATUS_SETTABLE_FLAGS); in loop_set_status()
1214 loff_t new_size = get_size(lo->lo_offset, lo->lo_sizelimit, in loop_set_status()
1215 lo->lo_backing_file); in loop_set_status()
1216 loop_set_size(lo, new_size); in loop_set_status()
1220 loop_update_dio(lo); in loop_set_status()
1223 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_set_status()
1225 clear_bit(GD_SUPPRESS_PART_SCAN, &lo->lo_disk->state); in loop_set_status()
1227 mutex_unlock(&lo->lo_mutex); in loop_set_status()
1229 loop_reread_partitions(lo); in loop_set_status()
1235 loop_get_status(struct loop_device *lo, struct loop_info64 *info) in loop_get_status() argument
1241 ret = mutex_lock_killable(&lo->lo_mutex); in loop_get_status()
1244 if (lo->lo_state != Lo_bound) { in loop_get_status()
1245 mutex_unlock(&lo->lo_mutex); in loop_get_status()
1250 info->lo_number = lo->lo_number; in loop_get_status()
1251 info->lo_offset = lo->lo_offset; in loop_get_status()
1252 info->lo_sizelimit = lo->lo_sizelimit; in loop_get_status()
1253 info->lo_flags = lo->lo_flags; in loop_get_status()
1254 memcpy(info->lo_file_name, lo->lo_file_name, LO_NAME_SIZE); in loop_get_status()
1257 path = lo->lo_backing_file->f_path; in loop_get_status()
1259 mutex_unlock(&lo->lo_mutex); in loop_get_status()
1307 loop_set_status_old(struct loop_device *lo, const struct loop_info __user *arg) in loop_set_status_old() argument
1315 return loop_set_status(lo, &info64); in loop_set_status_old()
1319 loop_set_status64(struct loop_device *lo, const struct loop_info64 __user *arg) in loop_set_status64() argument
1325 return loop_set_status(lo, &info64); in loop_set_status64()
1329 loop_get_status_old(struct loop_device *lo, struct loop_info __user *arg) { in loop_get_status_old() argument
1336 err = loop_get_status(lo, &info64); in loop_get_status_old()
1346 loop_get_status64(struct loop_device *lo, struct loop_info64 __user *arg) { in loop_get_status64() argument
1352 err = loop_get_status(lo, &info64); in loop_get_status64()
1359 static int loop_set_capacity(struct loop_device *lo) in loop_set_capacity() argument
1363 if (unlikely(lo->lo_state != Lo_bound)) in loop_set_capacity()
1366 size = get_loop_size(lo, lo->lo_backing_file); in loop_set_capacity()
1367 loop_set_size(lo, size); in loop_set_capacity()
1372 static int loop_set_dio(struct loop_device *lo, unsigned long arg) in loop_set_dio() argument
1377 if (lo->lo_state != Lo_bound) in loop_set_dio()
1379 if (use_dio == !!(lo->lo_flags & LO_FLAGS_DIRECT_IO)) in loop_set_dio()
1383 if (!lo_can_use_dio(lo)) in loop_set_dio()
1386 vfs_fsync(lo->lo_backing_file, 0); in loop_set_dio()
1389 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_set_dio()
1391 lo->lo_flags |= LO_FLAGS_DIRECT_IO; in loop_set_dio()
1393 lo->lo_flags &= ~LO_FLAGS_DIRECT_IO; in loop_set_dio()
1394 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_set_dio()
1398 static int loop_set_block_size(struct loop_device *lo, unsigned long arg) in loop_set_block_size() argument
1404 if (lo->lo_state != Lo_bound) in loop_set_block_size()
1407 if (lo->lo_queue->limits.logical_block_size == arg) in loop_set_block_size()
1410 sync_blockdev(lo->lo_device); in loop_set_block_size()
1411 invalidate_bdev(lo->lo_device); in loop_set_block_size()
1413 lim = queue_limits_start_update(lo->lo_queue); in loop_set_block_size()
1414 loop_update_limits(lo, &lim, arg); in loop_set_block_size()
1416 memflags = blk_mq_freeze_queue(lo->lo_queue); in loop_set_block_size()
1417 err = queue_limits_commit_update(lo->lo_queue, &lim); in loop_set_block_size()
1418 loop_update_dio(lo); in loop_set_block_size()
1419 blk_mq_unfreeze_queue(lo->lo_queue, memflags); in loop_set_block_size()
1424 static int lo_simple_ioctl(struct loop_device *lo, unsigned int cmd, in lo_simple_ioctl() argument
1429 err = mutex_lock_killable(&lo->lo_mutex); in lo_simple_ioctl()
1434 err = loop_set_capacity(lo); in lo_simple_ioctl()
1437 err = loop_set_dio(lo, arg); in lo_simple_ioctl()
1440 err = loop_set_block_size(lo, arg); in lo_simple_ioctl()
1445 mutex_unlock(&lo->lo_mutex); in lo_simple_ioctl()
1452 struct loop_device *lo = bdev->bd_disk->private_data; in lo_ioctl() local
1468 return loop_configure(lo, mode, bdev, &config); in lo_ioctl()
1476 return loop_configure(lo, mode, bdev, &config); in lo_ioctl()
1479 return loop_change_fd(lo, bdev, arg); in lo_ioctl()
1481 return loop_clr_fd(lo); in lo_ioctl()
1485 err = loop_set_status_old(lo, argp); in lo_ioctl()
1488 return loop_get_status_old(lo, argp); in lo_ioctl()
1492 err = loop_set_status64(lo, argp); in lo_ioctl()
1495 return loop_get_status64(lo, argp); in lo_ioctl()
1503 err = lo_simple_ioctl(lo, cmd, arg); in lo_ioctl()
1583 loop_set_status_compat(struct loop_device *lo, in loop_set_status_compat() argument
1592 return loop_set_status(lo, &info64); in loop_set_status_compat()
1596 loop_get_status_compat(struct loop_device *lo, in loop_get_status_compat() argument
1604 err = loop_get_status(lo, &info64); in loop_get_status_compat()
1613 struct loop_device *lo = bdev->bd_disk->private_data; in lo_compat_ioctl() local
1618 err = loop_set_status_compat(lo, in lo_compat_ioctl()
1622 err = loop_get_status_compat(lo, in lo_compat_ioctl()
1648 struct loop_device *lo = disk->private_data; in lo_open() local
1651 err = mutex_lock_killable(&lo->lo_mutex); in lo_open()
1655 if (lo->lo_state == Lo_deleting || lo->lo_state == Lo_rundown) in lo_open()
1657 mutex_unlock(&lo->lo_mutex); in lo_open()
1663 struct loop_device *lo = disk->private_data; in lo_release() local
1674 mutex_lock(&lo->lo_mutex); in lo_release()
1675 if (lo->lo_state == Lo_bound && (lo->lo_flags & LO_FLAGS_AUTOCLEAR)) in lo_release()
1676 lo->lo_state = Lo_rundown; in lo_release()
1678 need_clear = (lo->lo_state == Lo_rundown); in lo_release()
1679 mutex_unlock(&lo->lo_mutex); in lo_release()
1682 __loop_clr_fd(lo); in lo_release()
1687 struct loop_device *lo = disk->private_data; in lo_free_disk() local
1689 if (lo->workqueue) in lo_free_disk()
1690 destroy_workqueue(lo->workqueue); in lo_free_disk()
1691 loop_free_idle_workers(lo, true); in lo_free_disk()
1692 timer_shutdown_sync(&lo->timer); in lo_free_disk()
1693 mutex_destroy(&lo->lo_mutex); in lo_free_disk()
1694 kfree(lo); in lo_free_disk()
1787 struct loop_device *lo = rq->q->queuedata; in loop_queue_rq() local
1791 if (lo->lo_state != Lo_bound) in loop_queue_rq()
1801 cmd->use_aio = lo->lo_flags & LO_FLAGS_DIRECT_IO; in loop_queue_rq()
1820 loop_queue_work(lo, cmd); in loop_queue_rq()
1831 struct loop_device *lo = rq->q->queuedata; in loop_handle_cmd() local
1835 if (write && (lo->lo_flags & LO_FLAGS_READ_ONLY)) { in loop_handle_cmd()
1852 ret = do_req_filebacked(lo, rq); in loop_handle_cmd()
1874 struct list_head *cmd_list, struct loop_device *lo) in loop_process_work() argument
1880 spin_lock_irq(&lo->lo_work_lock); in loop_process_work()
1885 spin_unlock_irq(&lo->lo_work_lock); in loop_process_work()
1890 spin_lock_irq(&lo->lo_work_lock); in loop_process_work()
1900 list_add_tail(&worker->idle_list, &lo->idle_worker_list); in loop_process_work()
1901 loop_set_timer(lo); in loop_process_work()
1903 spin_unlock_irq(&lo->lo_work_lock); in loop_process_work()
1911 loop_process_work(worker, &worker->cmd_list, worker->lo); in loop_workfn()
1916 struct loop_device *lo = in loop_rootcg_workfn() local
1918 loop_process_work(NULL, &lo->rootcg_cmd_list, lo); in loop_rootcg_workfn()
1934 struct loop_device *lo; in loop_add() local
1939 lo = kzalloc(sizeof(*lo), GFP_KERNEL); in loop_add()
1940 if (!lo) in loop_add()
1942 lo->worker_tree = RB_ROOT; in loop_add()
1943 INIT_LIST_HEAD(&lo->idle_worker_list); in loop_add()
1944 timer_setup(&lo->timer, loop_free_idle_workers_timer, TIMER_DEFERRABLE); in loop_add()
1945 lo->lo_state = Lo_unbound; in loop_add()
1953 err = idr_alloc(&loop_index_idr, lo, i, i + 1, GFP_KERNEL); in loop_add()
1957 err = idr_alloc(&loop_index_idr, lo, 0, 0, GFP_KERNEL); in loop_add()
1964 lo->tag_set.ops = &loop_mq_ops; in loop_add()
1965 lo->tag_set.nr_hw_queues = 1; in loop_add()
1966 lo->tag_set.queue_depth = hw_queue_depth; in loop_add()
1967 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add()
1968 lo->tag_set.cmd_size = sizeof(struct loop_cmd); in loop_add()
1969 lo->tag_set.flags = BLK_MQ_F_STACKING | BLK_MQ_F_NO_SCHED_BY_DEFAULT; in loop_add()
1970 lo->tag_set.driver_data = lo; in loop_add()
1972 err = blk_mq_alloc_tag_set(&lo->tag_set); in loop_add()
1976 disk = lo->lo_disk = blk_mq_alloc_disk(&lo->tag_set, &lim, lo); in loop_add()
1981 lo->lo_queue = lo->lo_disk->queue; in loop_add()
2003 mutex_init(&lo->lo_mutex); in loop_add()
2004 lo->lo_number = i; in loop_add()
2005 spin_lock_init(&lo->lo_lock); in loop_add()
2006 spin_lock_init(&lo->lo_work_lock); in loop_add()
2007 INIT_WORK(&lo->rootcg_work, loop_rootcg_workfn); in loop_add()
2008 INIT_LIST_HEAD(&lo->rootcg_cmd_list); in loop_add()
2013 disk->private_data = lo; in loop_add()
2014 disk->queue = lo->lo_queue; in loop_add()
2025 lo->idr_visible = true; in loop_add()
2033 blk_mq_free_tag_set(&lo->tag_set); in loop_add()
2039 kfree(lo); in loop_add()
2044 static void loop_remove(struct loop_device *lo) in loop_remove() argument
2047 del_gendisk(lo->lo_disk); in loop_remove()
2048 blk_mq_free_tag_set(&lo->tag_set); in loop_remove()
2051 idr_remove(&loop_index_idr, lo->lo_number); in loop_remove()
2054 put_disk(lo->lo_disk); in loop_remove()
2072 struct loop_device *lo; in loop_control_remove() local
2084 lo = idr_find(&loop_index_idr, idx); in loop_control_remove()
2085 if (!lo || !lo->idr_visible) in loop_control_remove()
2088 lo->idr_visible = false; in loop_control_remove()
2094 ret = mutex_lock_killable(&lo->lo_mutex); in loop_control_remove()
2097 if (lo->lo_state != Lo_unbound || disk_openers(lo->lo_disk) > 0) { in loop_control_remove()
2098 mutex_unlock(&lo->lo_mutex); in loop_control_remove()
2103 lo->lo_state = Lo_deleting; in loop_control_remove()
2104 mutex_unlock(&lo->lo_mutex); in loop_control_remove()
2106 loop_remove(lo); in loop_control_remove()
2112 lo->idr_visible = true; in loop_control_remove()
2119 struct loop_device *lo; in loop_control_get_free() local
2125 idr_for_each_entry(&loop_index_idr, lo, id) { in loop_control_get_free()
2127 if (lo->idr_visible && data_race(lo->lo_state) == Lo_unbound) in loop_control_get_free()
2224 struct loop_device *lo; in loop_exit() local
2236 idr_for_each_entry(&loop_index_idr, lo, id) in loop_exit()
2237 loop_remove(lo); in loop_exit()