Lines Matching refs:lo
105 static int loop_global_lock_killable(struct loop_device *lo, bool global) in loop_global_lock_killable() argument
114 err = mutex_lock_killable(&lo->lo_mutex); in loop_global_lock_killable()
126 static void loop_global_unlock(struct loop_device *lo, bool global) in loop_global_unlock() argument
128 mutex_unlock(&lo->lo_mutex); in loop_global_unlock()
157 static loff_t get_loop_size(struct loop_device *lo, struct file *file) in get_loop_size() argument
159 return get_size(lo->lo_offset, lo->lo_sizelimit, file); in get_loop_size()
162 static void __loop_update_dio(struct loop_device *lo, bool dio) in __loop_update_dio() argument
164 struct file *file = lo->lo_backing_file; in __loop_update_dio()
186 if (queue_logical_block_size(lo->lo_queue) >= sb_bsize && in __loop_update_dio()
187 !(lo->lo_offset & dio_align) && in __loop_update_dio()
196 if (lo->use_dio == use_dio) in __loop_update_dio()
207 if (lo->lo_state == Lo_bound) in __loop_update_dio()
208 blk_mq_freeze_queue(lo->lo_queue); in __loop_update_dio()
209 lo->use_dio = use_dio; in __loop_update_dio()
211 blk_queue_flag_clear(QUEUE_FLAG_NOMERGES, lo->lo_queue); in __loop_update_dio()
212 lo->lo_flags |= LO_FLAGS_DIRECT_IO; in __loop_update_dio()
214 blk_queue_flag_set(QUEUE_FLAG_NOMERGES, lo->lo_queue); in __loop_update_dio()
215 lo->lo_flags &= ~LO_FLAGS_DIRECT_IO; in __loop_update_dio()
217 if (lo->lo_state == Lo_bound) in __loop_update_dio()
218 blk_mq_unfreeze_queue(lo->lo_queue); in __loop_update_dio()
229 static void loop_set_size(struct loop_device *lo, loff_t size) in loop_set_size() argument
231 if (!set_capacity_and_notify(lo->lo_disk, size)) in loop_set_size()
232 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in loop_set_size()
257 static int lo_write_simple(struct loop_device *lo, struct request *rq, in lo_write_simple() argument
265 ret = lo_write_bvec(lo->lo_backing_file, &bvec, &pos); in lo_write_simple()
274 static int lo_read_simple(struct loop_device *lo, struct request *rq, in lo_read_simple() argument
284 len = vfs_iter_read(lo->lo_backing_file, &i, &pos, 0); in lo_read_simple()
303 static int lo_fallocate(struct loop_device *lo, struct request *rq, loff_t pos, in lo_fallocate() argument
310 struct file *file = lo->lo_backing_file; in lo_fallocate()
311 struct request_queue *q = lo->lo_queue; in lo_fallocate()
328 static int lo_req_flush(struct loop_device *lo, struct request *rq) in lo_req_flush() argument
330 struct file *file = lo->lo_backing_file; in lo_req_flush()
393 static int lo_rw_aio(struct loop_device *lo, struct loop_cmd *cmd, in lo_rw_aio() argument
401 struct file *file = lo->lo_backing_file; in lo_rw_aio()
462 static int do_req_filebacked(struct loop_device *lo, struct request *rq) in do_req_filebacked() argument
465 loff_t pos = ((loff_t) blk_rq_pos(rq) << 9) + lo->lo_offset; in do_req_filebacked()
478 return lo_req_flush(lo, rq); in do_req_filebacked()
484 return lo_fallocate(lo, rq, pos, in do_req_filebacked()
489 return lo_fallocate(lo, rq, pos, FALLOC_FL_PUNCH_HOLE); in do_req_filebacked()
492 return lo_rw_aio(lo, cmd, pos, WRITE); in do_req_filebacked()
494 return lo_write_simple(lo, rq, pos); in do_req_filebacked()
497 return lo_rw_aio(lo, cmd, pos, READ); in do_req_filebacked()
499 return lo_read_simple(lo, rq, pos); in do_req_filebacked()
506 static inline void loop_update_dio(struct loop_device *lo) in loop_update_dio() argument
508 __loop_update_dio(lo, (lo->lo_backing_file->f_flags & O_DIRECT) | in loop_update_dio()
509 lo->use_dio); in loop_update_dio()
512 static void loop_reread_partitions(struct loop_device *lo) in loop_reread_partitions() argument
516 mutex_lock(&lo->lo_disk->open_mutex); in loop_reread_partitions()
517 rc = bdev_disk_changed(lo->lo_disk, false); in loop_reread_partitions()
518 mutex_unlock(&lo->lo_disk->open_mutex); in loop_reread_partitions()
521 __func__, lo->lo_number, lo->lo_file_name, rc); in loop_reread_partitions()
564 static int loop_change_fd(struct loop_device *lo, struct block_device *bdev, in loop_change_fd() argument
576 error = loop_global_lock_killable(lo, is_loop); in loop_change_fd()
580 if (lo->lo_state != Lo_bound) in loop_change_fd()
585 if (!(lo->lo_flags & LO_FLAGS_READ_ONLY)) in loop_change_fd()
592 old_file = lo->lo_backing_file; in loop_change_fd()
597 if (get_loop_size(lo, file) != get_loop_size(lo, old_file)) in loop_change_fd()
601 disk_force_media_change(lo->lo_disk, DISK_EVENT_MEDIA_CHANGE); in loop_change_fd()
602 blk_mq_freeze_queue(lo->lo_queue); in loop_change_fd()
603 mapping_set_gfp_mask(old_file->f_mapping, lo->old_gfp_mask); in loop_change_fd()
604 lo->lo_backing_file = file; in loop_change_fd()
605 lo->old_gfp_mask = mapping_gfp_mask(file->f_mapping); in loop_change_fd()
607 lo->old_gfp_mask & ~(__GFP_IO|__GFP_FS)); in loop_change_fd()
608 loop_update_dio(lo); in loop_change_fd()
609 blk_mq_unfreeze_queue(lo->lo_queue); in loop_change_fd()
610 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in loop_change_fd()
611 loop_global_unlock(lo, is_loop); in loop_change_fd()
628 loop_reread_partitions(lo); in loop_change_fd()
632 loop_global_unlock(lo, is_loop); in loop_change_fd()
644 struct loop_device *lo = disk->private_data; in loop_attr_show() local
646 return callback(lo, page); in loop_attr_show()
659 static ssize_t loop_attr_backing_file_show(struct loop_device *lo, char *buf) in loop_attr_backing_file_show() argument
664 spin_lock_irq(&lo->lo_lock); in loop_attr_backing_file_show()
665 if (lo->lo_backing_file) in loop_attr_backing_file_show()
666 p = file_path(lo->lo_backing_file, buf, PAGE_SIZE - 1); in loop_attr_backing_file_show()
667 spin_unlock_irq(&lo->lo_lock); in loop_attr_backing_file_show()
681 static ssize_t loop_attr_offset_show(struct loop_device *lo, char *buf) in loop_attr_offset_show() argument
683 return sprintf(buf, "%llu\n", (unsigned long long)lo->lo_offset); in loop_attr_offset_show()
686 static ssize_t loop_attr_sizelimit_show(struct loop_device *lo, char *buf) in loop_attr_sizelimit_show() argument
688 return sprintf(buf, "%llu\n", (unsigned long long)lo->lo_sizelimit); in loop_attr_sizelimit_show()
691 static ssize_t loop_attr_autoclear_show(struct loop_device *lo, char *buf) in loop_attr_autoclear_show() argument
693 int autoclear = (lo->lo_flags & LO_FLAGS_AUTOCLEAR); in loop_attr_autoclear_show()
698 static ssize_t loop_attr_partscan_show(struct loop_device *lo, char *buf) in loop_attr_partscan_show() argument
700 int partscan = (lo->lo_flags & LO_FLAGS_PARTSCAN); in loop_attr_partscan_show()
705 static ssize_t loop_attr_dio_show(struct loop_device *lo, char *buf) in loop_attr_dio_show() argument
707 int dio = (lo->lo_flags & LO_FLAGS_DIRECT_IO); in loop_attr_dio_show()
734 static void loop_sysfs_init(struct loop_device *lo) in loop_sysfs_init() argument
736 lo->sysfs_inited = !sysfs_create_group(&disk_to_dev(lo->lo_disk)->kobj, in loop_sysfs_init()
740 static void loop_sysfs_exit(struct loop_device *lo) in loop_sysfs_exit() argument
742 if (lo->sysfs_inited) in loop_sysfs_exit()
743 sysfs_remove_group(&disk_to_dev(lo->lo_disk)->kobj, in loop_sysfs_exit()
747 static void loop_config_discard(struct loop_device *lo) in loop_config_discard() argument
749 struct file *file = lo->lo_backing_file; in loop_config_discard()
751 struct request_queue *q = lo->lo_queue; in loop_config_discard()
800 struct loop_device *lo; member
821 static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd) in loop_queue_work() argument
823 struct rb_node **node = &(lo->worker_tree.rb_node), *parent = NULL; in loop_queue_work()
828 spin_lock_irq(&lo->lo_work_lock); in loop_queue_work()
833 node = &lo->worker_tree.rb_node; in loop_queue_work()
868 worker->lo = lo; in loop_queue_work()
870 rb_insert_color(&worker->rb_node, &lo->worker_tree); in loop_queue_work()
883 work = &lo->rootcg_work; in loop_queue_work()
884 cmd_list = &lo->rootcg_cmd_list; in loop_queue_work()
887 queue_work(lo->workqueue, work); in loop_queue_work()
888 spin_unlock_irq(&lo->lo_work_lock); in loop_queue_work()
891 static void loop_update_rotational(struct loop_device *lo) in loop_update_rotational() argument
893 struct file *file = lo->lo_backing_file; in loop_update_rotational()
896 struct request_queue *q = lo->lo_queue; in loop_update_rotational()
918 loop_set_status_from_info(struct loop_device *lo, in loop_set_status_from_info() argument
937 lo->lo_offset = info->lo_offset; in loop_set_status_from_info()
938 lo->lo_sizelimit = info->lo_sizelimit; in loop_set_status_from_info()
939 memcpy(lo->lo_file_name, info->lo_file_name, LO_NAME_SIZE); in loop_set_status_from_info()
940 lo->lo_file_name[LO_NAME_SIZE-1] = 0; in loop_set_status_from_info()
941 lo->lo_flags = info->lo_flags; in loop_set_status_from_info()
945 static int loop_configure(struct loop_device *lo, fmode_t mode, in loop_configure() argument
975 error = loop_global_lock_killable(lo, is_loop); in loop_configure()
980 if (lo->lo_state != Lo_unbound) in loop_configure()
1001 error = loop_set_status_from_info(lo, &config->info); in loop_configure()
1007 lo->lo_flags |= LO_FLAGS_READ_ONLY; in loop_configure()
1009 lo->workqueue = alloc_workqueue("loop%d", in loop_configure()
1012 lo->lo_number); in loop_configure()
1013 if (!lo->workqueue) { in loop_configure()
1018 disk_force_media_change(lo->lo_disk, DISK_EVENT_MEDIA_CHANGE); in loop_configure()
1019 set_disk_ro(lo->lo_disk, (lo->lo_flags & LO_FLAGS_READ_ONLY) != 0); in loop_configure()
1021 INIT_WORK(&lo->rootcg_work, loop_rootcg_workfn); in loop_configure()
1022 INIT_LIST_HEAD(&lo->rootcg_cmd_list); in loop_configure()
1023 INIT_LIST_HEAD(&lo->idle_worker_list); in loop_configure()
1024 lo->worker_tree = RB_ROOT; in loop_configure()
1025 timer_setup(&lo->timer, loop_free_idle_workers, in loop_configure()
1027 lo->use_dio = lo->lo_flags & LO_FLAGS_DIRECT_IO; in loop_configure()
1028 lo->lo_device = bdev; in loop_configure()
1029 lo->lo_backing_file = file; in loop_configure()
1030 lo->old_gfp_mask = mapping_gfp_mask(mapping); in loop_configure()
1031 mapping_set_gfp_mask(mapping, lo->old_gfp_mask & ~(__GFP_IO|__GFP_FS)); in loop_configure()
1033 if (!(lo->lo_flags & LO_FLAGS_READ_ONLY) && file->f_op->fsync) in loop_configure()
1034 blk_queue_write_cache(lo->lo_queue, true, false); in loop_configure()
1038 else if ((lo->lo_backing_file->f_flags & O_DIRECT) && inode->i_sb->s_bdev) in loop_configure()
1044 blk_queue_logical_block_size(lo->lo_queue, bsize); in loop_configure()
1045 blk_queue_physical_block_size(lo->lo_queue, bsize); in loop_configure()
1046 blk_queue_io_min(lo->lo_queue, bsize); in loop_configure()
1048 loop_config_discard(lo); in loop_configure()
1049 loop_update_rotational(lo); in loop_configure()
1050 loop_update_dio(lo); in loop_configure()
1051 loop_sysfs_init(lo); in loop_configure()
1053 size = get_loop_size(lo, file); in loop_configure()
1054 loop_set_size(lo, size); in loop_configure()
1059 lo->lo_state = Lo_bound; in loop_configure()
1061 lo->lo_flags |= LO_FLAGS_PARTSCAN; in loop_configure()
1062 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in loop_configure()
1064 lo->lo_disk->flags &= ~GENHD_FL_NO_PART_SCAN; in loop_configure()
1066 loop_global_unlock(lo, is_loop); in loop_configure()
1068 loop_reread_partitions(lo); in loop_configure()
1074 loop_global_unlock(lo, is_loop); in loop_configure()
1085 static int __loop_clr_fd(struct loop_device *lo, bool release) in __loop_clr_fd() argument
1088 gfp_t gfp = lo->old_gfp_mask; in __loop_clr_fd()
1106 mutex_lock(&lo->lo_mutex); in __loop_clr_fd()
1107 if (WARN_ON_ONCE(lo->lo_state != Lo_rundown)) { in __loop_clr_fd()
1112 filp = lo->lo_backing_file; in __loop_clr_fd()
1118 if (test_bit(QUEUE_FLAG_WC, &lo->lo_queue->queue_flags)) in __loop_clr_fd()
1119 blk_queue_write_cache(lo->lo_queue, false, false); in __loop_clr_fd()
1122 blk_mq_freeze_queue(lo->lo_queue); in __loop_clr_fd()
1124 destroy_workqueue(lo->workqueue); in __loop_clr_fd()
1125 spin_lock_irq(&lo->lo_work_lock); in __loop_clr_fd()
1126 list_for_each_entry_safe(worker, pos, &lo->idle_worker_list, in __loop_clr_fd()
1129 rb_erase(&worker->rb_node, &lo->worker_tree); in __loop_clr_fd()
1133 spin_unlock_irq(&lo->lo_work_lock); in __loop_clr_fd()
1134 del_timer_sync(&lo->timer); in __loop_clr_fd()
1136 spin_lock_irq(&lo->lo_lock); in __loop_clr_fd()
1137 lo->lo_backing_file = NULL; in __loop_clr_fd()
1138 spin_unlock_irq(&lo->lo_lock); in __loop_clr_fd()
1140 lo->lo_device = NULL; in __loop_clr_fd()
1141 lo->lo_offset = 0; in __loop_clr_fd()
1142 lo->lo_sizelimit = 0; in __loop_clr_fd()
1143 memset(lo->lo_file_name, 0, LO_NAME_SIZE); in __loop_clr_fd()
1144 blk_queue_logical_block_size(lo->lo_queue, 512); in __loop_clr_fd()
1145 blk_queue_physical_block_size(lo->lo_queue, 512); in __loop_clr_fd()
1146 blk_queue_io_min(lo->lo_queue, 512); in __loop_clr_fd()
1147 invalidate_disk(lo->lo_disk); in __loop_clr_fd()
1148 loop_sysfs_exit(lo); in __loop_clr_fd()
1150 kobject_uevent(&disk_to_dev(lo->lo_disk)->kobj, KOBJ_CHANGE); in __loop_clr_fd()
1154 blk_mq_unfreeze_queue(lo->lo_queue); in __loop_clr_fd()
1156 partscan = lo->lo_flags & LO_FLAGS_PARTSCAN; in __loop_clr_fd()
1157 lo_number = lo->lo_number; in __loop_clr_fd()
1158 disk_force_media_change(lo->lo_disk, DISK_EVENT_MEDIA_CHANGE); in __loop_clr_fd()
1160 mutex_unlock(&lo->lo_mutex); in __loop_clr_fd()
1171 mutex_lock(&lo->lo_disk->open_mutex); in __loop_clr_fd()
1172 err = bdev_disk_changed(lo->lo_disk, false); in __loop_clr_fd()
1174 mutex_unlock(&lo->lo_disk->open_mutex); in __loop_clr_fd()
1191 mutex_lock(&lo->lo_mutex); in __loop_clr_fd()
1192 lo->lo_flags = 0; in __loop_clr_fd()
1194 lo->lo_disk->flags |= GENHD_FL_NO_PART_SCAN; in __loop_clr_fd()
1195 lo->lo_state = Lo_unbound; in __loop_clr_fd()
1196 mutex_unlock(&lo->lo_mutex); in __loop_clr_fd()
1208 static int loop_clr_fd(struct loop_device *lo) in loop_clr_fd() argument
1212 err = mutex_lock_killable(&lo->lo_mutex); in loop_clr_fd()
1215 if (lo->lo_state != Lo_bound) { in loop_clr_fd()
1216 mutex_unlock(&lo->lo_mutex); in loop_clr_fd()
1229 if (atomic_read(&lo->lo_refcnt) > 1) { in loop_clr_fd()
1230 lo->lo_flags |= LO_FLAGS_AUTOCLEAR; in loop_clr_fd()
1231 mutex_unlock(&lo->lo_mutex); in loop_clr_fd()
1234 lo->lo_state = Lo_rundown; in loop_clr_fd()
1235 mutex_unlock(&lo->lo_mutex); in loop_clr_fd()
1237 return __loop_clr_fd(lo, false); in loop_clr_fd()
1241 loop_set_status(struct loop_device *lo, const struct loop_info64 *info) in loop_set_status() argument
1248 err = mutex_lock_killable(&lo->lo_mutex); in loop_set_status()
1251 if (lo->lo_state != Lo_bound) { in loop_set_status()
1256 if (lo->lo_offset != info->lo_offset || in loop_set_status()
1257 lo->lo_sizelimit != info->lo_sizelimit) { in loop_set_status()
1259 sync_blockdev(lo->lo_device); in loop_set_status()
1260 invalidate_bdev(lo->lo_device); in loop_set_status()
1264 blk_mq_freeze_queue(lo->lo_queue); in loop_set_status()
1266 if (size_changed && lo->lo_device->bd_inode->i_mapping->nrpages) { in loop_set_status()
1270 __func__, lo->lo_number, lo->lo_file_name, in loop_set_status()
1271 lo->lo_device->bd_inode->i_mapping->nrpages); in loop_set_status()
1275 prev_lo_flags = lo->lo_flags; in loop_set_status()
1277 err = loop_set_status_from_info(lo, info); in loop_set_status()
1282 lo->lo_flags &= LOOP_SET_STATUS_SETTABLE_FLAGS; in loop_set_status()
1284 lo->lo_flags |= prev_lo_flags & ~LOOP_SET_STATUS_SETTABLE_FLAGS; in loop_set_status()
1286 lo->lo_flags |= prev_lo_flags & ~LOOP_SET_STATUS_CLEARABLE_FLAGS; in loop_set_status()
1289 loff_t new_size = get_size(lo->lo_offset, lo->lo_sizelimit, in loop_set_status()
1290 lo->lo_backing_file); in loop_set_status()
1291 loop_set_size(lo, new_size); in loop_set_status()
1294 loop_config_discard(lo); in loop_set_status()
1297 __loop_update_dio(lo, lo->use_dio); in loop_set_status()
1300 blk_mq_unfreeze_queue(lo->lo_queue); in loop_set_status()
1302 if (!err && (lo->lo_flags & LO_FLAGS_PARTSCAN) && in loop_set_status()
1304 lo->lo_disk->flags &= ~GENHD_FL_NO_PART_SCAN; in loop_set_status()
1308 mutex_unlock(&lo->lo_mutex); in loop_set_status()
1310 loop_reread_partitions(lo); in loop_set_status()
1316 loop_get_status(struct loop_device *lo, struct loop_info64 *info) in loop_get_status() argument
1322 ret = mutex_lock_killable(&lo->lo_mutex); in loop_get_status()
1325 if (lo->lo_state != Lo_bound) { in loop_get_status()
1326 mutex_unlock(&lo->lo_mutex); in loop_get_status()
1331 info->lo_number = lo->lo_number; in loop_get_status()
1332 info->lo_offset = lo->lo_offset; in loop_get_status()
1333 info->lo_sizelimit = lo->lo_sizelimit; in loop_get_status()
1334 info->lo_flags = lo->lo_flags; in loop_get_status()
1335 memcpy(info->lo_file_name, lo->lo_file_name, LO_NAME_SIZE); in loop_get_status()
1338 path = lo->lo_backing_file->f_path; in loop_get_status()
1340 mutex_unlock(&lo->lo_mutex); in loop_get_status()
1388 loop_set_status_old(struct loop_device *lo, const struct loop_info __user *arg) in loop_set_status_old() argument
1396 return loop_set_status(lo, &info64); in loop_set_status_old()
1400 loop_set_status64(struct loop_device *lo, const struct loop_info64 __user *arg) in loop_set_status64() argument
1406 return loop_set_status(lo, &info64); in loop_set_status64()
1410 loop_get_status_old(struct loop_device *lo, struct loop_info __user *arg) { in loop_get_status_old() argument
1417 err = loop_get_status(lo, &info64); in loop_get_status_old()
1427 loop_get_status64(struct loop_device *lo, struct loop_info64 __user *arg) { in loop_get_status64() argument
1433 err = loop_get_status(lo, &info64); in loop_get_status64()
1440 static int loop_set_capacity(struct loop_device *lo) in loop_set_capacity() argument
1444 if (unlikely(lo->lo_state != Lo_bound)) in loop_set_capacity()
1447 size = get_loop_size(lo, lo->lo_backing_file); in loop_set_capacity()
1448 loop_set_size(lo, size); in loop_set_capacity()
1453 static int loop_set_dio(struct loop_device *lo, unsigned long arg) in loop_set_dio() argument
1456 if (lo->lo_state != Lo_bound) in loop_set_dio()
1459 __loop_update_dio(lo, !!arg); in loop_set_dio()
1460 if (lo->use_dio == !!arg) in loop_set_dio()
1467 static int loop_set_block_size(struct loop_device *lo, unsigned long arg) in loop_set_block_size() argument
1471 if (lo->lo_state != Lo_bound) in loop_set_block_size()
1478 if (lo->lo_queue->limits.logical_block_size == arg) in loop_set_block_size()
1481 sync_blockdev(lo->lo_device); in loop_set_block_size()
1482 invalidate_bdev(lo->lo_device); in loop_set_block_size()
1484 blk_mq_freeze_queue(lo->lo_queue); in loop_set_block_size()
1487 if (lo->lo_device->bd_inode->i_mapping->nrpages) { in loop_set_block_size()
1490 __func__, lo->lo_number, lo->lo_file_name, in loop_set_block_size()
1491 lo->lo_device->bd_inode->i_mapping->nrpages); in loop_set_block_size()
1495 blk_queue_logical_block_size(lo->lo_queue, arg); in loop_set_block_size()
1496 blk_queue_physical_block_size(lo->lo_queue, arg); in loop_set_block_size()
1497 blk_queue_io_min(lo->lo_queue, arg); in loop_set_block_size()
1498 loop_update_dio(lo); in loop_set_block_size()
1500 blk_mq_unfreeze_queue(lo->lo_queue); in loop_set_block_size()
1505 static int lo_simple_ioctl(struct loop_device *lo, unsigned int cmd, in lo_simple_ioctl() argument
1510 err = mutex_lock_killable(&lo->lo_mutex); in lo_simple_ioctl()
1515 err = loop_set_capacity(lo); in lo_simple_ioctl()
1518 err = loop_set_dio(lo, arg); in lo_simple_ioctl()
1521 err = loop_set_block_size(lo, arg); in lo_simple_ioctl()
1526 mutex_unlock(&lo->lo_mutex); in lo_simple_ioctl()
1533 struct loop_device *lo = bdev->bd_disk->private_data; in lo_ioctl() local
1549 return loop_configure(lo, mode, bdev, &config); in lo_ioctl()
1557 return loop_configure(lo, mode, bdev, &config); in lo_ioctl()
1560 return loop_change_fd(lo, bdev, arg); in lo_ioctl()
1562 return loop_clr_fd(lo); in lo_ioctl()
1566 err = loop_set_status_old(lo, argp); in lo_ioctl()
1570 return loop_get_status_old(lo, argp); in lo_ioctl()
1574 err = loop_set_status64(lo, argp); in lo_ioctl()
1578 return loop_get_status64(lo, argp); in lo_ioctl()
1586 err = lo_simple_ioctl(lo, cmd, arg); in lo_ioctl()
1665 loop_set_status_compat(struct loop_device *lo, in loop_set_status_compat() argument
1674 return loop_set_status(lo, &info64); in loop_set_status_compat()
1678 loop_get_status_compat(struct loop_device *lo, in loop_get_status_compat() argument
1686 err = loop_get_status(lo, &info64); in loop_get_status_compat()
1695 struct loop_device *lo = bdev->bd_disk->private_data; in lo_compat_ioctl() local
1700 err = loop_set_status_compat(lo, in lo_compat_ioctl()
1704 err = loop_get_status_compat(lo, in lo_compat_ioctl()
1730 struct loop_device *lo = bdev->bd_disk->private_data; in lo_open() local
1733 err = mutex_lock_killable(&lo->lo_mutex); in lo_open()
1736 if (lo->lo_state == Lo_deleting) in lo_open()
1739 atomic_inc(&lo->lo_refcnt); in lo_open()
1740 mutex_unlock(&lo->lo_mutex); in lo_open()
1746 struct loop_device *lo = disk->private_data; in lo_release() local
1748 mutex_lock(&lo->lo_mutex); in lo_release()
1749 if (atomic_dec_return(&lo->lo_refcnt)) in lo_release()
1752 if (lo->lo_flags & LO_FLAGS_AUTOCLEAR) { in lo_release()
1753 if (lo->lo_state != Lo_bound) in lo_release()
1755 lo->lo_state = Lo_rundown; in lo_release()
1756 mutex_unlock(&lo->lo_mutex); in lo_release()
1761 __loop_clr_fd(lo, true); in lo_release()
1763 } else if (lo->lo_state == Lo_bound) { in lo_release()
1768 blk_mq_freeze_queue(lo->lo_queue); in lo_release()
1769 blk_mq_unfreeze_queue(lo->lo_queue); in lo_release()
1773 mutex_unlock(&lo->lo_mutex); in lo_release()
1802 struct loop_device *lo = rq->q->queuedata; in loop_queue_rq() local
1806 if (lo->lo_state != Lo_bound) in loop_queue_rq()
1816 cmd->use_aio = lo->use_dio; in loop_queue_rq()
1833 loop_queue_work(lo, cmd); in loop_queue_rq()
1842 struct loop_device *lo = rq->q->queuedata; in loop_handle_cmd() local
1846 if (write && (lo->lo_flags & LO_FLAGS_READ_ONLY)) { in loop_handle_cmd()
1857 ret = do_req_filebacked(lo, rq); in loop_handle_cmd()
1878 static void loop_set_timer(struct loop_device *lo) in loop_set_timer() argument
1880 timer_reduce(&lo->timer, jiffies + LOOP_IDLE_WORKER_TIMEOUT); in loop_set_timer()
1884 struct list_head *cmd_list, struct loop_device *lo) in loop_process_work() argument
1890 spin_lock_irq(&lo->lo_work_lock); in loop_process_work()
1895 spin_unlock_irq(&lo->lo_work_lock); in loop_process_work()
1900 spin_lock_irq(&lo->lo_work_lock); in loop_process_work()
1910 list_add_tail(&worker->idle_list, &lo->idle_worker_list); in loop_process_work()
1911 loop_set_timer(lo); in loop_process_work()
1913 spin_unlock_irq(&lo->lo_work_lock); in loop_process_work()
1921 loop_process_work(worker, &worker->cmd_list, worker->lo); in loop_workfn()
1926 struct loop_device *lo = in loop_rootcg_workfn() local
1928 loop_process_work(NULL, &lo->rootcg_cmd_list, lo); in loop_rootcg_workfn()
1933 struct loop_device *lo = container_of(timer, struct loop_device, timer); in loop_free_idle_workers() local
1936 spin_lock_irq(&lo->lo_work_lock); in loop_free_idle_workers()
1937 list_for_each_entry_safe(worker, pos, &lo->idle_worker_list, in loop_free_idle_workers()
1943 rb_erase(&worker->rb_node, &lo->worker_tree); in loop_free_idle_workers()
1947 if (!list_empty(&lo->idle_worker_list)) in loop_free_idle_workers()
1948 loop_set_timer(lo); in loop_free_idle_workers()
1949 spin_unlock_irq(&lo->lo_work_lock); in loop_free_idle_workers()
1959 struct loop_device *lo; in loop_add() local
1964 lo = kzalloc(sizeof(*lo), GFP_KERNEL); in loop_add()
1965 if (!lo) in loop_add()
1967 lo->lo_state = Lo_unbound; in loop_add()
1975 err = idr_alloc(&loop_index_idr, lo, i, i + 1, GFP_KERNEL); in loop_add()
1979 err = idr_alloc(&loop_index_idr, lo, 0, 0, GFP_KERNEL); in loop_add()
1986 lo->tag_set.ops = &loop_mq_ops; in loop_add()
1987 lo->tag_set.nr_hw_queues = 1; in loop_add()
1988 lo->tag_set.queue_depth = 128; in loop_add()
1989 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add()
1990 lo->tag_set.cmd_size = sizeof(struct loop_cmd); in loop_add()
1991 lo->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_STACKING | in loop_add()
1993 lo->tag_set.driver_data = lo; in loop_add()
1995 err = blk_mq_alloc_tag_set(&lo->tag_set); in loop_add()
1999 disk = lo->lo_disk = blk_mq_alloc_disk(&lo->tag_set, lo); in loop_add()
2004 lo->lo_queue = lo->lo_disk->queue; in loop_add()
2006 blk_queue_max_hw_sectors(lo->lo_queue, BLK_DEF_MAX_SECTORS); in loop_add()
2014 blk_queue_flag_set(QUEUE_FLAG_NOMERGES, lo->lo_queue); in loop_add()
2037 atomic_set(&lo->lo_refcnt, 0); in loop_add()
2038 mutex_init(&lo->lo_mutex); in loop_add()
2039 lo->lo_number = i; in loop_add()
2040 spin_lock_init(&lo->lo_lock); in loop_add()
2041 spin_lock_init(&lo->lo_work_lock); in loop_add()
2046 disk->private_data = lo; in loop_add()
2047 disk->queue = lo->lo_queue; in loop_add()
2058 lo->idr_visible = true; in loop_add()
2066 blk_mq_free_tag_set(&lo->tag_set); in loop_add()
2072 kfree(lo); in loop_add()
2077 static void loop_remove(struct loop_device *lo) in loop_remove() argument
2080 del_gendisk(lo->lo_disk); in loop_remove()
2081 blk_cleanup_disk(lo->lo_disk); in loop_remove()
2082 blk_mq_free_tag_set(&lo->tag_set); in loop_remove()
2084 idr_remove(&loop_index_idr, lo->lo_number); in loop_remove()
2087 mutex_destroy(&lo->lo_mutex); in loop_remove()
2088 kfree(lo); in loop_remove()
2102 struct loop_device *lo; in loop_control_remove() local
2114 lo = idr_find(&loop_index_idr, idx); in loop_control_remove()
2115 if (!lo || !lo->idr_visible) in loop_control_remove()
2118 lo->idr_visible = false; in loop_control_remove()
2124 ret = mutex_lock_killable(&lo->lo_mutex); in loop_control_remove()
2127 if (lo->lo_state != Lo_unbound || in loop_control_remove()
2128 atomic_read(&lo->lo_refcnt) > 0) { in loop_control_remove()
2129 mutex_unlock(&lo->lo_mutex); in loop_control_remove()
2134 lo->lo_state = Lo_deleting; in loop_control_remove()
2135 mutex_unlock(&lo->lo_mutex); in loop_control_remove()
2137 loop_remove(lo); in loop_control_remove()
2143 lo->idr_visible = true; in loop_control_remove()
2150 struct loop_device *lo; in loop_control_get_free() local
2156 idr_for_each_entry(&loop_index_idr, lo, id) { in loop_control_get_free()
2158 if (lo->idr_visible && data_race(lo->lo_state) == Lo_unbound) in loop_control_get_free()
2268 struct loop_device *lo; in loop_exit() local
2280 idr_for_each_entry(&loop_index_idr, lo, id) in loop_exit()
2281 loop_remove(lo); in loop_exit()