Lines Matching full:sb
42 static int thaw_super_locked(struct super_block *sb, enum freeze_holder who);
53 static inline void __super_lock(struct super_block *sb, bool excl) in __super_lock() argument
56 down_write(&sb->s_umount); in __super_lock()
58 down_read(&sb->s_umount); in __super_lock()
61 static inline void super_unlock(struct super_block *sb, bool excl) in super_unlock() argument
64 up_write(&sb->s_umount); in super_unlock()
66 up_read(&sb->s_umount); in super_unlock()
69 static inline void __super_lock_excl(struct super_block *sb) in __super_lock_excl() argument
71 __super_lock(sb, true); in __super_lock_excl()
74 static inline void super_unlock_excl(struct super_block *sb) in super_unlock_excl() argument
76 super_unlock(sb, true); in super_unlock_excl()
79 static inline void super_unlock_shared(struct super_block *sb) in super_unlock_shared() argument
81 super_unlock(sb, false); in super_unlock_shared()
84 static bool super_flags(const struct super_block *sb, unsigned int flags) in super_flags() argument
90 return smp_load_acquire(&sb->s_flags) & flags; in super_flags()
95 * @sb: superblock to wait for
103 * The caller must have acquired a temporary reference on @sb->s_count.
109 static __must_check bool super_lock(struct super_block *sb, bool excl) in super_lock() argument
111 lockdep_assert_not_held(&sb->s_umount); in super_lock()
114 wait_var_event(&sb->s_flags, super_flags(sb, SB_BORN | SB_DYING)); in super_lock()
117 if (super_flags(sb, SB_DYING)) in super_lock()
120 __super_lock(sb, excl); in super_lock()
124 * @sb->s_root is NULL and @sb->s_active is 0. No one needs to in super_lock()
127 if (sb->s_flags & SB_DYING) { in super_lock()
128 super_unlock(sb, excl); in super_lock()
132 WARN_ON_ONCE(!(sb->s_flags & SB_BORN)); in super_lock()
136 /* wait and try to acquire read-side of @sb->s_umount */
137 static inline bool super_lock_shared(struct super_block *sb) in super_lock_shared() argument
139 return super_lock(sb, false); in super_lock_shared()
142 /* wait and try to acquire write-side of @sb->s_umount */
143 static inline bool super_lock_excl(struct super_block *sb) in super_lock_excl() argument
145 return super_lock(sb, true); in super_lock_excl()
150 static void super_wake(struct super_block *sb, unsigned int flag) in super_wake() argument
160 smp_store_release(&sb->s_flags, sb->s_flags | flag); in super_wake()
167 wake_up_var(&sb->s_flags); in super_wake()
171 * One thing we have to be careful of with a per-sb shrinker is that we don't
180 struct super_block *sb; in super_cache_scan() local
187 sb = shrink->private_data; in super_cache_scan()
196 if (!super_trylock_shared(sb)) in super_cache_scan()
199 if (sb->s_op->nr_cached_objects) in super_cache_scan()
200 fs_objects = sb->s_op->nr_cached_objects(sb, sc); in super_cache_scan()
202 inodes = list_lru_shrink_count(&sb->s_inode_lru, sc); in super_cache_scan()
203 dentries = list_lru_shrink_count(&sb->s_dentry_lru, sc); in super_cache_scan()
221 freed = prune_dcache_sb(sb, sc); in super_cache_scan()
223 freed += prune_icache_sb(sb, sc); in super_cache_scan()
227 freed += sb->s_op->free_cached_objects(sb, sc); in super_cache_scan()
230 super_unlock_shared(sb); in super_cache_scan()
237 struct super_block *sb; in super_cache_count() local
240 sb = shrink->private_data; in super_cache_count()
256 if (!(sb->s_flags & SB_BORN)) in super_cache_count()
260 if (sb->s_op && sb->s_op->nr_cached_objects) in super_cache_count()
261 total_objects = sb->s_op->nr_cached_objects(sb, sc); in super_cache_count()
263 total_objects += list_lru_shrink_count(&sb->s_dentry_lru, sc); in super_cache_count()
264 total_objects += list_lru_shrink_count(&sb->s_inode_lru, sc); in super_cache_count()
331 * When it cannot find a suitable sb, it allocates a new in alloc_super()
378 "sb-%s", type->name); in alloc_super()
416 * @sb: superblock in question
421 void put_super(struct super_block *sb) in put_super() argument
424 __put_super(sb); in put_super()
428 static void kill_super_notify(struct super_block *sb) in kill_super_notify() argument
430 lockdep_assert_not_held(&sb->s_umount); in kill_super_notify()
433 if (sb->s_flags & SB_DEAD) in kill_super_notify()
444 hlist_del_init(&sb->s_instances); in kill_super_notify()
449 * We don't need @sb->s_umount here as every concurrent caller in kill_super_notify()
453 super_wake(sb, SB_DEAD); in kill_super_notify()
513 * @sb: superblock to acquire
518 * sb->kill() and be marked as SB_DEAD.
523 static bool grab_super(struct super_block *sb) in grab_super() argument
527 sb->s_count++; in grab_super()
529 locked = super_lock_excl(sb); in grab_super()
531 if (atomic_inc_not_zero(&sb->s_active)) { in grab_super()
532 put_super(sb); in grab_super()
535 super_unlock_excl(sb); in grab_super()
537 wait_var_event(&sb->s_flags, super_flags(sb, SB_DEAD)); in grab_super()
538 put_super(sb); in grab_super()
544 * @sb: reference we are trying to grab
559 bool super_trylock_shared(struct super_block *sb) in super_trylock_shared() argument
561 if (down_read_trylock(&sb->s_umount)) { in super_trylock_shared()
562 if (!(sb->s_flags & SB_DYING) && sb->s_root && in super_trylock_shared()
563 (sb->s_flags & SB_BORN)) in super_trylock_shared()
565 super_unlock_shared(sb); in super_trylock_shared()
573 * @sb: superblock to retire
587 void retire_super(struct super_block *sb) in retire_super() argument
589 WARN_ON(!sb->s_bdev); in retire_super()
590 __super_lock_excl(sb); in retire_super()
591 if (sb->s_iflags & SB_I_PERSB_BDI) { in retire_super()
592 bdi_unregister(sb->s_bdi); in retire_super()
593 sb->s_iflags &= ~SB_I_PERSB_BDI; in retire_super()
595 sb->s_iflags |= SB_I_RETIRED; in retire_super()
596 super_unlock_excl(sb); in retire_super()
602 * @sb: superblock to kill
614 void generic_shutdown_super(struct super_block *sb) in generic_shutdown_super() argument
616 const struct super_operations *sop = sb->s_op; in generic_shutdown_super()
618 if (sb->s_root) { in generic_shutdown_super()
619 shrink_dcache_for_umount(sb); in generic_shutdown_super()
620 sync_filesystem(sb); in generic_shutdown_super()
621 sb->s_flags &= ~SB_ACTIVE; in generic_shutdown_super()
626 evict_inodes(sb); in generic_shutdown_super()
632 fsnotify_sb_delete(sb); in generic_shutdown_super()
633 security_sb_delete(sb); in generic_shutdown_super()
635 if (sb->s_dio_done_wq) { in generic_shutdown_super()
636 destroy_workqueue(sb->s_dio_done_wq); in generic_shutdown_super()
637 sb->s_dio_done_wq = NULL; in generic_shutdown_super()
641 sop->put_super(sb); in generic_shutdown_super()
647 fscrypt_destroy_keyring(sb); in generic_shutdown_super()
649 if (CHECK_DATA_CORRUPTION(!list_empty(&sb->s_inodes), in generic_shutdown_super()
651 sb->s_id, sb->s_type->name)) { in generic_shutdown_super()
659 spin_lock(&sb->s_inode_list_lock); in generic_shutdown_super()
660 list_for_each_entry(inode, &sb->s_inodes, i_sb_list) { in generic_shutdown_super()
665 spin_unlock(&sb->s_inode_list_lock); in generic_shutdown_super()
675 * sget{_fc}() until we passed sb->kill_sb(). in generic_shutdown_super()
677 super_wake(sb, SB_DYING); in generic_shutdown_super()
678 super_unlock_excl(sb); in generic_shutdown_super()
679 if (sb->s_bdi != &noop_backing_dev_info) { in generic_shutdown_super()
680 if (sb->s_iflags & SB_I_PERSB_BDI) in generic_shutdown_super()
681 bdi_unregister(sb->s_bdi); in generic_shutdown_super()
682 bdi_put(sb->s_bdi); in generic_shutdown_super()
683 sb->s_bdi = &noop_backing_dev_info; in generic_shutdown_super()
863 void drop_super(struct super_block *sb) in drop_super() argument
865 super_unlock_shared(sb); in drop_super()
866 put_super(sb); in drop_super()
871 void drop_super_exclusive(struct super_block *sb) in drop_super_exclusive() argument
873 super_unlock_excl(sb); in drop_super_exclusive()
874 put_super(sb); in drop_super_exclusive()
880 struct super_block *sb, *p = NULL; in __iterate_supers() local
883 list_for_each_entry(sb, &super_blocks, s_list) { in __iterate_supers()
884 if (super_flags(sb, SB_DYING)) in __iterate_supers()
886 sb->s_count++; in __iterate_supers()
889 f(sb); in __iterate_supers()
894 p = sb; in __iterate_supers()
910 struct super_block *sb, *p = NULL; in iterate_supers() local
913 list_for_each_entry(sb, &super_blocks, s_list) { in iterate_supers()
916 sb->s_count++; in iterate_supers()
919 locked = super_lock_shared(sb); in iterate_supers()
921 if (sb->s_root) in iterate_supers()
922 f(sb, arg); in iterate_supers()
923 super_unlock_shared(sb); in iterate_supers()
929 p = sb; in iterate_supers()
948 struct super_block *sb, *p = NULL; in iterate_supers_type() local
951 hlist_for_each_entry(sb, &type->fs_supers, s_instances) { in iterate_supers_type()
954 sb->s_count++; in iterate_supers_type()
957 locked = super_lock_shared(sb); in iterate_supers_type()
959 if (sb->s_root) in iterate_supers_type()
960 f(sb, arg); in iterate_supers_type()
961 super_unlock_shared(sb); in iterate_supers_type()
967 p = sb; in iterate_supers_type()
978 struct super_block *sb; in user_get_super() local
981 list_for_each_entry(sb, &super_blocks, s_list) { in user_get_super()
982 if (sb->s_dev == dev) { in user_get_super()
985 sb->s_count++; in user_get_super()
988 locked = super_lock(sb, excl); in user_get_super()
990 if (sb->s_root) in user_get_super()
991 return sb; in user_get_super()
992 super_unlock(sb, excl); in user_get_super()
996 __put_super(sb); in user_get_super()
1012 struct super_block *sb = fc->root->d_sb; in reconfigure_super() local
1020 if (sb->s_writers.frozen != SB_UNFROZEN) in reconfigure_super()
1023 retval = security_sb_remount(sb, fc->security); in reconfigure_super()
1029 if (!(fc->sb_flags & SB_RDONLY) && sb->s_bdev && in reconfigure_super()
1030 bdev_read_only(sb->s_bdev)) in reconfigure_super()
1033 remount_rw = !(fc->sb_flags & SB_RDONLY) && sb_rdonly(sb); in reconfigure_super()
1034 remount_ro = (fc->sb_flags & SB_RDONLY) && !sb_rdonly(sb); in reconfigure_super()
1038 if (!hlist_empty(&sb->s_pins)) { in reconfigure_super()
1039 super_unlock_excl(sb); in reconfigure_super()
1040 group_pin_kill(&sb->s_pins); in reconfigure_super()
1041 __super_lock_excl(sb); in reconfigure_super()
1042 if (!sb->s_root) in reconfigure_super()
1044 if (sb->s_writers.frozen != SB_UNFROZEN) in reconfigure_super()
1046 remount_ro = !sb_rdonly(sb); in reconfigure_super()
1049 shrink_dcache_sb(sb); in reconfigure_super()
1051 /* If we are reconfiguring to RDONLY and current sb is read/write, in reconfigure_super()
1056 sb_start_ro_state_change(sb); in reconfigure_super()
1058 retval = sb_prepare_remount_readonly(sb); in reconfigure_super()
1067 sb_start_ro_state_change(sb); in reconfigure_super()
1077 sb->s_type->name, retval); in reconfigure_super()
1081 WRITE_ONCE(sb->s_flags, ((sb->s_flags & ~fc->sb_flags_mask) | in reconfigure_super()
1083 sb_end_ro_state_change(sb); in reconfigure_super()
1093 if (remount_ro && sb->s_bdev) in reconfigure_super()
1094 invalidate_bdev(sb->s_bdev); in reconfigure_super()
1098 sb_end_ro_state_change(sb); in reconfigure_super()
1102 static void do_emergency_remount_callback(struct super_block *sb) in do_emergency_remount_callback() argument
1104 bool locked = super_lock_excl(sb); in do_emergency_remount_callback()
1106 if (locked && sb->s_root && sb->s_bdev && !sb_rdonly(sb)) { in do_emergency_remount_callback()
1109 fc = fs_context_for_reconfigure(sb->s_root, in do_emergency_remount_callback()
1118 super_unlock_excl(sb); in do_emergency_remount_callback()
1139 static void do_thaw_all_callback(struct super_block *sb) in do_thaw_all_callback() argument
1141 bool locked = super_lock_excl(sb); in do_thaw_all_callback()
1143 if (locked && sb->s_root) { in do_thaw_all_callback()
1145 while (sb->s_bdev && !bdev_thaw(sb->s_bdev)) in do_thaw_all_callback()
1146 pr_warn("Emergency Thaw on %pg\n", sb->s_bdev); in do_thaw_all_callback()
1147 thaw_super_locked(sb, FREEZE_HOLDER_USERSPACE); in do_thaw_all_callback()
1151 super_unlock_excl(sb); in do_thaw_all_callback()
1222 void kill_anon_super(struct super_block *sb) in kill_anon_super() argument
1224 dev_t dev = sb->s_dev; in kill_anon_super()
1225 generic_shutdown_super(sb); in kill_anon_super()
1226 kill_super_notify(sb); in kill_anon_super()
1231 void kill_litter_super(struct super_block *sb) in kill_litter_super() argument
1233 if (sb->s_root) in kill_litter_super()
1234 d_genocide(sb->s_root); in kill_litter_super()
1235 kill_anon_super(sb); in kill_litter_super()
1239 int set_anon_super_fc(struct super_block *sb, struct fs_context *fc) in set_anon_super_fc() argument
1241 return set_anon_super(sb, NULL); in set_anon_super_fc()
1245 static int test_keyed_super(struct super_block *sb, struct fs_context *fc) in test_keyed_super() argument
1247 return sb->s_fs_info == fc->s_fs_info; in test_keyed_super()
1257 int (*fill_super)(struct super_block *sb, in vfs_get_super() argument
1260 struct super_block *sb; in vfs_get_super() local
1263 sb = sget_fc(fc, test, set_anon_super_fc); in vfs_get_super()
1264 if (IS_ERR(sb)) in vfs_get_super()
1265 return PTR_ERR(sb); in vfs_get_super()
1267 if (!sb->s_root) { in vfs_get_super()
1268 err = fill_super(sb, fc); in vfs_get_super()
1272 sb->s_flags |= SB_ACTIVE; in vfs_get_super()
1275 fc->root = dget(sb->s_root); in vfs_get_super()
1279 deactivate_locked_super(sb); in vfs_get_super()
1284 int (*fill_super)(struct super_block *sb, in get_tree_nodev() argument
1292 int (*fill_super)(struct super_block *sb, in get_tree_single() argument
1300 int (*fill_super)(struct super_block *sb, in get_tree_keyed() argument
1364 struct super_block *sb = bdev->bd_holder; in bdev_super_lock() local
1368 lockdep_assert_not_held(&sb->s_umount); in bdev_super_lock()
1371 /* Make sure sb doesn't go away from under us */ in bdev_super_lock()
1373 sb->s_count++; in bdev_super_lock()
1378 locked = super_lock(sb, excl); in bdev_super_lock()
1384 put_super(sb); in bdev_super_lock()
1389 if (!sb->s_root || !(sb->s_flags & SB_ACTIVE)) { in bdev_super_lock()
1390 super_unlock(sb, excl); in bdev_super_lock()
1394 return sb; in bdev_super_lock()
1399 struct super_block *sb; in fs_bdev_mark_dead() local
1401 sb = bdev_super_lock(bdev, false); in fs_bdev_mark_dead()
1402 if (!sb) in fs_bdev_mark_dead()
1406 sync_filesystem(sb); in fs_bdev_mark_dead()
1407 shrink_dcache_sb(sb); in fs_bdev_mark_dead()
1408 invalidate_inodes(sb); in fs_bdev_mark_dead()
1409 if (sb->s_op->shutdown) in fs_bdev_mark_dead()
1410 sb->s_op->shutdown(sb); in fs_bdev_mark_dead()
1412 super_unlock_shared(sb); in fs_bdev_mark_dead()
1417 struct super_block *sb; in fs_bdev_sync() local
1419 sb = bdev_super_lock(bdev, false); in fs_bdev_sync()
1420 if (!sb) in fs_bdev_sync()
1423 sync_filesystem(sb); in fs_bdev_sync()
1424 super_unlock_shared(sb); in fs_bdev_sync()
1430 struct super_block *sb; in get_bdev_super() local
1432 sb = bdev_super_lock(bdev, true); in get_bdev_super()
1433 if (sb) { in get_bdev_super()
1434 active = atomic_inc_not_zero(&sb->s_active); in get_bdev_super()
1435 super_unlock_excl(sb); in get_bdev_super()
1439 return sb; in get_bdev_super()
1459 struct super_block *sb; in fs_bdev_freeze() local
1464 sb = get_bdev_super(bdev); in fs_bdev_freeze()
1465 if (!sb) in fs_bdev_freeze()
1468 if (sb->s_op->freeze_super) in fs_bdev_freeze()
1469 error = sb->s_op->freeze_super(sb, in fs_bdev_freeze()
1472 error = freeze_super(sb, in fs_bdev_freeze()
1476 deactivate_super(sb); in fs_bdev_freeze()
1499 struct super_block *sb; in fs_bdev_thaw() local
1504 sb = get_bdev_super(bdev); in fs_bdev_thaw()
1505 if (WARN_ON_ONCE(!sb)) in fs_bdev_thaw()
1508 if (sb->s_op->thaw_super) in fs_bdev_thaw()
1509 error = sb->s_op->thaw_super(sb, in fs_bdev_thaw()
1512 error = thaw_super(sb, in fs_bdev_thaw()
1514 deactivate_super(sb); in fs_bdev_thaw()
1526 int setup_bdev_super(struct super_block *sb, int sb_flags, in setup_bdev_super() argument
1533 bdev_handle = bdev_open_by_dev(sb->s_dev, mode, sb, &fs_holder_ops); in setup_bdev_super()
1562 sb->s_bdev_handle = bdev_handle; in setup_bdev_super()
1563 sb->s_bdev = bdev; in setup_bdev_super()
1564 sb->s_bdi = bdi_get(bdev->bd_disk->bdi); in setup_bdev_super()
1566 sb->s_iflags |= SB_I_STABLE_WRITES; in setup_bdev_super()
1569 snprintf(sb->s_id, sizeof(sb->s_id), "%pg", bdev); in setup_bdev_super()
1570 shrinker_debugfs_rename(sb->s_shrink, "sb-%s:%s", sb->s_type->name, in setup_bdev_super()
1571 sb->s_id); in setup_bdev_super()
1572 sb_set_blocksize(sb, block_size(bdev)); in setup_bdev_super()
1671 void kill_block_super(struct super_block *sb) in kill_block_super() argument
1673 struct block_device *bdev = sb->s_bdev; in kill_block_super()
1675 generic_shutdown_super(sb); in kill_block_super()
1678 bdev_release(sb->s_bdev_handle); in kill_block_super()
1770 struct super_block *sb; in vfs_get_tree() local
1792 sb = fc->root->d_sb; in vfs_get_tree()
1793 WARN_ON(!sb->s_bdi); in vfs_get_tree()
1802 super_wake(sb, SB_BORN); in vfs_get_tree()
1804 error = security_sb_set_mnt_opts(sb, fc->security, 0, NULL); in vfs_get_tree()
1816 WARN((sb->s_maxbytes < 0), "%s set sb->s_maxbytes to " in vfs_get_tree()
1817 "negative value (%lld)\n", fc->fs_type->name, sb->s_maxbytes); in vfs_get_tree()
1827 int super_setup_bdi_name(struct super_block *sb, char *fmt, ...) in super_setup_bdi_name() argument
1844 WARN_ON(sb->s_bdi != &noop_backing_dev_info); in super_setup_bdi_name()
1845 sb->s_bdi = bdi; in super_setup_bdi_name()
1846 sb->s_iflags |= SB_I_PERSB_BDI; in super_setup_bdi_name()
1856 int super_setup_bdi(struct super_block *sb) in super_setup_bdi() argument
1860 return super_setup_bdi_name(sb, "%.28s-%ld", sb->s_type->name, in super_setup_bdi()
1867 * @sb: the super for which we wait
1873 static void sb_wait_write(struct super_block *sb, int level) in sb_wait_write() argument
1875 percpu_down_write(sb->s_writers.rw_sem + level-1); in sb_wait_write()
1882 static void lockdep_sb_freeze_release(struct super_block *sb) in lockdep_sb_freeze_release() argument
1887 percpu_rwsem_release(sb->s_writers.rw_sem + level, 0, _THIS_IP_); in lockdep_sb_freeze_release()
1891 * Tell lockdep we are holding these locks before we call ->unfreeze_fs(sb).
1893 static void lockdep_sb_freeze_acquire(struct super_block *sb) in lockdep_sb_freeze_acquire() argument
1898 percpu_rwsem_acquire(sb->s_writers.rw_sem + level, 0, _THIS_IP_); in lockdep_sb_freeze_acquire()
1901 static void sb_freeze_unlock(struct super_block *sb, int level) in sb_freeze_unlock() argument
1904 percpu_up_write(sb->s_writers.rw_sem + level); in sb_freeze_unlock()
1907 static int wait_for_partially_frozen(struct super_block *sb) in wait_for_partially_frozen() argument
1912 unsigned short old = sb->s_writers.frozen; in wait_for_partially_frozen()
1914 up_write(&sb->s_umount); in wait_for_partially_frozen()
1915 ret = wait_var_event_killable(&sb->s_writers.frozen, in wait_for_partially_frozen()
1916 sb->s_writers.frozen != old); in wait_for_partially_frozen()
1917 down_write(&sb->s_umount); in wait_for_partially_frozen()
1919 sb->s_writers.frozen != SB_UNFROZEN && in wait_for_partially_frozen()
1920 sb->s_writers.frozen != SB_FREEZE_COMPLETE); in wait_for_partially_frozen()
1928 static inline int freeze_inc(struct super_block *sb, enum freeze_holder who) in freeze_inc() argument
1934 ++sb->s_writers.freeze_kcount; in freeze_inc()
1936 ++sb->s_writers.freeze_ucount; in freeze_inc()
1937 return sb->s_writers.freeze_kcount + sb->s_writers.freeze_ucount; in freeze_inc()
1940 static inline int freeze_dec(struct super_block *sb, enum freeze_holder who) in freeze_dec() argument
1945 if ((who & FREEZE_HOLDER_KERNEL) && sb->s_writers.freeze_kcount) in freeze_dec()
1946 --sb->s_writers.freeze_kcount; in freeze_dec()
1947 if ((who & FREEZE_HOLDER_USERSPACE) && sb->s_writers.freeze_ucount) in freeze_dec()
1948 --sb->s_writers.freeze_ucount; in freeze_dec()
1949 return sb->s_writers.freeze_kcount + sb->s_writers.freeze_ucount; in freeze_dec()
1952 static inline bool may_freeze(struct super_block *sb, enum freeze_holder who) in may_freeze() argument
1959 sb->s_writers.freeze_kcount == 0; in may_freeze()
1962 sb->s_writers.freeze_ucount == 0; in may_freeze()
1968 * @sb: the super to lock
1993 * During this function, sb->s_writers.frozen goes through these values:
2016 * sb->s_writers.frozen is protected by sb->s_umount.
2021 int freeze_super(struct super_block *sb, enum freeze_holder who) in freeze_super() argument
2025 if (!super_lock_excl(sb)) { in freeze_super()
2029 atomic_inc(&sb->s_active); in freeze_super()
2032 if (sb->s_writers.frozen == SB_FREEZE_COMPLETE) { in freeze_super()
2033 if (may_freeze(sb, who)) in freeze_super()
2034 ret = !!WARN_ON_ONCE(freeze_inc(sb, who) == 1); in freeze_super()
2038 deactivate_locked_super(sb); in freeze_super()
2042 if (sb->s_writers.frozen != SB_UNFROZEN) { in freeze_super()
2043 ret = wait_for_partially_frozen(sb); in freeze_super()
2045 deactivate_locked_super(sb); in freeze_super()
2052 if (sb_rdonly(sb)) { in freeze_super()
2054 WARN_ON_ONCE(freeze_inc(sb, who) > 1); in freeze_super()
2055 sb->s_writers.frozen = SB_FREEZE_COMPLETE; in freeze_super()
2056 wake_up_var(&sb->s_writers.frozen); in freeze_super()
2057 super_unlock_excl(sb); in freeze_super()
2061 sb->s_writers.frozen = SB_FREEZE_WRITE; in freeze_super()
2063 super_unlock_excl(sb); in freeze_super()
2064 sb_wait_write(sb, SB_FREEZE_WRITE); in freeze_super()
2065 __super_lock_excl(sb); in freeze_super()
2068 sb->s_writers.frozen = SB_FREEZE_PAGEFAULT; in freeze_super()
2069 sb_wait_write(sb, SB_FREEZE_PAGEFAULT); in freeze_super()
2072 ret = sync_filesystem(sb); in freeze_super()
2074 sb->s_writers.frozen = SB_UNFROZEN; in freeze_super()
2075 sb_freeze_unlock(sb, SB_FREEZE_PAGEFAULT); in freeze_super()
2076 wake_up_var(&sb->s_writers.frozen); in freeze_super()
2077 deactivate_locked_super(sb); in freeze_super()
2082 sb->s_writers.frozen = SB_FREEZE_FS; in freeze_super()
2083 sb_wait_write(sb, SB_FREEZE_FS); in freeze_super()
2085 if (sb->s_op->freeze_fs) { in freeze_super()
2086 ret = sb->s_op->freeze_fs(sb); in freeze_super()
2090 sb->s_writers.frozen = SB_UNFROZEN; in freeze_super()
2091 sb_freeze_unlock(sb, SB_FREEZE_FS); in freeze_super()
2092 wake_up_var(&sb->s_writers.frozen); in freeze_super()
2093 deactivate_locked_super(sb); in freeze_super()
2101 WARN_ON_ONCE(freeze_inc(sb, who) > 1); in freeze_super()
2102 sb->s_writers.frozen = SB_FREEZE_COMPLETE; in freeze_super()
2103 wake_up_var(&sb->s_writers.frozen); in freeze_super()
2104 lockdep_sb_freeze_release(sb); in freeze_super()
2105 super_unlock_excl(sb); in freeze_super()
2116 static int thaw_super_locked(struct super_block *sb, enum freeze_holder who) in thaw_super_locked() argument
2120 if (sb->s_writers.frozen != SB_FREEZE_COMPLETE) in thaw_super_locked()
2127 if (freeze_dec(sb, who)) in thaw_super_locked()
2130 if (sb_rdonly(sb)) { in thaw_super_locked()
2131 sb->s_writers.frozen = SB_UNFROZEN; in thaw_super_locked()
2132 wake_up_var(&sb->s_writers.frozen); in thaw_super_locked()
2136 lockdep_sb_freeze_acquire(sb); in thaw_super_locked()
2138 if (sb->s_op->unfreeze_fs) { in thaw_super_locked()
2139 error = sb->s_op->unfreeze_fs(sb); in thaw_super_locked()
2142 freeze_inc(sb, who); in thaw_super_locked()
2143 lockdep_sb_freeze_release(sb); in thaw_super_locked()
2148 sb->s_writers.frozen = SB_UNFROZEN; in thaw_super_locked()
2149 wake_up_var(&sb->s_writers.frozen); in thaw_super_locked()
2150 sb_freeze_unlock(sb, SB_FREEZE_FS); in thaw_super_locked()
2152 deactivate_locked_super(sb); in thaw_super_locked()
2156 super_unlock_excl(sb); in thaw_super_locked()
2162 * @sb: the super to thaw
2177 int thaw_super(struct super_block *sb, enum freeze_holder who) in thaw_super() argument
2179 if (!super_lock_excl(sb)) { in thaw_super()
2183 return thaw_super_locked(sb, who); in thaw_super()
2193 int sb_init_dio_done_wq(struct super_block *sb) in sb_init_dio_done_wq() argument
2198 sb->s_id); in sb_init_dio_done_wq()
2204 old = cmpxchg(&sb->s_dio_done_wq, NULL, wq); in sb_init_dio_done_wq()