[v6] f2fs: support FAULT_LOCK type fault injection
Commit Message
Similar to FAULT_LOCK_OP, this patch supports to inject fault into
variant of _trylock(), including:
- inode_trylock
- down_read_trylock
- down_write_trylock
- mutex_trylock
- sb_start_write_trylock
- trylock_page
Usage:
a) echo 524288 > /sys/fs/f2fs/<dev>/inject_type or
b) mount -o fault_type=524288 <dev> <mountpoint>
Reported-by: kernel test robot <lkp@intel.com>
Link: https://lore.kernel.org/oe-kbuild-all/202304031801.hGbHuuBP-lkp@intel.com/
Suggested-by: Chao Yu <chao@kernel.org>
Signed-off-by: Yangtao Li <frank.li@vivo.com>
---
v6:
-rebase
Documentation/ABI/testing/sysfs-fs-f2fs | 1 +
Documentation/filesystems/f2fs.rst | 1 +
fs/f2fs/checkpoint.c | 6 +-
fs/f2fs/compress.c | 2 +-
fs/f2fs/data.c | 8 +--
fs/f2fs/extent_cache.c | 6 +-
fs/f2fs/f2fs.h | 73 ++++++++++++++++++++++---
fs/f2fs/file.c | 31 ++++++-----
fs/f2fs/gc.c | 12 ++--
fs/f2fs/namei.c | 4 +-
fs/f2fs/node.c | 28 +++++-----
fs/f2fs/segment.c | 27 ++++-----
fs/f2fs/shrinker.c | 10 ++--
fs/f2fs/super.c | 7 ++-
fs/f2fs/sysfs.c | 2 +-
fs/f2fs/xattr.c | 15 ++---
16 files changed, 146 insertions(+), 87 deletions(-)
Comments
On 2023/4/20 14:22, Yangtao Li wrote:
> Similar to FAULT_LOCK_OP, this patch supports to inject fault into
> variant of _trylock(), including:
> - inode_trylock
> - down_read_trylock
> - down_write_trylock
> - mutex_trylock
> - sb_start_write_trylock
> - trylock_page
>
> Usage:
> a) echo 524288 > /sys/fs/f2fs/<dev>/inject_type or
> b) mount -o fault_type=524288 <dev> <mountpoint>
>
> Reported-by: kernel test robot <lkp@intel.com>
> Link: https://lore.kernel.org/oe-kbuild-all/202304031801.hGbHuuBP-lkp@intel.com/
> Suggested-by: Chao Yu <chao@kernel.org>
> Signed-off-by: Yangtao Li <frank.li@vivo.com>
> ---
> v6:
> -rebase
> Documentation/ABI/testing/sysfs-fs-f2fs | 1 +
> Documentation/filesystems/f2fs.rst | 1 +
> fs/f2fs/checkpoint.c | 6 +-
> fs/f2fs/compress.c | 2 +-
> fs/f2fs/data.c | 8 +--
> fs/f2fs/extent_cache.c | 6 +-
> fs/f2fs/f2fs.h | 73 ++++++++++++++++++++++---
> fs/f2fs/file.c | 31 ++++++-----
> fs/f2fs/gc.c | 12 ++--
> fs/f2fs/namei.c | 4 +-
> fs/f2fs/node.c | 28 +++++-----
> fs/f2fs/segment.c | 27 ++++-----
> fs/f2fs/shrinker.c | 10 ++--
> fs/f2fs/super.c | 7 ++-
> fs/f2fs/sysfs.c | 2 +-
> fs/f2fs/xattr.c | 15 ++---
> 16 files changed, 146 insertions(+), 87 deletions(-)
>
> diff --git a/Documentation/ABI/testing/sysfs-fs-f2fs b/Documentation/ABI/testing/sysfs-fs-f2fs
> index 8140fc98f5ae..5ada174cf30c 100644
> --- a/Documentation/ABI/testing/sysfs-fs-f2fs
> +++ b/Documentation/ABI/testing/sysfs-fs-f2fs
> @@ -708,6 +708,7 @@ Description: Support configuring fault injection type, should be
> FAULT_DQUOT_INIT 0x000010000
> FAULT_LOCK_OP 0x000020000
> FAULT_BLKADDR 0x000040000
> + FAULT_LOCK 0x000080000
> =================== ===========
>
> What: /sys/fs/f2fs/<disk>/discard_io_aware_gran
> diff --git a/Documentation/filesystems/f2fs.rst b/Documentation/filesystems/f2fs.rst
> index c57745375edb..fafae551df9e 100644
> --- a/Documentation/filesystems/f2fs.rst
> +++ b/Documentation/filesystems/f2fs.rst
> @@ -206,6 +206,7 @@ fault_type=%d Support configuring fault injection type, should be
> FAULT_DQUOT_INIT 0x000010000
> FAULT_LOCK_OP 0x000020000
> FAULT_BLKADDR 0x000040000
> + FAULT_LOCK 0x000080000
> =================== ===========
> mode=%s Control block allocation mode which supports "adaptive"
> and "lfs". In "lfs" mode, there should be no random
> diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c
> index 64b3860f50ee..2887d62b4a17 100644
> --- a/fs/f2fs/checkpoint.c
> +++ b/fs/f2fs/checkpoint.c
> @@ -389,7 +389,7 @@ static int f2fs_write_meta_pages(struct address_space *mapping,
> goto skip_write;
>
> /* if locked failed, cp will flush dirty pages instead */
> - if (!f2fs_down_write_trylock(&sbi->cp_global_sem))
> + if (!f2fs_down_write_trylock(sbi, &sbi->cp_global_sem))
> goto skip_write;
>
> trace_f2fs_writepages(mapping->host, wbc, META);
> @@ -1182,7 +1182,7 @@ static bool __need_flush_quota(struct f2fs_sb_info *sbi)
> if (!is_journalled_quota(sbi))
> return false;
>
> - if (!f2fs_down_write_trylock(&sbi->quota_sem))
> + if (!f2fs_down_write_trylock(sbi, &sbi->quota_sem))
> return true;
> if (is_sbi_flag_set(sbi, SBI_QUOTA_SKIP_FLUSH)) {
> ret = false;
> @@ -1228,7 +1228,7 @@ static int block_operations(struct f2fs_sb_info *sbi)
> f2fs_unlock_all(sbi);
>
> /* only failed during mount/umount/freeze/quotactl */
> - locked = down_read_trylock(&sbi->sb->s_umount);
> + locked = f2fs_raw_down_read_trylock(sbi, &sbi->sb->s_umount);
> f2fs_quota_sync(sbi->sb, -1);
> if (locked)
> up_read(&sbi->sb->s_umount);
> diff --git a/fs/f2fs/compress.c b/fs/f2fs/compress.c
> index 11653fa79289..b94c3d5087f3 100644
> --- a/fs/f2fs/compress.c
> +++ b/fs/f2fs/compress.c
> @@ -1228,7 +1228,7 @@ static int f2fs_write_compressed_pages(struct compress_ctx *cc,
> * checkpoint. This can only happen to quota writes which can cause
> * the below discard race condition.
> */
> - f2fs_down_read(&sbi->node_write);
> + f2fs_down_read(sbi, &sbi->node_write);
> } else if (!f2fs_trylock_op(sbi)) {
> goto out_free;
> }
> diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c
> index 8df225014304..872ed5554c26 100644
> --- a/fs/f2fs/data.c
> +++ b/fs/f2fs/data.c
> @@ -680,7 +680,7 @@ static void __submit_merged_write_cond(struct f2fs_sb_info *sbi,
> enum page_type btype = PAGE_TYPE_OF_BIO(type);
> struct f2fs_bio_info *io = sbi->write_io[btype] + temp;
>
> - f2fs_down_read(&io->io_rwsem);
> + f2fs_down_read(sbi, &io->io_rwsem);
> ret = __has_merged_page(io->bio, inode, page, ino);
> f2fs_up_read(&io->io_rwsem);
> }
> @@ -884,7 +884,7 @@ void f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi,
> if (list_empty(head))
> continue;
>
> - f2fs_down_read(&io->bio_list_lock);
> + f2fs_down_read(sbi, &io->bio_list_lock);
> list_for_each_entry(be, head, list) {
> if (target)
> found = (target == be->bio);
> @@ -1447,7 +1447,7 @@ static int __allocate_data_block(struct dnode_of_data *dn, int seg_type)
> static void f2fs_map_lock(struct f2fs_sb_info *sbi, int flag)
> {
> if (flag == F2FS_GET_BLOCK_PRE_AIO)
> - f2fs_down_read(&sbi->node_change);
> + f2fs_down_read(sbi, &sbi->node_change);
> else
> f2fs_lock_op(sbi);
> }
> @@ -2839,7 +2839,7 @@ int f2fs_write_single_data_page(struct page *page, int *submitted,
> * the below discard race condition.
> */
> if (IS_NOQUOTA(inode))
> - f2fs_down_read(&sbi->node_write);
> + f2fs_down_read(sbi, &sbi->node_write);
>
> fio.need_lock = LOCK_DONE;
> err = f2fs_do_write_data_page(&fio);
> diff --git a/fs/f2fs/extent_cache.c b/fs/f2fs/extent_cache.c
> index bea6ab9d846a..bab36a42bacf 100644
> --- a/fs/f2fs/extent_cache.c
> +++ b/fs/f2fs/extent_cache.c
> @@ -917,7 +917,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
> if (!atomic_read(&eti->total_zombie_tree))
> goto free_node;
>
> - if (!mutex_trylock(&eti->extent_tree_lock))
> + if (!f2fs_mutex_trylock(sbi, &eti->extent_tree_lock))
> goto out;
>
> /* 1. remove unreferenced extent tree */
> @@ -943,7 +943,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
>
> free_node:
> /* 2. remove LRU extent entries */
> - if (!mutex_trylock(&eti->extent_tree_lock))
> + if (!f2fs_mutex_trylock(sbi, &eti->extent_tree_lock))
> goto out;
>
> remained = nr_shrink - (node_cnt + tree_cnt);
> @@ -955,7 +955,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
> en = list_first_entry(&eti->extent_list,
> struct extent_node, list);
> et = en->et;
> - if (!write_trylock(&et->lock)) {
> + if (!f2fs_write_trylock(sbi, &et->lock)) {
> /* refresh this extent node's position in extent list */
> list_move_tail(&en->list, &eti->extent_list);
> continue;
> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> index d8cb1dc09f9f..2bbac3a15c5a 100644
> --- a/fs/f2fs/f2fs.h
> +++ b/fs/f2fs/f2fs.h
> @@ -61,6 +61,7 @@ enum {
> FAULT_DQUOT_INIT,
> FAULT_LOCK_OP,
> FAULT_BLKADDR,
> + FAULT_LOCK,
> FAULT_MAX,
> };
>
> @@ -2079,27 +2080,76 @@ static inline int f2fs_rwsem_is_contended(struct f2fs_rwsem *sem)
> return rwsem_is_contended(&sem->internal_rwsem);
> }
>
> -static inline void f2fs_down_read(struct f2fs_rwsem *sem)
> +static inline int f2fs_down_read_trylock(struct f2fs_sb_info *sbi,
> + struct f2fs_rwsem *sem)
> +{
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return down_read_trylock(&sem->internal_rwsem);
> +}
> +
> +static inline void f2fs_down_read(struct f2fs_sb_info *sbi, struct f2fs_rwsem *sem)
> {
> #ifdef CONFIG_F2FS_UNFAIR_RWSEM
> - wait_event(sem->read_waiters, down_read_trylock(&sem->internal_rwsem));
> + wait_event(sem->read_waiters, f2fs_down_read_trylock(sbi, sem));
It may hang if we wait on sem->read_waiters w/ TASK_UNINTERRUPTIBLE once we
inject FAULT_LOCK to f2fs_down_read_trylock(), it is due to FAULT_LOCK simulates
there is a write lock user, but if there isn't any write lock user, nobody will
call f2fs_up_write() to wake up the waiter.
Thanks,
> #else
> down_read(&sem->internal_rwsem);
> #endif
> }
>
> -static inline int f2fs_down_read_trylock(struct f2fs_rwsem *sem)
> +static inline int f2fs_raw_down_read_trylock(struct f2fs_sb_info *sbi,
> + struct rw_semaphore *sem)
> {
> - return down_read_trylock(&sem->internal_rwsem);
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return down_read_trylock(sem);
> +}
> +
> +static inline int f2fs_inode_trylock(struct f2fs_sb_info *sbi,
> + struct inode *inode)
> +{
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return down_write_trylock(&inode->i_rwsem);
> +}
> +
> +static inline int f2fs_mutex_trylock(struct f2fs_sb_info *sbi,
> + struct mutex *lock)
> +{
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return mutex_trylock(lock);
> +}
> +
> +static inline int f2fs_trylock_page(struct f2fs_sb_info *sbi, struct page *page)
> +{
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return trylock_page(page);
> +}
> +
> +static inline int f2fs_write_trylock(struct f2fs_sb_info *sbi, rwlock_t *lock)
> +{
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return write_trylock(lock);
> +}
> +
> +static inline bool f2fs_sb_start_write_trylock(struct f2fs_sb_info *sbi)
> +{
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> + return sb_start_write_trylock(sbi->sb);
> }
>
> #ifdef CONFIG_DEBUG_LOCK_ALLOC
> -static inline void f2fs_down_read_nested(struct f2fs_rwsem *sem, int subclass)
> +static inline void f2fs_down_read_nested(struct f2fs_sb_info *sbi,
> + struct f2fs_rwsem *sem, int subclass)
> {
> down_read_nested(&sem->internal_rwsem, subclass);
> }
> #else
> -#define f2fs_down_read_nested(sem, subclass) f2fs_down_read(sem)
> +#define f2fs_down_read_nested(sbi, sem, subclass) f2fs_down_read(sbi, sem)
> #endif
>
> static inline void f2fs_up_read(struct f2fs_rwsem *sem)
> @@ -2112,8 +2162,11 @@ static inline void f2fs_down_write(struct f2fs_rwsem *sem)
> down_write(&sem->internal_rwsem);
> }
>
> -static inline int f2fs_down_write_trylock(struct f2fs_rwsem *sem)
> +static inline int f2fs_down_write_trylock(struct f2fs_sb_info *sbi,
> + struct f2fs_rwsem *sem)
> {
> + if (time_to_inject(sbi, FAULT_LOCK))
> + return 0;
> return down_write_trylock(&sem->internal_rwsem);
> }
>
> @@ -2127,14 +2180,16 @@ static inline void f2fs_up_write(struct f2fs_rwsem *sem)
>
> static inline void f2fs_lock_op(struct f2fs_sb_info *sbi)
> {
> - f2fs_down_read(&sbi->cp_rwsem);
> + f2fs_down_read(sbi, &sbi->cp_rwsem);
> }
>
> static inline int f2fs_trylock_op(struct f2fs_sb_info *sbi)
> {
> + struct f2fs_rwsem *sem = &sbi->cp_rwsem;
> +
> if (time_to_inject(sbi, FAULT_LOCK_OP))
> return 0;
> - return f2fs_down_read_trylock(&sbi->cp_rwsem);
> + return down_read_trylock(&sem->internal_rwsem);
> }
>
> static inline void f2fs_unlock_op(struct f2fs_sb_info *sbi)
> diff --git a/fs/f2fs/file.c b/fs/f2fs/file.c
> index 720be239955d..ad4a3d79ddf5 100644
> --- a/fs/f2fs/file.c
> +++ b/fs/f2fs/file.c
> @@ -316,7 +316,7 @@ static int f2fs_do_sync_file(struct file *file, loff_t start, loff_t end,
> * Both of fdatasync() and fsync() are able to be recovered from
> * sudden-power-off.
> */
> - f2fs_down_read(&F2FS_I(inode)->i_sem);
> + f2fs_down_read(sbi, &F2FS_I(inode)->i_sem);
> cp_reason = need_do_checkpoint(inode);
> f2fs_up_read(&F2FS_I(inode)->i_sem);
>
> @@ -2466,7 +2466,7 @@ static int f2fs_ioc_gc(struct file *filp, unsigned long arg)
> return ret;
>
> if (!sync) {
> - if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
> + if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
> ret = -EBUSY;
> goto out;
> }
> @@ -2510,7 +2510,7 @@ static int __f2fs_ioc_gc_range(struct file *filp, struct f2fs_gc_range *range)
>
> do_more:
> if (!range->sync) {
> - if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
> + if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
> ret = -EBUSY;
> goto out;
> }
> @@ -2805,7 +2805,7 @@ static int f2fs_move_file_range(struct file *file_in, loff_t pos_in,
> inode_lock(src);
> if (src != dst) {
> ret = -EBUSY;
> - if (!inode_trylock(dst))
> + if (!f2fs_inode_trylock(sbi, dst))
> goto out;
> }
>
> @@ -2855,7 +2855,7 @@ static int f2fs_move_file_range(struct file *file_in, loff_t pos_in,
> f2fs_down_write(&F2FS_I(src)->i_gc_rwsem[WRITE]);
> if (src != dst) {
> ret = -EBUSY;
> - if (!f2fs_down_write_trylock(&F2FS_I(dst)->i_gc_rwsem[WRITE]))
> + if (!f2fs_down_write_trylock(sbi, &F2FS_I(dst)->i_gc_rwsem[WRITE]))
> goto out_src;
> }
>
> @@ -2975,7 +2975,7 @@ static int f2fs_ioc_flush_device(struct file *filp, unsigned long arg)
> end_segno = min(start_segno + range.segments, dev_end_segno);
>
> while (start_segno < end_segno) {
> - if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
> + if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
> ret = -EBUSY;
> goto out;
> }
> @@ -3325,7 +3325,7 @@ static int f2fs_ioc_getfslabel(struct file *filp, unsigned long arg)
> if (!vbuf)
> return -ENOMEM;
>
> - f2fs_down_read(&sbi->sb_lock);
> + f2fs_down_read(sbi, &sbi->sb_lock);
> count = utf16s_to_utf8s(sbi->raw_super->volume_name,
> ARRAY_SIZE(sbi->raw_super->volume_name),
> UTF16_LITTLE_ENDIAN, vbuf, MAX_VOLUME_NAME);
> @@ -4335,12 +4335,12 @@ static ssize_t f2fs_dio_read_iter(struct kiocb *iocb, struct iov_iter *to)
> trace_f2fs_direct_IO_enter(inode, iocb, count, READ);
>
> if (iocb->ki_flags & IOCB_NOWAIT) {
> - if (!f2fs_down_read_trylock(&fi->i_gc_rwsem[READ])) {
> + if (!f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[READ])) {
> ret = -EAGAIN;
> goto out;
> }
> } else {
> - f2fs_down_read(&fi->i_gc_rwsem[READ]);
> + f2fs_down_read(sbi, &fi->i_gc_rwsem[READ]);
> }
>
> /*
> @@ -4582,11 +4582,11 @@ static ssize_t f2fs_dio_write_iter(struct kiocb *iocb, struct iov_iter *from,
> goto out;
> }
>
> - if (!f2fs_down_read_trylock(&fi->i_gc_rwsem[WRITE])) {
> + if (!f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[WRITE])) {
> ret = -EAGAIN;
> goto out;
> }
> - if (do_opu && !f2fs_down_read_trylock(&fi->i_gc_rwsem[READ])) {
> + if (do_opu && !f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[READ])) {
> f2fs_up_read(&fi->i_gc_rwsem[WRITE]);
> ret = -EAGAIN;
> goto out;
> @@ -4596,9 +4596,9 @@ static ssize_t f2fs_dio_write_iter(struct kiocb *iocb, struct iov_iter *from,
> if (ret)
> goto out;
>
> - f2fs_down_read(&fi->i_gc_rwsem[WRITE]);
> + f2fs_down_read(sbi, &fi->i_gc_rwsem[WRITE]);
> if (do_opu)
> - f2fs_down_read(&fi->i_gc_rwsem[READ]);
> + f2fs_down_read(sbi, &fi->i_gc_rwsem[READ]);
> }
>
> /*
> @@ -4675,13 +4675,14 @@ static ssize_t f2fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
> struct inode *inode = file_inode(iocb->ki_filp);
> const loff_t orig_pos = iocb->ki_pos;
> const size_t orig_count = iov_iter_count(from);
> + struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
> loff_t target_size;
> bool dio;
> bool may_need_sync = true;
> int preallocated;
> ssize_t ret;
>
> - if (unlikely(f2fs_cp_error(F2FS_I_SB(inode)))) {
> + if (unlikely(f2fs_cp_error(sbi))) {
> ret = -EIO;
> goto out;
> }
> @@ -4692,7 +4693,7 @@ static ssize_t f2fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
> }
>
> if (iocb->ki_flags & IOCB_NOWAIT) {
> - if (!inode_trylock(inode)) {
> + if (!f2fs_inode_trylock(sbi, inode)) {
> ret = -EAGAIN;
> goto out;
> }
> diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c
> index 61c5f9d26018..3a7947f7895a 100644
> --- a/fs/f2fs/gc.c
> +++ b/fs/f2fs/gc.c
> @@ -76,7 +76,7 @@ static int gc_thread_func(void *data)
> f2fs_stop_checkpoint(sbi, false,
> STOP_CP_REASON_FAULT_INJECT);
>
> - if (!sb_start_write_trylock(sbi->sb)) {
> + if (!f2fs_sb_start_write_trylock(sbi)) {
> stat_other_skip_bggc_count(sbi);
> continue;
> }
> @@ -104,7 +104,7 @@ static int gc_thread_func(void *data)
> if (foreground) {
> f2fs_down_write(&sbi->gc_lock);
> goto do_gc;
> - } else if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
> + } else if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
> stat_other_skip_bggc_count(sbi);
> goto next;
> }
> @@ -1573,7 +1573,7 @@ static int gc_data_segment(struct f2fs_sb_info *sbi, struct f2fs_summary *sum,
> return submitted;
> }
>
> - if (!f2fs_down_write_trylock(
> + if (!f2fs_down_write_trylock(sbi,
> &F2FS_I(inode)->i_gc_rwsem[WRITE])) {
> iput(inode);
> sbi->skipped_gc_rwsem++;
> @@ -1616,11 +1616,11 @@ static int gc_data_segment(struct f2fs_sb_info *sbi, struct f2fs_summary *sum,
> int err;
>
> if (S_ISREG(inode->i_mode)) {
> - if (!f2fs_down_write_trylock(&fi->i_gc_rwsem[WRITE])) {
> + if (!f2fs_down_write_trylock(sbi, &fi->i_gc_rwsem[WRITE])) {
> sbi->skipped_gc_rwsem++;
> continue;
> }
> - if (!f2fs_down_write_trylock(
> + if (!f2fs_down_write_trylock(sbi,
> &fi->i_gc_rwsem[READ])) {
> sbi->skipped_gc_rwsem++;
> f2fs_up_write(&fi->i_gc_rwsem[WRITE]);
> @@ -2142,7 +2142,7 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
> secs = div_u64(shrunk_blocks, BLKS_PER_SEC(sbi));
>
> /* stop other GC */
> - if (!f2fs_down_write_trylock(&sbi->gc_lock))
> + if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock))
> return -EAGAIN;
>
> /* stop CP to protect MAIN_SEC in free_segment_range */
> diff --git a/fs/f2fs/namei.c b/fs/f2fs/namei.c
> index 77a71276ecb1..4d4426cda93a 100644
> --- a/fs/f2fs/namei.c
> +++ b/fs/f2fs/namei.c
> @@ -144,7 +144,7 @@ static void set_compress_new_inode(struct f2fs_sb_info *sbi, struct inode *dir,
> return;
>
> /* Don't compress hot files. */
> - f2fs_down_read(&sbi->sb_lock);
> + f2fs_down_read(sbi, &sbi->sb_lock);
> cold_count = le32_to_cpu(sbi->raw_super->extension_count);
> hot_count = sbi->raw_super->hot_ext_count;
> for (i = cold_count; i < cold_count + hot_count; i++)
> @@ -185,7 +185,7 @@ static void set_file_temperature(struct f2fs_sb_info *sbi, struct inode *inode,
> __u8 (*extlist)[F2FS_EXTENSION_LEN] = sbi->raw_super->extension_list;
> int i, cold_count, hot_count;
>
> - f2fs_down_read(&sbi->sb_lock);
> + f2fs_down_read(sbi, &sbi->sb_lock);
> cold_count = le32_to_cpu(sbi->raw_super->extension_count);
> hot_count = sbi->raw_super->hot_ext_count;
> for (i = 0; i < cold_count + hot_count; i++)
> diff --git a/fs/f2fs/node.c b/fs/f2fs/node.c
> index bd1dad523796..bd108b0feb45 100644
> --- a/fs/f2fs/node.c
> +++ b/fs/f2fs/node.c
> @@ -383,7 +383,7 @@ int f2fs_need_dentry_mark(struct f2fs_sb_info *sbi, nid_t nid)
> struct nat_entry *e;
> bool need = false;
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
> e = __lookup_nat_cache(nm_i, nid);
> if (e) {
> if (!get_nat_flag(e, IS_CHECKPOINTED) &&
> @@ -400,7 +400,7 @@ bool f2fs_is_checkpointed_node(struct f2fs_sb_info *sbi, nid_t nid)
> struct nat_entry *e;
> bool is_cp = true;
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
> e = __lookup_nat_cache(nm_i, nid);
> if (e && !get_nat_flag(e, IS_CHECKPOINTED))
> is_cp = false;
> @@ -414,7 +414,7 @@ bool f2fs_need_inode_block_update(struct f2fs_sb_info *sbi, nid_t ino)
> struct nat_entry *e;
> bool need_update = true;
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
> e = __lookup_nat_cache(nm_i, ino);
> if (e && get_nat_flag(e, HAS_LAST_FSYNC) &&
> (get_nat_flag(e, IS_CHECKPOINTED) ||
> @@ -517,7 +517,7 @@ int f2fs_try_to_free_nats(struct f2fs_sb_info *sbi, int nr_shrink)
> struct f2fs_nm_info *nm_i = NM_I(sbi);
> int nr = nr_shrink;
>
> - if (!f2fs_down_write_trylock(&nm_i->nat_tree_lock))
> + if (!f2fs_down_write_trylock(sbi, &nm_i->nat_tree_lock))
> return 0;
>
> spin_lock(&nm_i->nat_list_lock);
> @@ -561,7 +561,7 @@ int f2fs_get_node_info(struct f2fs_sb_info *sbi, nid_t nid,
> ni->nid = nid;
> retry:
> /* Check nat cache */
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
> e = __lookup_nat_cache(nm_i, nid);
> if (e) {
> ni->ino = nat_get_ino(e);
> @@ -580,7 +580,7 @@ int f2fs_get_node_info(struct f2fs_sb_info *sbi, nid_t nid,
> if (!f2fs_rwsem_is_locked(&sbi->cp_global_sem) || checkpoint_context) {
> down_read(&curseg->journal_rwsem);
> } else if (f2fs_rwsem_is_contended(&nm_i->nat_tree_lock) ||
> - !down_read_trylock(&curseg->journal_rwsem)) {
> + !f2fs_raw_down_read_trylock(sbi, &curseg->journal_rwsem)) {
> f2fs_up_read(&nm_i->nat_tree_lock);
> goto retry;
> }
> @@ -1618,10 +1618,10 @@ static int __write_node_page(struct page *page, bool atomic, bool *submitted,
> goto redirty_out;
>
> if (wbc->for_reclaim) {
> - if (!f2fs_down_read_trylock(&sbi->node_write))
> + if (!f2fs_down_read_trylock(sbi, &sbi->node_write))
> goto redirty_out;
> } else {
> - f2fs_down_read(&sbi->node_write);
> + f2fs_down_read(sbi, &sbi->node_write);
> }
>
> /* This page is already truncated */
> @@ -1982,7 +1982,7 @@ int f2fs_sync_node_pages(struct f2fs_sb_info *sbi,
> lock_node:
> if (wbc->sync_mode == WB_SYNC_ALL)
> lock_page(page);
> - else if (!trylock_page(page))
> + else if (!f2fs_trylock_page(sbi, page))
> continue;
>
> if (unlikely(page->mapping != NODE_MAPPING(sbi))) {
> @@ -2233,7 +2233,7 @@ bool f2fs_nat_bitmap_enabled(struct f2fs_sb_info *sbi)
> unsigned int i;
> bool ret = true;
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
> for (i = 0; i < nm_i->nat_blocks; i++) {
> if (!test_bit_le(i, nm_i->nat_block_bitmap)) {
> ret = false;
> @@ -2423,7 +2423,7 @@ static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
> unsigned int i, idx;
> nid_t nid;
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
>
> for (i = 0; i < nm_i->nat_blocks; i++) {
> if (!test_bit_le(i, nm_i->nat_block_bitmap))
> @@ -2481,7 +2481,7 @@ static int __f2fs_build_free_nids(struct f2fs_sb_info *sbi,
> f2fs_ra_meta_pages(sbi, NAT_BLOCK_OFFSET(nid), FREE_NID_PAGES,
> META_NAT, true);
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
>
> while (1) {
> if (!test_bit_le(NAT_BLOCK_OFFSET(nid),
> @@ -2636,7 +2636,7 @@ int f2fs_try_to_free_nids(struct f2fs_sb_info *sbi, int nr_shrink)
> if (nm_i->nid_cnt[FREE_NID] <= MAX_FREE_NIDS)
> return 0;
>
> - if (!mutex_trylock(&nm_i->build_lock))
> + if (!f2fs_mutex_trylock(sbi, &nm_i->build_lock))
> return 0;
>
> while (nr_shrink && nm_i->nid_cnt[FREE_NID] > MAX_FREE_NIDS) {
> @@ -2959,7 +2959,7 @@ void f2fs_enable_nat_bits(struct f2fs_sb_info *sbi)
> struct f2fs_nm_info *nm_i = NM_I(sbi);
> unsigned int nat_ofs;
>
> - f2fs_down_read(&nm_i->nat_tree_lock);
> + f2fs_down_read(sbi, &nm_i->nat_tree_lock);
>
> for (nat_ofs = 0; nat_ofs < nm_i->nat_blocks; nat_ofs++) {
> unsigned int valid = 0, nid_ofs = 0;
> diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c
> index 641bd8cc54da..77bee39498cd 100644
> --- a/fs/f2fs/segment.c
> +++ b/fs/f2fs/segment.c
> @@ -2807,7 +2807,7 @@ static void __f2fs_init_atgc_curseg(struct f2fs_sb_info *sbi)
> if (!sbi->am.atgc_enabled)
> return;
>
> - f2fs_down_read(&SM_I(sbi)->curseg_lock);
> + f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
>
> mutex_lock(&curseg->curseg_mutex);
> down_write(&SIT_I(sbi)->sentry_lock);
> @@ -2956,7 +2956,7 @@ void f2fs_allocate_segment_for_resize(struct f2fs_sb_info *sbi, int type,
> struct curseg_info *curseg = CURSEG_I(sbi, type);
> unsigned int segno;
>
> - f2fs_down_read(&SM_I(sbi)->curseg_lock);
> + f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
> mutex_lock(&curseg->curseg_mutex);
> down_write(&SIT_I(sbi)->sentry_lock);
>
> @@ -3003,7 +3003,7 @@ static void __allocate_new_segment(struct f2fs_sb_info *sbi, int type,
>
> void f2fs_allocate_new_section(struct f2fs_sb_info *sbi, int type, bool force)
> {
> - f2fs_down_read(&SM_I(sbi)->curseg_lock);
> + f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
> down_write(&SIT_I(sbi)->sentry_lock);
> __allocate_new_segment(sbi, type, true, force);
> up_write(&SIT_I(sbi)->sentry_lock);
> @@ -3014,7 +3014,7 @@ void f2fs_allocate_new_segments(struct f2fs_sb_info *sbi)
> {
> int i;
>
> - f2fs_down_read(&SM_I(sbi)->curseg_lock);
> + f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
> down_write(&SIT_I(sbi)->sentry_lock);
> for (i = CURSEG_HOT_DATA; i <= CURSEG_COLD_DATA; i++)
> __allocate_new_segment(sbi, i, false, false);
> @@ -3322,7 +3322,7 @@ void f2fs_allocate_data_block(struct f2fs_sb_info *sbi, struct page *page,
> struct seg_entry *se = NULL;
> bool segment_full = false;
>
> - f2fs_down_read(&SM_I(sbi)->curseg_lock);
> + f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
>
> mutex_lock(&curseg->curseg_mutex);
> down_write(&sit_i->sentry_lock);
> @@ -3451,17 +3451,18 @@ void f2fs_update_device_state(struct f2fs_sb_info *sbi, nid_t ino,
> static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
> {
> int type = __get_segment_type(fio);
> - bool keep_order = (f2fs_lfs_mode(fio->sbi) && type == CURSEG_COLD_DATA);
> + struct f2fs_sb_info *sbi = fio->sbi;
> + bool keep_order = (f2fs_lfs_mode(sbi) && type == CURSEG_COLD_DATA);
>
> if (keep_order)
> - f2fs_down_read(&fio->sbi->io_order_lock);
> + f2fs_down_read(sbi, &sbi->io_order_lock);
> reallocate:
> - f2fs_allocate_data_block(fio->sbi, fio->page, fio->old_blkaddr,
> + f2fs_allocate_data_block(sbi, fio->page, fio->old_blkaddr,
> &fio->new_blkaddr, sum, type, fio);
> - if (GET_SEGNO(fio->sbi, fio->old_blkaddr) != NULL_SEGNO) {
> - invalidate_mapping_pages(META_MAPPING(fio->sbi),
> + if (GET_SEGNO(sbi, fio->old_blkaddr) != NULL_SEGNO) {
> + invalidate_mapping_pages(META_MAPPING(sbi),
> fio->old_blkaddr, fio->old_blkaddr);
> - f2fs_invalidate_compress_page(fio->sbi, fio->old_blkaddr);
> + f2fs_invalidate_compress_page(sbi, fio->old_blkaddr);
> }
>
> /* writeout dirty page into bdev */
> @@ -3471,10 +3472,10 @@ static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
> goto reallocate;
> }
>
> - f2fs_update_device_state(fio->sbi, fio->ino, fio->new_blkaddr, 1);
> + f2fs_update_device_state(sbi, fio->ino, fio->new_blkaddr, 1);
>
> if (keep_order)
> - f2fs_up_read(&fio->sbi->io_order_lock);
> + f2fs_up_read(&sbi->io_order_lock);
> }
>
> void f2fs_do_write_meta_page(struct f2fs_sb_info *sbi, struct page *page,
> diff --git a/fs/f2fs/shrinker.c b/fs/f2fs/shrinker.c
> index 83d6fb97dcae..675758b509ee 100644
> --- a/fs/f2fs/shrinker.c
> +++ b/fs/f2fs/shrinker.c
> @@ -40,17 +40,16 @@ static unsigned long __count_extent_cache(struct f2fs_sb_info *sbi,
> unsigned long f2fs_shrink_count(struct shrinker *shrink,
> struct shrink_control *sc)
> {
> - struct f2fs_sb_info *sbi;
> struct list_head *p;
> unsigned long count = 0;
>
> spin_lock(&f2fs_list_lock);
> p = f2fs_list.next;
> while (p != &f2fs_list) {
> - sbi = list_entry(p, struct f2fs_sb_info, s_list);
> + struct f2fs_sb_info *sbi = list_entry(p, struct f2fs_sb_info, s_list);
>
> /* stop f2fs_put_super */
> - if (!mutex_trylock(&sbi->umount_mutex)) {
> + if (!f2fs_mutex_trylock(sbi, &sbi->umount_mutex)) {
> p = p->next;
> continue;
> }
> @@ -80,7 +79,6 @@ unsigned long f2fs_shrink_scan(struct shrinker *shrink,
> struct shrink_control *sc)
> {
> unsigned long nr = sc->nr_to_scan;
> - struct f2fs_sb_info *sbi;
> struct list_head *p;
> unsigned int run_no;
> unsigned long freed = 0;
> @@ -91,13 +89,13 @@ unsigned long f2fs_shrink_scan(struct shrinker *shrink,
> } while (run_no == 0);
> p = f2fs_list.next;
> while (p != &f2fs_list) {
> - sbi = list_entry(p, struct f2fs_sb_info, s_list);
> + struct f2fs_sb_info *sbi = list_entry(p, struct f2fs_sb_info, s_list);
>
> if (sbi->shrinker_run_no == run_no)
> break;
>
> /* stop f2fs_put_super */
> - if (!mutex_trylock(&sbi->umount_mutex)) {
> + if (!f2fs_mutex_trylock(sbi, &sbi->umount_mutex)) {
> p = p->next;
> continue;
> }
> diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c
> index 9f15b03037db..4aa4b12e455d 100644
> --- a/fs/f2fs/super.c
> +++ b/fs/f2fs/super.c
> @@ -62,6 +62,7 @@ const char *f2fs_fault_name[FAULT_MAX] = {
> [FAULT_DQUOT_INIT] = "dquot initialize",
> [FAULT_LOCK_OP] = "lock_op",
> [FAULT_BLKADDR] = "invalid blkaddr",
> + [FAULT_LOCK] = "lock",
> };
>
> void f2fs_build_fault_attr(struct f2fs_sb_info *sbi, unsigned int rate,
> @@ -2831,7 +2832,7 @@ int f2fs_quota_sync(struct super_block *sb, int type)
> * f2fs_down_read(quota_sem)
> */
> f2fs_lock_op(sbi);
> - f2fs_down_read(&sbi->quota_sem);
> + f2fs_down_read(sbi, &sbi->quota_sem);
>
> ret = f2fs_quota_sync_file(sbi, cnt);
>
> @@ -2961,7 +2962,7 @@ static int f2fs_dquot_commit(struct dquot *dquot)
> struct f2fs_sb_info *sbi = F2FS_SB(dquot->dq_sb);
> int ret;
>
> - f2fs_down_read_nested(&sbi->quota_sem, SINGLE_DEPTH_NESTING);
> + f2fs_down_read_nested(sbi, &sbi->quota_sem, SINGLE_DEPTH_NESTING);
> ret = dquot_commit(dquot);
> if (ret < 0)
> set_sbi_flag(sbi, SBI_QUOTA_NEED_REPAIR);
> @@ -2974,7 +2975,7 @@ static int f2fs_dquot_acquire(struct dquot *dquot)
> struct f2fs_sb_info *sbi = F2FS_SB(dquot->dq_sb);
> int ret;
>
> - f2fs_down_read(&sbi->quota_sem);
> + f2fs_down_read(sbi, &sbi->quota_sem);
> ret = dquot_acquire(dquot);
> if (ret < 0)
> set_sbi_flag(sbi, SBI_QUOTA_NEED_REPAIR);
> diff --git a/fs/f2fs/sysfs.c b/fs/f2fs/sysfs.c
> index 8ea05340bad9..dccc0a122c25 100644
> --- a/fs/f2fs/sysfs.c
> +++ b/fs/f2fs/sysfs.c
> @@ -738,7 +738,7 @@ static ssize_t f2fs_sbi_store(struct f2fs_attr *a,
> a->struct_type == GC_THREAD);
>
> if (gc_entry) {
> - if (!down_read_trylock(&sbi->sb->s_umount))
> + if (!f2fs_raw_down_read_trylock(sbi, &sbi->sb->s_umount))
> return -EAGAIN;
> }
> ret = __sbi_store(a, sbi, buf, count);
> diff --git a/fs/f2fs/xattr.c b/fs/f2fs/xattr.c
> index d92edbbdc30e..bb00c7e889da 100644
> --- a/fs/f2fs/xattr.c
> +++ b/fs/f2fs/xattr.c
> @@ -513,6 +513,7 @@ static inline int write_all_xattrs(struct inode *inode, __u32 hsize,
> int f2fs_getxattr(struct inode *inode, int index, const char *name,
> void *buffer, size_t buffer_size, struct page *ipage)
> {
> + struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
> struct f2fs_xattr_entry *entry = NULL;
> int error;
> unsigned int size, len;
> @@ -527,7 +528,7 @@ int f2fs_getxattr(struct inode *inode, int index, const char *name,
> if (len > F2FS_NAME_LEN)
> return -ERANGE;
>
> - f2fs_down_read(&F2FS_I(inode)->i_xattr_sem);
> + f2fs_down_read(sbi, &F2FS_I(inode)->i_xattr_sem);
> error = lookup_all_xattrs(inode, ipage, index, len, name,
> &entry, &base_addr, &base_size, &is_inline);
> f2fs_up_read(&F2FS_I(inode)->i_xattr_sem);
> @@ -552,19 +553,20 @@ int f2fs_getxattr(struct inode *inode, int index, const char *name,
> }
> error = size;
> out:
> - xattr_free(F2FS_I_SB(inode), base_addr, is_inline);
> + xattr_free(sbi, base_addr, is_inline);
> return error;
> }
>
> ssize_t f2fs_listxattr(struct dentry *dentry, char *buffer, size_t buffer_size)
> {
> struct inode *inode = d_inode(dentry);
> + struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
> struct f2fs_xattr_entry *entry;
> void *base_addr, *last_base_addr;
> int error;
> size_t rest = buffer_size;
>
> - f2fs_down_read(&F2FS_I(inode)->i_xattr_sem);
> + f2fs_down_read(sbi, &F2FS_I(inode)->i_xattr_sem);
> error = read_all_xattrs(inode, NULL, &base_addr);
> f2fs_up_read(&F2FS_I(inode)->i_xattr_sem);
> if (error)
> @@ -581,12 +583,11 @@ ssize_t f2fs_listxattr(struct dentry *dentry, char *buffer, size_t buffer_size)
>
> if ((void *)(entry) + sizeof(__u32) > last_base_addr ||
> (void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) {
> - f2fs_err(F2FS_I_SB(inode), "inode (%lu) has corrupted xattr",
> + f2fs_err(sbi, "inode (%lu) has corrupted xattr",
> inode->i_ino);
> - set_sbi_flag(F2FS_I_SB(inode), SBI_NEED_FSCK);
> + set_sbi_flag(sbi, SBI_NEED_FSCK);
> error = -EFSCORRUPTED;
> - f2fs_handle_error(F2FS_I_SB(inode),
> - ERROR_CORRUPTED_XATTR);
> + f2fs_handle_error(sbi, ERROR_CORRUPTED_XATTR);
> goto cleanup;
> }
>
@@ -708,6 +708,7 @@ Description: Support configuring fault injection type, should be
FAULT_DQUOT_INIT 0x000010000
FAULT_LOCK_OP 0x000020000
FAULT_BLKADDR 0x000040000
+ FAULT_LOCK 0x000080000
=================== ===========
What: /sys/fs/f2fs/<disk>/discard_io_aware_gran
@@ -206,6 +206,7 @@ fault_type=%d Support configuring fault injection type, should be
FAULT_DQUOT_INIT 0x000010000
FAULT_LOCK_OP 0x000020000
FAULT_BLKADDR 0x000040000
+ FAULT_LOCK 0x000080000
=================== ===========
mode=%s Control block allocation mode which supports "adaptive"
and "lfs". In "lfs" mode, there should be no random
@@ -389,7 +389,7 @@ static int f2fs_write_meta_pages(struct address_space *mapping,
goto skip_write;
/* if locked failed, cp will flush dirty pages instead */
- if (!f2fs_down_write_trylock(&sbi->cp_global_sem))
+ if (!f2fs_down_write_trylock(sbi, &sbi->cp_global_sem))
goto skip_write;
trace_f2fs_writepages(mapping->host, wbc, META);
@@ -1182,7 +1182,7 @@ static bool __need_flush_quota(struct f2fs_sb_info *sbi)
if (!is_journalled_quota(sbi))
return false;
- if (!f2fs_down_write_trylock(&sbi->quota_sem))
+ if (!f2fs_down_write_trylock(sbi, &sbi->quota_sem))
return true;
if (is_sbi_flag_set(sbi, SBI_QUOTA_SKIP_FLUSH)) {
ret = false;
@@ -1228,7 +1228,7 @@ static int block_operations(struct f2fs_sb_info *sbi)
f2fs_unlock_all(sbi);
/* only failed during mount/umount/freeze/quotactl */
- locked = down_read_trylock(&sbi->sb->s_umount);
+ locked = f2fs_raw_down_read_trylock(sbi, &sbi->sb->s_umount);
f2fs_quota_sync(sbi->sb, -1);
if (locked)
up_read(&sbi->sb->s_umount);
@@ -1228,7 +1228,7 @@ static int f2fs_write_compressed_pages(struct compress_ctx *cc,
* checkpoint. This can only happen to quota writes which can cause
* the below discard race condition.
*/
- f2fs_down_read(&sbi->node_write);
+ f2fs_down_read(sbi, &sbi->node_write);
} else if (!f2fs_trylock_op(sbi)) {
goto out_free;
}
@@ -680,7 +680,7 @@ static void __submit_merged_write_cond(struct f2fs_sb_info *sbi,
enum page_type btype = PAGE_TYPE_OF_BIO(type);
struct f2fs_bio_info *io = sbi->write_io[btype] + temp;
- f2fs_down_read(&io->io_rwsem);
+ f2fs_down_read(sbi, &io->io_rwsem);
ret = __has_merged_page(io->bio, inode, page, ino);
f2fs_up_read(&io->io_rwsem);
}
@@ -884,7 +884,7 @@ void f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi,
if (list_empty(head))
continue;
- f2fs_down_read(&io->bio_list_lock);
+ f2fs_down_read(sbi, &io->bio_list_lock);
list_for_each_entry(be, head, list) {
if (target)
found = (target == be->bio);
@@ -1447,7 +1447,7 @@ static int __allocate_data_block(struct dnode_of_data *dn, int seg_type)
static void f2fs_map_lock(struct f2fs_sb_info *sbi, int flag)
{
if (flag == F2FS_GET_BLOCK_PRE_AIO)
- f2fs_down_read(&sbi->node_change);
+ f2fs_down_read(sbi, &sbi->node_change);
else
f2fs_lock_op(sbi);
}
@@ -2839,7 +2839,7 @@ int f2fs_write_single_data_page(struct page *page, int *submitted,
* the below discard race condition.
*/
if (IS_NOQUOTA(inode))
- f2fs_down_read(&sbi->node_write);
+ f2fs_down_read(sbi, &sbi->node_write);
fio.need_lock = LOCK_DONE;
err = f2fs_do_write_data_page(&fio);
@@ -917,7 +917,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
if (!atomic_read(&eti->total_zombie_tree))
goto free_node;
- if (!mutex_trylock(&eti->extent_tree_lock))
+ if (!f2fs_mutex_trylock(sbi, &eti->extent_tree_lock))
goto out;
/* 1. remove unreferenced extent tree */
@@ -943,7 +943,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
free_node:
/* 2. remove LRU extent entries */
- if (!mutex_trylock(&eti->extent_tree_lock))
+ if (!f2fs_mutex_trylock(sbi, &eti->extent_tree_lock))
goto out;
remained = nr_shrink - (node_cnt + tree_cnt);
@@ -955,7 +955,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
en = list_first_entry(&eti->extent_list,
struct extent_node, list);
et = en->et;
- if (!write_trylock(&et->lock)) {
+ if (!f2fs_write_trylock(sbi, &et->lock)) {
/* refresh this extent node's position in extent list */
list_move_tail(&en->list, &eti->extent_list);
continue;
@@ -61,6 +61,7 @@ enum {
FAULT_DQUOT_INIT,
FAULT_LOCK_OP,
FAULT_BLKADDR,
+ FAULT_LOCK,
FAULT_MAX,
};
@@ -2079,27 +2080,76 @@ static inline int f2fs_rwsem_is_contended(struct f2fs_rwsem *sem)
return rwsem_is_contended(&sem->internal_rwsem);
}
-static inline void f2fs_down_read(struct f2fs_rwsem *sem)
+static inline int f2fs_down_read_trylock(struct f2fs_sb_info *sbi,
+ struct f2fs_rwsem *sem)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return down_read_trylock(&sem->internal_rwsem);
+}
+
+static inline void f2fs_down_read(struct f2fs_sb_info *sbi, struct f2fs_rwsem *sem)
{
#ifdef CONFIG_F2FS_UNFAIR_RWSEM
- wait_event(sem->read_waiters, down_read_trylock(&sem->internal_rwsem));
+ wait_event(sem->read_waiters, f2fs_down_read_trylock(sbi, sem));
#else
down_read(&sem->internal_rwsem);
#endif
}
-static inline int f2fs_down_read_trylock(struct f2fs_rwsem *sem)
+static inline int f2fs_raw_down_read_trylock(struct f2fs_sb_info *sbi,
+ struct rw_semaphore *sem)
{
- return down_read_trylock(&sem->internal_rwsem);
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return down_read_trylock(sem);
+}
+
+static inline int f2fs_inode_trylock(struct f2fs_sb_info *sbi,
+ struct inode *inode)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return down_write_trylock(&inode->i_rwsem);
+}
+
+static inline int f2fs_mutex_trylock(struct f2fs_sb_info *sbi,
+ struct mutex *lock)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return mutex_trylock(lock);
+}
+
+static inline int f2fs_trylock_page(struct f2fs_sb_info *sbi, struct page *page)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return trylock_page(page);
+}
+
+static inline int f2fs_write_trylock(struct f2fs_sb_info *sbi, rwlock_t *lock)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return write_trylock(lock);
+}
+
+static inline bool f2fs_sb_start_write_trylock(struct f2fs_sb_info *sbi)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return sb_start_write_trylock(sbi->sb);
}
#ifdef CONFIG_DEBUG_LOCK_ALLOC
-static inline void f2fs_down_read_nested(struct f2fs_rwsem *sem, int subclass)
+static inline void f2fs_down_read_nested(struct f2fs_sb_info *sbi,
+ struct f2fs_rwsem *sem, int subclass)
{
down_read_nested(&sem->internal_rwsem, subclass);
}
#else
-#define f2fs_down_read_nested(sem, subclass) f2fs_down_read(sem)
+#define f2fs_down_read_nested(sbi, sem, subclass) f2fs_down_read(sbi, sem)
#endif
static inline void f2fs_up_read(struct f2fs_rwsem *sem)
@@ -2112,8 +2162,11 @@ static inline void f2fs_down_write(struct f2fs_rwsem *sem)
down_write(&sem->internal_rwsem);
}
-static inline int f2fs_down_write_trylock(struct f2fs_rwsem *sem)
+static inline int f2fs_down_write_trylock(struct f2fs_sb_info *sbi,
+ struct f2fs_rwsem *sem)
{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
return down_write_trylock(&sem->internal_rwsem);
}
@@ -2127,14 +2180,16 @@ static inline void f2fs_up_write(struct f2fs_rwsem *sem)
static inline void f2fs_lock_op(struct f2fs_sb_info *sbi)
{
- f2fs_down_read(&sbi->cp_rwsem);
+ f2fs_down_read(sbi, &sbi->cp_rwsem);
}
static inline int f2fs_trylock_op(struct f2fs_sb_info *sbi)
{
+ struct f2fs_rwsem *sem = &sbi->cp_rwsem;
+
if (time_to_inject(sbi, FAULT_LOCK_OP))
return 0;
- return f2fs_down_read_trylock(&sbi->cp_rwsem);
+ return down_read_trylock(&sem->internal_rwsem);
}
static inline void f2fs_unlock_op(struct f2fs_sb_info *sbi)
@@ -316,7 +316,7 @@ static int f2fs_do_sync_file(struct file *file, loff_t start, loff_t end,
* Both of fdatasync() and fsync() are able to be recovered from
* sudden-power-off.
*/
- f2fs_down_read(&F2FS_I(inode)->i_sem);
+ f2fs_down_read(sbi, &F2FS_I(inode)->i_sem);
cp_reason = need_do_checkpoint(inode);
f2fs_up_read(&F2FS_I(inode)->i_sem);
@@ -2466,7 +2466,7 @@ static int f2fs_ioc_gc(struct file *filp, unsigned long arg)
return ret;
if (!sync) {
- if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
ret = -EBUSY;
goto out;
}
@@ -2510,7 +2510,7 @@ static int __f2fs_ioc_gc_range(struct file *filp, struct f2fs_gc_range *range)
do_more:
if (!range->sync) {
- if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
ret = -EBUSY;
goto out;
}
@@ -2805,7 +2805,7 @@ static int f2fs_move_file_range(struct file *file_in, loff_t pos_in,
inode_lock(src);
if (src != dst) {
ret = -EBUSY;
- if (!inode_trylock(dst))
+ if (!f2fs_inode_trylock(sbi, dst))
goto out;
}
@@ -2855,7 +2855,7 @@ static int f2fs_move_file_range(struct file *file_in, loff_t pos_in,
f2fs_down_write(&F2FS_I(src)->i_gc_rwsem[WRITE]);
if (src != dst) {
ret = -EBUSY;
- if (!f2fs_down_write_trylock(&F2FS_I(dst)->i_gc_rwsem[WRITE]))
+ if (!f2fs_down_write_trylock(sbi, &F2FS_I(dst)->i_gc_rwsem[WRITE]))
goto out_src;
}
@@ -2975,7 +2975,7 @@ static int f2fs_ioc_flush_device(struct file *filp, unsigned long arg)
end_segno = min(start_segno + range.segments, dev_end_segno);
while (start_segno < end_segno) {
- if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
ret = -EBUSY;
goto out;
}
@@ -3325,7 +3325,7 @@ static int f2fs_ioc_getfslabel(struct file *filp, unsigned long arg)
if (!vbuf)
return -ENOMEM;
- f2fs_down_read(&sbi->sb_lock);
+ f2fs_down_read(sbi, &sbi->sb_lock);
count = utf16s_to_utf8s(sbi->raw_super->volume_name,
ARRAY_SIZE(sbi->raw_super->volume_name),
UTF16_LITTLE_ENDIAN, vbuf, MAX_VOLUME_NAME);
@@ -4335,12 +4335,12 @@ static ssize_t f2fs_dio_read_iter(struct kiocb *iocb, struct iov_iter *to)
trace_f2fs_direct_IO_enter(inode, iocb, count, READ);
if (iocb->ki_flags & IOCB_NOWAIT) {
- if (!f2fs_down_read_trylock(&fi->i_gc_rwsem[READ])) {
+ if (!f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[READ])) {
ret = -EAGAIN;
goto out;
}
} else {
- f2fs_down_read(&fi->i_gc_rwsem[READ]);
+ f2fs_down_read(sbi, &fi->i_gc_rwsem[READ]);
}
/*
@@ -4582,11 +4582,11 @@ static ssize_t f2fs_dio_write_iter(struct kiocb *iocb, struct iov_iter *from,
goto out;
}
- if (!f2fs_down_read_trylock(&fi->i_gc_rwsem[WRITE])) {
+ if (!f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[WRITE])) {
ret = -EAGAIN;
goto out;
}
- if (do_opu && !f2fs_down_read_trylock(&fi->i_gc_rwsem[READ])) {
+ if (do_opu && !f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[READ])) {
f2fs_up_read(&fi->i_gc_rwsem[WRITE]);
ret = -EAGAIN;
goto out;
@@ -4596,9 +4596,9 @@ static ssize_t f2fs_dio_write_iter(struct kiocb *iocb, struct iov_iter *from,
if (ret)
goto out;
- f2fs_down_read(&fi->i_gc_rwsem[WRITE]);
+ f2fs_down_read(sbi, &fi->i_gc_rwsem[WRITE]);
if (do_opu)
- f2fs_down_read(&fi->i_gc_rwsem[READ]);
+ f2fs_down_read(sbi, &fi->i_gc_rwsem[READ]);
}
/*
@@ -4675,13 +4675,14 @@ static ssize_t f2fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
struct inode *inode = file_inode(iocb->ki_filp);
const loff_t orig_pos = iocb->ki_pos;
const size_t orig_count = iov_iter_count(from);
+ struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
loff_t target_size;
bool dio;
bool may_need_sync = true;
int preallocated;
ssize_t ret;
- if (unlikely(f2fs_cp_error(F2FS_I_SB(inode)))) {
+ if (unlikely(f2fs_cp_error(sbi))) {
ret = -EIO;
goto out;
}
@@ -4692,7 +4693,7 @@ static ssize_t f2fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
}
if (iocb->ki_flags & IOCB_NOWAIT) {
- if (!inode_trylock(inode)) {
+ if (!f2fs_inode_trylock(sbi, inode)) {
ret = -EAGAIN;
goto out;
}
@@ -76,7 +76,7 @@ static int gc_thread_func(void *data)
f2fs_stop_checkpoint(sbi, false,
STOP_CP_REASON_FAULT_INJECT);
- if (!sb_start_write_trylock(sbi->sb)) {
+ if (!f2fs_sb_start_write_trylock(sbi)) {
stat_other_skip_bggc_count(sbi);
continue;
}
@@ -104,7 +104,7 @@ static int gc_thread_func(void *data)
if (foreground) {
f2fs_down_write(&sbi->gc_lock);
goto do_gc;
- } else if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ } else if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
stat_other_skip_bggc_count(sbi);
goto next;
}
@@ -1573,7 +1573,7 @@ static int gc_data_segment(struct f2fs_sb_info *sbi, struct f2fs_summary *sum,
return submitted;
}
- if (!f2fs_down_write_trylock(
+ if (!f2fs_down_write_trylock(sbi,
&F2FS_I(inode)->i_gc_rwsem[WRITE])) {
iput(inode);
sbi->skipped_gc_rwsem++;
@@ -1616,11 +1616,11 @@ static int gc_data_segment(struct f2fs_sb_info *sbi, struct f2fs_summary *sum,
int err;
if (S_ISREG(inode->i_mode)) {
- if (!f2fs_down_write_trylock(&fi->i_gc_rwsem[WRITE])) {
+ if (!f2fs_down_write_trylock(sbi, &fi->i_gc_rwsem[WRITE])) {
sbi->skipped_gc_rwsem++;
continue;
}
- if (!f2fs_down_write_trylock(
+ if (!f2fs_down_write_trylock(sbi,
&fi->i_gc_rwsem[READ])) {
sbi->skipped_gc_rwsem++;
f2fs_up_write(&fi->i_gc_rwsem[WRITE]);
@@ -2142,7 +2142,7 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
secs = div_u64(shrunk_blocks, BLKS_PER_SEC(sbi));
/* stop other GC */
- if (!f2fs_down_write_trylock(&sbi->gc_lock))
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock))
return -EAGAIN;
/* stop CP to protect MAIN_SEC in free_segment_range */
@@ -144,7 +144,7 @@ static void set_compress_new_inode(struct f2fs_sb_info *sbi, struct inode *dir,
return;
/* Don't compress hot files. */
- f2fs_down_read(&sbi->sb_lock);
+ f2fs_down_read(sbi, &sbi->sb_lock);
cold_count = le32_to_cpu(sbi->raw_super->extension_count);
hot_count = sbi->raw_super->hot_ext_count;
for (i = cold_count; i < cold_count + hot_count; i++)
@@ -185,7 +185,7 @@ static void set_file_temperature(struct f2fs_sb_info *sbi, struct inode *inode,
__u8 (*extlist)[F2FS_EXTENSION_LEN] = sbi->raw_super->extension_list;
int i, cold_count, hot_count;
- f2fs_down_read(&sbi->sb_lock);
+ f2fs_down_read(sbi, &sbi->sb_lock);
cold_count = le32_to_cpu(sbi->raw_super->extension_count);
hot_count = sbi->raw_super->hot_ext_count;
for (i = 0; i < cold_count + hot_count; i++)
@@ -383,7 +383,7 @@ int f2fs_need_dentry_mark(struct f2fs_sb_info *sbi, nid_t nid)
struct nat_entry *e;
bool need = false;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, nid);
if (e) {
if (!get_nat_flag(e, IS_CHECKPOINTED) &&
@@ -400,7 +400,7 @@ bool f2fs_is_checkpointed_node(struct f2fs_sb_info *sbi, nid_t nid)
struct nat_entry *e;
bool is_cp = true;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, nid);
if (e && !get_nat_flag(e, IS_CHECKPOINTED))
is_cp = false;
@@ -414,7 +414,7 @@ bool f2fs_need_inode_block_update(struct f2fs_sb_info *sbi, nid_t ino)
struct nat_entry *e;
bool need_update = true;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, ino);
if (e && get_nat_flag(e, HAS_LAST_FSYNC) &&
(get_nat_flag(e, IS_CHECKPOINTED) ||
@@ -517,7 +517,7 @@ int f2fs_try_to_free_nats(struct f2fs_sb_info *sbi, int nr_shrink)
struct f2fs_nm_info *nm_i = NM_I(sbi);
int nr = nr_shrink;
- if (!f2fs_down_write_trylock(&nm_i->nat_tree_lock))
+ if (!f2fs_down_write_trylock(sbi, &nm_i->nat_tree_lock))
return 0;
spin_lock(&nm_i->nat_list_lock);
@@ -561,7 +561,7 @@ int f2fs_get_node_info(struct f2fs_sb_info *sbi, nid_t nid,
ni->nid = nid;
retry:
/* Check nat cache */
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, nid);
if (e) {
ni->ino = nat_get_ino(e);
@@ -580,7 +580,7 @@ int f2fs_get_node_info(struct f2fs_sb_info *sbi, nid_t nid,
if (!f2fs_rwsem_is_locked(&sbi->cp_global_sem) || checkpoint_context) {
down_read(&curseg->journal_rwsem);
} else if (f2fs_rwsem_is_contended(&nm_i->nat_tree_lock) ||
- !down_read_trylock(&curseg->journal_rwsem)) {
+ !f2fs_raw_down_read_trylock(sbi, &curseg->journal_rwsem)) {
f2fs_up_read(&nm_i->nat_tree_lock);
goto retry;
}
@@ -1618,10 +1618,10 @@ static int __write_node_page(struct page *page, bool atomic, bool *submitted,
goto redirty_out;
if (wbc->for_reclaim) {
- if (!f2fs_down_read_trylock(&sbi->node_write))
+ if (!f2fs_down_read_trylock(sbi, &sbi->node_write))
goto redirty_out;
} else {
- f2fs_down_read(&sbi->node_write);
+ f2fs_down_read(sbi, &sbi->node_write);
}
/* This page is already truncated */
@@ -1982,7 +1982,7 @@ int f2fs_sync_node_pages(struct f2fs_sb_info *sbi,
lock_node:
if (wbc->sync_mode == WB_SYNC_ALL)
lock_page(page);
- else if (!trylock_page(page))
+ else if (!f2fs_trylock_page(sbi, page))
continue;
if (unlikely(page->mapping != NODE_MAPPING(sbi))) {
@@ -2233,7 +2233,7 @@ bool f2fs_nat_bitmap_enabled(struct f2fs_sb_info *sbi)
unsigned int i;
bool ret = true;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
for (i = 0; i < nm_i->nat_blocks; i++) {
if (!test_bit_le(i, nm_i->nat_block_bitmap)) {
ret = false;
@@ -2423,7 +2423,7 @@ static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
unsigned int i, idx;
nid_t nid;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
for (i = 0; i < nm_i->nat_blocks; i++) {
if (!test_bit_le(i, nm_i->nat_block_bitmap))
@@ -2481,7 +2481,7 @@ static int __f2fs_build_free_nids(struct f2fs_sb_info *sbi,
f2fs_ra_meta_pages(sbi, NAT_BLOCK_OFFSET(nid), FREE_NID_PAGES,
META_NAT, true);
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
while (1) {
if (!test_bit_le(NAT_BLOCK_OFFSET(nid),
@@ -2636,7 +2636,7 @@ int f2fs_try_to_free_nids(struct f2fs_sb_info *sbi, int nr_shrink)
if (nm_i->nid_cnt[FREE_NID] <= MAX_FREE_NIDS)
return 0;
- if (!mutex_trylock(&nm_i->build_lock))
+ if (!f2fs_mutex_trylock(sbi, &nm_i->build_lock))
return 0;
while (nr_shrink && nm_i->nid_cnt[FREE_NID] > MAX_FREE_NIDS) {
@@ -2959,7 +2959,7 @@ void f2fs_enable_nat_bits(struct f2fs_sb_info *sbi)
struct f2fs_nm_info *nm_i = NM_I(sbi);
unsigned int nat_ofs;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
for (nat_ofs = 0; nat_ofs < nm_i->nat_blocks; nat_ofs++) {
unsigned int valid = 0, nid_ofs = 0;
@@ -2807,7 +2807,7 @@ static void __f2fs_init_atgc_curseg(struct f2fs_sb_info *sbi)
if (!sbi->am.atgc_enabled)
return;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
mutex_lock(&curseg->curseg_mutex);
down_write(&SIT_I(sbi)->sentry_lock);
@@ -2956,7 +2956,7 @@ void f2fs_allocate_segment_for_resize(struct f2fs_sb_info *sbi, int type,
struct curseg_info *curseg = CURSEG_I(sbi, type);
unsigned int segno;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
mutex_lock(&curseg->curseg_mutex);
down_write(&SIT_I(sbi)->sentry_lock);
@@ -3003,7 +3003,7 @@ static void __allocate_new_segment(struct f2fs_sb_info *sbi, int type,
void f2fs_allocate_new_section(struct f2fs_sb_info *sbi, int type, bool force)
{
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
down_write(&SIT_I(sbi)->sentry_lock);
__allocate_new_segment(sbi, type, true, force);
up_write(&SIT_I(sbi)->sentry_lock);
@@ -3014,7 +3014,7 @@ void f2fs_allocate_new_segments(struct f2fs_sb_info *sbi)
{
int i;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
down_write(&SIT_I(sbi)->sentry_lock);
for (i = CURSEG_HOT_DATA; i <= CURSEG_COLD_DATA; i++)
__allocate_new_segment(sbi, i, false, false);
@@ -3322,7 +3322,7 @@ void f2fs_allocate_data_block(struct f2fs_sb_info *sbi, struct page *page,
struct seg_entry *se = NULL;
bool segment_full = false;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
mutex_lock(&curseg->curseg_mutex);
down_write(&sit_i->sentry_lock);
@@ -3451,17 +3451,18 @@ void f2fs_update_device_state(struct f2fs_sb_info *sbi, nid_t ino,
static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
{
int type = __get_segment_type(fio);
- bool keep_order = (f2fs_lfs_mode(fio->sbi) && type == CURSEG_COLD_DATA);
+ struct f2fs_sb_info *sbi = fio->sbi;
+ bool keep_order = (f2fs_lfs_mode(sbi) && type == CURSEG_COLD_DATA);
if (keep_order)
- f2fs_down_read(&fio->sbi->io_order_lock);
+ f2fs_down_read(sbi, &sbi->io_order_lock);
reallocate:
- f2fs_allocate_data_block(fio->sbi, fio->page, fio->old_blkaddr,
+ f2fs_allocate_data_block(sbi, fio->page, fio->old_blkaddr,
&fio->new_blkaddr, sum, type, fio);
- if (GET_SEGNO(fio->sbi, fio->old_blkaddr) != NULL_SEGNO) {
- invalidate_mapping_pages(META_MAPPING(fio->sbi),
+ if (GET_SEGNO(sbi, fio->old_blkaddr) != NULL_SEGNO) {
+ invalidate_mapping_pages(META_MAPPING(sbi),
fio->old_blkaddr, fio->old_blkaddr);
- f2fs_invalidate_compress_page(fio->sbi, fio->old_blkaddr);
+ f2fs_invalidate_compress_page(sbi, fio->old_blkaddr);
}
/* writeout dirty page into bdev */
@@ -3471,10 +3472,10 @@ static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
goto reallocate;
}
- f2fs_update_device_state(fio->sbi, fio->ino, fio->new_blkaddr, 1);
+ f2fs_update_device_state(sbi, fio->ino, fio->new_blkaddr, 1);
if (keep_order)
- f2fs_up_read(&fio->sbi->io_order_lock);
+ f2fs_up_read(&sbi->io_order_lock);
}
void f2fs_do_write_meta_page(struct f2fs_sb_info *sbi, struct page *page,
@@ -40,17 +40,16 @@ static unsigned long __count_extent_cache(struct f2fs_sb_info *sbi,
unsigned long f2fs_shrink_count(struct shrinker *shrink,
struct shrink_control *sc)
{
- struct f2fs_sb_info *sbi;
struct list_head *p;
unsigned long count = 0;
spin_lock(&f2fs_list_lock);
p = f2fs_list.next;
while (p != &f2fs_list) {
- sbi = list_entry(p, struct f2fs_sb_info, s_list);
+ struct f2fs_sb_info *sbi = list_entry(p, struct f2fs_sb_info, s_list);
/* stop f2fs_put_super */
- if (!mutex_trylock(&sbi->umount_mutex)) {
+ if (!f2fs_mutex_trylock(sbi, &sbi->umount_mutex)) {
p = p->next;
continue;
}
@@ -80,7 +79,6 @@ unsigned long f2fs_shrink_scan(struct shrinker *shrink,
struct shrink_control *sc)
{
unsigned long nr = sc->nr_to_scan;
- struct f2fs_sb_info *sbi;
struct list_head *p;
unsigned int run_no;
unsigned long freed = 0;
@@ -91,13 +89,13 @@ unsigned long f2fs_shrink_scan(struct shrinker *shrink,
} while (run_no == 0);
p = f2fs_list.next;
while (p != &f2fs_list) {
- sbi = list_entry(p, struct f2fs_sb_info, s_list);
+ struct f2fs_sb_info *sbi = list_entry(p, struct f2fs_sb_info, s_list);
if (sbi->shrinker_run_no == run_no)
break;
/* stop f2fs_put_super */
- if (!mutex_trylock(&sbi->umount_mutex)) {
+ if (!f2fs_mutex_trylock(sbi, &sbi->umount_mutex)) {
p = p->next;
continue;
}
@@ -62,6 +62,7 @@ const char *f2fs_fault_name[FAULT_MAX] = {
[FAULT_DQUOT_INIT] = "dquot initialize",
[FAULT_LOCK_OP] = "lock_op",
[FAULT_BLKADDR] = "invalid blkaddr",
+ [FAULT_LOCK] = "lock",
};
void f2fs_build_fault_attr(struct f2fs_sb_info *sbi, unsigned int rate,
@@ -2831,7 +2832,7 @@ int f2fs_quota_sync(struct super_block *sb, int type)
* f2fs_down_read(quota_sem)
*/
f2fs_lock_op(sbi);
- f2fs_down_read(&sbi->quota_sem);
+ f2fs_down_read(sbi, &sbi->quota_sem);
ret = f2fs_quota_sync_file(sbi, cnt);
@@ -2961,7 +2962,7 @@ static int f2fs_dquot_commit(struct dquot *dquot)
struct f2fs_sb_info *sbi = F2FS_SB(dquot->dq_sb);
int ret;
- f2fs_down_read_nested(&sbi->quota_sem, SINGLE_DEPTH_NESTING);
+ f2fs_down_read_nested(sbi, &sbi->quota_sem, SINGLE_DEPTH_NESTING);
ret = dquot_commit(dquot);
if (ret < 0)
set_sbi_flag(sbi, SBI_QUOTA_NEED_REPAIR);
@@ -2974,7 +2975,7 @@ static int f2fs_dquot_acquire(struct dquot *dquot)
struct f2fs_sb_info *sbi = F2FS_SB(dquot->dq_sb);
int ret;
- f2fs_down_read(&sbi->quota_sem);
+ f2fs_down_read(sbi, &sbi->quota_sem);
ret = dquot_acquire(dquot);
if (ret < 0)
set_sbi_flag(sbi, SBI_QUOTA_NEED_REPAIR);
@@ -738,7 +738,7 @@ static ssize_t f2fs_sbi_store(struct f2fs_attr *a,
a->struct_type == GC_THREAD);
if (gc_entry) {
- if (!down_read_trylock(&sbi->sb->s_umount))
+ if (!f2fs_raw_down_read_trylock(sbi, &sbi->sb->s_umount))
return -EAGAIN;
}
ret = __sbi_store(a, sbi, buf, count);
@@ -513,6 +513,7 @@ static inline int write_all_xattrs(struct inode *inode, __u32 hsize,
int f2fs_getxattr(struct inode *inode, int index, const char *name,
void *buffer, size_t buffer_size, struct page *ipage)
{
+ struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
struct f2fs_xattr_entry *entry = NULL;
int error;
unsigned int size, len;
@@ -527,7 +528,7 @@ int f2fs_getxattr(struct inode *inode, int index, const char *name,
if (len > F2FS_NAME_LEN)
return -ERANGE;
- f2fs_down_read(&F2FS_I(inode)->i_xattr_sem);
+ f2fs_down_read(sbi, &F2FS_I(inode)->i_xattr_sem);
error = lookup_all_xattrs(inode, ipage, index, len, name,
&entry, &base_addr, &base_size, &is_inline);
f2fs_up_read(&F2FS_I(inode)->i_xattr_sem);
@@ -552,19 +553,20 @@ int f2fs_getxattr(struct inode *inode, int index, const char *name,
}
error = size;
out:
- xattr_free(F2FS_I_SB(inode), base_addr, is_inline);
+ xattr_free(sbi, base_addr, is_inline);
return error;
}
ssize_t f2fs_listxattr(struct dentry *dentry, char *buffer, size_t buffer_size)
{
struct inode *inode = d_inode(dentry);
+ struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
struct f2fs_xattr_entry *entry;
void *base_addr, *last_base_addr;
int error;
size_t rest = buffer_size;
- f2fs_down_read(&F2FS_I(inode)->i_xattr_sem);
+ f2fs_down_read(sbi, &F2FS_I(inode)->i_xattr_sem);
error = read_all_xattrs(inode, NULL, &base_addr);
f2fs_up_read(&F2FS_I(inode)->i_xattr_sem);
if (error)
@@ -581,12 +583,11 @@ ssize_t f2fs_listxattr(struct dentry *dentry, char *buffer, size_t buffer_size)
if ((void *)(entry) + sizeof(__u32) > last_base_addr ||
(void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) {
- f2fs_err(F2FS_I_SB(inode), "inode (%lu) has corrupted xattr",
+ f2fs_err(sbi, "inode (%lu) has corrupted xattr",
inode->i_ino);
- set_sbi_flag(F2FS_I_SB(inode), SBI_NEED_FSCK);
+ set_sbi_flag(sbi, SBI_NEED_FSCK);
error = -EFSCORRUPTED;
- f2fs_handle_error(F2FS_I_SB(inode),
- ERROR_CORRUPTED_XATTR);
+ f2fs_handle_error(sbi, ERROR_CORRUPTED_XATTR);
goto cleanup;
}