@@ -714,6 +714,7 @@ Description: Support configuring fault injection type, should be
FAULT_DQUOT_INIT 0x000010000
FAULT_LOCK_OP 0x000020000
FAULT_BLKADDR 0x000040000
+ FAULT_LOCK 0x000080000
=================== ===========
What: /sys/fs/f2fs/<disk>/discard_io_aware_gran
@@ -206,6 +206,7 @@ fault_type=%d Support configuring fault injection type, should be
FAULT_DQUOT_INIT 0x000010000
FAULT_LOCK_OP 0x000020000
FAULT_BLKADDR 0x000040000
+ FAULT_LOCK 0x000080000
=================== ===========
mode=%s Control block allocation mode which supports "adaptive"
and "lfs". In "lfs" mode, there should be no random
@@ -372,7 +372,7 @@ static int f2fs_write_meta_pages(struct address_space *mapping,
goto skip_write;
/* if locked failed, cp will flush dirty pages instead */
- if (!f2fs_down_write_trylock(&sbi->cp_global_sem))
+ if (!f2fs_down_write_trylock(sbi, &sbi->cp_global_sem))
goto skip_write;
trace_f2fs_writepages(mapping->host, wbc, META);
@@ -1185,7 +1185,7 @@ static bool __need_flush_quota(struct f2fs_sb_info *sbi)
if (!is_journalled_quota(sbi))
return false;
- if (!f2fs_down_write_trylock(&sbi->quota_sem))
+ if (!f2fs_down_write_trylock(sbi, &sbi->quota_sem))
return true;
if (is_sbi_flag_set(sbi, SBI_QUOTA_SKIP_FLUSH)) {
ret = false;
@@ -1231,7 +1231,7 @@ static int block_operations(struct f2fs_sb_info *sbi)
f2fs_unlock_all(sbi);
/* only failed during mount/umount/freeze/quotactl */
- locked = down_read_trylock(&sbi->sb->s_umount);
+ locked = f2fs_raw_down_read_trylock(sbi, &sbi->sb->s_umount);
f2fs_quota_sync(sbi->sb, -1);
if (locked)
up_read(&sbi->sb->s_umount);
@@ -1242,7 +1242,7 @@ static int f2fs_write_compressed_pages(struct compress_ctx *cc,
* checkpoint. This can only happen to quota writes which can cause
* the below discard race condition.
*/
- f2fs_down_read(&sbi->node_write);
+ f2fs_down_read(sbi, &sbi->node_write);
} else if (!f2fs_trylock_op(sbi)) {
goto out_free;
}
@@ -680,7 +680,7 @@ static void __submit_merged_write_cond(struct f2fs_sb_info *sbi,
enum page_type btype = PAGE_TYPE_OF_BIO(type);
struct f2fs_bio_info *io = sbi->write_io[btype] + temp;
- f2fs_down_read(&io->io_rwsem);
+ f2fs_down_read(sbi, &io->io_rwsem);
ret = __has_merged_page(io->bio, inode, page, ino);
f2fs_up_read(&io->io_rwsem);
}
@@ -882,7 +882,7 @@ void f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi,
if (list_empty(head))
continue;
- f2fs_down_read(&io->bio_list_lock);
+ f2fs_down_read(sbi, &io->bio_list_lock);
list_for_each_entry(be, head, list) {
if (target)
found = (target == be->bio);
@@ -1441,7 +1441,7 @@ static int __allocate_data_block(struct dnode_of_data *dn, int seg_type)
static void f2fs_map_lock(struct f2fs_sb_info *sbi, int flag)
{
if (flag == F2FS_GET_BLOCK_PRE_AIO)
- f2fs_down_read(&sbi->node_change);
+ f2fs_down_read(sbi, &sbi->node_change);
else
f2fs_lock_op(sbi);
}
@@ -2828,7 +2828,7 @@ int f2fs_write_single_data_page(struct page *page, int *submitted,
* the below discard race condition.
*/
if (IS_NOQUOTA(inode))
- f2fs_down_read(&sbi->node_write);
+ f2fs_down_read(sbi, &sbi->node_write);
fio.need_lock = LOCK_DONE;
err = f2fs_do_write_data_page(&fio);
@@ -909,7 +909,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
if (!atomic_read(&eti->total_zombie_tree))
goto free_node;
- if (!mutex_trylock(&eti->extent_tree_lock))
+ if (!f2fs_mutex_trylock(sbi, &eti->extent_tree_lock))
goto out;
/* 1. remove unreferenced extent tree */
@@ -935,7 +935,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
free_node:
/* 2. remove LRU extent entries */
- if (!mutex_trylock(&eti->extent_tree_lock))
+ if (!f2fs_mutex_trylock(sbi, &eti->extent_tree_lock))
goto out;
remained = nr_shrink - (node_cnt + tree_cnt);
@@ -947,7 +947,7 @@ static unsigned int __shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink
en = list_first_entry(&eti->extent_list,
struct extent_node, list);
et = en->et;
- if (!write_trylock(&et->lock)) {
+ if (!f2fs_write_trylock(sbi, &et->lock)) {
/* refresh this extent node's position in extent list */
list_move_tail(&en->list, &eti->extent_list);
continue;
@@ -61,6 +61,7 @@ enum {
FAULT_DQUOT_INIT,
FAULT_LOCK_OP,
FAULT_BLKADDR,
+ FAULT_LOCK,
FAULT_MAX,
};
@@ -2160,27 +2161,76 @@ static inline int f2fs_rwsem_is_contended(struct f2fs_rwsem *sem)
return rwsem_is_contended(&sem->internal_rwsem);
}
-static inline void f2fs_down_read(struct f2fs_rwsem *sem)
+static inline int f2fs_down_read_trylock(struct f2fs_sb_info *sbi,
+ struct f2fs_rwsem *sem)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return down_read_trylock(&sem->internal_rwsem);
+}
+
+static inline void f2fs_down_read(struct f2fs_sb_info *sbi, struct f2fs_rwsem *sem)
{
#ifdef CONFIG_F2FS_UNFAIR_RWSEM
- wait_event(sem->read_waiters, down_read_trylock(&sem->internal_rwsem));
+ wait_event(sem->read_waiters, f2fs_down_read_trylock(sbi, sem));
#else
down_read(&sem->internal_rwsem);
#endif
}
-static inline int f2fs_down_read_trylock(struct f2fs_rwsem *sem)
+static inline int f2fs_raw_down_read_trylock(struct f2fs_sb_info *sbi,
+ struct rw_semaphore *sem)
{
- return down_read_trylock(&sem->internal_rwsem);
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return down_read_trylock(sem);
+}
+
+static inline int f2fs_inode_trylock(struct f2fs_sb_info *sbi,
+ struct inode *inode)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return down_write_trylock(&inode->i_rwsem);
+}
+
+static inline int f2fs_mutex_trylock(struct f2fs_sb_info *sbi,
+ struct mutex *lock)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return mutex_trylock(lock);
+}
+
+static inline int f2fs_trylock_page(struct f2fs_sb_info *sbi, struct page *page)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return trylock_page(page);
+}
+
+static inline int f2fs_write_trylock(struct f2fs_sb_info *sbi, rwlock_t *lock)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return write_trylock(lock);
+}
+
+static inline bool f2fs_sb_start_write_trylock(struct f2fs_sb_info *sbi)
+{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
+ return sb_start_write_trylock(sbi->sb);
}
#ifdef CONFIG_DEBUG_LOCK_ALLOC
-static inline void f2fs_down_read_nested(struct f2fs_rwsem *sem, int subclass)
+static inline void f2fs_down_read_nested(struct f2fs_sb_info *sbi,
+ struct f2fs_rwsem *sem, int subclass)
{
down_read_nested(&sem->internal_rwsem, subclass);
}
#else
-#define f2fs_down_read_nested(sem, subclass) f2fs_down_read(sem)
+#define f2fs_down_read_nested(sbi, sem, subclass) f2fs_down_read(sbi, sem)
#endif
static inline void f2fs_up_read(struct f2fs_rwsem *sem)
@@ -2193,8 +2243,11 @@ static inline void f2fs_down_write(struct f2fs_rwsem *sem)
down_write(&sem->internal_rwsem);
}
-static inline int f2fs_down_write_trylock(struct f2fs_rwsem *sem)
+static inline int f2fs_down_write_trylock(struct f2fs_sb_info *sbi,
+ struct f2fs_rwsem *sem)
{
+ if (time_to_inject(sbi, FAULT_LOCK))
+ return 0;
return down_write_trylock(&sem->internal_rwsem);
}
@@ -2208,14 +2261,16 @@ static inline void f2fs_up_write(struct f2fs_rwsem *sem)
static inline void f2fs_lock_op(struct f2fs_sb_info *sbi)
{
- f2fs_down_read(&sbi->cp_rwsem);
+ f2fs_down_read(sbi, &sbi->cp_rwsem);
}
static inline int f2fs_trylock_op(struct f2fs_sb_info *sbi)
{
+ struct f2fs_rwsem *sem = &sbi->cp_rwsem;
+
if (time_to_inject(sbi, FAULT_LOCK_OP))
return 0;
- return f2fs_down_read_trylock(&sbi->cp_rwsem);
+ return down_read_trylock(&sem->internal_rwsem);
}
static inline void f2fs_unlock_op(struct f2fs_sb_info *sbi)
@@ -316,7 +316,7 @@ static int f2fs_do_sync_file(struct file *file, loff_t start, loff_t end,
* Both of fdatasync() and fsync() are able to be recovered from
* sudden-power-off.
*/
- f2fs_down_read(&F2FS_I(inode)->i_sem);
+ f2fs_down_read(sbi, &F2FS_I(inode)->i_sem);
cp_reason = need_do_checkpoint(inode);
f2fs_up_read(&F2FS_I(inode)->i_sem);
@@ -2462,7 +2462,7 @@ static int f2fs_ioc_gc(struct file *filp, unsigned long arg)
return ret;
if (!sync) {
- if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
ret = -EBUSY;
goto out;
}
@@ -2506,7 +2506,7 @@ static int __f2fs_ioc_gc_range(struct file *filp, struct f2fs_gc_range *range)
do_more:
if (!range->sync) {
- if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
ret = -EBUSY;
goto out;
}
@@ -2801,7 +2801,7 @@ static int f2fs_move_file_range(struct file *file_in, loff_t pos_in,
inode_lock(src);
if (src != dst) {
ret = -EBUSY;
- if (!inode_trylock(dst))
+ if (!f2fs_inode_trylock(sbi, dst))
goto out;
}
@@ -2851,7 +2851,7 @@ static int f2fs_move_file_range(struct file *file_in, loff_t pos_in,
f2fs_down_write(&F2FS_I(src)->i_gc_rwsem[WRITE]);
if (src != dst) {
ret = -EBUSY;
- if (!f2fs_down_write_trylock(&F2FS_I(dst)->i_gc_rwsem[WRITE]))
+ if (!f2fs_down_write_trylock(sbi, &F2FS_I(dst)->i_gc_rwsem[WRITE]))
goto out_src;
}
@@ -2971,7 +2971,7 @@ static int f2fs_ioc_flush_device(struct file *filp, unsigned long arg)
end_segno = min(start_segno + range.segments, dev_end_segno);
while (start_segno < end_segno) {
- if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
ret = -EBUSY;
goto out;
}
@@ -3321,7 +3321,7 @@ static int f2fs_ioc_getfslabel(struct file *filp, unsigned long arg)
if (!vbuf)
return -ENOMEM;
- f2fs_down_read(&sbi->sb_lock);
+ f2fs_down_read(sbi, &sbi->sb_lock);
count = utf16s_to_utf8s(sbi->raw_super->volume_name,
ARRAY_SIZE(sbi->raw_super->volume_name),
UTF16_LITTLE_ENDIAN, vbuf, MAX_VOLUME_NAME);
@@ -4331,12 +4331,12 @@ static ssize_t f2fs_dio_read_iter(struct kiocb *iocb, struct iov_iter *to)
trace_f2fs_direct_IO_enter(inode, iocb, count, READ);
if (iocb->ki_flags & IOCB_NOWAIT) {
- if (!f2fs_down_read_trylock(&fi->i_gc_rwsem[READ])) {
+ if (!f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[READ])) {
ret = -EAGAIN;
goto out;
}
} else {
- f2fs_down_read(&fi->i_gc_rwsem[READ]);
+ f2fs_down_read(sbi, &fi->i_gc_rwsem[READ]);
}
/*
@@ -4578,11 +4578,11 @@ static ssize_t f2fs_dio_write_iter(struct kiocb *iocb, struct iov_iter *from,
goto out;
}
- if (!f2fs_down_read_trylock(&fi->i_gc_rwsem[WRITE])) {
+ if (!f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[WRITE])) {
ret = -EAGAIN;
goto out;
}
- if (do_opu && !f2fs_down_read_trylock(&fi->i_gc_rwsem[READ])) {
+ if (do_opu && !f2fs_down_read_trylock(sbi, &fi->i_gc_rwsem[READ])) {
f2fs_up_read(&fi->i_gc_rwsem[WRITE]);
ret = -EAGAIN;
goto out;
@@ -4592,9 +4592,9 @@ static ssize_t f2fs_dio_write_iter(struct kiocb *iocb, struct iov_iter *from,
if (ret)
goto out;
- f2fs_down_read(&fi->i_gc_rwsem[WRITE]);
+ f2fs_down_read(sbi, &fi->i_gc_rwsem[WRITE]);
if (do_opu)
- f2fs_down_read(&fi->i_gc_rwsem[READ]);
+ f2fs_down_read(sbi, &fi->i_gc_rwsem[READ]);
}
/*
@@ -4671,13 +4671,14 @@ static ssize_t f2fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
struct inode *inode = file_inode(iocb->ki_filp);
const loff_t orig_pos = iocb->ki_pos;
const size_t orig_count = iov_iter_count(from);
+ struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
loff_t target_size;
bool dio;
bool may_need_sync = true;
int preallocated;
ssize_t ret;
- if (unlikely(f2fs_cp_error(F2FS_I_SB(inode)))) {
+ if (unlikely(f2fs_cp_error(sbi))) {
ret = -EIO;
goto out;
}
@@ -4688,7 +4689,7 @@ static ssize_t f2fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
}
if (iocb->ki_flags & IOCB_NOWAIT) {
- if (!inode_trylock(inode)) {
+ if (!f2fs_inode_trylock(sbi, inode)) {
ret = -EAGAIN;
goto out;
}
@@ -76,7 +76,7 @@ static int gc_thread_func(void *data)
f2fs_stop_checkpoint(sbi, false,
STOP_CP_REASON_FAULT_INJECT);
- if (!sb_start_write_trylock(sbi->sb)) {
+ if (!f2fs_sb_start_write_trylock(sbi)) {
stat_other_skip_bggc_count(sbi);
continue;
}
@@ -104,7 +104,7 @@ static int gc_thread_func(void *data)
if (foreground) {
f2fs_down_write(&sbi->gc_lock);
goto do_gc;
- } else if (!f2fs_down_write_trylock(&sbi->gc_lock)) {
+ } else if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock)) {
stat_other_skip_bggc_count(sbi);
goto next;
}
@@ -1577,7 +1577,7 @@ static int gc_data_segment(struct f2fs_sb_info *sbi, struct f2fs_summary *sum,
return submitted;
}
- if (!f2fs_down_write_trylock(
+ if (!f2fs_down_write_trylock(sbi,
&F2FS_I(inode)->i_gc_rwsem[WRITE])) {
iput(inode);
sbi->skipped_gc_rwsem++;
@@ -1620,11 +1620,11 @@ static int gc_data_segment(struct f2fs_sb_info *sbi, struct f2fs_summary *sum,
int err;
if (S_ISREG(inode->i_mode)) {
- if (!f2fs_down_write_trylock(&fi->i_gc_rwsem[READ])) {
+ if (!f2fs_down_write_trylock(sbi, &fi->i_gc_rwsem[READ])) {
sbi->skipped_gc_rwsem++;
continue;
}
- if (!f2fs_down_write_trylock(
+ if (!f2fs_down_write_trylock(sbi,
&fi->i_gc_rwsem[WRITE])) {
sbi->skipped_gc_rwsem++;
f2fs_up_write(&fi->i_gc_rwsem[READ]);
@@ -2150,7 +2150,7 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
secs = div_u64(shrunk_blocks, BLKS_PER_SEC(sbi));
/* stop other GC */
- if (!f2fs_down_write_trylock(&sbi->gc_lock))
+ if (!f2fs_down_write_trylock(sbi, &sbi->gc_lock))
return -EAGAIN;
/* stop CP to protect MAIN_SEC in free_segment_range */
@@ -144,7 +144,7 @@ static void set_compress_new_inode(struct f2fs_sb_info *sbi, struct inode *dir,
return;
/* Don't compress hot files. */
- f2fs_down_read(&sbi->sb_lock);
+ f2fs_down_read(sbi, &sbi->sb_lock);
cold_count = le32_to_cpu(sbi->raw_super->extension_count);
hot_count = sbi->raw_super->hot_ext_count;
for (i = cold_count; i < cold_count + hot_count; i++)
@@ -185,7 +185,7 @@ static void set_file_temperature(struct f2fs_sb_info *sbi, struct inode *inode,
__u8 (*extlist)[F2FS_EXTENSION_LEN] = sbi->raw_super->extension_list;
int i, cold_count, hot_count;
- f2fs_down_read(&sbi->sb_lock);
+ f2fs_down_read(sbi, &sbi->sb_lock);
cold_count = le32_to_cpu(sbi->raw_super->extension_count);
hot_count = sbi->raw_super->hot_ext_count;
for (i = 0; i < cold_count + hot_count; i++)
@@ -383,7 +383,7 @@ int f2fs_need_dentry_mark(struct f2fs_sb_info *sbi, nid_t nid)
struct nat_entry *e;
bool need = false;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, nid);
if (e) {
if (!get_nat_flag(e, IS_CHECKPOINTED) &&
@@ -400,7 +400,7 @@ bool f2fs_is_checkpointed_node(struct f2fs_sb_info *sbi, nid_t nid)
struct nat_entry *e;
bool is_cp = true;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, nid);
if (e && !get_nat_flag(e, IS_CHECKPOINTED))
is_cp = false;
@@ -414,7 +414,7 @@ bool f2fs_need_inode_block_update(struct f2fs_sb_info *sbi, nid_t ino)
struct nat_entry *e;
bool need_update = true;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, ino);
if (e && get_nat_flag(e, HAS_LAST_FSYNC) &&
(get_nat_flag(e, IS_CHECKPOINTED) ||
@@ -517,7 +517,7 @@ int f2fs_try_to_free_nats(struct f2fs_sb_info *sbi, int nr_shrink)
struct f2fs_nm_info *nm_i = NM_I(sbi);
int nr = nr_shrink;
- if (!f2fs_down_write_trylock(&nm_i->nat_tree_lock))
+ if (!f2fs_down_write_trylock(sbi, &nm_i->nat_tree_lock))
return 0;
spin_lock(&nm_i->nat_list_lock);
@@ -561,7 +561,7 @@ int f2fs_get_node_info(struct f2fs_sb_info *sbi, nid_t nid,
ni->nid = nid;
retry:
/* Check nat cache */
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
e = __lookup_nat_cache(nm_i, nid);
if (e) {
ni->ino = nat_get_ino(e);
@@ -580,7 +580,7 @@ int f2fs_get_node_info(struct f2fs_sb_info *sbi, nid_t nid,
if (!f2fs_rwsem_is_locked(&sbi->cp_global_sem) || checkpoint_context) {
down_read(&curseg->journal_rwsem);
} else if (f2fs_rwsem_is_contended(&nm_i->nat_tree_lock) ||
- !down_read_trylock(&curseg->journal_rwsem)) {
+ !f2fs_raw_down_read_trylock(sbi, &curseg->journal_rwsem)) {
f2fs_up_read(&nm_i->nat_tree_lock);
goto retry;
}
@@ -1618,10 +1618,10 @@ static int __write_node_page(struct page *page, bool atomic, bool *submitted,
goto redirty_out;
if (wbc->for_reclaim) {
- if (!f2fs_down_read_trylock(&sbi->node_write))
+ if (!f2fs_down_read_trylock(sbi, &sbi->node_write))
goto redirty_out;
} else {
- f2fs_down_read(&sbi->node_write);
+ f2fs_down_read(sbi, &sbi->node_write);
}
/* This page is already truncated */
@@ -1982,7 +1982,7 @@ int f2fs_sync_node_pages(struct f2fs_sb_info *sbi,
lock_node:
if (wbc->sync_mode == WB_SYNC_ALL)
lock_page(page);
- else if (!trylock_page(page))
+ else if (!f2fs_trylock_page(sbi, page))
continue;
if (unlikely(page->mapping != NODE_MAPPING(sbi))) {
@@ -2233,7 +2233,7 @@ bool f2fs_nat_bitmap_enabled(struct f2fs_sb_info *sbi)
unsigned int i;
bool ret = true;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
for (i = 0; i < nm_i->nat_blocks; i++) {
if (!test_bit_le(i, nm_i->nat_block_bitmap)) {
ret = false;
@@ -2423,7 +2423,7 @@ static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
unsigned int i, idx;
nid_t nid;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
for (i = 0; i < nm_i->nat_blocks; i++) {
if (!test_bit_le(i, nm_i->nat_block_bitmap))
@@ -2481,7 +2481,7 @@ static int __f2fs_build_free_nids(struct f2fs_sb_info *sbi,
f2fs_ra_meta_pages(sbi, NAT_BLOCK_OFFSET(nid), FREE_NID_PAGES,
META_NAT, true);
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
while (1) {
if (!test_bit_le(NAT_BLOCK_OFFSET(nid),
@@ -2636,7 +2636,7 @@ int f2fs_try_to_free_nids(struct f2fs_sb_info *sbi, int nr_shrink)
if (nm_i->nid_cnt[FREE_NID] <= MAX_FREE_NIDS)
return 0;
- if (!mutex_trylock(&nm_i->build_lock))
+ if (!f2fs_mutex_trylock(sbi, &nm_i->build_lock))
return 0;
while (nr_shrink && nm_i->nid_cnt[FREE_NID] > MAX_FREE_NIDS) {
@@ -2959,7 +2959,7 @@ void f2fs_enable_nat_bits(struct f2fs_sb_info *sbi)
struct f2fs_nm_info *nm_i = NM_I(sbi);
unsigned int nat_ofs;
- f2fs_down_read(&nm_i->nat_tree_lock);
+ f2fs_down_read(sbi, &nm_i->nat_tree_lock);
for (nat_ofs = 0; nat_ofs < nm_i->nat_blocks; nat_ofs++) {
unsigned int valid = 0, nid_ofs = 0;
@@ -2806,7 +2806,7 @@ static void __f2fs_init_atgc_curseg(struct f2fs_sb_info *sbi)
if (!sbi->am.atgc_enabled)
return;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
mutex_lock(&curseg->curseg_mutex);
down_write(&SIT_I(sbi)->sentry_lock);
@@ -2956,7 +2956,7 @@ void f2fs_allocate_segment_for_resize(struct f2fs_sb_info *sbi, int type,
struct curseg_info *curseg = CURSEG_I(sbi, type);
unsigned int segno;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
mutex_lock(&curseg->curseg_mutex);
down_write(&SIT_I(sbi)->sentry_lock);
@@ -3003,7 +3003,7 @@ static void __allocate_new_segment(struct f2fs_sb_info *sbi, int type,
void f2fs_allocate_new_section(struct f2fs_sb_info *sbi, int type, bool force)
{
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
down_write(&SIT_I(sbi)->sentry_lock);
__allocate_new_segment(sbi, type, true, force);
up_write(&SIT_I(sbi)->sentry_lock);
@@ -3014,7 +3014,7 @@ void f2fs_allocate_new_segments(struct f2fs_sb_info *sbi)
{
int i;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
down_write(&SIT_I(sbi)->sentry_lock);
for (i = CURSEG_HOT_DATA; i <= CURSEG_COLD_DATA; i++)
__allocate_new_segment(sbi, i, false, false);
@@ -3322,7 +3322,7 @@ void f2fs_allocate_data_block(struct f2fs_sb_info *sbi, struct page *page,
struct seg_entry *se = NULL;
bool segment_full = false;
- f2fs_down_read(&SM_I(sbi)->curseg_lock);
+ f2fs_down_read(sbi, &SM_I(sbi)->curseg_lock);
mutex_lock(&curseg->curseg_mutex);
down_write(&sit_i->sentry_lock);
@@ -3451,17 +3451,18 @@ void f2fs_update_device_state(struct f2fs_sb_info *sbi, nid_t ino,
static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
{
int type = __get_segment_type(fio);
- bool keep_order = (f2fs_lfs_mode(fio->sbi) && type == CURSEG_COLD_DATA);
+ struct f2fs_sb_info *sbi = fio->sbi;
+ bool keep_order = (f2fs_lfs_mode(sbi) && type == CURSEG_COLD_DATA);
if (keep_order)
- f2fs_down_read(&fio->sbi->io_order_lock);
+ f2fs_down_read(sbi, &sbi->io_order_lock);
reallocate:
- f2fs_allocate_data_block(fio->sbi, fio->page, fio->old_blkaddr,
+ f2fs_allocate_data_block(sbi, fio->page, fio->old_blkaddr,
&fio->new_blkaddr, sum, type, fio);
- if (GET_SEGNO(fio->sbi, fio->old_blkaddr) != NULL_SEGNO) {
- invalidate_mapping_pages(META_MAPPING(fio->sbi),
+ if (GET_SEGNO(sbi, fio->old_blkaddr) != NULL_SEGNO) {
+ invalidate_mapping_pages(META_MAPPING(sbi),
fio->old_blkaddr, fio->old_blkaddr);
- f2fs_invalidate_compress_page(fio->sbi, fio->old_blkaddr);
+ f2fs_invalidate_compress_page(sbi, fio->old_blkaddr);
}
/* writeout dirty page into bdev */
@@ -3471,10 +3472,10 @@ static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
goto reallocate;
}
- f2fs_update_device_state(fio->sbi, fio->ino, fio->new_blkaddr, 1);
+ f2fs_update_device_state(sbi, fio->ino, fio->new_blkaddr, 1);
if (keep_order)
- f2fs_up_read(&fio->sbi->io_order_lock);
+ f2fs_up_read(&sbi->io_order_lock);
}
void f2fs_do_write_meta_page(struct f2fs_sb_info *sbi, struct page *page,
@@ -40,17 +40,16 @@ static unsigned long __count_extent_cache(struct f2fs_sb_info *sbi,
unsigned long f2fs_shrink_count(struct shrinker *shrink,
struct shrink_control *sc)
{
- struct f2fs_sb_info *sbi;
struct list_head *p;
unsigned long count = 0;
spin_lock(&f2fs_list_lock);
p = f2fs_list.next;
while (p != &f2fs_list) {
- sbi = list_entry(p, struct f2fs_sb_info, s_list);
+ struct f2fs_sb_info *sbi = list_entry(p, struct f2fs_sb_info, s_list);
/* stop f2fs_put_super */
- if (!mutex_trylock(&sbi->umount_mutex)) {
+ if (!f2fs_mutex_trylock(sbi, &sbi->umount_mutex)) {
p = p->next;
continue;
}
@@ -80,7 +79,6 @@ unsigned long f2fs_shrink_scan(struct shrinker *shrink,
struct shrink_control *sc)
{
unsigned long nr = sc->nr_to_scan;
- struct f2fs_sb_info *sbi;
struct list_head *p;
unsigned int run_no;
unsigned long freed = 0;
@@ -91,13 +89,13 @@ unsigned long f2fs_shrink_scan(struct shrinker *shrink,
} while (run_no == 0);
p = f2fs_list.next;
while (p != &f2fs_list) {
- sbi = list_entry(p, struct f2fs_sb_info, s_list);
+ struct f2fs_sb_info *sbi = list_entry(p, struct f2fs_sb_info, s_list);
if (sbi->shrinker_run_no == run_no)
break;
/* stop f2fs_put_super */
- if (!mutex_trylock(&sbi->umount_mutex)) {
+ if (!f2fs_mutex_trylock(sbi, &sbi->umount_mutex)) {
p = p->next;
continue;
}
@@ -62,6 +62,7 @@ const char *f2fs_fault_name[FAULT_MAX] = {
[FAULT_DQUOT_INIT] = "dquot initialize",
[FAULT_LOCK_OP] = "lock_op",
[FAULT_BLKADDR] = "invalid blkaddr",
+ [FAULT_LOCK] = "lock",
};
void f2fs_build_fault_attr(struct f2fs_sb_info *sbi, unsigned int rate,
@@ -2781,7 +2782,7 @@ int f2fs_quota_sync(struct super_block *sb, int type)
* f2fs_down_read(quota_sem)
*/
f2fs_lock_op(sbi);
- f2fs_down_read(&sbi->quota_sem);
+ f2fs_down_read(sbi, &sbi->quota_sem);
ret = f2fs_quota_sync_file(sbi, cnt);
@@ -2911,7 +2912,7 @@ static int f2fs_dquot_commit(struct dquot *dquot)
struct f2fs_sb_info *sbi = F2FS_SB(dquot->dq_sb);
int ret;
- f2fs_down_read_nested(&sbi->quota_sem, SINGLE_DEPTH_NESTING);
+ f2fs_down_read_nested(sbi, &sbi->quota_sem, SINGLE_DEPTH_NESTING);
ret = dquot_commit(dquot);
if (ret < 0)
set_sbi_flag(sbi, SBI_QUOTA_NEED_REPAIR);
@@ -2924,7 +2925,7 @@ static int f2fs_dquot_acquire(struct dquot *dquot)
struct f2fs_sb_info *sbi = F2FS_SB(dquot->dq_sb);
int ret;
- f2fs_down_read(&sbi->quota_sem);
+ f2fs_down_read(sbi, &sbi->quota_sem);
ret = dquot_acquire(dquot);
if (ret < 0)
set_sbi_flag(sbi, SBI_QUOTA_NEED_REPAIR);
@@ -743,7 +743,7 @@ static ssize_t f2fs_sbi_store(struct f2fs_attr *a,
a->struct_type == GC_THREAD);
if (gc_entry) {
- if (!down_read_trylock(&sbi->sb->s_umount))
+ if (!f2fs_raw_down_read_trylock(sbi, &sbi->sb->s_umount))
return -EAGAIN;
}
ret = __sbi_store(a, sbi, buf, count);
@@ -513,6 +513,7 @@ static inline int write_all_xattrs(struct inode *inode, __u32 hsize,
int f2fs_getxattr(struct inode *inode, int index, const char *name,
void *buffer, size_t buffer_size, struct page *ipage)
{
+ struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
struct f2fs_xattr_entry *entry = NULL;
int error;
unsigned int size, len;
@@ -527,7 +528,7 @@ int f2fs_getxattr(struct inode *inode, int index, const char *name,
if (len > F2FS_NAME_LEN)
return -ERANGE;
- f2fs_down_read(&F2FS_I(inode)->i_xattr_sem);
+ f2fs_down_read(sbi, &F2FS_I(inode)->i_xattr_sem);
error = lookup_all_xattrs(inode, ipage, index, len, name,
&entry, &base_addr, &base_size, &is_inline);
f2fs_up_read(&F2FS_I(inode)->i_xattr_sem);
@@ -552,19 +553,20 @@ int f2fs_getxattr(struct inode *inode, int index, const char *name,
}
error = size;
out:
- xattr_free(F2FS_I_SB(inode), base_addr, is_inline);
+ xattr_free(sbi, base_addr, is_inline);
return error;
}
ssize_t f2fs_listxattr(struct dentry *dentry, char *buffer, size_t buffer_size)
{
struct inode *inode = d_inode(dentry);
+ struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
struct f2fs_xattr_entry *entry;
void *base_addr, *last_base_addr;
int error;
size_t rest = buffer_size;
- f2fs_down_read(&F2FS_I(inode)->i_xattr_sem);
+ f2fs_down_read(sbi, &F2FS_I(inode)->i_xattr_sem);
error = read_all_xattrs(inode, NULL, &base_addr);
f2fs_up_read(&F2FS_I(inode)->i_xattr_sem);
if (error)
@@ -581,12 +583,11 @@ ssize_t f2fs_listxattr(struct dentry *dentry, char *buffer, size_t buffer_size)
if ((void *)(entry) + sizeof(__u32) > last_base_addr ||
(void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) {
- f2fs_err(F2FS_I_SB(inode), "inode (%lu) has corrupted xattr",
+ f2fs_err(sbi, "inode (%lu) has corrupted xattr",
inode->i_ino);
- set_sbi_flag(F2FS_I_SB(inode), SBI_NEED_FSCK);
+ set_sbi_flag(sbi, SBI_NEED_FSCK);
error = -EFSCORRUPTED;
- f2fs_handle_error(F2FS_I_SB(inode),
- ERROR_CORRUPTED_XATTR);
+ f2fs_handle_error(sbi, ERROR_CORRUPTED_XATTR);
goto cleanup;
}