The semantic is reset now, rename the function to reflect that.
Signed-off-by: Alan Huang <[email protected]>
---
fs/bcachefs/btree_key_cache.c | 6 +++---
fs/bcachefs/btree_locking.c | 6 +++---
fs/bcachefs/btree_locking.h | 10 +++++-----
fs/bcachefs/btree_update_interior.c | 4 ++--
4 files changed, 13 insertions(+), 13 deletions(-)
diff --git a/fs/bcachefs/btree_key_cache.c b/fs/bcachefs/btree_key_cache.c
index 2b186584a291..c7ad01c1355c 100644
--- a/fs/bcachefs/btree_key_cache.c
+++ b/fs/bcachefs/btree_key_cache.c
@@ -240,7 +240,7 @@ static int btree_key_cache_create(struct btree_trans *trans,
ck->flags = 1U << BKEY_CACHED_ACCESSED;
if (unlikely(key_u64s > ck->u64s)) {
- mark_btree_node_locked_noreset(ck_path, 0, BTREE_NODE_UNLOCKED);
+ mark_btree_node_locked_reset(ck_path, 0, BTREE_NODE_UNLOCKED);
struct bkey_i *new_k = allocate_dropping_locks(trans, ret,
kmalloc(key_u64s * sizeof(u64), _gfp));
@@ -282,7 +282,7 @@ static int btree_key_cache_create(struct btree_trans *trans,
return 0;
err:
bkey_cached_free(bc, ck);
- mark_btree_node_locked_noreset(ck_path, 0, BTREE_NODE_UNLOCKED);
+ mark_btree_node_locked_reset(ck_path, 0, BTREE_NODE_UNLOCKED);
return ret;
}
@@ -500,7 +500,7 @@ static int btree_key_cache_flush_pos(struct btree_trans
*trans,
atomic_long_dec(&c->btree_key_cache.nr_dirty);
}
- mark_btree_node_locked_noreset(path, 0, BTREE_NODE_UNLOCKED);
+ mark_btree_node_locked_reset(path, 0, BTREE_NODE_UNLOCKED);
if (bkey_cached_evict(&c->btree_key_cache, ck)) {
bkey_cached_free(&c->btree_key_cache, ck);
} else {
diff --git a/fs/bcachefs/btree_locking.c b/fs/bcachefs/btree_locking.c
index f4f563944340..71dbac0bcd58 100644
--- a/fs/bcachefs/btree_locking.c
+++ b/fs/bcachefs/btree_locking.c
@@ -435,7 +435,7 @@ int __bch2_btree_node_lock_write(struct btree_trans *trans,
struct btree_path *p
six_lock_readers_add(&b->lock, readers);
if (ret)
- mark_btree_node_locked_noreset(path, b->level,
BTREE_NODE_INTENT_LOCKED);
+ mark_btree_node_locked_reset(path, b->level,
BTREE_NODE_INTENT_LOCKED);
return ret;
}
@@ -564,7 +564,7 @@ bool bch2_btree_node_upgrade(struct btree_trans *trans,
trace_and_count(trans->c, btree_path_upgrade_fail, trans, _RET_IP_,
path, level);
return false;
success:
- mark_btree_node_locked_noreset(path, level, BTREE_NODE_INTENT_LOCKED);
+ mark_btree_node_locked_reset(path, level, BTREE_NODE_INTENT_LOCKED);
return true;
}
@@ -693,7 +693,7 @@ void __bch2_btree_path_downgrade(struct btree_trans *trans,
} else {
if (btree_node_intent_locked(path, l)) {
six_lock_downgrade(&path->l[l].b->c.lock);
- mark_btree_node_locked_noreset(path, l,
BTREE_NODE_READ_LOCKED);
+ mark_btree_node_locked_reset(path, l,
BTREE_NODE_READ_LOCKED);
}
break;
}
diff --git a/fs/bcachefs/btree_locking.h b/fs/bcachefs/btree_locking.h
index 66b27c0853a5..8978f7969bef 100644
--- a/fs/bcachefs/btree_locking.h
+++ b/fs/bcachefs/btree_locking.h
@@ -63,7 +63,7 @@ static inline bool btree_node_locked(struct btree_path *path,
unsigned level)
return btree_node_locked_type(path, level) != BTREE_NODE_UNLOCKED;
}
-static inline void mark_btree_node_locked_noreset(struct btree_path *path,
+static inline void mark_btree_node_locked_reset(struct btree_path *path,
unsigned level,
enum btree_node_locked_type
type)
{
@@ -80,7 +80,7 @@ static inline void mark_btree_node_locked(struct btree_trans
*trans,
unsigned level,
enum btree_node_locked_type type)
{
- mark_btree_node_locked_noreset(path, level, (enum
btree_node_locked_type) type);
+ mark_btree_node_locked_reset(path, level, (enum btree_node_locked_type)
type);
#ifdef CONFIG_BCACHEFS_LOCK_TIME_STATS
path->l[level].lock_taken_time = local_clock();
#endif
@@ -134,7 +134,7 @@ static inline void btree_node_unlock(struct btree_trans
*trans,
}
six_unlock_type(&path->l[level].b->c.lock, lock_type);
btree_trans_lock_hold_time_update(trans, path, level);
- mark_btree_node_locked_noreset(path, level,
BTREE_NODE_UNLOCKED);
+ mark_btree_node_locked_reset(path, level, BTREE_NODE_UNLOCKED);
}
}
@@ -183,7 +183,7 @@ bch2_btree_node_unlock_write_inlined(struct btree_trans
*trans, struct btree_pat
EBUG_ON(path->l[b->c.level].lock_seq != six_lock_seq(&b->c.lock));
EBUG_ON(btree_node_locked_type(path, b->c.level) != SIX_LOCK_write);
- mark_btree_node_locked_noreset(path, b->c.level,
BTREE_NODE_INTENT_LOCKED);
+ mark_btree_node_locked_reset(path, b->c.level,
BTREE_NODE_INTENT_LOCKED);
__bch2_btree_node_unlock_write(trans, b);
}
@@ -315,7 +315,7 @@ static inline int __btree_node_lock_write(struct
btree_trans *trans,
* write lock: thus, we need to tell the cycle detector we have a write
* lock _before_ taking the lock:
*/
- mark_btree_node_locked_noreset(path, b->level, BTREE_NODE_WRITE_LOCKED);
+ mark_btree_node_locked_reset(path, b->level, BTREE_NODE_WRITE_LOCKED);
return likely(six_trylock_write(&b->lock))
? 0
diff --git a/fs/bcachefs/btree_update_interior.c
b/fs/bcachefs/btree_update_interior.c
index 55fbeeb8eaaa..29e03408a019 100644
--- a/fs/bcachefs/btree_update_interior.c
+++ b/fs/bcachefs/btree_update_interior.c
@@ -245,7 +245,7 @@ static void bch2_btree_node_free_inmem(struct btree_trans
*trans,
mutex_unlock(&c->btree_cache.lock);
six_unlock_write(&b->c.lock);
- mark_btree_node_locked_noreset(path, b->c.level,
BTREE_NODE_INTENT_LOCKED);
+ mark_btree_node_locked_reset(path, b->c.level,
BTREE_NODE_INTENT_LOCKED);
bch2_trans_node_drop(trans, b);
}
@@ -788,7 +788,7 @@ static void btree_update_nodes_written(struct btree_update
*as)
mutex_unlock(&c->btree_interior_update_lock);
- mark_btree_node_locked_noreset(path, b->c.level,
BTREE_NODE_INTENT_LOCKED);
+ mark_btree_node_locked_reset(path, b->c.level,
BTREE_NODE_INTENT_LOCKED);
six_unlock_write(&b->c.lock);
btree_node_write_if_need(trans, b, SIX_LOCK_intent);
--
2.48.1