Lines Matching full:path
26 static inline bool is_btree_node(struct btree_path *path, unsigned l) in is_btree_node() argument
28 return l < BTREE_MAX_DEPTH && !IS_ERR_OR_NULL(path->l[l].b); in is_btree_node()
46 static inline int btree_node_locked_type(struct btree_path *path, in btree_node_locked_type() argument
49 return BTREE_NODE_UNLOCKED + ((path->nodes_locked >> (level << 1)) & 3); in btree_node_locked_type()
52 static inline bool btree_node_write_locked(struct btree_path *path, unsigned l) in btree_node_write_locked() argument
54 return btree_node_locked_type(path, l) == BTREE_NODE_WRITE_LOCKED; in btree_node_write_locked()
57 static inline bool btree_node_intent_locked(struct btree_path *path, unsigned l) in btree_node_intent_locked() argument
59 return btree_node_locked_type(path, l) == BTREE_NODE_INTENT_LOCKED; in btree_node_intent_locked()
62 static inline bool btree_node_read_locked(struct btree_path *path, unsigned l) in btree_node_read_locked() argument
64 return btree_node_locked_type(path, l) == BTREE_NODE_READ_LOCKED; in btree_node_read_locked()
67 static inline bool btree_node_locked(struct btree_path *path, unsigned level) in btree_node_locked() argument
69 return btree_node_locked_type(path, level) != BTREE_NODE_UNLOCKED; in btree_node_locked()
72 static inline void mark_btree_node_locked_noreset(struct btree_path *path, in mark_btree_node_locked_noreset() argument
80 path->nodes_locked &= ~(3U << (level << 1)); in mark_btree_node_locked_noreset()
81 path->nodes_locked |= (type + 1) << (level << 1); in mark_btree_node_locked_noreset()
84 static inline void mark_btree_node_unlocked(struct btree_path *path, in mark_btree_node_unlocked() argument
87 EBUG_ON(btree_node_write_locked(path, level)); in mark_btree_node_unlocked()
88 mark_btree_node_locked_noreset(path, level, BTREE_NODE_UNLOCKED); in mark_btree_node_unlocked()
92 struct btree_path *path, in mark_btree_node_locked() argument
96 mark_btree_node_locked_noreset(path, level, (enum btree_node_locked_type) type); in mark_btree_node_locked()
98 path->l[level].lock_taken_time = local_clock(); in mark_btree_node_locked()
102 static inline enum six_lock_type __btree_lock_want(struct btree_path *path, int level) in __btree_lock_want() argument
104 return level < path->locks_want in __btree_lock_want()
110 btree_lock_want(struct btree_path *path, int level) in btree_lock_want() argument
112 if (level < path->level) in btree_lock_want()
114 if (level < path->locks_want) in btree_lock_want()
116 if (level == path->level) in btree_lock_want()
122 struct btree_path *path, unsigned level) in btree_trans_lock_hold_time_update() argument
126 path->l[level].lock_taken_time, in btree_trans_lock_hold_time_update()
134 struct btree_path *path, unsigned level) in btree_node_unlock() argument
136 int lock_type = btree_node_locked_type(path, level); in btree_node_unlock()
141 six_unlock_type(&path->l[level].b->c.lock, lock_type); in btree_node_unlock()
142 btree_trans_lock_hold_time_update(trans, path, level); in btree_node_unlock()
144 mark_btree_node_unlocked(path, level); in btree_node_unlock()
147 static inline int btree_path_lowest_level_locked(struct btree_path *path) in btree_path_lowest_level_locked() argument
149 return __ffs(path->nodes_locked) >> 1; in btree_path_lowest_level_locked()
152 static inline int btree_path_highest_level_locked(struct btree_path *path) in btree_path_highest_level_locked() argument
154 return __fls(path->nodes_locked) >> 1; in btree_path_highest_level_locked()
158 struct btree_path *path) in __bch2_btree_path_unlock() argument
160 btree_path_set_dirty(path, BTREE_ITER_NEED_RELOCK); in __bch2_btree_path_unlock()
162 while (path->nodes_locked) in __bch2_btree_path_unlock()
163 btree_node_unlock(trans, path, btree_path_lowest_level_locked(path)); in __bch2_btree_path_unlock()
171 bch2_btree_node_unlock_write_inlined(struct btree_trans *trans, struct btree_path *path, in bch2_btree_node_unlock_write_inlined() argument
177 EBUG_ON(path->l[b->c.level].b != b); in bch2_btree_node_unlock_write_inlined()
178 EBUG_ON(path->l[b->c.level].lock_seq != six_lock_seq(&b->c.lock)); in bch2_btree_node_unlock_write_inlined()
179 EBUG_ON(btree_node_locked_type(path, b->c.level) != SIX_LOCK_write); in bch2_btree_node_unlock_write_inlined()
181 mark_btree_node_locked_noreset(path, b->c.level, BTREE_NODE_INTENT_LOCKED); in bch2_btree_node_unlock_write_inlined()
242 struct btree_path *path; in btree_node_lock_increment() local
245 trans_for_each_path(trans, path, i) in btree_node_lock_increment()
246 if (&path->l[level].b->c == b && in btree_node_lock_increment()
247 btree_node_locked_type(path, level) >= want) { in btree_node_lock_increment()
256 struct btree_path *path, in btree_node_lock() argument
268 !(ret = btree_node_lock_nopath(trans, b, type, btree_path_ip_allocated(path)))) { in btree_node_lock()
270 path->l[b->level].lock_taken_time = local_clock(); in btree_node_lock()
281 struct btree_path *path, in __btree_node_lock_write() argument
285 EBUG_ON(&path->l[b->level].b->c != b); in __btree_node_lock_write()
286 EBUG_ON(path->l[b->level].lock_seq != six_lock_seq(&b->lock)); in __btree_node_lock_write()
287 EBUG_ON(!btree_node_intent_locked(path, b->level)); in __btree_node_lock_write()
294 mark_btree_node_locked_noreset(path, b->level, BTREE_NODE_WRITE_LOCKED); in __btree_node_lock_write()
298 : __bch2_btree_node_lock_write(trans, path, b, lock_may_not_fail); in __btree_node_lock_write()
303 struct btree_path *path, in bch2_btree_node_lock_write() argument
306 return __btree_node_lock_write(trans, path, b, false); in bch2_btree_node_lock_write()
320 struct btree_path *path, unsigned long trace_ip) in bch2_btree_path_relock() argument
322 return btree_node_locked(path, path->level) in bch2_btree_path_relock()
324 : __bch2_btree_path_relock(trans, path, trace_ip); in bch2_btree_path_relock()
330 struct btree_path *path, unsigned level) in bch2_btree_node_relock() argument
332 EBUG_ON(btree_node_locked(path, level) && in bch2_btree_node_relock()
333 !btree_node_write_locked(path, level) && in bch2_btree_node_relock()
334 btree_node_locked_type(path, level) != __btree_lock_want(path, level)); in bch2_btree_node_relock()
336 return likely(btree_node_locked(path, level)) || in bch2_btree_node_relock()
337 (!IS_ERR_OR_NULL(path->l[level].b) && in bch2_btree_node_relock()
338 __bch2_btree_node_relock(trans, path, level, true)); in bch2_btree_node_relock()
342 struct btree_path *path, unsigned level) in bch2_btree_node_relock_notrace() argument
344 EBUG_ON(btree_node_locked(path, level) && in bch2_btree_node_relock_notrace()
345 !btree_node_write_locked(path, level) && in bch2_btree_node_relock_notrace()
346 btree_node_locked_type(path, level) != __btree_lock_want(path, level)); in bch2_btree_node_relock_notrace()
348 return likely(btree_node_locked(path, level)) || in bch2_btree_node_relock_notrace()
349 (!IS_ERR_OR_NULL(path->l[level].b) && in bch2_btree_node_relock_notrace()
350 __bch2_btree_node_relock(trans, path, level, false)); in bch2_btree_node_relock_notrace()
364 struct btree_path *path, in bch2_btree_path_upgrade() argument
368 unsigned old_locks_want = path->locks_want; in bch2_btree_path_upgrade()
372 if (path->locks_want < new_locks_want in bch2_btree_path_upgrade()
373 ? __bch2_btree_path_upgrade(trans, path, new_locks_want, &f) in bch2_btree_path_upgrade()
374 : path->uptodate == BTREE_ITER_UPTODATE) in bch2_btree_path_upgrade()
377 trace_and_count(trans->c, trans_restart_upgrade, trans, _THIS_IP_, path, in bch2_btree_path_upgrade()
384 static inline void btree_path_set_should_be_locked(struct btree_path *path) in btree_path_set_should_be_locked() argument
386 EBUG_ON(!btree_node_locked(path, path->level)); in btree_path_set_should_be_locked()
387 EBUG_ON(path->uptodate); in btree_path_set_should_be_locked()
389 path->should_be_locked = true; in btree_path_set_should_be_locked()
393 struct btree_path *path, in __btree_path_set_level_up() argument
396 btree_node_unlock(trans, path, l); in __btree_path_set_level_up()
397 path->l[l].b = ERR_PTR(-BCH_ERR_no_btree_node_up); in __btree_path_set_level_up()
401 struct btree_path *path) in btree_path_set_level_up() argument
403 __btree_path_set_level_up(trans, path, path->level++); in btree_path_set_level_up()
404 btree_path_set_dirty(path, BTREE_ITER_NEED_TRAVERSE); in btree_path_set_level_up()
420 static inline void bch2_btree_path_verify_locks(struct btree_path *path) {} in bch2_btree_path_verify_locks() argument