Lines Matching full:path
21 static inline bool is_btree_node(struct btree_path *path, unsigned l) in is_btree_node() argument
23 return l < BTREE_MAX_DEPTH && !IS_ERR_OR_NULL(path->l[l].b); in is_btree_node()
41 static inline int btree_node_locked_type(struct btree_path *path, in btree_node_locked_type() argument
44 return BTREE_NODE_UNLOCKED + ((path->nodes_locked >> (level << 1)) & 3); in btree_node_locked_type()
47 static inline bool btree_node_write_locked(struct btree_path *path, unsigned l) in btree_node_write_locked() argument
49 return btree_node_locked_type(path, l) == BTREE_NODE_WRITE_LOCKED; in btree_node_write_locked()
52 static inline bool btree_node_intent_locked(struct btree_path *path, unsigned l) in btree_node_intent_locked() argument
54 return btree_node_locked_type(path, l) == BTREE_NODE_INTENT_LOCKED; in btree_node_intent_locked()
57 static inline bool btree_node_read_locked(struct btree_path *path, unsigned l) in btree_node_read_locked() argument
59 return btree_node_locked_type(path, l) == BTREE_NODE_READ_LOCKED; in btree_node_read_locked()
62 static inline bool btree_node_locked(struct btree_path *path, unsigned level) in btree_node_locked() argument
64 return btree_node_locked_type(path, level) != BTREE_NODE_UNLOCKED; in btree_node_locked()
67 static inline void mark_btree_node_locked_noreset(struct btree_path *path, in mark_btree_node_locked_noreset() argument
75 path->nodes_locked &= ~(3U << (level << 1)); in mark_btree_node_locked_noreset()
76 path->nodes_locked |= (type + 1) << (level << 1); in mark_btree_node_locked_noreset()
80 struct btree_path *path, in mark_btree_node_locked() argument
84 mark_btree_node_locked_noreset(path, level, (enum btree_node_locked_type) type); in mark_btree_node_locked()
86 path->l[level].lock_taken_time = local_clock(); in mark_btree_node_locked()
90 static inline enum six_lock_type __btree_lock_want(struct btree_path *path, int level) in __btree_lock_want() argument
92 return level < path->locks_want in __btree_lock_want()
98 btree_lock_want(struct btree_path *path, int level) in btree_lock_want() argument
100 if (level < path->level) in btree_lock_want()
102 if (level < path->locks_want) in btree_lock_want()
104 if (level == path->level) in btree_lock_want()
110 struct btree_path *path, unsigned level) in btree_trans_lock_hold_time_update() argument
114 path->l[level].lock_taken_time, in btree_trans_lock_hold_time_update()
125 struct btree_path *path, unsigned level) in btree_node_unlock() argument
127 int lock_type = btree_node_locked_type(path, level); in btree_node_unlock()
133 bch2_btree_node_unlock_write(trans, path, path->l[level].b); in btree_node_unlock()
136 six_unlock_type(&path->l[level].b->c.lock, lock_type); in btree_node_unlock()
137 btree_trans_lock_hold_time_update(trans, path, level); in btree_node_unlock()
138 mark_btree_node_locked_noreset(path, level, BTREE_NODE_UNLOCKED); in btree_node_unlock()
142 static inline int btree_path_lowest_level_locked(struct btree_path *path) in btree_path_lowest_level_locked() argument
144 return __ffs(path->nodes_locked) >> 1; in btree_path_lowest_level_locked()
147 static inline int btree_path_highest_level_locked(struct btree_path *path) in btree_path_highest_level_locked() argument
149 return __fls(path->nodes_locked) >> 1; in btree_path_highest_level_locked()
153 struct btree_path *path) in __bch2_btree_path_unlock() argument
155 btree_path_set_dirty(path, BTREE_ITER_NEED_RELOCK); in __bch2_btree_path_unlock()
157 while (path->nodes_locked) in __bch2_btree_path_unlock()
158 btree_node_unlock(trans, path, btree_path_lowest_level_locked(path)); in __bch2_btree_path_unlock()
180 bch2_btree_node_unlock_write_inlined(struct btree_trans *trans, struct btree_path *path, in bch2_btree_node_unlock_write_inlined() argument
183 EBUG_ON(path->l[b->c.level].b != b); in bch2_btree_node_unlock_write_inlined()
184 EBUG_ON(path->l[b->c.level].lock_seq != six_lock_seq(&b->c.lock)); in bch2_btree_node_unlock_write_inlined()
185 EBUG_ON(btree_node_locked_type(path, b->c.level) != SIX_LOCK_write); in bch2_btree_node_unlock_write_inlined()
187 mark_btree_node_locked_noreset(path, b->c.level, BTREE_NODE_INTENT_LOCKED); in bch2_btree_node_unlock_write_inlined()
266 struct btree_path *path; in btree_node_lock_increment() local
269 trans_for_each_path(trans, path, i) in btree_node_lock_increment()
270 if (&path->l[level].b->c == b && in btree_node_lock_increment()
271 btree_node_locked_type(path, level) >= want) { in btree_node_lock_increment()
280 struct btree_path *path, in btree_node_lock() argument
293 !(ret = btree_node_lock_nopath(trans, b, type, btree_path_ip_allocated(path)))) { in btree_node_lock()
295 path->l[b->level].lock_taken_time = local_clock(); in btree_node_lock()
306 struct btree_path *path, in __btree_node_lock_write() argument
310 EBUG_ON(&path->l[b->level].b->c != b); in __btree_node_lock_write()
311 EBUG_ON(path->l[b->level].lock_seq != six_lock_seq(&b->lock)); in __btree_node_lock_write()
312 EBUG_ON(!btree_node_intent_locked(path, b->level)); in __btree_node_lock_write()
319 mark_btree_node_locked_noreset(path, b->level, BTREE_NODE_WRITE_LOCKED); in __btree_node_lock_write()
323 : __bch2_btree_node_lock_write(trans, path, b, lock_may_not_fail); in __btree_node_lock_write()
328 struct btree_path *path, in bch2_btree_node_lock_write() argument
331 return __btree_node_lock_write(trans, path, b, false); in bch2_btree_node_lock_write()
345 struct btree_path *path, unsigned long trace_ip) in bch2_btree_path_relock() argument
347 return btree_node_locked(path, path->level) in bch2_btree_path_relock()
349 : __bch2_btree_path_relock(trans, path, trace_ip); in bch2_btree_path_relock()
355 struct btree_path *path, unsigned level) in bch2_btree_node_relock() argument
357 EBUG_ON(btree_node_locked(path, level) && in bch2_btree_node_relock()
358 !btree_node_write_locked(path, level) && in bch2_btree_node_relock()
359 btree_node_locked_type(path, level) != __btree_lock_want(path, level)); in bch2_btree_node_relock()
361 return likely(btree_node_locked(path, level)) || in bch2_btree_node_relock()
362 (!IS_ERR_OR_NULL(path->l[level].b) && in bch2_btree_node_relock()
363 __bch2_btree_node_relock(trans, path, level, true)); in bch2_btree_node_relock()
367 struct btree_path *path, unsigned level) in bch2_btree_node_relock_notrace() argument
369 EBUG_ON(btree_node_locked(path, level) && in bch2_btree_node_relock_notrace()
370 !btree_node_write_locked(path, level) && in bch2_btree_node_relock_notrace()
371 btree_node_locked_type(path, level) != __btree_lock_want(path, level)); in bch2_btree_node_relock_notrace()
373 return likely(btree_node_locked(path, level)) || in bch2_btree_node_relock_notrace()
374 (!IS_ERR_OR_NULL(path->l[level].b) && in bch2_btree_node_relock_notrace()
375 __bch2_btree_node_relock(trans, path, level, false)); in bch2_btree_node_relock_notrace()
389 struct btree_path *path, in bch2_btree_path_upgrade() argument
393 unsigned old_locks_want = path->locks_want; in bch2_btree_path_upgrade()
397 if (path->locks_want < new_locks_want in bch2_btree_path_upgrade()
398 ? __bch2_btree_path_upgrade(trans, path, new_locks_want, &f) in bch2_btree_path_upgrade()
399 : path->nodes_locked) in bch2_btree_path_upgrade()
402 trace_and_count(trans->c, trans_restart_upgrade, trans, _THIS_IP_, path, in bch2_btree_path_upgrade()
409 …tic inline void btree_path_set_should_be_locked(struct btree_trans *trans, struct btree_path *path) in btree_path_set_should_be_locked() argument
411 EBUG_ON(!btree_node_locked(path, path->level)); in btree_path_set_should_be_locked()
412 EBUG_ON(path->uptodate); in btree_path_set_should_be_locked()
414 path->should_be_locked = true; in btree_path_set_should_be_locked()
415 trace_btree_path_should_be_locked(trans, path); in btree_path_set_should_be_locked()
419 struct btree_path *path, in __btree_path_set_level_up() argument
422 btree_node_unlock(trans, path, l); in __btree_path_set_level_up()
423 path->l[l].b = ERR_PTR(-BCH_ERR_no_btree_node_up); in __btree_path_set_level_up()
427 struct btree_path *path) in btree_path_set_level_up() argument
429 __btree_path_set_level_up(trans, path, path->level++); in btree_path_set_level_up()
430 btree_path_set_dirty(path, BTREE_ITER_NEED_TRAVERSE); in btree_path_set_level_up()
446 static inline void bch2_btree_path_verify_locks(struct btree_path *path) {} in bch2_btree_path_verify_locks() argument