Lines Matching full:upper
3060 INIT_LIST_HEAD(&node->upper); in btrfs_backref_alloc_node()
3130 ASSERT(list_empty(&node->upper)); in btrfs_backref_drop_node()
3142 * upper edges and any uncached nodes in the path.
3155 while (!list_empty(&node->upper)) { in btrfs_backref_cleanup_node()
3156 edge = list_first_entry(&node->upper, struct btrfs_backref_edge, in btrfs_backref_cleanup_node()
3159 list_del(&edge->list[UPPER]); in btrfs_backref_cleanup_node()
3185 struct btrfs_backref_node *upper) in btrfs_backref_link_edge() argument
3187 ASSERT(upper && lower && upper->level == lower->level + 1); in btrfs_backref_link_edge()
3189 edge->node[UPPER] = upper; in btrfs_backref_link_edge()
3190 list_add_tail(&edge->list[LOWER], &lower->upper); in btrfs_backref_link_edge()
3209 struct btrfs_backref_node *upper; in handle_direct_tree_backref() local
3242 upper = btrfs_backref_alloc_node(cache, ref_key->offset, in handle_direct_tree_backref()
3244 if (!upper) { in handle_direct_tree_backref()
3250 * Backrefs for the upper level block isn't cached, add the in handle_direct_tree_backref()
3253 list_add_tail(&edge->list[UPPER], &cache->pending_edge); in handle_direct_tree_backref()
3256 upper = rb_entry(rb_node, struct btrfs_backref_node, rb_node); in handle_direct_tree_backref()
3257 ASSERT(upper->checked); in handle_direct_tree_backref()
3258 INIT_LIST_HEAD(&edge->list[UPPER]); in handle_direct_tree_backref()
3260 btrfs_backref_link_edge(edge, cur, upper); in handle_direct_tree_backref()
3285 struct btrfs_backref_node *upper; in handle_indirect_tree_backref() local
3380 upper = btrfs_backref_alloc_node(cache, eb->start, in handle_indirect_tree_backref()
3382 if (!upper) { in handle_indirect_tree_backref()
3388 upper->owner = btrfs_header_owner(eb); in handle_indirect_tree_backref()
3394 btrfs_backref_free_node(cache, upper); in handle_indirect_tree_backref()
3404 upper->checked = 0; in handle_indirect_tree_backref()
3406 upper->checked = 1; in handle_indirect_tree_backref()
3413 if (!upper->checked && need_check) { in handle_indirect_tree_backref()
3415 list_add_tail(&edge->list[UPPER], in handle_indirect_tree_backref()
3418 if (upper->checked) in handle_indirect_tree_backref()
3420 INIT_LIST_HEAD(&edge->list[UPPER]); in handle_indirect_tree_backref()
3423 upper = rb_entry(rb_node, struct btrfs_backref_node, in handle_indirect_tree_backref()
3425 ASSERT(upper->checked); in handle_indirect_tree_backref()
3426 INIT_LIST_HEAD(&edge->list[UPPER]); in handle_indirect_tree_backref()
3427 if (!upper->owner) in handle_indirect_tree_backref()
3428 upper->owner = btrfs_header_owner(eb); in handle_indirect_tree_backref()
3430 btrfs_backref_link_edge(edge, lower, upper); in handle_indirect_tree_backref()
3436 lower = upper; in handle_indirect_tree_backref()
3437 upper = NULL; in handle_indirect_tree_backref()
3447 * NOTE: Even if the function returned 0, @cur is not yet cached as its upper
3485 if (!list_empty(&cur->upper)) { in btrfs_backref_add_tree_node()
3490 ASSERT(list_is_singular(&cur->upper)); in btrfs_backref_add_tree_node()
3491 edge = list_first_entry(&cur->upper, struct btrfs_backref_edge, in btrfs_backref_add_tree_node()
3493 ASSERT(list_empty(&edge->list[UPPER])); in btrfs_backref_add_tree_node()
3494 exist = edge->node[UPPER]; in btrfs_backref_add_tree_node()
3496 * Add the upper level block to pending list if we need check in btrfs_backref_add_tree_node()
3500 list_add_tail(&edge->list[UPPER], &cache->pending_edge); in btrfs_backref_add_tree_node()
3597 list_for_each_entry(edge, &start->upper, list[LOWER]) in btrfs_backref_finish_upper_links()
3598 list_add_tail(&edge->list[UPPER], &pending_edge); in btrfs_backref_finish_upper_links()
3601 struct btrfs_backref_node *upper; in btrfs_backref_finish_upper_links() local
3605 struct btrfs_backref_edge, list[UPPER]); in btrfs_backref_finish_upper_links()
3606 list_del_init(&edge->list[UPPER]); in btrfs_backref_finish_upper_links()
3607 upper = edge->node[UPPER]; in btrfs_backref_finish_upper_links()
3611 if (upper->detached) { in btrfs_backref_finish_upper_links()
3616 if (list_empty(&lower->upper)) in btrfs_backref_finish_upper_links()
3624 * So if we have upper->rb_node populated, this means a cache in btrfs_backref_finish_upper_links()
3625 * hit. We only need to link the edge, as @upper and all its in btrfs_backref_finish_upper_links()
3628 if (!RB_EMPTY_NODE(&upper->rb_node)) { in btrfs_backref_finish_upper_links()
3629 list_add_tail(&edge->list[UPPER], &upper->lower); in btrfs_backref_finish_upper_links()
3634 if (unlikely(!upper->checked)) { in btrfs_backref_finish_upper_links()
3639 rb_node = rb_simple_insert(&cache->rb_root, &upper->simple_node); in btrfs_backref_finish_upper_links()
3641 btrfs_backref_panic(cache->fs_info, upper->bytenr, -EEXIST); in btrfs_backref_finish_upper_links()
3643 list_add_tail(&edge->list[UPPER], &upper->lower); in btrfs_backref_finish_upper_links()
3647 * to finish the upper linkage in btrfs_backref_finish_upper_links()
3649 list_for_each_entry(edge, &upper->upper, list[LOWER]) in btrfs_backref_finish_upper_links()
3650 list_add_tail(&edge->list[UPPER], &pending_edge); in btrfs_backref_finish_upper_links()
3659 struct btrfs_backref_node *upper; in btrfs_backref_error_cleanup() local
3669 struct btrfs_backref_edge, list[UPPER]); in btrfs_backref_error_cleanup()
3670 list_del(&edge->list[UPPER]); in btrfs_backref_error_cleanup()
3673 upper = edge->node[UPPER]; in btrfs_backref_error_cleanup()
3677 * Lower is no longer linked to any upper backref nodes and in btrfs_backref_error_cleanup()
3680 if (list_empty(&lower->upper) && in btrfs_backref_error_cleanup()
3684 if (!RB_EMPTY_NODE(&upper->rb_node)) in btrfs_backref_error_cleanup()
3687 /* Add this guy's upper edges to the list to process */ in btrfs_backref_error_cleanup()
3688 list_for_each_entry(edge, &upper->upper, list[LOWER]) in btrfs_backref_error_cleanup()
3689 list_add_tail(&edge->list[UPPER], in btrfs_backref_error_cleanup()
3691 if (list_empty(&upper->upper)) in btrfs_backref_error_cleanup()
3692 list_add(&upper->list, &cache->useless_node); in btrfs_backref_error_cleanup()