Lines Matching refs:key_tmp

240 	struct pcache_cache_key *key_tmp, *key;  in cache_subtree_walk()  local
251 key_tmp = CACHE_KEY(node_tmp); in cache_subtree_walk()
257 if (cache_key_lend(key_tmp) <= cache_key_lstart(key)) { in cache_subtree_walk()
259 ret = ctx->after(key, key_tmp, ctx); in cache_subtree_walk()
271 if (cache_key_lstart(key_tmp) >= cache_key_lend(key)) { in cache_subtree_walk()
273 ret = ctx->before(key, key_tmp, ctx); in cache_subtree_walk()
281 if (cache_key_lstart(key_tmp) >= cache_key_lstart(key)) { in cache_subtree_walk()
287 if (cache_key_lend(key_tmp) >= cache_key_lend(key)) { in cache_subtree_walk()
289 ret = ctx->overlap_tail(key, key_tmp, ctx); in cache_subtree_walk()
302 ret = ctx->overlap_contain(key, key_tmp, ctx); in cache_subtree_walk()
315 if (cache_key_lend(key_tmp) > cache_key_lend(key)) { in cache_subtree_walk()
317 ret = ctx->overlap_contained(key, key_tmp, ctx); in cache_subtree_walk()
330 ret = ctx->overlap_head(key, key_tmp, ctx); in cache_subtree_walk()
365 struct pcache_cache_key *key_tmp; in cache_subtree_search() local
370 key_tmp = container_of(*new, struct pcache_cache_key, rb_node); in cache_subtree_search()
371 if (cache_key_invalid(key_tmp)) in cache_subtree_search()
372 list_add(&key_tmp->list_node, delete_key_list); in cache_subtree_search()
375 if (key_tmp->off >= key->off) { in cache_subtree_search()
420 struct pcache_cache_key *key_tmp, in fixup_overlap_tail() argument
428 if (cache_key_empty(key_tmp)) { in fixup_overlap_tail()
429 cache_key_delete(key_tmp); in fixup_overlap_tail()
433 cache_key_cutfront(key_tmp, cache_key_lend(key) - cache_key_lstart(key_tmp)); in fixup_overlap_tail()
434 if (key_tmp->len == 0) { in fixup_overlap_tail()
435 cache_key_delete(key_tmp); in fixup_overlap_tail()
454 struct pcache_cache_key *key_tmp, in fixup_overlap_contain() argument
462 cache_key_delete(key_tmp); in fixup_overlap_contain()
482 struct pcache_cache_key *key_tmp, struct pcache_cache_subtree_walk_ctx *ctx) in fixup_overlap_contained() argument
491 if (cache_key_empty(key_tmp)) { in fixup_overlap_contained()
495 cache_key_cutback(key_tmp, cache_key_lend(key_tmp) - cache_key_lstart(key)); in fixup_overlap_contained()
496 if (key_tmp->len == 0) { in fixup_overlap_contained()
497 cache_key_delete(key_tmp); in fixup_overlap_contained()
508 cache_key_copy(key_fixup, key_tmp); in fixup_overlap_contained()
511 cache_key_cutback(key_tmp, cache_key_lend(key_tmp) - cache_key_lstart(key)); in fixup_overlap_contained()
512 if (key_tmp->len == 0) { in fixup_overlap_contained()
513 cache_key_delete(key_tmp); in fixup_overlap_contained()
518 cache_key_cutfront(key_fixup, cache_key_lend(key) - cache_key_lstart(key_tmp)); in fixup_overlap_contained()
547 struct pcache_cache_key *key_tmp, struct pcache_cache_subtree_walk_ctx *ctx) in fixup_overlap_head() argument
555 cache_key_cutback(key_tmp, cache_key_lend(key_tmp) - cache_key_lstart(key)); in fixup_overlap_head()
556 if (key_tmp->len == 0) { in fixup_overlap_head()
558 cache_key_delete(key_tmp); in fixup_overlap_head()
581 struct pcache_cache_key *key_tmp = NULL, *key_next; in cache_key_insert() local
592 list_for_each_entry_safe(key_tmp, key_next, &delete_key_list, list_node) { in cache_key_insert()
593 list_del_init(&key_tmp->list_node); in cache_key_insert()
594 cache_key_delete(key_tmp); in cache_key_insert()