| /linux/include/linux/ |
| H A D | rbtree.h | 138 bool leftmost) in rb_insert_color_cached() argument 140 if (leftmost) in rb_insert_color_cached() 149 struct rb_node *leftmost = NULL; in rb_erase_cached() local 152 leftmost = root->rb_leftmost = rb_next(node); in rb_erase_cached() 156 return leftmost; in rb_erase_cached() 198 bool leftmost = true; in rb_add_cached() local 206 leftmost = false; in rb_add_cached() 211 rb_insert_color_cached(node, tree, leftmost); in rb_add_cached() 213 return leftmost ? node : NULL; in rb_add_cached() 254 bool leftmost = true; in rb_find_add_cached() local [all …]
|
| H A D | timerqueue.h | 25 struct rb_node *leftmost = rb_first_cached(&head->rb_root); in timerqueue_getnext() local 27 return rb_entry_safe(leftmost, struct timerqueue_node, node); in timerqueue_getnext()
|
| H A D | rbtree_augmented.h | 70 bool leftmost = true; in rb_add_augmented_cached() local 78 leftmost = false; in rb_add_augmented_cached() 84 rb_insert_augmented_cached(node, tree, leftmost, augment); in rb_add_augmented_cached() 86 return leftmost ? node : NULL; in rb_add_augmented_cached()
|
| /linux/tools/perf/util/ |
| H A D | rblist.c | 17 bool leftmost = true; in rblist__add_node() local 29 leftmost = false; in rblist__add_node() 40 rb_insert_color_cached(new_node, &rblist->entries, leftmost); in rblist__add_node() 59 bool leftmost = true; in __rblist__findnew() local 71 leftmost = false; in __rblist__findnew() 82 &rblist->entries, leftmost); in __rblist__findnew()
|
| H A D | srcline.c | 268 bool leftmost = true; in srcline__tree_insert() local 286 leftmost = false; in srcline__tree_insert() 290 rb_insert_color_cached(&node->rb_node, tree, leftmost); in srcline__tree_insert() 361 bool leftmost = true; in inlines__tree_insert() local 370 leftmost = false; in inlines__tree_insert() 374 rb_insert_color_cached(&inlines->rb_node, tree, leftmost); in inlines__tree_insert()
|
| H A D | hist.c | 690 bool leftmost = true; in hists__findnew_entry() local 739 leftmost = false; in hists__findnew_entry() 752 rb_insert_color_cached(&he->rb_node_in, hists->entries_in, leftmost); in hists__findnew_entry() 1634 bool leftmost = true; in hierarchy_insert_entry() local 1650 leftmost = false; in hierarchy_insert_entry() 1690 rb_insert_color_cached(&new->rb_node_in, root, leftmost); in hierarchy_insert_entry() 1754 bool leftmost = true; in hists__collapse_insert_entry() local 1792 leftmost = false; in hists__collapse_insert_entry() 1798 rb_insert_color_cached(&he->rb_node_in, root, leftmost); in hists__collapse_insert_entry() 1937 bool leftmost = true; in hierarchy_insert_output_entry() local [all …]
|
| H A D | symbol.c | 378 bool leftmost = true; in __symbols__insert() local 398 leftmost = false; in __symbols__insert() 402 rb_insert_color_cached(&sym->rb_node, symbols, leftmost); in __symbols__insert()
|
| H A D | machine.c | 250 bool leftmost = true; in machines__add() local 267 leftmost = false; in machines__add() 272 rb_insert_color_cached(&machine->rb_node, &machines->guests, leftmost); in machines__add()
|
| /linux/tools/include/linux/ |
| H A D | rbtree.h | 131 bool leftmost) in rb_insert_color_cached() argument 133 if (leftmost) in rb_insert_color_cached() 183 bool leftmost = true; in rb_add_cached() local 191 leftmost = false; in rb_add_cached() 196 rb_insert_color_cached(node, tree, leftmost); in rb_add_cached()
|
| /linux/fs/f2fs/ |
| H A D | extent_cache.c | 223 bool *leftmost) in __lookup_extent_node_ret() argument 240 *leftmost = true; in __lookup_extent_node_ret() 250 *leftmost = false; in __lookup_extent_node_ret() 293 bool leftmost) in __attach_extent_node() argument 307 rb_insert_color_cached(&en->rb_node, &et->root, leftmost); in __attach_extent_node() 581 bool leftmost) in __insert_extent_tree() argument 594 leftmost = true; in __insert_extent_tree() 605 leftmost = false; in __insert_extent_tree() 618 en = __attach_extent_node(sbi, et, ei, parent, p, leftmost); in __insert_extent_tree() 668 bool leftmost = false; in __update_extent_tree_range() local [all …]
|
| /linux/kernel/bpf/ |
| H A D | range_tree.c | 90 bool leftmost = true; in __range_size_insert() 98 leftmost = false; in __range_size_insert() 103 rb_insert_color_cached(&rn->rb_range_size, root, leftmost); in __range_size_insert() 91 bool leftmost = true; __range_size_insert() local
|
| /linux/Documentation/translations/zh_CN/dev-tools/ |
| H A D | gdb-kernel-debugging.rst | 130 (gdb) set $leftmost = $lx_per_cpu(hrtimer_bases).clock_base[0].active.rb_root.rb_leftmost 131 (gdb) p *$container_of($leftmost, "struct hrtimer", "node")
|
| /linux/Documentation/translations/zh_TW/dev-tools/ |
| H A D | gdb-kernel-debugging.rst | 126 (gdb) set $leftmost = $lx_per_cpu(hrtimer_bases).clock_base[0].active.rb_root.rb_leftmost 127 (gdb) p *$container_of($leftmost, "struct hrtimer", "node")
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | pixfmt-intro.rst | 30 leftmost pixel of the topmost row. Following that is the pixel 35 leftmost pixel of the second row from the top, and so on. The last row
|
| H A D | pixfmt-v4l2-mplane.rst | 44 - Distance in bytes between the leftmost pixels in two adjacent
|
| H A D | vidioc-g-fbuf.rst | 155 - Distance in bytes between the leftmost pixels in two adjacent
|
| /linux/net/sched/ |
| H A D | sch_etf.c | 168 bool leftmost = true; in etf_enqueue_timesortedlist() local 183 leftmost = false; in etf_enqueue_timesortedlist() 189 rb_insert_color_cached(&nskb->rbnode, &q->head, leftmost); in etf_enqueue_timesortedlist()
|
| /linux/kernel/sched/ |
| H A D | deadline.c | 580 struct rb_node *leftmost; in enqueue_pushable_dl_task() local 584 leftmost = rb_add_cached(&p->pushable_dl_tasks, in enqueue_pushable_dl_task() 587 if (leftmost) in enqueue_pushable_dl_task() 600 struct rb_node *leftmost; in dequeue_pushable_dl_task() local 605 leftmost = rb_erase_cached(&p->pushable_dl_tasks, root); in dequeue_pushable_dl_task() 606 if (leftmost) in dequeue_pushable_dl_task() 607 dl_rq->earliest_dl.next = __node_2_pdl(leftmost)->dl.deadline; in dequeue_pushable_dl_task() 2028 struct rb_node *leftmost = rb_first_cached(&dl_rq->root); in dec_dl_deadline() local 2029 struct sched_dl_entity *entry = __node_2_dle(leftmost); in dec_dl_deadline()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | rbtree.rst | 281 * Iterate to find the leftmost such node N. 293 return node; /* node is leftmost match */
|
| /linux/Documentation/process/debugging/ |
| H A D | gdb-kernel-debugging.rst | 137 (gdb) set $leftmost = $lx_per_cpu(hrtimer_bases).clock_base[0].active.rb_root.rb_leftmost 138 (gdb) p *$container_of($leftmost, "struct hrtimer", "node")
|
| /linux/Documentation/core-api/ |
| H A D | rbtree.rst | 199 Computing the leftmost (smallest) node is quite a common task for binary 212 leftmost node. This allows rb_root_cached to exist wherever rb_root does, 319 * Iterate to find the leftmost such node N. 331 return node; /* node is leftmost match */
|
| /linux/Documentation/bpf/ |
| H A D | map_lpm_trie.rst | 105 from leftmost leaf first. This means that iteration will return more
|
| /linux/Documentation/RCU/Design/Memory-Ordering/ |
| H A D | Tree-RCU-Memory-Ordering.rst | 360 The diagram below shows the path of ordering if the leftmost 363 leftmost ``rcu_node`` structure offlines its last CPU and if the next 547 | The leftmost drive to root stopped before it reached the root | 624 running on a CPU corresponding to the leftmost leaf ``rcu_node``
|
| /linux/tools/perf/ |
| H A D | builtin-sched.c | 1009 bool leftmost = true; in __thread_latency_insert() local 1024 leftmost = false; in __thread_latency_insert() 1029 rb_insert_color_cached(&data->node, root, leftmost); in __thread_latency_insert() 3420 bool leftmost = true; in __merge_work_atoms() local 3434 leftmost = false; in __merge_work_atoms() 3453 rb_insert_color_cached(&data->node, root, leftmost); in __merge_work_atoms()
|
| H A D | builtin-kwork.c | 338 bool leftmost = true; in work_insert() local 351 leftmost = false; in work_insert() 356 rb_insert_color_cached(&key->node, root, leftmost); in work_insert()
|