/linux/tools/perf/util/ |
H A D | rblist.c | 112 struct rb_node *pos, *next = rb_first_cached(&rblist->entries); in rblist__exit() 133 for (node = rb_first_cached(&rblist->entries); node; in rblist__entry()
|
H A D | intlist.h | 48 struct rb_node *rn = rb_first_cached(&ilist->rblist.entries); in intlist__first()
|
H A D | strlist.h | 60 struct rb_node *rn = rb_first_cached(&slist->rblist.entries); in strlist__first()
|
H A D | build-id.c | 359 for (nd = rb_first_cached(&session->machines.guests); nd; in perf_session__write_buildid_table() 907 for (nd = rb_first_cached(&machines->guests); nd; in machines__for_each_dso() 943 for (nd = rb_first_cached(&session->machines.guests); nd; in perf_session__read_build_ids()
|
H A D | symbol.h | 85 for (nd = rb_first_cached(symbols); \
|
H A D | machine.c | 247 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__set_comm_exec() 415 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__process_guests() 428 for (node = rb_first_cached(&machines->guests); node; in machines__set_id_hdr_size() 862 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos() 882 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos_buildid() 1236 struct rb_node *next = rb_first_cached(&machines->guests); in machines__destroy_kernel_maps() 3002 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__for_each_thread()
|
H A D | stream.c | 114 struct rb_node *next = rb_first_cached(&hists->entries); in init_hot_callchain()
|
H A D | srcline.c | 1029 struct rb_node *next = rb_first_cached(tree); in srcline__tree_delete() 1113 struct rb_node *next = rb_first_cached(tree); in inlines__tree_delete()
|
/linux/tools/perf/tests/ |
H A D | hists_output.c | 110 node = rb_first_cached(root_out); in del_hist_entries() 179 node = rb_first_cached(root); in test1() 279 node = rb_first_cached(root); in test2() 333 node = rb_first_cached(root); in test3() 411 node = rb_first_cached(root); in test4() 514 node = rb_first_cached(root); in test5()
|
H A D | hists_common.c | 175 node = rb_first_cached(root); in print_hists_in() 204 node = rb_first_cached(root); in print_hists_out()
|
H A D | hists_link.c | 174 node = rb_first_cached(root); in __validate_match() 226 node = rb_first_cached(root); in __validate_link()
|
H A D | symbols.c | 119 for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) { in test_dso()
|
H A D | hists_cumulate.c | 145 node = rb_first_cached(root_out); in del_hist_entries()
|
/linux/fs/btrfs/tests/ |
H A D | free-space-tests.c | 859 for (node = rb_first_cached(&ctl->free_space_bytes), i = 9; node; in test_bytes_index() 882 for (node = rb_first_cached(&ctl->free_space_bytes), i = 1; node; in test_bytes_index() 939 entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), in test_bytes_index() 964 entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), in test_bytes_index() 979 entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), in test_bytes_index()
|
/linux/include/linux/ |
H A D | timerqueue.h | 25 struct rb_node *leftmost = rb_first_cached(&head->rb_root); in timerqueue_getnext()
|
H A D | rbtree.h | 106 #define rb_first_cached(root) (root)->rb_leftmost macro
|
/linux/drivers/gpu/drm/i915/ |
H A D | i915_scheduler.c | 47 GEM_BUG_ON(rb_first_cached(&sched_engine->queue) != in assert_priolists() 51 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { in assert_priolists()
|
/linux/net/sched/ |
H A D | sch_etf.c | 113 p = rb_first_cached(&q->head); in etf_peek_timesortedlist() 420 struct rb_node *p = rb_first_cached(&q->head); in timesortedlist_clear()
|
/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_execlists_submission.c | 283 rb = rb_first_cached(&sched_engine->queue); in queue_prio() 292 struct rb_node *rb = rb_first_cached(&el->virtual); in virtual_prio() 1019 struct rb_node *rb = rb_first_cached(&el->virtual); in first_virtual_engine() 1030 rb = rb_first_cached(&el->virtual); in first_virtual_engine() 1482 while ((rb = rb_first_cached(&sched_engine->queue))) { in execlists_dequeue() 3173 while ((rb = rb_first_cached(&sched_engine->queue))) { in execlists_reset_cancel() 3192 while ((rb = rb_first_cached(&execlists->virtual))) { in execlists_reset_cancel() 3842 first = rb_first_cached(&sibling->execlists.virtual) == in virtual_submission_tasklet() 4107 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { in intel_execlists_show_requests() 4128 for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) { in intel_execlists_show_requests()
|
/linux/tools/include/linux/ |
H A D | rbtree.h | 127 #define rb_first_cached(root) (root)->rb_leftmost macro
|
/linux/tools/perf/ui/gtk/ |
H A D | hists.c | 358 for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) { in perf_gtk__show_hists() 420 for (node = rb_first_cached(root); node; node = rb_next(node)) { in perf_gtk__add_hierarchy_entries()
|
/linux/fs/btrfs/ |
H A D | delayed-ref.c | 453 for (node = rb_first_cached(&head->ref_tree); node; in btrfs_merge_delayed_refs() 572 ref = rb_entry(rb_first_cached(&head->ref_tree), in btrfs_select_delayed_ref() 1273 while ((n = rb_first_cached(&head->ref_tree)) != NULL) { in btrfs_destroy_delayed_refs()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | mmu_rb.c | 100 while ((node = rb_first_cached(&handler->root))) { in hfi1_mmu_rb_unregister()
|
/linux/tools/perf/ui/browsers/ |
H A D | hists.c | 68 for (nd = rb_first_cached(&hists->entries); in hist_browser__get_folding() 286 node = rb_first_cached(&he->hroot_out); in hierarchy_count_rows() 573 for (nd = rb_first_cached(&he->hroot_out); nd; nd = rb_next(nd)) { in hierarchy_set_folding() 609 nd = rb_first_cached(&browser->hists->entries); in __hist_browser__set_folding() 1811 browser->top = rb_first_cached(&hb->hists->entries); in ui_browser__hists_init_top() 2910 struct rb_node *nd = rb_first_cached(&hb->hists->entries); in hist_browser__update_nr_entries() 2930 struct rb_node *nd = rb_first_cached(&hb->hists->entries); in hist_browser__update_percent_limit()
|
/linux/kernel/sched/ |
H A D | deadline.c | 536 return rb_first_cached(&dl_rq->root) == &dl_se->rb_node; in is_leftmost() 1878 struct rb_node *leftmost = rb_first_cached(&dl_rq->root); in dec_dl_deadline() 2412 struct rb_node *left = rb_first_cached(&dl_rq->root); in pick_next_dl_entity() 2522 next_node = rb_first_cached(&rq->dl.pushable_dl_tasks_root); in pick_earliest_pushable_dl_task() 2686 p = __node_2_pdl(rb_first_cached(&rq->dl.pushable_dl_tasks_root)); in pick_next_pushable_dl_task()
|