| /linux/tools/perf/util/ |
| A D | rblist.c | 112 struct rb_node *pos, *next = rb_first_cached(&rblist->entries); in rblist__exit() 133 for (node = rb_first_cached(&rblist->entries); node; in rblist__entry()
|
| A D | hist.c | 256 struct rb_node *next = rb_first_cached(&hists->entries); in hists__output_recalc_col_len() 1703 next = rb_first_cached(root); in hists__collapse_resort() 1782 node = rb_first_cached(&hists->entries); in hierarchy_recalc_total_periods() 1844 node = rb_first_cached(root_in); in hists__hierarchy_output_resort() 1967 next = rb_first_cached(root); in output_resort() 2046 node = rb_first_cached(&he->hroot_out); in __rb_hierarchy_next() 2084 node = rb_first_cached(&he->hroot_out); in hist_entry__has_hierarchy_children() 2240 nd = rb_first_cached(&he->hroot_out); in resort_filtered_entry() 2263 nd = rb_first_cached(&hists->entries); in hists__filter_hierarchy() 2307 nd = rb_first_cached(&hists->entries); in hists__filter_hierarchy() [all …]
|
| A D | intlist.h | 48 struct rb_node *rn = rb_first_cached(&ilist->rblist.entries); in intlist__first()
|
| A D | strlist.h | 60 struct rb_node *rn = rb_first_cached(&slist->rblist.entries); in strlist__first()
|
| A D | symbol.c | 200 nd = rb_first_cached(symbols); in symbols__fixup_duplicate() 232 struct rb_node *nd, *prevnd = rb_first_cached(symbols); in symbols__fixup_end() 335 struct rb_node *next = rb_first_cached(symbols); in symbols__delete() 409 struct rb_node *n = rb_first_cached(symbols); in symbols__first() 451 for (nd = rb_first_cached(source); nd; nd = rb_next(nd)) in symbols__sort_by_name() 458 for (nd = rb_first_cached(source); nd; nd = rb_next(nd)) { in symbols__sort_by_name() 794 struct rb_node *next = rb_first_cached(root); in maps__split_kallsyms_for_kcore() 851 struct rb_node *next = rb_first_cached(root); in maps__split_kallsyms()
|
| A D | build-id.c | 359 for (nd = rb_first_cached(&session->machines.guests); nd; in perf_session__write_buildid_table() 907 for (nd = rb_first_cached(&machines->guests); nd; in machines__for_each_dso() 943 for (nd = rb_first_cached(&session->machines.guests); nd; in perf_session__read_build_ids()
|
| A D | symbol.h | 85 for (nd = rb_first_cached(symbols); \
|
| A D | stream.c | 115 struct rb_node *next = rb_first_cached(&hists->entries); in init_hot_callchain()
|
| /linux/tools/perf/tests/ |
| A D | hists_output.c | 110 node = rb_first_cached(root_out); in del_hist_entries() 179 node = rb_first_cached(root); in test1() 279 node = rb_first_cached(root); in test2() 333 node = rb_first_cached(root); in test3() 411 node = rb_first_cached(root); in test4() 514 node = rb_first_cached(root); in test5()
|
| A D | hists_common.c | 175 node = rb_first_cached(root); in print_hists_in() 204 node = rb_first_cached(root); in print_hists_out()
|
| A D | hists_link.c | 174 node = rb_first_cached(root); in __validate_match() 226 node = rb_first_cached(root); in __validate_link()
|
| A D | symbols.c | 119 for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) { in test_dso()
|
| /linux/kernel/locking/ |
| A D | rtmutex_common.h | 123 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_waiter_is_top_waiter() 130 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_top_waiter()
|
| /linux/fs/btrfs/tests/ |
| A D | free-space-tests.c | 859 for (node = rb_first_cached(&ctl->free_space_bytes), i = 9; node; in test_bytes_index() 882 for (node = rb_first_cached(&ctl->free_space_bytes), i = 1; node; in test_bytes_index() 939 entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), in test_bytes_index() 964 entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), in test_bytes_index() 979 entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), in test_bytes_index()
|
| /linux/include/linux/ |
| A D | timerqueue.h | 25 struct rb_node *leftmost = rb_first_cached(&head->rb_root); in timerqueue_getnext()
|
| A D | rbtree.h | 106 #define rb_first_cached(root) (root)->rb_leftmost macro
|
| /linux/drivers/gpu/drm/i915/ |
| A D | i915_scheduler.c | 47 GEM_BUG_ON(rb_first_cached(&sched_engine->queue) != in assert_priolists() 51 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { in assert_priolists()
|
| /linux/net/sched/ |
| A D | sch_etf.c | 113 p = rb_first_cached(&q->head); in etf_peek_timesortedlist() 420 struct rb_node *p = rb_first_cached(&q->head); in timesortedlist_clear()
|
| /linux/tools/perf/ |
| A D | builtin-kwork.c | 552 next = rb_first_cached(root); in find_work_by_id() 1691 node = rb_first_cached(root); in work_sort() 1920 next = rb_first_cached(&kwork->sorted_work_root); in perf_kwork__report() 2006 next = rb_first_cached(&class->work_root); in top_calc_total_runtime() 2078 next = rb_first_cached(&class->work_root); in top_calc_cpu_usage() 2122 node = rb_first_cached(&class->work_root); in top_merge_tasks() 2157 next = rb_first_cached(&kwork->sorted_work_root); in perf_kwork__top_report()
|
| A D | builtin-diff.c | 537 next = rb_first_cached(root); in hists__baseline_only() 588 struct rb_node *next = rb_first_cached(root); in get_block_pair() 655 struct rb_node *next = rb_first_cached(root); in block_hists_match() 681 next = rb_first_cached(root); in hists__precompute()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| A D | intel_execlists_submission.c | 283 rb = rb_first_cached(&sched_engine->queue); in queue_prio() 292 struct rb_node *rb = rb_first_cached(&el->virtual); in virtual_prio() 1028 struct rb_node *rb = rb_first_cached(&el->virtual); in first_virtual_engine() 1039 rb = rb_first_cached(&el->virtual); in first_virtual_engine() 1491 while ((rb = rb_first_cached(&sched_engine->queue))) { in execlists_dequeue() 3182 while ((rb = rb_first_cached(&sched_engine->queue))) { in execlists_reset_cancel() 3201 while ((rb = rb_first_cached(&execlists->virtual))) { in execlists_reset_cancel() 3851 first = rb_first_cached(&sibling->execlists.virtual) == in virtual_submission_tasklet() 4116 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { in intel_execlists_show_requests() 4137 for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) { in intel_execlists_show_requests()
|
| /linux/fs/ |
| A D | eventpoll.c | 882 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_clear_and_put() 897 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = next) { in ep_clear_and_put() 1049 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_show_fdinfo() 1175 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_find_tfd() 2110 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_loop_check_proc()
|
| /linux/tools/include/linux/ |
| A D | rbtree.h | 127 #define rb_first_cached(root) (root)->rb_leftmost macro
|
| /linux/tools/perf/ui/gtk/ |
| A D | hists.c | 358 for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) { in perf_gtk__show_hists() 420 for (node = rb_first_cached(root); node; node = rb_next(node)) { in perf_gtk__add_hierarchy_entries()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| A D | mmu_rb.c | 100 while ((node = rb_first_cached(&handler->root))) { in hfi1_mmu_rb_unregister()
|