Lines Matching refs:kwork_work
60 int (*cmp)(struct kwork_work *l, struct kwork_work *r);
64 static int id_cmp(struct kwork_work *l, struct kwork_work *r) in id_cmp()
79 static int count_cmp(struct kwork_work *l, struct kwork_work *r) in count_cmp()
89 static int runtime_cmp(struct kwork_work *l, struct kwork_work *r) in runtime_cmp()
99 static int max_runtime_cmp(struct kwork_work *l, struct kwork_work *r) in max_runtime_cmp()
109 static int avg_latency_cmp(struct kwork_work *l, struct kwork_work *r) in avg_latency_cmp()
129 static int max_latency_cmp(struct kwork_work *l, struct kwork_work *r) in max_latency_cmp()
139 static int cpu_usage_cmp(struct kwork_work *l, struct kwork_work *r) in cpu_usage_cmp()
149 static int id_or_cpu_r_cmp(struct kwork_work *l, struct kwork_work *r) in id_or_cpu_r_cmp()
293 struct kwork_work *l, struct kwork_work *r) in work_cmp()
309 static struct kwork_work *work_search(struct rb_root_cached *root, in work_search()
310 struct kwork_work *key, in work_search()
314 struct kwork_work *work; in work_search()
318 work = container_of(node, struct kwork_work, node); in work_search()
334 struct kwork_work *key, struct list_head *sort_list) in work_insert()
338 struct kwork_work *cur; in work_insert()
342 cur = container_of(*new, struct kwork_work, node); in work_insert()
358 static struct kwork_work *work_new(struct kwork_work *key) in work_new()
361 struct kwork_work *work = zalloc(sizeof(*work)); in work_new()
378 static struct kwork_work *work_findnew(struct rb_root_cached *root, in work_findnew()
379 struct kwork_work *key, in work_findnew()
382 struct kwork_work *work = work_search(root, key, sort_list); in work_findnew()
408 struct kwork_work *work) in profile_name_match()
419 struct kwork_work *work, in profile_event_match()
453 struct kwork_work **ret_work, in work_push_atom()
457 struct kwork_work *work, key; in work_push_atom()
512 struct kwork_work **ret_work) in work_pop_atom()
515 struct kwork_work *work, key; in work_pop_atom()
546 static struct kwork_work *find_work_by_id(struct rb_root_cached *root, in find_work_by_id()
550 struct kwork_work *work; in find_work_by_id()
554 work = rb_entry(next, struct kwork_work, node); in find_work_by_id()
578 static void report_update_exit_event(struct kwork_work *work, in report_update_exit_event()
617 struct kwork_work *work = NULL; in report_exit_event()
633 static void latency_update_entry_event(struct kwork_work *work, in latency_update_entry_event()
672 struct kwork_work *work = NULL; in latency_entry_event()
738 struct kwork_work *work, in timehist_print_event()
831 struct kwork_work *work = NULL; in timehist_entry_event()
852 struct kwork_work *work = NULL; in timehist_exit_event()
882 static void top_update_runtime(struct kwork_work *work, in top_update_runtime()
913 struct kwork_work *work, *sched_work; in top_exit_event()
944 struct kwork_work *work; in top_sched_switch_event()
1006 struct kwork_work *work, in irq_work_init()
1024 static void irq_work_name(struct kwork_work *work, char *buf, int len) in irq_work_name()
1133 struct kwork_work *work, in softirq_work_init()
1154 static void softirq_work_name(struct kwork_work *work, char *buf, int len) in softirq_work_name()
1233 struct kwork_work *work, in workqueue_work_init()
1250 static void workqueue_work_name(struct kwork_work *work, char *buf, int len) in workqueue_work_name()
1301 struct kwork_work *work, in sched_work_init()
1319 static void sched_work_name(struct kwork_work *work, char *buf, int len) in sched_work_name()
1346 static int report_print_work(struct perf_kwork *kwork, struct kwork_work *work) in report_print_work()
1637 static int top_print_work(struct perf_kwork *kwork __maybe_unused, struct kwork_work *work) in top_print_work()
1687 struct kwork_work *data; in work_sort()
1696 data = rb_entry(node, struct kwork_work, node); in work_sort()
1837 struct kwork_work *work) in process_skipped_events()
1849 struct kwork_work *perf_kwork_add_work(struct perf_kwork *kwork, in perf_kwork_add_work()
1851 struct kwork_work *key) in perf_kwork_add_work()
1853 struct kwork_work *work = NULL; in perf_kwork_add_work()
1905 struct kwork_work *work; in perf_kwork__report()
1922 work = rb_entry(next, struct kwork_work, node); in perf_kwork__report()
1998 struct kwork_work *work; in top_calc_total_runtime()
2008 work = rb_entry(next, struct kwork_work, node); in top_calc_total_runtime()
2017 struct kwork_work *work) in top_calc_idle_time()
2029 struct kwork_work *work) in top_calc_irq_runtime()
2043 struct kwork_work *work) in top_subtract_irq_runtime()
2046 struct kwork_work *data; in top_subtract_irq_runtime()
2070 struct kwork_work *work; in top_calc_cpu_usage()
2080 work = rb_entry(next, struct kwork_work, node); in top_calc_cpu_usage()
2099 struct kwork_work *work) in top_calc_load_runtime()
2111 struct kwork_work *merged_work, *data; in top_merge_tasks()
2127 data = rb_entry(node, struct kwork_work, node); in top_merge_tasks()
2150 struct kwork_work *work; in perf_kwork__top_report()
2159 work = rb_entry(next, struct kwork_work, node); in perf_kwork__top_report()