Lines Matching refs:kwork_work

61 	int             (*cmp)(struct kwork_work *l, struct kwork_work *r);
65 static int id_cmp(struct kwork_work *l, struct kwork_work *r) in id_cmp()
80 static int count_cmp(struct kwork_work *l, struct kwork_work *r) in count_cmp()
90 static int runtime_cmp(struct kwork_work *l, struct kwork_work *r) in runtime_cmp()
100 static int max_runtime_cmp(struct kwork_work *l, struct kwork_work *r) in max_runtime_cmp()
110 static int avg_latency_cmp(struct kwork_work *l, struct kwork_work *r) in avg_latency_cmp()
130 static int max_latency_cmp(struct kwork_work *l, struct kwork_work *r) in max_latency_cmp()
140 static int cpu_usage_cmp(struct kwork_work *l, struct kwork_work *r) in cpu_usage_cmp()
150 static int id_or_cpu_r_cmp(struct kwork_work *l, struct kwork_work *r) in id_or_cpu_r_cmp()
294 struct kwork_work *l, struct kwork_work *r) in work_cmp()
310 static struct kwork_work *work_search(struct rb_root_cached *root, in work_search()
311 struct kwork_work *key, in work_search()
315 struct kwork_work *work; in work_search()
319 work = container_of(node, struct kwork_work, node); in work_search()
335 struct kwork_work *key, struct list_head *sort_list) in work_insert()
339 struct kwork_work *cur; in work_insert()
343 cur = container_of(*new, struct kwork_work, node); in work_insert()
359 static struct kwork_work *work_new(struct kwork_work *key) in work_new()
362 struct kwork_work *work = zalloc(sizeof(*work)); in work_new()
379 static struct kwork_work *work_findnew(struct rb_root_cached *root, in work_findnew()
380 struct kwork_work *key, in work_findnew()
383 struct kwork_work *work = work_search(root, key, sort_list); in work_findnew()
409 struct kwork_work *work) in profile_name_match()
420 struct kwork_work *work, in profile_event_match()
454 struct kwork_work **ret_work, in work_push_atom()
458 struct kwork_work *work, key; in work_push_atom()
513 struct kwork_work **ret_work) in work_pop_atom()
516 struct kwork_work *work, key; in work_pop_atom()
547 static struct kwork_work *find_work_by_id(struct rb_root_cached *root, in find_work_by_id()
551 struct kwork_work *work; in find_work_by_id()
555 work = rb_entry(next, struct kwork_work, node); in find_work_by_id()
579 static void report_update_exit_event(struct kwork_work *work, in report_update_exit_event()
618 struct kwork_work *work = NULL; in report_exit_event()
634 static void latency_update_entry_event(struct kwork_work *work, in latency_update_entry_event()
673 struct kwork_work *work = NULL; in latency_entry_event()
739 struct kwork_work *work, in timehist_print_event()
832 struct kwork_work *work = NULL; in timehist_entry_event()
853 struct kwork_work *work = NULL; in timehist_exit_event()
883 static void top_update_runtime(struct kwork_work *work, in top_update_runtime()
914 struct kwork_work *work, *sched_work; in top_exit_event()
945 struct kwork_work *work; in top_sched_switch_event()
1007 struct kwork_work *work, in irq_work_init()
1025 static void irq_work_name(struct kwork_work *work, char *buf, int len) in irq_work_name()
1135 struct kwork_work *work, in softirq_work_init()
1156 static void softirq_work_name(struct kwork_work *work, char *buf, int len) in softirq_work_name()
1235 struct kwork_work *work, in workqueue_work_init()
1252 static void workqueue_work_name(struct kwork_work *work, char *buf, int len) in workqueue_work_name()
1303 struct kwork_work *work, in sched_work_init()
1321 static void sched_work_name(struct kwork_work *work, char *buf, int len) in sched_work_name()
1348 static int report_print_work(struct perf_kwork *kwork, struct kwork_work *work) in report_print_work()
1639 static int top_print_work(struct perf_kwork *kwork __maybe_unused, struct kwork_work *work) in top_print_work()
1689 struct kwork_work *data; in work_sort()
1698 data = rb_entry(node, struct kwork_work, node); in work_sort()
1839 struct kwork_work *work) in process_skipped_events()
1851 static struct kwork_work *perf_kwork_add_work(struct perf_kwork *kwork, in perf_kwork_add_work()
1853 struct kwork_work *key) in perf_kwork_add_work()
1855 struct kwork_work *work = NULL; in perf_kwork_add_work()
1907 struct kwork_work *work; in perf_kwork__report()
1924 work = rb_entry(next, struct kwork_work, node); in perf_kwork__report()
2000 struct kwork_work *work; in top_calc_total_runtime()
2010 work = rb_entry(next, struct kwork_work, node); in top_calc_total_runtime()
2019 struct kwork_work *work) in top_calc_idle_time()
2031 struct kwork_work *work) in top_calc_irq_runtime()
2045 struct kwork_work *work) in top_subtract_irq_runtime()
2048 struct kwork_work *data; in top_subtract_irq_runtime()
2072 struct kwork_work *work; in top_calc_cpu_usage()
2082 work = rb_entry(next, struct kwork_work, node); in top_calc_cpu_usage()
2101 struct kwork_work *work) in top_calc_load_runtime()
2113 struct kwork_work *merged_work, *data; in top_merge_tasks()
2129 data = rb_entry(node, struct kwork_work, node); in top_merge_tasks()
2152 struct kwork_work *work; in perf_kwork__top_report()
2161 work = rb_entry(next, struct kwork_work, node); in perf_kwork__top_report()