Searched refs:kwork (Results 1 – 6 of 6) sorted by relevance
| /linux/tools/perf/ ! |
| H A D | builtin-kwork.c | 168 static int sort_dimension__add(struct perf_kwork *kwork __maybe_unused, in sort_dimension__add() 210 if (kwork->report == KWORK_REPORT_LATENCY) in sort_dimension__add() 223 static void setup_sorting(struct perf_kwork *kwork, in setup_sorting() argument 227 char *tmp, *tok, *str = strdup(kwork->sort_order); in setup_sorting() 231 if (sort_dimension__add(kwork, tok, &kwork->sort_list) < 0) in setup_sorting() 236 pr_debug("Sort order: %s\n", kwork->sort_order); in setup_sorting() 240 static struct kwork_atom *atom_new(struct perf_kwork *kwork, in atom_new() argument 247 list_for_each_entry(page, &kwork->atom_page_list, list) { in atom_new() 267 list_add_tail(&page->list, &kwork->atom_page_list); in atom_new() 395 static void profile_update_timespan(struct perf_kwork *kwork, in profile_update_timespan() argument [all …]
|
| H A D | Build | 27 perf-$(CONFIG_LIBTRACEEVENT) += builtin-kwork.o
|
| /linux/tools/perf/util/ ! |
| H A D | bpf_kwork.c | 47 void (*load_prepare)(struct perf_kwork *kwork); 91 static void irq_load_prepare(struct perf_kwork *kwork) in irq_load_prepare() argument 93 if (kwork->report == KWORK_REPORT_RUNTIME) { in irq_load_prepare() 104 static void softirq_load_prepare(struct perf_kwork *kwork) in softirq_load_prepare() argument 106 if (kwork->report == KWORK_REPORT_RUNTIME) { in softirq_load_prepare() 109 } else if (kwork->report == KWORK_REPORT_LATENCY) { in softirq_load_prepare() 120 static void workqueue_load_prepare(struct perf_kwork *kwork) in workqueue_load_prepare() argument 122 if (kwork->report == KWORK_REPORT_RUNTIME) { in workqueue_load_prepare() 125 } else if (kwork->report == KWORK_REPORT_LATENCY) { in workqueue_load_prepare() 148 static int setup_filters(struct perf_kwork *kwork) in setup_filters() argument [all …]
|
| H A D | kwork.h | 156 void (*work_init)(struct perf_kwork *kwork, 169 int (*raise_event)(struct perf_kwork *kwork, 173 int (*entry_event)(struct perf_kwork *kwork, 177 int (*exit_event)(struct perf_kwork *kwork, 181 int (*sched_switch_event)(struct perf_kwork *kwork, 257 struct kwork_work *(*add_work)(struct perf_kwork *kwork, 265 int perf_kwork__trace_prepare_bpf(struct perf_kwork *kwork); 266 int perf_kwork__report_read_bpf(struct perf_kwork *kwork); 272 int perf_kwork__top_prepare_bpf(struct perf_kwork *kwork); 273 int perf_kwork__top_read_bpf(struct perf_kwork *kwork); [all …]
|
| H A D | bpf_kwork_top.c | 123 static int setup_filters(struct perf_kwork *kwork) in setup_filters() argument 125 if (kwork->cpu_list) { in setup_filters() 136 map = perf_cpu_map__new(kwork->cpu_list); in setup_filters() 159 int perf_kwork__top_prepare_bpf(struct perf_kwork *kwork) in perf_kwork__top_prepare_bpf() argument 179 list_for_each_entry(class, &kwork->class_list, list) { in perf_kwork__top_prepare_bpf() 194 if (kwork->cpu_list) in perf_kwork__top_prepare_bpf() 202 if (setup_filters(kwork)) in perf_kwork__top_prepare_bpf() 238 static int add_work(struct perf_kwork *kwork, struct work_key *key, in add_work() argument 258 work = kwork->add_work(kwork, tmp.class, &tmp); in add_work() 268 int perf_kwork__top_read_bpf(struct perf_kwork *kwork) in perf_kwork__top_read_bpf() argument [all …]
|
| /linux/tools/perf/Documentation/ ! |
| H A D | perf-kwork.txt | 1 perf-kwork(1) 6 perf-kwork - Tool to trace/measure kernel work properties (latencies) 11 'perf kwork' {record|report|latency|timehist|top} 15 There are several variants of 'perf kwork': 17 'perf kwork record <command>' to record the kernel work 20 'perf kwork report' to report the per kwork runtime. 22 'perf kwork latency' to report the per kwork latencies. 24 'perf kwork timehist' provides an analysis of kernel work events. 26 'perf kwork top' to report the task cpu usage. 29 perf kwork record -- sleep 1 [all …]
|