Lines Matching defs:ftrace

8 #include "util/ftrace.h"
20 int perf_ftrace__latency_prepare_bpf(struct perf_ftrace *ftrace)
26 if (!list_empty(&ftrace->filters)) {
27 if (!list_is_singular(&ftrace->filters)) {
31 func = list_first_entry(&ftrace->filters, struct filter_entry, list);
36 list_for_each(pos, &ftrace->event_pair)
51 skel->rodata->bucket_range = ftrace->bucket_range;
52 skel->rodata->min_latency = ftrace->min_latency;
53 skel->rodata->bucket_num = ftrace->bucket_num;
54 if (ftrace->bucket_range && ftrace->bucket_num) {
55 bpf_map__set_max_entries(skel->maps.latency, ftrace->bucket_num);
59 if (ftrace->target.cpu_list) {
60 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus);
65 if (target__has_task(&ftrace->target) || target__none(&ftrace->target)) {
66 ntasks = perf_thread_map__nr(ftrace->evlist->core.threads);
71 skel->rodata->use_nsec = ftrace->use_nsec;
81 if (ftrace->target.cpu_list) {
88 cpu = perf_cpu_map__cpu(ftrace->evlist->core.user_requested_cpus, i).cpu;
93 if (target__has_task(&ftrace->target) || target__none(&ftrace->target)) {
100 pid = perf_thread_map__pid(ftrace->evlist->core.threads, i);
126 event = list_first_entry(&ftrace->event_pair, struct filter_entry, list);
154 int perf_ftrace__latency_start_bpf(struct perf_ftrace *ftrace __maybe_unused)
160 int perf_ftrace__latency_stop_bpf(struct perf_ftrace *ftrace __maybe_unused)
166 int perf_ftrace__latency_read_bpf(struct perf_ftrace *ftrace,
197 if (!ftrace->use_nsec) {
208 int perf_ftrace__latency_cleanup_bpf(struct perf_ftrace *ftrace __maybe_unused)