| /linux/tools/bpf/bpftool/ |
| H A D | map_perf_ring.c | 52 bool all_cpus; member 67 int idx = ctx->all_cpus ? cpu : ctx->idx; in print_bpf_output() 122 .all_cpus = true, in do_event_pipe() 174 ctx.all_cpus = false; in do_event_pipe() 177 if (!ctx.all_cpus) { in do_event_pipe() 187 opts.cpu_cnt = ctx.all_cpus ? 0 : 1; in do_event_pipe()
|
| /linux/tools/lib/perf/ |
| H A D | evlist.c | 112 evlist->all_cpus, in __perf_evlist__propagate_maps() 153 perf_cpu_map__merge(&evlist->all_cpus, evsel->cpus); in __perf_evlist__propagate_maps() 161 perf_cpu_map__put(evlist->all_cpus); in perf_evlist__propagate_maps() 162 evlist->all_cpus = NULL; in perf_evlist__propagate_maps() 239 perf_cpu_map__put(evlist->all_cpus); in perf_evlist__exit() 242 evlist->all_cpus = NULL; in perf_evlist__exit() 412 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__alloc_pollfd() 544 struct perf_cpu evlist_cpu = perf_cpu_map__cpu(evlist->all_cpus, cpu_idx); in mmap_per_evsel() 638 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_thread() 680 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_cpu() [all …]
|
| /linux/lib/ |
| H A D | test_lockup.c | 39 static bool all_cpus; variable 40 module_param(all_cpus, bool, 0400); 41 MODULE_PARM_DESC(all_cpus, "trigger lockup at all cpus at once"); 559 all_cpus ? "all_cpus " : "", in test_lockup_init() 576 if (all_cpus) { in test_lockup_init()
|
| /linux/arch/riscv/kernel/vdso/ |
| H A D | hwprobe.c | 20 bool all_cpus = !cpusetsize && !cpus; in riscv_vdso_get_values() local 30 if (flags != 0 || (!all_cpus && !avd->homogeneous_cpus) || unlikely(!avd->ready)) in riscv_vdso_get_values()
|
| /linux/tools/perf/util/ |
| H A D | bpf_counter_cgroup.c | 126 if (evsel__open_per_cpu(cgrp_switch, evlist->core.all_cpus, -1) < 0) { in bperf_load_program() 131 perf_cpu_map__for_each_cpu(cpu, i, evlist->core.all_cpus) { in bperf_load_program() 240 perf_cpu_map__for_each_cpu(cpu, idx, evlist->core.all_cpus) in bperf_cgrp__sync_counters()
|
| H A D | evlist.c | 428 .evlist_cpu_map_nr = perf_cpu_map__nr(evlist->core.all_cpus), in evlist_cpu_iterator__init() 444 itr->cpu = perf_cpu_map__cpu(evlist->core.all_cpus, 0); in evlist_cpu_iterator__init() 479 perf_cpu_map__cpu(evlist_cpu_itr->container->core.all_cpus, in evlist_cpu_iterator__next()
|
| H A D | python.c | 1302 pcpu_map->cpus = perf_cpu_map__get(pevlist->evlist.core.all_cpus); in pyrf_evlist__all_cpus() 2009 evlist__init(&pevlist->evlist, evlist->core.all_cpus, evlist->core.threads); in pyrf_evlist__from_evlist()
|
| H A D | auxtrace.c | 204 mp->cpu = perf_cpu_map__cpu(evlist->core.all_cpus, idx); in auxtrace_mmap_params__set_idx() 673 struct perf_cpu evlist_cpu = perf_cpu_map__cpu(evlist->core.all_cpus, idx); in evlist__enable_event_idx()
|
| H A D | header.c | 905 struct perf_cpu cpu = perf_cpu_map__min(evlist->core.all_cpus); in write_cpuid()
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | auxtrace.c | 58 struct perf_cpu cpu = perf_cpu_map__min(evlist->core.all_cpus); in auxtrace_record__init()
|
| /linux/arch/x86/kvm/ |
| H A D | hyperv.c | 2027 bool all_cpus; in kvm_hv_flush_tlb() local 2077 all_cpus = (flush.flags & HV_FLUSH_ALL_PROCESSORS) || in kvm_hv_flush_tlb() 2099 all_cpus = flush_ex.hv_vp_set.format != in kvm_hv_flush_tlb() 2105 if (!all_cpus) { in kvm_hv_flush_tlb() 2140 if (all_cpus && !is_guest_mode(vcpu)) { in kvm_hv_flush_tlb() 2182 if (!all_cpus && in kvm_hv_flush_tlb() 2232 bool all_cpus; in kvm_hv_send_ipi() local 2251 all_cpus = false; in kvm_hv_send_ipi() 2272 all_cpus = send_ipi_ex.vp_set.format == HV_GENERIC_SET_ALL; in kvm_hv_send_ipi() 2277 if (all_cpus) in kvm_hv_send_ipi() [all …]
|
| /linux/tools/lib/perf/include/internal/ |
| H A D | evlist.h | 28 struct perf_cpu_map *all_cpus; member
|
| /linux/drivers/platform/x86/intel/ifs/ |
| H A D | runtest.c | 166 int all_cpus = cpumask_weight(smt_mask); in wait_for_sibling_cpu() local 169 while (atomic_read(t) < all_cpus) { in wait_for_sibling_cpu()
|
| /linux/tools/perf/python/ |
| H A D | ilist.py | 429 for cpu in self.evlist.all_cpus(): 434 for cpu in self.evlist.all_cpus():
|
| /linux/tools/perf/ |
| H A D | builtin-record.c | 1064 struct perf_cpu_map *cpus = evlist->core.all_cpus; in record__thread_data_init_maps() 2158 err = perf_event__synthesize_cpu_map(&rec->tool, rec->evlist->core.all_cpus, in record__synthesize() 4064 struct perf_cpu_map *cpus = rec->evlist->core.all_cpus; in record__init_thread_masks()
|
| H A D | builtin-stat.c | 1603 nr = perf_cpu_map__max(evsel_list->core.all_cpus).cpu + 1; in perf_stat_init_aggr_mode()
|