Home
last modified time | relevance | path

Searched refs:nr_cpus (Results 1 – 25 of 90) sorted by relevance

1234

/linux/tools/lib/perf/
H A Dcpumap.c18 void perf_cpu_map__set_nr(struct perf_cpu_map *map, int nr_cpus) in perf_cpu_map__set_nr() argument
20 RC_CHK_ACCESS(map)->nr = nr_cpus; in perf_cpu_map__set_nr()
23 struct perf_cpu_map *perf_cpu_map__alloc(int nr_cpus) in perf_cpu_map__alloc() argument
28 if (nr_cpus == 0) in perf_cpu_map__alloc()
31 cpus = malloc(sizeof(*cpus) + sizeof(struct perf_cpu) * nr_cpus); in perf_cpu_map__alloc()
33 cpus->nr = nr_cpus; in perf_cpu_map__alloc()
81 int nr_cpus, nr_cpus_conf; in cpu_map__new_sysconf() local
83 nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); in cpu_map__new_sysconf()
84 if (nr_cpus < 0) in cpu_map__new_sysconf()
88 if (nr_cpus ! in cpu_map__new_sysconf()
140 cpu_map__trim_new(int nr_cpus,const struct perf_cpu * tmp_cpus) cpu_map__trim_new() argument
170 int i, nr_cpus = 0; perf_cpu_map__new() local
[all...]
/linux/tools/testing/selftests/kvm/arm64/
H A Dvgic_lpi_stress.c34 u32 nr_cpus; member
47 .nr_cpus = 1,
68 u32 nr_cpus = test_data.nr_cpus; in guest_setup_its_mappings() local
70 for (coll_id = 0; coll_id < nr_cpus; coll_id++) in guest_setup_its_mappings()
85 coll_id = (coll_id + 1) % test_data.nr_cpus; in guest_setup_its_mappings()
94 for (i = 0; i < test_data.nr_cpus; i++) in guest_invalidate_all_rdists()
103 gic_init(GIC_V3, test_data.nr_cpus); in guest_setup_gic()
112 while (atomic_load(&nr_cpus_ready) < test_data.nr_cpus) in guest_setup_gic()
158 sz += (1 + test_data.nr_cpus) * SZ_64 in setup_memslot()
185 u32 nr_cpus = test_data.nr_cpus; setup_test_data() local
[all...]
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dmap_lookup_percpu_elem.c11 int ret, i, nr_cpus = libbpf_num_possible_cpus(); in test_map_lookup_percpu_elem() local
14 buf = malloc(nr_cpus*sizeof(__u64)); in test_map_lookup_percpu_elem()
18 for (i = 0; i < nr_cpus; i++) in test_map_lookup_percpu_elem()
20 sum = (nr_cpus - 1) * nr_cpus / 2; in test_map_lookup_percpu_elem()
27 skel->rodata->nr_cpus = nr_cpus; in test_map_lookup_percpu_elem()
H A Dmap_init.c10 static int nr_cpus; variable
22 pcpu_map_value_t value[nr_cpus]; in map_populate()
26 for (i = 0; i < nr_cpus; i++) in map_populate()
105 for (i = 0; i < nr_cpus; i++) { in check_values_one_cpu()
131 pcpu_map_value_t value[nr_cpus]; in test_pcpu_map_init()
171 pcpu_map_value_t value[nr_cpus]; in test_pcpu_lru_map_init()
203 nr_cpus = bpf_num_possible_cpus(); in test_map_init()
204 if (nr_cpus <= 1) { in test_map_init()
H A Dlookup_and_delete.c11 static int nr_cpus; variable
29 __u64 key, value[nr_cpus]; in fill_values_percpu()
32 for (i = 0; i < nr_cpus; i++) in fill_values_percpu()
137 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_percpu_hash()
156 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_percpu_hash()
223 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_lru_percpu_hash()
242 for (i = 0; i < nr_cpus; i++) in test_lookup_and_delete_lru_percpu_hash()
252 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_lru_percpu_hash()
281 nr_cpus = bpf_num_possible_cpus(); in test_lookup_and_delete()
H A Dperf_buffer.c48 int err, on_len, nr_on_cpus = 0, nr_cpus, i, j; in serial_test_perf_buffer() local
56 nr_cpus = libbpf_num_possible_cpus(); in serial_test_perf_buffer()
57 if (CHECK(nr_cpus < 0, "nr_cpus", "err %d\n", nr_cpus)) in serial_test_perf_buffer()
94 for (i = 0; i < nr_cpus; i++) { in serial_test_perf_buffer()
117 for (i = 0, j = 0; i < nr_cpus; i++) { in serial_test_perf_buffer()
/linux/tools/tracing/rtla/src/
H A Dtimerlat_u.c106 static int timerlat_u_send_kill(pid_t *procs, int nr_cpus) in timerlat_u_send_kill() argument
111 for (i = 0; i < nr_cpus; i++) { in timerlat_u_send_kill()
138 int nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in timerlat_u_dispatcher() local
150 procs = calloc(nr_cpus, sizeof(pid_t)); in timerlat_u_dispatcher()
154 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher()
177 timerlat_u_send_kill(procs, nr_cpus); in timerlat_u_dispatcher()
190 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher()
204 timerlat_u_send_kill(procs, nr_cpus); in timerlat_u_dispatcher()
212 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher()
H A Dtimerlat_hist.c48 int nr_cpus; member
60 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in timerlat_free_histogram()
83 *timerlat_alloc_histogram(int nr_cpus, int entries, int bucket_size) in timerlat_alloc_histogram() argument
94 data->nr_cpus = nr_cpus; in timerlat_alloc_histogram()
97 data->hist = calloc(1, sizeof(*data->hist) * nr_cpus); in timerlat_alloc_histogram()
102 for (cpu = 0; cpu < nr_cpus; cpu++) { in timerlat_alloc_histogram()
117 for (cpu = 0; cpu < nr_cpus; cpu++) { in timerlat_alloc_histogram()
204 long long value_irq[data->nr_cpus], in timerlat_hist_bpf_pull_data()
205 value_thread[data->nr_cpus], in timerlat_hist_bpf_pull_data()
1122 int nr_cpus; timerlat_init_hist() local
1188 int nr_cpus, i; timerlat_hist_main() local
[all...]
H A Dosnoise_top.c34 int nr_cpus; member
50 static struct osnoise_top_data *osnoise_alloc_top(int nr_cpus) in osnoise_alloc_top() argument
58 data->nr_cpus = nr_cpus; in osnoise_alloc_top()
61 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); in osnoise_alloc_top()
229 static int nr_cpus = -1; in osnoise_print_stats() local
232 if (nr_cpus == -1) in osnoise_print_stats()
233 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in osnoise_print_stats()
240 for (i = 0; i < nr_cpus; i++) { in osnoise_print_stats()
552 int nr_cpus; in osnoise_init_top() local
[all...]
H A Dutils.c106 int nr_cpus; in parse_cpu_set() local
112 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in parse_cpu_set()
116 if (cpu < 0 || (!cpu && *p != '0') || cpu >= nr_cpus) in parse_cpu_set()
124 if (end_cpu < cpu || (!end_cpu && *p != '0') || end_cpu >= nr_cpus) in parse_cpu_set()
540 int nr_cpus; in save_cpu_idle_disable_state() local
548 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in save_cpu_idle_disable_state()
549 saved_cpu_idle_disable_state = calloc(nr_cpus, sizeof(unsigned int *)); in save_cpu_idle_disable_state()
625 int nr_cpus; in free_cpu_idle_disable_states() local
630 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in free_cpu_idle_disable_states()
632 for (cpu = 0; cpu < nr_cpus; cp in free_cpu_idle_disable_states()
[all...]
/linux/tools/testing/selftests/bpf/map_tests/
H A Darray_map_batch_ops.c13 static int nr_cpus; variable
28 cpu_offset = i * nr_cpus; in map_batch_update()
29 for (j = 0; j < nr_cpus; j++) in map_batch_update()
49 cpu_offset = i * nr_cpus; in map_batch_verify()
50 for (j = 0; j < nr_cpus; j++) { in map_batch_verify()
90 value_size *= nr_cpus; in __test_map_lookup_and_update_batch()
158 nr_cpus = libbpf_num_possible_cpus(); in test_array_map_batch_ops()
160 CHECK(nr_cpus < 0, "nr_cpus checking", in test_array_map_batch_ops()
/linux/tools/testing/selftests/bpf/
H A Dtest_lru_map.c26 static int nr_cpus; variable
96 unsigned long long value0[nr_cpus], value1[nr_cpus]; in map_subset()
127 while (next < nr_cpus) { in sched_next_online()
144 return (map_size / nr_cpus) / 2; in __tgt_size()
150 return tgt_free * nr_cpus * 2; in __map_size()
163 unsigned long long key, value[nr_cpus]; in test_lru_sanity0()
173 lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); in test_lru_sanity0()
254 unsigned long long key, end_key, value[nr_cpus]; in test_lru_sanity1()
322 unsigned long long key, value[nr_cpus]; in test_lru_sanity2()
[all...]
/linux/samples/bpf/
H A Dtest_lru_dist.c31 static int nr_cpus; variable
222 if (next_to_try == nr_cpus) in sched_next_online()
225 while (next_to_try < nr_cpus) { in sched_next_online()
318 nr_cpus * lru_size); in test_parallel_lru_dist()
334 unsigned long long key, value[nr_cpus]; in test_lru_loss0()
346 map_fd = create_map(map_type, map_flags, 900 * nr_cpus); in test_lru_loss0()
388 unsigned long long key, value[nr_cpus]; in test_lru_loss1()
398 map_fd = create_map(map_type, map_flags, 1000 * nr_cpus); in test_lru_loss1()
426 unsigned long long key, value[nr_cpus]; in do_test_parallel_lru_loss()
475 nr_cpus * (100 in test_parallel_lru_loss()
[all...]
H A Dtracex3_user.c19 unsigned int nr_cpus = bpf_num_possible_cpus(); in clear_stats() local
20 __u64 values[nr_cpus]; in clear_stats()
76 unsigned int nr_cpus = bpf_num_possible_cpus(); in print_hist() local
78 long values[nr_cpus]; in print_hist()
88 for (i = 0; i < nr_cpus; i++) in print_hist()
H A Dxdp_sample_user.c317 unsigned int nr_cpus = libbpf_num_possible_cpus(); in alloc_record_per_cpu() local
320 array = calloc(nr_cpus, sizeof(*array)); in alloc_record_per_cpu()
322 fprintf(stderr, "Failed to allocate memory (nr_cpus: %u)\n", in alloc_record_per_cpu()
323 nr_cpus); in alloc_record_per_cpu()
343 unsigned int nr_cpus = libbpf_num_possible_cpus(); in map_collect_percpu() local
356 for (i = 0; i < nr_cpus; i++) { in map_collect_percpu()
382 unsigned int nr_cpus = bpf_num_possible_cpus(); in map_collect_percpu_devmap() local
392 values = calloc(count * nr_cpus, sizeof(struct datarec)); in map_collect_percpu_devmap()
414 arr = &values[i * nr_cpus]; in map_collect_percpu_devmap()
650 unsigned int nr_cpus, struc in stats_get_rx_cnt() argument
693 stats_get_cpumap_enqueue(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus) stats_get_cpumap_enqueue() argument
747 stats_get_cpumap_remote(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus) stats_get_cpumap_remote() argument
786 stats_get_cpumap_kthread(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus) stats_get_cpumap_kthread() argument
824 stats_get_redirect_cnt(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus,struct sample_output * out) stats_get_redirect_cnt() argument
856 stats_get_redirect_err_cnt(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus,struct sample_output * out) stats_get_redirect_err_cnt() argument
905 stats_get_exception_cnt(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus,struct sample_output * out) stats_get_exception_cnt() argument
950 stats_get_devmap_xmit(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus,struct sample_output * out) stats_get_devmap_xmit() argument
1002 stats_get_devmap_xmit_multi(struct stats_record * stats_rec,struct stats_record * stats_prev,unsigned int nr_cpus,struct sample_output * out,bool xmit_total) stats_get_devmap_xmit_multi() argument
1105 int nr_cpus = libbpf_num_possible_cpus(); stats_print() local
[all...]
H A Dsampleip_user.c26 static int nr_cpus; variable
49 for (i = 0; i < nr_cpus; i++) { in sampling_start()
72 for (i = 0; i < nr_cpus; i++) in sampling_end()
180 nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); in main()
181 links = calloc(nr_cpus, sizeof(struct bpf_link *)); in main()
/linux/tools/testing/selftests/kvm/lib/arm64/
H A Dgic.c25 static void gic_dist_init(enum gic_type type, unsigned int nr_cpus) in gic_dist_init() argument
42 gic_ops->gic_init(nr_cpus); in gic_dist_init()
51 void gic_init(enum gic_type type, unsigned int nr_cpus) in gic_init() argument
56 GUEST_ASSERT(nr_cpus); in gic_init()
58 gic_dist_init(type, nr_cpus); in gic_init()
/linux/drivers/powercap/
H A Ddtpm_cpu.c48 int i, nr_cpus; in set_pd_power_limit() local
50 nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(pd->cpus)); in set_pd_power_limit()
56 power = table[i].power * nr_cpus; in set_pd_power_limit()
63 power_limit = table[i - 1].power * nr_cpus; in set_pd_power_limit()
126 int nr_cpus; in update_pd_power_uw() local
128 nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(em->cpus)); in update_pd_power_uw()
134 dtpm->power_min *= nr_cpus; in update_pd_power_uw()
137 dtpm->power_max *= nr_cpus; in update_pd_power_uw()
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_map_lookup_percpu_elem.c10 const volatile int nr_cpus; variable
60 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid()
65 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid()
70 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid()
/linux/tools/perf/util/
H A Dbpf_kwork_top.c126 int idx, nr_cpus, fd; in setup_filters() local
142 nr_cpus = libbpf_num_possible_cpus(); in setup_filters()
146 if (cpu.cpu >= nr_cpus) { in setup_filters()
270 int i, fd, nr_cpus; in perf_kwork__top_read_bpf() local
280 nr_cpus = libbpf_num_possible_cpus(); in perf_kwork__top_read_bpf()
281 data = calloc(nr_cpus, sizeof(struct work_data)); in perf_kwork__top_read_bpf()
292 for (i = 0; i < nr_cpus; i++) { in perf_kwork__top_read_bpf()
H A Dtop.c31 int nr_cpus; in perf_top__header_snprintf() local
97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf()
100 nr_cpus > 1 ? "s" : "", in perf_top__header_snprintf()
107 nr_cpus, nr_cpus > 1 ? "s" : ""); in perf_top__header_snprintf()
H A Dsvghelper.c701 int *pos, int nr_cpus) in scan_thread_topology() argument
710 for_each_set_bit(thr, cpumask_bits(&t->sib_thr[i]), nr_cpus) in scan_thread_topology()
716 static void scan_core_topology(int *map, struct topology *t, int nr_cpus) in scan_core_topology() argument
723 for_each_set_bit(cpu, cpumask_bits(&t->sib_core[i]), nr_cpus) in scan_core_topology()
724 scan_thread_topology(map, t, cpu, &pos, nr_cpus); in scan_core_topology()
727 static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus) in str_to_bitmap() argument
738 if (cpu.cpu >= nr_cpus) { in str_to_bitmap()
753 int i, nr_cpus; in svg_build_topology_map() local
758 nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS); in svg_build_topology_map()
774 if (str_to_bitmap(sib_core, &t.sib_core[i], nr_cpus)) { in svg_build_topology_map()
[all...]
/linux/arch/mips/loongson64/
H A Denv.c148 loongson_sysconf.nr_cpus = ecpu->nr_cpus; in prom_lefi_init_env()
151 if (ecpu->nr_cpus > NR_CPUS || ecpu->nr_cpus == 0) in prom_lefi_init_env()
152 loongson_sysconf.nr_cpus = NR_CPUS; in prom_lefi_init_env()
153 loongson_sysconf.nr_nodes = (loongson_sysconf.nr_cpus + in prom_lefi_init_env()
/linux/drivers/pci/controller/
H A Dpcie-iproc-msi.c78 * @nr_cpus: number of toal CPUs
98 int nr_cpus; member
198 return (hwirq % msi->nr_cpus); in hwirq_to_cpu()
254 if (msi->nr_cpus > 1 && nr_irqs > 1) in iproc_msi_irq_domain_alloc()
260 * Allocate 'nr_irqs' multiplied by 'nr_cpus' number of MSI vectors in iproc_msi_irq_domain_alloc()
264 order_base_2(msi->nr_cpus * nr_irqs)); in iproc_msi_irq_domain_alloc()
292 order_base_2(msi->nr_cpus * nr_irqs)); in iproc_msi_irq_domain_free()
474 for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { in iproc_msi_irq_free()
486 for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { in iproc_msi_irq_setup()
538 msi->nr_cpus in iproc_msi_init()
[all...]
/linux/tools/testing/selftests/sched_ext/
H A Ddsp_local_on.bpf.c9 const volatile s32 nr_cpus; variable
46 if (p->nr_cpus_allowed == nr_cpus && !is_migration_disabled(p)) in BPF_STRUCT_OPS()
47 target = bpf_get_prandom_u32() % nr_cpus; in BPF_STRUCT_OPS()

1234