Lines Matching full:cpus

23 			    const struct cpumask *cpus)  in hwprobe_arch_id()  argument
29 for_each_cpu(cpu, cpus) { in hwprobe_arch_id()
63 const struct cpumask *cpus) in hwprobe_isa_ext0() argument
82 for_each_cpu(cpu, cpus) { in hwprobe_isa_ext0()
172 static bool hwprobe_ext0_has(const struct cpumask *cpus, u64 ext) in hwprobe_ext0_has() argument
176 hwprobe_isa_ext0(&pair, cpus); in hwprobe_ext0_has()
181 static u64 hwprobe_misaligned(const struct cpumask *cpus) in hwprobe_misaligned() argument
186 for_each_cpu(cpu, cpus) { in hwprobe_misaligned()
204 static u64 hwprobe_misaligned(const struct cpumask *cpus) in hwprobe_misaligned() argument
217 static u64 hwprobe_vec_misaligned(const struct cpumask *cpus) in hwprobe_vec_misaligned() argument
223 for_each_cpu(cpu, cpus) { in hwprobe_vec_misaligned()
241 static u64 hwprobe_vec_misaligned(const struct cpumask *cpus) in hwprobe_vec_misaligned() argument
254 const struct cpumask *cpus) in hwprobe_one_pair() argument
260 hwprobe_arch_id(pair, cpus); in hwprobe_one_pair()
273 hwprobe_isa_ext0(pair, cpus); in hwprobe_one_pair()
278 pair->value = hwprobe_misaligned(cpus); in hwprobe_one_pair()
282 pair->value = hwprobe_vec_misaligned(cpus); in hwprobe_one_pair()
287 if (hwprobe_ext0_has(cpus, RISCV_HWPROBE_EXT_ZICBOZ)) in hwprobe_one_pair()
292 if (hwprobe_ext0_has(cpus, RISCV_HWPROBE_EXT_ZICBOM)) in hwprobe_one_pair()
304 hwprobe_isa_vendor_ext_thead_0(pair, cpus); in hwprobe_one_pair()
326 cpumask_t cpus; in hwprobe_get_values() local
335 * 0 as a shortcut to all online CPUs. in hwprobe_get_values()
337 cpumask_clear(&cpus); in hwprobe_get_values()
339 cpumask_copy(&cpus, cpu_online_mask); in hwprobe_get_values()
344 ret = copy_from_user(&cpus, cpus_user, cpusetsize); in hwprobe_get_values()
352 cpumask_and(&cpus, &cpus, cpu_online_mask); in hwprobe_get_values()
353 if (cpumask_empty(&cpus)) in hwprobe_get_values()
364 hwprobe_one_pair(&pair, &cpus); in hwprobe_get_values()
381 cpumask_t cpus, one_cpu; in hwprobe_get_cpus() local
395 ret = copy_from_user(&cpus, cpus_user, cpusetsize); in hwprobe_get_cpus()
399 if (cpumask_empty(&cpus)) in hwprobe_get_cpus()
400 cpumask_copy(&cpus, cpu_online_mask); in hwprobe_get_cpus()
402 cpumask_and(&cpus, &cpus, cpu_online_mask); in hwprobe_get_cpus()
427 for_each_cpu(cpu, &cpus) { in hwprobe_get_cpus()
433 cpumask_clear_cpu(cpu, &cpus); in hwprobe_get_cpus()
440 cpumask_clear(&cpus); in hwprobe_get_cpus()
442 ret = copy_to_user(cpus_user, &cpus, cpusetsize); in hwprobe_get_cpus()
472 * Initialize vDSO data with the answers for the "all CPUs" case, to in init_hwprobe_vdso_data()
492 * all harts, then assume all CPUs are the same, and allow the vDSO to in init_hwprobe_vdso_data()
494 * populated) or any value returns -1 (varies across CPUs), then the in init_hwprobe_vdso_data()
507 cpus, unsigned int, flags) in SYSCALL_DEFINE5() argument
510 cpus, flags); in SYSCALL_DEFINE5()