Lines Matching full:cpu
3 * Arch specific cpu topology information
11 #include <linux/cpu.h>
65 int cpu; in topology_set_scale_freq_source() local
76 for_each_cpu(cpu, cpus) { in topology_set_scale_freq_source()
77 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); in topology_set_scale_freq_source()
81 rcu_assign_pointer(per_cpu(sft_data, cpu), data); in topology_set_scale_freq_source()
82 cpumask_set_cpu(cpu, &scale_freq_counters_mask); in topology_set_scale_freq_source()
96 int cpu; in topology_clear_scale_freq_source() local
100 for_each_cpu(cpu, cpus) { in topology_clear_scale_freq_source()
101 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); in topology_clear_scale_freq_source()
104 rcu_assign_pointer(per_cpu(sft_data, cpu), NULL); in topology_clear_scale_freq_source()
105 cpumask_clear_cpu(cpu, &scale_freq_counters_mask); in topology_clear_scale_freq_source()
158 void topology_set_cpu_scale(unsigned int cpu, unsigned long capacity) in topology_set_cpu_scale() argument
160 per_cpu(cpu_scale, cpu) = capacity; in topology_set_cpu_scale()
184 int cpu; in topology_update_thermal_pressure() local
186 cpu = cpumask_first(cpus); in topology_update_thermal_pressure()
187 max_capacity = arch_scale_cpu_capacity(cpu); in topology_update_thermal_pressure()
188 max_freq = arch_scale_freq_ref(cpu); in topology_update_thermal_pressure()
201 trace_thermal_pressure_update(cpu, th_pressure); in topology_update_thermal_pressure()
203 for_each_cpu(cpu, cpus) in topology_update_thermal_pressure()
204 WRITE_ONCE(per_cpu(thermal_pressure, cpu), th_pressure); in topology_update_thermal_pressure()
212 struct cpu *cpu = container_of(dev, struct cpu, dev); in cpu_capacity_show() local
214 return sysfs_emit(buf, "%lu\n", topology_get_cpu_scale(cpu->dev.id)); in cpu_capacity_show()
222 static int cpu_capacity_sysctl_add(unsigned int cpu) in cpu_capacity_sysctl_add() argument
224 struct device *cpu_dev = get_cpu_device(cpu); in cpu_capacity_sysctl_add()
234 static int cpu_capacity_sysctl_remove(unsigned int cpu) in cpu_capacity_sysctl_remove() argument
236 struct device *cpu_dev = get_cpu_device(cpu); in cpu_capacity_sysctl_remove()
248 cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "topology/cpu-capacity", in register_cpu_capacity_sysctl()
288 int cpu; in topology_normalize_cpu_scale() local
294 for_each_possible_cpu(cpu) { in topology_normalize_cpu_scale()
295 capacity = raw_capacity[cpu] * per_cpu(capacity_freq_ref, cpu); in topology_normalize_cpu_scale()
300 for_each_possible_cpu(cpu) { in topology_normalize_cpu_scale()
301 capacity = raw_capacity[cpu] * per_cpu(capacity_freq_ref, cpu); in topology_normalize_cpu_scale()
304 topology_set_cpu_scale(cpu, capacity); in topology_normalize_cpu_scale()
305 pr_debug("cpu_capacity: CPU%d cpu_capacity=%lu\n", in topology_normalize_cpu_scale()
306 cpu, topology_get_cpu_scale(cpu)); in topology_normalize_cpu_scale()
310 bool __init topology_parse_cpu_capacity(struct device_node *cpu_node, int cpu) in topology_parse_cpu_capacity() argument
332 raw_capacity[cpu] = cpu_capacity; in topology_parse_cpu_capacity()
334 cpu_node, raw_capacity[cpu]); in topology_parse_cpu_capacity()
337 * Update capacity_freq_ref for calculating early boot CPU capacities. in topology_parse_cpu_capacity()
338 * For non-clk CPU DVFS mechanism, there's no way to get the in topology_parse_cpu_capacity()
344 per_cpu(capacity_freq_ref, cpu) = in topology_parse_cpu_capacity()
361 void __weak freq_inv_set_max_ratio(int cpu, u64 max_rate) in freq_inv_set_max_ratio() argument
372 int cpu; in topology_init_cpu_capacity_cppc() local
382 for_each_possible_cpu(cpu) { in topology_init_cpu_capacity_cppc()
383 if (!cppc_get_perf_caps(cpu, &perf_caps) && in topology_init_cpu_capacity_cppc()
386 raw_capacity[cpu] = perf_caps.highest_perf; in topology_init_cpu_capacity_cppc()
387 capacity_scale = max_t(u64, capacity_scale, raw_capacity[cpu]); in topology_init_cpu_capacity_cppc()
389 per_cpu(capacity_freq_ref, cpu) = cppc_perf_to_khz(&perf_caps, raw_capacity[cpu]); in topology_init_cpu_capacity_cppc()
391 pr_debug("cpu_capacity: CPU%d cpu_capacity=%u (raw).\n", in topology_init_cpu_capacity_cppc()
392 cpu, raw_capacity[cpu]); in topology_init_cpu_capacity_cppc()
396 pr_err("cpu_capacity: CPU%d missing/invalid highest performance.\n", cpu); in topology_init_cpu_capacity_cppc()
401 for_each_possible_cpu(cpu) { in topology_init_cpu_capacity_cppc()
402 freq_inv_set_max_ratio(cpu, in topology_init_cpu_capacity_cppc()
403 per_cpu(capacity_freq_ref, cpu) * HZ_PER_KHZ); in topology_init_cpu_capacity_cppc()
405 capacity = raw_capacity[cpu]; in topology_init_cpu_capacity_cppc()
408 topology_set_cpu_scale(cpu, capacity); in topology_init_cpu_capacity_cppc()
409 pr_debug("cpu_capacity: CPU%d cpu_capacity=%lu\n", in topology_init_cpu_capacity_cppc()
410 cpu, topology_get_cpu_scale(cpu)); in topology_init_cpu_capacity_cppc()
432 int cpu; in init_cpu_capacity_callback() local
437 pr_debug("cpu_capacity: init cpu capacity for CPUs [%*pbl] (to_visit=%*pbl)\n", in init_cpu_capacity_callback()
443 for_each_cpu(cpu, policy->related_cpus) { in init_cpu_capacity_callback()
444 per_cpu(capacity_freq_ref, cpu) = policy->cpuinfo.max_freq; in init_cpu_capacity_callback()
445 freq_inv_set_max_ratio(cpu, in init_cpu_capacity_callback()
446 per_cpu(capacity_freq_ref, cpu) * HZ_PER_KHZ); in init_cpu_capacity_callback()
472 * information is not needed for cpu capacity initialization. in register_cpufreq_notifier()
505 * This function returns the logic cpu number of the node.
507 * (1) logic cpu number which is > 0.
509 * there is no possible logical CPU in the kernel to match. This happens
511 * CPU nodes in DT. We need to just ignore this case.
517 int cpu; in get_cpu_for_node() local
519 cpu_node = of_parse_phandle(node, "cpu", 0); in get_cpu_for_node()
523 cpu = of_cpu_node_to_id(cpu_node); in get_cpu_for_node()
524 if (cpu >= 0) in get_cpu_for_node()
525 topology_parse_cpu_capacity(cpu_node, cpu); in get_cpu_for_node()
527 pr_info("CPU node for %pOF exist but the possible cpu range is :%*pbl\n", in get_cpu_for_node()
531 return cpu; in get_cpu_for_node()
540 int cpu; in parse_core() local
548 cpu = get_cpu_for_node(t); in parse_core()
549 if (cpu >= 0) { in parse_core()
550 cpu_topology[cpu].package_id = package_id; in parse_core()
551 cpu_topology[cpu].cluster_id = cluster_id; in parse_core()
552 cpu_topology[cpu].core_id = core_id; in parse_core()
553 cpu_topology[cpu].thread_id = i; in parse_core()
554 } else if (cpu != -ENODEV) { in parse_core()
555 pr_err("%pOF: Can't get CPU for thread\n", t); in parse_core()
564 cpu = get_cpu_for_node(core); in parse_core()
565 if (cpu >= 0) { in parse_core()
567 pr_err("%pOF: Core has both threads and CPU\n", in parse_core()
572 cpu_topology[cpu].package_id = package_id; in parse_core()
573 cpu_topology[cpu].cluster_id = cluster_id; in parse_core()
574 cpu_topology[cpu].core_id = core_id; in parse_core()
575 } else if (leaf && cpu != -ENODEV) { in parse_core()
576 pr_err("%pOF: Can't get CPU for leaf core\n", core); in parse_core()
623 pr_err("%pOF: cpu-map children should be clusters\n", in parse_cluster()
681 int cpu; in parse_dt_topology() local
685 pr_err("No CPU information found in DT\n"); in parse_dt_topology()
690 * When topology is provided cpu-map is essentially a root in parse_dt_topology()
693 map = of_get_child_by_name(cn, "cpu-map"); in parse_dt_topology()
707 for_each_possible_cpu(cpu) in parse_dt_topology()
708 if (cpu_topology[cpu].package_id < 0) { in parse_dt_topology()
722 * cpu topology table
727 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask() argument
729 const cpumask_t *core_mask = cpumask_of_node(cpu_to_node(cpu)); in cpu_coregroup_mask()
732 if (cpumask_subset(&cpu_topology[cpu].core_sibling, core_mask)) { in cpu_coregroup_mask()
734 core_mask = &cpu_topology[cpu].core_sibling; in cpu_coregroup_mask()
737 if (last_level_cache_is_valid(cpu)) { in cpu_coregroup_mask()
738 if (cpumask_subset(&cpu_topology[cpu].llc_sibling, core_mask)) in cpu_coregroup_mask()
739 core_mask = &cpu_topology[cpu].llc_sibling; in cpu_coregroup_mask()
743 * For systems with no shared cpu-side LLC but with clusters defined, in cpu_coregroup_mask()
748 cpumask_subset(core_mask, &cpu_topology[cpu].cluster_sibling)) in cpu_coregroup_mask()
749 core_mask = &cpu_topology[cpu].cluster_sibling; in cpu_coregroup_mask()
754 const struct cpumask *cpu_clustergroup_mask(int cpu) in cpu_clustergroup_mask() argument
760 if (cpumask_subset(cpu_coregroup_mask(cpu), in cpu_clustergroup_mask()
761 &cpu_topology[cpu].cluster_sibling)) in cpu_clustergroup_mask()
762 return topology_sibling_cpumask(cpu); in cpu_clustergroup_mask()
764 return &cpu_topology[cpu].cluster_sibling; in cpu_clustergroup_mask()
770 int cpu, ret; in update_siblings_masks() local
777 for_each_online_cpu(cpu) { in update_siblings_masks()
778 cpu_topo = &cpu_topology[cpu]; in update_siblings_masks()
780 if (last_level_cache_is_shared(cpu, cpuid)) { in update_siblings_masks()
781 cpumask_set_cpu(cpu, &cpuid_topo->llc_sibling); in update_siblings_masks()
789 cpumask_set_cpu(cpu, &cpuid_topo->core_sibling); in update_siblings_masks()
795 cpumask_set_cpu(cpu, &cpuid_topo->cluster_sibling); in update_siblings_masks()
803 cpumask_set_cpu(cpu, &cpuid_topo->thread_sibling); in update_siblings_masks()
807 static void clear_cpu_topology(int cpu) in clear_cpu_topology() argument
809 struct cpu_topology *cpu_topo = &cpu_topology[cpu]; in clear_cpu_topology()
812 cpumask_set_cpu(cpu, &cpu_topo->llc_sibling); in clear_cpu_topology()
815 cpumask_set_cpu(cpu, &cpu_topo->cluster_sibling); in clear_cpu_topology()
818 cpumask_set_cpu(cpu, &cpu_topo->core_sibling); in clear_cpu_topology()
820 cpumask_set_cpu(cpu, &cpu_topo->thread_sibling); in clear_cpu_topology()
825 unsigned int cpu; in reset_cpu_topology() local
827 for_each_possible_cpu(cpu) { in reset_cpu_topology()
828 struct cpu_topology *cpu_topo = &cpu_topology[cpu]; in reset_cpu_topology()
835 clear_cpu_topology(cpu); in reset_cpu_topology()
839 void remove_cpu_topology(unsigned int cpu) in remove_cpu_topology() argument
843 for_each_cpu(sibling, topology_core_cpumask(cpu)) in remove_cpu_topology()
844 cpumask_clear_cpu(cpu, topology_core_cpumask(sibling)); in remove_cpu_topology()
845 for_each_cpu(sibling, topology_sibling_cpumask(cpu)) in remove_cpu_topology()
846 cpumask_clear_cpu(cpu, topology_sibling_cpumask(sibling)); in remove_cpu_topology()
847 for_each_cpu(sibling, topology_cluster_cpumask(cpu)) in remove_cpu_topology()
848 cpumask_clear_cpu(cpu, topology_cluster_cpumask(sibling)); in remove_cpu_topology()
849 for_each_cpu(sibling, topology_llc_cpumask(cpu)) in remove_cpu_topology()
850 cpumask_clear_cpu(cpu, topology_llc_cpumask(sibling)); in remove_cpu_topology()
852 clear_cpu_topology(cpu); in remove_cpu_topology()
863 int cpu, ret; in init_cpu_topology() local
879 for_each_possible_cpu(cpu) { in init_cpu_topology()
880 ret = fetch_cache_info(cpu); in init_cpu_topology()
900 pr_debug("CPU%u: package %d core %d thread %d\n", in store_cpu_topology()