Searched refs:cpu_active_mask (Results 1 – 16 of 16) sorted by relevance
| /linux/kernel/cgroup/ |
| H A D | cpuset.c | 407 if (WARN_ON(!cpumask_and(pmask, possible_mask, cpu_active_mask))) in guarantee_active_cpus() 408 cpumask_copy(pmask, cpu_active_mask); in guarantee_active_cpus() 1100 !cpumask_equal(top_cpuset.effective_cpus, cpu_active_mask)) in rebuild_sched_domains_locked() 1116 cpu_active_mask)) { in rebuild_sched_domains_locked() 1292 (!cpumask_intersects(xcpus, cpu_active_mask) && in tasks_nocpu_error() 1389 cpumask_and(xcpus, xcpus, cpu_active_mask); in partition_xcpus_del() 1549 if (!cpumask_intersects(tmp->new_cpus, cpu_active_mask) || in remote_partition_enable() 1828 cpumask_and(tmp->new_cpus, xcpus, cpu_active_mask); in update_parent_effective_cpumask() 1890 cpumask_and(tmp->new_cpus, tmp->delmask, cpu_active_mask); in update_parent_effective_cpumask() 1899 !cpumask_intersects(tmp->addmask, cpu_active_mask))) { in update_parent_effective_cpumask() [all …]
|
| /linux/include/linux/ |
| H A D | cpumask.h | 125 #define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask) macro 1201 #define num_active_cpus() cpumask_weight(cpu_active_mask) 1225 return cpumask_test_cpu(cpu, cpu_active_mask); in cpu_active()
|
| /linux/arch/mips/kernel/ |
| H A D | mips-mt-fpaff.c | 182 cpumask_and(&mask, &allowed, cpu_active_mask); in mipsmt_sys_sched_getaffinity()
|
| /linux/kernel/sched/ |
| H A D | cpupri.c | 105 cpumask_and(lowest_mask, lowest_mask, cpu_active_mask); in __cpupri_find()
|
| H A D | deadline.c | 133 if (cpumask_subset(rd->span, cpu_active_mask)) in dl_bw_cpus() 138 for_each_cpu_and(i, rd->span, cpu_active_mask) in dl_bw_cpus() 149 for_each_cpu_and(i, mask, cpu_active_mask) in __dl_bw_capacity() 191 for_each_cpu_and(i, rd->span, cpu_active_mask) { in __dl_update() 667 cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr); in dl_task_offline_migration() 680 cpu = cpumask_any(cpu_active_mask); in dl_task_offline_migration()
|
| H A D | topology.c | 280 if (!sched_is_eas_possible(cpu_active_mask)) { in sched_energy_aware_handler() 500 if (cpumask_test_cpu(rq->cpu, cpu_active_mask)) in rq_attach_root() 2791 cpumask_and(doms_new[0], cpu_active_mask, in partition_sched_domains_locked() 2815 cpumask_and(doms_new[0], cpu_active_mask, in partition_sched_domains_locked()
|
| H A D | syscalls.c | 1304 cpumask_and(mask, &p->cpus_mask, cpu_active_mask); in sched_getaffinity()
|
| H A D | core.c | 3062 const struct cpumask *cpu_valid_mask = cpu_active_mask; in __set_cpus_allowed_ptr_locked() 8600 sched_init_domains(cpu_active_mask); in sched_init_smp()
|
| H A D | fair.c | 6642 if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask)) in unthrottle_offline_cfs_rqs() 11758 cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask); in sched_balance_rq()
|
| /linux/kernel/ |
| H A D | stop_machine.c | 697 queue_stop_cpus_work(cpu_active_mask, multi_cpu_stop, &msdata, in stop_machine_from_inactive_cpu()
|
| H A D | workqueue.c | 2706 if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask)) in unbind_worker()
|
| /linux/kernel/power/ |
| H A D | energy_model.c | 462 cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask); in em_cpufreq_update_efficiencies()
|
| /linux/arch/powerpc/perf/ |
| H A D | hv-gpci.c | 862 target = cpumask_last(cpu_active_mask); in ppc_hv_gpci_cpu_offline()
|
| H A D | hv-24x7.c | 1672 target = cpumask_last(cpu_active_mask); in ppc_hv_24x7_cpu_offline()
|
| /linux/kernel/time/ |
| H A D | hrtimer.c | 2296 int i, ncpu = cpumask_any_and(cpu_active_mask, housekeeping_cpumask(HK_TYPE_TIMER)); in hrtimers_cpu_dying()
|
| /linux/arch/arm64/kernel/ |
| H A D | cpufeature.c | 3954 cpu_active_mask); in enable_mismatched_32bit_el0()
|