Searched refs:cpu_active_mask (Results 1 – 12 of 12) sorted by relevance
/linux/kernel/cgroup/ |
H A D | cpuset.c | 196 * The top_cpuset is always synchronized to cpu_active_mask and we should avoid 198 * CPU, but not vice versa. cpu_active_mask and cpu_online_mask can differ 372 * of cpu_active_mask. 382 if (WARN_ON(!cpumask_and(pmask, possible_mask, cpu_active_mask))) in guarantee_active_cpus() 383 cpumask_copy(pmask, cpu_active_mask); in guarantee_active_cpus() 1048 !cpumask_equal(top_cpuset.effective_cpus, cpu_active_mask)) in rebuild_sched_domains_locked() 1064 cpu_active_mask)) { in rebuild_sched_domains_locked() 1240 (!cpumask_intersects(xcpus, cpu_active_mask) && in tasks_nocpu_error() 1337 cpumask_and(xcpus, xcpus, cpu_active_mask); in partition_xcpus_del() 1464 if (!cpumask_intersects(tmp->new_cpus, cpu_active_mask) || in remote_partition_enable() [all...] |
/linux/kernel/sched/ |
H A D | cpupri.c | 105 cpumask_and(lowest_mask, lowest_mask, cpu_active_mask); in __cpupri_find()
|
H A D | deadline.c | 133 if (cpumask_subset(rd->span, cpu_active_mask)) in dl_bw_cpus() 138 for_each_cpu_and(i, rd->span, cpu_active_mask) in dl_bw_cpus() 149 for_each_cpu_and(i, mask, cpu_active_mask) in __dl_bw_capacity() 191 for_each_cpu_and(i, rd->span, cpu_active_mask) { in __dl_update() 667 cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr); in dl_task_offline_migration() 680 cpu = cpumask_any(cpu_active_mask); in dl_task_offline_migration()
|
H A D | topology.c | 280 if (!sched_is_eas_possible(cpu_active_mask)) { in sched_energy_aware_handler() 500 if (cpumask_test_cpu(rq->cpu, cpu_active_mask)) in rq_attach_root() 2752 cpumask_and(doms_new[0], cpu_active_mask, in partition_sched_domains_locked() 2776 cpumask_and(doms_new[0], cpu_active_mask, in partition_sched_domains_locked()
|
H A D | syscalls.c | 1304 cpumask_and(mask, &p->cpus_mask, cpu_active_mask); in sched_getaffinity()
|
H A D | fair.c | 6557 if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask)) in unthrottle_offline_cfs_rqs() 11665 cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask); in sched_balance_rq()
|
/linux/arch/mips/kernel/ |
H A D | mips-mt-fpaff.c | 182 cpumask_and(&mask, &allowed, cpu_active_mask); in mipsmt_sys_sched_getaffinity()
|
/linux/include/linux/ |
H A D | cpumask.h | 86 * cpu_active_mask - has bit 'cpu' set iff cpu available to migration 125 #define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask) macro 1201 #define num_active_cpus() cpumask_weight(cpu_active_mask) 1225 return cpumask_test_cpu(cpu, cpu_active_mask); in cpu_active()
|
/linux/kernel/ |
H A D | stop_machine.c | 697 queue_stop_cpus_work(cpu_active_mask, multi_cpu_stop, &msdata, in stop_machine_from_inactive_cpu()
|
H A D | workqueue.c | 2704 if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask)) in unbind_worker()
|
/linux/kernel/power/ |
H A D | energy_model.c | 462 cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask); in em_cpufreq_update_efficiencies()
|
/linux/kernel/time/ |
H A D | hrtimer.c | 2280 int i, ncpu = cpumask_any_and(cpu_active_mask, housekeeping_cpumask(HK_TYPE_TIMER)); in hrtimers_cpu_dying()
|