Home
last modified time | relevance | path

Searched refs:cpu_smt_mask (Results 1 – 11 of 11) sorted by relevance

/linux/drivers/platform/x86/intel/ifs/
H A Druntest.c89 cpumask_pr_args(cpu_smt_mask(cpu)), ifsd->cur_batch, ifsd->loaded_version); in message_not_tested()
95 cpumask_pr_args(cpu_smt_mask(cpu)), in message_not_tested()
99 cpumask_pr_args(cpu_smt_mask(cpu))); in message_not_tested()
102 cpumask_pr_args(cpu_smt_mask(cpu)), in message_not_tested()
106 cpumask_pr_args(cpu_smt_mask(cpu)), status.data); in message_not_tested()
123 cpumask_pr_args(cpu_smt_mask(cpu)), ifsd->cur_batch, ifsd->loaded_version); in message_fail()
165 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in wait_for_sibling_cpu()
201 first = cpumask_first(cpu_smt_mask(cpu)); in doscan()
322 first = cpumask_first(cpu_smt_mask(cpu)); in do_array_test()
375 first = cpumask_first(cpu_smt_mask(cpu)); in do_array_test_gen1()
[all …]
/linux/arch/powerpc/include/asm/
H A Dsmp.h140 #define cpu_smt_mask cpu_smt_mask macro
142 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask() function
/linux/include/linux/
H A Dtopology.h233 #if defined(CONFIG_SCHED_SMT) && !defined(cpu_smt_mask)
234 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask() function
/linux/kernel/sched/
H A Dext_idle.c89 const struct cpumask *smt = cpu_smt_mask(cpu); in scx_idle_test_and_clear_cpu()
630 for_each_cpu_and(cpu, cpu_smt_mask(prev_cpu), allowed) { in scx_select_cpu_dfl()
719 const struct cpumask *smt = cpu_smt_mask(cpu); in update_builtin_idle()
H A Dcore_sched.c244 const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq)); in __sched_core_account_forceidle()
H A Dcore.c405 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_lock()
417 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_unlock()
437 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in __sched_core_flip()
6146 smt_mask = cpu_smt_mask(cpu); in pick_next_task()
6446 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_cpu_starting()
6485 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_cpu_deactivate()
8602 if (cpumask_weight(cpu_smt_mask(cpu)) == 2) in sched_smt_present_inc()
8610 if (cpumask_weight(cpu_smt_mask(cpu)) == 2) in sched_smt_present_dec()
H A Dtopology.c1314 cpumask_andnot(mask, mask, cpu_smt_mask(cpu)); in init_sched_groups_capacity()
1767 return cpu_smt_mask(cpu); in tl_smt_mask()
H A Dfair.c1561 for_each_cpu(sibling, cpu_smt_mask(cpu)) { in is_core_idle()
7825 for_each_cpu(cpu, cpu_smt_mask(core)) { in __update_idle_core()
7848 for_each_cpu(cpu, cpu_smt_mask(core)) { in select_idle_core()
7868 cpumask_andnot(cpus, cpus, cpu_smt_mask(core)); in select_idle_core()
7879 for_each_cpu_and(cpu, cpu_smt_mask(target), p->cpus_ptr) { in select_idle_smt()
12004 cpumask_andnot(swb_cpus, swb_cpus, cpu_smt_mask(cpu)); in should_we_balance()
H A Dsched.h1507 for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) { in sched_core_cookie_match()
/linux/kernel/
H A Dstop_machine.c638 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in stop_core_cpuslocked()
H A Dworkqueue.c8191 return cpumask_test_cpu(cpu0, cpu_smt_mask(cpu1)); in cpus_share_smt()