Searched refs:cpu_smt_mask (Results 1 – 11 of 11) sorted by relevance
| /linux/drivers/platform/x86/intel/ifs/ |
| H A D | runtest.c | 89 cpumask_pr_args(cpu_smt_mask(cpu)), ifsd->cur_batch, ifsd->loaded_version); in message_not_tested() 95 cpumask_pr_args(cpu_smt_mask(cpu)), in message_not_tested() 99 cpumask_pr_args(cpu_smt_mask(cpu))); in message_not_tested() 102 cpumask_pr_args(cpu_smt_mask(cpu)), in message_not_tested() 106 cpumask_pr_args(cpu_smt_mask(cpu)), status.data); in message_not_tested() 123 cpumask_pr_args(cpu_smt_mask(cpu)), ifsd->cur_batch, ifsd->loaded_version); in message_fail() 165 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in wait_for_sibling_cpu() 201 first = cpumask_first(cpu_smt_mask(cpu)); in doscan() 322 first = cpumask_first(cpu_smt_mask(cpu)); in do_array_test() 375 first = cpumask_first(cpu_smt_mask(cpu)); in do_array_test_gen1() [all …]
|
| /linux/arch/powerpc/include/asm/ |
| H A D | smp.h | 140 #define cpu_smt_mask cpu_smt_mask macro 142 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask() function
|
| /linux/include/linux/ |
| H A D | topology.h | 233 #if defined(CONFIG_SCHED_SMT) && !defined(cpu_smt_mask) 234 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask() function
|
| /linux/kernel/sched/ |
| H A D | core_sched.c | 244 const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq)); in __sched_core_account_forceidle()
|
| H A D | ext_idle.c | 89 const struct cpumask *smt = cpu_smt_mask(cpu); in scx_idle_test_and_clear_cpu() 699 const struct cpumask *smt = cpu_smt_mask(cpu); in update_builtin_idle()
|
| H A D | core.c | 405 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_lock() 417 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_unlock() 437 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in __sched_core_flip() 6062 smt_mask = cpu_smt_mask(cpu); in pick_next_task() 6362 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_cpu_starting() 6401 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in sched_core_cpu_deactivate() 8333 if (cpumask_weight(cpu_smt_mask(cpu)) == 2) in sched_smt_present_inc() 8341 if (cpumask_weight(cpu_smt_mask(cpu)) == 2) in sched_smt_present_dec()
|
| H A D | topology.c | 1302 cpumask_andnot(mask, mask, cpu_smt_mask(cpu)); in init_sched_groups_capacity() 1753 return cpu_smt_mask(cpu); in tl_smt_mask()
|
| H A D | fair.c | 1469 for_each_cpu(sibling, cpu_smt_mask(cpu)) { in is_core_idle() 7604 for_each_cpu(cpu, cpu_smt_mask(core)) { in __update_idle_core() 7627 for_each_cpu(cpu, cpu_smt_mask(core)) { in select_idle_core() 7646 cpumask_andnot(cpus, cpus, cpu_smt_mask(core)); in select_idle_core() 7657 for_each_cpu_and(cpu, cpu_smt_mask(target), p->cpus_ptr) { in select_idle_smt() 11804 cpumask_andnot(swb_cpus, swb_cpus, cpu_smt_mask(cpu)); in should_we_balance()
|
| H A D | sched.h | 1498 for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) { in sched_core_cookie_match()
|
| /linux/kernel/ |
| H A D | stop_machine.c | 638 const struct cpumask *smt_mask = cpu_smt_mask(cpu); in stop_core_cpuslocked()
|
| H A D | workqueue.c | 8124 return cpumask_test_cpu(cpu0, cpu_smt_mask(cpu1)); in cpus_share_smt()
|