/linux/arch/riscv/kernel/ |
H A D | sbi.c | 23 static int (*__sbi_rfence)(int fid, const struct cpumask *cpu_mask, 28 static unsigned long __sbi_v01_cpumask_to_hartmask(const struct cpumask *cpu_mask) in __sbi_v01_cpumask_to_hartmask() argument 40 for_each_cpu(cpuid, cpu_mask) { in __sbi_v01_cpumask_to_hartmask() 114 static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask, in __sbi_rfence_v01() argument 121 if (!cpu_mask || cpumask_empty(cpu_mask)) in __sbi_rfence_v01() 122 cpu_mask = cpu_online_mask; in __sbi_rfence_v01() 123 hart_mask = __sbi_v01_cpumask_to_hartmask(cpu_mask); in __sbi_rfence_v01() 166 static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask, in __sbi_rfence_v01() argument 257 static int __sbi_rfence_v02(int fid, const struct cpumask *cpu_mask, in __sbi_rfence_v02() argument 399 sbi_remote_fence_i(const struct cpumask * cpu_mask) sbi_remote_fence_i() argument 418 sbi_remote_sfence_vma_asid(const struct cpumask * cpu_mask,unsigned long start,unsigned long size,unsigned long asid) sbi_remote_sfence_vma_asid() argument 441 sbi_remote_hfence_gvma(const struct cpumask * cpu_mask,unsigned long start,unsigned long size) sbi_remote_hfence_gvma() argument 461 sbi_remote_hfence_gvma_vmid(const struct cpumask * cpu_mask,unsigned long start,unsigned long size,unsigned long vmid) sbi_remote_hfence_gvma_vmid() argument 480 sbi_remote_hfence_vvma(const struct cpumask * cpu_mask,unsigned long start,unsigned long size) sbi_remote_hfence_vvma() argument 501 sbi_remote_hfence_vvma_asid(const struct cpumask * cpu_mask,unsigned long start,unsigned long size,unsigned long asid) sbi_remote_hfence_vvma_asid() argument [all...] |
/linux/arch/x86/kernel/cpu/resctrl/ |
H A D | rdtgroup.c | 137 cpumask_var_t cpu_mask; in set_cache_qos_cfg() local 150 if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL)) in set_cache_qos_cfg() 157 for_each_cpu(cpu, &d->hdr.cpu_mask) in set_cache_qos_cfg() 158 cpumask_set_cpu(cpu, cpu_mask); in set_cache_qos_cfg() 161 cpumask_set_cpu(cpumask_any(&d->hdr.cpu_mask), cpu_mask); in set_cache_qos_cfg() 164 /* Update QOS_CFG MSR on all the CPUs in cpu_mask */ in set_cache_qos_cfg() 165 on_each_cpu_mask(cpu_mask, update, &enable, 1); in set_cache_qos_cfg() 167 free_cpumask_var(cpu_mask); in set_cache_qos_cfg() 258 smp_call_function_any(&d->hdr.cpu_mask, rdt_ctrl_updat in resctrl_arch_reset_all_ctrls() [all...] |
H A D | ctrlmondata.c | 30 if (!cpumask_test_cpu(smp_processor_id(), &d->hdr.cpu_mask)) in resctrl_arch_update_one() 80 smp_call_function_any(&d->hdr.cpu_mask, rdt_ctrl_update, &msr_param, 1); in resctrl_arch_update_domains()
|
H A D | core.c | 462 cpumask_set_cpu(cpu, &d->hdr.cpu_mask); in domain_add_cpu_ctrl() 475 cpumask_set_cpu(cpu, &d->hdr.cpu_mask); in domain_add_cpu_ctrl() 518 cpumask_set_cpu(cpu, &d->hdr.cpu_mask); in domain_add_cpu_mon() 536 cpumask_set_cpu(cpu, &d->hdr.cpu_mask); in domain_add_cpu_mon() 591 cpumask_clear_cpu(cpu, &d->hdr.cpu_mask); in domain_remove_cpu_ctrl() 592 if (cpumask_empty(&d->hdr.cpu_mask)) { in domain_remove_cpu_ctrl() 637 cpumask_clear_cpu(cpu, &d->hdr.cpu_mask); in domain_remove_cpu_mon() 638 if (cpumask_empty(&d->hdr.cpu_mask)) { in domain_remove_cpu_mon()
|
/linux/arch/powerpc/platforms/pseries/ |
H A D | hotplug-cpu.c | 152 * @cpu_mask: the returned CPU mask. 157 cpumask_var_t *cpu_mask) in find_cpu_id_range() argument 166 cpumask_clear(*cpu_mask); in find_cpu_id_range() 168 cpumask_set_cpu(cpu, *cpu_mask); in find_cpu_id_range() 192 while (!cpumask_empty(*cpu_mask)) { in find_cpu_id_range() 193 if (cpumask_subset(*cpu_mask, candidate_mask)) in find_cpu_id_range() 196 cpumask_shift_left(*cpu_mask, *cpu_mask, nthreads); in find_cpu_id_range() 199 if (!cpumask_empty(*cpu_mask)) in find_cpu_id_range() 218 cpumask_var_t cpu_mask; in pseries_add_processor() local [all...] |
/linux/scripts/gdb/linux/ |
H A D | cpus.py | 52 cpu_mask = {} variable 56 global cpu_mask 57 cpu_mask = {} 64 global cpu_mask 66 if mask_name in cpu_mask: 67 mask = cpu_mask[mask_name] 71 cpu_mask[mask_name] = mask
|
/linux/lib/ |
H A D | group_cpus.c | 127 * Active CPUs means the CPUs in '@cpu_mask AND @node_to_cpumask[]' 132 const struct cpumask *cpu_mask, in alloc_nodes_groups() argument 147 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in alloc_nodes_groups() 251 const struct cpumask *cpu_mask, in __group_cpus_evenly() argument 260 if (cpumask_empty(cpu_mask)) in __group_cpus_evenly() 263 nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk); in __group_cpus_evenly() 272 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in __group_cpus_evenly() 287 alloc_nodes_groups(numgrps, node_to_cpumask, cpu_mask, in __group_cpus_evenly() 297 cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]); in __group_cpus_evenly()
|
/linux/tools/power/x86/amd_pstate_tracer/ |
H A D | amd_pstate_trace.py | 155 def store_csv(cpu_int, time_pre_dec, time_post_dec, min_perf, des_perf, max_perf, freq_ghz, mperf, aperf, tsc, common_comm, load, duration_ms, sample_num, elapsed_time, cpu_mask): argument 160 if cpu_mask[cpu_int] == 0: 185 def read_trace_data(file_name, cpu_mask): argument 234 store_csv(cpu_int, time_pre_dec, time_post_dec, min_perf, des_perf, max_perf, freq_ghz, mperf, aperf, tsc, common_comm, load, duration_ms, sample_num, elapsed_time, cpu_mask) 239 ipt.split_csv(current_max_cpu, cpu_mask) 263 cpu_mask = zeros((MAX_CPUS,), dtype=int) variable 297 cpu_mask[int(p)] = 1 300 cpu_mask[i] = 1 330 read_trace_data(file_name, cpu_mask)
|
/linux/tools/power/x86/intel_pstate_tracer/ |
H A D | intel_pstate_tracer.py | 326 def store_csv(cpu_int, time_pre_dec, time_post_dec, core_busy, scaled, _from, _to, mperf, aperf, tsc, freq_ghz, io_boost, common_comm, load, duration_ms, sample_num, elapsed_time, tsc_ghz, cpu_mask): argument 331 if cpu_mask[cpu_int] == 0: 345 def split_csv(current_max_cpu, cpu_mask): argument 350 if cpu_mask[int(index)] != 0: 420 def read_trace_data(filename, cpu_mask): argument 480 store_csv(cpu_int, time_pre_dec, time_post_dec, core_busy, scaled, _from, _to, mperf, aperf, tsc, freq_ghz, io_boost, common_comm, load, duration_ms, sample_num, elapsed_time, tsc_ghz, cpu_mask) 486 split_csv(current_max_cpu, cpu_mask) 511 cpu_mask = zeros((MAX_CPUS,), dtype=int) variable 544 cpu_mask[int(p)] = 1 547 cpu_mask[ [all...] |
/linux/arch/x86/events/amd/ |
H A D | power.c | 39 * MSR_F15H_CU_PWR_ACCUMULATOR. cpu_mask represents CPU bit map of all cores 42 static cpumask_t cpu_mask; variable 150 return cpumap_print_to_pagebuf(true, buf, &cpu_mask); in get_attr_cpumask() 224 if (!cpumask_test_and_clear_cpu(cpu, &cpu_mask)) in power_cpu_exit() 234 cpumask_set_cpu(target, &cpu_mask); in power_cpu_exit() 245 * 1) If any CPU is set at cpu_mask in the same compute unit, do in power_cpu_init() 247 * 2) If no CPU is set at cpu_mask in the same compute unit, in power_cpu_init() 251 * sibling mask, then it is also in cpu_mask. in power_cpu_init() 255 cpumask_set_cpu(cpu, &cpu_mask); in power_cpu_init()
|
/linux/arch/riscv/include/asm/ |
H A D | sbi.h | 484 int sbi_remote_fence_i(const struct cpumask *cpu_mask); 486 int sbi_remote_sfence_vma_asid(const struct cpumask *cpu_mask, 490 int sbi_remote_hfence_gvma(const struct cpumask *cpu_mask, 493 int sbi_remote_hfence_gvma_vmid(const struct cpumask *cpu_mask, 497 int sbi_remote_hfence_vvma(const struct cpumask *cpu_mask, 500 int sbi_remote_hfence_vvma_asid(const struct cpumask *cpu_mask, 583 static inline int sbi_remote_fence_i(const struct cpumask *cpu_mask) { return -1; } in sbi_remote_fence_i() argument
|
/linux/arch/arm/mach-omap2/ |
H A D | clkt2xxx_virt_prcm_set.c | 42 static u16 cpu_mask; variable 82 if (!(ptr->flags & cpu_mask)) in omap2_round_to_table_rate() 106 if (!(prcm->flags & cpu_mask)) in omap2_select_table_rate() 179 if (!(prcm->flags & cpu_mask)) in omap2xxx_clkt_vps_check_bootloader_rates()
|
/linux/fs/resctrl/ |
H A D | rdtgroup.c | 366 mask = &rdtgrp->plr->d->hdr.cpu_mask; in rdtgroup_cpus_show() 373 cpumask_pr_args(&rdtgrp->cpu_mask)); in rdtgroup_cpus_show() 384 * Update the PGR_ASSOC MSR on all cpus in @cpu_mask, 390 update_closid_rmid(const struct cpumask *cpu_mask, struct rdtgroup *r) in update_closid_rmid() argument 400 on_each_cpu_mask(cpu_mask, resctrl_arch_sync_cpu_closid_rmid, p, 1); in update_closid_rmid() 410 cpumask_andnot(tmpmask, newmask, &prgrp->cpu_mask); in cpus_mon_write() 417 cpumask_andnot(tmpmask, &rdtgrp->cpu_mask, newmask); in cpus_mon_write() 420 cpumask_or(&prgrp->cpu_mask, &prgrp->cpu_mask, tmpmask); in cpus_mon_write() 428 cpumask_andnot(tmpmask, newmask, &rdtgrp->cpu_mask); in cpus_mon_write() [all...] |
H A D | pseudo_lock.c | 156 for_each_cpu(cpu, &plr->d->hdr.cpu_mask) { in pseudo_lock_cstates_constrain() 235 plr->cpu = cpumask_first(&plr->d->hdr.cpu_mask); in pseudo_lock_region_init() 534 if (!cpumask_empty(&rdtgrp->cpu_mask)) { in rdtgroup_locksetup_enter() 662 &d_i->hdr.cpu_mask); in rdtgroup_pseudo_locked_in_hierarchy() 670 if (cpumask_intersects(&d->hdr.cpu_mask, cpu_with_psl)) in rdtgroup_pseudo_locked_in_hierarchy() 710 cpu = cpumask_first(&plr->d->hdr.cpu_mask); in pseudo_lock_measure_cycles() 1029 if (!cpumask_subset(current->cpus_ptr, &plr->d->hdr.cpu_mask)) { in pseudo_lock_dev_mmap()
|
H A D | monitor.c | 379 if (!cpumask_test_cpu(cpu, &rr->d->hdr.cpu_mask)) in __mon_event_count() 499 if (cpumask_test_cpu(cpu, &d->hdr.cpu_mask)) in get_ctrl_domain_from_cpu() 664 d->cqm_work_cpu = cpumask_any_housekeeping(&d->hdr.cpu_mask, in cqm_handle_limbo() 688 cpu = cpumask_any_housekeeping(&dom->hdr.cpu_mask, exclude_cpu); in cqm_setup_limbo_handler() 731 d->mbm_work_cpu = cpumask_any_housekeeping(&d->hdr.cpu_mask, in mbm_handle_overflow() 760 cpu = cpumask_any_housekeeping(&dom->hdr.cpu_mask, exclude_cpu); in mbm_setup_overflow_handler()
|
/linux/drivers/irqchip/ |
H A D | irq-vf610-mscm-ir.c | 43 u16 cpu_mask; member 95 WARN_ON(irsprc & ~chip_data->cpu_mask); in vf610_mscm_ir_enable() 97 writew_relaxed(chip_data->cpu_mask, in vf610_mscm_ir_enable() 210 mscm_ir_data->cpu_mask = 0x1 << cpuid; in vf610_mscm_ir_of_init()
|
/linux/tools/testing/selftests/ring-buffer/ |
H A D | map_test.c | 140 cpu_set_t cpu_mask; in FIXTURE_SETUP() local 168 CPU_ZERO(&cpu_mask); in FIXTURE_SETUP() 169 CPU_SET(cpu, &cpu_mask); in FIXTURE_SETUP() 170 ASSERT_EQ(sched_setaffinity(0, sizeof(cpu_mask), &cpu_mask), 0); in FIXTURE_SETUP()
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | cpu_mask.c | 33 const char *cpu_mask; member 65 err = parse_cpu_mask_str(test_cases[i].cpu_mask, &mask, &n); in test_cpu_mask()
|
/linux/tools/perf/tests/ |
H A D | perf-record.c | 53 cpu_set_t cpu_mask; in test__PERF_RECORD() local 54 size_t cpu_mask_size = sizeof(cpu_mask); in test__PERF_RECORD() 114 err = sched__get_first_possible_cpu(evlist->workload.pid, &cpu_mask); in test__PERF_RECORD() 126 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, &cpu_mask) < 0) { in test__PERF_RECORD()
|
/linux/drivers/soc/ti/ |
H A D | knav_qmss_acc.c | 200 struct cpumask *cpu_mask; in knav_range_setup_acc_irq() local 207 cpu_mask = range->irqs[0].cpu_mask; in knav_range_setup_acc_irq() 211 cpu_mask = range->irqs[queue].cpu_mask; in knav_range_setup_acc_irq() 234 if (!ret && cpu_mask) { in knav_range_setup_acc_irq() 235 ret = irq_set_affinity_hint(irq, cpu_mask); in knav_range_setup_acc_irq()
|
/linux/arch/arm/mm/ |
H A D | proc-arm7tdmi.S | 89 .macro arm7tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req, \ 94 .long \cpu_mask
|
/linux/arch/x86/hyperv/ |
H A D | hv_apic.c | 206 ipi_arg.cpu_mask = 0; in __send_ipi_mask() 222 __set_bit(vcpu, (unsigned long *)&ipi_arg.cpu_mask); in __send_ipi_mask() 226 ipi_arg.cpu_mask); in __send_ipi_mask()
|
/linux/io_uring/ |
H A D | io-wq.c | 132 cpumask_var_t cpu_mask; member 733 wq->cpu_mask); in io_wq_worker() 785 set_cpus_allowed_ptr(tsk, wq->cpu_mask); in io_init_new_worker() 1245 if (!alloc_cpumask_var(&wq->cpu_mask, GFP_KERNEL)) in io_wq_create() 1247 cpuset_cpus_allowed(data->task, wq->cpu_mask); in io_wq_create() 1278 free_cpumask_var(wq->cpu_mask); in io_wq_create() 1346 free_cpumask_var(wq->cpu_mask); in io_wq_destroy() 1369 cpumask_set_cpu(od->cpu, worker->wq->cpu_mask); in io_wq_worker_affinity() 1371 cpumask_clear_cpu(od->cpu, worker->wq->cpu_mask); in io_wq_worker_affinity() 1417 cpumask_copy(tctx->io_wq->cpu_mask, mask); in io_wq_cpu_affinity() [all …]
|
/linux/arch/arm/mach-omap1/ |
H A D | clock_data.c | 692 u32 cpu_mask; variable 716 cpu_mask = 0; in omap1_clk_init() 718 cpu_mask |= CK_1710; in omap1_clk_init() 720 cpu_mask |= CK_16XX; in omap1_clk_init() 722 cpu_mask |= CK_1510; in omap1_clk_init() 724 cpu_mask |= CK_310; in omap1_clk_init() 791 if (!(c->cpu & cpu_mask)) in omap1_clk_init()
|
/linux/arch/arm/include/asm/ |
H A D | procinfo.h | 28 unsigned int cpu_mask; member
|