/linux/kernel/irq/ |
H A D | migration.c | 29 if (!cpumask_intersects(desc->pending_mask, cpu_online_mask)) { in irq_fixup_move_pending() 87 if (cpumask_intersects(desc->pending_mask, cpu_online_mask)) { in irq_move_masked_irq()
|
H A D | cpuhotplug.c | 40 !cpumask_intersects(m, cpu_online_mask)) { in irq_needs_fixup() 113 if (!cpumask_intersects(affinity, cpu_online_mask)) { in migrate_one_irq()
|
H A D | manage.c | 258 if (!cpumask_intersects(tmp_mask, cpu_online_mask)) in irq_do_set_affinity() 590 if (cpumask_intersects(desc->irq_common_data.affinity, in irq_setup_affinity() 605 if (cpumask_intersects(&mask, nodemask)) in irq_setup_affinity()
|
H A D | chip.c | 180 if (!cpumask_intersects(aff, cpu_online_mask)) { in __irq_startup_managed()
|
H A D | msi.c | 1252 !cpumask_intersects(irq_data_get_affinity_mask(irqd), in msi_init_virq()
|
/linux/kernel/cgroup/ |
H A D | cpuset.c | 388 while (!cpumask_intersects(cs->effective_cpus, pmask)) in guarantee_active_cpus() 538 if (cpumask_intersects(xcpus1, xcpus2)) in cpusets_are_exclusive() 672 return cpumask_intersects(a->effective_cpus, b->effective_cpus); in cpusets_overlap() 825 cpumask_intersects(cp->cpus_allowed, in generate_sched_domains() 1240 (!cpumask_intersects(xcpus, cpu_active_mask) && in tasks_nocpu_error() 1407 if (cpumask_intersects(xcpus, sibling->exclusive_cpus)) { in compute_effective_exclusive_cpumask() 1412 if (cpumask_intersects(xcpus, sibling->effective_xcpus)) { in compute_effective_exclusive_cpumask() 1463 WARN_ON_ONCE(cpumask_intersects(tmp->new_cpus, subpartitions_cpus)); in remote_partition_enable() 1464 if (!cpumask_intersects(tmp->new_cpus, cpu_active_mask) || in remote_partition_enable() 1560 WARN_ON_ONCE(cpumask_intersects(tm in remote_cpus_update() [all...] |
/linux/arch/mips/kernel/ |
H A D | mips-mt-fpaff.c | 122 cpumask_intersects(new_mask, &mt_fpu_cpumask)) { in mipsmt_sys_sched_setaffinity()
|
H A D | traps.c | 945 if (cpumask_intersects(¤t->cpus_mask, &mt_fpu_cpumask)) { in mt_ase_fp_affinity()
|
/linux/arch/mips/mm/ |
H A D | context.c | 249 cpumask_intersects(&tlb_flush_pending, &cpu_sibling_map[cpu])) { in check_switch_mmu_context()
|
/linux/drivers/sh/intc/ |
H A D | chip.c | 72 if (!cpumask_intersects(cpumask, cpu_online_mask)) in intc_set_affinity()
|
/linux/lib/ |
H A D | group_cpus.c | 92 if (cpumask_intersects(mask, node_to_cpumask[n])) { in get_nodes_in_cpumask()
|
/linux/arch/arm64/kernel/ |
H A D | topology.c | 316 if (!cpumask_intersects(policy->related_cpus, in arch_freq_get_on_cpu()
|
/linux/arch/arm/kernel/ |
H A D | hw_breakpoint.c | 1108 if (cpumask_intersects(&debug_err_mask, cpumask_of(cpu))) { in reset_ctrl_regs() 1128 if (cpumask_intersects(&debug_err_mask, cpumask_of(cpu))) { in reset_ctrl_regs()
|
/linux/kernel/sched/ |
H A D | topology.c | 93 cpumask_intersects(groupmask, sched_group_span(group))) { in sched_domain_debug_one() 1403 if (cpumask_intersects(sd_span, cpu_capacity_span(entry))) in asym_cpu_capacity_classify() 1405 else if (cpumask_intersects(cpu_map, cpu_capacity_span(entry))) in asym_cpu_capacity_classify() 2426 if (cpumask_intersects(tl_cpu_mask, covered)) in topology_span_sane()
|
H A D | ext_idle.c | 98 * NOTE: Use cpumask_intersects() and cpumask_test_cpu() to in scx_idle_test_and_clear_cpu() 102 if (cpumask_intersects(smt, idle_smts)) in scx_idle_test_and_clear_cpu()
|
H A D | deadline.c | 2858 if (!cpumask_intersects(src_rd->span, ctx->new_mask)) { in set_cpus_allowed_dl()
|
H A D | fair.c | 7399 if (!cpumask_intersects(sched_domain_span(sd), p->cpus_ptr)) in sched_balance_find_dst_cpu() 8608 if (!cpumask_intersects(p->cpus_ptr, cpumask)) in set_task_max_allowed_capacity() 10705 if (!cpumask_intersects(sched_group_span(group), in sched_balance_find_dst_group()
|
/linux/fs/resctrl/ |
H A D | pseudo_lock.c | 670 if (cpumask_intersects(&d->hdr.cpu_mask, cpu_with_psl)) in rdtgroup_pseudo_locked_in_hierarchy()
|
/linux/drivers/hwmon/ |
H A D | coretemp.c | 716 if (!cpumask_intersects(&pdata->cpumask, topology_sibling_cpumask(cpu))) in coretemp_cpu_online()
|
/linux/include/linux/ |
H A D | cpumask.h | 793 * cpumask_intersects - (*src1p & *src2p) != 0 801 bool cpumask_intersects(const struct cpumask *src1p, const struct cpumask *src2p) in cpumask_intersects() function
|
/linux/arch/powerpc/platforms/cell/spufs/ |
H A D | sched.c | 160 if (cpumask_intersects(mask, &ctx->cpus_allowed)) in __node_allowed()
|
/linux/arch/powerpc/sysdev/xive/ |
H A D | common.c | 729 if (!cpumask_intersects(cpumask, cpu_online_mask)) in xive_irq_set_affinity()
|
/linux/kernel/ |
H A D | workqueue.c | 2704 if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask)) in unbind_worker() 5209 if (!cpumask_intersects(attrs->__pod_cpumask, wq_online_cpumask)) { in wq_calc_pod_cpumask() 7685 if (!cpumask_intersects(wq_unbound_cpumask, mask)) { in restrict_unbound_cpumask()
|
/linux/kernel/trace/ |
H A D | trace_events_filter.c | 661 return cpumask_intersects(mask, cmp); in do_filter_cpumask()
|
/linux/drivers/vdpa/vdpa_user/ |
H A D | vduse_dev.c | 1497 if (!cpumask_intersects(new_value, cpu_online_mask)) in irq_cb_affinity_store()
|