| /linux/kernel/irq/ |
| H A D | cpuhotplug.c | 58 const struct cpumask *affinity; in migrate_one_irq() local 105 affinity = irq_desc_get_pending_mask(desc); in migrate_one_irq() 107 affinity = irq_data_get_affinity_mask(d); in migrate_one_irq() 113 if (!cpumask_intersects(affinity, cpu_online_mask)) { in migrate_one_irq() 123 affinity = cpu_online_mask; in migrate_one_irq() 132 err = irq_do_set_affinity(d, affinity, false); in migrate_one_irq() 139 if (err == -ENOSPC && !irqd_affinity_is_managed(d) && affinity != cpu_online_mask) { in migrate_one_irq() 141 d->irq, cpumask_pr_args(affinity)); in migrate_one_irq() 143 affinity = cpu_online_mask; in migrate_one_irq() 146 err = irq_do_set_affinity(d, affinity, false); in migrate_one_irq() [all …]
|
| H A D | irqdesc.c | 57 if (!zalloc_cpumask_var_node(&desc->irq_common_data.affinity, in alloc_masks() 64 free_cpumask_var(desc->irq_common_data.affinity); in alloc_masks() 74 free_cpumask_var(desc->irq_common_data.affinity); in alloc_masks() 82 const struct cpumask *affinity) in desc_smp_init() argument 84 if (!affinity) in desc_smp_init() 85 affinity = irq_default_affinity; in desc_smp_init() 86 cpumask_copy(desc->irq_common_data.affinity, affinity); in desc_smp_init() 101 free_cpumask_var(desc->irq_common_data.affinity); in free_masks() 111 desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { } in desc_smp_init() argument 116 const struct cpumask *affinity, struct module *owner) in desc_set_defaults() argument [all …]
|
| H A D | irq_test.c | 120 struct irq_affinity_desc affinity = { in irq_shutdown_depth_test() local 128 virq = irq_test_setup_fake_irq(test, &affinity); in irq_shutdown_depth_test() 172 struct irq_affinity_desc affinity = { in irq_cpuhotplug_test() local 185 cpumask_copy(&affinity.mask, cpumask_of(1)); in irq_cpuhotplug_test() 187 virq = irq_test_setup_fake_irq(test, &affinity); in irq_cpuhotplug_test()
|
| H A D | irqdomain.c | 30 bool realloc, const struct irq_affinity_desc *affinity); 771 const struct irq_affinity_desc *affinity) in irq_create_mapping_affinity_locked() argument 780 affinity); in irq_create_mapping_affinity_locked() 810 const struct irq_affinity_desc *affinity) in irq_create_mapping_affinity() argument 831 virq = irq_create_mapping_affinity_locked(domain, hwirq, affinity); in irq_create_mapping_affinity() 1250 int node, const struct irq_affinity_desc *affinity) in irq_domain_alloc_descs() argument 1256 affinity); in irq_domain_alloc_descs() 1262 affinity); in irq_domain_alloc_descs() 1265 affinity); in irq_domain_alloc_descs() 1595 bool realloc, const struct irq_affinity_desc *affinity) in irq_domain_alloc_irqs_locked() argument [all …]
|
| /linux/tools/testing/selftests/rseq/ |
| H A D | basic_test.c | 18 cpu_set_t affinity, test_affinity; in test_cpu_pointer() local 21 sched_getaffinity(0, sizeof(affinity), &affinity); in test_cpu_pointer() 24 if (CPU_ISSET(i, &affinity)) { in test_cpu_pointer() 39 sched_setaffinity(0, sizeof(affinity), &affinity); in test_cpu_pointer()
|
| /linux/tools/perf/util/ |
| H A D | affinity.h | 8 struct affinity { struct 14 void affinity__cleanup(struct affinity *a); argument 15 void affinity__set(struct affinity *a, int cpu); 16 int affinity__setup(struct affinity *a);
|
| H A D | affinity.c | 25 int affinity__setup(struct affinity *a) in affinity__setup() 49 void affinity__set(struct affinity *a, int cpu) in affinity__set() 72 static void __affinity__cleanup(struct affinity *a) in __affinity__cleanup() 82 void affinity__cleanup(struct affinity *a) in affinity__cleanup()
|
| H A D | mmap.c | 97 static int perf_mmap__aio_bind(struct mmap *map, int idx, struct perf_cpu cpu, int affinity) in perf_mmap__aio_bind() argument 105 if (affinity != PERF_AFFINITY_SYS && cpu__max_node() > 1) { in perf_mmap__aio_bind() 141 struct perf_cpu cpu __maybe_unused, int affinity __maybe_unused) in perf_mmap__aio_bind() 175 ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity); in perf_mmap__aio_mmap() 268 if (mp->affinity == PERF_AFFINITY_NODE && cpu__max_node() > 1) in perf_mmap__setup_affinity_mask() 270 else if (mp->affinity == PERF_AFFINITY_CPU) in perf_mmap__setup_affinity_mask() 284 if (mp->affinity != PERF_AFFINITY_SYS && in mmap__mmap()
|
| H A D | evlist.h | 196 int affinity, int flush, int comp_level); 363 struct affinity *affinity; member 376 #define evlist__for_each_cpu(evlist_cpu_itr, evlist, affinity) \ argument 377 for ((evlist_cpu_itr) = evlist__cpu_begin(evlist, affinity); \ 382 struct evlist_cpu_iterator evlist__cpu_begin(struct evlist *evlist, struct affinity *affinity);
|
| H A D | evlist.c | 353 struct evlist_cpu_iterator evlist__cpu_begin(struct evlist *evlist, struct affinity *affinity) in evlist__cpu_begin() argument 362 .affinity = affinity, in evlist__cpu_begin() 370 if (itr.affinity) { in evlist__cpu_begin() 372 affinity__set(itr.affinity, itr.cpu.cpu); in evlist__cpu_begin() 401 if (evlist_cpu_itr->affinity) in evlist_cpu_iterator__next() 402 affinity__set(evlist_cpu_itr->affinity, evlist_cpu_itr->cpu.cpu); in evlist_cpu_iterator__next() 447 struct affinity saved_affinity, *affinity = NULL; in __evlist__disable() local 454 affinity = &saved_affinity; in __evlist__disable() 459 evlist__for_each_cpu(evlist_cpu_itr, evlist, affinity) { in __evlist__disable() 477 affinity__cleanup(affinity); in __evlist__disable() [all …]
|
| /linux/Documentation/arch/arm64/ |
| H A D | asymmetric-32bit.rst | 51 CPU affinity. 68 On a homogeneous system, the CPU affinity of a task is preserved across 71 affinity mask contains 64-bit-only CPUs. In this situation, the kernel 72 determines the new affinity mask as follows: 74 1. If the 32-bit-capable subset of the affinity mask is not empty, 75 then the affinity is restricted to that subset and the old affinity 84 affinity of the task is then changed to match the 32-bit-capable 87 3. On failure (i.e. out of memory), the affinity is changed to the set 91 invalidate the affinity mask saved in (1) and attempt to restore the CPU 92 affinity of the task using the saved mask if it was previously valid. [all …]
|
| /linux/tools/virtio/ringtest/ |
| H A D | run-on-all.sh | 20 "$@" --host-affinity $HOST_AFFINITY --guest-affinity $cpu 24 "$@" --host-affinity $HOST_AFFINITY
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | affinity.c | 956 struct hfi1_affinity_node_list *affinity) in find_hw_thread_mask() argument 961 cpumask_copy(hw_thread_mask, &affinity->proc.mask); in find_hw_thread_mask() 963 if (affinity->num_core_siblings == 0) in find_hw_thread_mask() 966 num_cores = rounddown(node_affinity.num_online_cpus / affinity->num_core_siblings, in find_hw_thread_mask() 984 struct hfi1_affinity_node_list *affinity = &node_affinity; in hfi1_get_proc_affinity() local 985 struct cpu_mask_set *set = &affinity->proc; in hfi1_get_proc_affinity() 1043 mutex_lock(&affinity->lock); in hfi1_get_proc_affinity() 1073 for (i = 0; i < affinity->num_core_siblings; i++) { in hfi1_get_proc_affinity() 1074 find_hw_thread_mask(i, hw_thread_mask, affinity); in hfi1_get_proc_affinity() 1144 mutex_unlock(&affinity->lock); in hfi1_get_proc_affinity() [all …]
|
| /linux/arch/arm64/kernel/ |
| H A D | setup.c | 113 u32 i, affinity, fs[4], bits[4], ls; in smp_build_mpidr_hash() local 127 affinity = MPIDR_AFFINITY_LEVEL(mask, i); in smp_build_mpidr_hash() 133 ls = fls(affinity); in smp_build_mpidr_hash() 134 fs[i] = affinity ? ffs(affinity) - 1 : 0; in smp_build_mpidr_hash()
|
| /linux/kernel/ |
| H A D | kthread.c | 381 cpumask_var_t affinity; in kthread_affine_node() local 388 if (!zalloc_cpumask_var(&affinity, GFP_KERNEL)) { in kthread_affine_node() 403 kthread_fetch_affinity(kthread, affinity); in kthread_affine_node() 404 set_cpus_allowed_ptr(current, affinity); in kthread_affine_node() 407 free_cpumask_var(affinity); in kthread_affine_node() 859 cpumask_var_t affinity; in kthread_affine_preferred() local 870 if (!zalloc_cpumask_var(&affinity, GFP_KERNEL)) in kthread_affine_preferred() 883 kthread_fetch_affinity(kthread, affinity); in kthread_affine_preferred() 887 do_set_cpus_allowed(p, affinity); in kthread_affine_preferred() 892 free_cpumask_var(affinity); in kthread_affine_preferred() [all …]
|
| /linux/Documentation/core-api/irq/ |
| H A D | irq-affinity.rst | 2 SMP IRQ affinity 14 IRQ affinity then the value will not change from the default of all CPUs. 16 /proc/irq/default_smp_affinity specifies default affinity mask that applies 17 to all non-active IRQs. Once IRQ is allocated/activated its affinity bitmask
|
| /linux/Documentation/translations/zh_CN/core-api/irq/ |
| H A D | irq-affinity.rst | 3 :Original: Documentation/core-api/irq/irq-affinity.rst 9 .. _cn_irq-affinity.rst: 23 (IRQ affinity),那么所有cpu的默认值将保持不变(即关联到所有CPU).
|
| /linux/arch/alpha/kernel/ |
| H A D | sys_dp264.c | 136 cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity) in cpu_set_irq_affinity() argument 142 if (cpumask_test_cpu(cpu, &affinity)) in cpu_set_irq_affinity() 151 dp264_set_affinity(struct irq_data *d, const struct cpumask *affinity, in dp264_set_affinity() argument 155 cpu_set_irq_affinity(d->irq, *affinity); in dp264_set_affinity() 163 clipper_set_affinity(struct irq_data *d, const struct cpumask *affinity, in clipper_set_affinity() argument 167 cpu_set_irq_affinity(d->irq - 16, *affinity); in clipper_set_affinity()
|
| H A D | sys_titan.c | 135 titan_cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity) in titan_cpu_set_irq_affinity() argument 140 if (cpumask_test_cpu(cpu, &affinity)) in titan_cpu_set_irq_affinity() 149 titan_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, in titan_set_irq_affinity() argument 154 titan_cpu_set_irq_affinity(irq - 16, *affinity); in titan_set_irq_affinity()
|
| /linux/drivers/irqchip/ |
| H A D | irq-bcm7038-l1.c | 47 u8 affinity[MAX_WORDS * IRQS_PER_WORD]; member 179 __bcm7038_l1_unmask(d, intc->affinity[d->hwirq]); in bcm7038_l1_unmask() 189 __bcm7038_l1_mask(d, intc->affinity[d->hwirq]); in bcm7038_l1_mask() 208 was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] & in bcm7038_l1_set_affinity() 210 __bcm7038_l1_mask(d, intc->affinity[hw]); in bcm7038_l1_set_affinity() 211 intc->affinity[hw] = first_cpu; in bcm7038_l1_set_affinity()
|
| /linux/tools/testing/selftests/rcutorture/bin/ |
| H A D | kvm-test-1-run-batch.sh | 65 print "echo No CPU-affinity information, so no taskset command."; 71 print "echo " scenario ": Bogus CPU-affinity information, so no taskset command.";
|
| /linux/tools/perf/ |
| H A D | builtin-record.c | 100 struct mmap_cpu_mask affinity; member 1328 if (opts->affinity != PERF_AFFINITY_SYS) in record__mmap_evlist() 1334 opts->nr_cblocks, opts->affinity, in record__mmap_evlist() 1539 if (rec->opts.affinity != PERF_AFFINITY_SYS && in record__adjust_affinity() 1540 !bitmap_equal(thread->mask->affinity.bits, map->affinity_mask.bits, in record__adjust_affinity() 1541 thread->mask->affinity.nbits)) { in record__adjust_affinity() 1542 bitmap_zero(thread->mask->affinity.bits, thread->mask->affinity.nbits); in record__adjust_affinity() 1543 bitmap_or(thread->mask->affinity.bits, thread->mask->affinity.bits, in record__adjust_affinity() 1544 map->affinity_mask.bits, thread->mask->affinity.nbits); in record__adjust_affinity() 1545 sched_setaffinity(0, MMAP_CPU_MASK_BYTES(&thread->mask->affinity), in record__adjust_affinity() [all …]
|
| /linux/arch/arm64/boot/dts/apple/ |
| H A D | t6021.dtsi | 54 e-core-pmu-affinity { 59 p-core-pmu-affinity {
|
| H A D | t6001.dtsi | 53 e-core-pmu-affinity { 58 p-core-pmu-affinity {
|
| /linux/include/ras/ |
| H A D | ras_event.h | 182 __field(u8, affinity) 187 __entry->affinity = proc->affinity_level; 189 __entry->affinity = ~0; 206 __entry->affinity, __entry->mpidr, __entry->midr,
|