/linux/tools/lib/ |
H A D | rbtree.c | 230 struct rb_node *node = NULL, *sibling, *tmp1, *tmp2; in ____rb_erase_color() local 240 sibling = parent->rb_right; in ____rb_erase_color() 241 if (node != sibling) { /* node == parent->rb_left */ in ____rb_erase_color() 242 if (rb_is_red(sibling)) { in ____rb_erase_color() 252 tmp1 = sibling->rb_left; in ____rb_erase_color() 254 WRITE_ONCE(sibling->rb_left, parent); in ____rb_erase_color() 256 __rb_rotate_set_parents(parent, sibling, root, in ____rb_erase_color() 258 augment_rotate(parent, sibling); in ____rb_erase_color() 259 sibling = tmp1; in ____rb_erase_color() 261 tmp1 = sibling in ____rb_erase_color() [all...] |
/linux/lib/ |
H A D | rbtree.c | 230 struct rb_node *node = NULL, *sibling, *tmp1, *tmp2; in ____rb_erase_color() local 240 sibling = parent->rb_right; in ____rb_erase_color() 241 if (node != sibling) { /* node == parent->rb_left */ in ____rb_erase_color() 242 if (rb_is_red(sibling)) { in ____rb_erase_color() 252 tmp1 = sibling->rb_left; in ____rb_erase_color() 254 WRITE_ONCE(sibling->rb_left, parent); in ____rb_erase_color() 256 __rb_rotate_set_parents(parent, sibling, root, in ____rb_erase_color() 258 augment_rotate(parent, sibling); in ____rb_erase_color() 259 sibling = tmp1; in ____rb_erase_color() 261 tmp1 = sibling in ____rb_erase_color() [all...] |
/linux/kernel/ |
H A D | resource.c | 64 while (!p->sibling && p->parent) { in next_resource() 69 return p->sibling; in next_resource() 197 new->sibling = tmp; in __request_resource() 202 p = &tmp->sibling; in __request_resource() 220 *p = tmp->sibling; in __release_resource() 222 for (chd = tmp->child;; chd = chd->sibling) { in __release_resource() 224 if (!(chd->sibling)) in __release_resource() 228 chd->sibling = tmp->sibling; in __release_resource() 233 p = &tmp->sibling; in __release_resource() [all...] |
/linux/drivers/media/platform/amphion/ |
H A D | vpu_helpers.c | 66 const struct vpu_format *sibling; in vpu_helper_find_sibling() local 69 if (!fmt || !fmt->sibling) in vpu_helper_find_sibling() 72 sibling = vpu_helper_find_format(inst, type, fmt->sibling); in vpu_helper_find_sibling() 73 if (!sibling || sibling->sibling != fmt->pixfmt || in vpu_helper_find_sibling() 74 sibling->comp_planes != fmt->comp_planes) in vpu_helper_find_sibling() 77 return sibling; in vpu_helper_find_sibling() 82 const struct vpu_format *sibling; in vpu_helper_match_format() local [all...] |
/linux/arch/x86/kernel/ |
H A D | smpboot.c | 109 /* Representing CPUs for which sibling maps can be computed */ 333 "sched: CPU #%d's %s-sibling CPU #%d is not on the same node! " in topology_sane() 1140 int sibling; in remove_siblinginfo() local 1143 for_each_cpu(sibling, topology_core_cpumask(cpu)) { in remove_siblinginfo() 1144 cpumask_clear_cpu(cpu, topology_core_cpumask(sibling)); in remove_siblinginfo() 1146 * last thread sibling in this cpu core going down in remove_siblinginfo() 1149 cpu_data(sibling).booted_cores--; in remove_siblinginfo() 1152 for_each_cpu(sibling, topology_die_cpumask(cpu)) in remove_siblinginfo() 1153 cpumask_clear_cpu(cpu, topology_die_cpumask(sibling)); in remove_siblinginfo() 1155 for_each_cpu(sibling, topology_sibling_cpumas in remove_siblinginfo() [all...] |
/linux/arch/sparc/kernel/ |
H A D | pci_psycho.c | 191 if (pbm->sibling) in psycho_ue_intr() 192 psycho_check_iommu_error(pbm->sibling, afsr, afar, UE_ERR); in psycho_ue_intr() 528 pbm->sibling = psycho_find_sibling(upa_portid); in psycho_probe() 529 if (pbm->sibling) { in psycho_probe() 530 iommu = pbm->sibling->iommu; in psycho_probe() 565 if (!pbm->sibling) { in psycho_probe() 581 if (pbm->sibling) in psycho_probe() 582 pbm->sibling->sibling = pbm; in psycho_probe() 589 if (!pbm->sibling) in psycho_probe() [all...] |
/linux/drivers/base/ |
H A D | cacheinfo.c | 455 unsigned int sibling, index, sib_index; in cache_shared_cpu_map_remove() local 459 for_each_cpu(sibling, &this_leaf->shared_cpu_map) { in cache_shared_cpu_map_remove() 460 if (sibling == cpu || !per_cpu_cacheinfo(sibling)) in cache_shared_cpu_map_remove() 463 for (sib_index = 0; sib_index < cache_leaves(sibling); sib_index++) { in cache_shared_cpu_map_remove() 464 sib_leaf = per_cpu_cacheinfo_idx(sibling, sib_index); in cache_shared_cpu_map_remove() 477 cpumask_clear_cpu(sibling, &this_leaf->shared_cpu_map); in cache_shared_cpu_map_remove() 937 unsigned int sibling; in cpu_map_shared_cache() local 952 /* shared_cpu_map of offlined CPU will be cleared, so use sibling map */ in cpu_map_shared_cache() 953 for_each_cpu(sibling, in cpu_map_shared_cache() [all...] |
/linux/tools/testing/selftests/seccomp/ |
H A D | seccomp_bpf.c | 2549 struct tsync_sibling sibling[TSYNC_SIBLINGS]; in FIXTURE() local 2571 memset(&self->sibling, 0, sizeof(self->sibling)); in FIXTURE_SETUP() 2586 self->sibling[0].tid = 0; in FIXTURE_SETUP() 2587 self->sibling[0].cond = &self->cond; in FIXTURE_SETUP() 2588 self->sibling[0].started = &self->started; in FIXTURE_SETUP() 2589 self->sibling[0].mutex = &self->mutex; in FIXTURE_SETUP() 2590 self->sibling[0].diverge = 0; in FIXTURE_SETUP() 2591 self->sibling[0].num_waits = 1; in FIXTURE_SETUP() 2592 self->sibling[ in FIXTURE_SETUP() 2661 tsync_start_sibling(struct tsync_sibling * sibling) tsync_start_sibling() argument 4878 pthread_t sibling; TEST() local [all...] |
/linux/drivers/powercap/ |
H A D | dtpm.c | 72 list_for_each_entry(child, &dtpm->children, sibling) { in __get_power_uw() 91 list_for_each_entry(child, &dtpm->children, sibling) { in __dtpm_rebalance_weight() 178 list_del(&dtpm->sibling); in dtpm_release_zone() 231 list_for_each_entry(child, &dtpm->children, sibling) { in __set_power_limit_uw() 323 INIT_LIST_HEAD(&dtpm->sibling); in dtpm_init() 398 list_add_tail(&dtpm->sibling, &parent->children); in dtpm_register() 622 list_for_each_entry_safe(child, aux, &dtpm->children, sibling) in __dtpm_destroy_hierarchy()
|
/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_execlists_submission.c | 557 * After this point, the rq may be transferred to a new sibling, so in kick_siblings() 1411 goto unlock; /* lost the race to a sibling */ in execlists_dequeue() 1424 return; /* leave this for another sibling */ in execlists_dequeue() 3621 * the rb_node into a sibling. in rcu_virtual_context_destroy() 3627 struct intel_engine_cs *sibling = ve->siblings[n]; in rcu_virtual_context_destroy() local 3628 struct rb_node *node = &ve->nodes[sibling->id].rb; in rcu_virtual_context_destroy() 3633 spin_lock_irq(&sibling->sched_engine->lock); in rcu_virtual_context_destroy() 3637 rb_erase_cached(node, &sibling->execlists.virtual); in rcu_virtual_context_destroy() 3639 spin_unlock_irq(&sibling->sched_engine->lock); in rcu_virtual_context_destroy() 3682 * Pick a random sibling o in virtual_engine_initial_hint() 3746 virtual_get_sibling(struct intel_engine_cs * engine,unsigned int sibling) virtual_get_sibling() argument 3817 struct intel_engine_cs *sibling = READ_ONCE(ve->siblings[n]); virtual_submission_tasklet() local 3990 struct intel_engine_cs *sibling = siblings[n]; execlists_create_virtual() local [all...] |
/linux/Documentation/devicetree/ |
H A D | of_unittest.rst | 91 struct device_node *sibling; 96 considering only child and sibling pointers. There exists another pointer, 98 a particular level the child node and all the sibling nodes will have a parent 161 replaces the current child and turns it into its sibling. So, when the testcase 202 sibling compared to the earlier structure (Figure 2). After attaching first 204 (i.e. test-child0) to become a sibling and makes itself a child node, 223 node's parent to its sibling or attaches the previous sibling to the given 224 node's sibling, as appropriate. That is it :)
|
/linux/kernel/sched/ |
H A D | topology.c | 915 struct sched_domain *sibling; in build_balance_mask() local 921 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask() 928 if (!sibling->child) in build_balance_mask() 932 if (!cpumask_equal(sg_span, sched_domain_span(sibling->child))) in build_balance_mask() 1000 find_descended_sibling(struct sched_domain *sd, struct sched_domain *sibling) in find_descended_sibling() argument 1006 while (sibling->child && in find_descended_sibling() 1007 !cpumask_subset(sched_domain_span(sibling->child), in find_descended_sibling() 1009 sibling = sibling->child; in find_descended_sibling() 1016 while (sibling in find_descended_sibling() 1031 struct sched_domain *sibling; build_overlap_sched_groups() local [all...] |
/linux/Documentation/admin-guide/hw-vuln/ |
H A D | cross-thread-rsb.rst | 8 predictions vulnerability. When running in SMT mode and one sibling thread 9 transitions out of C0 state, the other sibling thread could use return target 10 predictions from the sibling thread that transitioned out of C0. 16 being consumed by the sibling thread. 52 used by RET predictions in the sibling thread following a 1T/2T switch. In
|
H A D | core-scheduling.rst | 105 During a schedule() event on any sibling of a core, the highest priority task on 106 the sibling's core is picked and assigned to the sibling calling schedule(), if 107 the sibling has the task enqueued. For rest of the siblings in the core, 114 switch to the new task immediately. If an idle task is selected for a sibling, 115 then the sibling is considered to be in a `forced idle` state. I.e., it may 127 task. If a sibling does not have a trusted task to run, it will be forced idle 131 the sibling to force it into idle. This results in 4 cases which need to be 189 sibling. Such attacks are possible for any combination of sibling CP [all...] |
/linux/drivers/pci/hotplug/ |
H A D | acpiphp_glue.c | 161 list_for_each_entry_safe(func, tmp, &slot->funcs, sibling) in free_bridge() 330 list_add_tail(&newfunc->sibling, &slot->funcs); in acpiphp_add_context() 345 list_for_each_entry(func, &slot->funcs, sibling) { in cleanup_bridge() 398 list_for_each_entry(func, &slot->funcs, sibling) { in acpiphp_set_acpi_region() 422 list_for_each_entry(func, &slot->funcs, sibling) { in check_hotplug_bridge() 434 list_for_each_entry(func, &slot->funcs, sibling) { in acpiphp_rescan_slot() 536 list_for_each_entry(func, &slot->funcs, sibling) { in enable_slot() 569 list_for_each_entry(func, &slot->funcs, sibling) in disable_slot() 605 list_for_each_entry(func, &slot->funcs, sibling) { in get_slot_status() 1014 list_for_each_entry(func, &slot->funcs, sibling) in acpiphp_disable_and_eject_slot() [all...] |
/linux/drivers/sh/clk/ |
H A D | core.c | 217 list_del_init(&child->sibling); in clk_reparent() 219 list_add(&child->sibling, &parent->children); in clk_reparent() 230 list_for_each_entry(clkp, &tclk->children, sibling) { in propagate_rate() 321 list_for_each_entry(clkp, &root_clks, sibling) { in recalculate_root_clocks() 434 list_add(&clk->sibling, &clk->parent->children); in clk_register() 436 list_add(&clk->sibling, &root_clks); in clk_register() 455 list_del(&clk->sibling); in clk_unregister()
|
/linux/drivers/perf/hisilicon/ |
H A D | hisi_pcie_pmu.c | 336 struct perf_event *sibling, *leader = event->group_leader; in hisi_pcie_pmu_validate_event_group() local 350 for_each_sibling_event(sibling, event->group_leader) { in hisi_pcie_pmu_validate_event_group() 351 if (is_software_event(sibling)) in hisi_pcie_pmu_validate_event_group() 354 if (sibling->pmu != event->pmu) in hisi_pcie_pmu_validate_event_group() 362 if (hisi_pcie_pmu_cmp_event(event_group[num], sibling)) in hisi_pcie_pmu_validate_event_group() 375 event_group[counters++] = sibling; in hisi_pcie_pmu_validate_event_group() 426 struct perf_event *sibling; in hisi_pcie_pmu_get_event_idx() local 430 sibling = pcie_pmu->hw_events[idx]; in hisi_pcie_pmu_get_event_idx() 431 if (!sibling) { in hisi_pcie_pmu_get_event_idx() 438 if (hisi_pcie_pmu_cmp_event(sibling, even in hisi_pcie_pmu_get_event_idx() [all...] |
/linux/arch/s390/kernel/ |
H A D | Makefile | 35 CFLAGS_stacktrace.o += -fno-optimize-sibling-calls 36 CFLAGS_dumpstack.o += -fno-optimize-sibling-calls 37 CFLAGS_unwind_bc.o += -fno-optimize-sibling-calls
|
/linux/net/netfilter/ |
H A D | nf_conntrack_pptp.c | 133 struct nf_conn *sibling; in destroy_sibling_or_exp() local 141 sibling = nf_ct_tuplehash_to_ctrack(h); in destroy_sibling_or_exp() 142 pr_debug("setting timeout of conntrack %p to 0\n", sibling); in destroy_sibling_or_exp() 143 sibling->proto.gre.timeout = 0; in destroy_sibling_or_exp() 144 sibling->proto.gre.stream_timeout = 0; in destroy_sibling_or_exp() 145 nf_ct_kill(sibling); in destroy_sibling_or_exp() 146 nf_ct_put(sibling); in destroy_sibling_or_exp()
|
/linux/arch/powerpc/platforms/85xx/ |
H A D | smp.c | 390 int sibling = cpu_last_thread_sibling(cpu); in mpc85xx_smp_kexec_cpu_down() local 410 } else if (sibling == crashing_cpu) { in mpc85xx_smp_kexec_cpu_down() 414 if (cpu_thread_in_core(cpu) == 0 && cpu_thread_in_core(sibling) != 0) { in mpc85xx_smp_kexec_cpu_down() 416 disable_cpu = sibling; in mpc85xx_smp_kexec_cpu_down()
|
/linux/drivers/acpi/ |
H A D | dock.c | 36 struct list_head sibling; member 120 list_for_each_entry(ds, &dock_stations, sibling) in find_dock_station() 178 list_for_each_entry(dock_station, &dock_stations, sibling) in is_dock_device() 599 INIT_LIST_HEAD(&dock_station->sibling); in acpi_dock_add() 622 list_add(&dock_station->sibling, &dock_stations); in acpi_dock_add()
|
/linux/drivers/perf/ |
H A D | marvell_pem_pmu.c | 193 struct perf_event *sibling; in pem_perf_event_init() local 214 for_each_sibling_event(sibling, event->group_leader) { in pem_perf_event_init() 215 if (sibling->pmu != event->pmu && in pem_perf_event_init() 216 !is_software_event(sibling)) in pem_perf_event_init()
|
/linux/net/ipv6/ |
H A D | ip6_fib.c | 547 * last sibling of this route (no need to dump the in fib6_dump_node() 548 * sibling routes again) in fib6_dump_node() 1149 * add is sibling to this route, increment our counter in fib6_add_rt2node() 1188 struct fib6_info *sibling, *temp_sibling; in fib6_add_rt2node() local 1191 sibling = leaf; in fib6_add_rt2node() 1193 while (sibling) { in fib6_add_rt2node() 1194 if (sibling->fib6_metric == rt->fib6_metric && in fib6_add_rt2node() 1195 rt6_qualify_for_ecmp(sibling)) { in fib6_add_rt2node() 1197 &sibling->fib6_siblings); in fib6_add_rt2node() 1200 sibling in fib6_add_rt2node() 1248 struct fib6_info *sibling, *next_sibling; fib6_add_rt2node() local 1998 struct fib6_info *sibling, *next_sibling; fib6_del_route() local [all...] |
/linux/arch/arm/mm/ |
H A D | cache-l2x0-pmu.c | 276 struct perf_event *sibling; in l2x0_pmu_group_is_valid() local 284 for_each_sibling_event(sibling, leader) { in l2x0_pmu_group_is_valid() 285 if (sibling->pmu == pmu) in l2x0_pmu_group_is_valid() 287 else if (!is_software_event(sibling)) in l2x0_pmu_group_is_valid()
|
/linux/arch/powerpc/kernel/ |
H A D | pci_dn.c | 430 else if (dn->sibling) in pci_traverse_device_nodes() 431 /* ok, try next sibling instead. */ in pci_traverse_device_nodes() 432 nextdn = dn->sibling; in pci_traverse_device_nodes() 434 /* Walk up to next valid sibling. */ in pci_traverse_device_nodes() 439 } while (dn->sibling == NULL); in pci_traverse_device_nodes() 440 nextdn = dn->sibling; in pci_traverse_device_nodes()
|