| /src/sys/dev/xen/control/ |
| H A D | control.c | 187 MPASS(mp_ncpus == 1 || smp_started); in xctrl_suspend() 192 if (smp_started) { in xctrl_suspend() 223 if (smp_started) { in xctrl_suspend() 291 if (smp_started) { in xctrl_suspend()
|
| /src/sys/arm/arm/ |
| H A D | mp_machdep.c | 197 atomic_store_rel_int(&smp_started, 1); in init_secondary() 203 while (smp_started == 0) { in init_secondary() 315 if (smp_started) in release_aps()
|
| /src/sys/kern/ |
| H A D | subr_smp.c | 108 volatile int smp_started; variable 233 if (!smp_started || cold || KERNEL_PANICKED()) in forward_signal() 283 if (!smp_started) in generic_stop_cpus() 398 if (!smp_started) in generic_restart_cpus() 442 if (!smp_started) in generic_restart_cpus() 594 if (!smp_started) { in smp_rendezvous_cpus() 936 KASSERT((!smp_started),("smp_no_rendezvous called and smp is started")); in smp_no_rendezvous_barrier() 955 if (!smp_started) { in smp_rendezvous_cpus_retry() 1124 active = smp_started; in sysctl_kern_smp_active()
|
| H A D | subr_atomic64.c | 67 if (smp_started) mtx_lock(_amtx) 69 #define UNLOCK_A64() if (smp_started) mtx_unlock(_amtx)
|
| H A D | kern_clocksource.c | 327 MPASS(mp_ncpus == 1 || smp_started); 334 if ((et->et_flags & ET_FLAGS_PERCPU) == 0 && smp_started) { 497 MPASS(mp_ncpus == 1 || smp_started); 503 if (!smp_started && cpu != CPU_FIRST()) 534 if ((timer->et_flags & ET_FLAGS_PERCPU) == 0 || !smp_started) {
|
| H A D | kern_cpu.c | 264 MPASS(mp_ncpus == 1 || smp_started); in cf_set_method() 273 if (mp_ncpus > 1 && !smp_started) { in cf_set_method()
|
| H A D | sched_ule.c | 1077 if (smp_started == 0 || steal_idle == 0 || tdq->tdq_cg == NULL) in tdq_idled() 1381 if (smp_started == 0) in sched_pickcpu() 2145 if (smp_started == 0 || steal_idle == 0 || trysteal_limit == 0 || in tdq_trysteal() 2695 if (balance_tdq == tdq && smp_started != 0 && rebalance != 0 && in sched_ule_clock()
|
| H A D | sched_4bsd.c | 1172 if (!smp_started || KERNEL_PANICKED()) in forward_wakeup() 1355 if (smp_started && (td->td_pinned != 0 || td->td_flags & TDF_BOUND || in sched_4bsd_add()
|
| /src/sys/riscv/riscv/ |
| H A D | mp_machdep.c | 141 if (atomic_load_acq_int(&smp_started)) in release_aps() 200 atomic_store_rel_int(&smp_started, 1); in init_secondary() 223 KASSERT(smp_started != 0, ("%s: SMP not started yet", __func__)); in smp_after_idle_runnable()
|
| H A D | pmap.c | 1083 if (!CPU_EMPTY(&mask) && smp_started) in pmap_invalidate_page() 1098 if (!CPU_EMPTY(&mask) && smp_started) in pmap_invalidate_range() 1124 if (!CPU_EMPTY(&mask) && smp_started) in pmap_invalidate_all() 5253 if (!CPU_EMPTY(&mask) && smp_started) { in pmap_sync_icache()
|
| /src/sys/powerpc/powerpc/ |
| H A D | mp_machdep.c | 104 while(smp_started == 0) in machdep_ap_bootstrap() 284 atomic_store_rel_int(&smp_started, 1); in cpu_mp_unleash()
|
| H A D | cpu.c | 838 if (smp_started == 0) in cpu_idle_powerx()
|
| /src/sys/x86/x86/ |
| H A D | mp_x86.c | 1124 atomic_store_rel_int(&smp_started, 1); in init_secondary_tail() 1138 while (atomic_load_acq_int(&smp_started) == 0) in init_secondary_tail() 1157 KASSERT(smp_started != 0, ("%s: SMP not started yet", __func__)); in smp_after_idle_runnable() 1749 while (smp_started == 0) in release_aps()
|
| H A D | intr_machdep.c | 424 MPASS(mp_ncpus == 1 || smp_started); in intr_assign_cpu() 633 MPASS(mp_ncpus == 1 || smp_started); in intr_next_cpu()
|
| H A D | local_apic.c | 1043 MPASS(mp_ncpus == 1 || smp_started); in lapic_enable_pcint() 1070 KASSERT(mp_ncpus == 1 || smp_started, ("hwpmc unloaded too early")); in lapic_disable_pcint()
|
| H A D | cpu_machdep.c | 596 if (smp_started) { in cpu_reset()
|
| /src/sys/sys/ |
| H A D | smp.h | 177 extern volatile int smp_started;
|
| /src/sys/arm64/arm64/ |
| H A D | mp_machdep.c | 216 atomic_store_rel_int(&smp_started, 1); in release_aps() 322 KASSERT(smp_started != 0, ("%s: SMP not started yet", __func__)); in smp_after_idle_runnable()
|
| /src/sys/dev/hyperv/vmbus/ |
| H A D | hyperv_mmu.c | 155 if (__predict_false(kdb_active || KERNEL_PANICKED() || !smp_started)) in hv_vm_tlb_flush()
|
| H A D | vmbus.c | 1538 device_printf(sc->vmbus_dev, "smp_started = %d\n", smp_started); in vmbus_doattach()
|
| /src/sys/i386/i386/ |
| H A D | mp_machdep.c | 559 if (kdb_active || KERNEL_PANICKED() || !smp_started) { in smp_targeted_tlb_shootdown()
|
| /src/sys/geom/eli/ |
| H A D | g_eli.c | 668 MPASS(!sc->sc_cpubind || smp_started); in g_eli_worker() 672 while (!smp_started) in g_eli_worker()
|
| /src/sys/powerpc/booke/ |
| H A D | pmap.c | 475 if (!smp_started) in tlb_miss_lock() 500 if (!smp_started) in tlb_miss_unlock() 2677 if ((e->mas2 & _TLB_ENTRY_SHARED) && smp_started) { in tlb1_write_entry()
|
| /src/sys/amd64/amd64/ |
| H A D | mp_machdep.c | 613 if (__predict_false(kdb_active || KERNEL_PANICKED() || !smp_started)) in smp_targeted_tlb_shootdown_native()
|
| /src/sys/dev/acpica/ |
| H A D | acpi.c | 3650 MPASS(mp_ncpus == 1 || smp_started); in acpi_EnterSleepState() 3655 if (smp_started) { in acpi_EnterSleepState() 3762 if (smp_started) { in acpi_EnterSleepState()
|