/linux/arch/powerpc/kernel/ |
H A D | misc_64.S | 84 ori r0,r7,MSR_DR 85 xori r0,r0,MSR_DR 115 ori r0,r7,MSR_DR 116 xori r0,r0,MSR_DR 146 ori r0,r7,MSR_DR 147 xori r0,r0,MSR_DR 161 ori r0,r7,MSR_DR 162 xori r0,r0,MSR_DR 363 li r10,MSR_DR|MSR_IR
|
H A D | rtas_entry.S | 27 li r9,MSR_KERNEL & ~(MSR_IR|MSR_DR) 130 LOAD_REG_IMMEDIATE(r6, MSR_KERNEL & ~(MSR_IR|MSR_DR))
|
H A D | head_book3s_32.S | 202 ori r0,r0,MSR_DR|MSR_IR|MSR_RI 986 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 1077 li r3, MSR_KERNEL & ~(MSR_IR | MSR_DR) 1100 li r3, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI) 1118 andi. r0,r3,MSR_DR|MSR_IR /* MMU enabled? */
|
H A D | paca.c | 199 new_paca->kernel_msr = MSR_KERNEL & ~(MSR_IR | MSR_DR); in initialise_paca()
|
H A D | head_8xx.S | 88 ori r0,r0,MSR_DR|MSR_IR 568 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 697 li r12, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI)
|
H A D | kvm_emul.S | 300 andi. r31, r31, MSR_DR | MSR_IR
|
H A D | head_64.S | 479 andi. r0,r3,MSR_IR|MSR_DR
|
/linux/arch/powerpc/kvm/ |
H A D | book3s_32_sr.S | 127 ori r11, r9, MSR_DR 143 li r6, MSR_DR
|
H A D | book3s_32_mmu.c | 356 if (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_32_esid_to_vsid() 362 /* In case we only have one of MSR_IR or MSR_DR set, let's put in kvmppc_mmu_book3s_32_esid_to_vsid() 365 switch (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_32_esid_to_vsid() 372 case MSR_DR: in kvmppc_mmu_book3s_32_esid_to_vsid() 375 case MSR_DR|MSR_IR: in kvmppc_mmu_book3s_32_esid_to_vsid()
|
H A D | book3s_64_mmu.c | 582 if (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_64_esid_to_vsid() 595 switch (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_64_esid_to_vsid() 602 case MSR_DR: in kvmppc_mmu_book3s_64_esid_to_vsid() 605 case MSR_DR|MSR_IR: in kvmppc_mmu_book3s_64_esid_to_vsid()
|
H A D | book3s_rmhandlers.S | 152 li r6, MSR_IR | MSR_DR
|
H A D | book3s.c | 456 int relocated = (kvmppc_get_msr(vcpu) & (data ? MSR_DR : MSR_IR)); in kvmppc_xlate() 470 if ((kvmppc_get_msr(vcpu) & (MSR_IR | MSR_DR)) == MSR_DR && in kvmppc_xlate()
|
H A D | book3s_segment.S | 299 ori r11, r9, MSR_DR /* Enable paging for data */
|
/linux/arch/powerpc/platforms/82xx/ |
H A D | pq2.c | 29 mtmsr(mfmsr() & ~(MSR_ME | MSR_EE | MSR_IR | MSR_DR)); in pq2_restart()
|
/linux/arch/powerpc/platforms/powernv/ |
H A D | opal-wrappers.S | 27 li r0,MSR_IR|MSR_DR|MSR_LE
|
H A D | subcore-asm.S | 31 li r5, MSR_IR|MSR_DR
|
H A D | rng.c | 88 if (mfmsr() & MSR_DR) { in pnv_get_random_long()
|
H A D | idle.c | 390 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power7_idle_insn() 702 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power9_idle_stop() 938 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power10_idle_stop()
|
H A D | opal-call.c | 101 bool mmu = (msr & (MSR_IR|MSR_DR)); in opal_call()
|
/linux/arch/powerpc/platforms/pasemi/ |
H A D | powersave.S | 62 LOAD_REG_IMMEDIATE(r6,MSR_DR|MSR_IR|MSR_ME|MSR_EE)
|
/linux/arch/powerpc/platforms/powermac/ |
H A D | cache.S | 44 rlwinm r0,r0,0,~MSR_DR 194 rlwinm r0,r0,0,~MSR_DR
|
/linux/arch/powerpc/mm/ |
H A D | cacheflush.c | 93 msr = msr0 & ~MSR_DR; in flush_dcache_icache_phys()
|
/linux/arch/powerpc/platforms/pseries/ |
H A D | ras.c | 496 (MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR| in pSeries_system_reset_exception() 497 MSR_ILE|MSR_HV|MSR_SF)) == (MSR_DR|MSR_SF)) { in pSeries_system_reset_exception()
|
/linux/arch/powerpc/include/asm/ |
H A D | reg.h | 106 #define MSR_DR __MASK(MSR_DR_LG) /* Data Relocate */ macro 134 #define __MSR (MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_HV) 147 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR)
|
/linux/arch/powerpc/mm/book3s64/ |
H A D | hash_pgtable.c | 467 [p] "b" (p), [MSR_IR_DR] "i" (MSR_IR | MSR_DR) in chmem_secondary_loop()
|