| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | svm.c | 96 efer = rdmsr(MSR_EFER); in generic_svm_setup() 97 wrmsr(MSR_EFER, efer | EFER_SVME); in generic_svm_setup() 111 save->efer = rdmsr(MSR_EFER); in generic_svm_setup()
|
| H A D | vmx.c | 275 vmwrite(HOST_IA32_EFER, rdmsr(MSR_EFER)); in init_vmcs_host_state()
|
| /linux/arch/x86/kernel/acpi/ |
| H A D | sleep.c | 85 if (!rdmsr_safe(MSR_EFER, in x86_acpi_suspend_lowlevel() 88 !wrmsr_safe(MSR_EFER, in x86_acpi_suspend_lowlevel()
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | nested_set_state_test.c | 253 uint64_t old_efer = vcpu_get_msr(vcpu, MSR_EFER); in vcpu_efer_enable_svm() 255 vcpu_set_msr(vcpu, MSR_EFER, old_efer | EFER_SVME); in vcpu_efer_enable_svm() 260 uint64_t old_efer = vcpu_get_msr(vcpu, MSR_EFER); in vcpu_efer_disable_svm() 262 vcpu_set_msr(vcpu, MSR_EFER, old_efer & ~EFER_SVME); in vcpu_efer_disable_svm()
|
| H A D | sev_smoke_test.c | 34 guest_sev_test_msr(MSR_EFER); in guest_sev_test_regs()
|
| /linux/arch/x86/boot/startup/ |
| H A D | efi-mixed.S | 74 movl $MSR_EFER, %ecx 114 movl $MSR_EFER, %ecx
|
| H A D | la57toggle.S | 74 movl $MSR_EFER, %ecx
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_trampoline.S | 57 movl $MSR_EFER, %ecx
|
| H A D | hv_crash.c | 176 hv_wrmsr(MSR_EFER, hv_crash_ctxt.efer); in hv_crash_c_entry() 221 ctxt->efer = __rdmsr(MSR_EFER); in hv_hvcrash_ctxt_save()
|
| H A D | hv_vtl.c | 165 input->vp_context.efer = native_rdmsrq(MSR_EFER); in hv_vtl_bringup_vcpu()
|
| H A D | ivm.c | 331 vmsa->efer = native_read_msr(MSR_EFER); in hv_snp_boot_ap()
|
| /linux/arch/x86/power/ |
| H A D | cpu.c | 119 rdmsrq(MSR_EFER, ctxt->efer); in __save_processor_state() 212 wrmsrq(MSR_EFER, ctxt->efer); in __restore_processor_state()
|
| /linux/arch/x86/realmode/rm/ |
| H A D | reboot.S | 40 movl $MSR_EFER, %ecx
|
| H A D | wakeup_asm.S | 123 movl $MSR_EFER, %ecx
|
| H A D | trampoline_64.S | 164 movl $MSR_EFER, %ecx
|
| /linux/arch/x86/realmode/ |
| H A D | init.c | 151 rdmsrq(MSR_EFER, efer); in setup_real_mode()
|
| /linux/arch/x86/platform/pvh/ |
| H A D | head.S | 103 mov $MSR_EFER, %ecx
|
| /linux/arch/x86/kernel/ |
| H A D | head_32.S | 216 movl $MSR_EFER, %ecx
|
| H A D | head_64.S | 384 movl $MSR_EFER, %ecx
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | amd.c | 1154 WARN_ON_ONCE(msr_set_bit(MSR_EFER, _EFER_AUTOIBRS) < 0); in init_amd() 1161 msr_set_bit(MSR_EFER, _EFER_TCE); in init_amd()
|
| /linux/arch/x86/kvm/ |
| H A D | smm.c | 536 if (__kvm_emulate_msr_write(vcpu, MSR_EFER, smstate->efer & ~EFER_LMA)) in rsm_load_state_64() 631 __kvm_emulate_msr_write(vcpu, MSR_EFER, efer); in emulator_leave_smm()
|
| /linux/arch/x86/boot/compressed/ |
| H A D | head_64.S | 238 movl $MSR_EFER, %ecx
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | vmx.c | 1062 case MSR_EFER: in clear_atomic_switch_msr() 1120 case MSR_EFER: in add_atomic_switch_msr() 1186 add_atomic_switch_msr(vmx, MSR_EFER, guest_efer, kvm_host.efer); in update_transition_efer() 1188 clear_atomic_switch_msr(vmx, MSR_EFER); in update_transition_efer() 1192 i = kvm_find_user_return_msr(MSR_EFER); in update_transition_efer() 1196 clear_atomic_switch_msr(vmx, MSR_EFER); in update_transition_efer() 2017 vmx_setup_uret_msr(vmx, MSR_EFER, update_transition_efer(vmx)); in vmx_setup_uret_msrs() 2142 case MSR_EFER: in vmx_get_msr() 2340 case MSR_EFER: in vmx_set_msr() 3314 if (!vmx_find_uret_msr(vmx, MSR_EFER)) in vmx_set_efer() [all …]
|
| /linux/tools/arch/x86/include/asm/ |
| H A D | msr-index.h | 10 #define MSR_EFER 0xc0000080 /* extended feature register */ macro
|
| /linux/arch/x86/include/asm/ |
| H A D | msr-index.h | 10 #define MSR_EFER 0xc0000080 /* extended feature register */ macro
|