Searched refs:msr_autoload (Results 1 – 3 of 3) sorted by relevance
| /linux/arch/x86/kvm/vmx/ |
| H A D | vmx.h | 235 struct msr_autoload { struct 238 } msr_autoload; member
|
| H A D | vmx.c | 1059 struct msr_autoload *m = &vmx->msr_autoload; in clear_atomic_switch_msr() 1116 struct msr_autoload *m = &vmx->msr_autoload; in add_atomic_switch_msr() 5005 vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val)); in init_vmcs() 5007 vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val)); in init_vmcs() 6638 efer_slot = vmx_find_loadstore_msr_slot(&vmx->msr_autoload.guest, MSR_EFER); in dump_vmcs() 6643 vmx->msr_autoload.guest.val[efer_slot].value); in dump_vmcs() 6668 vmx_dump_msrs("guest autoload", &vmx->msr_autoload.guest); in dump_vmcs() 6705 vmx_dump_msrs("host autoload", &vmx->msr_autoload.host); in dump_vmcs()
|
| H A D | nested.c | 2334 vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val)); in prepare_vmcs02_constant_state() 2335 vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val)); in prepare_vmcs02_constant_state() 2684 vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, vmx->msr_autoload.host.nr); in prepare_vmcs02_rare() 2685 vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, vmx->msr_autoload.guest.nr); in prepare_vmcs02_rare() 4923 for (i = 0; i < vmx->msr_autoload.guest.nr; ++i) { in nested_vmx_get_vmcs01_guest_efer() 4924 if (vmx->msr_autoload.guest.val[i].index == MSR_EFER) in nested_vmx_get_vmcs01_guest_efer() 4925 return vmx->msr_autoload.guest.val[i].value; in nested_vmx_get_vmcs01_guest_efer() 5146 vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, vmx->msr_autoload.host.nr); in __nested_vmx_vmexit() 5147 vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, vmx->msr_autoload.guest.nr); in __nested_vmx_vmexit()
|