| /linux/arch/x86/lib/ |
| H A D | msr-smp.c | 10 struct msr_info *rv = info; in __rdmsr_on_cpu() 23 struct msr_info *rv = info; in __wrmsr_on_cpu() 37 struct msr_info rv; in rdmsr_on_cpu() 53 struct msr_info rv; in rdmsrq_on_cpu() 68 struct msr_info rv; in wrmsr_on_cpu() 84 struct msr_info rv; in wrmsrq_on_cpu() 101 struct msr_info rv; in __rwmsr_on_cpus() 146 struct msr_info msr; 162 struct msr_info *rv = info; in __wrmsr_safe_on_cpu() 194 struct msr_info rv; in wrmsr_safe_on_cpu() [all …]
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | pmu_intel.c | 292 struct msr_data *msr_info, bool read) in intel_pmu_handle_lbr_msrs_access() argument 295 u32 index = msr_info->index; in intel_pmu_handle_lbr_msrs_access() 312 rdmsrq(index, msr_info->data); in intel_pmu_handle_lbr_msrs_access() 314 wrmsrq(index, msr_info->data); in intel_pmu_handle_lbr_msrs_access() 324 msr_info->data = 0; in intel_pmu_handle_lbr_msrs_access() 328 static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in intel_pmu_get_msr() argument 332 u32 msr = msr_info->index; in intel_pmu_get_msr() 336 msr_info->data = pmu->fixed_ctr_ctrl; in intel_pmu_get_msr() 339 msr_info->data = pmu->pebs_enable; in intel_pmu_get_msr() 342 msr_info->data = pmu->ds_area; in intel_pmu_get_msr() [all …]
|
| H A D | vmx.c | 2124 int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in vmx_get_msr() argument 2130 switch (msr_info->index) { in vmx_get_msr() 2133 msr_info->data = vmcs_readl(GUEST_FS_BASE); in vmx_get_msr() 2136 msr_info->data = vmcs_readl(GUEST_GS_BASE); in vmx_get_msr() 2139 msr_info->data = vmx_read_guest_kernel_gs_base(vmx); in vmx_get_msr() 2143 return kvm_get_msr_common(vcpu, msr_info); in vmx_get_msr() 2145 if (!msr_info->host_initiated && in vmx_get_msr() 2150 if (!msr_info->host_initiated && !vmx_has_waitpkg(vmx)) in vmx_get_msr() 2153 msr_info->data = vmx->msr_ia32_umwait_control; in vmx_get_msr() 2156 if (!msr_info->host_initiated && in vmx_get_msr() [all …]
|
| H A D | main.c | 163 static int vt_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in vt_set_msr() argument 166 return tdx_set_msr(vcpu, msr_info); in vt_set_msr() 168 return vmx_set_msr(vcpu, msr_info); in vt_set_msr() 183 static int vt_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in vt_get_msr() argument 186 return tdx_get_msr(vcpu, msr_info); in vt_get_msr() 188 return vmx_get_msr(vcpu, msr_info); in vt_get_msr()
|
| H A D | x86_ops.h | 33 int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info); 59 int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
|
| /linux/arch/x86/kvm/svm/ |
| H A D | pmu.c | 128 static int amd_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in amd_pmu_get_msr() argument 132 u32 msr = msr_info->index; in amd_pmu_get_msr() 137 msr_info->data = pmc_read_counter(pmc); in amd_pmu_get_msr() 143 msr_info->data = pmc->eventsel; in amd_pmu_get_msr() 150 static int amd_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in amd_pmu_set_msr() argument 154 u32 msr = msr_info->index; in amd_pmu_set_msr() 155 u64 data = msr_info->data; in amd_pmu_set_msr()
|
| H A D | svm.c | 2693 struct msr_data msr_info; in efer_trap() local 2702 msr_info.host_initiated = false; in efer_trap() 2703 msr_info.index = MSR_EFER; in efer_trap() 2704 msr_info.data = to_svm(vcpu)->vmcb->control.exit_info_1 & ~EFER_SVME; in efer_trap() 2705 ret = kvm_set_msr_common(vcpu, &msr_info); in efer_trap() 2727 struct msr_data *msr_info) in sev_es_prevent_msr_access() argument 2730 msr_info->index != MSR_IA32_XSS && in sev_es_prevent_msr_access() 2731 !msr_write_intercepted(vcpu, msr_info->index); in sev_es_prevent_msr_access() 2734 static int svm_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in svm_get_msr() argument 2738 if (sev_es_prevent_msr_access(vcpu, msr_info)) { in svm_get_msr() [all …]
|
| /linux/arch/x86/kvm/ |
| H A D | pmu.c | 820 int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in kvm_pmu_get_msr() argument 823 u32 msr = msr_info->index; in kvm_pmu_get_msr() 828 msr_info->data = pmu->global_status; in kvm_pmu_get_msr() 832 msr_info->data = pmu->global_ctrl; in kvm_pmu_get_msr() 837 msr_info->data = 0; in kvm_pmu_get_msr() 840 return kvm_pmu_call(get_msr)(vcpu, msr_info); in kvm_pmu_get_msr() 846 int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in kvm_pmu_set_msr() argument 849 u32 msr = msr_info->index; in kvm_pmu_set_msr() 850 u64 data = msr_info->data; in kvm_pmu_set_msr() 859 if (!msr_info->host_initiated) in kvm_pmu_set_msr() [all …]
|
| H A D | pmu.h | 32 int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info); 33 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info); 252 int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info); 253 int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
|
| H A D | x86.c | 1756 static int set_efer(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in set_efer() argument 1759 u64 efer = msr_info->data; in set_efer() 1765 if (!msr_info->host_initiated) { in set_efer() 3557 static int set_msr_mce(struct kvm_vcpu *vcpu, struct msr_data *msr_info) in set_msr_mce() argument 3561 u32 msr = msr_info->index; in set_msr_mce() 3562 u64 data = msr_info->data; in set_msr_mce() 3571 (data || !msr_info->host_initiated)) in set_msr_mce() 3582 if (!(mcg_cap & MCG_CMCI_P) && (data || !msr_info->host_initiated)) in set_msr_mce() 3615 if (!msr_info->host_initiated && is_mci_status_msr(msr) && in set_msr_mce() 3878 struct msr_data *msr_info, in kvm_access_xstate_msr() argument [all …]
|
| /linux/arch/x86/include/asm/ |
| H A D | msr.h | 18 struct msr_info { struct 32 struct msr_info info; argument
|