Home
last modified time | relevance | path

Searched refs:svm (Results 1 – 25 of 57) sorted by relevance

123

/linux/arch/x86/kvm/svm/
H A Dnested.c39 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_inject_npf_exit() local
40 struct vmcb *vmcb = svm->vmcb; in nested_svm_inject_npf_exit()
55 nested_svm_vmexit(svm); in nested_svm_inject_npf_exit()
60 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_get_tdp_pdptr() local
61 u64 cr3 = svm->nested.ctl.nested_cr3; in nested_svm_get_tdp_pdptr()
78 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_get_tdp_cr3() local
80 return svm->nested.ctl.nested_cr3; in nested_svm_get_tdp_cr3()
85 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_init_mmu_context() local
96 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context()
97 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
[all …]
H A Dsvm.c205 struct vcpu_svm *svm = to_svm(vcpu); in svm_set_efer() local
222 clr_exception_intercept(svm, GP_VECTOR); in svm_set_efer()
230 svm_free_nested(svm); in svm_set_efer()
233 int ret = svm_allocate_nested(svm); in svm_set_efer()
245 set_exception_intercept(svm, GP_VECTOR); in svm_set_efer()
249 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer()
250 vmcb_mark_dirty(svm->vmcb, VMCB_CR); in svm_set_efer()
256 struct vcpu_svm *svm = to_svm(vcpu); in svm_get_interrupt_shadow() local
259 if (svm->vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK) in svm_get_interrupt_shadow()
266 struct vcpu_svm *svm = to_svm(vcpu); in svm_set_interrupt_shadow() local
[all …]
H A Dsvm.h360 void recalc_intercepts(struct vcpu_svm *svm);
397 static inline bool ghcb_gpa_is_registered(struct vcpu_svm *svm, u64 val) in ghcb_gpa_is_registered() argument
399 return svm->sev_es.ghcb_registered_gpa == val; in ghcb_gpa_is_registered()
489 static inline void set_exception_intercept(struct vcpu_svm *svm, u32 bit) in set_exception_intercept() argument
491 struct vmcb *vmcb = svm->vmcb01.ptr; in set_exception_intercept()
496 recalc_intercepts(svm); in set_exception_intercept()
499 static inline void clr_exception_intercept(struct vcpu_svm *svm, u32 bit) in clr_exception_intercept() argument
501 struct vmcb *vmcb = svm->vmcb01.ptr; in clr_exception_intercept()
506 recalc_intercepts(svm); in clr_exception_intercept()
509 static inline void svm_set_intercept(struct vcpu_svm *svm, int bit) in svm_set_intercept() argument
[all …]
H A Dsev.c159 static bool sev_vcpu_has_debug_swap(struct vcpu_svm *svm) in sev_vcpu_has_debug_swap() argument
161 struct kvm_vcpu *vcpu = &svm->vcpu; in sev_vcpu_has_debug_swap()
875 static int sev_es_sync_vmsa(struct vcpu_svm *svm) in sev_es_sync_vmsa() argument
877 struct kvm_vcpu *vcpu = &svm->vcpu; in sev_es_sync_vmsa()
879 struct sev_es_save_area *save = svm->sev_es.vmsa; in sev_es_sync_vmsa()
886 if (svm->vcpu.guest_debug || (svm->vmcb->save.dr7 & ~DR7_FIXED_1)) in sev_es_sync_vmsa()
895 memcpy(save, &svm->vmcb->save, sizeof(svm->vmcb->save)); in sev_es_sync_vmsa()
898 save->rax = svm->vcpu.arch.regs[VCPU_REGS_RAX]; in sev_es_sync_vmsa()
899 save->rbx = svm->vcpu.arch.regs[VCPU_REGS_RBX]; in sev_es_sync_vmsa()
900 save->rcx = svm->vcpu.arch.regs[VCPU_REGS_RCX]; in sev_es_sync_vmsa()
[all …]
H A Davic.c111 static void avic_set_x2apic_msr_interception(struct vcpu_svm *svm, in avic_set_x2apic_msr_interception() argument
148 if (intercept == svm->x2avic_msrs_intercepted) in avic_set_x2apic_msr_interception()
155 svm_set_intercept_for_msr(&svm->vcpu, x2avic_passthrough_msrs[i], in avic_set_x2apic_msr_interception()
158 svm->x2avic_msrs_intercepted = intercept; in avic_set_x2apic_msr_interception()
186 static void avic_activate_vmcb(struct vcpu_svm *svm) in avic_activate_vmcb() argument
188 struct vmcb *vmcb = svm->vmcb01.ptr; in avic_activate_vmcb()
189 struct kvm_vcpu *vcpu = &svm->vcpu; in avic_activate_vmcb()
196 svm_clr_intercept(svm, INTERCEPT_CR8_WRITE); in avic_activate_vmcb()
205 if (x2avic_enabled && apic_x2apic_mode(svm->vcpu.arch.apic)) { in avic_activate_vmcb()
209 avic_set_x2apic_msr_interception(svm, false); in avic_activate_vmcb()
[all …]
H A Dhyperv.c11 struct vcpu_svm *svm = to_svm(vcpu); in svm_hv_inject_synthetic_vmexit_post_tlb_flush() local
19 svm->vmcb->control.exit_code = HV_SVM_EXITCODE_ENL; in svm_hv_inject_synthetic_vmexit_post_tlb_flush()
20 svm->vmcb->control.exit_info_1 = HV_SVM_ENL_EXITCODE_TRAP_AFTER_FLUSH; in svm_hv_inject_synthetic_vmexit_post_tlb_flush()
21 svm->vmcb->control.exit_info_2 = 0; in svm_hv_inject_synthetic_vmexit_post_tlb_flush()
22 nested_svm_vmexit(svm); in svm_hv_inject_synthetic_vmexit_post_tlb_flush()
H A Dhyperv.h17 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_hv_update_vm_vp_ids() local
18 struct hv_vmcb_enlightenments *hve = &svm->nested.ctl.hv_enlightenments; in nested_svm_hv_update_vm_vp_ids()
31 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_l2_tlb_flush_enabled() local
32 struct hv_vmcb_enlightenments *hve = &svm->nested.ctl.hv_enlightenments; in nested_svm_l2_tlb_flush_enabled()
/linux/tools/testing/selftests/kvm/x86/
H A Dnested_vmsave_vmload_test.c71 static void l1_guest_code(struct svm_test_data *svm) in l1_guest_code() argument
76 generic_svm_setup(svm, NULL, &l2_guest_stack[L2_GUEST_STACK_SIZE]); in l1_guest_code()
79 svm->vmcb->control.intercept |= (BIT_ULL(INTERCEPT_VMSAVE) | in l1_guest_code()
83 svm->vmcb->control.virt_ext &= ~VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK; in l1_guest_code()
85 svm->vmcb->save.rip = (u64)l2_guest_code_vmsave; in l1_guest_code()
86 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_guest_code()
87 GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMSAVE); in l1_guest_code()
89 svm->vmcb->save.rip = (u64)l2_guest_code_vmload; in l1_guest_code()
90 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_guest_code()
91 GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMLOAD); in l1_guest_code()
[all …]
H A Dnested_invalid_cr3_test.c21 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
26 generic_svm_setup(svm, l2_guest_code, in l1_svm_code()
30 save_cr3 = svm->vmcb->save.cr3; in l1_svm_code()
31 svm->vmcb->save.cr3 = -1ull; in l1_svm_code()
32 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
33 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_ERR); in l1_svm_code()
36 svm->vmcb->save.cr3 = save_cr3; in l1_svm_code()
37 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
38 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL); in l1_svm_code()
H A Dhyperv_svm_test.c68 static void __attribute__((__flatten__)) guest_code(struct svm_test_data *svm, in guest_code() argument
73 struct vmcb *vmcb = svm->vmcb; in guest_code()
82 GUEST_ASSERT(svm->vmcb_gpa); in guest_code()
84 generic_svm_setup(svm, l2_guest_code, in guest_code()
96 run_guest(vmcb, svm->vmcb_gpa); in guest_code()
103 __set_bit(2 * (MSR_FS_BASE & 0x1fff), svm->msr + 0x800); in guest_code()
104 run_guest(vmcb, svm->vmcb_gpa); in guest_code()
110 run_guest(vmcb, svm->vmcb_gpa); in guest_code()
115 __set_bit(2 * (MSR_GS_BASE & 0x1fff), svm->msr + 0x800); in guest_code()
118 run_guest(vmcb, svm->vmcb_gpa); in guest_code()
[all …]
H A Dnested_exceptions_test.c74 static void svm_run_l2(struct svm_test_data *svm, void *l2_code, int vector, in svm_run_l2() argument
77 struct vmcb *vmcb = svm->vmcb; in svm_run_l2()
81 run_guest(vmcb, svm->vmcb_gpa); in svm_run_l2()
91 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
93 struct vmcb_control_area *ctrl = &svm->vmcb->control; in l1_svm_code()
96 generic_svm_setup(svm, NULL, &l2_guest_stack[L2_GUEST_STACK_SIZE]); in l1_svm_code()
97 svm->vmcb->save.idtr.limit = 0; in l1_svm_code()
101 svm_run_l2(svm, l2_ss_pending_test, SS_VECTOR, SS_ERROR_CODE); in l1_svm_code()
102 svm_run_l2(svm, l2_ss_injected_gp_test, GP_VECTOR, GP_ERROR_CODE_AMD); in l1_svm_code()
105 svm_run_l2(svm, l2_ss_injected_df_test, DF_VECTOR, DF_ERROR_CODE); in l1_svm_code()
[all …]
H A Dsvm_nested_shutdown_test.c15 static void l2_guest_code(struct svm_test_data *svm) in l2_guest_code() argument
20 static void l1_guest_code(struct svm_test_data *svm, struct idt_entry *idt) in l1_guest_code() argument
24 struct vmcb *vmcb = svm->vmcb; in l1_guest_code()
26 generic_svm_setup(svm, l2_guest_code, in l1_guest_code()
36 run_guest(vmcb, svm->vmcb_gpa); in l1_guest_code()
H A Dsvm_vmcall_test.c15 static void l2_guest_code(struct svm_test_data *svm) in l2_guest_code() argument
20 static void l1_guest_code(struct svm_test_data *svm) in l1_guest_code() argument
24 struct vmcb *vmcb = svm->vmcb; in l1_guest_code()
27 generic_svm_setup(svm, l2_guest_code, in l1_guest_code()
30 run_guest(vmcb, svm->vmcb_gpa); in l1_guest_code()
H A Dnested_tsc_adjust_test.c106 struct svm_test_data *svm = data; in l1_guest_code() local
108 generic_svm_setup(svm, l2_guest_code, in l1_guest_code()
111 svm->vmcb->control.tsc_offset = TSC_OFFSET_VALUE; in l1_guest_code()
112 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_guest_code()
113 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL); in l1_guest_code()
H A Dsmm_test.c69 struct svm_test_data *svm = arg; in guest_code() local
84 generic_svm_setup(svm, l2_guest_code, in guest_code()
100 run_guest(svm->vmcb, svm->vmcb_gpa); in guest_code()
101 run_guest(svm->vmcb, svm->vmcb_gpa); in guest_code()
H A Dsvm_int_ctl_test.c33 static void l2_guest_code(struct svm_test_data *svm) in l2_guest_code() argument
55 static void l1_guest_code(struct svm_test_data *svm) in l1_guest_code() argument
59 struct vmcb *vmcb = svm->vmcb; in l1_guest_code()
64 generic_svm_setup(svm, l2_guest_code, in l1_guest_code()
77 run_guest(vmcb, svm->vmcb_gpa); in l1_guest_code()
H A Dnested_emulation_test.c57 struct svm_test_data *svm = test_data; in guest_code() local
58 struct vmcb *vmcb = svm->vmcb; in guest_code()
60 generic_svm_setup(svm, NULL, NULL); in guest_code()
97 struct svm_test_data *svm = test_data; in guest_code() local
98 struct vmcb *vmcb = svm->vmcb; in guest_code()
100 run_guest(vmcb, svm->vmcb_gpa); in guest_code()
H A Dnested_close_kvm_test.c48 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
53 generic_svm_setup(svm, l2_guest_code, in l1_svm_code()
56 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
/linux/tools/testing/selftests/kvm/lib/x86/
H A Dsvm.c34 struct svm_test_data *svm = addr_gva2hva(vm, svm_gva); in vcpu_alloc_svm() local
36 svm->vmcb = (void *)vm_vaddr_alloc_page(vm); in vcpu_alloc_svm()
37 svm->vmcb_hva = addr_gva2hva(vm, (uintptr_t)svm->vmcb); in vcpu_alloc_svm()
38 svm->vmcb_gpa = addr_gva2gpa(vm, (uintptr_t)svm->vmcb); in vcpu_alloc_svm()
40 svm->save_area = (void *)vm_vaddr_alloc_page(vm); in vcpu_alloc_svm()
41 svm->save_area_hva = addr_gva2hva(vm, (uintptr_t)svm->save_area); in vcpu_alloc_svm()
42 svm->save_area_gpa = addr_gva2gpa(vm, (uintptr_t)svm->save_area); in vcpu_alloc_svm()
44 svm->msr = (void *)vm_vaddr_alloc_page(vm); in vcpu_alloc_svm()
45 svm->msr_hva = addr_gva2hva(vm, (uintptr_t)svm->msr); in vcpu_alloc_svm()
46 svm->msr_gpa = addr_gva2gpa(vm, (uintptr_t)svm->msr); in vcpu_alloc_svm()
[all …]
H A Dmemstress.c54 static void l1_svm_code(struct svm_test_data *svm, uint64_t vcpu_id) in l1_svm_code() argument
62 generic_svm_setup(svm, memstress_l2_guest_entry, rsp); in l1_svm_code()
64 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
65 GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMMCALL); in l1_svm_code()
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_svm.c94 nouveau_ivmm_find(struct nouveau_svm *svm, u64 inst) in nouveau_ivmm_find() argument
97 list_for_each_entry(ivmm, &svm->inst, head) { in nouveau_ivmm_find()
169 if (!cli->svm.svmm) { in nouveau_svmm_bind()
186 nouveau_dmem_migrate_vma(cli->drm, cli->svm.svmm, vma, addr, in nouveau_svmm_bind()
210 mutex_lock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_part()
211 ivmm = nouveau_ivmm_find(svmm->vmm->cli->drm->svm, inst); in nouveau_svmm_part()
216 mutex_unlock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_part()
231 mutex_lock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_join()
232 list_add(&ivmm->head, &svmm->vmm->cli->drm->svm->inst); in nouveau_svmm_join()
233 mutex_unlock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_join()
[all …]
/linux/arch/arm/mm/
H A Dioremap.c53 struct static_vm *svm; in find_static_vm_paddr() local
56 list_for_each_entry(svm, &static_vmlist, list) { in find_static_vm_paddr()
57 vm = &svm->vm; in find_static_vm_paddr()
67 return svm; in find_static_vm_paddr()
75 struct static_vm *svm; in find_static_vm_vaddr() local
78 list_for_each_entry(svm, &static_vmlist, list) { in find_static_vm_vaddr()
79 vm = &svm->vm; in find_static_vm_vaddr()
86 return svm; in find_static_vm_vaddr()
92 void __init add_static_vm_early(struct static_vm *svm) in add_static_vm_early() argument
98 vm = &svm->vm; in add_static_vm_early()
[all …]
/linux/arch/x86/kvm/
H A DMakefile24 kvm-amd-y += svm/svm.o svm/vmenter.o svm/pmu.o svm/nested.o svm/avic.o
26 kvm-amd-$(CONFIG_KVM_AMD_SEV) += svm/sev.o
27 kvm-amd-$(CONFIG_KVM_HYPERV) += svm/hyperv.o
32 kvm-amd-y += svm/svm_onhyperv.o
40 $(obj)/svm/vmenter.o: $(obj)/kvm-asm-offsets.h
/linux/drivers/gpu/drm/xe/
H A Dxe_svm.c56 return force_smem ? NULL : vm->svm.peer.owner; in xe_svm_private_page_owner()
82 return container_of(gpusvm, struct xe_vm, svm.gpusvm); in gpusvm_to_vm()
138 spin_lock(&vm->svm.garbage_collector.lock); in xe_svm_garbage_collector_add_range()
141 &vm->svm.garbage_collector.range_list); in xe_svm_garbage_collector_add_range()
142 spin_unlock(&vm->svm.garbage_collector.lock); in xe_svm_garbage_collector_add_range()
144 queue_work(xe->usm.pf_wq, &vm->svm.garbage_collector.work); in xe_svm_garbage_collector_add_range()
210 drm_gpusvm_range_unmap_pages(&vm->svm.gpusvm, r, &ctx); in xe_svm_range_notifier_event_end()
309 drm_gpusvm_range_remove(&vm->svm.gpusvm, &range->base); in __xe_svm_garbage_collector()
387 spin_lock(&vm->svm.garbage_collector.lock); in xe_svm_garbage_collector()
388 range = list_first_entry_or_null(&vm->svm.garbage_collector.range_list, in xe_svm_garbage_collector()
[all …]
/linux/tools/testing/vsock/
H A Dvsock_perf.c104 struct sockaddr_vm svm; in vsock_connect() member
106 .svm = { in vsock_connect()
121 if (connect(fd, &addr.sa, sizeof(addr.svm)) < 0) { in vsock_connect()
147 struct sockaddr_vm svm; in run_receiver() member
149 .svm = { in run_receiver()
157 struct sockaddr_vm svm; in run_receiver() member
160 socklen_t clientaddr_len = sizeof(clientaddr.svm); in run_receiver()
173 if (bind(fd, &addr.sa, sizeof(addr.svm)) < 0) in run_receiver()

123