| /linux/arch/arm64/kvm/ |
| H A D | pvtime.c | 22 if (base == INVALID_GPA) in kvm_update_stolen_time() 43 if (vcpu->arch.steal.base != INVALID_GPA) in kvm_hypercall_pv_features() 57 if (base == INVALID_GPA) in kvm_init_stolen_time() 92 if (vcpu->arch.steal.base != INVALID_GPA) in kvm_arm_pvtime_set_attr()
|
| H A D | inject_fault.c | 121 if (hpfar == INVALID_GPA) in inject_abt64()
|
| H A D | hypercalls.c | 355 if (gpa != INVALID_GPA) in kvm_smccc_call_handler()
|
| H A D | mmu.c | 2030 if (ipa != INVALID_GPA) { in kvm_handle_guest_sea() 2071 if (KVM_BUG_ON(ipa == INVALID_GPA, vcpu->kvm)) in kvm_handle_guest_abort()
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_sbi_sta.c | 21 vcpu->arch.sta.shmem = INVALID_GPA; in kvm_riscv_vcpu_sbi_sta_reset() 38 if (shmem == INVALID_GPA) in kvm_riscv_vcpu_record_steal_time() 50 vcpu->arch.sta.shmem = INVALID_GPA; in kvm_riscv_vcpu_record_steal_time() 96 vcpu->arch.sta.shmem = INVALID_GPA; in kvm_sbi_sta_steal_time_set_shmem()
|
| H A D | vcpu_pmu.c | 405 kvpmu->snapshot_addr = INVALID_GPA; in kvm_pmu_clear_snapshot_area() 554 if (kvpmu->snapshot_addr == INVALID_GPA) { in kvm_riscv_vcpu_pmu_ctr_start() 632 if (snap_flag_set && kvpmu->snapshot_addr == INVALID_GPA) { in kvm_riscv_vcpu_pmu_ctr_stop() 843 kvpmu->snapshot_addr = INVALID_GPA; in kvm_riscv_vcpu_pmu_init()
|
| /linux/virt/kvm/ |
| H A D | pfncache.c | 286 gpc->gpa = INVALID_GPA; in __kvm_gpc_refresh() 392 gpc->gpa = INVALID_GPA; in kvm_gpc_init() 444 return __kvm_gpc_activate(gpc, INVALID_GPA, uhva, len); in kvm_gpc_activate_hva()
|
| H A D | kvm_main.c | 689 kvm->mmu_invalidate_range_start = INVALID_GPA; in kvm_mmu_invalidate_begin() 690 kvm->mmu_invalidate_range_end = INVALID_GPA; in kvm_mmu_invalidate_begin() 700 if (likely(kvm->mmu_invalidate_range_start == INVALID_GPA)) { in kvm_mmu_invalidate_range_add() 799 WARN_ON_ONCE(kvm->mmu_invalidate_range_start == INVALID_GPA); in kvm_mmu_invalidate_end()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | nested.c | 202 if (vmx->nested.current_vmptr == INVALID_GPA && in nested_vmx_fail() 229 vmcs_write64(VMCS_LINK_POINTER, INVALID_GPA); in vmx_disable_shadow_vmcs() 244 hv_vcpu->nested.pa_page_gpa = INVALID_GPA; in nested_release_evmcs() 350 vmx->nested.vmxon_ptr = INVALID_GPA; in free_nested() 353 vmx->nested.current_vmptr = INVALID_GPA; in free_nested() 796 vmcs12->vmcs_link_pointer == INVALID_GPA) in nested_cache_shadow_vmcs12() 815 vmcs12->vmcs_link_pointer == INVALID_GPA) in nested_flush_cached_shadow_vmcs12() 2137 vmx->nested.current_vmptr = INVALID_GPA; in nested_vmx_handle_enlightened_vmptrld() 2323 vmcs_write64(ENCLS_EXITING_BITMAP, INVALID_GPA); in prepare_vmcs02_constant_state() 2345 vmcs_write64(VMCS_LINK_POINTER, INVALID_GPA); in prepare_vmcs02_early_rare() [all …]
|
| H A D | sgx.c | 84 if (*gpa == INVALID_GPA) { in sgx_gva_to_gpa()
|
| H A D | vmx.c | 4946 vmcs_write64(VMCS_LINK_POINTER, INVALID_GPA); /* 22.3.1.5 */ in init_vmcs() 5072 vmx->nested.vmxon_ptr = INVALID_GPA; in __vmx_vcpu_reset() 5073 vmx->nested.current_vmptr = INVALID_GPA; in __vmx_vcpu_reset() 6879 kvm_prepare_event_vectoring_exit(vcpu, INVALID_GPA); in __vmx_handle_exit()
|
| /linux/include/linux/ |
| H A D | kvm_types.h | 68 #define INVALID_GPA (~(gpa_t)0) macro
|
| H A D | kvm_host.h | 155 return gpa == INVALID_GPA; in kvm_is_error_gpa() 2143 if (WARN_ON_ONCE(kvm->mmu_invalidate_range_start == INVALID_GPA || in mmu_invalidate_retry_gfn() 2144 kvm->mmu_invalidate_range_end == INVALID_GPA)) in mmu_invalidate_retry_gfn()
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | paging_tmpl.h | 392 if (unlikely(real_gpa == INVALID_GPA)) in walk_addr_generic() 449 if (real_gpa == INVALID_GPA) in walk_addr_generic() 865 gpa_t gpa = INVALID_GPA; in gva_to_gpa()
|
| H A D | mmu.c | 2890 if (gpa == INVALID_GPA) in __kvm_mmu_unprotect_gfn_and_retry() 4528 arch.cr3 = (unsigned long)INVALID_GPA; in kvm_arch_setup_async_pf()
|
| /linux/arch/loongarch/kvm/ |
| H A D | vcpu.c | 178 ghc->gpa = INVALID_GPA; in kvm_update_stolen_time() 236 if (vcpu->arch.flush_gpa != INVALID_GPA) { in kvm_late_check_requests() 238 vcpu->arch.flush_gpa = INVALID_GPA; in kvm_late_check_requests() 1537 vcpu->arch.flush_gpa = INVALID_GPA; in kvm_arch_vcpu_create() 1830 ghc->gpa = INVALID_GPA; in kvm_vcpu_set_pv_preempted()
|
| /linux/arch/arm64/include/asm/ |
| H A D | kvm_emulate.h | 353 return INVALID_GPA; in kvm_vcpu_get_fault_ipa()
|
| H A D | kvm_host.h | 1339 vcpu_arch->steal.base = INVALID_GPA; in kvm_arm_pvtime_vcpu_init() 1344 return (vcpu_arch->steal.base != INVALID_GPA); in kvm_arm_is_pvtime_enabled()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | nested.c | 1209 svm->nested.ctl.bus_lock_rip = INVALID_GPA; in nested_svm_vmexit() 1400 svm->nested.last_vmcb12_gpa = INVALID_GPA; in svm_free_nested() 1411 svm->nested.vmcb12_gpa = INVALID_GPA; in svm_leave_nested()
|
| H A D | svm.c | 1176 svm->nested.vmcb12_gpa = INVALID_GPA; in init_vmcb() 1177 svm->nested.last_vmcb12_gpa = INVALID_GPA; in init_vmcb()
|
| /linux/arch/x86/kvm/ |
| H A D | x86.c | 1074 if (real_gpa == INVALID_GPA) in load_pdptrs() 7884 if (gpa == INVALID_GPA) in kvm_read_guest_virt_helper() 7915 if (unlikely(gpa == INVALID_GPA)) in kvm_fetch_guest_virt() 7976 if (gpa == INVALID_GPA) in kvm_write_guest_virt_helper() 8092 if (*gpa == INVALID_GPA) in vcpu_mmio_gva_to_gpa() 8329 if (gpa == INVALID_GPA || in emulator_cmpxchg_emulated() 12596 tr->valid = gpa != INVALID_GPA; in kvm_arch_vcpu_ioctl_translate() 14118 mmu->gva_to_gpa(vcpu, mmu, gva, access, &fault) != INVALID_GPA) { in kvm_fixup_and_inject_pf_error()
|
| H A D | hyperv.c | 2045 if (unlikely(hc->ingpa == INVALID_GPA)) in kvm_hv_flush_tlb()
|