Home
last modified time | relevance | path

Searched refs:kvm_vcpu_get_esr (Results 1 – 12 of 12) sorted by relevance

/linux/arch/arm64/include/asm/
H A Dkvm_emulate.h315 static __always_inline u64 kvm_vcpu_get_esr(const struct kvm_vcpu *vcpu) in kvm_vcpu_get_esr() function
322 u64 esr = kvm_vcpu_get_esr(vcpu); in guest_hyp_wfx_traps_enabled()
335 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_get_condition()
365 return kvm_vcpu_get_esr(vcpu) & ESR_ELx_xVC_IMM_MASK; in kvm_vcpu_hvc_get_imm()
370 return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_ISV); in kvm_vcpu_dabt_isvalid()
375 return kvm_vcpu_get_esr(vcpu) & (ESR_ELx_CM | ESR_ELx_WNR | ESR_ELx_FSC); in kvm_vcpu_dabt_iss_nisv_sanitized()
380 return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_SSE); in kvm_vcpu_dabt_issext()
385 return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_SF); in kvm_vcpu_dabt_issf()
390 return (kvm_vcpu_get_esr(vcpu) & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT; in kvm_vcpu_dabt_get_rd()
395 return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_S1PTW); in kvm_vcpu_abt_iss1tw()
[all …]
/linux/arch/arm64/kvm/
H A Dhandle_exit.c49 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in handle_hvc()
108 return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_fpasimd()
132 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_wfx()
136 return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_wfx()
189 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_guest_debug()
213 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_unknown_ec()
229 return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in handle_sve()
256 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_ptrauth()
268 if (esr_iss_is_eretax(kvm_vcpu_get_esr(vcpu)) && in kvm_handle_eret()
286 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in kvm_handle_eret()
[all …]
H A Dpauth.c139 if (esr_iss_is_eretab(kvm_vcpu_get_esr(vcpu))) in corrupt_addr()
159 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_auth_eretax()
H A Dmmio.c171 trace_kvm_mmio_nisv(*vcpu_pc(vcpu), kvm_vcpu_get_esr(vcpu), in io_mem_abort()
H A Dnested.c146 esr = kvm_vcpu_get_esr(vcpu) & ~ESR_ELx_FSC; in esr_s2_fault()
377 result->esr |= (kvm_vcpu_get_esr(vcpu) & ~ESR_ELx_FSC); in kvm_walk_nested_s2()
1245 u64 esr = kvm_vcpu_get_esr(vcpu); in inject_vncr_perm()
1288 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_vncr_abort()
H A Dhypercalls.c248 u8 ec = ESR_ELx_EC(kvm_vcpu_get_esr(vcpu)); in kvm_prepare_hypercall_exit()
H A Dmmu.c1849 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_guest_abort()
1876 trace_kvm_guest_fault(*vcpu_pc(vcpu), kvm_vcpu_get_esr(vcpu), in kvm_handle_guest_abort()
1886 (unsigned long)kvm_vcpu_get_esr(vcpu)); in kvm_handle_guest_abort()
H A Dsys_regs.c4478 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_cp_64()
4576 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_cp10_id()
4670 params = esr_cp1x_32_to_params(kvm_vcpu_get_esr(vcpu)); in kvm_handle_cp15_32()
4697 params = esr_cp1x_32_to_params(kvm_vcpu_get_esr(vcpu)); in kvm_handle_cp14_32()
4895 unsigned long esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_sys_reg()
/linux/arch/arm64/kvm/hyp/vhe/
H A Dswitch.c264 esr = kvm_vcpu_get_esr(vcpu); in kvm_hyp_handle_timer()
329 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_hyp_handle_eret()
405 instr = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu)); in kvm_hyp_handle_tlbi_el2()
431 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_hyp_handle_cpacr_el1()
453 u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu)); in kvm_hyp_handle_zcr_el2()
/linux/arch/arm64/kvm/hyp/include/hyp/
H A Dswitch.h738 u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu)); in handle_tx2_tvm()
826 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu)); in kvm_handle_cntxct()
868 u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu)); in handle_ampere1_tcr()
/linux/arch/arm64/kvm/hyp/
H A Dvgic-v3-sr.c546 u64 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_get_group()
1164 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_perform_cpuif_access()
/linux/arch/arm64/kvm/hyp/nvhe/
H A Dsys_regs.c528 unsigned long esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_pvm_sysreg()