Home
last modified time | relevance | path

Searched refs:vcpu (Results 1 – 25 of 555) sorted by relevance

12345678910>>...23

/linux/arch/arm64/include/asm/
H A Dkvm_emulate.h44 bool kvm_condition_valid32(const struct kvm_vcpu *vcpu);
45 void kvm_skip_instr32(struct kvm_vcpu *vcpu);
47 void kvm_inject_undefined(struct kvm_vcpu *vcpu);
48 void kvm_inject_sync(struct kvm_vcpu *vcpu, u64 esr);
49 int kvm_inject_serror_esr(struct kvm_vcpu *vcpu, u64 esr);
50 int kvm_inject_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr);
51 int kvm_inject_dabt_excl_atomic(struct kvm_vcpu *vcpu, u64 addr);
52 void kvm_inject_size_fault(struct kvm_vcpu *vcpu);
54 static inline int kvm_inject_sea_dabt(struct kvm_vcpu *vcpu, u64 addr) in kvm_inject_sea_dabt() argument
56 return kvm_inject_sea(vcpu, false, addr); in kvm_inject_sea_dabt()
[all …]
/linux/arch/arm64/kvm/
H A Dinject_fault.c18 static unsigned int exception_target_el(struct kvm_vcpu *vcpu) in exception_target_el() argument
21 if (likely(!vcpu_has_nv(vcpu))) in exception_target_el()
30 switch(*vcpu_cpsr(vcpu) & PSR_MODE_MASK) { in exception_target_el()
38 return vcpu_el2_tge_is_set(vcpu) ? PSR_MODE_EL2h : PSR_MODE_EL1h; in exception_target_el()
44 static enum vcpu_sysreg exception_esr_elx(struct kvm_vcpu *vcpu) in exception_esr_elx() argument
46 if (exception_target_el(vcpu) == PSR_MODE_EL2h) in exception_esr_elx()
52 static enum vcpu_sysreg exception_far_elx(struct kvm_vcpu *vcpu) in exception_far_elx() argument
54 if (exception_target_el(vcpu) == PSR_MODE_EL2h) in exception_far_elx()
60 static void pend_sync_exception(struct kvm_vcpu *vcpu) in pend_sync_exception() argument
62 if (exception_target_el(vcpu) == PSR_MODE_EL1h) in pend_sync_exception()
[all …]
H A Dhandle_exit.c32 static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u64 esr) in kvm_handle_guest_serror() argument
35 kvm_inject_serror(vcpu); in kvm_handle_guest_serror()
38 static int handle_hvc(struct kvm_vcpu *vcpu) in handle_hvc() argument
40 trace_kvm_hvc_arm64(*vcpu_pc(vcpu), vcpu_get_reg(vcpu, 0), in handle_hvc()
41 kvm_vcpu_hvc_get_imm(vcpu)); in handle_hvc()
42 vcpu->stat.hvc_exit_stat++; in handle_hvc()
45 if (vcpu_has_nv(vcpu)) { in handle_hvc()
46 if (vcpu_read_sys_reg(vcpu, HCR_EL2) & HCR_HCD) in handle_hvc()
47 kvm_inject_undefined(vcpu); in handle_hvc()
49 kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu)); in handle_hvc()
[all …]
/linux/arch/s390/kvm/
H A Dpriv.c33 static int handle_ri(struct kvm_vcpu *vcpu) in handle_ri() argument
35 vcpu->stat.instruction_ri++; in handle_ri()
37 if (test_kvm_facility(vcpu->kvm, 64)) { in handle_ri()
38 VCPU_EVENT(vcpu, 3, "%s", "ENABLE: RI (lazy)"); in handle_ri()
39 vcpu->arch.sie_block->ecb3 |= ECB3_RI; in handle_ri()
40 kvm_s390_retry_instr(vcpu); in handle_ri()
43 return kvm_s390_inject_program_int(vcpu, PGM_OPERATION); in handle_ri()
46 int kvm_s390_handle_aa(struct kvm_vcpu *vcpu) in kvm_s390_handle_aa() argument
48 if ((vcpu->arch.sie_block->ipa & 0xf) <= 4) in kvm_s390_handle_aa()
49 return handle_ri(vcpu); in kvm_s390_handle_aa()
[all …]
H A Dintercept.c26 u8 kvm_s390_get_ilen(struct kvm_vcpu *vcpu) in kvm_s390_get_ilen() argument
28 struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block; in kvm_s390_get_ilen()
31 switch (vcpu->arch.sie_block->icptcode) { in kvm_s390_get_ilen()
38 ilen = insn_length(vcpu->arch.sie_block->ipa >> 8); in kvm_s390_get_ilen()
48 ilen = vcpu->arch.sie_block->pgmilc & 0x6; in kvm_s390_get_ilen()
54 static int handle_stop(struct kvm_vcpu *vcpu) in handle_stop() argument
56 struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int; in handle_stop()
60 vcpu->stat.exit_stop_request++; in handle_stop()
63 if (kvm_s390_vcpu_has_irq(vcpu, 1)) in handle_stop()
69 stop_pending = kvm_s390_is_stop_irq_pending(vcpu); in handle_stop()
[all …]
H A Ddiag.c21 static void do_discard_gfn_range(struct kvm_vcpu *vcpu, gfn_t gfn_start, gfn_t gfn_end) in do_discard_gfn_range() argument
28 slots = kvm_vcpu_memslots(vcpu); in do_discard_gfn_range()
34 gmap_helper_discard(vcpu->kvm->mm, start, end); in do_discard_gfn_range()
38 static int diag_release_pages(struct kvm_vcpu *vcpu) in diag_release_pages() argument
41 unsigned long prefix = kvm_s390_get_prefix(vcpu); in diag_release_pages()
43 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages()
44 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages()
45 vcpu->stat.instruction_diagnose_10++; in diag_release_pages()
49 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION); in diag_release_pages()
51 VCPU_EVENT(vcpu, 5, "diag release pages %lX %lX", start, end); in diag_release_pages()
[all …]
H A Dguestdbg.c59 static void enable_all_hw_bp(struct kvm_vcpu *vcpu) in enable_all_hw_bp() argument
62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_bp()
63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_bp()
64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_bp()
67 if (vcpu->arch.guestdbg.nr_hw_bp <= 0 || in enable_all_hw_bp()
68 vcpu->arch.guestdbg.hw_bp_info == NULL) in enable_all_hw_bp()
79 for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) { in enable_all_hw_bp()
80 start = vcpu->arch.guestdbg.hw_bp_info[i].addr; in enable_all_hw_bp()
81 len = vcpu->arch.guestdbg.hw_bp_info[i].len; in enable_all_hw_bp()
99 static void enable_all_hw_wp(struct kvm_vcpu *vcpu) in enable_all_hw_wp() argument
[all …]
/linux/arch/x86/kvm/vmx/
H A Dmain.c68 static int vt_vcpu_create(struct kvm_vcpu *vcpu) in vt_vcpu_create() argument
70 if (is_td_vcpu(vcpu)) in vt_vcpu_create()
71 return tdx_vcpu_create(vcpu); in vt_vcpu_create()
73 return vmx_vcpu_create(vcpu); in vt_vcpu_create()
76 static void vt_vcpu_free(struct kvm_vcpu *vcpu) in vt_vcpu_free() argument
78 if (is_td_vcpu(vcpu)) { in vt_vcpu_free()
79 tdx_vcpu_free(vcpu); in vt_vcpu_free()
83 vmx_vcpu_free(vcpu); in vt_vcpu_free()
86 static void vt_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) in vt_vcpu_reset() argument
88 if (is_td_vcpu(vcpu)) { in vt_vcpu_reset()
[all …]
H A Dx86_ops.h22 int vmx_vcpu_create(struct kvm_vcpu *vcpu);
23 int vmx_vcpu_pre_run(struct kvm_vcpu *vcpu);
24 fastpath_t vmx_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags);
25 void vmx_vcpu_free(struct kvm_vcpu *vcpu);
26 void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
27 void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
28 void vmx_vcpu_put(struct kvm_vcpu *vcpu);
29 int vmx_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath);
30 void vmx_handle_exit_irqoff(struct kvm_vcpu *vcpu);
31 int vmx_skip_emulated_instruction(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/powerpc/kvm/
H A Dbooke.c90 void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu) in kvmppc_dump_vcpu() argument
94 printk("pc: %08lx msr: %08llx\n", vcpu->arch.regs.nip, in kvmppc_dump_vcpu()
95 vcpu->arch.shared->msr); in kvmppc_dump_vcpu()
96 printk("lr: %08lx ctr: %08lx\n", vcpu->arch.regs.link, in kvmppc_dump_vcpu()
97 vcpu->arch.regs.ctr); in kvmppc_dump_vcpu()
98 printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0, in kvmppc_dump_vcpu()
99 vcpu->arch.shared->srr1); in kvmppc_dump_vcpu()
101 printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions); in kvmppc_dump_vcpu()
105 kvmppc_get_gpr(vcpu, i), in kvmppc_dump_vcpu()
106 kvmppc_get_gpr(vcpu, i+1), in kvmppc_dump_vcpu()
[all …]
H A Dbook3s_emulate.c70 static bool spr_allowed(struct kvm_vcpu *vcpu, enum priv_level level) in spr_allowed() argument
73 if (vcpu->arch.papr_enabled && (level > PRIV_SUPER)) in spr_allowed()
77 if ((kvmppc_get_msr(vcpu) & MSR_PR) && level > PRIV_PROBLEM) in spr_allowed()
84 static inline void kvmppc_copyto_vcpu_tm(struct kvm_vcpu *vcpu) in kvmppc_copyto_vcpu_tm() argument
86 memcpy(&vcpu->arch.gpr_tm[0], &vcpu->arch.regs.gpr[0], in kvmppc_copyto_vcpu_tm()
87 sizeof(vcpu->arch.gpr_tm)); in kvmppc_copyto_vcpu_tm()
88 memcpy(&vcpu->arch.fp_tm, &vcpu->arch.fp, in kvmppc_copyto_vcpu_tm()
90 memcpy(&vcpu->arch.vr_tm, &vcpu->arch.vr, in kvmppc_copyto_vcpu_tm()
92 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm()
93 vcpu->arch.dscr_tm = vcpu->arch.dscr; in kvmppc_copyto_vcpu_tm()
[all …]
H A Dbooke_emulate.c24 static void kvmppc_emul_rfi(struct kvm_vcpu *vcpu) in kvmppc_emul_rfi() argument
26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi()
27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi()
30 static void kvmppc_emul_rfdi(struct kvm_vcpu *vcpu) in kvmppc_emul_rfdi() argument
32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi()
33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi()
36 static void kvmppc_emul_rfci(struct kvm_vcpu *vcpu) in kvmppc_emul_rfci() argument
38 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci()
39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci()
42 int kvmppc_booke_emulate_op(struct kvm_vcpu *vcpu, in kvmppc_booke_emulate_op() argument
[all …]
H A Dbook3s_pr.c53 static int kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr,
56 static int kvmppc_handle_fac(struct kvm_vcpu *vcpu, ulong fac);
67 static bool kvmppc_is_split_real(struct kvm_vcpu *vcpu) in kvmppc_is_split_real() argument
69 ulong msr = kvmppc_get_msr(vcpu); in kvmppc_is_split_real()
73 static void kvmppc_fixup_split_real(struct kvm_vcpu *vcpu) in kvmppc_fixup_split_real() argument
75 ulong msr = kvmppc_get_msr(vcpu); in kvmppc_fixup_split_real()
76 ulong pc = kvmppc_get_pc(vcpu); in kvmppc_fixup_split_real()
83 if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) in kvmppc_fixup_split_real()
90 vcpu->arch.hflags |= BOOK3S_HFLAG_SPLIT_HACK; in kvmppc_fixup_split_real()
91 kvmppc_set_pc(vcpu, pc | SPLIT_HACK_OFFS); in kvmppc_fixup_split_real()
[all …]
H A Demulate_loadstore.c28 static bool kvmppc_check_fp_disabled(struct kvm_vcpu *vcpu) in kvmppc_check_fp_disabled() argument
30 if (!(kvmppc_get_msr(vcpu) & MSR_FP)) { in kvmppc_check_fp_disabled()
31 kvmppc_core_queue_fpunavail(vcpu, kvmppc_get_msr(vcpu) & SRR1_PREFIXED); in kvmppc_check_fp_disabled()
40 static bool kvmppc_check_vsx_disabled(struct kvm_vcpu *vcpu) in kvmppc_check_vsx_disabled() argument
42 if (!(kvmppc_get_msr(vcpu) & MSR_VSX)) { in kvmppc_check_vsx_disabled()
43 kvmppc_core_queue_vsx_unavail(vcpu, kvmppc_get_msr(vcpu) & SRR1_PREFIXED); in kvmppc_check_vsx_disabled()
52 static bool kvmppc_check_altivec_disabled(struct kvm_vcpu *vcpu) in kvmppc_check_altivec_disabled() argument
54 if (!(kvmppc_get_msr(vcpu) & MSR_VEC)) { in kvmppc_check_altivec_disabled()
55 kvmppc_core_queue_vec_unavail(vcpu, kvmppc_get_msr(vcpu) & SRR1_PREFIXED); in kvmppc_check_altivec_disabled()
72 int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu) in kvmppc_emulate_loadstore() argument
[all …]
H A Dbook3s.c98 static inline void kvmppc_update_int_pending(struct kvm_vcpu *vcpu, in kvmppc_update_int_pending() argument
101 if (is_kvmppc_hv_enabled(vcpu->kvm)) in kvmppc_update_int_pending()
104 kvmppc_set_int_pending(vcpu, 1); in kvmppc_update_int_pending()
106 kvmppc_set_int_pending(vcpu, 0); in kvmppc_update_int_pending()
109 static inline bool kvmppc_critical_section(struct kvm_vcpu *vcpu) in kvmppc_critical_section() argument
115 if (is_kvmppc_hv_enabled(vcpu->kvm)) in kvmppc_critical_section()
118 crit_raw = kvmppc_get_critical(vcpu); in kvmppc_critical_section()
119 crit_r1 = kvmppc_get_gpr(vcpu, 1); in kvmppc_critical_section()
122 if (!(kvmppc_get_msr(vcpu) & MSR_SF)) { in kvmppc_critical_section()
130 crit = crit && !(kvmppc_get_msr(vcpu) & MSR_PR); in kvmppc_critical_section()
[all …]
H A Dbook3s_hv_tm.c16 static void emulate_tx_failure(struct kvm_vcpu *vcpu, u64 failure_cause) in emulate_tx_failure() argument
19 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure()
21 tfiar = vcpu->arch.regs.nip & ~0x3ull; in emulate_tx_failure()
23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure()
29 vcpu->arch.tfiar = tfiar; in emulate_tx_failure()
31 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure()
42 int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu) in kvmhv_p9_tm_emulation() argument
44 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation()
45 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation()
56 vcpu->arch.regs.nip -= 4; in kvmhv_p9_tm_emulation()
[all …]
/linux/arch/riscv/kvm/
H A Dvcpu.c54 static void kvm_riscv_vcpu_context_reset(struct kvm_vcpu *vcpu, in kvm_riscv_vcpu_context_reset() argument
57 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_context_reset()
58 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_context_reset()
63 memset(&vcpu->arch.smstateen_csr, 0, sizeof(vcpu->arch.smstateen_csr)); in kvm_riscv_vcpu_context_reset()
69 kvm_riscv_vcpu_sbi_load_reset_state(vcpu); in kvm_riscv_vcpu_context_reset()
79 static void kvm_riscv_reset_vcpu(struct kvm_vcpu *vcpu, bool kvm_sbi_reset) in kvm_riscv_reset_vcpu() argument
89 loaded = (vcpu->cpu != -1); in kvm_riscv_reset_vcpu()
91 kvm_arch_vcpu_put(vcpu); in kvm_riscv_reset_vcpu()
93 vcpu->arch.last_exit_cpu = -1; in kvm_riscv_reset_vcpu()
95 kvm_riscv_vcpu_context_reset(vcpu, kvm_sbi_reset); in kvm_riscv_reset_vcpu()
[all …]
/linux/arch/arm64/kvm/hyp/vhe/
H A Dswitch.c52 static u64 __compute_hcr(struct kvm_vcpu *vcpu) in __compute_hcr() argument
54 u64 guest_hcr, hcr = vcpu->arch.hcr_el2; in __compute_hcr()
56 if (!vcpu_has_nv(vcpu)) in __compute_hcr()
65 if (is_hyp_ctxt(vcpu)) { in __compute_hcr()
70 if (!vcpu_el2_e2h_is_set(vcpu)) in __compute_hcr()
81 guest_hcr = kvm_vcpu_apply_reg_masks(vcpu, HCR_EL2, 0); in __compute_hcr()
83 write_sysreg_s(vcpu->arch.ctxt.vncr_array, SYS_VNCR_EL2); in __compute_hcr()
87 guest_hcr = __vcpu_sys_reg(vcpu, HCR_EL2); in __compute_hcr()
92 va |= __vcpu_sys_reg(vcpu, VNCR_EL2) & GENMASK(PAGE_SHIFT - 1, 0); in __compute_hcr()
113 static void __activate_traps(struct kvm_vcpu *vcpu) in __activate_traps() argument
[all …]
/linux/arch/mips/kvm/
H A Demulate.c40 static int kvm_compute_return_epc(struct kvm_vcpu *vcpu, unsigned long instpc, in kvm_compute_return_epc() argument
45 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_compute_return_epc()
56 err = kvm_get_badinstrp((u32 *)epc, vcpu, &insn.word); in kvm_compute_return_epc()
243 enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause) in update_pc() argument
248 err = kvm_compute_return_epc(vcpu, vcpu->arch.pc, in update_pc()
249 &vcpu->arch.pc); in update_pc()
253 vcpu->arch.pc += 4; in update_pc()
256 kvm_debug("update_pc(): New PC: %#lx\n", vcpu->arch.pc); in update_pc()
272 int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out) in kvm_get_badinstr() argument
275 *out = vcpu->arch.host_cp0_badinstr; in kvm_get_badinstr()
[all …]
/linux/arch/arm64/kvm/hyp/include/hyp/
H A Dswitch.h43 static inline void __fpsimd_save_fpexc32(struct kvm_vcpu *vcpu) in __fpsimd_save_fpexc32() argument
45 if (!vcpu_el1_is_32bit(vcpu)) in __fpsimd_save_fpexc32()
48 __vcpu_assign_sys_reg(vcpu, FPEXC32_EL2, read_sysreg(fpexc32_el2)); in __fpsimd_save_fpexc32()
51 static inline void __activate_traps_fpsimd32(struct kvm_vcpu *vcpu) in __activate_traps_fpsimd32() argument
62 if (vcpu_el1_is_32bit(vcpu) && system_supports_fpsimd()) in __activate_traps_fpsimd32()
66 static inline void __activate_cptr_traps_nvhe(struct kvm_vcpu *vcpu) in __activate_cptr_traps_nvhe() argument
76 if (!vcpu_has_sve(vcpu) || !guest_owns_fp_regs()) in __activate_cptr_traps_nvhe()
85 static inline void __activate_cptr_traps_vhe(struct kvm_vcpu *vcpu) in __activate_cptr_traps_vhe() argument
100 if (vcpu_has_sve(vcpu)) in __activate_cptr_traps_vhe()
104 if (!vcpu_has_nv(vcpu)) in __activate_cptr_traps_vhe()
[all …]
/linux/arch/x86/kvm/
H A Dx86.c107 ((struct kvm_vcpu *)(ctxt)->vcpu)
129 static void update_cr8_intercept(struct kvm_vcpu *vcpu);
130 static void process_nmi(struct kvm_vcpu *vcpu);
131 static void __kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
132 static void store_regs(struct kvm_vcpu *vcpu);
133 static int sync_regs(struct kvm_vcpu *vcpu);
134 static int kvm_vcpu_do_singlestep(struct kvm_vcpu *vcpu);
136 static int __set_sregs2(struct kvm_vcpu *vcpu, struct kvm_sregs2 *sregs2);
137 static void __get_sregs2(struct kvm_vcpu *vcpu, struct kvm_sregs2 *sregs2);
140 static void kvm_load_guest_fpu(struct kvm_vcpu *vcpu);
[all …]
H A Dhyperv.h64 static inline struct kvm_vcpu_hv *to_hv_vcpu(struct kvm_vcpu *vcpu) in to_hv_vcpu() argument
66 return vcpu->arch.hyperv; in to_hv_vcpu()
69 static inline struct kvm_vcpu_hv_synic *to_hv_synic(struct kvm_vcpu *vcpu) in to_hv_synic() argument
71 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in to_hv_synic()
80 return hv_vcpu->vcpu; in hv_synic_to_vcpu()
83 static inline struct kvm_hv_syndbg *to_hv_syndbg(struct kvm_vcpu *vcpu) in to_hv_syndbg() argument
85 return &vcpu->kvm->arch.hyperv.hv_syndbg; in to_hv_syndbg()
88 static inline u32 kvm_hv_get_vpindex(struct kvm_vcpu *vcpu) in kvm_hv_get_vpindex() argument
90 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_get_vpindex()
92 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex()
[all …]
H A Dkvm_cache_regs.h20 static __always_inline unsigned long kvm_##lname##_read(struct kvm_vcpu *vcpu)\
22 return vcpu->arch.regs[VCPU_REGS_##uname]; \
24 static __always_inline void kvm_##lname##_write(struct kvm_vcpu *vcpu, \
27 vcpu->arch.regs[VCPU_REGS_##uname] = val; \
56 #define kvm_assert_register_caching_allowed(vcpu) \ in BUILD_KVM_GPR_ACCESSORS() argument
57 lockdep_assert_once(in_task() || kvm_arch_pmi_in_guest(vcpu)) in BUILD_KVM_GPR_ACCESSORS()
66 static inline bool kvm_register_is_available(struct kvm_vcpu *vcpu, in BUILD_KVM_GPR_ACCESSORS()
69 kvm_assert_register_caching_allowed(vcpu); in BUILD_KVM_GPR_ACCESSORS()
70 return test_bit(reg, (unsigned long *)&vcpu->arch.regs_avail); in BUILD_KVM_GPR_ACCESSORS()
73 static inline bool kvm_register_is_dirty(struct kvm_vcpu *vcpu, in kvm_register_is_dirty() argument
[all …]
/linux/arch/loongarch/kvm/
H A Dvcpu.c41 static inline void kvm_save_host_pmu(struct kvm_vcpu *vcpu) in kvm_save_host_pmu() argument
45 context = this_cpu_ptr(vcpu->kvm->arch.vmcs); in kvm_save_host_pmu()
56 static inline void kvm_restore_host_pmu(struct kvm_vcpu *vcpu) in kvm_restore_host_pmu() argument
60 context = this_cpu_ptr(vcpu->kvm->arch.vmcs); in kvm_restore_host_pmu()
72 static inline void kvm_save_guest_pmu(struct kvm_vcpu *vcpu) in kvm_save_guest_pmu() argument
74 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_save_guest_pmu()
86 static inline void kvm_restore_guest_pmu(struct kvm_vcpu *vcpu) in kvm_restore_guest_pmu() argument
88 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_restore_guest_pmu()
100 static int kvm_own_pmu(struct kvm_vcpu *vcpu) in kvm_own_pmu() argument
104 if (!kvm_guest_has_pmu(&vcpu->arch)) in kvm_own_pmu()
[all …]
H A Dexit.c24 static int kvm_emu_cpucfg(struct kvm_vcpu *vcpu, larch_inst inst) in kvm_emu_cpucfg() argument
34 ++vcpu->stat.cpucfg_exits; in kvm_emu_cpucfg()
35 index = vcpu->arch.gprs[rj]; in kvm_emu_cpucfg()
46 vcpu->arch.gprs[rd] = vcpu->arch.cpucfg[index]; in kvm_emu_cpucfg()
50 vcpu->arch.gprs[rd] = *(unsigned int *)KVM_SIGNATURE; in kvm_emu_cpucfg()
53 ret = vcpu->kvm->arch.pv_features & LOONGARCH_PV_FEAT_MASK; in kvm_emu_cpucfg()
54 vcpu->arch.gprs[rd] = ret; in kvm_emu_cpucfg()
57 vcpu->arch.gprs[rd] = 0; in kvm_emu_cpucfg()
65 static unsigned long kvm_emu_read_csr(struct kvm_vcpu *vcpu, int csrid) in kvm_emu_read_csr() argument
68 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_emu_read_csr()
[all …]

12345678910>>...23