Home
last modified time | relevance | path

Searched refs:kvm_vcpu (Results 1 – 25 of 463) sorted by relevance

12345678910>>...19

/linux/arch/x86/kvm/vmx/
H A Dx86_ops.h22 int vmx_vcpu_create(struct kvm_vcpu *vcpu);
23 int vmx_vcpu_pre_run(struct kvm_vcpu *vcpu);
24 fastpath_t vmx_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags);
25 void vmx_vcpu_free(struct kvm_vcpu *vcpu);
26 void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
27 void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
28 void vmx_vcpu_put(struct kvm_vcpu *vcpu);
29 int vmx_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath);
30 void vmx_handle_exit_irqoff(struct kvm_vcpu *vcpu);
31 int vmx_skip_emulated_instruction(struct kvm_vcpu *vcpu);
[all …]
H A Dmain.c68 static int vt_vcpu_create(struct kvm_vcpu *vcpu) in vt_vcpu_create()
76 static void vt_vcpu_free(struct kvm_vcpu *vcpu) in vt_vcpu_free()
86 static void vt_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) in vt_vcpu_reset()
96 static void vt_vcpu_load(struct kvm_vcpu *vcpu, int cpu) in vt_vcpu_load()
106 static void vt_update_cpu_dirty_logging(struct kvm_vcpu *vcpu) in vt_update_cpu_dirty_logging()
118 static void vt_prepare_switch_to_guest(struct kvm_vcpu *vcpu) in vt_prepare_switch_to_guest()
128 static void vt_vcpu_put(struct kvm_vcpu *vcpu) in vt_vcpu_put()
138 static int vt_vcpu_pre_run(struct kvm_vcpu *vcpu) in vt_vcpu_pre_run()
146 static fastpath_t vt_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags) in vt_vcpu_run()
154 static int vt_handle_exit(struct kvm_vcpu *vcpu, in vt_handle_exit()
[all …]
H A Dvmx.h204 struct kvm_vcpu vcpu;
304 static __always_inline struct vcpu_vt *to_vt(struct kvm_vcpu *vcpu) in to_vt()
309 static __always_inline struct kvm_vcpu *vt_to_vcpu(struct vcpu_vt *vt) in vt_to_vcpu()
314 static __always_inline union vmx_exit_reason vmx_get_exit_reason(struct kvm_vcpu *vcpu) in vmx_get_exit_reason()
319 static __always_inline unsigned long vmx_get_exit_qual(struct kvm_vcpu *vcpu) in vmx_get_exit_qual()
330 static __always_inline u32 vmx_get_intr_info(struct kvm_vcpu *vcpu) in vmx_get_intr_info()
341 void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu);
345 void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
348 int vmx_get_cpl(struct kvm_vcpu *vcpu);
349 int vmx_get_cpl_no_cache(struct kvm_vcpu *vcpu);
[all …]
/linux/include/kvm/
H A Darm_pmu.h42 u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx);
43 void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
44 void kvm_pmu_set_counter_value_user(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
45 u64 kvm_pmu_implemented_counter_mask(struct kvm_vcpu *vcpu);
46 u64 kvm_pmu_accessible_counter_mask(struct kvm_vcpu *vcpu);
47 u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1);
48 void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu);
49 void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu);
50 void kvm_pmu_reprogram_counter_mask(struct kvm_vcpu *vcpu, u64 val);
51 void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/powerpc/include/asm/
H A Dkvm_ppc.h62 extern int kvmppc_vcpu_run(struct kvm_vcpu *vcpu);
63 extern int __kvmppc_vcpu_run(struct kvm_vcpu *vcpu);
66 extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu);
67 extern int kvmppc_handle_load(struct kvm_vcpu *vcpu,
70 extern int kvmppc_handle_loads(struct kvm_vcpu *vcpu,
73 extern int kvmppc_handle_vsx_load(struct kvm_vcpu *vcpu,
76 extern int kvmppc_handle_vmx_load(struct kvm_vcpu *vcpu,
78 extern int kvmppc_handle_vmx_store(struct kvm_vcpu *vcpu,
80 extern int kvmppc_handle_store(struct kvm_vcpu *vcpu,
83 extern int kvmppc_handle_vsx_store(struct kvm_vcpu *vcpu,
[all …]
H A Dkvm_book3s.h79 struct kvm_vcpu *runnable_threads[MAX_SMT_THREADS];
86 struct kvm_vcpu *runner;
146 extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask);
147 extern void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 vp, u64 vp_mask);
148 extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end);
149 extern void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 new_msr);
150 extern void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu);
151 extern void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu);
152 extern void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu);
153 extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte,
[all …]
H A Dkvm_booke.h26 static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val) in kvmppc_set_gpr()
31 static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) in kvmppc_get_gpr()
36 static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val) in kvmppc_set_cr()
41 static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu) in kvmppc_get_cr()
46 static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_xer()
51 static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu) in kvmppc_get_xer()
56 static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) in kvmppc_need_byteswap()
62 static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_ctr()
67 static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu) in kvmppc_get_ctr()
72 static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_lr()
[all …]
/linux/arch/arm64/include/asm/
H A Dkvm_emulate.h44 bool kvm_condition_valid32(const struct kvm_vcpu *vcpu);
45 void kvm_skip_instr32(struct kvm_vcpu *vcpu);
47 void kvm_inject_undefined(struct kvm_vcpu *vcpu);
48 void kvm_inject_sync(struct kvm_vcpu *vcpu, u64 esr);
49 int kvm_inject_serror_esr(struct kvm_vcpu *vcpu, u64 esr);
50 int kvm_inject_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr);
51 int kvm_inject_dabt_excl_atomic(struct kvm_vcpu *vcpu, u64 addr);
52 void kvm_inject_size_fault(struct kvm_vcpu *vcpu);
54 static inline int kvm_inject_sea_dabt(struct kvm_vcpu *vcpu, u64 addr) in kvm_inject_sea_dabt()
59 static inline int kvm_inject_sea_iabt(struct kvm_vcpu *vcpu, u64 addr) in kvm_inject_sea_iabt()
[all …]
/linux/arch/x86/include/asm/
H A Dkvm_host.h464 unsigned long (*get_guest_pgd)(struct kvm_vcpu *vcpu);
465 u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index);
466 int (*page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
467 void (*inject_page_fault)(struct kvm_vcpu *vcpu,
469 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
472 int (*sync_spte)(struct kvm_vcpu *vcpu,
542 struct kvm_vcpu *vcpu;
696 struct kvm_vcpu *vcpu;
945 int (*complete_userspace_io)(struct kvm_vcpu *vcpu);
1742 void (*vcpu_after_set_cpuid)(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/x86/kvm/
H A Dhyperv.h64 static inline struct kvm_vcpu_hv *to_hv_vcpu(struct kvm_vcpu *vcpu) in to_hv_vcpu()
69 static inline struct kvm_vcpu_hv_synic *to_hv_synic(struct kvm_vcpu *vcpu) in to_hv_synic()
76 static inline struct kvm_vcpu *hv_synic_to_vcpu(struct kvm_vcpu_hv_synic *synic) in hv_synic_to_vcpu()
83 static inline struct kvm_hv_syndbg *to_hv_syndbg(struct kvm_vcpu *vcpu) in to_hv_syndbg()
88 static inline u32 kvm_hv_get_vpindex(struct kvm_vcpu *vcpu) in kvm_hv_get_vpindex()
95 int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host);
96 int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host);
98 static inline bool kvm_hv_hypercall_enabled(struct kvm_vcpu *vcpu) in kvm_hv_hypercall_enabled()
103 int kvm_hv_hypercall(struct kvm_vcpu *vcpu);
108 void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector);
[all …]
H A Dlapic.h67 struct kvm_vcpu *vcpu;
93 int kvm_create_lapic(struct kvm_vcpu *vcpu);
94 void kvm_free_lapic(struct kvm_vcpu *vcpu);
96 int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu);
97 void kvm_apic_ack_interrupt(struct kvm_vcpu *vcpu, int vector);
98 int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu);
99 int kvm_apic_accept_events(struct kvm_vcpu *vcpu);
100 void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event);
101 u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu);
102 void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8);
[all …]
H A Dx86.h148 void kvm_service_local_tlb_flush_requests(struct kvm_vcpu *vcpu);
149 int kvm_check_nested_events(struct kvm_vcpu *vcpu);
152 static inline void kvm_leave_nested(struct kvm_vcpu *vcpu) in kvm_leave_nested()
165 static inline void kvm_nested_vmexit_handle_ibrs(struct kvm_vcpu *vcpu) in kvm_nested_vmexit_handle_ibrs()
186 static inline bool kvm_can_set_cpuid_and_feature_msrs(struct kvm_vcpu *vcpu) in kvm_can_set_cpuid_and_feature_msrs()
191 static inline void kvm_set_mp_state(struct kvm_vcpu *vcpu, int mp_state) in kvm_set_mp_state()
198 static inline bool kvm_is_exception_pending(struct kvm_vcpu *vcpu) in kvm_is_exception_pending()
205 static inline void kvm_clear_exception_queue(struct kvm_vcpu *vcpu) in kvm_clear_exception_queue()
212 static inline void kvm_queue_interrupt(struct kvm_vcpu *vcpu, u8 vector, in kvm_queue_interrupt()
220 static inline void kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu) in kvm_clear_interrupt_queue()
[all …]
H A Dxen.h20 int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu);
21 void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu);
22 void kvm_xen_inject_vcpu_vector(struct kvm_vcpu *vcpu);
23 int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
24 int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
28 int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data);
32 void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu);
33 void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu);
40 static inline void kvm_xen_sw_enable_lapic(struct kvm_vcpu *vcpu) in kvm_xen_sw_enable_lapic()
53 static inline bool kvm_xen_is_tsc_leaf(struct kvm_vcpu *vcpu, u32 function) in kvm_xen_is_tsc_leaf()
[all …]
H A Dpmu.h10 #define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu))
27 struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu,
29 struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr);
30 int (*check_rdpmc_early)(struct kvm_vcpu *vcpu, unsigned int idx);
31 bool (*is_valid_msr)(struct kvm_vcpu *vcpu, u32 msr);
32 int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
33 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
34 void (*refresh)(struct kvm_vcpu *vcpu);
35 void (*init)(struct kvm_vcpu *vcpu);
36 void (*reset)(struct kvm_vcpu *vcpu);
[all …]
H A Dcpuid.h21 void kvm_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu);
36 static inline struct kvm_cpuid_entry2 *kvm_find_cpuid_entry_index(struct kvm_vcpu *vcpu, in kvm_find_cpuid_entry_index()
43 static inline struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu, in kvm_find_cpuid_entry()
53 int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
56 int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu,
59 int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu,
62 bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx,
68 int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu);
69 int cpuid_query_maxguestphyaddr(struct kvm_vcpu *vcpu);
70 u64 kvm_vcpu_reserved_gpa_bits_raw(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/s390/kvm/
H A Dkvm-s390.h91 static inline void kvm_s390_set_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_set_cpuflags()
96 static inline void kvm_s390_clear_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_clear_cpuflags()
101 static inline bool kvm_s390_test_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_test_cpuflags()
106 static inline int is_vcpu_stopped(struct kvm_vcpu *vcpu) in is_vcpu_stopped()
111 static inline int is_vcpu_idle(struct kvm_vcpu *vcpu) in is_vcpu_idle()
126 static inline u32 kvm_s390_get_prefix(struct kvm_vcpu *vcpu) in kvm_s390_get_prefix()
131 static inline void kvm_s390_set_prefix(struct kvm_vcpu *vcpu, u32 prefix) in kvm_s390_set_prefix()
140 static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar) in kvm_s390_get_base_disp_s()
151 static inline u64 kvm_s390_get_base_disp_siy(struct kvm_vcpu *vcpu, u8 *ar) in kvm_s390_get_base_disp_siy()
166 static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu, in kvm_s390_get_base_disp_sse()
[all …]
/linux/arch/powerpc/kernel/
H A Dasm-offsets.c155 OFFSET(THREAD_KVM_VCPU, thread_struct, kvm_vcpu); in main()
355 OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); in main()
356 OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); in main()
357 OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); in main()
358 OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); in main()
359 OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); in main()
360 OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); in main()
362 OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); in main()
364 OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); in main()
365 OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); in main()
[all …]
/linux/arch/loongarch/include/asm/
H A Dkvm_vcpu.h41 typedef int (*exit_handle_fn)(struct kvm_vcpu *, int);
43 int kvm_emu_mmio_read(struct kvm_vcpu *vcpu, larch_inst inst);
44 int kvm_emu_mmio_write(struct kvm_vcpu *vcpu, larch_inst inst);
45 int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
46 int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
47 int kvm_complete_user_service(struct kvm_vcpu *vcpu, struct kvm_run *run);
48 int kvm_emu_idle(struct kvm_vcpu *vcpu);
49 int kvm_pending_timer(struct kvm_vcpu *vcpu);
50 int kvm_handle_fault(struct kvm_vcpu *vcpu, int fault);
51 void kvm_deliver_intr(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/riscv/include/asm/
H A Dkvm_vcpu_sbi.h47 int (*handler)(struct kvm_vcpu *vcpu, struct kvm_run *run,
51 unsigned long (*probe)(struct kvm_vcpu *vcpu);
58 int (*init)(struct kvm_vcpu *vcpu);
59 void (*deinit)(struct kvm_vcpu *vcpu);
61 void (*reset)(struct kvm_vcpu *vcpu);
64 unsigned long (*get_state_reg_count)(struct kvm_vcpu *vcpu);
65 int (*get_state_reg_id)(struct kvm_vcpu *vcpu, int index, u64 *reg_id);
66 int (*get_state_reg)(struct kvm_vcpu *vcpu, unsigned long reg_num,
68 int (*set_state_reg)(struct kvm_vcpu *vcpu, unsigned long reg_num,
72 int kvm_riscv_vcpu_sbi_forward_handler(struct kvm_vcpu *vcpu,
[all …]
H A Dkvm_aia.h90 bool kvm_riscv_vcpu_aia_imsic_has_interrupt(struct kvm_vcpu *vcpu);
91 void kvm_riscv_vcpu_aia_imsic_load(struct kvm_vcpu *vcpu, int cpu);
92 void kvm_riscv_vcpu_aia_imsic_put(struct kvm_vcpu *vcpu);
93 void kvm_riscv_vcpu_aia_imsic_release(struct kvm_vcpu *vcpu);
94 int kvm_riscv_vcpu_aia_imsic_update(struct kvm_vcpu *vcpu);
97 int kvm_riscv_vcpu_aia_imsic_rmw(struct kvm_vcpu *vcpu, unsigned long isel,
103 void kvm_riscv_vcpu_aia_imsic_reset(struct kvm_vcpu *vcpu);
104 int kvm_riscv_vcpu_aia_imsic_inject(struct kvm_vcpu *vcpu,
106 int kvm_riscv_vcpu_aia_imsic_init(struct kvm_vcpu *vcpu);
107 void kvm_riscv_vcpu_aia_imsic_cleanup(struct kvm_vcpu *vcpu);
[all …]
H A Dkvm_vcpu_pmu.h40 struct kvm_vcpu *vcpu;
64 #define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu_context))
75 int kvm_riscv_vcpu_pmu_incr_fw(struct kvm_vcpu *vcpu, unsigned long fid);
76 int kvm_riscv_vcpu_pmu_read_hpm(struct kvm_vcpu *vcpu, unsigned int csr_num,
80 int kvm_riscv_vcpu_pmu_num_ctrs(struct kvm_vcpu *vcpu, struct kvm_vcpu_sbi_return *retdata);
81 int kvm_riscv_vcpu_pmu_ctr_info(struct kvm_vcpu *vcpu, unsigned long cidx,
83 int kvm_riscv_vcpu_pmu_ctr_start(struct kvm_vcpu *vcpu, unsigned long ctr_base,
86 int kvm_riscv_vcpu_pmu_ctr_stop(struct kvm_vcpu *vcpu, unsigned long ctr_base,
89 int kvm_riscv_vcpu_pmu_ctr_cfg_match(struct kvm_vcpu *vcpu, unsigned long ctr_base,
93 int kvm_riscv_vcpu_pmu_fw_ctr_read(struct kvm_vcpu *vcpu, unsigned long cidx,
[all …]
/linux/arch/arm64/kvm/vgic/
H A Dvgic-mmio.h14 unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr,
20 void (*write)(struct kvm_vcpu *vcpu, gpa_t addr,
26 unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr,
29 int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr,
112 unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu,
115 unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu,
118 void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
121 int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
124 unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr,
127 void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr,
[all …]
H A Dvgic.h121 static inline u32 vgic_get_implementation_rev(struct kvm_vcpu *vcpu) in vgic_get_implementation_rev()
205 struct kvm_vcpu *vcpu;
259 vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev,
262 struct vgic_irq *vgic_get_vcpu_irq(struct kvm_vcpu *vcpu, u32 intid);
264 struct kvm_vcpu *vgic_target_oracle(struct vgic_irq *irq);
278 void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu);
279 void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr);
280 void vgic_v2_deactivate(struct kvm_vcpu *vcpu, u32 val);
281 void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr);
282 void vgic_v2_configure_hcr(struct kvm_vcpu *vcpu, struct ap_list_summary *als);
[all …]
/linux/arch/powerpc/kvm/
H A Dbooke.h65 void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr);
66 void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr);
68 void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr);
69 void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr);
70 void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
71 void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
73 int kvmppc_booke_emulate_op(struct kvm_vcpu *vcpu,
75 int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val);
76 int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val);
79 void kvmppc_load_guest_spe(struct kvm_vcpu *vcpu);
[all …]
/linux/arch/mips/include/asm/
H A Dkvm_host.h292 int (*vcpu_run)(struct kvm_vcpu *vcpu);
717 int (*handle_cop_unusable)(struct kvm_vcpu *vcpu);
718 int (*handle_tlb_mod)(struct kvm_vcpu *vcpu);
719 int (*handle_tlb_ld_miss)(struct kvm_vcpu *vcpu);
720 int (*handle_tlb_st_miss)(struct kvm_vcpu *vcpu);
721 int (*handle_addr_err_st)(struct kvm_vcpu *vcpu);
722 int (*handle_addr_err_ld)(struct kvm_vcpu *vcpu);
723 int (*handle_syscall)(struct kvm_vcpu *vcpu);
724 int (*handle_res_inst)(struct kvm_vcpu *vcpu);
725 int (*handle_break)(struct kvm_vcpu *vcpu);
[all …]

12345678910>>...19