Searched refs:is_guest_mode (Results 1 – 19 of 19) sorted by relevance
198 bool is_guest_mode) in kvm_hv_get_tlb_flush_fifo() argument201 int i = is_guest_mode ? HV_L2_TLB_FLUSH_FIFO : in kvm_hv_get_tlb_flush_fifo()214 tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu)); in kvm_hv_vcpu_purge_flush_tlb()
1976 tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu)); in kvm_hv_vcpu_flush_tlb()2043 if (!hc->fast && is_guest_mode(vcpu)) { in kvm_hv_flush_tlb()2065 is_guest_mode(vcpu)); in kvm_hv_flush_tlb()2096 flush_ex.flags, is_guest_mode(vcpu)); in kvm_hv_flush_tlb()2140 if (all_cpus && !is_guest_mode(vcpu)) { in kvm_hv_flush_tlb()2148 } else if (!is_guest_mode(vcpu)) { in kvm_hv_flush_tlb()2390 if (hv_result_success(result) && is_guest_mode(vcpu) && in kvm_hv_hypercall_complete()
244 static inline bool is_guest_mode(struct kvm_vcpu *vcpu) in is_guest_mode() function
850 if (is_guest_mode(vcpu) && in kvm_multiple_exception()867 if (!is_guest_mode(vcpu)) in kvm_multiple_exception()979 if (is_guest_mode(vcpu) && fault->async_page_fault) in kvm_inject_page_fault()2724 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_offset()2740 if (is_guest_mode(vcpu)) in kvm_vcpu_write_tsc_multiplier()9164 if (!is_guest_mode(vcpu) && kvm_x86_call(get_cpl)(vcpu) == 0) { in handle_emulation_failure()9424 (WARN_ON_ONCE(is_guest_mode(vcpu)) || in x86_emulate_instruction()9558 r = x86_emulate_insn(ctxt, is_guest_mode(vcpu) && in x86_emulate_instruction()10542 if (is_guest_mode(vcpu)) in post_kvm_run_save()10652 if (is_guest_mode(vcpu)) in kvm_check_and_inject_events()[all …]
101 if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v)) in kvm_cpu_has_injectable_intr()
188 return vcpu->arch.last_vmentry_cpu == -1 && !is_guest_mode(vcpu); in kvm_can_set_cpuid_and_feature_msrs()
660 if (ret != X86EMUL_CONTINUE && is_guest_mode(vcpu)) in emulator_leave_smm()
3533 if (is_guest_mode(vcpu)) { in kvm_apic_accept_events()
961 if (is_guest_mode(vcpu)) in vmx_update_exception_bitmap()1619 if (!is_guest_mode(vcpu)) { in vmx_load_vmcs01()1630 if (!is_guest_mode(vcpu)) in vmx_put_vmcs01()1893 if (!is_guest_mode(vcpu)) in vmx_update_emulated_instruction()2375 if (is_guest_mode(vcpu)) in vmx_set_msr()2380 if (is_guest_mode(vcpu)) { in vmx_set_msr()2387 if (is_guest_mode(vcpu)) { in vmx_set_msr()2399 if (is_guest_mode(vcpu) && get_vmcs12(vcpu)->vm_exit_controls & in vmx_set_msr()2418 if (is_guest_mode(vcpu) && in vmx_set_msr()2475 if (is_guest_mode(vcpu) && in vmx_set_msr()[all …]
3528 if (is_guest_mode(vcpu) && !nested_get_vmcs12_pages(vcpu)) in vmx_get_nested_state_pages()3540 if (WARN_ON_ONCE(!is_guest_mode(vcpu))) in nested_vmx_write_pml_buffer()5633 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmread()5657 (is_guest_mode(vcpu) && in handle_vmread()5665 if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field)) in handle_vmread()5681 if (WARN_ON_ONCE(is_guest_mode(vcpu))) in handle_vmread()5739 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmwrite()5767 (is_guest_mode(vcpu) && in handle_vmwrite()5801 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) in handle_vmwrite()5823 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) { in handle_vmwrite()[all …]
504 if (!vmcs12 && is_guest_mode(vcpu)) in vmx_write_encls_bitmap()
697 return enable_unrestricted_guest && (!is_guest_mode(vcpu) || in is_unrestricted_guest()
40 bool is_guest_mode; member
661 void *msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm : in msr_write_intercepted()877 (is_guest_mode(vcpu) && guest_cpu_cap_has(vcpu, X86_FEATURE_LBRV) && in svm_update_lbrv()1536 if (is_guest_mode(&svm->vcpu)) { in svm_clear_vintr()1940 if ((error_code & PFERR_RSVD_MASK) && !is_guest_mode(vcpu)) { in npf_interception()2249 if (is_guest_mode(vcpu)) { in emulate_svm_instr()2291 if (!is_guest_mode(vcpu)) in gp_interception()2509 if (!is_guest_mode(vcpu) || in check_selective_cr0_intercepted()2934 is_guest_mode(vcpu)) in svm_set_msr()2944 if (is_guest_mode(vcpu)) in svm_set_msr()3226 if (is_guest_mode(vcpu)) in bus_lock_exit()[all …]
543 if (is_guest_mode(&svm->vcpu) && !nested_vgif_enabled(svm)) in get_vgif_vmcb()604 if (is_guest_mode(&svm->vcpu)) in get_vnmi_vmcb_l1()759 return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK); in nested_svm_virtualize_tpr()
133 if (!is_guest_mode(&svm->vcpu)) in recalc_intercepts()1409 if (is_guest_mode(vcpu)) { in svm_leave_nested()1767 if (is_guest_mode(vcpu)) { in svm_get_nested_state()1782 if (!is_guest_mode(vcpu)) in svm_get_nested_state()1902 if (is_guest_mode(vcpu)) in svm_set_nested_state()1944 if (WARN_ON(!is_guest_mode(vcpu))) in svm_get_nested_state_pages()
236 if (is_guest_mode(&svm->vcpu) && in avic_deactivate_vmcb()663 if (is_guest_mode(vcpu)) in avic_vcpu_get_apicv_inhibit_reasons()894 .is_guest_mode = kvm_vcpu_apicv_active(vcpu), in avic_pi_update_irte()
4699 if (is_guest_mode(vcpu)) in kvm_mmu_faultin_pfn()5705 role.base.guest_mode = is_guest_mode(vcpu); in kvm_calc_cpu_role()6392 if (!is_guest_mode(vcpu)) in kvm_mmu_write_protect_fault()
4081 if (pi_data->is_guest_mode) in amd_ir_set_vcpu_affinity()