/linux/arch/x86/kvm/vmx/ |
H A D | nested.c | 192 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_fail() local 222 static void vmx_disable_shadow_vmcs(struct vcpu_vmx *vmx) in vmx_disable_shadow_vmcs() argument 233 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_evmcs() local 250 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_evmcs_handle_vmclear() local 274 vmx_sync_vmcs_host_state(struct vcpu_vmx * vmx,struct loaded_vmcs * prev) vmx_sync_vmcs_host_state() argument 295 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_switch_vmcs() local 320 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_put_vmcs12_pages() local 334 struct vcpu_vmx *vmx = to_vmx(vcpu); free_nested() local 413 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_ept_inject_page_fault() local 466 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_ept_new_eptp() local 624 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_prepare_msr_bitmap() local 734 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_cache_shadow_vmcs12() local 753 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_flush_cached_shadow_vmcs12() local 836 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_max_atomic_switch_msrs() local 1017 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_get_vmexit_msr_value() local 1115 struct vcpu_vmx *vmx = to_vmx(vcpu); prepare_vmx_msr_autostore_list() local 1210 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_transition_tlb_flush() local 1271 vmx_restore_vmx_basic(struct vcpu_vmx * vmx,u64 data) vmx_restore_vmx_basic() argument 1342 vmx_restore_control_msr(struct vcpu_vmx * vmx,u32 msr_index,u64 data) vmx_restore_control_msr() argument 1365 vmx_restore_vmx_misc(struct vcpu_vmx * vmx,u64 data) vmx_restore_vmx_misc() argument 1413 vmx_restore_vmx_ept_vpid_cap(struct vcpu_vmx * vmx,u64 data) vmx_restore_vmx_ept_vpid_cap() argument 1439 vmx_restore_fixed0_msr(struct vcpu_vmx * vmx,u32 msr_index,u64 data) vmx_restore_fixed0_msr() argument 1461 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_vmx_msr() local 1609 copy_shadow_to_vmcs12(struct vcpu_vmx * vmx) copy_shadow_to_vmcs12() argument 1636 copy_vmcs12_to_shadow(struct vcpu_vmx * vmx) copy_vmcs12_to_shadow() argument 1670 copy_enlightened_to_vmcs12(struct vcpu_vmx * vmx,u32 hv_clean_fields) copy_enlightened_to_vmcs12() argument 1918 copy_vmcs12_to_enlightened(struct vcpu_vmx * vmx) copy_vmcs12_to_enlightened() argument 2104 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_handle_enlightened_vmptrld() local 2192 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_sync_vmcs12_to_shadow() local 2204 struct vcpu_vmx *vmx = vmx_preemption_timer_fn() local 2216 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_calc_preemption_timer_value() local 2233 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_start_preemption_timer() local 2255 nested_vmx_calc_efer(struct vcpu_vmx * vmx,struct vmcs12 * vmcs12) nested_vmx_calc_efer() argument 2266 prepare_vmcs02_constant_state(struct vcpu_vmx * vmx) prepare_vmcs02_constant_state() argument 2330 prepare_vmcs02_early_rare(struct vcpu_vmx * vmx,struct vmcs12 * vmcs12) prepare_vmcs02_early_rare() argument 2356 prepare_vmcs02_early(struct vcpu_vmx * vmx,struct loaded_vmcs * vmcs01,struct vmcs12 * vmcs12) prepare_vmcs02_early() argument 2523 prepare_vmcs02_rare(struct vcpu_vmx * vmx,struct vmcs12 * vmcs12) prepare_vmcs02_rare() argument 2657 struct vcpu_vmx *vmx = to_vmx(vcpu); prepare_vmcs02() local 2809 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_check_eptp() local 2858 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_check_vm_execution_controls() local 2917 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_check_vm_exit_controls() local 2934 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_check_vm_entry_controls() local 3113 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_check_vmcs_link_ptr() local 3222 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_check_vmentry_hw() local 3301 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_get_evmcs_page() local 3331 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_get_vmcs12_pages() local 3446 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_write_pml_buffer() local 3518 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_enter_non_root_mode() local 3680 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_run() local 3937 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_complete_nested_posted_interrupt() local 4087 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_has_nested_events() local 4213 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_check_nested_events() local 4485 struct vcpu_vmx *vmx = to_vmx(vcpu); sync_vmcs02_to_vmcs12_rare() local 4530 struct vcpu_vmx *vmx = to_vmx(vcpu); copy_vmcs02_to_vmcs12_rare() local 4558 struct vcpu_vmx *vmx = to_vmx(vcpu); sync_vmcs02_to_vmcs12() local 4824 nested_vmx_get_vmcs01_guest_efer(struct vcpu_vmx * vmx) nested_vmx_get_vmcs01_guest_efer() argument 4850 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_restore_host_state() local 4964 struct vcpu_vmx *vmx = to_vmx(vcpu); __nested_vmx_vmexit() local 5308 struct vcpu_vmx *vmx = to_vmx(vcpu); alloc_shadow_vmcs() local 5331 struct vcpu_vmx *vmx = to_vmx(vcpu); enter_vmx_operation() local 5384 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmxon() local 5472 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_release_vmcs12() local 5514 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmclear() local 5572 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmread() local 5678 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmwrite() local 5779 set_current_vmptr(struct vcpu_vmx * vmx,gpa_t vmptr) set_current_vmptr() argument 5795 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmptrld() local 5893 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_invept() local 5973 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_invvpid() local 6105 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmfunc() local 6592 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_reflect_vmexit() local 6646 struct vcpu_vmx *vmx; vmx_get_nested_state() local 6778 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_nested_state() local [all...] |
H A D | vmx.c | 356 static __always_inline void vmx_disable_fb_clear(struct vcpu_vmx *vmx) in vmx_disable_fb_clear() argument 370 static __always_inline void vmx_enable_fb_clear(struct vcpu_vmx *vmx) in vmx_enable_fb_clear() argument 379 static void vmx_update_fb_clear_dis(struct kvm_vcpu *vcpu, struct vcpu_vmx *vmx) in vmx_update_fb_clear_dis() argument 652 vmx_find_uret_msr(struct vcpu_vmx * vmx,u32 msr) vmx_find_uret_msr() argument 662 vmx_set_guest_uret_msr(struct vcpu_vmx * vmx,struct vmx_uret_msr * msr,u64 data) vmx_set_guest_uret_msr() argument 765 vmx_segment_cache_test_set(struct vcpu_vmx * vmx,unsigned seg,unsigned field) vmx_segment_cache_test_set() argument 780 vmx_read_guest_seg_selector(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_selector() argument 789 vmx_read_guest_seg_base(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_base() argument 798 vmx_read_guest_seg_limit(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_limit() argument 807 vmx_read_guest_seg_ar(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_ar() argument 884 msr_write_intercepted(struct vcpu_vmx * vmx,u32 msr) msr_write_intercepted() argument 892 __vmx_vcpu_run_flags(struct vcpu_vmx * vmx) __vmx_vcpu_run_flags() argument 914 clear_atomic_switch_msr_special(struct vcpu_vmx * vmx,unsigned long entry,unsigned long exit) clear_atomic_switch_msr_special() argument 932 clear_atomic_switch_msr(struct vcpu_vmx * vmx,unsigned msr) clear_atomic_switch_msr() argument 972 add_atomic_switch_msr_special(struct vcpu_vmx * vmx,unsigned long entry,unsigned long exit,unsigned long guest_val_vmcs,unsigned long host_val_vmcs,u64 guest_val,u64 host_val) add_atomic_switch_msr_special() argument 984 add_atomic_switch_msr(struct vcpu_vmx * vmx,unsigned msr,u64 guest_val,u64 host_val,bool entry_only) add_atomic_switch_msr() argument 1050 update_transition_efer(struct vcpu_vmx * vmx) update_transition_efer() argument 1132 pt_can_write_msr(struct vcpu_vmx * vmx) pt_can_write_msr() argument 1172 pt_guest_enter(struct vcpu_vmx * vmx) pt_guest_enter() argument 1189 pt_guest_exit(struct vcpu_vmx * vmx) pt_guest_exit() argument 1236 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_prepare_switch_to_guest() local 1307 vmx_prepare_switch_to_host(struct vcpu_vmx * vmx) vmx_prepare_switch_to_host() argument 1347 vmx_read_guest_kernel_gs_base(struct vcpu_vmx * vmx) vmx_read_guest_kernel_gs_base() argument 1356 vmx_write_guest_kernel_gs_base(struct vcpu_vmx * vmx,u64 data) vmx_write_guest_kernel_gs_base() argument 1368 struct vcpu_vmx *vmx = to_vmx(vcpu); grow_ple_window() local 1384 struct vcpu_vmx *vmx = to_vmx(vcpu); shrink_ple_window() local 1400 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_load_vmcs() local 1482 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_rflags() local 1500 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_rflags() local 1563 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_rtit_ctl_check() local 1728 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_emulated_instruction() local 1779 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_inject_exception() local 1820 vmx_setup_uret_msr(struct vcpu_vmx * vmx,unsigned int msr,bool load_into_hardware) vmx_setup_uret_msr() argument 1838 vmx_setup_uret_msrs(struct vcpu_vmx * vmx) vmx_setup_uret_msrs() argument 1920 is_vmx_feature_control_msr_valid(struct vcpu_vmx * vmx,struct msr_data * msr) is_vmx_feature_control_msr_valid() argument 1963 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_msr() local 2160 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_msr() local 2997 struct vcpu_vmx *vmx = to_vmx(vcpu); enter_pmode() local 3067 struct vcpu_vmx *vmx = to_vmx(vcpu); enter_rmode() local 3114 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_efer() local 3163 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_flush_tlb_all() local 3274 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_cr0() local 3430 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_cr4() local 3493 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_segment() local 3540 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_get_cpl() local 3582 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_set_segment() local 3952 vmx_msr_bitmap_l01_changed(struct vcpu_vmx * vmx) vmx_msr_bitmap_l01_changed() argument 3972 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_intercept_for_msr() local 4004 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_msr_bitmap_x2apic() local 4057 struct vcpu_vmx *vmx = to_vmx(vcpu); pt_update_intercept_for_msr() local 4127 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_deliver_nested_posted_interrupt() local 4208 vmx_set_constant_host_state(struct vcpu_vmx * vmx) vmx_set_constant_host_state() argument 4275 set_cr4_guest_host_mask(struct vcpu_vmx * vmx) set_cr4_guest_host_mask() argument 4291 vmx_pin_based_exec_ctrl(struct vcpu_vmx * vmx) vmx_pin_based_exec_ctrl() argument 4345 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_refresh_apicv_exec_ctrl() local 4371 vmx_exec_control(struct vcpu_vmx * vmx) vmx_exec_control() argument 4415 vmx_tertiary_exec_control(struct vcpu_vmx * vmx) vmx_tertiary_exec_control() argument 4435 vmx_adjust_secondary_exec_control(struct vcpu_vmx * vmx,u32 * exec_control,u32 control,bool enabled,bool exiting) vmx_adjust_secondary_exec_control() argument 4474 vmx_adjust_sec_exec_control(vmx,exec_control,name,feat_name,ctrl_name,exiting) global() argument 4487 vmx_adjust_sec_exec_feature(vmx,exec_control,lname,uname) global() argument 4490 vmx_adjust_sec_exec_exiting(vmx,exec_control,lname,uname) global() argument 4493 vmx_secondary_exec_control(struct vcpu_vmx * vmx) vmx_secondary_exec_control() argument 4613 init_vmcs(struct vcpu_vmx * vmx) init_vmcs() argument 4738 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_vcpu_reset() local 4770 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_reset() local 4845 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_inject_irq() local 4873 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_inject_nmi() local 4904 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_nmi_mask() local 4918 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_nmi_mask() local 5107 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_exception_nmi() local 5619 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_task_switch() local 5747 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_unhandleable_emulation_required() local 5774 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_invalid_guest_state() local 5906 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_fastpath_preemption_timer() local 6072 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_exit_info() local 6099 vmx_destroy_pml_buffer(struct vcpu_vmx * vmx) vmx_destroy_pml_buffer() argument 6109 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_flush_pml_buffer() local 6174 struct vcpu_vmx *vmx = to_vmx(vcpu); dump_vmcs() local 6371 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_handle_exit() local 6637 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_virtual_apic_mode() local 6963 vmx_recover_nmi_blocking(struct vcpu_vmx * vmx) vmx_recover_nmi_blocking() argument 7061 vmx_complete_interrupts(struct vcpu_vmx * vmx) vmx_complete_interrupts() argument 7078 atomic_switch_perf_msrs(struct vcpu_vmx * vmx) atomic_switch_perf_msrs() argument 7103 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_hv_timer() local 7127 vmx_update_host_rsp(struct vcpu_vmx * vmx,unsigned long host_rsp) vmx_update_host_rsp() argument 7135 vmx_spec_ctrl_restore_host(struct vcpu_vmx * vmx,unsigned int flags) vmx_spec_ctrl_restore_host() argument 7200 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_enter_exit() local 7253 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_run() local 7410 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_free() local 7423 struct vcpu_vmx *vmx; vmx_vcpu_create() local 7585 vmcs_set_secondary_exec_control(struct vcpu_vmx * vmx,u32 new_ctl) vmcs_set_secondary_exec_control() argument 7610 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_cr_fixed1_bits_update() local 7653 struct vcpu_vmx *vmx = to_vmx(vcpu); update_intel_pt_cfg() local 7722 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_after_set_cpuid() local 8053 struct vcpu_vmx *vmx; vmx_set_hv_timer() local 8098 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_cpu_dirty_logging() local 8140 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_enter_smm() local 8161 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_leave_smm() local [all...] |
H A D | nested.h | 81 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_has_valid_vmcs12() local 90 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vpid02() local
|
H A D | sgx.c | 300 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_encls_einit() local 430 struct vcpu_vmx *vmx = to_vmx(vcpu); in vcpu_setup_sgx_lepubkeyhash() local
|
H A D | vmx.h | 679 vmx_has_waitpkg(struct vcpu_vmx * vmx) vmx_has_waitpkg() argument 719 vmx_segment_cache_clear(struct vcpu_vmx * vmx) vmx_segment_cache_clear() argument [all...] |
/linux/tools/testing/selftests/kvm/lib/x86/ |
H A D | vmx.c | 76 struct vmx_pages *vmx = addr_gva2hva(vm, vmx_gva); in vcpu_alloc_vmx() local 114 prepare_for_vmx_operation(struct vmx_pages * vmx) prepare_for_vmx_operation() argument 158 load_vmcs(struct vmx_pages * vmx) load_vmcs() argument 189 init_vmcs_control_fields(struct vmx_pages * vmx) init_vmcs_control_fields() argument 358 prepare_vmcs(struct vmx_pages * vmx,void * guest_rip,void * guest_rsp) prepare_vmcs() argument 397 __nested_pg_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr,int target_level) __nested_pg_map() argument 448 nested_pg_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr) nested_pg_map() argument 471 __nested_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr,uint64_t size,int level) __nested_map() argument 488 nested_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr,uint64_t size) nested_map() argument 497 nested_map_memslot(struct vmx_pages * vmx,struct kvm_vm * vm,uint32_t memslot) nested_map_memslot() argument 519 nested_identity_map_1g(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t addr,uint64_t size) nested_identity_map_1g() argument 537 prepare_eptp(struct vmx_pages * vmx,struct kvm_vm * vm,uint32_t eptp_memslot) prepare_eptp() argument 547 prepare_virtualize_apic_accesses(struct vmx_pages * vmx,struct kvm_vm * vm) prepare_virtualize_apic_accesses() argument [all...] |
H A D | memstress.c | 32 static void memstress_l1_guest_code(struct vmx_pages *vmx, uint64_t vcpu_id) in memstress_l1_guest_code() argument 62 void memstress_setup_ept(struct vmx_pages *vmx, struct kvm_vm *vm) in memstress_setup_ept() argument 82 struct vmx_pages *vmx, *vmx0 = NULL; memstress_setup_nested() local [all...] |
/linux/tools/testing/selftests/kvm/x86/ |
H A D | vmx_dirty_log_test.c | 56 void l1_guest_code(struct vmx_pages *vmx) in l1_guest_code() argument 83 struct vmx_pages *vmx; in test_vmx_dirty_log() local [all...] |
H A D | vmx_apic_access_test.c | 77 struct vmx_pages *vmx; in main() local
|
H A D | triple_fault_event_test.c | 27 void l1_guest_code_vmx(struct vmx_pages *vmx) in l1_guest_code_vmx() argument
|
H A D | kvm_buslock_test.c | 46 static void l1_vmx_code(struct vmx_pages *vmx) in l1_vmx_code() argument
|
H A D | aperfmperf_test.c | 74 static void l1_vmx_code(struct vmx_pages *vmx) in l1_vmx_code() argument
|
H A D | nested_exceptions_test.c | 129 static void l1_vmx_code(struct vmx_pages *vmx) in l1_vmx_code() argument
|
/linux/tools/testing/selftests/powerpc/ptrace/ |
H A D | ptrace-vsx.c | 40 unsigned long vmx[VMX_MAX + 2][2]; in trace_vsx() local
|
H A D | ptrace-tm-vsx.c | 87 unsigned long vmx[VMX_MAX + 2][2]; in trace_tm_vsx() local
|
H A D | ptrace-tm-spd-vsx.c | 99 unsigned long vmx[VMX_MAX + 2][2]; in trace_tm_spd_vsx() local
|
/linux/arch/x86/events/intel/ |
H A D | pt.h | 43 bool vmx; member
|
/linux/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 507 struct kvm_vmx_nested_state_hdr vmx; member
|
/linux/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 507 struct kvm_vmx_nested_state_hdr vmx; member
|