Home
last modified time | relevance | path

Searched defs:vmx (Results 1 – 19 of 19) sorted by relevance

/linux/arch/x86/kvm/vmx/
H A Dnested.c192 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_fail() local
222 static void vmx_disable_shadow_vmcs(struct vcpu_vmx *vmx) in vmx_disable_shadow_vmcs() argument
233 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_evmcs() local
250 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_evmcs_handle_vmclear() local
274 vmx_sync_vmcs_host_state(struct vcpu_vmx * vmx,struct loaded_vmcs * prev) vmx_sync_vmcs_host_state() argument
295 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_switch_vmcs() local
320 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_put_vmcs12_pages() local
334 struct vcpu_vmx *vmx = to_vmx(vcpu); free_nested() local
413 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_ept_inject_page_fault() local
466 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_ept_new_eptp() local
624 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_prepare_msr_bitmap() local
734 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_cache_shadow_vmcs12() local
753 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_flush_cached_shadow_vmcs12() local
836 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_max_atomic_switch_msrs() local
1017 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_get_vmexit_msr_value() local
1115 struct vcpu_vmx *vmx = to_vmx(vcpu); prepare_vmx_msr_autostore_list() local
1210 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_transition_tlb_flush() local
1271 vmx_restore_vmx_basic(struct vcpu_vmx * vmx,u64 data) vmx_restore_vmx_basic() argument
1342 vmx_restore_control_msr(struct vcpu_vmx * vmx,u32 msr_index,u64 data) vmx_restore_control_msr() argument
1365 vmx_restore_vmx_misc(struct vcpu_vmx * vmx,u64 data) vmx_restore_vmx_misc() argument
1413 vmx_restore_vmx_ept_vpid_cap(struct vcpu_vmx * vmx,u64 data) vmx_restore_vmx_ept_vpid_cap() argument
1439 vmx_restore_fixed0_msr(struct vcpu_vmx * vmx,u32 msr_index,u64 data) vmx_restore_fixed0_msr() argument
1461 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_vmx_msr() local
1609 copy_shadow_to_vmcs12(struct vcpu_vmx * vmx) copy_shadow_to_vmcs12() argument
1636 copy_vmcs12_to_shadow(struct vcpu_vmx * vmx) copy_vmcs12_to_shadow() argument
1670 copy_enlightened_to_vmcs12(struct vcpu_vmx * vmx,u32 hv_clean_fields) copy_enlightened_to_vmcs12() argument
1918 copy_vmcs12_to_enlightened(struct vcpu_vmx * vmx) copy_vmcs12_to_enlightened() argument
2104 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_handle_enlightened_vmptrld() local
2192 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_sync_vmcs12_to_shadow() local
2204 struct vcpu_vmx *vmx = vmx_preemption_timer_fn() local
2216 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_calc_preemption_timer_value() local
2233 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_start_preemption_timer() local
2255 nested_vmx_calc_efer(struct vcpu_vmx * vmx,struct vmcs12 * vmcs12) nested_vmx_calc_efer() argument
2266 prepare_vmcs02_constant_state(struct vcpu_vmx * vmx) prepare_vmcs02_constant_state() argument
2330 prepare_vmcs02_early_rare(struct vcpu_vmx * vmx,struct vmcs12 * vmcs12) prepare_vmcs02_early_rare() argument
2356 prepare_vmcs02_early(struct vcpu_vmx * vmx,struct loaded_vmcs * vmcs01,struct vmcs12 * vmcs12) prepare_vmcs02_early() argument
2523 prepare_vmcs02_rare(struct vcpu_vmx * vmx,struct vmcs12 * vmcs12) prepare_vmcs02_rare() argument
2657 struct vcpu_vmx *vmx = to_vmx(vcpu); prepare_vmcs02() local
2809 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_check_eptp() local
2858 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_check_vm_execution_controls() local
2917 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_check_vm_exit_controls() local
2934 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_check_vm_entry_controls() local
3113 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_check_vmcs_link_ptr() local
3222 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_check_vmentry_hw() local
3301 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_get_evmcs_page() local
3331 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_get_vmcs12_pages() local
3446 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_write_pml_buffer() local
3518 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_enter_non_root_mode() local
3680 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_run() local
3937 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_complete_nested_posted_interrupt() local
4087 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_has_nested_events() local
4213 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_check_nested_events() local
4485 struct vcpu_vmx *vmx = to_vmx(vcpu); sync_vmcs02_to_vmcs12_rare() local
4530 struct vcpu_vmx *vmx = to_vmx(vcpu); copy_vmcs02_to_vmcs12_rare() local
4558 struct vcpu_vmx *vmx = to_vmx(vcpu); sync_vmcs02_to_vmcs12() local
4824 nested_vmx_get_vmcs01_guest_efer(struct vcpu_vmx * vmx) nested_vmx_get_vmcs01_guest_efer() argument
4850 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_restore_host_state() local
4964 struct vcpu_vmx *vmx = to_vmx(vcpu); __nested_vmx_vmexit() local
5308 struct vcpu_vmx *vmx = to_vmx(vcpu); alloc_shadow_vmcs() local
5331 struct vcpu_vmx *vmx = to_vmx(vcpu); enter_vmx_operation() local
5384 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmxon() local
5472 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_release_vmcs12() local
5514 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmclear() local
5572 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmread() local
5678 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmwrite() local
5779 set_current_vmptr(struct vcpu_vmx * vmx,gpa_t vmptr) set_current_vmptr() argument
5795 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmptrld() local
5893 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_invept() local
5973 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_invvpid() local
6105 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_vmfunc() local
6592 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_reflect_vmexit() local
6646 struct vcpu_vmx *vmx; vmx_get_nested_state() local
6778 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_nested_state() local
[all...]
H A Dvmx.c356 static __always_inline void vmx_disable_fb_clear(struct vcpu_vmx *vmx) in vmx_disable_fb_clear() argument
370 static __always_inline void vmx_enable_fb_clear(struct vcpu_vmx *vmx) in vmx_enable_fb_clear() argument
379 static void vmx_update_fb_clear_dis(struct kvm_vcpu *vcpu, struct vcpu_vmx *vmx) in vmx_update_fb_clear_dis() argument
652 vmx_find_uret_msr(struct vcpu_vmx * vmx,u32 msr) vmx_find_uret_msr() argument
662 vmx_set_guest_uret_msr(struct vcpu_vmx * vmx,struct vmx_uret_msr * msr,u64 data) vmx_set_guest_uret_msr() argument
765 vmx_segment_cache_test_set(struct vcpu_vmx * vmx,unsigned seg,unsigned field) vmx_segment_cache_test_set() argument
780 vmx_read_guest_seg_selector(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_selector() argument
789 vmx_read_guest_seg_base(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_base() argument
798 vmx_read_guest_seg_limit(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_limit() argument
807 vmx_read_guest_seg_ar(struct vcpu_vmx * vmx,unsigned seg) vmx_read_guest_seg_ar() argument
884 msr_write_intercepted(struct vcpu_vmx * vmx,u32 msr) msr_write_intercepted() argument
892 __vmx_vcpu_run_flags(struct vcpu_vmx * vmx) __vmx_vcpu_run_flags() argument
914 clear_atomic_switch_msr_special(struct vcpu_vmx * vmx,unsigned long entry,unsigned long exit) clear_atomic_switch_msr_special() argument
932 clear_atomic_switch_msr(struct vcpu_vmx * vmx,unsigned msr) clear_atomic_switch_msr() argument
972 add_atomic_switch_msr_special(struct vcpu_vmx * vmx,unsigned long entry,unsigned long exit,unsigned long guest_val_vmcs,unsigned long host_val_vmcs,u64 guest_val,u64 host_val) add_atomic_switch_msr_special() argument
984 add_atomic_switch_msr(struct vcpu_vmx * vmx,unsigned msr,u64 guest_val,u64 host_val,bool entry_only) add_atomic_switch_msr() argument
1050 update_transition_efer(struct vcpu_vmx * vmx) update_transition_efer() argument
1132 pt_can_write_msr(struct vcpu_vmx * vmx) pt_can_write_msr() argument
1172 pt_guest_enter(struct vcpu_vmx * vmx) pt_guest_enter() argument
1189 pt_guest_exit(struct vcpu_vmx * vmx) pt_guest_exit() argument
1236 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_prepare_switch_to_guest() local
1307 vmx_prepare_switch_to_host(struct vcpu_vmx * vmx) vmx_prepare_switch_to_host() argument
1347 vmx_read_guest_kernel_gs_base(struct vcpu_vmx * vmx) vmx_read_guest_kernel_gs_base() argument
1356 vmx_write_guest_kernel_gs_base(struct vcpu_vmx * vmx,u64 data) vmx_write_guest_kernel_gs_base() argument
1368 struct vcpu_vmx *vmx = to_vmx(vcpu); grow_ple_window() local
1384 struct vcpu_vmx *vmx = to_vmx(vcpu); shrink_ple_window() local
1400 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_load_vmcs() local
1482 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_rflags() local
1500 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_rflags() local
1563 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_rtit_ctl_check() local
1728 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_emulated_instruction() local
1779 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_inject_exception() local
1820 vmx_setup_uret_msr(struct vcpu_vmx * vmx,unsigned int msr,bool load_into_hardware) vmx_setup_uret_msr() argument
1838 vmx_setup_uret_msrs(struct vcpu_vmx * vmx) vmx_setup_uret_msrs() argument
1920 is_vmx_feature_control_msr_valid(struct vcpu_vmx * vmx,struct msr_data * msr) is_vmx_feature_control_msr_valid() argument
1963 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_msr() local
2160 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_msr() local
2997 struct vcpu_vmx *vmx = to_vmx(vcpu); enter_pmode() local
3067 struct vcpu_vmx *vmx = to_vmx(vcpu); enter_rmode() local
3114 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_efer() local
3163 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_flush_tlb_all() local
3274 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_cr0() local
3430 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_cr4() local
3493 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_segment() local
3540 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_get_cpl() local
3582 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_set_segment() local
3952 vmx_msr_bitmap_l01_changed(struct vcpu_vmx * vmx) vmx_msr_bitmap_l01_changed() argument
3972 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_intercept_for_msr() local
4004 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_msr_bitmap_x2apic() local
4057 struct vcpu_vmx *vmx = to_vmx(vcpu); pt_update_intercept_for_msr() local
4127 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_deliver_nested_posted_interrupt() local
4208 vmx_set_constant_host_state(struct vcpu_vmx * vmx) vmx_set_constant_host_state() argument
4275 set_cr4_guest_host_mask(struct vcpu_vmx * vmx) set_cr4_guest_host_mask() argument
4291 vmx_pin_based_exec_ctrl(struct vcpu_vmx * vmx) vmx_pin_based_exec_ctrl() argument
4345 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_refresh_apicv_exec_ctrl() local
4371 vmx_exec_control(struct vcpu_vmx * vmx) vmx_exec_control() argument
4415 vmx_tertiary_exec_control(struct vcpu_vmx * vmx) vmx_tertiary_exec_control() argument
4435 vmx_adjust_secondary_exec_control(struct vcpu_vmx * vmx,u32 * exec_control,u32 control,bool enabled,bool exiting) vmx_adjust_secondary_exec_control() argument
4474 vmx_adjust_sec_exec_control(vmx,exec_control,name,feat_name,ctrl_name,exiting) global() argument
4487 vmx_adjust_sec_exec_feature(vmx,exec_control,lname,uname) global() argument
4490 vmx_adjust_sec_exec_exiting(vmx,exec_control,lname,uname) global() argument
4493 vmx_secondary_exec_control(struct vcpu_vmx * vmx) vmx_secondary_exec_control() argument
4613 init_vmcs(struct vcpu_vmx * vmx) init_vmcs() argument
4738 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_vcpu_reset() local
4770 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_reset() local
4845 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_inject_irq() local
4873 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_inject_nmi() local
4904 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_nmi_mask() local
4918 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_nmi_mask() local
5107 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_exception_nmi() local
5619 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_task_switch() local
5747 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_unhandleable_emulation_required() local
5774 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_invalid_guest_state() local
5906 struct vcpu_vmx *vmx = to_vmx(vcpu); handle_fastpath_preemption_timer() local
6072 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_get_exit_info() local
6099 vmx_destroy_pml_buffer(struct vcpu_vmx * vmx) vmx_destroy_pml_buffer() argument
6109 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_flush_pml_buffer() local
6174 struct vcpu_vmx *vmx = to_vmx(vcpu); dump_vmcs() local
6371 struct vcpu_vmx *vmx = to_vmx(vcpu); __vmx_handle_exit() local
6637 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_set_virtual_apic_mode() local
6963 vmx_recover_nmi_blocking(struct vcpu_vmx * vmx) vmx_recover_nmi_blocking() argument
7061 vmx_complete_interrupts(struct vcpu_vmx * vmx) vmx_complete_interrupts() argument
7078 atomic_switch_perf_msrs(struct vcpu_vmx * vmx) atomic_switch_perf_msrs() argument
7103 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_hv_timer() local
7127 vmx_update_host_rsp(struct vcpu_vmx * vmx,unsigned long host_rsp) vmx_update_host_rsp() argument
7135 vmx_spec_ctrl_restore_host(struct vcpu_vmx * vmx,unsigned int flags) vmx_spec_ctrl_restore_host() argument
7200 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_enter_exit() local
7253 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_run() local
7410 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_free() local
7423 struct vcpu_vmx *vmx; vmx_vcpu_create() local
7585 vmcs_set_secondary_exec_control(struct vcpu_vmx * vmx,u32 new_ctl) vmcs_set_secondary_exec_control() argument
7610 struct vcpu_vmx *vmx = to_vmx(vcpu); nested_vmx_cr_fixed1_bits_update() local
7653 struct vcpu_vmx *vmx = to_vmx(vcpu); update_intel_pt_cfg() local
7722 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_vcpu_after_set_cpuid() local
8053 struct vcpu_vmx *vmx; vmx_set_hv_timer() local
8098 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_update_cpu_dirty_logging() local
8140 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_enter_smm() local
8161 struct vcpu_vmx *vmx = to_vmx(vcpu); vmx_leave_smm() local
[all...]
H A Dnested.h81 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_has_valid_vmcs12() local
90 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vpid02() local
H A Dsgx.c300 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_encls_einit() local
430 struct vcpu_vmx *vmx = to_vmx(vcpu); in vcpu_setup_sgx_lepubkeyhash() local
H A Dvmx.h679 vmx_has_waitpkg(struct vcpu_vmx * vmx) vmx_has_waitpkg() argument
719 vmx_segment_cache_clear(struct vcpu_vmx * vmx) vmx_segment_cache_clear() argument
[all...]
/linux/tools/testing/selftests/kvm/lib/x86/
H A Dvmx.c76 struct vmx_pages *vmx = addr_gva2hva(vm, vmx_gva); in vcpu_alloc_vmx() local
114 prepare_for_vmx_operation(struct vmx_pages * vmx) prepare_for_vmx_operation() argument
158 load_vmcs(struct vmx_pages * vmx) load_vmcs() argument
189 init_vmcs_control_fields(struct vmx_pages * vmx) init_vmcs_control_fields() argument
358 prepare_vmcs(struct vmx_pages * vmx,void * guest_rip,void * guest_rsp) prepare_vmcs() argument
397 __nested_pg_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr,int target_level) __nested_pg_map() argument
448 nested_pg_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr) nested_pg_map() argument
471 __nested_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr,uint64_t size,int level) __nested_map() argument
488 nested_map(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t nested_paddr,uint64_t paddr,uint64_t size) nested_map() argument
497 nested_map_memslot(struct vmx_pages * vmx,struct kvm_vm * vm,uint32_t memslot) nested_map_memslot() argument
519 nested_identity_map_1g(struct vmx_pages * vmx,struct kvm_vm * vm,uint64_t addr,uint64_t size) nested_identity_map_1g() argument
537 prepare_eptp(struct vmx_pages * vmx,struct kvm_vm * vm,uint32_t eptp_memslot) prepare_eptp() argument
547 prepare_virtualize_apic_accesses(struct vmx_pages * vmx,struct kvm_vm * vm) prepare_virtualize_apic_accesses() argument
[all...]
H A Dmemstress.c32 static void memstress_l1_guest_code(struct vmx_pages *vmx, uint64_t vcpu_id) in memstress_l1_guest_code() argument
62 void memstress_setup_ept(struct vmx_pages *vmx, struct kvm_vm *vm) in memstress_setup_ept() argument
82 struct vmx_pages *vmx, *vmx0 = NULL; memstress_setup_nested() local
[all...]
/linux/tools/testing/selftests/kvm/x86/
H A Dvmx_dirty_log_test.c56 void l1_guest_code(struct vmx_pages *vmx) in l1_guest_code() argument
83 struct vmx_pages *vmx; in test_vmx_dirty_log() local
[all...]
H A Dvmx_apic_access_test.c77 struct vmx_pages *vmx; in main() local
H A Dtriple_fault_event_test.c27 void l1_guest_code_vmx(struct vmx_pages *vmx) in l1_guest_code_vmx() argument
H A Dkvm_buslock_test.c46 static void l1_vmx_code(struct vmx_pages *vmx) in l1_vmx_code() argument
H A Daperfmperf_test.c74 static void l1_vmx_code(struct vmx_pages *vmx) in l1_vmx_code() argument
H A Dnested_exceptions_test.c129 static void l1_vmx_code(struct vmx_pages *vmx) in l1_vmx_code() argument
/linux/tools/testing/selftests/powerpc/ptrace/
H A Dptrace-vsx.c40 unsigned long vmx[VMX_MAX + 2][2]; in trace_vsx() local
H A Dptrace-tm-vsx.c87 unsigned long vmx[VMX_MAX + 2][2]; in trace_tm_vsx() local
H A Dptrace-tm-spd-vsx.c99 unsigned long vmx[VMX_MAX + 2][2]; in trace_tm_spd_vsx() local
/linux/arch/x86/events/intel/
H A Dpt.h43 bool vmx; member
/linux/tools/arch/x86/include/uapi/asm/
H A Dkvm.h507 struct kvm_vmx_nested_state_hdr vmx; member
/linux/arch/x86/include/uapi/asm/
H A Dkvm.h507 struct kvm_vmx_nested_state_hdr vmx; member