Lines Matching +full:1 +full:f
49 /* ECX[1]: 64B alignment in compacted form */ in xstate_required_size()
57 xstate_bv >>= 1; in xstate_required_size()
64 #define F feature_bit macro
70 (boot_cpu_has(X86_FEATURE_##name) ? F(name) : 0); \
80 #define KVM_CPUID_INDEX_NOT_SIGNIFICANT -1ull
206 signature[1] = entry->ecx; in kvm_get_hypervisor_cpuid()
270 best = cpuid_entry2_find(entries, nent, 1, KVM_CPUID_INDEX_NOT_SIGNIFICANT); in __kvm_update_cpuid_runtime()
290 best = cpuid_entry2_find(entries, nent, 0xD, 1); in __kvm_update_cpuid_runtime()
297 (best->eax & (1 << KVM_FEATURE_PV_UNHALT))) in __kvm_update_cpuid_runtime()
298 best->eax &= ~(1 << KVM_FEATURE_PV_UNHALT); in __kvm_update_cpuid_runtime()
347 * and can install smaller shadow pages if the host lacks 1GiB support. in kvm_vcpu_after_set_cpuid()
354 best = kvm_find_cpuid_entry(vcpu, 1); in kvm_vcpu_after_set_cpuid()
359 apic->lapic_timer.timer_mode_mask = 1 << 17; in kvm_vcpu_after_set_cpuid()
497 e2[i].padding[1] = 0; in kvm_vcpu_ioctl_set_cpuid()
587 unsigned int f_gbpages = F(GBPAGES); in kvm_set_cpu_caps()
588 unsigned int f_lm = F(LM); in kvm_set_cpu_caps()
589 unsigned int f_xfd = F(XFD); in kvm_set_cpu_caps()
608 F(XMM3) | F(PCLMULQDQ) | 0 /* DTES64, MONITOR */ | in kvm_set_cpu_caps()
610 0 /* TM2 */ | F(SSSE3) | 0 /* CNXT-ID */ | 0 /* Reserved */ | in kvm_set_cpu_caps()
611 F(FMA) | F(CX16) | 0 /* xTPR Update */ | F(PDCM) | in kvm_set_cpu_caps()
612 F(PCID) | 0 /* Reserved, DCA */ | F(XMM4_1) | in kvm_set_cpu_caps()
613 F(XMM4_2) | F(X2APIC) | F(MOVBE) | F(POPCNT) | in kvm_set_cpu_caps()
614 0 /* Reserved*/ | F(AES) | F(XSAVE) | 0 /* OSXSAVE */ | F(AVX) | in kvm_set_cpu_caps()
615 F(F16C) | F(RDRAND) in kvm_set_cpu_caps()
621 F(FPU) | F(VME) | F(DE) | F(PSE) | in kvm_set_cpu_caps()
622 F(TSC) | F(MSR) | F(PAE) | F(MCE) | in kvm_set_cpu_caps()
623 F(CX8) | F(APIC) | 0 /* Reserved */ | F(SEP) | in kvm_set_cpu_caps()
624 F(MTRR) | F(PGE) | F(MCA) | F(CMOV) | in kvm_set_cpu_caps()
625 F(PAT) | F(PSE36) | 0 /* PSN */ | F(CLFLUSH) | in kvm_set_cpu_caps()
626 0 /* Reserved, DS, ACPI */ | F(MMX) | in kvm_set_cpu_caps()
627 F(FXSR) | F(XMM) | F(XMM2) | F(SELFSNOOP) | in kvm_set_cpu_caps()
632 F(FSGSBASE) | F(SGX) | F(BMI1) | F(HLE) | F(AVX2) | in kvm_set_cpu_caps()
633 F(FDP_EXCPTN_ONLY) | F(SMEP) | F(BMI2) | F(ERMS) | F(INVPCID) | in kvm_set_cpu_caps()
634 F(RTM) | F(ZERO_FCS_FDS) | 0 /*MPX*/ | F(AVX512F) | in kvm_set_cpu_caps()
635 F(AVX512DQ) | F(RDSEED) | F(ADX) | F(SMAP) | F(AVX512IFMA) | in kvm_set_cpu_caps()
636 F(CLFLUSHOPT) | F(CLWB) | 0 /*INTEL_PT*/ | F(AVX512PF) | in kvm_set_cpu_caps()
637 F(AVX512ER) | F(AVX512CD) | F(SHA_NI) | F(AVX512BW) | in kvm_set_cpu_caps()
638 F(AVX512VL)); in kvm_set_cpu_caps()
641 F(AVX512VBMI) | F(LA57) | F(PKU) | 0 /*OSPKE*/ | F(RDPID) | in kvm_set_cpu_caps()
642 F(AVX512_VPOPCNTDQ) | F(UMIP) | F(AVX512_VBMI2) | F(GFNI) | in kvm_set_cpu_caps()
643 F(VAES) | F(VPCLMULQDQ) | F(AVX512_VNNI) | F(AVX512_BITALG) | in kvm_set_cpu_caps()
644 F(CLDEMOTE) | F(MOVDIRI) | F(MOVDIR64B) | 0 /*WAITPKG*/ | in kvm_set_cpu_caps()
645 F(SGX_LC) | F(BUS_LOCK_DETECT) in kvm_set_cpu_caps()
648 if (cpuid_ecx(7) & F(LA57)) in kvm_set_cpu_caps()
659 F(AVX512_4VNNIW) | F(AVX512_4FMAPS) | F(SPEC_CTRL) | in kvm_set_cpu_caps()
660 F(SPEC_CTRL_SSBD) | F(ARCH_CAPABILITIES) | F(INTEL_STIBP) | in kvm_set_cpu_caps()
661 F(MD_CLEAR) | F(AVX512_VP2INTERSECT) | F(FSRM) | in kvm_set_cpu_caps()
662 F(SERIALIZE) | F(TSXLDTRK) | F(AVX512_FP16) | in kvm_set_cpu_caps()
663 F(AMX_TILE) | F(AMX_INT8) | F(AMX_BF16) | F(FLUSH_L1D) in kvm_set_cpu_caps()
678 F(AVX_VNNI) | F(AVX512_BF16) | F(CMPCCXADD) | in kvm_set_cpu_caps()
679 F(FZRM) | F(FSRS) | F(FSRC) | in kvm_set_cpu_caps()
680 F(AMX_FP16) | F(AVX_IFMA) | F(LAM) in kvm_set_cpu_caps()
684 F(AVX_VNNI_INT8) | F(AVX_NE_CONVERT) | F(PREFETCHITI) | in kvm_set_cpu_caps()
685 F(AMX_COMPLEX) in kvm_set_cpu_caps()
689 F(INTEL_PSFD) | F(IPRED_CTRL) | F(RRSBA_CTRL) | F(DDPD_U) | in kvm_set_cpu_caps()
690 F(BHI_CTRL) | F(MCDT_NO) in kvm_set_cpu_caps()
694 F(XSAVEOPT) | F(XSAVEC) | F(XGETBV1) | F(XSAVES) | f_xfd in kvm_set_cpu_caps()
702 F(LAHF_LM) | F(CMP_LEGACY) | 0 /*SVM*/ | 0 /* ExtApicSpace */ | in kvm_set_cpu_caps()
703 F(CR8_LEGACY) | F(ABM) | F(SSE4A) | F(MISALIGNSSE) | in kvm_set_cpu_caps()
704 F(3DNOWPREFETCH) | F(OSVW) | 0 /* IBS */ | F(XOP) | in kvm_set_cpu_caps()
705 0 /* SKINIT, WDT, LWP */ | F(FMA4) | F(TBM) | in kvm_set_cpu_caps()
706 F(TOPOEXT) | 0 /* PERFCTR_CORE */ in kvm_set_cpu_caps()
710 F(FPU) | F(VME) | F(DE) | F(PSE) | in kvm_set_cpu_caps()
711 F(TSC) | F(MSR) | F(PAE) | F(MCE) | in kvm_set_cpu_caps()
712 F(CX8) | F(APIC) | 0 /* Reserved */ | F(SYSCALL) | in kvm_set_cpu_caps()
713 F(MTRR) | F(PGE) | F(MCA) | F(CMOV) | in kvm_set_cpu_caps()
714 F(PAT) | F(PSE36) | 0 /* Reserved */ | in kvm_set_cpu_caps()
715 F(NX) | 0 /* Reserved */ | F(MMXEXT) | F(MMX) | in kvm_set_cpu_caps()
716 F(FXSR) | F(FXSR_OPT) | f_gbpages | F(RDTSCP) | in kvm_set_cpu_caps()
717 0 /* Reserved */ | f_lm | F(3DNOWEXT) | F(3DNOW) in kvm_set_cpu_caps()
728 F(CLZERO) | F(XSAVEERPTR) | in kvm_set_cpu_caps()
729 F(WBNOINVD) | F(AMD_IBPB) | F(AMD_IBRS) | F(AMD_SSBD) | F(VIRT_SSBD) | in kvm_set_cpu_caps()
730 F(AMD_SSB_NO) | F(AMD_STIBP) | F(AMD_STIBP_ALWAYS_ON) | in kvm_set_cpu_caps()
731 F(AMD_PSFD) in kvm_set_cpu_caps()
764 0 /* SME */ | F(SEV) | 0 /* VM_PAGE_FLUSH */ | F(SEV_ES) | in kvm_set_cpu_caps()
765 F(SME_COHERENT)); in kvm_set_cpu_caps()
768 F(NO_NESTED_DATA_BP) | F(LFENCE_RDTSC) | 0 /* SmmPgCfgLock */ | in kvm_set_cpu_caps()
769 F(NULL_SEL_CLR_BASE) | F(AUTOIBRS) | 0 /* PrefetchCtlMsr */ | in kvm_set_cpu_caps()
770 F(WRMSR_XX_BASE_NS) in kvm_set_cpu_caps()
778 F(PERFMON_V2) in kvm_set_cpu_caps()
798 F(XSTORE) | F(XSTORE_EN) | F(XCRYPT) | F(XCRYPT_EN) | in kvm_set_cpu_caps()
799 F(ACE2) | F(ACE2_EN) | F(PHE) | F(PHE_EN) | in kvm_set_cpu_caps()
800 F(PMM) | F(PMM_EN) in kvm_set_cpu_caps()
894 case 1: in __do_cpuid_func_emulated()
895 entry->ecx = F(MOVBE); in __do_cpuid_func_emulated()
902 entry->ecx = F(RDPID); in __do_cpuid_func_emulated()
931 case 1: in __do_cpuid_func()
951 WARN_ON_ONCE((entry->eax & 0xff) > 1); in __do_cpuid_func()
960 for (i = 1; entry->eax & 0x1f; ++i) { in __do_cpuid_func()
980 if (max_idx >= 1) { in __do_cpuid_func()
981 entry = do_host_cpuid(array, function, 1); in __do_cpuid_func()
1018 edx.split.anythread_deprecated = 1; in __do_cpuid_func()
1032 * of subleaf 1. in __do_cpuid_func()
1047 entry = do_host_cpuid(array, function, 1); in __do_cpuid_func()
1052 if (entry->eax & (F(XSAVES)|F(XSAVEC))) in __do_cpuid_func()
1110 entry = do_host_cpuid(array, function, 1); in __do_cpuid_func()
1115 * Index 1: SECS.ATTRIBUTES. ATTRIBUTES are restricted a la in __do_cpuid_func()
1131 for (i = 1, max_idx = entry->eax; i <= max_idx; ++i) { in __do_cpuid_func()
1143 for (i = 1, max_idx = entry->eax; i <= max_idx; ++i) { in __do_cpuid_func()
1158 entry->ecx = sigptr[1]; in __do_cpuid_func()
1163 entry->eax = (1 << KVM_FEATURE_CLOCKSOURCE) | in __do_cpuid_func()
1164 (1 << KVM_FEATURE_NOP_IO_DELAY) | in __do_cpuid_func()
1165 (1 << KVM_FEATURE_CLOCKSOURCE2) | in __do_cpuid_func()
1166 (1 << KVM_FEATURE_ASYNC_PF) | in __do_cpuid_func()
1167 (1 << KVM_FEATURE_PV_EOI) | in __do_cpuid_func()
1168 (1 << KVM_FEATURE_CLOCKSOURCE_STABLE_BIT) | in __do_cpuid_func()
1169 (1 << KVM_FEATURE_PV_UNHALT) | in __do_cpuid_func()
1170 (1 << KVM_FEATURE_PV_TLB_FLUSH) | in __do_cpuid_func()
1171 (1 << KVM_FEATURE_ASYNC_PF_VMEXIT) | in __do_cpuid_func()
1172 (1 << KVM_FEATURE_PV_SEND_IPI) | in __do_cpuid_func()
1173 (1 << KVM_FEATURE_POLL_CONTROL) | in __do_cpuid_func()
1174 (1 << KVM_FEATURE_PV_SCHED_YIELD) | in __do_cpuid_func()
1175 (1 << KVM_FEATURE_ASYNC_PF_INT); in __do_cpuid_func()
1178 entry->eax |= (1 << KVM_FEATURE_STEAL_TIME); in __do_cpuid_func()
1254 entry->eax = 1; /* SVM revision 1 */ in __do_cpuid_func()
1366 limit = array->entries[array->nent - 1].eax; in get_cpuid_func()
1367 for (func = func + 1; func <= limit; ++func) { in get_cpuid_func()
1397 if (pad[0] || pad[1] || pad[2]) in sanity_check_entries()
1416 if (cpuid->nent < 1) in kvm_dev_ioctl_get_cpuid()
1518 * needs to be pulled from CPUID.0xb.1. in get_out_of_range_cpuid_entry()
1554 *ebx &= ~(F(RTM) | F(HLE)); in kvm_cpuid()
1562 * When leaf 0BH or 1FH is defined, CL is pass-through in kvm_cpuid()
1564 * subleaves. Index 1 will exist iff the leaf is in kvm_cpuid()
1565 * implemented, so we pass through CL iff leaf 1 in kvm_cpuid()
1569 entry = kvm_find_cpuid_entry_index(vcpu, function, 1); in kvm_cpuid()
1587 return 1; in kvm_emulate_cpuid()