Lines Matching +full:1 +full:f

70 			/* ECX[1]: 64B alignment in compacted form */  in xstate_required_size()
78 xstate_bv >>= 1; in xstate_required_size()
85 #define F feature_bit macro
91 (boot_cpu_has(X86_FEATURE_##name) ? F(name) : 0); \
101 #define KVM_CPUID_INDEX_NOT_SIGNIFICANT -1ull
227 signature[1] = entry->ecx; in __kvm_get_hypervisor_cpuid()
299 best = cpuid_entry2_find(entries, nent, 1, KVM_CPUID_INDEX_NOT_SIGNIFICANT); in __kvm_update_cpuid_runtime()
319 best = cpuid_entry2_find(entries, nent, 0xD, 1); in __kvm_update_cpuid_runtime()
328 best->eax &= ~(1 << KVM_FEATURE_PV_UNHALT); in __kvm_update_cpuid_runtime()
390 * and can install smaller shadow pages if the host lacks 1GiB support. in kvm_vcpu_after_set_cpuid()
397 best = kvm_find_cpuid_entry(vcpu, 1); in kvm_vcpu_after_set_cpuid()
402 apic->lapic_timer.timer_mode_mask = 1 << 17; in kvm_vcpu_after_set_cpuid()
541 e2[i].padding[1] = 0; in kvm_vcpu_ioctl_set_cpuid()
631 unsigned int f_gbpages = F(GBPAGES); in kvm_set_cpu_caps()
632 unsigned int f_lm = F(LM); in kvm_set_cpu_caps()
633 unsigned int f_xfd = F(XFD); in kvm_set_cpu_caps()
652 F(XMM3) | F(PCLMULQDQ) | 0 /* DTES64, MONITOR */ | in kvm_set_cpu_caps()
654 0 /* TM2 */ | F(SSSE3) | 0 /* CNXT-ID */ | 0 /* Reserved */ | in kvm_set_cpu_caps()
655 F(FMA) | F(CX16) | 0 /* xTPR Update */ | F(PDCM) | in kvm_set_cpu_caps()
656 F(PCID) | 0 /* Reserved, DCA */ | F(XMM4_1) | in kvm_set_cpu_caps()
657 F(XMM4_2) | F(X2APIC) | F(MOVBE) | F(POPCNT) | in kvm_set_cpu_caps()
658 0 /* Reserved*/ | F(AES) | F(XSAVE) | 0 /* OSXSAVE */ | F(AVX) | in kvm_set_cpu_caps()
659 F(F16C) | F(RDRAND) in kvm_set_cpu_caps()
665 F(FPU) | F(VME) | F(DE) | F(PSE) | in kvm_set_cpu_caps()
666 F(TSC) | F(MSR) | F(PAE) | F(MCE) | in kvm_set_cpu_caps()
667 F(CX8) | F(APIC) | 0 /* Reserved */ | F(SEP) | in kvm_set_cpu_caps()
668 F(MTRR) | F(PGE) | F(MCA) | F(CMOV) | in kvm_set_cpu_caps()
669 F(PAT) | F(PSE36) | 0 /* PSN */ | F(CLFLUSH) | in kvm_set_cpu_caps()
670 0 /* Reserved, DS, ACPI */ | F(MMX) | in kvm_set_cpu_caps()
671 F(FXSR) | F(XMM) | F(XMM2) | F(SELFSNOOP) | in kvm_set_cpu_caps()
676 F(FSGSBASE) | F(SGX) | F(BMI1) | F(HLE) | F(AVX2) | in kvm_set_cpu_caps()
677 F(FDP_EXCPTN_ONLY) | F(SMEP) | F(BMI2) | F(ERMS) | F(INVPCID) | in kvm_set_cpu_caps()
678 F(RTM) | F(ZERO_FCS_FDS) | 0 /*MPX*/ | F(AVX512F) | in kvm_set_cpu_caps()
679 F(AVX512DQ) | F(RDSEED) | F(ADX) | F(SMAP) | F(AVX512IFMA) | in kvm_set_cpu_caps()
680 F(CLFLUSHOPT) | F(CLWB) | 0 /*INTEL_PT*/ | F(AVX512PF) | in kvm_set_cpu_caps()
681 F(AVX512ER) | F(AVX512CD) | F(SHA_NI) | F(AVX512BW) | in kvm_set_cpu_caps()
682 F(AVX512VL)); in kvm_set_cpu_caps()
685 F(AVX512VBMI) | F(LA57) | F(PKU) | 0 /*OSPKE*/ | F(RDPID) | in kvm_set_cpu_caps()
686 F(AVX512_VPOPCNTDQ) | F(UMIP) | F(AVX512_VBMI2) | F(GFNI) | in kvm_set_cpu_caps()
687 F(VAES) | F(VPCLMULQDQ) | F(AVX512_VNNI) | F(AVX512_BITALG) | in kvm_set_cpu_caps()
688 F(CLDEMOTE) | F(MOVDIRI) | F(MOVDIR64B) | 0 /*WAITPKG*/ | in kvm_set_cpu_caps()
689 F(SGX_LC) | F(BUS_LOCK_DETECT) in kvm_set_cpu_caps()
692 if (cpuid_ecx(7) & F(LA57)) in kvm_set_cpu_caps()
703 F(AVX512_4VNNIW) | F(AVX512_4FMAPS) | F(SPEC_CTRL) | in kvm_set_cpu_caps()
704 F(SPEC_CTRL_SSBD) | F(ARCH_CAPABILITIES) | F(INTEL_STIBP) | in kvm_set_cpu_caps()
705 F(MD_CLEAR) | F(AVX512_VP2INTERSECT) | F(FSRM) | in kvm_set_cpu_caps()
706 F(SERIALIZE) | F(TSXLDTRK) | F(AVX512_FP16) | in kvm_set_cpu_caps()
707 F(AMX_TILE) | F(AMX_INT8) | F(AMX_BF16) | F(FLUSH_L1D) in kvm_set_cpu_caps()
724 F(SHA512) | F(SM3) | F(SM4) | F(AVX_VNNI) | F(AVX512_BF16) | in kvm_set_cpu_caps()
725 F(CMPCCXADD) | F(FZRM) | F(FSRS) | F(FSRC) | F(AMX_FP16) | in kvm_set_cpu_caps()
726 F(AVX_IFMA) | F(LAM) in kvm_set_cpu_caps()
730 F(AVX_VNNI_INT8) | F(AVX_NE_CONVERT) | F(AMX_COMPLEX) | in kvm_set_cpu_caps()
731 F(AVX_VNNI_INT16) | F(PREFETCHITI) | F(AVX10) in kvm_set_cpu_caps()
735 F(INTEL_PSFD) | F(IPRED_CTRL) | F(RRSBA_CTRL) | F(DDPD_U) | in kvm_set_cpu_caps()
736 F(BHI_CTRL) | F(MCDT_NO) in kvm_set_cpu_caps()
740 F(XSAVEOPT) | F(XSAVEC) | F(XGETBV1) | F(XSAVES) | f_xfd in kvm_set_cpu_caps()
748 F(AVX10_128) | F(AVX10_256) | F(AVX10_512) in kvm_set_cpu_caps()
752 F(LAHF_LM) | F(CMP_LEGACY) | 0 /*SVM*/ | 0 /* ExtApicSpace */ | in kvm_set_cpu_caps()
753 F(CR8_LEGACY) | F(ABM) | F(SSE4A) | F(MISALIGNSSE) | in kvm_set_cpu_caps()
754 F(3DNOWPREFETCH) | F(OSVW) | 0 /* IBS */ | F(XOP) | in kvm_set_cpu_caps()
755 0 /* SKINIT, WDT, LWP */ | F(FMA4) | F(TBM) | in kvm_set_cpu_caps()
756 F(TOPOEXT) | 0 /* PERFCTR_CORE */ in kvm_set_cpu_caps()
760 F(FPU) | F(VME) | F(DE) | F(PSE) | in kvm_set_cpu_caps()
761 F(TSC) | F(MSR) | F(PAE) | F(MCE) | in kvm_set_cpu_caps()
762 F(CX8) | F(APIC) | 0 /* Reserved */ | F(SYSCALL) | in kvm_set_cpu_caps()
763 F(MTRR) | F(PGE) | F(MCA) | F(CMOV) | in kvm_set_cpu_caps()
764 F(PAT) | F(PSE36) | 0 /* Reserved */ | in kvm_set_cpu_caps()
765 F(NX) | 0 /* Reserved */ | F(MMXEXT) | F(MMX) | in kvm_set_cpu_caps()
766 F(FXSR) | F(FXSR_OPT) | f_gbpages | F(RDTSCP) | in kvm_set_cpu_caps()
767 0 /* Reserved */ | f_lm | F(3DNOWEXT) | F(3DNOW) in kvm_set_cpu_caps()
778 F(CLZERO) | F(XSAVEERPTR) | in kvm_set_cpu_caps()
779 F(WBNOINVD) | F(AMD_IBPB) | F(AMD_IBRS) | F(AMD_SSBD) | F(VIRT_SSBD) | in kvm_set_cpu_caps()
780 F(AMD_SSB_NO) | F(AMD_STIBP) | F(AMD_STIBP_ALWAYS_ON) | in kvm_set_cpu_caps()
781 F(AMD_PSFD) | F(AMD_IBPB_RET) in kvm_set_cpu_caps()
819 F(SME_COHERENT)); in kvm_set_cpu_caps()
822 F(NO_NESTED_DATA_BP) | F(LFENCE_RDTSC) | 0 /* SmmPgCfgLock */ | in kvm_set_cpu_caps()
823 F(NULL_SEL_CLR_BASE) | F(AUTOIBRS) | 0 /* PrefetchCtlMsr */ | in kvm_set_cpu_caps()
824 F(WRMSR_XX_BASE_NS) in kvm_set_cpu_caps()
832 F(PERFMON_V2) in kvm_set_cpu_caps()
852 F(XSTORE) | F(XSTORE_EN) | F(XCRYPT) | F(XCRYPT_EN) | in kvm_set_cpu_caps()
853 F(ACE2) | F(ACE2_EN) | F(PHE) | F(PHE_EN) | in kvm_set_cpu_caps()
854 F(PMM) | F(PMM_EN) in kvm_set_cpu_caps()
948 case 1: in __do_cpuid_func_emulated()
949 entry->ecx = F(MOVBE); in __do_cpuid_func_emulated()
956 entry->ecx = F(RDPID); in __do_cpuid_func_emulated()
985 case 1: in __do_cpuid_func()
1005 WARN_ON_ONCE((entry->eax & 0xff) > 1); in __do_cpuid_func()
1014 for (i = 1; entry->eax & 0x1f; ++i) { in __do_cpuid_func()
1034 if (max_idx >= 1) { in __do_cpuid_func()
1035 entry = do_host_cpuid(array, function, 1); in __do_cpuid_func()
1072 edx.split.anythread_deprecated = 1; in __do_cpuid_func()
1086 * of subleaf 1. in __do_cpuid_func()
1101 entry = do_host_cpuid(array, function, 1); in __do_cpuid_func()
1106 if (entry->eax & (F(XSAVES)|F(XSAVEC))) in __do_cpuid_func()
1164 entry = do_host_cpuid(array, function, 1); in __do_cpuid_func()
1169 * Index 1: SECS.ATTRIBUTES. ATTRIBUTES are restricted a la in __do_cpuid_func()
1185 for (i = 1, max_idx = entry->eax; i <= max_idx; ++i) { in __do_cpuid_func()
1197 for (i = 1, max_idx = entry->eax; i <= max_idx; ++i) { in __do_cpuid_func()
1218 * is guaranteed to be >=1 if AVX10 is supported. Note #2, the in __do_cpuid_func()
1221 avx10_version = min_t(u8, entry->ebx & 0xff, 1); in __do_cpuid_func()
1234 entry->ecx = sigptr[1]; in __do_cpuid_func()
1239 entry->eax = (1 << KVM_FEATURE_CLOCKSOURCE) | in __do_cpuid_func()
1240 (1 << KVM_FEATURE_NOP_IO_DELAY) | in __do_cpuid_func()
1241 (1 << KVM_FEATURE_CLOCKSOURCE2) | in __do_cpuid_func()
1242 (1 << KVM_FEATURE_ASYNC_PF) | in __do_cpuid_func()
1243 (1 << KVM_FEATURE_PV_EOI) | in __do_cpuid_func()
1244 (1 << KVM_FEATURE_CLOCKSOURCE_STABLE_BIT) | in __do_cpuid_func()
1245 (1 << KVM_FEATURE_PV_UNHALT) | in __do_cpuid_func()
1246 (1 << KVM_FEATURE_PV_TLB_FLUSH) | in __do_cpuid_func()
1247 (1 << KVM_FEATURE_ASYNC_PF_VMEXIT) | in __do_cpuid_func()
1248 (1 << KVM_FEATURE_PV_SEND_IPI) | in __do_cpuid_func()
1249 (1 << KVM_FEATURE_POLL_CONTROL) | in __do_cpuid_func()
1250 (1 << KVM_FEATURE_PV_SCHED_YIELD) | in __do_cpuid_func()
1251 (1 << KVM_FEATURE_ASYNC_PF_INT); in __do_cpuid_func()
1254 entry->eax |= (1 << KVM_FEATURE_STEAL_TIME); in __do_cpuid_func()
1351 entry->eax = 1; /* SVM revision 1 */ in __do_cpuid_func()
1463 limit = array->entries[array->nent - 1].eax; in get_cpuid_func()
1464 for (func = func + 1; func <= limit; ++func) { in get_cpuid_func()
1494 if (pad[0] || pad[1] || pad[2]) in sanity_check_entries()
1513 if (cpuid->nent < 1) in kvm_dev_ioctl_get_cpuid()
1615 * needs to be pulled from CPUID.0xb.1. in get_out_of_range_cpuid_entry()
1651 *ebx &= ~(F(RTM) | F(HLE)); in kvm_cpuid()
1659 * When leaf 0BH or 1FH is defined, CL is pass-through in kvm_cpuid()
1661 * subleaves. Index 1 will exist iff the leaf is in kvm_cpuid()
1662 * implemented, so we pass through CL iff leaf 1 in kvm_cpuid()
1666 entry = kvm_find_cpuid_entry_index(vcpu, function, 1); in kvm_cpuid()
1684 return 1; in kvm_emulate_cpuid()