| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | state_test.c | 142 if (this_cpu_has(X86_FEATURE_XSAVE)) { in guest_code() 170 if (this_cpu_has(X86_FEATURE_MPX)) { in guest_code() 204 if (this_cpu_has(X86_FEATURE_PKU)) { in guest_code() 207 GUEST_ASSERT(this_cpu_has(X86_FEATURE_OSPKE)); in guest_code() 216 if (this_cpu_has(X86_FEATURE_SVM)) in guest_code()
|
| H A D | pmu_counters_test.c | 181 if (!this_cpu_has(X86_FEATURE_CLFLUSHOPT) && in guest_assert_event_count() 182 !this_cpu_has(X86_FEATURE_CLFLUSH)) in guest_assert_event_count() 248 if (this_cpu_has(X86_FEATURE_CLFLUSHOPT)) \ 250 else if (this_cpu_has(X86_FEATURE_CLFLUSH)) \ 280 if (this_cpu_has(X86_FEATURE_PDCM) && in guest_test_arch_event() 488 if (this_cpu_has(X86_FEATURE_PDCM) && in guest_test_gp_counters()
|
| H A D | monitor_mwait_test.c | 51 __GUEST_ASSERT(this_cpu_has(X86_FEATURE_MWAIT) == !(testcase & MWAIT_DISABLED), in guest_monitor_wait() 77 TEST_REQUIRE(this_cpu_has(X86_FEATURE_MWAIT)); in main()
|
| H A D | aperfmperf_test.c | 98 if (this_cpu_has(X86_FEATURE_SVM)) in guest_code() 100 else if (this_cpu_has(X86_FEATURE_VMX)) in guest_code()
|
| H A D | msrs_test.c | 147 if (this_cpu_has(msr->feature2)) { in guest_test_unsupported_msr() 178 if (!this_cpu_has(msr->feature) || in guest_test_reserved_val() 196 if (this_cpu_has(msr->feature)) in guest_main()
|
| H A D | nested_close_kvm_test.c | 62 if (this_cpu_has(X86_FEATURE_VMX)) in l1_guest_code()
|
| H A D | amx_test.c | 153 GUEST_ASSERT(this_cpu_has(X86_FEATURE_XSAVE) && in guest_code() 154 this_cpu_has(X86_FEATURE_OSXSAVE)); in guest_code()
|
| H A D | hyperv_features.c | 74 if (!this_cpu_has(HV_ACCESS_TSC_INVARIANT)) in guest_msr() 75 GUEST_ASSERT(this_cpu_has(X86_FEATURE_INVTSC)); in guest_msr() 77 GUEST_ASSERT(this_cpu_has(X86_FEATURE_INVTSC) == in guest_msr()
|
| H A D | nested_invalid_cr3_test.c | 71 if (this_cpu_has(X86_FEATURE_VMX)) in l1_guest_code()
|
| H A D | nested_tsc_adjust_test.c | 89 if (this_cpu_has(X86_FEATURE_VMX)) { in l1_guest_code()
|
| H A D | nested_tsc_scaling_test.c | 145 if (this_cpu_has(X86_FEATURE_VMX)) in l1_guest_code()
|
| H A D | sev_smoke_test.c | 47 GUEST_ASSERT(this_cpu_has(X86_FEATURE_SEV)); in guest_sev_code()
|
| H A D | nested_exceptions_test.c | 161 if (this_cpu_has(X86_FEATURE_SVM)) in l1_guest_code()
|
| H A D | xapic_ipi_test.c | 474 if (this_cpu_has(X86_FEATURE_IDLE_HLT)) in main()
|
| /linux/arch/um/include/asm/ |
| H A D | cpufeature.h | 39 #define this_cpu_has(bit) \ macro
|
| /linux/arch/x86/include/asm/ |
| H A D | cpufeature.h | 58 #define this_cpu_has(bit) \ macro
|
| /linux/arch/x86/kernel/apic/ |
| H A D | apic.c | 578 if (this_cpu_has(X86_FEATURE_ARAT)) { in setup_APIC_timer() 587 if (this_cpu_has(X86_FEATURE_TSC_DEADLINE_TIMER)) { in setup_APIC_timer() 609 if (!this_cpu_has(X86_FEATURE_TSC_DEADLINE_TIMER)) in __lapic_update_tsc_freq()
|
| /linux/tools/testing/selftests/kvm/include/x86/ |
| H A D | processor.h | 716 static inline bool this_cpu_has(struct kvm_x86_cpu_feature feature) in this_cpu_has() function 754 return nr_bits > feature.f.bit && !this_cpu_has(feature.f); in this_pmu_has() 759 return nr_bits > feature.f.bit || this_cpu_has(feature.f); in this_pmu_has()
|
| /linux/tools/testing/selftests/kvm/ |
| H A D | access_tracking_perf_test.c | 469 if (this_cpu_has(X86_FEATURE_HYPERVISOR)) { in access_tracking_unreliable()
|
| /linux/drivers/acpi/ |
| H A D | processor_throttling.c | 708 !this_cpu_has(X86_FEATURE_ACPI)) { in acpi_throttling_rdmsr() 728 !this_cpu_has(X86_FEATURE_ACPI)) { in acpi_throttling_wrmsr()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | vmx.c | 2889 if (!this_cpu_has(X86_FEATURE_MSR_IA32_FEAT_CTL) || in __kvm_is_vmx_supported() 2890 !this_cpu_has(X86_FEATURE_VMX)) { in __kvm_is_vmx_supported()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | sev.c | 745 if (this_cpu_has(X86_FEATURE_SME_COHERENT) || npages == 0 || in sev_clflush_pages()
|
| /linux/arch/x86/events/intel/ |
| H A D | core.c | 5763 if (this_cpu_has(X86_FEATURE_ARCH_PERFMON_EXT)) in init_hybrid_pmu()
|