xref: /linux/tools/testing/selftests/kvm/include/x86/pmu.h (revision 2922b595886575ebef46e65a0d4c8a857c5cd464)
1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3  * Copyright (C) 2023, Tencent, Inc.
4  */
5 #ifndef SELFTEST_KVM_PMU_H
6 #define SELFTEST_KVM_PMU_H
7 
8 #include <stdint.h>
9 
10 #define KVM_PMU_EVENT_FILTER_MAX_EVENTS			300
11 
12 /*
13  * Encode an eventsel+umask pair into event-select MSR format.  Note, this is
14  * technically AMD's format, as Intel's format only supports 8 bits for the
15  * event selector, i.e. doesn't use bits 24:16 for the selector.  But, OR-ing
16  * in '0' is a nop and won't clobber the CMASK.
17  */
18 #define RAW_EVENT(eventsel, umask) (((eventsel & 0xf00UL) << 24) |	\
19 				    ((eventsel) & 0xff) |		\
20 				    ((umask) & 0xff) << 8)
21 
22 /*
23  * These are technically Intel's definitions, but except for CMASK (see above),
24  * AMD's layout is compatible with Intel's.
25  */
26 #define ARCH_PERFMON_EVENTSEL_EVENT		GENMASK_ULL(7, 0)
27 #define ARCH_PERFMON_EVENTSEL_UMASK		GENMASK_ULL(15, 8)
28 #define ARCH_PERFMON_EVENTSEL_USR		BIT_ULL(16)
29 #define ARCH_PERFMON_EVENTSEL_OS		BIT_ULL(17)
30 #define ARCH_PERFMON_EVENTSEL_EDGE		BIT_ULL(18)
31 #define ARCH_PERFMON_EVENTSEL_PIN_CONTROL	BIT_ULL(19)
32 #define ARCH_PERFMON_EVENTSEL_INT		BIT_ULL(20)
33 #define ARCH_PERFMON_EVENTSEL_ANY		BIT_ULL(21)
34 #define ARCH_PERFMON_EVENTSEL_ENABLE		BIT_ULL(22)
35 #define ARCH_PERFMON_EVENTSEL_INV		BIT_ULL(23)
36 #define ARCH_PERFMON_EVENTSEL_CMASK		GENMASK_ULL(31, 24)
37 
38 /* RDPMC control flags, Intel only. */
39 #define INTEL_RDPMC_METRICS			BIT_ULL(29)
40 #define INTEL_RDPMC_FIXED			BIT_ULL(30)
41 #define INTEL_RDPMC_FAST			BIT_ULL(31)
42 
43 /* Fixed PMC controls, Intel only. */
44 #define FIXED_PMC_GLOBAL_CTRL_ENABLE(_idx)	BIT_ULL((32 + (_idx)))
45 
46 #define FIXED_PMC_KERNEL			BIT_ULL(0)
47 #define FIXED_PMC_USER				BIT_ULL(1)
48 #define FIXED_PMC_ANYTHREAD			BIT_ULL(2)
49 #define FIXED_PMC_ENABLE_PMI			BIT_ULL(3)
50 #define FIXED_PMC_NR_BITS			4
51 #define FIXED_PMC_CTRL(_idx, _val)		((_val) << ((_idx) * FIXED_PMC_NR_BITS))
52 
53 #define PMU_CAP_FW_WRITES			BIT_ULL(13)
54 #define PMU_CAP_LBR_FMT				0x3f
55 
56 #define	INTEL_ARCH_CPU_CYCLES			RAW_EVENT(0x3c, 0x00)
57 #define	INTEL_ARCH_INSTRUCTIONS_RETIRED		RAW_EVENT(0xc0, 0x00)
58 #define	INTEL_ARCH_REFERENCE_CYCLES		RAW_EVENT(0x3c, 0x01)
59 #define	INTEL_ARCH_LLC_REFERENCES		RAW_EVENT(0x2e, 0x4f)
60 #define	INTEL_ARCH_LLC_MISSES			RAW_EVENT(0x2e, 0x41)
61 #define	INTEL_ARCH_BRANCHES_RETIRED		RAW_EVENT(0xc4, 0x00)
62 #define	INTEL_ARCH_BRANCHES_MISPREDICTED	RAW_EVENT(0xc5, 0x00)
63 #define	INTEL_ARCH_TOPDOWN_SLOTS		RAW_EVENT(0xa4, 0x01)
64 #define	INTEL_ARCH_TOPDOWN_BE_BOUND		RAW_EVENT(0xa4, 0x02)
65 #define	INTEL_ARCH_TOPDOWN_BAD_SPEC		RAW_EVENT(0x73, 0x00)
66 #define	INTEL_ARCH_TOPDOWN_FE_BOUND		RAW_EVENT(0x9c, 0x01)
67 #define	INTEL_ARCH_TOPDOWN_RETIRING		RAW_EVENT(0xc2, 0x02)
68 #define	INTEL_ARCH_LBR_INSERTS			RAW_EVENT(0xe4, 0x01)
69 
70 #define	AMD_ZEN_CORE_CYCLES			RAW_EVENT(0x76, 0x00)
71 #define	AMD_ZEN_INSTRUCTIONS_RETIRED		RAW_EVENT(0xc0, 0x00)
72 #define	AMD_ZEN_BRANCHES_RETIRED		RAW_EVENT(0xc2, 0x00)
73 #define	AMD_ZEN_BRANCHES_MISPREDICTED		RAW_EVENT(0xc3, 0x00)
74 
75 /*
76  * Note!  The order and thus the index of the architectural events matters as
77  * support for each event is enumerated via CPUID using the index of the event.
78  */
79 enum intel_pmu_architectural_events {
80 	INTEL_ARCH_CPU_CYCLES_INDEX,
81 	INTEL_ARCH_INSTRUCTIONS_RETIRED_INDEX,
82 	INTEL_ARCH_REFERENCE_CYCLES_INDEX,
83 	INTEL_ARCH_LLC_REFERENCES_INDEX,
84 	INTEL_ARCH_LLC_MISSES_INDEX,
85 	INTEL_ARCH_BRANCHES_RETIRED_INDEX,
86 	INTEL_ARCH_BRANCHES_MISPREDICTED_INDEX,
87 	INTEL_ARCH_TOPDOWN_SLOTS_INDEX,
88 	INTEL_ARCH_TOPDOWN_BE_BOUND_INDEX,
89 	INTEL_ARCH_TOPDOWN_BAD_SPEC_INDEX,
90 	INTEL_ARCH_TOPDOWN_FE_BOUND_INDEX,
91 	INTEL_ARCH_TOPDOWN_RETIRING_INDEX,
92 	INTEL_ARCH_LBR_INSERTS_INDEX,
93 	NR_INTEL_ARCH_EVENTS,
94 };
95 
96 enum amd_pmu_zen_events {
97 	AMD_ZEN_CORE_CYCLES_INDEX,
98 	AMD_ZEN_INSTRUCTIONS_INDEX,
99 	AMD_ZEN_BRANCHES_INDEX,
100 	AMD_ZEN_BRANCH_MISSES_INDEX,
101 	NR_AMD_ZEN_EVENTS,
102 };
103 
104 extern const uint64_t intel_pmu_arch_events[];
105 extern const uint64_t amd_pmu_zen_events[];
106 
107 #endif /* SELFTEST_KVM_PMU_H */
108