xref: /freebsd/sys/dev/hwpmc/hwpmc_amd.c (revision 90a6ea5ca0b1b578e265d2479511bd04347f9510)
1ebccf1e3SJoseph Koshy /*-
24d846d26SWarner Losh  * SPDX-License-Identifier: BSD-2-Clause
3718cf2ccSPedro F. Giffuni  *
4122ccdc1SJoseph Koshy  * Copyright (c) 2003-2008 Joseph Koshy
5d07f36b0SJoseph Koshy  * Copyright (c) 2007 The FreeBSD Foundation
6ebccf1e3SJoseph Koshy  * All rights reserved.
7ebccf1e3SJoseph Koshy  *
8d07f36b0SJoseph Koshy  * Portions of this software were developed by A. Joseph Koshy under
9d07f36b0SJoseph Koshy  * sponsorship from the FreeBSD Foundation and Google, Inc.
10d07f36b0SJoseph Koshy  *
11ebccf1e3SJoseph Koshy  * Redistribution and use in source and binary forms, with or without
12ebccf1e3SJoseph Koshy  * modification, are permitted provided that the following conditions
13ebccf1e3SJoseph Koshy  * are met:
14ebccf1e3SJoseph Koshy  * 1. Redistributions of source code must retain the above copyright
15ebccf1e3SJoseph Koshy  *    notice, this list of conditions and the following disclaimer.
16ebccf1e3SJoseph Koshy  * 2. Redistributions in binary form must reproduce the above copyright
17ebccf1e3SJoseph Koshy  *    notice, this list of conditions and the following disclaimer in the
18ebccf1e3SJoseph Koshy  *    documentation and/or other materials provided with the distribution.
19ebccf1e3SJoseph Koshy  *
20ebccf1e3SJoseph Koshy  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
21ebccf1e3SJoseph Koshy  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22ebccf1e3SJoseph Koshy  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23ebccf1e3SJoseph Koshy  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
24ebccf1e3SJoseph Koshy  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25ebccf1e3SJoseph Koshy  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
26ebccf1e3SJoseph Koshy  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
27ebccf1e3SJoseph Koshy  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
28ebccf1e3SJoseph Koshy  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
29ebccf1e3SJoseph Koshy  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
30ebccf1e3SJoseph Koshy  * SUCH DAMAGE.
31ebccf1e3SJoseph Koshy  */
32ebccf1e3SJoseph Koshy 
33ebccf1e3SJoseph Koshy #include <sys/cdefs.h>
34ebccf1e3SJoseph Koshy /* Support for the AMD K7 and later processors */
35ebccf1e3SJoseph Koshy 
36ebccf1e3SJoseph Koshy #include <sys/param.h>
37ebccf1e3SJoseph Koshy #include <sys/lock.h>
38ebccf1e3SJoseph Koshy #include <sys/malloc.h>
39ebccf1e3SJoseph Koshy #include <sys/mutex.h>
40c5445f8bSAndrew Gallatin #include <sys/pcpu.h>
417ad17ef9SMarcel Moolenaar #include <sys/pmc.h>
42122ccdc1SJoseph Koshy #include <sys/pmckern.h>
43ebccf1e3SJoseph Koshy #include <sys/smp.h>
44ebccf1e3SJoseph Koshy #include <sys/systm.h>
45ebccf1e3SJoseph Koshy 
46d07f36b0SJoseph Koshy #include <machine/cpu.h>
47f263522aSJoseph Koshy #include <machine/cpufunc.h>
48ebccf1e3SJoseph Koshy #include <machine/md_var.h>
49f263522aSJoseph Koshy #include <machine/specialreg.h>
50ebccf1e3SJoseph Koshy 
51680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
52f263522aSJoseph Koshy enum pmc_class	amd_pmc_class;
53ebccf1e3SJoseph Koshy #endif
54ebccf1e3SJoseph Koshy 
55c5445f8bSAndrew Gallatin #define	OVERFLOW_WAIT_COUNT	50
56c5445f8bSAndrew Gallatin 
57c5445f8bSAndrew Gallatin DPCPU_DEFINE_STATIC(uint32_t, nmi_counter);
58c5445f8bSAndrew Gallatin 
59ebccf1e3SJoseph Koshy /* AMD K7 & K8 PMCs */
60ebccf1e3SJoseph Koshy struct amd_descr {
61ebccf1e3SJoseph Koshy 	struct pmc_descr pm_descr;  /* "base class" */
62ebccf1e3SJoseph Koshy 	uint32_t	pm_evsel;   /* address of EVSEL register */
63ebccf1e3SJoseph Koshy 	uint32_t	pm_perfctr; /* address of PERFCTR register */
64ebccf1e3SJoseph Koshy };
65ebccf1e3SJoseph Koshy 
66f263522aSJoseph Koshy static  struct amd_descr amd_pmcdesc[AMD_NPMCS] =
67ebccf1e3SJoseph Koshy {
68ebccf1e3SJoseph Koshy     {
69ebccf1e3SJoseph Koshy 	.pm_descr =
70ebccf1e3SJoseph Koshy 	{
71f263522aSJoseph Koshy 		.pd_name  = "",
72f263522aSJoseph Koshy 		.pd_class = -1,
73ebccf1e3SJoseph Koshy 		.pd_caps  = AMD_PMC_CAPS,
74ebccf1e3SJoseph Koshy 		.pd_width = 48
75ebccf1e3SJoseph Koshy 	},
76ebccf1e3SJoseph Koshy 	.pm_evsel   = AMD_PMC_EVSEL_0,
77ebccf1e3SJoseph Koshy 	.pm_perfctr = AMD_PMC_PERFCTR_0
78ebccf1e3SJoseph Koshy     },
79ebccf1e3SJoseph Koshy     {
80ebccf1e3SJoseph Koshy 	.pm_descr =
81ebccf1e3SJoseph Koshy 	{
82f263522aSJoseph Koshy 		.pd_name  = "",
83f263522aSJoseph Koshy 		.pd_class = -1,
84ebccf1e3SJoseph Koshy 		.pd_caps  = AMD_PMC_CAPS,
85ebccf1e3SJoseph Koshy 		.pd_width = 48
86ebccf1e3SJoseph Koshy 	},
87ebccf1e3SJoseph Koshy 	.pm_evsel   = AMD_PMC_EVSEL_1,
88ebccf1e3SJoseph Koshy 	.pm_perfctr = AMD_PMC_PERFCTR_1
89ebccf1e3SJoseph Koshy     },
90ebccf1e3SJoseph Koshy     {
91ebccf1e3SJoseph Koshy 	.pm_descr =
92ebccf1e3SJoseph Koshy 	{
93f263522aSJoseph Koshy 		.pd_name  = "",
94f263522aSJoseph Koshy 		.pd_class = -1,
95ebccf1e3SJoseph Koshy 		.pd_caps  = AMD_PMC_CAPS,
96ebccf1e3SJoseph Koshy 		.pd_width = 48
97ebccf1e3SJoseph Koshy 	},
98ebccf1e3SJoseph Koshy 	.pm_evsel   = AMD_PMC_EVSEL_2,
99ebccf1e3SJoseph Koshy 	.pm_perfctr = AMD_PMC_PERFCTR_2
100ebccf1e3SJoseph Koshy     },
101ebccf1e3SJoseph Koshy     {
102ebccf1e3SJoseph Koshy 	.pm_descr =
103ebccf1e3SJoseph Koshy 	{
104f263522aSJoseph Koshy 		.pd_name  = "",
105f263522aSJoseph Koshy 		.pd_class = -1,
106ebccf1e3SJoseph Koshy 		.pd_caps  = AMD_PMC_CAPS,
107ebccf1e3SJoseph Koshy 		.pd_width = 48
108ebccf1e3SJoseph Koshy 	},
109ebccf1e3SJoseph Koshy 	.pm_evsel   = AMD_PMC_EVSEL_3,
110ebccf1e3SJoseph Koshy 	.pm_perfctr = AMD_PMC_PERFCTR_3
111dacc43dfSMatt Macy      },
112dacc43dfSMatt Macy     {
113dacc43dfSMatt Macy 	.pm_descr =
114dacc43dfSMatt Macy 	{
115dacc43dfSMatt Macy 		.pd_name  = "",
116dacc43dfSMatt Macy 		.pd_class = -1,
117dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
118dacc43dfSMatt Macy 		.pd_width = 48
119dacc43dfSMatt Macy 	},
120dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_4,
121dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_4
122dacc43dfSMatt Macy     },
123dacc43dfSMatt Macy     {
124dacc43dfSMatt Macy 	.pm_descr =
125dacc43dfSMatt Macy 	{
126dacc43dfSMatt Macy 		.pd_name  = "",
127dacc43dfSMatt Macy 		.pd_class = -1,
128dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
129dacc43dfSMatt Macy 		.pd_width = 48
130dacc43dfSMatt Macy 	},
131dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_5,
132dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_5
133dacc43dfSMatt Macy     },
134dacc43dfSMatt Macy     {
135dacc43dfSMatt Macy 	.pm_descr =
136dacc43dfSMatt Macy 	{
137dacc43dfSMatt Macy 		.pd_name  = "",
138dacc43dfSMatt Macy 		.pd_class = -1,
139dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
140dacc43dfSMatt Macy 		.pd_width = 48
141dacc43dfSMatt Macy 	},
142dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_L3_0,
143dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_L3_0
144dacc43dfSMatt Macy     },
145dacc43dfSMatt Macy     {
146dacc43dfSMatt Macy 	.pm_descr =
147dacc43dfSMatt Macy 	{
148dacc43dfSMatt Macy 		.pd_name  = "",
149dacc43dfSMatt Macy 		.pd_class = -1,
150dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
151dacc43dfSMatt Macy 		.pd_width = 48
152dacc43dfSMatt Macy 	},
153dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_L3_1,
154dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_L3_1
155dacc43dfSMatt Macy     },
156dacc43dfSMatt Macy     {
157dacc43dfSMatt Macy 	.pm_descr =
158dacc43dfSMatt Macy 	{
159dacc43dfSMatt Macy 		.pd_name  = "",
160dacc43dfSMatt Macy 		.pd_class = -1,
161dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
162dacc43dfSMatt Macy 		.pd_width = 48
163dacc43dfSMatt Macy 	},
164dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_L3_2,
165dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_L3_2
166dacc43dfSMatt Macy     },
167dacc43dfSMatt Macy     {
168dacc43dfSMatt Macy 	.pm_descr =
169dacc43dfSMatt Macy 	{
170dacc43dfSMatt Macy 		.pd_name  = "",
171dacc43dfSMatt Macy 		.pd_class = -1,
172dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
173dacc43dfSMatt Macy 		.pd_width = 48
174dacc43dfSMatt Macy 	},
175dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_L3_3,
176dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_L3_3
177dacc43dfSMatt Macy     },
178dacc43dfSMatt Macy     {
179dacc43dfSMatt Macy 	.pm_descr =
180dacc43dfSMatt Macy 	{
181dacc43dfSMatt Macy 		.pd_name  = "",
182dacc43dfSMatt Macy 		.pd_class = -1,
183dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
184dacc43dfSMatt Macy 		.pd_width = 48
185dacc43dfSMatt Macy 	},
186dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_L3_4,
187dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_L3_4
188dacc43dfSMatt Macy     },
189dacc43dfSMatt Macy     {
190dacc43dfSMatt Macy 	.pm_descr =
191dacc43dfSMatt Macy 	{
192dacc43dfSMatt Macy 		.pd_name  = "",
193dacc43dfSMatt Macy 		.pd_class = -1,
194dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
195dacc43dfSMatt Macy 		.pd_width = 48
196dacc43dfSMatt Macy 	},
197dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_L3_5,
198dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_L3_5
199dacc43dfSMatt Macy     },
200dacc43dfSMatt Macy     {
201dacc43dfSMatt Macy 	.pm_descr =
202dacc43dfSMatt Macy 	{
203dacc43dfSMatt Macy 		.pd_name  = "",
204dacc43dfSMatt Macy 		.pd_class = -1,
205dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
206dacc43dfSMatt Macy 		.pd_width = 48
207dacc43dfSMatt Macy 	},
208dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_DF_0,
209dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_DF_0
210dacc43dfSMatt Macy     },
211dacc43dfSMatt Macy     {
212dacc43dfSMatt Macy 	.pm_descr =
213dacc43dfSMatt Macy 	{
214dacc43dfSMatt Macy 		.pd_name  = "",
215dacc43dfSMatt Macy 		.pd_class = -1,
216dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
217dacc43dfSMatt Macy 		.pd_width = 48
218dacc43dfSMatt Macy 	},
219dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_DF_1,
220dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_DF_1
221dacc43dfSMatt Macy     },
222dacc43dfSMatt Macy     {
223dacc43dfSMatt Macy 	.pm_descr =
224dacc43dfSMatt Macy 	{
225dacc43dfSMatt Macy 		.pd_name  = "",
226dacc43dfSMatt Macy 		.pd_class = -1,
227dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
228dacc43dfSMatt Macy 		.pd_width = 48
229dacc43dfSMatt Macy 	},
230dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_DF_2,
231dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_DF_2
232dacc43dfSMatt Macy     },
233dacc43dfSMatt Macy     {
234dacc43dfSMatt Macy 	.pm_descr =
235dacc43dfSMatt Macy 	{
236dacc43dfSMatt Macy 		.pd_name  = "",
237dacc43dfSMatt Macy 		.pd_class = -1,
238dacc43dfSMatt Macy 		.pd_caps  = AMD_PMC_CAPS,
239dacc43dfSMatt Macy 		.pd_width = 48
240dacc43dfSMatt Macy 	},
241dacc43dfSMatt Macy 	.pm_evsel   = AMD_PMC_EVSEL_EP_DF_3,
242dacc43dfSMatt Macy 	.pm_perfctr = AMD_PMC_PERFCTR_EP_DF_3
243ebccf1e3SJoseph Koshy      }
244ebccf1e3SJoseph Koshy };
245ebccf1e3SJoseph Koshy 
246ebccf1e3SJoseph Koshy struct amd_event_code_map {
247ebccf1e3SJoseph Koshy 	enum pmc_event	pe_ev;	 /* enum value */
2481d3aa362SConrad Meyer 	uint16_t	pe_code; /* encoded event mask */
249ebccf1e3SJoseph Koshy 	uint8_t		pe_mask; /* bits allowed in unit mask */
250ebccf1e3SJoseph Koshy };
251ebccf1e3SJoseph Koshy 
252ebccf1e3SJoseph Koshy const struct amd_event_code_map amd_event_codes[] = {
253f263522aSJoseph Koshy #if	defined(__i386__)	/* 32 bit Athlon (K7) only */
254ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_DC_ACCESSES, 		0x40, 0 },
255ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_DC_MISSES,			0x41, 0 },
256f263522aSJoseph Koshy 	{ PMC_EV_K7_DC_REFILLS_FROM_L2,		0x42, AMD_PMC_UNITMASK_MOESI },
257f263522aSJoseph Koshy 	{ PMC_EV_K7_DC_REFILLS_FROM_SYSTEM,	0x43, AMD_PMC_UNITMASK_MOESI },
258f263522aSJoseph Koshy 	{ PMC_EV_K7_DC_WRITEBACKS,		0x44, AMD_PMC_UNITMASK_MOESI },
259ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_L1_DTLB_MISS_AND_L2_DTLB_HITS, 0x45, 0 },
260ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_L1_AND_L2_DTLB_MISSES,	0x46, 0 },
261ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_MISALIGNED_REFERENCES,	0x47, 0 },
262ebccf1e3SJoseph Koshy 
263ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_IC_FETCHES,			0x80, 0 },
264ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_IC_MISSES,			0x81, 0 },
265ebccf1e3SJoseph Koshy 
266ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_L1_ITLB_MISSES,		0x84, 0 },
267ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_L1_L2_ITLB_MISSES,		0x85, 0 },
268ebccf1e3SJoseph Koshy 
269ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_INSTRUCTIONS,	0xC0, 0 },
270ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_OPS,		0xC1, 0 },
271ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_BRANCHES,		0xC2, 0 },
272ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_BRANCHES_MISPREDICTED, 0xC3, 0 },
273ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_TAKEN_BRANCHES, 	0xC4, 0 },
274ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_TAKEN_BRANCHES_MISPREDICTED, 0xC5, 0 },
275ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_FAR_CONTROL_TRANSFERS, 0xC6, 0 },
276ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_RETIRED_RESYNC_BRANCHES,	0xC7, 0 },
277ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_INTERRUPTS_MASKED_CYCLES,	0xCD, 0 },
278ebccf1e3SJoseph Koshy 	{ PMC_EV_K7_INTERRUPTS_MASKED_WHILE_PENDING_CYCLES, 0xCE, 0 },
279f263522aSJoseph Koshy 	{ PMC_EV_K7_HARDWARE_INTERRUPTS,	0xCF, 0 },
280ebccf1e3SJoseph Koshy #endif
281ebccf1e3SJoseph Koshy 
282ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FP_DISPATCHED_FPU_OPS,		0x00, 0x3F },
283ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FP_CYCLES_WITH_NO_FPU_OPS_RETIRED,	0x01, 0x00 },
284ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FP_DISPATCHED_FPU_FAST_FLAG_OPS,	0x02, 0x00 },
285ebccf1e3SJoseph Koshy 
286ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_SEGMENT_REGISTER_LOAD, 		0x20, 0x7F },
287ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_MICROARCHITECTURAL_RESYNC_BY_SELF_MODIFYING_CODE,
288ebccf1e3SJoseph Koshy 	  						0x21, 0x00 },
289ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_MICROARCHITECTURAL_RESYNC_BY_SNOOP, 0x22, 0x00 },
290ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_BUFFER2_FULL,			0x23, 0x00 },
291ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_LOCKED_OPERATION,		0x24, 0x07 },
292ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_MICROARCHITECTURAL_LATE_CANCEL,	0x25, 0x00 },
293ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_RETIRED_CFLUSH_INSTRUCTIONS,	0x26, 0x00 },
294ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_LS_RETIRED_CPUID_INSTRUCTIONS,	0x27, 0x00 },
295ebccf1e3SJoseph Koshy 
296ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_ACCESS,				0x40, 0x00 },
297ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_MISS,				0x41, 0x00 },
298ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_REFILL_FROM_L2,			0x42, 0x1F },
299ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_REFILL_FROM_SYSTEM,		0x43, 0x1F },
300ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_COPYBACK,			0x44, 0x1F },
301ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_L1_DTLB_MISS_AND_L2_DTLB_HIT,	0x45, 0x00 },
302ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_L1_DTLB_MISS_AND_L2_DTLB_MISS,	0x46, 0x00 },
303ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_MISALIGNED_DATA_REFERENCE,	0x47, 0x00 },
304ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_MICROARCHITECTURAL_LATE_CANCEL,	0x48, 0x00 },
305ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_MICROARCHITECTURAL_EARLY_CANCEL, 0x49, 0x00 },
306ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_ONE_BIT_ECC_ERROR,		0x4A, 0x03 },
307ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_DISPATCHED_PREFETCH_INSTRUCTIONS, 0x4B, 0x07 },
308ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_DC_DCACHE_ACCESSES_BY_LOCKS,	0x4C, 0x03 },
309ebccf1e3SJoseph Koshy 
310ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_BU_CPU_CLK_UNHALTED,		0x76, 0x00 },
311ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_BU_INTERNAL_L2_REQUEST,		0x7D, 0x1F },
312ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_BU_FILL_REQUEST_L2_MISS,		0x7E, 0x07 },
313ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_BU_FILL_INTO_L2,			0x7F, 0x03 },
314ebccf1e3SJoseph Koshy 
315ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_FETCH,				0x80, 0x00 },
316ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_MISS,				0x81, 0x00 },
317ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_REFILL_FROM_L2,			0x82, 0x00 },
318ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_REFILL_FROM_SYSTEM,		0x83, 0x00 },
319ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_L1_ITLB_MISS_AND_L2_ITLB_HIT,	0x84, 0x00 },
320ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_L1_ITLB_MISS_AND_L2_ITLB_MISS,	0x85, 0x00 },
321ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_MICROARCHITECTURAL_RESYNC_BY_SNOOP, 0x86, 0x00 },
322ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_INSTRUCTION_FETCH_STALL,		0x87, 0x00 },
323ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_RETURN_STACK_HIT,		0x88, 0x00 },
324ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_IC_RETURN_STACK_OVERFLOW,		0x89, 0x00 },
325ebccf1e3SJoseph Koshy 
326ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_X86_INSTRUCTIONS,	0xC0, 0x00 },
327ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_UOPS,			0xC1, 0x00 },
328ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_BRANCHES,		0xC2, 0x00 },
329ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_BRANCHES_MISPREDICTED,	0xC3, 0x00 },
330ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_TAKEN_BRANCHES,		0xC4, 0x00 },
331ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_TAKEN_BRANCHES_MISPREDICTED, 0xC5, 0x00 },
332ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_FAR_CONTROL_TRANSFERS,	0xC6, 0x00 },
333ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_RESYNCS,			0xC7, 0x00 },
334ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_NEAR_RETURNS,		0xC8, 0x00 },
335ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_NEAR_RETURNS_MISPREDICTED, 0xC9, 0x00 },
336ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_TAKEN_BRANCHES_MISPREDICTED_BY_ADDR_MISCOMPARE,
337ebccf1e3SJoseph Koshy 							0xCA, 0x00 },
338ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_FPU_INSTRUCTIONS,	0xCB, 0x0F },
339ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_RETIRED_FASTPATH_DOUBLE_OP_INSTRUCTIONS,
340ebccf1e3SJoseph Koshy 							0xCC, 0x07 },
341ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_INTERRUPTS_MASKED_CYCLES,	0xCD, 0x00 },
342ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_INTERRUPTS_MASKED_WHILE_PENDING_CYCLES, 0xCE, 0x00 },
343ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_TAKEN_HARDWARE_INTERRUPTS,	0xCF, 0x00 },
344ebccf1e3SJoseph Koshy 
345ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DECODER_EMPTY,			0xD0, 0x00 },
346ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALLS,			0xD1, 0x00 },
347ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_FROM_BRANCH_ABORT_TO_RETIRE,
348ebccf1e3SJoseph Koshy 							0xD2, 0x00 },
349ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_FOR_SERIALIZATION, 0xD3, 0x00 },
350ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_FOR_SEGMENT_LOAD,	0xD4, 0x00 },
351ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_WHEN_REORDER_BUFFER_IS_FULL,
352ebccf1e3SJoseph Koshy 							0xD5, 0x00 },
353ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_WHEN_RESERVATION_STATIONS_ARE_FULL,
354ebccf1e3SJoseph Koshy 							0xD6, 0x00 },
355ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_WHEN_FPU_IS_FULL,	0xD7, 0x00 },
356ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_WHEN_LS_IS_FULL,	0xD8, 0x00 },
357ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_WHEN_WAITING_FOR_ALL_TO_BE_QUIET,
358ebccf1e3SJoseph Koshy 							0xD9, 0x00 },
359ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_DISPATCH_STALL_WHEN_FAR_XFER_OR_RESYNC_BRANCH_PENDING,
360ebccf1e3SJoseph Koshy 							0xDA, 0x00 },
361ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_FPU_EXCEPTIONS,			0xDB, 0x0F },
362ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_NUMBER_OF_BREAKPOINTS_FOR_DR0,	0xDC, 0x00 },
363ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_NUMBER_OF_BREAKPOINTS_FOR_DR1,	0xDD, 0x00 },
364ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_NUMBER_OF_BREAKPOINTS_FOR_DR2,	0xDE, 0x00 },
365ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_FR_NUMBER_OF_BREAKPOINTS_FOR_DR3,	0xDF, 0x00 },
366ebccf1e3SJoseph Koshy 
367ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_MEMORY_CONTROLLER_PAGE_ACCESS_EVENT, 0xE0, 0x7 },
368ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_MEMORY_CONTROLLER_PAGE_TABLE_OVERFLOW, 0xE1, 0x00 },
369ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_MEMORY_CONTROLLER_DRAM_COMMAND_SLOTS_MISSED,
370ebccf1e3SJoseph Koshy 							0xE2, 0x00 },
371ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_MEMORY_CONTROLLER_TURNAROUND,	0xE3, 0x07 },
372ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_MEMORY_CONTROLLER_BYPASS_SATURATION, 0xE4, 0x0F },
373ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_SIZED_COMMANDS,			0xEB, 0x7F },
374ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_PROBE_RESULT,			0xEC, 0x0F },
375ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_HT_BUS0_BANDWIDTH,		0xF6, 0x0F },
376ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_HT_BUS1_BANDWIDTH,		0xF7, 0x0F },
377ebccf1e3SJoseph Koshy 	{ PMC_EV_K8_NB_HT_BUS2_BANDWIDTH,		0xF8, 0x0F }
378ebccf1e3SJoseph Koshy 
379ebccf1e3SJoseph Koshy };
380ebccf1e3SJoseph Koshy 
381323b076eSPedro F. Giffuni const int amd_event_codes_size = nitems(amd_event_codes);
382ebccf1e3SJoseph Koshy 
383ebccf1e3SJoseph Koshy /*
384e829eb6dSJoseph Koshy  * Per-processor information
385e829eb6dSJoseph Koshy  */
386e829eb6dSJoseph Koshy 
387e829eb6dSJoseph Koshy struct amd_cpu {
388e829eb6dSJoseph Koshy 	struct pmc_hw	pc_amdpmcs[AMD_NPMCS];
389e829eb6dSJoseph Koshy };
390e829eb6dSJoseph Koshy 
391e829eb6dSJoseph Koshy static struct amd_cpu **amd_pcpu;
392e829eb6dSJoseph Koshy 
393e829eb6dSJoseph Koshy /*
394ebccf1e3SJoseph Koshy  * read a pmc register
395ebccf1e3SJoseph Koshy  */
396ebccf1e3SJoseph Koshy 
397ebccf1e3SJoseph Koshy static int
39839f92a76SMitchell Horne amd_read_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t *v)
399ebccf1e3SJoseph Koshy {
400ebccf1e3SJoseph Koshy 	enum pmc_mode mode;
401ebccf1e3SJoseph Koshy 	const struct amd_descr *pd;
402ebccf1e3SJoseph Koshy 	pmc_value_t tmp;
403ebccf1e3SJoseph Koshy 
404122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
405ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
406ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
407ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row-index %d", __LINE__, ri));
408e829eb6dSJoseph Koshy 	KASSERT(amd_pcpu[cpu],
409e829eb6dSJoseph Koshy 	    ("[amd,%d] null per-cpu, cpu %d", __LINE__, cpu));
410ebccf1e3SJoseph Koshy 
411ebccf1e3SJoseph Koshy 	pd = &amd_pmcdesc[ri];
412c5153e19SJoseph Koshy 	mode = PMC_TO_MODE(pm);
413ebccf1e3SJoseph Koshy 
4144a3690dfSJohn Baldwin 	PMCDBG2(MDP,REA,1,"amd-read id=%d class=%d", ri, pd->pm_descr.pd_class);
415ebccf1e3SJoseph Koshy 
416680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
417f263522aSJoseph Koshy 	KASSERT(pd->pm_descr.pd_class == amd_pmc_class,
418ebccf1e3SJoseph Koshy 	    ("[amd,%d] unknown PMC class (%d)", __LINE__,
419ebccf1e3SJoseph Koshy 		pd->pm_descr.pd_class));
420f263522aSJoseph Koshy #endif
421ebccf1e3SJoseph Koshy 
422ebccf1e3SJoseph Koshy 	tmp = rdmsr(pd->pm_perfctr); /* RDMSR serializes */
4234a3690dfSJohn Baldwin 	PMCDBG2(MDP,REA,2,"amd-read (pre-munge) id=%d -> %jd", ri, tmp);
42405e486c7SAdrian Chadd 	if (PMC_IS_SAMPLING_MODE(mode)) {
425e74c7ffcSJessica Clarke 		/*
426e74c7ffcSJessica Clarke 		 * Clamp value to 0 if the counter just overflowed,
427e74c7ffcSJessica Clarke 		 * otherwise the returned reload count would wrap to a
428e74c7ffcSJessica Clarke 		 * huge value.
429e74c7ffcSJessica Clarke 		 */
430e74c7ffcSJessica Clarke 		if ((tmp & (1ULL << 47)) == 0)
431e74c7ffcSJessica Clarke 			tmp = 0;
432e74c7ffcSJessica Clarke 		else {
43305e486c7SAdrian Chadd 			/* Sign extend 48 bit value to 64 bits. */
434e74c7ffcSJessica Clarke 			tmp = (pmc_value_t) ((int64_t)(tmp << 16) >> 16);
43505e486c7SAdrian Chadd 			tmp = AMD_PERFCTR_VALUE_TO_RELOAD_COUNT(tmp);
43605e486c7SAdrian Chadd 		}
437e74c7ffcSJessica Clarke 	}
438ebccf1e3SJoseph Koshy 	*v = tmp;
439ebccf1e3SJoseph Koshy 
4404a3690dfSJohn Baldwin 	PMCDBG2(MDP,REA,2,"amd-read (post-munge) id=%d -> %jd", ri, *v);
441ebccf1e3SJoseph Koshy 
442ebccf1e3SJoseph Koshy 	return 0;
443ebccf1e3SJoseph Koshy }
444ebccf1e3SJoseph Koshy 
445ebccf1e3SJoseph Koshy /*
446ebccf1e3SJoseph Koshy  * Write a PMC MSR.
447ebccf1e3SJoseph Koshy  */
448ebccf1e3SJoseph Koshy 
449ebccf1e3SJoseph Koshy static int
45039f92a76SMitchell Horne amd_write_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t v)
451ebccf1e3SJoseph Koshy {
452ebccf1e3SJoseph Koshy 	const struct amd_descr *pd;
453ebccf1e3SJoseph Koshy 	enum pmc_mode mode;
454ebccf1e3SJoseph Koshy 
455122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
456ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
457ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
458ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row-index %d", __LINE__, ri));
459ebccf1e3SJoseph Koshy 
460ebccf1e3SJoseph Koshy 	pd = &amd_pmcdesc[ri];
461c5153e19SJoseph Koshy 	mode = PMC_TO_MODE(pm);
462ebccf1e3SJoseph Koshy 
463680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
464f263522aSJoseph Koshy 	KASSERT(pd->pm_descr.pd_class == amd_pmc_class,
465ebccf1e3SJoseph Koshy 	    ("[amd,%d] unknown PMC class (%d)", __LINE__,
466ebccf1e3SJoseph Koshy 		pd->pm_descr.pd_class));
467f263522aSJoseph Koshy #endif
468ebccf1e3SJoseph Koshy 
469ebccf1e3SJoseph Koshy 	/* use 2's complement of the count for sampling mode PMCs */
470ebccf1e3SJoseph Koshy 	if (PMC_IS_SAMPLING_MODE(mode))
471f263522aSJoseph Koshy 		v = AMD_RELOAD_COUNT_TO_PERFCTR_VALUE(v);
472ebccf1e3SJoseph Koshy 
4734a3690dfSJohn Baldwin 	PMCDBG3(MDP,WRI,1,"amd-write cpu=%d ri=%d v=%jx", cpu, ri, v);
474ebccf1e3SJoseph Koshy 
475ebccf1e3SJoseph Koshy 	/* write the PMC value */
476ebccf1e3SJoseph Koshy 	wrmsr(pd->pm_perfctr, v);
477ebccf1e3SJoseph Koshy 	return 0;
478ebccf1e3SJoseph Koshy }
479ebccf1e3SJoseph Koshy 
480ebccf1e3SJoseph Koshy /*
481ebccf1e3SJoseph Koshy  * configure hardware pmc according to the configuration recorded in
482ebccf1e3SJoseph Koshy  * pmc 'pm'.
483ebccf1e3SJoseph Koshy  */
484ebccf1e3SJoseph Koshy 
485ebccf1e3SJoseph Koshy static int
486ebccf1e3SJoseph Koshy amd_config_pmc(int cpu, int ri, struct pmc *pm)
487ebccf1e3SJoseph Koshy {
488ebccf1e3SJoseph Koshy 	struct pmc_hw *phw;
489ebccf1e3SJoseph Koshy 
4904a3690dfSJohn Baldwin 	PMCDBG3(MDP,CFG,1, "cpu=%d ri=%d pm=%p", cpu, ri, pm);
4916b8c8cd8SJoseph Koshy 
492122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
493ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
494ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
495ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row-index %d", __LINE__, ri));
496ebccf1e3SJoseph Koshy 
497e829eb6dSJoseph Koshy 	phw = &amd_pcpu[cpu]->pc_amdpmcs[ri];
498ebccf1e3SJoseph Koshy 
499ebccf1e3SJoseph Koshy 	KASSERT(pm == NULL || phw->phw_pmc == NULL,
5006b8c8cd8SJoseph Koshy 	    ("[amd,%d] pm=%p phw->pm=%p hwpmc not unconfigured",
5016b8c8cd8SJoseph Koshy 		__LINE__, pm, phw->phw_pmc));
502ebccf1e3SJoseph Koshy 
503ebccf1e3SJoseph Koshy 	phw->phw_pmc = pm;
504ebccf1e3SJoseph Koshy 	return 0;
505ebccf1e3SJoseph Koshy }
506ebccf1e3SJoseph Koshy 
507ebccf1e3SJoseph Koshy /*
508c5153e19SJoseph Koshy  * Retrieve a configured PMC pointer from hardware state.
509c5153e19SJoseph Koshy  */
510c5153e19SJoseph Koshy 
511c5153e19SJoseph Koshy static int
512c5153e19SJoseph Koshy amd_get_config(int cpu, int ri, struct pmc **ppm)
513c5153e19SJoseph Koshy {
514e829eb6dSJoseph Koshy 	*ppm = amd_pcpu[cpu]->pc_amdpmcs[ri].phw_pmc;
515c5153e19SJoseph Koshy 
516c5153e19SJoseph Koshy 	return 0;
517c5153e19SJoseph Koshy }
518c5153e19SJoseph Koshy 
519c5153e19SJoseph Koshy /*
520ebccf1e3SJoseph Koshy  * Machine dependent actions taken during the context switch in of a
521ebccf1e3SJoseph Koshy  * thread.
522ebccf1e3SJoseph Koshy  */
523ebccf1e3SJoseph Koshy 
524ebccf1e3SJoseph Koshy static int
5256b8c8cd8SJoseph Koshy amd_switch_in(struct pmc_cpu *pc, struct pmc_process *pp)
526ebccf1e3SJoseph Koshy {
527ebccf1e3SJoseph Koshy 	(void) pc;
528ebccf1e3SJoseph Koshy 
5294a3690dfSJohn Baldwin 	PMCDBG3(MDP,SWI,1, "pc=%p pp=%p enable-msr=%d", pc, pp,
530c5153e19SJoseph Koshy 	    (pp->pp_flags & PMC_PP_ENABLE_MSR_ACCESS) != 0);
5316b8c8cd8SJoseph Koshy 
5326b8c8cd8SJoseph Koshy 	/* enable the RDPMC instruction if needed */
533c5153e19SJoseph Koshy 	if (pp->pp_flags & PMC_PP_ENABLE_MSR_ACCESS)
534ebccf1e3SJoseph Koshy 		load_cr4(rcr4() | CR4_PCE);
5356b8c8cd8SJoseph Koshy 
536ebccf1e3SJoseph Koshy 	return 0;
537ebccf1e3SJoseph Koshy }
538ebccf1e3SJoseph Koshy 
539ebccf1e3SJoseph Koshy /*
540ebccf1e3SJoseph Koshy  * Machine dependent actions taken during the context switch out of a
541ebccf1e3SJoseph Koshy  * thread.
542ebccf1e3SJoseph Koshy  */
543ebccf1e3SJoseph Koshy 
544ebccf1e3SJoseph Koshy static int
5456b8c8cd8SJoseph Koshy amd_switch_out(struct pmc_cpu *pc, struct pmc_process *pp)
546ebccf1e3SJoseph Koshy {
547ebccf1e3SJoseph Koshy 	(void) pc;
5486b8c8cd8SJoseph Koshy 	(void) pp;		/* can be NULL */
549ebccf1e3SJoseph Koshy 
5504a3690dfSJohn Baldwin 	PMCDBG3(MDP,SWO,1, "pc=%p pp=%p enable-msr=%d", pc, pp, pp ?
551c5153e19SJoseph Koshy 	    (pp->pp_flags & PMC_PP_ENABLE_MSR_ACCESS) == 1 : 0);
5526b8c8cd8SJoseph Koshy 
5536b8c8cd8SJoseph Koshy 	/* always turn off the RDPMC instruction */
554ebccf1e3SJoseph Koshy 	load_cr4(rcr4() & ~CR4_PCE);
5556b8c8cd8SJoseph Koshy 
556ebccf1e3SJoseph Koshy 	return 0;
557ebccf1e3SJoseph Koshy }
558ebccf1e3SJoseph Koshy 
559ebccf1e3SJoseph Koshy /*
560ebccf1e3SJoseph Koshy  * Check if a given allocation is feasible.
561ebccf1e3SJoseph Koshy  */
562ebccf1e3SJoseph Koshy 
563ebccf1e3SJoseph Koshy static int
564ebccf1e3SJoseph Koshy amd_allocate_pmc(int cpu, int ri, struct pmc *pm,
565ebccf1e3SJoseph Koshy     const struct pmc_op_pmcallocate *a)
566ebccf1e3SJoseph Koshy {
567ebccf1e3SJoseph Koshy 	int i;
568dacc43dfSMatt Macy 	uint64_t allowed_unitmask, caps, config, unitmask;
569ebccf1e3SJoseph Koshy 	enum pmc_event pe;
570ebccf1e3SJoseph Koshy 	const struct pmc_descr *pd;
571ebccf1e3SJoseph Koshy 
572ebccf1e3SJoseph Koshy 	(void) cpu;
573ebccf1e3SJoseph Koshy 
574122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
575ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
576ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
577ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row index %d", __LINE__, ri));
578ebccf1e3SJoseph Koshy 
579ebccf1e3SJoseph Koshy 	pd = &amd_pmcdesc[ri].pm_descr;
580ebccf1e3SJoseph Koshy 
581ebccf1e3SJoseph Koshy 	/* check class match */
582c5153e19SJoseph Koshy 	if (pd->pd_class != a->pm_class)
583ebccf1e3SJoseph Koshy 		return EINVAL;
584ebccf1e3SJoseph Koshy 
585c190fb35SMitchell Horne 	if ((a->pm_flags & PMC_F_EV_PMU) == 0)
586c190fb35SMitchell Horne 		return (EINVAL);
587c190fb35SMitchell Horne 
588ebccf1e3SJoseph Koshy 	caps = pm->pm_caps;
589ebccf1e3SJoseph Koshy 
5904a3690dfSJohn Baldwin 	PMCDBG2(MDP,ALL,1,"amd-allocate ri=%d caps=0x%x", ri, caps);
591ebccf1e3SJoseph Koshy 
592dacc43dfSMatt Macy 	if((ri >= 0 && ri < 6) && !(a->pm_md.pm_amd.pm_amd_sub_class == PMC_AMD_SUB_CLASS_CORE))
593dacc43dfSMatt Macy 		return EINVAL;
594dacc43dfSMatt Macy 	if((ri >= 6 && ri < 12) && !(a->pm_md.pm_amd.pm_amd_sub_class == PMC_AMD_SUB_CLASS_L3_CACHE))
595dacc43dfSMatt Macy 		return EINVAL;
596dacc43dfSMatt Macy 	if((ri >= 12 && ri < 16) && !(a->pm_md.pm_amd.pm_amd_sub_class == PMC_AMD_SUB_CLASS_DATA_FABRIC))
597dacc43dfSMatt Macy 		return EINVAL;
598dacc43dfSMatt Macy 
59981eb4dcfSMatt Macy 	if (strlen(pmc_cpuid) != 0) {
60081eb4dcfSMatt Macy 		pm->pm_md.pm_amd.pm_amd_evsel =
60181eb4dcfSMatt Macy 			a->pm_md.pm_amd.pm_amd_config;
60281eb4dcfSMatt Macy 		PMCDBG2(MDP,ALL,2,"amd-allocate ri=%d -> config=0x%x", ri, a->pm_md.pm_amd.pm_amd_config);
60381eb4dcfSMatt Macy 		return (0);
60481eb4dcfSMatt Macy 	}
605ebccf1e3SJoseph Koshy 
606ebccf1e3SJoseph Koshy 	pe = a->pm_ev;
607ebccf1e3SJoseph Koshy 
608ebccf1e3SJoseph Koshy 	/* map ev to the correct event mask code */
609ebccf1e3SJoseph Koshy 	config = allowed_unitmask = 0;
610ebccf1e3SJoseph Koshy 	for (i = 0; i < amd_event_codes_size; i++)
611ebccf1e3SJoseph Koshy 		if (amd_event_codes[i].pe_ev == pe) {
612ebccf1e3SJoseph Koshy 			config =
613ebccf1e3SJoseph Koshy 			    AMD_PMC_TO_EVENTMASK(amd_event_codes[i].pe_code);
614ebccf1e3SJoseph Koshy 			allowed_unitmask =
615ebccf1e3SJoseph Koshy 			    AMD_PMC_TO_UNITMASK(amd_event_codes[i].pe_mask);
616ebccf1e3SJoseph Koshy 			break;
617ebccf1e3SJoseph Koshy 		}
618ebccf1e3SJoseph Koshy 	if (i == amd_event_codes_size)
619ebccf1e3SJoseph Koshy 		return EINVAL;
620ebccf1e3SJoseph Koshy 
621f263522aSJoseph Koshy 	unitmask = a->pm_md.pm_amd.pm_amd_config & AMD_PMC_UNITMASK;
622ebccf1e3SJoseph Koshy 	if (unitmask & ~allowed_unitmask) /* disallow reserved bits */
623ebccf1e3SJoseph Koshy 		return EINVAL;
624ebccf1e3SJoseph Koshy 
625ebccf1e3SJoseph Koshy 	if (unitmask && (caps & PMC_CAP_QUALIFIER))
626ebccf1e3SJoseph Koshy 		config |= unitmask;
627ebccf1e3SJoseph Koshy 
628ebccf1e3SJoseph Koshy 	if (caps & PMC_CAP_THRESHOLD)
629f263522aSJoseph Koshy 		config |= a->pm_md.pm_amd.pm_amd_config & AMD_PMC_COUNTERMASK;
630ebccf1e3SJoseph Koshy 
631ebccf1e3SJoseph Koshy 	/* set at least one of the 'usr' or 'os' caps */
632ebccf1e3SJoseph Koshy 	if (caps & PMC_CAP_USER)
633ebccf1e3SJoseph Koshy 		config |= AMD_PMC_USR;
634ebccf1e3SJoseph Koshy 	if (caps & PMC_CAP_SYSTEM)
635ebccf1e3SJoseph Koshy 		config |= AMD_PMC_OS;
636ebccf1e3SJoseph Koshy 	if ((caps & (PMC_CAP_USER|PMC_CAP_SYSTEM)) == 0)
637ebccf1e3SJoseph Koshy 		config |= (AMD_PMC_USR|AMD_PMC_OS);
638ebccf1e3SJoseph Koshy 
639ebccf1e3SJoseph Koshy 	if (caps & PMC_CAP_EDGE)
640ebccf1e3SJoseph Koshy 		config |= AMD_PMC_EDGE;
641ebccf1e3SJoseph Koshy 	if (caps & PMC_CAP_INVERT)
642ebccf1e3SJoseph Koshy 		config |= AMD_PMC_INVERT;
643ebccf1e3SJoseph Koshy 	if (caps & PMC_CAP_INTERRUPT)
644ebccf1e3SJoseph Koshy 		config |= AMD_PMC_INT;
645ebccf1e3SJoseph Koshy 
646ebccf1e3SJoseph Koshy 	pm->pm_md.pm_amd.pm_amd_evsel = config; /* save config value */
647ebccf1e3SJoseph Koshy 
6484a3690dfSJohn Baldwin 	PMCDBG2(MDP,ALL,2,"amd-allocate ri=%d -> config=0x%x", ri, config);
649ebccf1e3SJoseph Koshy 
650ebccf1e3SJoseph Koshy 	return 0;
651ebccf1e3SJoseph Koshy }
652ebccf1e3SJoseph Koshy 
653ebccf1e3SJoseph Koshy /*
654ebccf1e3SJoseph Koshy  * Release machine dependent state associated with a PMC.  This is a
655ebccf1e3SJoseph Koshy  * no-op on this architecture.
656ebccf1e3SJoseph Koshy  *
657ebccf1e3SJoseph Koshy  */
658ebccf1e3SJoseph Koshy 
659ebccf1e3SJoseph Koshy /* ARGSUSED0 */
660ebccf1e3SJoseph Koshy static int
661ebccf1e3SJoseph Koshy amd_release_pmc(int cpu, int ri, struct pmc *pmc)
662ebccf1e3SJoseph Koshy {
663680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
664ebccf1e3SJoseph Koshy 	const struct amd_descr *pd;
665ebccf1e3SJoseph Koshy #endif
666aee6e7dcSMateusz Guzik 	struct pmc_hw *phw __diagused;
667ebccf1e3SJoseph Koshy 
668ebccf1e3SJoseph Koshy 	(void) pmc;
669ebccf1e3SJoseph Koshy 
670122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
671ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
672ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
673ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row-index %d", __LINE__, ri));
674ebccf1e3SJoseph Koshy 
675e829eb6dSJoseph Koshy 	phw = &amd_pcpu[cpu]->pc_amdpmcs[ri];
676ebccf1e3SJoseph Koshy 
677ebccf1e3SJoseph Koshy 	KASSERT(phw->phw_pmc == NULL,
678ebccf1e3SJoseph Koshy 	    ("[amd,%d] PHW pmc %p non-NULL", __LINE__, phw->phw_pmc));
679ebccf1e3SJoseph Koshy 
680680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
681ebccf1e3SJoseph Koshy 	pd = &amd_pmcdesc[ri];
682f263522aSJoseph Koshy 	if (pd->pm_descr.pd_class == amd_pmc_class)
683ebccf1e3SJoseph Koshy 		KASSERT(AMD_PMC_IS_STOPPED(pd->pm_evsel),
684ebccf1e3SJoseph Koshy 		    ("[amd,%d] PMC %d released while active", __LINE__, ri));
685ebccf1e3SJoseph Koshy #endif
686ebccf1e3SJoseph Koshy 
687ebccf1e3SJoseph Koshy 	return 0;
688ebccf1e3SJoseph Koshy }
689ebccf1e3SJoseph Koshy 
690ebccf1e3SJoseph Koshy /*
691ebccf1e3SJoseph Koshy  * start a PMC.
692ebccf1e3SJoseph Koshy  */
693ebccf1e3SJoseph Koshy 
694ebccf1e3SJoseph Koshy static int
69539f92a76SMitchell Horne amd_start_pmc(int cpu, int ri, struct pmc *pm)
696ebccf1e3SJoseph Koshy {
697dacc43dfSMatt Macy 	uint64_t config;
698ebccf1e3SJoseph Koshy 	const struct amd_descr *pd;
699ebccf1e3SJoseph Koshy 
700122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
701ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
702ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
703ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row-index %d", __LINE__, ri));
704ebccf1e3SJoseph Koshy 
705ebccf1e3SJoseph Koshy 	pd = &amd_pmcdesc[ri];
706ebccf1e3SJoseph Koshy 
7074a3690dfSJohn Baldwin 	PMCDBG2(MDP,STA,1,"amd-start cpu=%d ri=%d", cpu, ri);
708ebccf1e3SJoseph Koshy 
709ebccf1e3SJoseph Koshy 	KASSERT(AMD_PMC_IS_STOPPED(pd->pm_evsel),
710ebccf1e3SJoseph Koshy 	    ("[amd,%d] pmc%d,cpu%d: Starting active PMC \"%s\"", __LINE__,
711ebccf1e3SJoseph Koshy 	    ri, cpu, pd->pm_descr.pd_name));
712ebccf1e3SJoseph Koshy 
713ebccf1e3SJoseph Koshy 	/* turn on the PMC ENABLE bit */
714ebccf1e3SJoseph Koshy 	config = pm->pm_md.pm_amd.pm_amd_evsel | AMD_PMC_ENABLE;
715ebccf1e3SJoseph Koshy 
7164a3690dfSJohn Baldwin 	PMCDBG1(MDP,STA,2,"amd-start config=0x%x", config);
717ebccf1e3SJoseph Koshy 
718ebccf1e3SJoseph Koshy 	wrmsr(pd->pm_evsel, config);
719ebccf1e3SJoseph Koshy 	return 0;
720ebccf1e3SJoseph Koshy }
721ebccf1e3SJoseph Koshy 
722ebccf1e3SJoseph Koshy /*
723ebccf1e3SJoseph Koshy  * Stop a PMC.
724ebccf1e3SJoseph Koshy  */
725ebccf1e3SJoseph Koshy 
726ebccf1e3SJoseph Koshy static int
72739f92a76SMitchell Horne amd_stop_pmc(int cpu, int ri, struct pmc *pm)
728ebccf1e3SJoseph Koshy {
729ebccf1e3SJoseph Koshy 	const struct amd_descr *pd;
730ebccf1e3SJoseph Koshy 	uint64_t config;
731c5445f8bSAndrew Gallatin 	int i;
732ebccf1e3SJoseph Koshy 
733122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
734ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
735ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
736ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal row-index %d", __LINE__, ri));
737ebccf1e3SJoseph Koshy 
738ebccf1e3SJoseph Koshy 	pd = &amd_pmcdesc[ri];
739ebccf1e3SJoseph Koshy 
740ebccf1e3SJoseph Koshy 	KASSERT(!AMD_PMC_IS_STOPPED(pd->pm_evsel),
741ebccf1e3SJoseph Koshy 	    ("[amd,%d] PMC%d, CPU%d \"%s\" already stopped",
742ebccf1e3SJoseph Koshy 		__LINE__, ri, cpu, pd->pm_descr.pd_name));
743ebccf1e3SJoseph Koshy 
7444a3690dfSJohn Baldwin 	PMCDBG1(MDP,STO,1,"amd-stop ri=%d", ri);
745ebccf1e3SJoseph Koshy 
746ebccf1e3SJoseph Koshy 	/* turn off the PMC ENABLE bit */
747ebccf1e3SJoseph Koshy 	config = pm->pm_md.pm_amd.pm_amd_evsel & ~AMD_PMC_ENABLE;
748ebccf1e3SJoseph Koshy 	wrmsr(pd->pm_evsel, config);
749c5445f8bSAndrew Gallatin 
750c5445f8bSAndrew Gallatin 	/*
751c5445f8bSAndrew Gallatin 	 * Due to NMI latency on newer AMD processors
752c5445f8bSAndrew Gallatin 	 * NMI interrupts are ignored, which leads to
75304389c85SGordon Bergling 	 * panic or messages based on kernel configuration
754c5445f8bSAndrew Gallatin 	 */
755c5445f8bSAndrew Gallatin 
756c5445f8bSAndrew Gallatin 	/* Wait for the count to be reset */
757c5445f8bSAndrew Gallatin 	for (i = 0; i < OVERFLOW_WAIT_COUNT; i++) {
758c5445f8bSAndrew Gallatin 		if (rdmsr(pd->pm_perfctr) & (1 << (pd->pm_descr.pd_width - 1)))
759c5445f8bSAndrew Gallatin 			break;
760c5445f8bSAndrew Gallatin 
761c5445f8bSAndrew Gallatin 		DELAY(1);
762c5445f8bSAndrew Gallatin 	}
763c5445f8bSAndrew Gallatin 
764ebccf1e3SJoseph Koshy 	return 0;
765ebccf1e3SJoseph Koshy }
766ebccf1e3SJoseph Koshy 
767ebccf1e3SJoseph Koshy /*
768ebccf1e3SJoseph Koshy  * Interrupt handler.  This function needs to return '1' if the
769ebccf1e3SJoseph Koshy  * interrupt was this CPU's PMCs or '0' otherwise.  It is not allowed
770ebccf1e3SJoseph Koshy  * to sleep or do anything a 'fast' interrupt handler is not allowed
771ebccf1e3SJoseph Koshy  * to do.
772ebccf1e3SJoseph Koshy  */
773ebccf1e3SJoseph Koshy 
774ebccf1e3SJoseph Koshy static int
775eb7c9019SMatt Macy amd_intr(struct trapframe *tf)
776ebccf1e3SJoseph Koshy {
777eb7c9019SMatt Macy 	int i, error, retval, cpu;
778dacc43dfSMatt Macy 	uint64_t config, evsel, perfctr;
779ebccf1e3SJoseph Koshy 	struct pmc *pm;
780e829eb6dSJoseph Koshy 	struct amd_cpu *pac;
781f263522aSJoseph Koshy 	pmc_value_t v;
782c5445f8bSAndrew Gallatin 	uint32_t active = 0, count = 0;
78336c0fd9dSJoseph Koshy 
784eb7c9019SMatt Macy 	cpu = curcpu;
785122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
786ebccf1e3SJoseph Koshy 	    ("[amd,%d] out of range CPU %d", __LINE__, cpu));
787ebccf1e3SJoseph Koshy 
7884a3690dfSJohn Baldwin 	PMCDBG3(MDP,INT,1, "cpu=%d tf=%p um=%d", cpu, (void *) tf,
789d07f36b0SJoseph Koshy 	    TRAPF_USERMODE(tf));
790f263522aSJoseph Koshy 
791ebccf1e3SJoseph Koshy 	retval = 0;
792ebccf1e3SJoseph Koshy 
793e829eb6dSJoseph Koshy 	pac = amd_pcpu[cpu];
794ebccf1e3SJoseph Koshy 
795ebccf1e3SJoseph Koshy 	/*
796ebccf1e3SJoseph Koshy 	 * look for all PMCs that have interrupted:
797f263522aSJoseph Koshy 	 * - look for a running, sampling PMC which has overflowed
798f263522aSJoseph Koshy 	 *   and which has a valid 'struct pmc' association
799f263522aSJoseph Koshy 	 *
800f263522aSJoseph Koshy 	 * If found, we call a helper to process the interrupt.
801bebaef4aSJoseph Koshy 	 *
802c5445f8bSAndrew Gallatin 	 * PMCs interrupting at the same time are collapsed into
803c5445f8bSAndrew Gallatin 	 * a single interrupt. Check all the valid pmcs for
804c5445f8bSAndrew Gallatin 	 * overflow.
805ebccf1e3SJoseph Koshy 	 */
806ebccf1e3SJoseph Koshy 
807c5445f8bSAndrew Gallatin 	for (i = 0; i < AMD_CORE_NPMCS; i++) {
808f263522aSJoseph Koshy 
809dfd9bc23SJoseph Koshy 		if ((pm = pac->pc_amdpmcs[i].phw_pmc) == NULL ||
810f263522aSJoseph Koshy 		    !PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm))) {
811ebccf1e3SJoseph Koshy 			continue;
812ebccf1e3SJoseph Koshy 		}
813ebccf1e3SJoseph Koshy 
814c5445f8bSAndrew Gallatin 		/* Consider pmc with valid handle as active */
815c5445f8bSAndrew Gallatin 		active++;
816c5445f8bSAndrew Gallatin 
817dfd9bc23SJoseph Koshy 		if (!AMD_PMC_HAS_OVERFLOWED(i))
818dfd9bc23SJoseph Koshy 			continue;
819dfd9bc23SJoseph Koshy 
820d07f36b0SJoseph Koshy 		retval = 1;	/* Found an interrupting PMC. */
821bebaef4aSJoseph Koshy 
822dfd9bc23SJoseph Koshy 		if (pm->pm_state != PMC_STATE_RUNNING)
823dfd9bc23SJoseph Koshy 			continue;
824dfd9bc23SJoseph Koshy 
825d07f36b0SJoseph Koshy 		/* Stop the PMC, reload count. */
826c5445f8bSAndrew Gallatin 		evsel	= amd_pmcdesc[i].pm_evsel;
827c5445f8bSAndrew Gallatin 		perfctr	= amd_pmcdesc[i].pm_perfctr;
828f263522aSJoseph Koshy 		v       = pm->pm_sc.pm_reloadcount;
829f263522aSJoseph Koshy 		config  = rdmsr(evsel);
830f263522aSJoseph Koshy 
831f263522aSJoseph Koshy 		KASSERT((config & ~AMD_PMC_ENABLE) ==
832f263522aSJoseph Koshy 		    (pm->pm_md.pm_amd.pm_amd_evsel & ~AMD_PMC_ENABLE),
833dacc43dfSMatt Macy 		    ("[amd,%d] config mismatch reg=0x%jx pm=0x%jx", __LINE__,
834dacc43dfSMatt Macy 			 (uintmax_t)config, (uintmax_t)pm->pm_md.pm_amd.pm_amd_evsel));
835f263522aSJoseph Koshy 
836f263522aSJoseph Koshy 		wrmsr(evsel, config & ~AMD_PMC_ENABLE);
837f263522aSJoseph Koshy 		wrmsr(perfctr, AMD_RELOAD_COUNT_TO_PERFCTR_VALUE(v));
838f263522aSJoseph Koshy 
839d07f36b0SJoseph Koshy 		/* Restart the counter if logging succeeded. */
840eb7c9019SMatt Macy 		error = pmc_process_interrupt(PMC_HR, pm, tf);
841f263522aSJoseph Koshy 		if (error == 0)
8423c1f73b1SAndriy Gapon 			wrmsr(evsel, config);
843ebccf1e3SJoseph Koshy 	}
844f263522aSJoseph Koshy 
845c5445f8bSAndrew Gallatin 	/*
846c5445f8bSAndrew Gallatin 	 * Due to NMI latency, there can be a scenario in which
847c5445f8bSAndrew Gallatin 	 * multiple pmcs gets serviced in an earlier NMI and we
848c5445f8bSAndrew Gallatin 	 * do not find an overflow in the subsequent NMI.
849c5445f8bSAndrew Gallatin 	 *
850c5445f8bSAndrew Gallatin 	 * For such cases we keep a per-cpu count of active NMIs
851c5445f8bSAndrew Gallatin 	 * and compare it with min(active pmcs, 2) to determine
852c5445f8bSAndrew Gallatin 	 * if this NMI was for a pmc overflow which was serviced
853c5445f8bSAndrew Gallatin 	 * in an earlier request or should be ignored.
854c5445f8bSAndrew Gallatin 	 */
855c5445f8bSAndrew Gallatin 
856c5445f8bSAndrew Gallatin 	if (retval) {
857c5445f8bSAndrew Gallatin 		DPCPU_SET(nmi_counter, min(2, active));
858c5445f8bSAndrew Gallatin 	} else {
859c5445f8bSAndrew Gallatin 		if ((count = DPCPU_GET(nmi_counter))) {
860c5445f8bSAndrew Gallatin 			retval = 1;
861c5445f8bSAndrew Gallatin 			DPCPU_SET(nmi_counter, --count);
862c5445f8bSAndrew Gallatin 		}
863c5445f8bSAndrew Gallatin 	}
864c5445f8bSAndrew Gallatin 
865e6b475e0SMatt Macy 	if (retval)
866e6b475e0SMatt Macy 		counter_u64_add(pmc_stats.pm_intr_processed, 1);
867e6b475e0SMatt Macy 	else
868e6b475e0SMatt Macy 		counter_u64_add(pmc_stats.pm_intr_ignored, 1);
869fbf1556dSJoseph Koshy 
8703c1f73b1SAndriy Gapon 	PMCDBG1(MDP,INT,2, "retval=%d", retval);
871d07f36b0SJoseph Koshy 	return (retval);
872ebccf1e3SJoseph Koshy }
873ebccf1e3SJoseph Koshy 
874ebccf1e3SJoseph Koshy /*
875ebccf1e3SJoseph Koshy  * describe a PMC
876ebccf1e3SJoseph Koshy  */
877ebccf1e3SJoseph Koshy static int
878ebccf1e3SJoseph Koshy amd_describe(int cpu, int ri, struct pmc_info *pi, struct pmc **ppmc)
879ebccf1e3SJoseph Koshy {
880ebccf1e3SJoseph Koshy 	const struct amd_descr *pd;
881ebccf1e3SJoseph Koshy 	struct pmc_hw *phw;
882ebccf1e3SJoseph Koshy 
883122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
884ebccf1e3SJoseph Koshy 	    ("[amd,%d] illegal CPU %d", __LINE__, cpu));
885ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
886ebccf1e3SJoseph Koshy 	    ("[amd,%d] row-index %d out of range", __LINE__, ri));
887ebccf1e3SJoseph Koshy 
888e829eb6dSJoseph Koshy 	phw = &amd_pcpu[cpu]->pc_amdpmcs[ri];
889ebccf1e3SJoseph Koshy 	pd  = &amd_pmcdesc[ri];
890ebccf1e3SJoseph Koshy 
89131610e34SMitchell Horne 	strlcpy(pi->pm_name, pd->pm_descr.pd_name, sizeof(pi->pm_name));
892ebccf1e3SJoseph Koshy 	pi->pm_class = pd->pm_descr.pd_class;
893ebccf1e3SJoseph Koshy 
894ebccf1e3SJoseph Koshy 	if (phw->phw_state & PMC_PHW_FLAG_IS_ENABLED) {
895ebccf1e3SJoseph Koshy 		pi->pm_enabled = TRUE;
896ebccf1e3SJoseph Koshy 		*ppmc          = phw->phw_pmc;
897ebccf1e3SJoseph Koshy 	} else {
898ebccf1e3SJoseph Koshy 		pi->pm_enabled = FALSE;
899ebccf1e3SJoseph Koshy 		*ppmc          = NULL;
900ebccf1e3SJoseph Koshy 	}
901ebccf1e3SJoseph Koshy 
902ebccf1e3SJoseph Koshy 	return 0;
903ebccf1e3SJoseph Koshy }
904ebccf1e3SJoseph Koshy 
905ebccf1e3SJoseph Koshy /*
906ebccf1e3SJoseph Koshy  * i386 specific entry points
907ebccf1e3SJoseph Koshy  */
908ebccf1e3SJoseph Koshy 
909ebccf1e3SJoseph Koshy /*
910ebccf1e3SJoseph Koshy  * return the MSR address of the given PMC.
911ebccf1e3SJoseph Koshy  */
912ebccf1e3SJoseph Koshy 
913ebccf1e3SJoseph Koshy static int
914ebccf1e3SJoseph Koshy amd_get_msr(int ri, uint32_t *msr)
915ebccf1e3SJoseph Koshy {
916ebccf1e3SJoseph Koshy 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
917ebccf1e3SJoseph Koshy 	    ("[amd,%d] ri %d out of range", __LINE__, ri));
918ebccf1e3SJoseph Koshy 
9196b8c8cd8SJoseph Koshy 	*msr = amd_pmcdesc[ri].pm_perfctr - AMD_PMC_PERFCTR_0;
920e829eb6dSJoseph Koshy 
921e829eb6dSJoseph Koshy 	return (0);
922ebccf1e3SJoseph Koshy }
923ebccf1e3SJoseph Koshy 
924ebccf1e3SJoseph Koshy /*
925ebccf1e3SJoseph Koshy  * processor dependent initialization.
926ebccf1e3SJoseph Koshy  */
927ebccf1e3SJoseph Koshy 
928ebccf1e3SJoseph Koshy static int
929e829eb6dSJoseph Koshy amd_pcpu_init(struct pmc_mdep *md, int cpu)
930ebccf1e3SJoseph Koshy {
931e829eb6dSJoseph Koshy 	int classindex, first_ri, n;
932e829eb6dSJoseph Koshy 	struct pmc_cpu *pc;
933e829eb6dSJoseph Koshy 	struct amd_cpu *pac;
934ebccf1e3SJoseph Koshy 	struct pmc_hw  *phw;
935ebccf1e3SJoseph Koshy 
936122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
937ebccf1e3SJoseph Koshy 	    ("[amd,%d] insane cpu number %d", __LINE__, cpu));
938ebccf1e3SJoseph Koshy 
9394a3690dfSJohn Baldwin 	PMCDBG1(MDP,INI,1,"amd-init cpu=%d", cpu);
940ebccf1e3SJoseph Koshy 
941e829eb6dSJoseph Koshy 	amd_pcpu[cpu] = pac = malloc(sizeof(struct amd_cpu), M_PMC,
942ebccf1e3SJoseph Koshy 	    M_WAITOK|M_ZERO);
943ebccf1e3SJoseph Koshy 
944ebccf1e3SJoseph Koshy 	/*
945e829eb6dSJoseph Koshy 	 * Set the content of the hardware descriptors to a known
946e829eb6dSJoseph Koshy 	 * state and initialize pointers in the MI per-cpu descriptor.
947ebccf1e3SJoseph Koshy 	 */
948e829eb6dSJoseph Koshy 	pc = pmc_pcpu[cpu];
949e829eb6dSJoseph Koshy #if	defined(__amd64__)
950e829eb6dSJoseph Koshy 	classindex = PMC_MDEP_CLASS_INDEX_K8;
951e829eb6dSJoseph Koshy #elif	defined(__i386__)
952e829eb6dSJoseph Koshy 	classindex = md->pmd_cputype == PMC_CPU_AMD_K8 ?
953e829eb6dSJoseph Koshy 	    PMC_MDEP_CLASS_INDEX_K8 : PMC_MDEP_CLASS_INDEX_K7;
954e829eb6dSJoseph Koshy #endif
955e829eb6dSJoseph Koshy 	first_ri = md->pmd_classdep[classindex].pcd_ri;
956ebccf1e3SJoseph Koshy 
957e829eb6dSJoseph Koshy 	KASSERT(pc != NULL, ("[amd,%d] NULL per-cpu pointer", __LINE__));
958e829eb6dSJoseph Koshy 
959e829eb6dSJoseph Koshy 	for (n = 0, phw = pac->pc_amdpmcs; n < AMD_NPMCS; n++, phw++) {
960ebccf1e3SJoseph Koshy 		phw->phw_state 	  = PMC_PHW_FLAG_IS_ENABLED |
961ebccf1e3SJoseph Koshy 		    PMC_PHW_CPU_TO_STATE(cpu) | PMC_PHW_INDEX_TO_STATE(n);
962ebccf1e3SJoseph Koshy 		phw->phw_pmc	  = NULL;
963e829eb6dSJoseph Koshy 		pc->pc_hwpmcs[n + first_ri]  = phw;
964ebccf1e3SJoseph Koshy 	}
965ebccf1e3SJoseph Koshy 
966e829eb6dSJoseph Koshy 	return (0);
967ebccf1e3SJoseph Koshy }
968ebccf1e3SJoseph Koshy 
969ebccf1e3SJoseph Koshy 
970ebccf1e3SJoseph Koshy /*
971ebccf1e3SJoseph Koshy  * processor dependent cleanup prior to the KLD
972ebccf1e3SJoseph Koshy  * being unloaded
973ebccf1e3SJoseph Koshy  */
974ebccf1e3SJoseph Koshy 
975ebccf1e3SJoseph Koshy static int
976e829eb6dSJoseph Koshy amd_pcpu_fini(struct pmc_mdep *md, int cpu)
977ebccf1e3SJoseph Koshy {
978e829eb6dSJoseph Koshy 	int classindex, first_ri, i;
979ebccf1e3SJoseph Koshy 	uint32_t evsel;
980e829eb6dSJoseph Koshy 	struct pmc_cpu *pc;
981e829eb6dSJoseph Koshy 	struct amd_cpu *pac;
982ebccf1e3SJoseph Koshy 
983122ccdc1SJoseph Koshy 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
984ebccf1e3SJoseph Koshy 	    ("[amd,%d] insane cpu number (%d)", __LINE__, cpu));
985ebccf1e3SJoseph Koshy 
9864a3690dfSJohn Baldwin 	PMCDBG1(MDP,INI,1,"amd-cleanup cpu=%d", cpu);
987ebccf1e3SJoseph Koshy 
988ebccf1e3SJoseph Koshy 	/*
989ebccf1e3SJoseph Koshy 	 * First, turn off all PMCs on this CPU.
990ebccf1e3SJoseph Koshy 	 */
991ebccf1e3SJoseph Koshy 	for (i = 0; i < 4; i++) { /* XXX this loop is now not needed */
992ebccf1e3SJoseph Koshy 		evsel = rdmsr(AMD_PMC_EVSEL_0 + i);
993ebccf1e3SJoseph Koshy 		evsel &= ~AMD_PMC_ENABLE;
994ebccf1e3SJoseph Koshy 		wrmsr(AMD_PMC_EVSEL_0 + i, evsel);
995ebccf1e3SJoseph Koshy 	}
996ebccf1e3SJoseph Koshy 
997ebccf1e3SJoseph Koshy 	/*
998ebccf1e3SJoseph Koshy 	 * Next, free up allocated space.
999ebccf1e3SJoseph Koshy 	 */
1000e829eb6dSJoseph Koshy 	if ((pac = amd_pcpu[cpu]) == NULL)
1001e829eb6dSJoseph Koshy 		return (0);
1002ebccf1e3SJoseph Koshy 
1003e829eb6dSJoseph Koshy 	amd_pcpu[cpu] = NULL;
1004ebccf1e3SJoseph Koshy 
1005680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
1006e829eb6dSJoseph Koshy 	for (i = 0; i < AMD_NPMCS; i++) {
1007e829eb6dSJoseph Koshy 		KASSERT(pac->pc_amdpmcs[i].phw_pmc == NULL,
1008ebccf1e3SJoseph Koshy 		    ("[amd,%d] CPU%d/PMC%d in use", __LINE__, cpu, i));
1009e67c0426SAndriy Gapon 		KASSERT(AMD_PMC_IS_STOPPED(AMD_PMC_EVSEL_0 + i),
1010ebccf1e3SJoseph Koshy 		    ("[amd,%d] CPU%d/PMC%d not stopped", __LINE__, cpu, i));
1011ebccf1e3SJoseph Koshy 	}
1012ebccf1e3SJoseph Koshy #endif
1013ebccf1e3SJoseph Koshy 
1014e829eb6dSJoseph Koshy 	pc = pmc_pcpu[cpu];
1015e829eb6dSJoseph Koshy 	KASSERT(pc != NULL, ("[amd,%d] NULL per-cpu state", __LINE__));
1016e829eb6dSJoseph Koshy 
1017e829eb6dSJoseph Koshy #if	defined(__amd64__)
1018e829eb6dSJoseph Koshy 	classindex = PMC_MDEP_CLASS_INDEX_K8;
1019e829eb6dSJoseph Koshy #elif	defined(__i386__)
1020e829eb6dSJoseph Koshy 	classindex = md->pmd_cputype == PMC_CPU_AMD_K8 ? PMC_MDEP_CLASS_INDEX_K8 :
1021e829eb6dSJoseph Koshy 	    PMC_MDEP_CLASS_INDEX_K7;
1022e829eb6dSJoseph Koshy #endif
1023e829eb6dSJoseph Koshy 	first_ri = md->pmd_classdep[classindex].pcd_ri;
1024e829eb6dSJoseph Koshy 
1025e829eb6dSJoseph Koshy 	/*
1026e829eb6dSJoseph Koshy 	 * Reset pointers in the MI 'per-cpu' state.
1027e829eb6dSJoseph Koshy 	 */
1028e829eb6dSJoseph Koshy 	for (i = 0; i < AMD_NPMCS; i++) {
1029e829eb6dSJoseph Koshy 		pc->pc_hwpmcs[i + first_ri] = NULL;
1030e829eb6dSJoseph Koshy 	}
1031e829eb6dSJoseph Koshy 
1032e829eb6dSJoseph Koshy 
1033e829eb6dSJoseph Koshy 	free(pac, M_PMC);
1034e829eb6dSJoseph Koshy 
1035e829eb6dSJoseph Koshy 	return (0);
1036ebccf1e3SJoseph Koshy }
1037ebccf1e3SJoseph Koshy 
1038ebccf1e3SJoseph Koshy /*
1039ebccf1e3SJoseph Koshy  * Initialize ourselves.
1040ebccf1e3SJoseph Koshy  */
1041ebccf1e3SJoseph Koshy 
1042ebccf1e3SJoseph Koshy struct pmc_mdep *
1043ebccf1e3SJoseph Koshy pmc_amd_initialize(void)
1044ebccf1e3SJoseph Koshy {
1045f5f9340bSFabien Thomas 	int classindex, error, i, ncpus;
1046e829eb6dSJoseph Koshy 	struct pmc_classdep *pcd;
1047f263522aSJoseph Koshy 	enum pmc_cputype cputype;
1048ebccf1e3SJoseph Koshy 	struct pmc_mdep *pmc_mdep;
1049e829eb6dSJoseph Koshy 	enum pmc_class class;
1050ef013ceeSRyan Moeller 	int family, model, stepping;
1051f263522aSJoseph Koshy 	char *name;
1052ebccf1e3SJoseph Koshy 
1053f263522aSJoseph Koshy 	/*
1054f263522aSJoseph Koshy 	 * The presence of hardware performance counters on the AMD
1055f263522aSJoseph Koshy 	 * Athlon, Duron or later processors, is _not_ indicated by
1056f263522aSJoseph Koshy 	 * any of the processor feature flags set by the 'CPUID'
1057f263522aSJoseph Koshy 	 * instruction, so we only check the 'instruction family'
1058f263522aSJoseph Koshy 	 * field returned by CPUID for instruction family >= 6.
1059f263522aSJoseph Koshy 	 */
1060ebccf1e3SJoseph Koshy 
106154bad7c6SJoseph Koshy 	name = NULL;
1062ef013ceeSRyan Moeller 	family = CPUID_TO_FAMILY(cpu_id);
1063ef013ceeSRyan Moeller 	model = CPUID_TO_MODEL(cpu_id);
1064ef013ceeSRyan Moeller 	stepping = CPUID_TO_STEPPING(cpu_id);
1065ef013ceeSRyan Moeller 
1066ef013ceeSRyan Moeller 	if (family == 0x18)
10671791cad0SAlexander Motin 		snprintf(pmc_cpuid, sizeof(pmc_cpuid), "HygonGenuine-%d-%02X-%X",
1068ef013ceeSRyan Moeller 		    family, model, stepping);
1069ef013ceeSRyan Moeller 	else
1070ef013ceeSRyan Moeller 		snprintf(pmc_cpuid, sizeof(pmc_cpuid), "AuthenticAMD-%d-%02X-%X",
1071ef013ceeSRyan Moeller 		    family, model, stepping);
107281eb4dcfSMatt Macy 
1073f263522aSJoseph Koshy 	switch (cpu_id & 0xF00) {
1074e829eb6dSJoseph Koshy #if	defined(__i386__)
1075f263522aSJoseph Koshy 	case 0x600:		/* Athlon(tm) processor */
1076e829eb6dSJoseph Koshy 		classindex = PMC_MDEP_CLASS_INDEX_K7;
1077f263522aSJoseph Koshy 		cputype = PMC_CPU_AMD_K7;
1078f263522aSJoseph Koshy 		class = PMC_CLASS_K7;
1079f263522aSJoseph Koshy 		name = "K7";
1080f263522aSJoseph Koshy 		break;
1081e829eb6dSJoseph Koshy #endif
1082f263522aSJoseph Koshy 	case 0xF00:		/* Athlon64/Opteron processor */
1083e829eb6dSJoseph Koshy 		classindex = PMC_MDEP_CLASS_INDEX_K8;
1084f263522aSJoseph Koshy 		cputype = PMC_CPU_AMD_K8;
1085f263522aSJoseph Koshy 		class = PMC_CLASS_K8;
1086f263522aSJoseph Koshy 		name = "K8";
1087f263522aSJoseph Koshy 		break;
1088f263522aSJoseph Koshy 
1089b38c0519SDimitry Andric 	default:
109081eb4dcfSMatt Macy 		(void) printf("pmc: Unknown AMD CPU %x %d-%d.\n", cpu_id, (cpu_id & 0xF00) >> 8, model);
1091ebccf1e3SJoseph Koshy 		return NULL;
1092f263522aSJoseph Koshy 	}
1093f263522aSJoseph Koshy 
1094680f1afdSJohn Baldwin #ifdef	HWPMC_DEBUG
1095f263522aSJoseph Koshy 	amd_pmc_class = class;
1096f263522aSJoseph Koshy #endif
1097ebccf1e3SJoseph Koshy 
1098e829eb6dSJoseph Koshy 	/*
1099e829eb6dSJoseph Koshy 	 * Allocate space for pointers to PMC HW descriptors and for
1100e829eb6dSJoseph Koshy 	 * the MDEP structure used by MI code.
1101e829eb6dSJoseph Koshy 	 */
1102e829eb6dSJoseph Koshy 	amd_pcpu = malloc(sizeof(struct amd_cpu *) * pmc_cpu_max(), M_PMC,
1103e829eb6dSJoseph Koshy 	    M_WAITOK|M_ZERO);
1104e829eb6dSJoseph Koshy 
1105e829eb6dSJoseph Koshy 	/*
1106e829eb6dSJoseph Koshy 	 * These processors have two classes of PMCs: the TSC and
1107e829eb6dSJoseph Koshy 	 * programmable PMCs.
1108e829eb6dSJoseph Koshy 	 */
1109f5f9340bSFabien Thomas 	pmc_mdep = pmc_mdep_alloc(2);
1110ebccf1e3SJoseph Koshy 
1111f263522aSJoseph Koshy 	pmc_mdep->pmd_cputype = cputype;
1112ebccf1e3SJoseph Koshy 
1113e829eb6dSJoseph Koshy 	ncpus = pmc_cpu_max();
1114c5153e19SJoseph Koshy 
1115e829eb6dSJoseph Koshy 	/* Initialize the TSC. */
1116e829eb6dSJoseph Koshy 	error = pmc_tsc_initialize(pmc_mdep, ncpus);
1117e829eb6dSJoseph Koshy 	if (error)
1118e829eb6dSJoseph Koshy 		goto error;
1119c5153e19SJoseph Koshy 
1120e829eb6dSJoseph Koshy 	/* Initialize AMD K7 and K8 PMC handling. */
1121e829eb6dSJoseph Koshy 	pcd = &pmc_mdep->pmd_classdep[classindex];
1122c5153e19SJoseph Koshy 
1123e829eb6dSJoseph Koshy 	pcd->pcd_caps		= AMD_PMC_CAPS;
1124e829eb6dSJoseph Koshy 	pcd->pcd_class		= class;
1125e829eb6dSJoseph Koshy 	pcd->pcd_num		= AMD_NPMCS;
1126e829eb6dSJoseph Koshy 	pcd->pcd_ri		= pmc_mdep->pmd_npmc;
1127e829eb6dSJoseph Koshy 	pcd->pcd_width		= 48;
1128ebccf1e3SJoseph Koshy 
1129f263522aSJoseph Koshy 	/* fill in the correct pmc name and class */
1130e829eb6dSJoseph Koshy 	for (i = 0; i < AMD_NPMCS; i++) {
1131f263522aSJoseph Koshy 		(void) snprintf(amd_pmcdesc[i].pm_descr.pd_name,
1132f263522aSJoseph Koshy 		    sizeof(amd_pmcdesc[i].pm_descr.pd_name), "%s-%d",
11338cd64ec8SJoseph Koshy 		    name, i);
1134f263522aSJoseph Koshy 		amd_pmcdesc[i].pm_descr.pd_class = class;
1135f263522aSJoseph Koshy 	}
1136f263522aSJoseph Koshy 
1137e829eb6dSJoseph Koshy 	pcd->pcd_allocate_pmc	= amd_allocate_pmc;
1138e829eb6dSJoseph Koshy 	pcd->pcd_config_pmc	= amd_config_pmc;
1139e829eb6dSJoseph Koshy 	pcd->pcd_describe	= amd_describe;
1140e829eb6dSJoseph Koshy 	pcd->pcd_get_config	= amd_get_config;
1141e829eb6dSJoseph Koshy 	pcd->pcd_get_msr	= amd_get_msr;
1142e829eb6dSJoseph Koshy 	pcd->pcd_pcpu_fini	= amd_pcpu_fini;
1143e829eb6dSJoseph Koshy 	pcd->pcd_pcpu_init	= amd_pcpu_init;
1144e829eb6dSJoseph Koshy 	pcd->pcd_read_pmc	= amd_read_pmc;
1145e829eb6dSJoseph Koshy 	pcd->pcd_release_pmc	= amd_release_pmc;
1146e829eb6dSJoseph Koshy 	pcd->pcd_start_pmc	= amd_start_pmc;
1147e829eb6dSJoseph Koshy 	pcd->pcd_stop_pmc	= amd_stop_pmc;
1148e829eb6dSJoseph Koshy 	pcd->pcd_write_pmc	= amd_write_pmc;
1149e829eb6dSJoseph Koshy 
1150e829eb6dSJoseph Koshy 	pmc_mdep->pmd_intr	= amd_intr;
1151ebccf1e3SJoseph Koshy 	pmc_mdep->pmd_switch_in = amd_switch_in;
1152ebccf1e3SJoseph Koshy 	pmc_mdep->pmd_switch_out = amd_switch_out;
1153e829eb6dSJoseph Koshy 
1154e829eb6dSJoseph Koshy 	pmc_mdep->pmd_npmc     += AMD_NPMCS;
1155ebccf1e3SJoseph Koshy 
11564a3690dfSJohn Baldwin 	PMCDBG0(MDP,INI,0,"amd-initialize");
1157ebccf1e3SJoseph Koshy 
1158e829eb6dSJoseph Koshy 	return (pmc_mdep);
1159e829eb6dSJoseph Koshy 
1160e829eb6dSJoseph Koshy   error:
1161e829eb6dSJoseph Koshy 	if (error) {
1162e829eb6dSJoseph Koshy 		free(pmc_mdep, M_PMC);
1163e829eb6dSJoseph Koshy 		pmc_mdep = NULL;
1164e829eb6dSJoseph Koshy 	}
1165e829eb6dSJoseph Koshy 
1166e829eb6dSJoseph Koshy 	return (NULL);
1167e829eb6dSJoseph Koshy }
1168e829eb6dSJoseph Koshy 
1169e829eb6dSJoseph Koshy /*
1170e829eb6dSJoseph Koshy  * Finalization code for AMD CPUs.
1171e829eb6dSJoseph Koshy  */
1172e829eb6dSJoseph Koshy 
1173e829eb6dSJoseph Koshy void
1174e829eb6dSJoseph Koshy pmc_amd_finalize(struct pmc_mdep *md)
1175e829eb6dSJoseph Koshy {
1176*90a6ea5cSMitchell Horne 	PMCDBG0(MDP, INI, 1, "amd-finalize");
1177e829eb6dSJoseph Koshy 
1178e829eb6dSJoseph Koshy 	pmc_tsc_finalize(md);
1179e829eb6dSJoseph Koshy 
1180*90a6ea5cSMitchell Horne 	for (int i = 0; i < pmc_cpu_max(); i++)
1181*90a6ea5cSMitchell Horne 		KASSERT(amd_pcpu[i] == NULL,
1182*90a6ea5cSMitchell Horne 		    ("[amd,%d] non-null pcpu cpu %d", __LINE__, i));
1183e829eb6dSJoseph Koshy 
1184e829eb6dSJoseph Koshy 	free(amd_pcpu, M_PMC);
1185e829eb6dSJoseph Koshy 	amd_pcpu = NULL;
1186ebccf1e3SJoseph Koshy }
1187