xref: /linux/tools/perf/util/pmus.c (revision 520da457f9b883e541c36c8f55e7590848b7aa64)
1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/list.h>
3 #include <linux/zalloc.h>
4 #include <subcmd/pager.h>
5 #include <sys/types.h>
6 #include <dirent.h>
7 #include <pthread.h>
8 #include <string.h>
9 #include <unistd.h>
10 #include "debug.h"
11 #include "evsel.h"
12 #include "pmus.h"
13 #include "pmu.h"
14 #include "print-events.h"
15 
16 /*
17  * core_pmus:  A PMU belongs to core_pmus if it's name is "cpu" or it's sysfs
18  *             directory contains "cpus" file. All PMUs belonging to core_pmus
19  *             must have pmu->is_core=1. If there are more than one PMU in
20  *             this list, perf interprets it as a heterogeneous platform.
21  *             (FWIW, certain ARM platforms having heterogeneous cores uses
22  *             homogeneous PMU, and thus they are treated as homogeneous
23  *             platform by perf because core_pmus will have only one entry)
24  * other_pmus: All other PMUs which are not part of core_pmus list. It doesn't
25  *             matter whether PMU is present per SMT-thread or outside of the
26  *             core in the hw. For e.g., an instance of AMD ibs_fetch// and
27  *             ibs_op// PMUs is present in each hw SMT thread, however they
28  *             are captured under other_pmus. PMUs belonging to other_pmus
29  *             must have pmu->is_core=0 but pmu->is_uncore could be 0 or 1.
30  */
31 static LIST_HEAD(core_pmus);
32 static LIST_HEAD(other_pmus);
33 static bool read_sysfs_core_pmus;
34 static bool read_sysfs_all_pmus;
35 
36 void perf_pmus__destroy(void)
37 {
38 	struct perf_pmu *pmu, *tmp;
39 
40 	list_for_each_entry_safe(pmu, tmp, &core_pmus, list) {
41 		list_del(&pmu->list);
42 
43 		perf_pmu__delete(pmu);
44 	}
45 	list_for_each_entry_safe(pmu, tmp, &other_pmus, list) {
46 		list_del(&pmu->list);
47 
48 		perf_pmu__delete(pmu);
49 	}
50 	read_sysfs_core_pmus = false;
51 	read_sysfs_all_pmus = false;
52 }
53 
54 static struct perf_pmu *pmu_find(const char *name)
55 {
56 	struct perf_pmu *pmu;
57 
58 	list_for_each_entry(pmu, &core_pmus, list) {
59 		if (!strcmp(pmu->name, name) ||
60 		    (pmu->alias_name && !strcmp(pmu->alias_name, name)))
61 			return pmu;
62 	}
63 	list_for_each_entry(pmu, &other_pmus, list) {
64 		if (!strcmp(pmu->name, name) ||
65 		    (pmu->alias_name && !strcmp(pmu->alias_name, name)))
66 			return pmu;
67 	}
68 
69 	return NULL;
70 }
71 
72 struct perf_pmu *perf_pmus__find(const char *name)
73 {
74 	struct perf_pmu *pmu;
75 	int dirfd;
76 	bool core_pmu;
77 
78 	/*
79 	 * Once PMU is loaded it stays in the list,
80 	 * so we keep us from multiple reading/parsing
81 	 * the pmu format definitions.
82 	 */
83 	pmu = pmu_find(name);
84 	if (pmu)
85 		return pmu;
86 
87 	if (read_sysfs_all_pmus)
88 		return NULL;
89 
90 	core_pmu = is_pmu_core(name);
91 	if (core_pmu && read_sysfs_core_pmus)
92 		return NULL;
93 
94 	dirfd = perf_pmu__event_source_devices_fd();
95 	pmu = perf_pmu__lookup(core_pmu ? &core_pmus : &other_pmus, dirfd, name);
96 	close(dirfd);
97 
98 	return pmu;
99 }
100 
101 static struct perf_pmu *perf_pmu__find2(int dirfd, const char *name)
102 {
103 	struct perf_pmu *pmu;
104 	bool core_pmu;
105 
106 	/*
107 	 * Once PMU is loaded it stays in the list,
108 	 * so we keep us from multiple reading/parsing
109 	 * the pmu format definitions.
110 	 */
111 	pmu = pmu_find(name);
112 	if (pmu)
113 		return pmu;
114 
115 	if (read_sysfs_all_pmus)
116 		return NULL;
117 
118 	core_pmu = is_pmu_core(name);
119 	if (core_pmu && read_sysfs_core_pmus)
120 		return NULL;
121 
122 	return perf_pmu__lookup(core_pmu ? &core_pmus : &other_pmus, dirfd, name);
123 }
124 
125 /* Add all pmus in sysfs to pmu list: */
126 static void pmu_read_sysfs(bool core_only)
127 {
128 	int fd;
129 	DIR *dir;
130 	struct dirent *dent;
131 
132 	if (read_sysfs_all_pmus || (core_only && read_sysfs_core_pmus))
133 		return;
134 
135 	fd = perf_pmu__event_source_devices_fd();
136 	if (fd < 0)
137 		return;
138 
139 	dir = fdopendir(fd);
140 	if (!dir) {
141 		close(fd);
142 		return;
143 	}
144 
145 	while ((dent = readdir(dir))) {
146 		if (!strcmp(dent->d_name, ".") || !strcmp(dent->d_name, ".."))
147 			continue;
148 		if (core_only && !is_pmu_core(dent->d_name))
149 			continue;
150 		/* add to static LIST_HEAD(core_pmus) or LIST_HEAD(other_pmus): */
151 		perf_pmu__find2(fd, dent->d_name);
152 	}
153 
154 	closedir(dir);
155 	if (list_empty(&core_pmus)) {
156 		if (!perf_pmu__create_placeholder_core_pmu(&core_pmus))
157 			pr_err("Failure to set up any core PMUs\n");
158 	}
159 	if (!list_empty(&core_pmus)) {
160 		read_sysfs_core_pmus = true;
161 		if (!core_only)
162 			read_sysfs_all_pmus = true;
163 	}
164 }
165 
166 static struct perf_pmu *__perf_pmus__find_by_type(unsigned int type)
167 {
168 	struct perf_pmu *pmu;
169 
170 	list_for_each_entry(pmu, &core_pmus, list) {
171 		if (pmu->type == type)
172 			return pmu;
173 	}
174 
175 	list_for_each_entry(pmu, &other_pmus, list) {
176 		if (pmu->type == type)
177 			return pmu;
178 	}
179 	return NULL;
180 }
181 
182 struct perf_pmu *perf_pmus__find_by_type(unsigned int type)
183 {
184 	struct perf_pmu *pmu = __perf_pmus__find_by_type(type);
185 
186 	if (pmu || read_sysfs_all_pmus)
187 		return pmu;
188 
189 	pmu_read_sysfs(/*core_only=*/false);
190 	pmu = __perf_pmus__find_by_type(type);
191 	return pmu;
192 }
193 
194 /*
195  * pmu iterator: If pmu is NULL, we start at the begin, otherwise return the
196  * next pmu. Returns NULL on end.
197  */
198 struct perf_pmu *perf_pmus__scan(struct perf_pmu *pmu)
199 {
200 	bool use_core_pmus = !pmu || pmu->is_core;
201 
202 	if (!pmu) {
203 		pmu_read_sysfs(/*core_only=*/false);
204 		pmu = list_prepare_entry(pmu, &core_pmus, list);
205 	}
206 	if (use_core_pmus) {
207 		list_for_each_entry_continue(pmu, &core_pmus, list)
208 			return pmu;
209 
210 		pmu = NULL;
211 		pmu = list_prepare_entry(pmu, &other_pmus, list);
212 	}
213 	list_for_each_entry_continue(pmu, &other_pmus, list)
214 		return pmu;
215 	return NULL;
216 }
217 
218 struct perf_pmu *perf_pmus__scan_core(struct perf_pmu *pmu)
219 {
220 	if (!pmu) {
221 		pmu_read_sysfs(/*core_only=*/true);
222 		pmu = list_prepare_entry(pmu, &core_pmus, list);
223 	}
224 	list_for_each_entry_continue(pmu, &core_pmus, list)
225 		return pmu;
226 
227 	return NULL;
228 }
229 
230 const struct perf_pmu *perf_pmus__pmu_for_pmu_filter(const char *str)
231 {
232 	struct perf_pmu *pmu = NULL;
233 
234 	while ((pmu = perf_pmus__scan(pmu)) != NULL) {
235 		if (!strcmp(pmu->name, str))
236 			return pmu;
237 		/* Ignore "uncore_" prefix. */
238 		if (!strncmp(pmu->name, "uncore_", 7)) {
239 			if (!strcmp(pmu->name + 7, str))
240 				return pmu;
241 		}
242 		/* Ignore "cpu_" prefix on Intel hybrid PMUs. */
243 		if (!strncmp(pmu->name, "cpu_", 4)) {
244 			if (!strcmp(pmu->name + 4, str))
245 				return pmu;
246 		}
247 	}
248 	return NULL;
249 }
250 
251 int __weak perf_pmus__num_mem_pmus(void)
252 {
253 	/* All core PMUs are for mem events. */
254 	return perf_pmus__num_core_pmus();
255 }
256 
257 /** Struct for ordering events as output in perf list. */
258 struct sevent {
259 	/** PMU for event. */
260 	const struct perf_pmu *pmu;
261 	const char *name;
262 	const char* alias;
263 	const char *scale_unit;
264 	const char *desc;
265 	const char *long_desc;
266 	const char *encoding_desc;
267 	const char *topic;
268 	const char *pmu_name;
269 	bool deprecated;
270 };
271 
272 static int cmp_sevent(const void *a, const void *b)
273 {
274 	const struct sevent *as = a;
275 	const struct sevent *bs = b;
276 	bool a_iscpu, b_iscpu;
277 	int ret;
278 
279 	/* Put extra events last. */
280 	if (!!as->desc != !!bs->desc)
281 		return !!as->desc - !!bs->desc;
282 
283 	/* Order by topics. */
284 	ret = strcmp(as->topic ?: "", bs->topic ?: "");
285 	if (ret)
286 		return ret;
287 
288 	/* Order CPU core events to be first */
289 	a_iscpu = as->pmu ? as->pmu->is_core : true;
290 	b_iscpu = bs->pmu ? bs->pmu->is_core : true;
291 	if (a_iscpu != b_iscpu)
292 		return a_iscpu ? -1 : 1;
293 
294 	/* Order by PMU name. */
295 	if (as->pmu != bs->pmu) {
296 		ret = strcmp(as->pmu_name ?: "", bs->pmu_name ?: "");
297 		if (ret)
298 			return ret;
299 	}
300 
301 	/* Order by event name. */
302 	return strcmp(as->name, bs->name);
303 }
304 
305 static bool pmu_alias_is_duplicate(struct sevent *a, struct sevent *b)
306 {
307 	/* Different names -> never duplicates */
308 	if (strcmp(a->name ?: "//", b->name ?: "//"))
309 		return false;
310 
311 	/* Don't remove duplicates for different PMUs */
312 	return strcmp(a->pmu_name, b->pmu_name) == 0;
313 }
314 
315 struct events_callback_state {
316 	struct sevent *aliases;
317 	size_t aliases_len;
318 	size_t index;
319 };
320 
321 static int perf_pmus__print_pmu_events__callback(void *vstate,
322 						struct pmu_event_info *info)
323 {
324 	struct events_callback_state *state = vstate;
325 	struct sevent *s;
326 
327 	if (state->index >= state->aliases_len) {
328 		pr_err("Unexpected event %s/%s/\n", info->pmu->name, info->name);
329 		return 1;
330 	}
331 	s = &state->aliases[state->index];
332 	s->pmu = info->pmu;
333 #define COPY_STR(str) s->str = info->str ? strdup(info->str) : NULL
334 	COPY_STR(name);
335 	COPY_STR(alias);
336 	COPY_STR(scale_unit);
337 	COPY_STR(desc);
338 	COPY_STR(long_desc);
339 	COPY_STR(encoding_desc);
340 	COPY_STR(topic);
341 	COPY_STR(pmu_name);
342 #undef COPY_STR
343 	s->deprecated = info->deprecated;
344 	state->index++;
345 	return 0;
346 }
347 
348 void perf_pmus__print_pmu_events(const struct print_callbacks *print_cb, void *print_state)
349 {
350 	struct perf_pmu *pmu;
351 	int printed = 0;
352 	int len;
353 	struct sevent *aliases;
354 	struct events_callback_state state;
355 
356 	pmu = NULL;
357 	len = 0;
358 	while ((pmu = perf_pmus__scan(pmu)) != NULL)
359 		len += perf_pmu__num_events(pmu);
360 
361 	aliases = zalloc(sizeof(struct sevent) * len);
362 	if (!aliases) {
363 		pr_err("FATAL: not enough memory to print PMU events\n");
364 		return;
365 	}
366 	pmu = NULL;
367 	state = (struct events_callback_state) {
368 		.aliases = aliases,
369 		.aliases_len = len,
370 		.index = 0,
371 	};
372 	while ((pmu = perf_pmus__scan(pmu)) != NULL) {
373 		perf_pmu__for_each_event(pmu, &state, perf_pmus__print_pmu_events__callback);
374 	}
375 	qsort(aliases, len, sizeof(struct sevent), cmp_sevent);
376 	for (int j = 0; j < len; j++) {
377 		/* Skip duplicates */
378 		if (j > 0 && pmu_alias_is_duplicate(&aliases[j], &aliases[j - 1]))
379 			continue;
380 
381 		print_cb->print_event(print_state,
382 				aliases[j].pmu_name,
383 				aliases[j].topic,
384 				aliases[j].name,
385 				aliases[j].alias,
386 				aliases[j].scale_unit,
387 				aliases[j].deprecated,
388 				"Kernel PMU event",
389 				aliases[j].desc,
390 				aliases[j].long_desc,
391 				aliases[j].encoding_desc);
392 		zfree(&aliases[j].name);
393 		zfree(&aliases[j].alias);
394 		zfree(&aliases[j].scale_unit);
395 		zfree(&aliases[j].desc);
396 		zfree(&aliases[j].long_desc);
397 		zfree(&aliases[j].encoding_desc);
398 		zfree(&aliases[j].topic);
399 		zfree(&aliases[j].pmu_name);
400 	}
401 	if (printed && pager_in_use())
402 		printf("\n");
403 
404 	zfree(&aliases);
405 }
406 
407 bool perf_pmus__have_event(const char *pname, const char *name)
408 {
409 	struct perf_pmu *pmu = perf_pmus__find(pname);
410 
411 	return pmu && perf_pmu__have_event(pmu, name);
412 }
413 
414 int perf_pmus__num_core_pmus(void)
415 {
416 	static int count;
417 
418 	if (!count) {
419 		struct perf_pmu *pmu = NULL;
420 
421 		while ((pmu = perf_pmus__scan_core(pmu)) != NULL)
422 			count++;
423 	}
424 	return count;
425 }
426 
427 static bool __perf_pmus__supports_extended_type(void)
428 {
429 	struct perf_pmu *pmu = NULL;
430 
431 	if (perf_pmus__num_core_pmus() <= 1)
432 		return false;
433 
434 	while ((pmu = perf_pmus__scan_core(pmu)) != NULL) {
435 		if (!is_event_supported(PERF_TYPE_HARDWARE, PERF_COUNT_HW_CPU_CYCLES | ((__u64)pmu->type << PERF_PMU_TYPE_SHIFT)))
436 			return false;
437 	}
438 
439 	return true;
440 }
441 
442 static bool perf_pmus__do_support_extended_type;
443 
444 static void perf_pmus__init_supports_extended_type(void)
445 {
446 	perf_pmus__do_support_extended_type = __perf_pmus__supports_extended_type();
447 }
448 
449 bool perf_pmus__supports_extended_type(void)
450 {
451 	static pthread_once_t extended_type_once = PTHREAD_ONCE_INIT;
452 
453 	pthread_once(&extended_type_once, perf_pmus__init_supports_extended_type);
454 
455 	return perf_pmus__do_support_extended_type;
456 }
457 
458 char *perf_pmus__default_pmu_name(void)
459 {
460 	int fd;
461 	DIR *dir;
462 	struct dirent *dent;
463 	char *result = NULL;
464 
465 	if (!list_empty(&core_pmus))
466 		return strdup(list_first_entry(&core_pmus, struct perf_pmu, list)->name);
467 
468 	fd = perf_pmu__event_source_devices_fd();
469 	if (fd < 0)
470 		return strdup("cpu");
471 
472 	dir = fdopendir(fd);
473 	if (!dir) {
474 		close(fd);
475 		return strdup("cpu");
476 	}
477 
478 	while ((dent = readdir(dir))) {
479 		if (!strcmp(dent->d_name, ".") || !strcmp(dent->d_name, ".."))
480 			continue;
481 		if (is_pmu_core(dent->d_name)) {
482 			result = strdup(dent->d_name);
483 			break;
484 		}
485 	}
486 
487 	closedir(dir);
488 	return result ?: strdup("cpu");
489 }
490 
491 struct perf_pmu *evsel__find_pmu(const struct evsel *evsel)
492 {
493 	struct perf_pmu *pmu = evsel->pmu;
494 
495 	if (!pmu) {
496 		pmu = perf_pmus__find_by_type(evsel->core.attr.type);
497 		((struct evsel *)evsel)->pmu = pmu;
498 	}
499 	return pmu;
500 }
501