1 // SPDX-License-Identifier: GPL-2.0 2 #include <linux/list.h> 3 #include <linux/zalloc.h> 4 #include <subcmd/pager.h> 5 #include <sys/types.h> 6 #include <dirent.h> 7 #include <pthread.h> 8 #include <string.h> 9 #include <unistd.h> 10 #include "debug.h" 11 #include "evsel.h" 12 #include "pmus.h" 13 #include "pmu.h" 14 #include "print-events.h" 15 16 /* 17 * core_pmus: A PMU belongs to core_pmus if it's name is "cpu" or it's sysfs 18 * directory contains "cpus" file. All PMUs belonging to core_pmus 19 * must have pmu->is_core=1. If there are more than one PMU in 20 * this list, perf interprets it as a heterogeneous platform. 21 * (FWIW, certain ARM platforms having heterogeneous cores uses 22 * homogeneous PMU, and thus they are treated as homogeneous 23 * platform by perf because core_pmus will have only one entry) 24 * other_pmus: All other PMUs which are not part of core_pmus list. It doesn't 25 * matter whether PMU is present per SMT-thread or outside of the 26 * core in the hw. For e.g., an instance of AMD ibs_fetch// and 27 * ibs_op// PMUs is present in each hw SMT thread, however they 28 * are captured under other_pmus. PMUs belonging to other_pmus 29 * must have pmu->is_core=0 but pmu->is_uncore could be 0 or 1. 30 */ 31 static LIST_HEAD(core_pmus); 32 static LIST_HEAD(other_pmus); 33 static bool read_sysfs_core_pmus; 34 static bool read_sysfs_all_pmus; 35 36 void perf_pmus__destroy(void) 37 { 38 struct perf_pmu *pmu, *tmp; 39 40 list_for_each_entry_safe(pmu, tmp, &core_pmus, list) { 41 list_del(&pmu->list); 42 43 perf_pmu__delete(pmu); 44 } 45 list_for_each_entry_safe(pmu, tmp, &other_pmus, list) { 46 list_del(&pmu->list); 47 48 perf_pmu__delete(pmu); 49 } 50 read_sysfs_core_pmus = false; 51 read_sysfs_all_pmus = false; 52 } 53 54 static struct perf_pmu *pmu_find(const char *name) 55 { 56 struct perf_pmu *pmu; 57 58 list_for_each_entry(pmu, &core_pmus, list) { 59 if (!strcmp(pmu->name, name) || 60 (pmu->alias_name && !strcmp(pmu->alias_name, name))) 61 return pmu; 62 } 63 list_for_each_entry(pmu, &other_pmus, list) { 64 if (!strcmp(pmu->name, name) || 65 (pmu->alias_name && !strcmp(pmu->alias_name, name))) 66 return pmu; 67 } 68 69 return NULL; 70 } 71 72 struct perf_pmu *perf_pmus__find(const char *name) 73 { 74 struct perf_pmu *pmu; 75 int dirfd; 76 bool core_pmu; 77 78 /* 79 * Once PMU is loaded it stays in the list, 80 * so we keep us from multiple reading/parsing 81 * the pmu format definitions. 82 */ 83 pmu = pmu_find(name); 84 if (pmu) 85 return pmu; 86 87 if (read_sysfs_all_pmus) 88 return NULL; 89 90 core_pmu = is_pmu_core(name); 91 if (core_pmu && read_sysfs_core_pmus) 92 return NULL; 93 94 dirfd = perf_pmu__event_source_devices_fd(); 95 pmu = perf_pmu__lookup(core_pmu ? &core_pmus : &other_pmus, dirfd, name); 96 close(dirfd); 97 98 return pmu; 99 } 100 101 static struct perf_pmu *perf_pmu__find2(int dirfd, const char *name) 102 { 103 struct perf_pmu *pmu; 104 bool core_pmu; 105 106 /* 107 * Once PMU is loaded it stays in the list, 108 * so we keep us from multiple reading/parsing 109 * the pmu format definitions. 110 */ 111 pmu = pmu_find(name); 112 if (pmu) 113 return pmu; 114 115 if (read_sysfs_all_pmus) 116 return NULL; 117 118 core_pmu = is_pmu_core(name); 119 if (core_pmu && read_sysfs_core_pmus) 120 return NULL; 121 122 return perf_pmu__lookup(core_pmu ? &core_pmus : &other_pmus, dirfd, name); 123 } 124 125 /* Add all pmus in sysfs to pmu list: */ 126 static void pmu_read_sysfs(bool core_only) 127 { 128 int fd; 129 DIR *dir; 130 struct dirent *dent; 131 132 if (read_sysfs_all_pmus || (core_only && read_sysfs_core_pmus)) 133 return; 134 135 fd = perf_pmu__event_source_devices_fd(); 136 if (fd < 0) 137 return; 138 139 dir = fdopendir(fd); 140 if (!dir) { 141 close(fd); 142 return; 143 } 144 145 while ((dent = readdir(dir))) { 146 if (!strcmp(dent->d_name, ".") || !strcmp(dent->d_name, "..")) 147 continue; 148 if (core_only && !is_pmu_core(dent->d_name)) 149 continue; 150 /* add to static LIST_HEAD(core_pmus) or LIST_HEAD(other_pmus): */ 151 perf_pmu__find2(fd, dent->d_name); 152 } 153 154 closedir(dir); 155 if (list_empty(&core_pmus)) { 156 if (!perf_pmu__create_placeholder_core_pmu(&core_pmus)) 157 pr_err("Failure to set up any core PMUs\n"); 158 } 159 if (!list_empty(&core_pmus)) { 160 read_sysfs_core_pmus = true; 161 if (!core_only) 162 read_sysfs_all_pmus = true; 163 } 164 } 165 166 static struct perf_pmu *__perf_pmus__find_by_type(unsigned int type) 167 { 168 struct perf_pmu *pmu; 169 170 list_for_each_entry(pmu, &core_pmus, list) { 171 if (pmu->type == type) 172 return pmu; 173 } 174 175 list_for_each_entry(pmu, &other_pmus, list) { 176 if (pmu->type == type) 177 return pmu; 178 } 179 return NULL; 180 } 181 182 struct perf_pmu *perf_pmus__find_by_type(unsigned int type) 183 { 184 struct perf_pmu *pmu = __perf_pmus__find_by_type(type); 185 186 if (pmu || read_sysfs_all_pmus) 187 return pmu; 188 189 pmu_read_sysfs(/*core_only=*/false); 190 pmu = __perf_pmus__find_by_type(type); 191 return pmu; 192 } 193 194 /* 195 * pmu iterator: If pmu is NULL, we start at the begin, otherwise return the 196 * next pmu. Returns NULL on end. 197 */ 198 struct perf_pmu *perf_pmus__scan(struct perf_pmu *pmu) 199 { 200 bool use_core_pmus = !pmu || pmu->is_core; 201 202 if (!pmu) { 203 pmu_read_sysfs(/*core_only=*/false); 204 pmu = list_prepare_entry(pmu, &core_pmus, list); 205 } 206 if (use_core_pmus) { 207 list_for_each_entry_continue(pmu, &core_pmus, list) 208 return pmu; 209 210 pmu = NULL; 211 pmu = list_prepare_entry(pmu, &other_pmus, list); 212 } 213 list_for_each_entry_continue(pmu, &other_pmus, list) 214 return pmu; 215 return NULL; 216 } 217 218 struct perf_pmu *perf_pmus__scan_core(struct perf_pmu *pmu) 219 { 220 if (!pmu) { 221 pmu_read_sysfs(/*core_only=*/true); 222 pmu = list_prepare_entry(pmu, &core_pmus, list); 223 } 224 list_for_each_entry_continue(pmu, &core_pmus, list) 225 return pmu; 226 227 return NULL; 228 } 229 230 const struct perf_pmu *perf_pmus__pmu_for_pmu_filter(const char *str) 231 { 232 struct perf_pmu *pmu = NULL; 233 234 while ((pmu = perf_pmus__scan(pmu)) != NULL) { 235 if (!strcmp(pmu->name, str)) 236 return pmu; 237 /* Ignore "uncore_" prefix. */ 238 if (!strncmp(pmu->name, "uncore_", 7)) { 239 if (!strcmp(pmu->name + 7, str)) 240 return pmu; 241 } 242 /* Ignore "cpu_" prefix on Intel hybrid PMUs. */ 243 if (!strncmp(pmu->name, "cpu_", 4)) { 244 if (!strcmp(pmu->name + 4, str)) 245 return pmu; 246 } 247 } 248 return NULL; 249 } 250 251 int __weak perf_pmus__num_mem_pmus(void) 252 { 253 /* All core PMUs are for mem events. */ 254 return perf_pmus__num_core_pmus(); 255 } 256 257 /** Struct for ordering events as output in perf list. */ 258 struct sevent { 259 /** PMU for event. */ 260 const struct perf_pmu *pmu; 261 const char *name; 262 const char* alias; 263 const char *scale_unit; 264 const char *desc; 265 const char *long_desc; 266 const char *encoding_desc; 267 const char *topic; 268 const char *pmu_name; 269 bool deprecated; 270 }; 271 272 static int cmp_sevent(const void *a, const void *b) 273 { 274 const struct sevent *as = a; 275 const struct sevent *bs = b; 276 bool a_iscpu, b_iscpu; 277 int ret; 278 279 /* Put extra events last. */ 280 if (!!as->desc != !!bs->desc) 281 return !!as->desc - !!bs->desc; 282 283 /* Order by topics. */ 284 ret = strcmp(as->topic ?: "", bs->topic ?: ""); 285 if (ret) 286 return ret; 287 288 /* Order CPU core events to be first */ 289 a_iscpu = as->pmu ? as->pmu->is_core : true; 290 b_iscpu = bs->pmu ? bs->pmu->is_core : true; 291 if (a_iscpu != b_iscpu) 292 return a_iscpu ? -1 : 1; 293 294 /* Order by PMU name. */ 295 if (as->pmu != bs->pmu) { 296 ret = strcmp(as->pmu_name ?: "", bs->pmu_name ?: ""); 297 if (ret) 298 return ret; 299 } 300 301 /* Order by event name. */ 302 return strcmp(as->name, bs->name); 303 } 304 305 static bool pmu_alias_is_duplicate(struct sevent *a, struct sevent *b) 306 { 307 /* Different names -> never duplicates */ 308 if (strcmp(a->name ?: "//", b->name ?: "//")) 309 return false; 310 311 /* Don't remove duplicates for different PMUs */ 312 return strcmp(a->pmu_name, b->pmu_name) == 0; 313 } 314 315 struct events_callback_state { 316 struct sevent *aliases; 317 size_t aliases_len; 318 size_t index; 319 }; 320 321 static int perf_pmus__print_pmu_events__callback(void *vstate, 322 struct pmu_event_info *info) 323 { 324 struct events_callback_state *state = vstate; 325 struct sevent *s; 326 327 if (state->index >= state->aliases_len) { 328 pr_err("Unexpected event %s/%s/\n", info->pmu->name, info->name); 329 return 1; 330 } 331 s = &state->aliases[state->index]; 332 s->pmu = info->pmu; 333 #define COPY_STR(str) s->str = info->str ? strdup(info->str) : NULL 334 COPY_STR(name); 335 COPY_STR(alias); 336 COPY_STR(scale_unit); 337 COPY_STR(desc); 338 COPY_STR(long_desc); 339 COPY_STR(encoding_desc); 340 COPY_STR(topic); 341 COPY_STR(pmu_name); 342 #undef COPY_STR 343 s->deprecated = info->deprecated; 344 state->index++; 345 return 0; 346 } 347 348 void perf_pmus__print_pmu_events(const struct print_callbacks *print_cb, void *print_state) 349 { 350 struct perf_pmu *pmu; 351 int printed = 0; 352 int len; 353 struct sevent *aliases; 354 struct events_callback_state state; 355 356 pmu = NULL; 357 len = 0; 358 while ((pmu = perf_pmus__scan(pmu)) != NULL) 359 len += perf_pmu__num_events(pmu); 360 361 aliases = zalloc(sizeof(struct sevent) * len); 362 if (!aliases) { 363 pr_err("FATAL: not enough memory to print PMU events\n"); 364 return; 365 } 366 pmu = NULL; 367 state = (struct events_callback_state) { 368 .aliases = aliases, 369 .aliases_len = len, 370 .index = 0, 371 }; 372 while ((pmu = perf_pmus__scan(pmu)) != NULL) { 373 perf_pmu__for_each_event(pmu, &state, perf_pmus__print_pmu_events__callback); 374 } 375 qsort(aliases, len, sizeof(struct sevent), cmp_sevent); 376 for (int j = 0; j < len; j++) { 377 /* Skip duplicates */ 378 if (j > 0 && pmu_alias_is_duplicate(&aliases[j], &aliases[j - 1])) 379 continue; 380 381 print_cb->print_event(print_state, 382 aliases[j].pmu_name, 383 aliases[j].topic, 384 aliases[j].name, 385 aliases[j].alias, 386 aliases[j].scale_unit, 387 aliases[j].deprecated, 388 "Kernel PMU event", 389 aliases[j].desc, 390 aliases[j].long_desc, 391 aliases[j].encoding_desc); 392 zfree(&aliases[j].name); 393 zfree(&aliases[j].alias); 394 zfree(&aliases[j].scale_unit); 395 zfree(&aliases[j].desc); 396 zfree(&aliases[j].long_desc); 397 zfree(&aliases[j].encoding_desc); 398 zfree(&aliases[j].topic); 399 zfree(&aliases[j].pmu_name); 400 } 401 if (printed && pager_in_use()) 402 printf("\n"); 403 404 zfree(&aliases); 405 } 406 407 bool perf_pmus__have_event(const char *pname, const char *name) 408 { 409 struct perf_pmu *pmu = perf_pmus__find(pname); 410 411 return pmu && perf_pmu__have_event(pmu, name); 412 } 413 414 int perf_pmus__num_core_pmus(void) 415 { 416 static int count; 417 418 if (!count) { 419 struct perf_pmu *pmu = NULL; 420 421 while ((pmu = perf_pmus__scan_core(pmu)) != NULL) 422 count++; 423 } 424 return count; 425 } 426 427 static bool __perf_pmus__supports_extended_type(void) 428 { 429 struct perf_pmu *pmu = NULL; 430 431 if (perf_pmus__num_core_pmus() <= 1) 432 return false; 433 434 while ((pmu = perf_pmus__scan_core(pmu)) != NULL) { 435 if (!is_event_supported(PERF_TYPE_HARDWARE, PERF_COUNT_HW_CPU_CYCLES | ((__u64)pmu->type << PERF_PMU_TYPE_SHIFT))) 436 return false; 437 } 438 439 return true; 440 } 441 442 static bool perf_pmus__do_support_extended_type; 443 444 static void perf_pmus__init_supports_extended_type(void) 445 { 446 perf_pmus__do_support_extended_type = __perf_pmus__supports_extended_type(); 447 } 448 449 bool perf_pmus__supports_extended_type(void) 450 { 451 static pthread_once_t extended_type_once = PTHREAD_ONCE_INIT; 452 453 pthread_once(&extended_type_once, perf_pmus__init_supports_extended_type); 454 455 return perf_pmus__do_support_extended_type; 456 } 457 458 char *perf_pmus__default_pmu_name(void) 459 { 460 int fd; 461 DIR *dir; 462 struct dirent *dent; 463 char *result = NULL; 464 465 if (!list_empty(&core_pmus)) 466 return strdup(list_first_entry(&core_pmus, struct perf_pmu, list)->name); 467 468 fd = perf_pmu__event_source_devices_fd(); 469 if (fd < 0) 470 return strdup("cpu"); 471 472 dir = fdopendir(fd); 473 if (!dir) { 474 close(fd); 475 return strdup("cpu"); 476 } 477 478 while ((dent = readdir(dir))) { 479 if (!strcmp(dent->d_name, ".") || !strcmp(dent->d_name, "..")) 480 continue; 481 if (is_pmu_core(dent->d_name)) { 482 result = strdup(dent->d_name); 483 break; 484 } 485 } 486 487 closedir(dir); 488 return result ?: strdup("cpu"); 489 } 490 491 struct perf_pmu *evsel__find_pmu(const struct evsel *evsel) 492 { 493 struct perf_pmu *pmu = evsel->pmu; 494 495 if (!pmu) { 496 pmu = perf_pmus__find_by_type(evsel->core.attr.type); 497 ((struct evsel *)evsel)->pmu = pmu; 498 } 499 return pmu; 500 } 501