| /linux/tools/perf/tests/shell/ |
| H A D | stat.sh | 229 pmus=$(ls /sys/bus/event_source/devices/*/cpus 2>/dev/null | wc -l) 230 if [ "$pmus" -lt 1 ] 232 pmus=1 242 if [ "$pmus" -ne "$cycles_events" ] && [ "0" -ne "$cycles_events" ] 244 echo "hybrid test [Found $pmus PMUs but $cycles_events cycles events. Failed]"
|
| /linux/tools/perf/pmu-events/ |
| H A D | empty-pmu-events.c | 2775 const struct pmu_table_entry *pmus; member 2781 const struct pmu_table_entry *pmus; member 2809 .pmus = pmu_events__common, 2813 .pmus = pmu_metrics__common, 2821 .pmus = pmu_events__test_soc_cpu, 2825 .pmus = pmu_metrics__test_soc_cpu, 2846 .pmus = pmu_events__test_soc_sys, 2989 const struct pmu_table_entry *table_pmu = &table->pmus[i]; in pmu_events_table__for_each_event() 3012 const struct pmu_table_entry *table_pmu = &table->pmus[i]; in pmu_events_table__find_event() 3034 const struct pmu_table_entry *table_pmu = &table->pmus[i]; in pmu_events_table__num_events() [all …]
|
| H A D | jevents.py | 536 pmus = set() 548 pmus.add((event.pmu, pmu_name)) 559 for (pmu, tbl_pmu) in sorted(pmus): 594 pmus = set() 604 pmus.add((metric.pmu, pmu_name)) 614 for (pmu, tbl_pmu) in sorted(pmus):
|
| /linux/arch/x86/events/amd/ |
| H A D | uncore.c | 79 struct amd_uncore_pmu *pmus; member 482 pmu = &uncore->pmus[i]; in amd_uncore_ctx_free() 512 pmu = &uncore->pmus[i]; in amd_uncore_ctx_init() 579 pmu = &uncore->pmus[i]; in amd_uncore_ctx_move() 729 uncore->pmus = kzalloc(sizeof(*uncore->pmus), GFP_KERNEL); in amd_uncore_df_ctx_init() 730 if (!uncore->pmus) in amd_uncore_df_ctx_init() 738 pmu = &uncore->pmus[0]; in amd_uncore_df_ctx_init() 863 uncore->pmus = kzalloc(sizeof(*uncore->pmus), GFP_KERNEL); in amd_uncore_l3_ctx_init() 864 if (!uncore->pmus) in amd_uncore_l3_ctx_init() 872 pmu = &uncore->pmus[0]; in amd_uncore_l3_ctx_init() [all …]
|
| /linux/tools/perf/arch/arm/util/ |
| H A D | auxtrace.c | 103 static struct perf_pmu *find_pmu_for_event(struct perf_pmu **pmus, in find_pmu_for_event() argument 108 if (!pmus) in find_pmu_for_event() 112 if (evsel->core.attr.type == pmus[i]->type) in find_pmu_for_event() 113 return pmus[i]; in find_pmu_for_event()
|
| /linux/arch/x86/events/intel/ |
| H A D | uncore.c | 959 struct intel_uncore_pmu *pmu = type->pmus; in uncore_type_exit() 973 kfree(type->pmus); in uncore_type_exit() 974 type->pmus = NULL; in uncore_type_exit() 989 struct intel_uncore_pmu *pmus; in uncore_type_init() local 993 pmus = kcalloc(type->num_boxes, sizeof(*pmus), GFP_KERNEL); in uncore_type_init() 994 if (!pmus) in uncore_type_init() 1000 pmus[i].pmu_idx = i; in uncore_type_init() 1001 pmus[i].type = type; in uncore_type_init() 1002 pmus[i].boxes = kzalloc(size, GFP_KERNEL); in uncore_type_init() 1003 if (!pmus[i].boxes) in uncore_type_init() [all …]
|
| H A D | uncore.h | 80 struct intel_uncore_pmu *pmus; member
|
| H A D | core.c | 7109 static __always_inline int intel_pmu_init_hybrid(enum hybrid_pmu_type pmus) in intel_pmu_init_hybrid() argument 7111 unsigned long pmus_mask = pmus; in intel_pmu_init_hybrid()
|
| /linux/tools/perf/util/ |
| H A D | drm_pmu.c | 73 static struct drm_pmu *add_drm_pmu(struct list_head *pmus, char *line, size_t line_len) in add_drm_pmu() argument 92 list_for_each_entry(pmu, pmus, list) { in add_drm_pmu() 161 struct list_head *pmus = args; in read_drm_pmus_cb() local 178 drm = add_drm_pmu(pmus, line, line_len); in read_drm_pmus_cb() 245 list_add_tail(&drm->pmu.list, pmus); in read_drm_pmus_cb() 514 int perf_pmus__read_drm_pmus(struct list_head *pmus) in perf_pmus__read_drm_pmus() argument 516 return for_each_drm_fdinfo(/*skip_all_duplicates=*/true, read_drm_pmus_cb, pmus); in perf_pmus__read_drm_pmus()
|
| H A D | hwmon_pmu.h | 145 struct perf_pmu *hwmon_pmu__new(struct list_head *pmus, const char *hwmon_dir, 159 int perf_pmus__read_hwmon_pmus(struct list_head *pmus);
|
| H A D | drm_pmu.h | 32 int perf_pmus__read_drm_pmus(struct list_head *pmus);
|
| H A D | hwmon_pmu.c | 345 struct perf_pmu *hwmon_pmu__new(struct list_head *pmus, const char *hwmon_dir, in hwmon_pmu__new() argument 389 list_add_tail(&hwm->pmu.list, pmus); in hwmon_pmu__new() 704 int perf_pmus__read_hwmon_pmus(struct list_head *pmus) in perf_pmus__read_hwmon_pmus() argument 747 hwmon_pmu__new(pmus, buf, class_hwmon_ent->d_name, line); in perf_pmus__read_hwmon_pmus()
|
| H A D | pmu.h | 315 struct perf_pmu *perf_pmu__lookup(struct list_head *pmus, int dirfd, const char *lookup_name,
|
| H A D | pmu.c | 1248 struct perf_pmu *perf_pmu__lookup(struct list_head *pmus, int dirfd, const char *name, in perf_pmu__lookup() argument 1301 list_add_tail(&pmu->list, pmus); in perf_pmu__lookup()
|
| H A D | Build | 87 perf-util-y += pmus.o
|
| /linux/tools/perf/python/ |
| H A D | ilist.py | 102 for p in perf.pmus(): 443 pmus = tree.root.add("PMUs") 444 for pmu in perf.pmus(): 446 pmu_node = pmus.add(pmu_name)
|
| /linux/tools/perf/Documentation/ |
| H A D | intel-hybrid.txt | 10 Kernel exports two new cpu pmus via sysfs: 203 warning and disable grouping, because the pmus in group are
|
| /linux/kernel/events/ |
| H A D | core.c | 450 static LIST_HEAD(pmus); 12505 list_add_rcu(&pmu->entry, &pmus); in DEFINE_FREE() 12632 list_add_rcu(&pmu->entry, &pmus); in perf_pmu_unregister() 12794 list_for_each_entry_rcu(pmu, &pmus, entry, lockdep_is_held(&pmus_srcu)) { in perf_init_event() 14760 list_for_each_entry(pmu, &pmus, entry) { in perf_event_clear_cpumask() 14934 list_for_each_entry(pmu, &pmus, entry) { in perf_event_sysfs_init()
|