| /linux/tools/lib/perf/ |
| H A D | evlist.c | 52 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps() 75 assert(!perf_cpu_map__has_any_cpu(evlist->user_requested_cpus)); in __perf_evlist__propagate_maps() 76 if (!perf_cpu_map__equal(evsel->cpus, evlist->user_requested_cpus)) { in __perf_evlist__propagate_maps() 78 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps() 99 perf_cpu_map__has_any_cpu(evlist->user_requested_cpus)) { in __perf_evlist__propagate_maps() 101 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps() 113 evlist->user_requested_cpus, in __perf_evlist__propagate_maps() 238 perf_cpu_map__put(evlist->user_requested_cpus); in perf_evlist__exit() 241 evlist->user_requested_cpus = NULL; in perf_evlist__exit() 270 if (cpus != evlist->user_requested_cpus) { in perf_evlist__set_maps() [all …]
|
| /linux/tools/perf/util/ |
| H A D | cputopo.c | 179 struct perf_cpu_map *user_requested_cpus; in cpu_topology__core_wide() local 188 user_requested_cpus = perf_cpu_map__new(user_requested_cpu_list); in cpu_topology__core_wide() 199 has_first = perf_cpu_map__has(user_requested_cpus, cpu); in cpu_topology__core_wide() 209 if (perf_cpu_map__has(user_requested_cpus, cpu) != has_first) { in cpu_topology__core_wide() 211 perf_cpu_map__put(user_requested_cpus); in cpu_topology__core_wide() 218 perf_cpu_map__put(user_requested_cpus); in cpu_topology__core_wide()
|
| H A D | record.c | 102 if (perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0).cpu < 0) in evlist__config() 240 if (!evlist || perf_cpu_map__is_any_cpu_or_is_empty(evlist->core.user_requested_cpus)) { in evlist__can_select_event() 248 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0); in evlist__can_select_event()
|
| H A D | bpf_ftrace.c | 62 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf() 90 cpu = perf_cpu_map__cpu(ftrace->evlist->core.user_requested_cpus, i).cpu; in perf_ftrace__latency_prepare_bpf()
|
| H A D | sideband_evlist.c | 117 if (evsel__open(counter, evlist->core.user_requested_cpus, in evlist__start_sb_thread()
|
| H A D | bpf_off_cpu.c | 169 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare() 241 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in off_cpu_prepare()
|
| H A D | evlist.c | 382 if (evlist->no_affinity || !evlist->core.user_requested_cpus || in evlist__use_affinity() 383 cpu_map__is_dummy(evlist->core.user_requested_cpus)) in evlist__use_affinity() 1455 if (evlist->core.threads == NULL && evlist->core.user_requested_cpus == NULL) { in evlist__open() 2602 struct perf_cpu_map *user_requested_cpus; in evlist__warn_user_requested_cpus() 2608 user_requested_cpus = perf_cpu_map__new(cpu_list); in evlist__disable_uniquify() 2609 if (!user_requested_cpus) in evlist__disable_uniquify() 2613 evsel__warn_user_requested_cpus(pos, user_requested_cpus); in evlist__disable_uniquify() 2615 perf_cpu_map__put(user_requested_cpus); in evlist__disable_uniquify() 2590 struct perf_cpu_map *user_requested_cpus; evlist__warn_user_requested_cpus() local
|
| H A D | top.c | 97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf()
|
| H A D | intel-tpebs.c | 98 if (!perf_cpu_map__is_any_cpu_or_is_empty(evsel->evlist->core.user_requested_cpus)) { in evsel__tpebs_start_perf_record() 99 cpu_map__snprint(evsel->evlist->core.user_requested_cpus, cpumap_buf, in evsel__tpebs_start_perf_record()
|
| H A D | bpf_lock_contention.c | 225 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in lock_contention_prepare() 330 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in lock_contention_prepare()
|
| H A D | evsel.h | 588 void evsel__warn_user_requested_cpus(struct evsel *evsel, struct perf_cpu_map *user_requested_cpus);
|
| H A D | synthetic-events.c | 2376 err = perf_event__synthesize_cpu_map(tool, evlist->core.user_requested_cpus, process, NULL); in perf_event__synthesize_stat_events() 2656 struct perf_cpu_map *user_requested_cpus) in perf_event__synthesize_schedstat() argument 2714 if (user_requested_cpus && !perf_cpu_map__has(user_requested_cpus, this_cpu)) in perf_event__synthesize_schedstat()
|
| H A D | evsel.c | 4416 void evsel__warn_user_requested_cpus(struct evsel *evsel, struct perf_cpu_map *user_requested_cpus) 4422 intersect = perf_cpu_map__intersect(pmu->cpus, user_requested_cpus); 4425 intersect = perf_cpu_map__intersect(online, user_requested_cpus); 4427 if (!perf_cpu_map__equal(intersect, user_requested_cpus)) { 4431 cpu_map__snprint(user_requested_cpus, buf1, sizeof(buf1)); 4384 evsel__warn_user_requested_cpus(struct evsel * evsel,struct perf_cpu_map * user_requested_cpus) evsel__warn_user_requested_cpus() argument
|
| H A D | auxtrace.c | 194 bool per_cpu = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); in auxtrace_mmap_params__set_idx() 670 bool per_cpu_mmaps = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); in evlist__enable_event_idx()
|
| H A D | stat-display.c | 1131 perf_cpu_map__for_each_cpu(cpu, all_idx, evlist->core.user_requested_cpus) { in print_no_aggr_metric()
|
| /linux/tools/perf/arch/arm64/util/ |
| H A D | arm-spe.c | 63 struct perf_cpu_map *event_cpus = evlist->core.user_requested_cpus; in arm_spe_find_cpus() 366 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in arm_spe_setup_tracking_event() 399 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in arm_spe_recording_options()
|
| /linux/tools/perf/arch/arm/util/ |
| H A D | cs-etm.c | 202 struct perf_cpu_map *event_cpus = evsel->evlist->core.user_requested_cpus; in cs_etm_validate_config() 326 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in cs_etm_recording_options() 552 struct perf_cpu_map *event_cpus = evlist->core.user_requested_cpus; in cs_etm_info_priv_size() 791 struct perf_cpu_map *event_cpus = session->evlist->core.user_requested_cpus; in cs_etm_info_fill()
|
| /linux/tools/lib/perf/include/internal/ |
| H A D | evlist.h | 26 struct perf_cpu_map *user_requested_cpus; member
|
| /linux/tools/perf/bench/ |
| H A D | evlist-open-close.c | 148 printf(" Number of cpus:\t%d\n", perf_cpu_map__nr(evlist->core.user_requested_cpus)); in bench_evlist_open_close__run()
|
| /linux/tools/perf/ |
| H A D | builtin-sched.c | 61 static struct perf_cpu_map *user_requested_cpus; variable 3890 user_requested_cpus = evlist->core.user_requested_cpus; in perf_sched__schedstat_record() 3894 user_requested_cpus); in perf_sched__schedstat_record() 3916 user_requested_cpus); in perf_sched__schedstat_record() 4406 if (user_requested_cpus && !perf_cpu_map__has(user_requested_cpus, this_cpu)) in perf_sched__process_schedstat() 4512 user_requested_cpus = session->evlist->core.user_requested_cpus; in perf_sched__schedstat_report() 4678 user_requested_cpus in perf_sched__schedstat_live() [all...] |
| H A D | builtin-stat.c | 1589 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode() 1917 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in default_evlist_evsel_cmp()
|
| H A D | builtin-ftrace.c | 416 struct perf_cpu_map *cpumap = ftrace->evlist->core.user_requested_cpus; in set_tracing_cpu()
|
| H A D | builtin-record.c | 1056 return cpu_map__is_dummy(evlist->core.user_requested_cpus); in evlist__per_thread()
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | intel-bts.c | 113 const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in intel_bts_recording_options()
|
| H A D | intel-pt.c | 376 per_cpu_mmaps = !perf_cpu_map__is_any_cpu_or_is_empty(session->evlist->core.user_requested_cpus); in intel_pt_info_fill() 621 const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in intel_pt_recording_options()
|