| /linux/tools/lib/perf/ |
| H A D | evlist.c | 52 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps() 75 assert(!perf_cpu_map__has_any_cpu(evlist->user_requested_cpus)); in __perf_evlist__propagate_maps() 76 if (!perf_cpu_map__equal(evsel->cpus, evlist->user_requested_cpus)) { in __perf_evlist__propagate_maps() 78 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps() 99 perf_cpu_map__has_any_cpu(evlist->user_requested_cpus)) { in __perf_evlist__propagate_maps() 101 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps() 113 evlist->user_requested_cpus, in __perf_evlist__propagate_maps() 238 perf_cpu_map__put(evlist->user_requested_cpus); in perf_evlist__exit() 241 evlist->user_requested_cpus = NULL; in perf_evlist__exit() 270 if (cpus != evlist->user_requested_cpus) { in perf_evlist__set_maps() [all …]
|
| /linux/tools/perf/util/ |
| H A D | cputopo.c | 179 struct perf_cpu_map *user_requested_cpus; in cpu_topology__core_wide() local 188 user_requested_cpus = perf_cpu_map__new(user_requested_cpu_list); in cpu_topology__core_wide() 199 has_first = perf_cpu_map__has(user_requested_cpus, cpu); in cpu_topology__core_wide() 209 if (perf_cpu_map__has(user_requested_cpus, cpu) != has_first) { in cpu_topology__core_wide() 211 perf_cpu_map__put(user_requested_cpus); in cpu_topology__core_wide() 218 perf_cpu_map__put(user_requested_cpus); in cpu_topology__core_wide()
|
| H A D | record.c | 102 if (perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0).cpu < 0) in evlist__config() 240 if (!evlist || perf_cpu_map__is_any_cpu_or_is_empty(evlist->core.user_requested_cpus)) { in evlist__can_select_event() 248 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0); in evlist__can_select_event()
|
| H A D | bpf_ftrace.c | 62 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf() 90 cpu = perf_cpu_map__cpu(ftrace->evlist->core.user_requested_cpus, i).cpu; in perf_ftrace__latency_prepare_bpf()
|
| H A D | bpf_off_cpu.c | 169 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare() 241 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in off_cpu_prepare()
|
| H A D | sideband_evlist.c | 117 if (evsel__open(counter, evlist->core.user_requested_cpus, in evlist__start_sb_thread()
|
| H A D | evlist.c | 372 if (evlist->no_affinity || !evlist->core.user_requested_cpus || in evlist__use_affinity() 373 cpu_map__is_dummy(evlist->core.user_requested_cpus)) in evlist__use_affinity() 1446 if (evlist->core.threads == NULL && evlist->core.user_requested_cpus == NULL) { in evlist__open() 2590 struct perf_cpu_map *user_requested_cpus; in evlist__warn_user_requested_cpus() local 2596 user_requested_cpus = perf_cpu_map__new(cpu_list); in evlist__warn_user_requested_cpus() 2597 if (!user_requested_cpus) in evlist__warn_user_requested_cpus() 2601 evsel__warn_user_requested_cpus(pos, user_requested_cpus); in evlist__warn_user_requested_cpus() 2603 perf_cpu_map__put(user_requested_cpus); in evlist__warn_user_requested_cpus()
|
| H A D | top.c | 97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf()
|
| H A D | intel-tpebs.c | 99 if (!perf_cpu_map__is_any_cpu_or_is_empty(evsel->evlist->core.user_requested_cpus)) { in evsel__tpebs_start_perf_record() 100 cpu_map__snprint(evsel->evlist->core.user_requested_cpus, cpumap_buf, in evsel__tpebs_start_perf_record()
|
| H A D | bpf_lock_contention.c | 225 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in lock_contention_prepare() 330 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in lock_contention_prepare()
|
| H A D | evsel.h | 586 void evsel__warn_user_requested_cpus(struct evsel *evsel, struct perf_cpu_map *user_requested_cpus);
|
| H A D | synthetic-events.c | 2376 err = perf_event__synthesize_cpu_map(tool, evlist->core.user_requested_cpus, process, NULL); in perf_event__synthesize_stat_events() 2656 struct perf_cpu_map *user_requested_cpus) in perf_event__synthesize_schedstat() argument 2714 if (user_requested_cpus && !perf_cpu_map__has(user_requested_cpus, this_cpu)) in perf_event__synthesize_schedstat()
|
| H A D | evsel.c | 4384 void evsel__warn_user_requested_cpus(struct evsel *evsel, struct perf_cpu_map *user_requested_cpus) in evsel__warn_user_requested_cpus() argument 4390 intersect = perf_cpu_map__intersect(pmu->cpus, user_requested_cpus); in evsel__warn_user_requested_cpus() 4393 intersect = perf_cpu_map__intersect(online, user_requested_cpus); in evsel__warn_user_requested_cpus() 4395 if (!perf_cpu_map__equal(intersect, user_requested_cpus)) { in evsel__warn_user_requested_cpus() 4399 cpu_map__snprint(user_requested_cpus, buf1, sizeof(buf1)); in evsel__warn_user_requested_cpus()
|
| H A D | auxtrace.c | 194 bool per_cpu = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); in auxtrace_mmap_params__set_idx() 670 bool per_cpu_mmaps = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); in evlist__enable_event_idx()
|
| H A D | stat-display.c | 1131 perf_cpu_map__for_each_cpu(cpu, all_idx, evlist->core.user_requested_cpus) { in print_no_aggr_metric()
|
| /linux/tools/perf/arch/arm64/util/ |
| H A D | arm-spe.c | 63 struct perf_cpu_map *event_cpus = evlist->core.user_requested_cpus; in arm_spe_find_cpus() 365 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in arm_spe_setup_tracking_event() 398 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in arm_spe_recording_options()
|
| /linux/tools/perf/arch/arm/util/ |
| H A D | cs-etm.c | 201 struct perf_cpu_map *event_cpus = evsel->evlist->core.user_requested_cpus; in cs_etm_validate_config() 325 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in cs_etm_recording_options() 551 struct perf_cpu_map *event_cpus = evlist->core.user_requested_cpus; in cs_etm_info_priv_size() 790 struct perf_cpu_map *event_cpus = session->evlist->core.user_requested_cpus; in cs_etm_info_fill()
|
| /linux/tools/lib/perf/include/internal/ |
| H A D | evlist.h | 26 struct perf_cpu_map *user_requested_cpus; member
|
| /linux/tools/perf/ |
| H A D | builtin-sched.c | 61 static struct perf_cpu_map *user_requested_cpus; variable 3890 user_requested_cpus = evlist->core.user_requested_cpus; in perf_sched__schedstat_record() 3894 user_requested_cpus); in perf_sched__schedstat_record() 3916 user_requested_cpus); in perf_sched__schedstat_record() 4406 if (user_requested_cpus && !perf_cpu_map__has(user_requested_cpus, this_cpu)) in perf_sched__process_schedstat() 4512 user_requested_cpus = session->evlist->core.user_requested_cpus; in perf_sched__schedstat_report() 4678 user_requested_cpus = evlist->core.user_requested_cpus; in perf_sched__schedstat_live() 4682 user_requested_cpus); in perf_sched__schedstat_live() 4704 user_requested_cpus); in perf_sched__schedstat_live()
|
| H A D | builtin-stat.c | 1574 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode() 1902 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode_file()
|
| H A D | builtin-ftrace.c | 415 struct perf_cpu_map *cpumap = ftrace->evlist->core.user_requested_cpus; in set_tracing_cpu()
|
| H A D | builtin-record.c | 1056 return cpu_map__is_dummy(evlist->core.user_requested_cpus); in evlist__per_thread()
|
| /linux/tools/perf/bench/ |
| H A D | evlist-open-close.c | 148 printf(" Number of cpus:\t%d\n", perf_cpu_map__nr(evlist->core.user_requested_cpus)); in bench_evlist_open_close__run()
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | intel-bts.c | 113 const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in intel_bts_recording_options()
|
| H A D | intel-pt.c | 376 per_cpu_mmaps = !perf_cpu_map__is_any_cpu_or_is_empty(session->evlist->core.user_requested_cpus); in intel_pt_info_fill() 621 const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in intel_pt_recording_options()
|