| /linux/tools/lib/perf/include/perf/ |
| H A D | cpumap.h | 60 LIBPERF_API int perf_cpu_map__nr(const struct perf_cpu_map *cpus); 92 (idx) < perf_cpu_map__nr(cpus); \ 97 (idx) < perf_cpu_map__nr(cpus); \ 102 for ((idx) = 0; (idx) < perf_cpu_map__nr(cpus); (idx)++)
|
| /linux/tools/perf/util/ |
| H A D | tool_pmu.c | 211 evsel->start_times = xyarray__new(perf_cpu_map__nr(cpus), in evsel__tool_pmu_prepare_open() 359 *result = perf_cpu_map__nr(evsel->core.cpus); in tool_pmu__read_event() 367 *result = perf_cpu_map__nr(evsel->core.pmu_cpus); in tool_pmu__read_event() 381 *result = perf_cpu_map__nr(online); in tool_pmu__read_event() 387 *result = perf_cpu_map__nr(tmp); in tool_pmu__read_event() 400 *result = perf_cpu_map__nr(tmp); in tool_pmu__read_event() 403 *result = perf_cpu_map__nr(online); in tool_pmu__read_event()
|
| H A D | cpumap.c | 259 struct cpu_aggr_map *c = cpu_aggr_map__empty_new(perf_cpu_map__nr(cpus)); in cpu_aggr_map__new() 283 if (c->nr != perf_cpu_map__nr(cpus)) { in cpu_aggr_map__new() 634 for (i = 0; i < perf_cpu_map__nr(map) + 1; i++) { in cpu_map__snprint() 636 bool last = i == perf_cpu_map__nr(map); in cpu_map__snprint()
|
| H A D | counts.c | 63 evsel->counts = perf_counts__new(perf_cpu_map__nr(cpus), nthreads); in evsel__alloc_counts()
|
| H A D | cpumap.h | 71 return perf_cpu_map__nr(cpus) == 1 && perf_cpu_map__cpu(cpus, 0).cpu == -1; in cpu_map__is_dummy()
|
| H A D | top.c | 97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf()
|
| H A D | bpf_ftrace.c | 62 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf()
|
| H A D | mmap.c | 253 nr_cpus = perf_cpu_map__nr(cpu_map); in build_node_mask()
|
| H A D | cputopo.c | 404 nr = (u32) perf_cpu_map__nr(node_map); in numa_topology__new()
|
| H A D | bpf_off_cpu.c | 169 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare()
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | pmu.c | 66 snc_nodes = perf_cpu_map__nr(cache_cpus) / perf_cpu_map__nr(node_cpus); in snc_nodes_per_l3_cache() 248 adjusted[pmu_snc] = perf_cpu_map__empty_new(perf_cpu_map__nr(pmu->cpus)); in gnr_uncore_cha_imc_adjust_cpumask_for_snc()
|
| /linux/tools/lib/perf/ |
| H A D | evlist.c | 110 if (evsel->reads_only_on_cpu_idx0 && perf_cpu_map__nr(evsel->cpus) > 0) { in __perf_evlist__propagate_maps() 412 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__alloc_pollfd() 638 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_thread() 680 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_cpu() 712 nr_mmaps = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__nr_mmaps()
|
| H A D | evsel.c | 157 perf_evsel__alloc_fd(evsel, perf_cpu_map__nr(cpus), threads->nr) < 0) in perf_evsel__open() 504 for (i = 0; i < perf_cpu_map__nr(evsel->cpus) && !err; i++) in perf_evsel__apply_filter()
|
| /linux/tools/perf/bench/ |
| H A D | futex-wake.c | 119 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 170 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake()
|
| H A D | futex-lock-pi.c | 151 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in create_threads() 193 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_lock_pi()
|
| H A D | futex-hash.c | 160 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_hash() 195 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in bench_futex_hash()
|
| H A D | evlist-open-close.c | 56 cnt += evsel->core.threads->nr * perf_cpu_map__nr(evsel->core.cpus); in evlist__count_evsel_fds() 148 printf(" Number of cpus:\t%d\n", perf_cpu_map__nr(evlist->core.user_requested_cpus)); in bench_evlist_open_close__run()
|
| H A D | futex-requeue.c | 146 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 195 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_requeue()
|
| H A D | futex-wake-parallel.c | 170 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 276 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake_parallel()
|
| H A D | epoll-ctl.c | 264 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, in do_threads() 350 nthreads = perf_cpu_map__nr(cpu); in bench_epoll_ctl()
|
| /linux/tools/perf/tests/ |
| H A D | bitmap.c | |
| H A D | event_update.c | 72 TEST_ASSERT_VAL("wrong cpus", perf_cpu_map__nr(map) == 3); in process_event_cpus()
|
| /linux/tools/lib/perf/tests/ |
| H A D | test-evlist.c | 77 for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) { in test_stat_cpu() 414 __T("failed count", count >= perf_cpu_map__nr(cpus)); in test_mmap_cpus()
|
| H A D | test-evsel.c | 39 for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) { in test_stat_cpu()
|
| /linux/tools/perf/arch/arm64/util/ |
| H A D | arm-spe.c | 90 ARM_SPE_CPU_PRIV_MAX * perf_cpu_map__nr(cpu_map); in arm_spe_info_priv_size() 171 auxtrace_info->priv[ARM_SPE_CPUS_NUM] = perf_cpu_map__nr(cpu_map); in arm_spe_info_fill()
|