| /linux/tools/testing/selftests/kvm/arm64/ |
| H A D | vgic_lpi_stress.c | 34 u32 nr_cpus; member 47 .nr_cpus = 1, 68 u32 nr_cpus = test_data.nr_cpus; in guest_setup_its_mappings() local 70 for (coll_id = 0; coll_id < nr_cpus; coll_id++) in guest_setup_its_mappings() 85 coll_id = (coll_id + 1) % test_data.nr_cpus; in guest_setup_its_mappings() 94 for (i = 0; i < test_data.nr_cpus; i++) in guest_invalidate_all_rdists() 103 gic_init(GIC_V3, test_data.nr_cpus); in guest_setup_gic() 112 while (atomic_load(&nr_cpus_ready) < test_data.nr_cpus) in guest_setup_gic() 123 for (cpuid = 0; cpuid < test_data.nr_cpus; cpuid++) in guest_setup_gic() 163 sz += (1 + test_data.nr_cpus) * SZ_64K; in setup_memslot() [all …]
|
| /linux/tools/lib/perf/ |
| H A D | cpumap.c | 18 void perf_cpu_map__set_nr(struct perf_cpu_map *map, int nr_cpus) in perf_cpu_map__set_nr() argument 20 RC_CHK_ACCESS(map)->nr = nr_cpus; in perf_cpu_map__set_nr() 23 struct perf_cpu_map *perf_cpu_map__alloc(int nr_cpus) in perf_cpu_map__alloc() argument 28 if (nr_cpus == 0) in perf_cpu_map__alloc() 31 cpus = malloc(sizeof(*cpus) + sizeof(struct perf_cpu) * nr_cpus); in perf_cpu_map__alloc() 33 cpus->nr = nr_cpus; in perf_cpu_map__alloc() 81 int nr_cpus, nr_cpus_conf; in cpu_map__new_sysconf() local 83 nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); in cpu_map__new_sysconf() 84 if (nr_cpus < 0) in cpu_map__new_sysconf() 88 if (nr_cpus != nr_cpus_conf) { in cpu_map__new_sysconf() [all …]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | map_lookup_percpu_elem.c | 11 int ret, i, nr_cpus = libbpf_num_possible_cpus(); in test_map_lookup_percpu_elem() local 14 buf = malloc(nr_cpus*sizeof(__u64)); in test_map_lookup_percpu_elem() 18 for (i = 0; i < nr_cpus; i++) in test_map_lookup_percpu_elem() 20 sum = (nr_cpus - 1) * nr_cpus / 2; in test_map_lookup_percpu_elem() 27 skel->rodata->nr_cpus = nr_cpus; in test_map_lookup_percpu_elem()
|
| H A D | map_init.c | 10 static int nr_cpus; variable 22 pcpu_map_value_t value[nr_cpus]; in map_populate() 26 for (i = 0; i < nr_cpus; i++) in map_populate() 105 for (i = 0; i < nr_cpus; i++) { in check_values_one_cpu() 131 pcpu_map_value_t value[nr_cpus]; in test_pcpu_map_init() 171 pcpu_map_value_t value[nr_cpus]; in test_pcpu_lru_map_init() 203 nr_cpus = bpf_num_possible_cpus(); in test_map_init() 204 if (nr_cpus <= 1) { in test_map_init()
|
| H A D | lookup_and_delete.c | 11 static int nr_cpus; variable 29 __u64 key, value[nr_cpus]; in fill_values_percpu() 32 for (i = 0; i < nr_cpus; i++) in fill_values_percpu() 137 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_percpu_hash() 156 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_percpu_hash() 223 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_lru_percpu_hash() 242 for (i = 0; i < nr_cpus; i++) in test_lookup_and_delete_lru_percpu_hash() 252 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_lru_percpu_hash() 281 nr_cpus = bpf_num_possible_cpus(); in test_lookup_and_delete()
|
| H A D | perf_buffer.c | 48 int err, on_len, nr_on_cpus = 0, nr_cpus, i, j; in serial_test_perf_buffer() local 56 nr_cpus = libbpf_num_possible_cpus(); in serial_test_perf_buffer() 57 if (CHECK(nr_cpus < 0, "nr_cpus", "err %d\n", nr_cpus)) in serial_test_perf_buffer() 94 for (i = 0; i < nr_cpus; i++) { in serial_test_perf_buffer() 117 for (i = 0, j = 0; i < nr_cpus; i++) { in serial_test_perf_buffer()
|
| /linux/tools/tracing/rtla/src/ |
| H A D | timerlat_u.c | 102 static int timerlat_u_send_kill(pid_t *procs, int nr_cpus) in timerlat_u_send_kill() argument 107 for (i = 0; i < nr_cpus; i++) { in timerlat_u_send_kill() 134 int nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in timerlat_u_dispatcher() local 146 procs = calloc(nr_cpus, sizeof(pid_t)); in timerlat_u_dispatcher() 150 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher() 173 timerlat_u_send_kill(procs, nr_cpus); in timerlat_u_dispatcher() 186 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher() 200 timerlat_u_send_kill(procs, nr_cpus); in timerlat_u_dispatcher() 208 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher()
|
| H A D | timerlat_top.c | 45 int nr_cpus; member 66 static struct timerlat_top_data *timerlat_alloc_top(int nr_cpus) in timerlat_alloc_top() argument 75 data->nr_cpus = nr_cpus; in timerlat_alloc_top() 78 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); in timerlat_alloc_top() 83 for (cpu = 0; cpu < nr_cpus; cpu++) { in timerlat_alloc_top() 194 long long value_irq[data->nr_cpus], in timerlat_top_bpf_pull_data() 195 value_thread[data->nr_cpus], in timerlat_top_bpf_pull_data() 196 value_user[data->nr_cpus]; in timerlat_top_bpf_pull_data() 201 data->nr_cpus); in timerlat_top_bpf_pull_data() 204 for (i = 0; i < data->nr_cpus; i++) { in timerlat_top_bpf_pull_data() [all …]
|
| H A D | timerlat_hist.c | 47 int nr_cpus; member 59 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in timerlat_free_histogram() 86 *timerlat_alloc_histogram(int nr_cpus, int entries, int bucket_size) in timerlat_alloc_histogram() argument 97 data->nr_cpus = nr_cpus; in timerlat_alloc_histogram() 100 data->hist = calloc(1, sizeof(*data->hist) * nr_cpus); in timerlat_alloc_histogram() 105 for (cpu = 0; cpu < nr_cpus; cpu++) { in timerlat_alloc_histogram() 120 for (cpu = 0; cpu < nr_cpus; cpu++) { in timerlat_alloc_histogram() 207 long long value_irq[data->nr_cpus], in timerlat_hist_bpf_pull_data() 208 value_thread[data->nr_cpus], in timerlat_hist_bpf_pull_data() 209 value_user[data->nr_cpus]; in timerlat_hist_bpf_pull_data() [all …]
|
| H A D | osnoise_hist.c | 33 int nr_cpus; member 45 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in osnoise_free_histogram() 66 *osnoise_alloc_histogram(int nr_cpus, int entries, int bucket_size) in osnoise_alloc_histogram() argument 77 data->nr_cpus = nr_cpus; in osnoise_alloc_histogram() 79 data->hist = calloc(1, sizeof(*data->hist) * nr_cpus); in osnoise_alloc_histogram() 83 for (cpu = 0; cpu < nr_cpus; cpu++) { in osnoise_alloc_histogram() 90 for (cpu = 0; cpu < nr_cpus; cpu++) in osnoise_alloc_histogram() 250 for_each_monitored_cpu(cpu, data->nr_cpus, ¶ms->common) { in osnoise_hist_header() 279 for_each_monitored_cpu(cpu, data->nr_cpus, ¶ms->common) { in osnoise_print_summary() 291 for_each_monitored_cpu(cpu, data->nr_cpus, ¶ms->common) { in osnoise_print_summary() [all …]
|
| H A D | osnoise_top.c | 34 int nr_cpus; member 54 static struct osnoise_top_data *osnoise_alloc_top(int nr_cpus) in osnoise_alloc_top() argument 62 data->nr_cpus = nr_cpus; in osnoise_alloc_top() 65 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); in osnoise_alloc_top() 235 static int nr_cpus = -1; in osnoise_print_stats() local 238 if (nr_cpus == -1) in osnoise_print_stats() 239 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in osnoise_print_stats() 246 for_each_monitored_cpu(i, nr_cpus, ¶ms->common) { in osnoise_print_stats() 550 int nr_cpus; in osnoise_init_top() local 552 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in osnoise_init_top() [all …]
|
| /linux/tools/testing/selftests/bpf/map_tests/ |
| H A D | array_map_batch_ops.c | 13 static int nr_cpus; variable 28 cpu_offset = i * nr_cpus; in map_batch_update() 29 for (j = 0; j < nr_cpus; j++) in map_batch_update() 49 cpu_offset = i * nr_cpus; in map_batch_verify() 50 for (j = 0; j < nr_cpus; j++) { in map_batch_verify() 90 value_size *= nr_cpus; in __test_map_lookup_and_update_batch() 158 nr_cpus = libbpf_num_possible_cpus(); in test_array_map_batch_ops() 160 CHECK(nr_cpus < 0, "nr_cpus checking", in test_array_map_batch_ops()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | test_lru_map.c | 26 static int nr_cpus; variable 96 unsigned long long value0[nr_cpus], value1[nr_cpus]; in map_subset() 127 while (next < nr_cpus) { in sched_next_online() 144 return (map_size / nr_cpus) / 2; in __tgt_size() 150 return tgt_free * nr_cpus * 2; in __map_size() 163 unsigned long long key, value[nr_cpus]; in test_lru_sanity0() 173 lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); in test_lru_sanity0() 254 unsigned long long key, end_key, value[nr_cpus]; in test_lru_sanity1() 322 unsigned long long key, value[nr_cpus]; in test_lru_sanity2() 428 unsigned long long key, end_key, value[nr_cpus]; in test_lru_sanity3() [all …]
|
| /linux/include/trace/events/ |
| H A D | sched_ext.h | 50 TP_PROTO(__u32 node, __u32 nr_cpus, __u32 nr_tasks, __u32 nr_balanced, 54 TP_ARGS(node, nr_cpus, nr_tasks, nr_balanced, 59 __field( __u32, nr_cpus ) 70 __entry->nr_cpus = nr_cpus; 80 __entry->node, __entry->nr_cpus,
|
| /linux/samples/bpf/ |
| H A D | test_lru_dist.c | 31 static int nr_cpus; variable 222 if (next_to_try == nr_cpus) in sched_next_online() 225 while (next_to_try < nr_cpus) { in sched_next_online() 318 nr_cpus * lru_size); in test_parallel_lru_dist() 334 unsigned long long key, value[nr_cpus]; in test_lru_loss0() 346 map_fd = create_map(map_type, map_flags, 900 * nr_cpus); in test_lru_loss0() 388 unsigned long long key, value[nr_cpus]; in test_lru_loss1() 398 map_fd = create_map(map_type, map_flags, 1000 * nr_cpus); in test_lru_loss1() 426 unsigned long long key, value[nr_cpus]; in do_test_parallel_lru_loss() 475 nr_cpus * (1000 + 200)); in test_parallel_lru_loss() [all …]
|
| H A D | xdp_sample_user.c | 317 unsigned int nr_cpus = libbpf_num_possible_cpus(); in alloc_record_per_cpu() local 320 array = calloc(nr_cpus, sizeof(*array)); in alloc_record_per_cpu() 323 nr_cpus); in alloc_record_per_cpu() 343 unsigned int nr_cpus = libbpf_num_possible_cpus(); in map_collect_percpu() local 356 for (i = 0; i < nr_cpus; i++) { in map_collect_percpu() 382 unsigned int nr_cpus = bpf_num_possible_cpus(); in map_collect_percpu_devmap() local 392 values = calloc(count * nr_cpus, sizeof(struct datarec)); in map_collect_percpu_devmap() 414 arr = &values[i * nr_cpus]; in map_collect_percpu_devmap() 650 unsigned int nr_cpus, struct sample_output *out) in stats_get_rx_cnt() argument 660 for (i = 0; i < nr_cpus; i++) { in stats_get_rx_cnt() [all …]
|
| H A D | tracex3_user.c | 19 unsigned int nr_cpus = bpf_num_possible_cpus(); in clear_stats() local 20 __u64 values[nr_cpus]; in clear_stats() 76 unsigned int nr_cpus = bpf_num_possible_cpus(); in print_hist() local 78 long values[nr_cpus]; in print_hist() 88 for (i = 0; i < nr_cpus; i++) in print_hist()
|
| H A D | sampleip_user.c | 26 static int nr_cpus; variable 49 for (i = 0; i < nr_cpus; i++) { in sampling_start() 72 for (i = 0; i < nr_cpus; i++) in sampling_end() 180 nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); in main() 181 links = calloc(nr_cpus, sizeof(struct bpf_link *)); in main()
|
| /linux/tools/sched_ext/ |
| H A D | scx_cpu0.c | 42 int nr_cpus = libbpf_num_possible_cpus(); in read_stats() local 43 assert(nr_cpus > 0); in read_stats() 44 __u64 cnts[2][nr_cpus]; in read_stats() 56 for (cpu = 0; cpu < nr_cpus; cpu++) in read_stats() 74 skel->rodata->nr_cpus = libbpf_num_possible_cpus(); in main()
|
| H A D | scx_simple.c | 44 int nr_cpus = libbpf_num_possible_cpus(); in read_stats() local 45 assert(nr_cpus > 0); in read_stats() 46 __u64 cnts[2][nr_cpus]; in read_stats() 58 for (cpu = 0; cpu < nr_cpus; cpu++) in read_stats()
|
| /linux/drivers/powercap/ |
| H A D | dtpm_cpu.c | 48 int i, nr_cpus; in set_pd_power_limit() local 50 nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(pd->cpus)); in set_pd_power_limit() 56 power = table[i].power * nr_cpus; in set_pd_power_limit() 63 power_limit = table[i - 1].power * nr_cpus; in set_pd_power_limit() 126 int nr_cpus; in update_pd_power_uw() local 128 nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(em->cpus)); in update_pd_power_uw() 134 dtpm->power_min *= nr_cpus; in update_pd_power_uw() 137 dtpm->power_max *= nr_cpus; in update_pd_power_uw()
|
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | gic.c | 25 static void gic_dist_init(enum gic_type type, unsigned int nr_cpus) in gic_dist_init() argument 42 gic_ops->gic_init(nr_cpus); in gic_dist_init() 51 void gic_init(enum gic_type type, unsigned int nr_cpus) in gic_init() argument 56 GUEST_ASSERT(nr_cpus); in gic_init() 58 gic_dist_init(type, nr_cpus); in gic_init()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_map_lookup_percpu_elem.c | 10 const volatile int nr_cpus; variable 60 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid() 65 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid() 70 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid()
|
| /linux/tools/perf/util/ |
| H A D | top.c | 31 int nr_cpus; in perf_top__header_snprintf() local 97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf() 100 nr_cpus > 1 ? "s" : "", in perf_top__header_snprintf() 107 nr_cpus, nr_cpus > 1 ? "s" : ""); in perf_top__header_snprintf()
|
| /linux/drivers/pci/controller/ |
| H A D | pcie-iproc-msi.c | 98 int nr_cpus; member 198 return (hwirq % msi->nr_cpus); in hwirq_to_cpu() 254 if (msi->nr_cpus > 1 && nr_irqs > 1) in iproc_msi_irq_domain_alloc() 264 order_base_2(msi->nr_cpus * nr_irqs)); in iproc_msi_irq_domain_alloc() 292 order_base_2(msi->nr_cpus * nr_irqs)); in iproc_msi_irq_domain_free() 474 for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { in iproc_msi_irq_free() 486 for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { in iproc_msi_irq_setup() 538 msi->nr_cpus = num_possible_cpus(); in iproc_msi_init() 540 if (msi->nr_cpus == 1) in iproc_msi_init() 555 if (msi->nr_irqs < msi->nr_cpus) { in iproc_msi_init() [all …]
|