/linux/tools/perf/util/ |
H A D | bpf_lock_contention.c | 58 bpf_map__set_max_entries(skel->maps.slab_caches, con->map_nr_entries); in check_slab_cache_iter() 194 bpf_map__set_max_entries(skel->maps.lock_stat, con->map_nr_entries); in lock_contention_prepare() 195 bpf_map__set_max_entries(skel->maps.tstamp, con->map_nr_entries); in lock_contention_prepare() 198 bpf_map__set_max_entries(skel->maps.task_data, con->map_nr_entries); in lock_contention_prepare() 200 bpf_map__set_max_entries(skel->maps.task_data, 1); in lock_contention_prepare() 203 bpf_map__set_max_entries(skel->maps.stacks, con->map_nr_entries); in lock_contention_prepare() 208 bpf_map__set_max_entries(skel->maps.owner_stacks, con->map_nr_entries); in lock_contention_prepare() 209 bpf_map__set_max_entries(skel->maps.owner_data, con->map_nr_entries); in lock_contention_prepare() 210 bpf_map__set_max_entries(skel->maps.owner_stat, con->map_nr_entries); in lock_contention_prepare() 214 bpf_map__set_max_entries(skel->maps.stacks, 1); in lock_contention_prepare() [all …]
|
H A D | bpf_counter_cgroup.c | 71 bpf_map__set_max_entries(skel->maps.events, map_size); in bperf_load_program() 72 bpf_map__set_max_entries(skel->maps.cgrp_idx, nr_cgroups); in bperf_load_program() 75 bpf_map__set_max_entries(skel->maps.prev_readings, map_size); in bperf_load_program() 78 bpf_map__set_max_entries(skel->maps.cgrp_readings, map_size); in bperf_load_program()
|
H A D | bpf_off_cpu.c | 169 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in off_cpu_prepare() 195 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in off_cpu_prepare() 200 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in off_cpu_prepare() 203 bpf_map__set_max_entries(skel->maps.task_filter, MAX_PROC); in off_cpu_prepare() 210 bpf_map__set_max_entries(skel->maps.cgroup_filter, ncgrps); in off_cpu_prepare()
|
H A D | bpf_ftrace.c | 55 bpf_map__set_max_entries(skel->maps.latency, ftrace->bucket_num); in perf_ftrace__latency_prepare_bpf() 61 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in perf_ftrace__latency_prepare_bpf() 67 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in perf_ftrace__latency_prepare_bpf()
|
H A D | bpf-filter.c | 682 bpf_map__set_max_entries(skel->maps.filters, MAX_FILTERS); in perf_bpf_filter__pin() 683 bpf_map__set_max_entries(skel->maps.event_hash, MAX_EVT_HASH); in perf_bpf_filter__pin() 684 bpf_map__set_max_entries(skel->maps.idx_hash, MAX_IDX_HASH); in perf_bpf_filter__pin() 685 bpf_map__set_max_entries(skel->maps.dropped, MAX_FILTERS); in perf_bpf_filter__pin()
|
/linux/tools/tracing/rtla/src/ |
H A D | timerlat_bpf.c | 35 bpf_map__set_max_entries(bpf->maps.hist_irq, params->entries); in timerlat_bpf_init() 36 bpf_map__set_max_entries(bpf->maps.hist_thread, params->entries); in timerlat_bpf_init() 37 bpf_map__set_max_entries(bpf->maps.hist_user, params->entries); in timerlat_bpf_init()
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | ringbuf_multi.c | 57 ASSERT_OK(bpf_map__set_max_entries(skel->maps.ringbuf1, page_size + 1), "rb1_resize"); in test_ringbuf_multi() 59 ASSERT_OK(bpf_map__set_max_entries(skel->maps.ringbuf1, page_size), "rb1_reset"); in test_ringbuf_multi() 78 …if (!ASSERT_ERR(bpf_map__set_max_entries(skel->maps.ringbuf1, 3 * page_size), "rb1_resize_after_lo… in test_ringbuf_multi()
|
H A D | mmap.c | 36 err = bpf_map__set_max_entries(skel->maps.rdonly_map, page_size); in test_mmap() 41 err = bpf_map__set_max_entries(skel->maps.data_map, in test_mmap()
|
H A D | map_init.c | 52 err = bpf_map__set_max_entries(skel->maps.hashmap1, map_sz); in setup()
|
H A D | user_ringbuf.c | 79 err = bpf_map__set_max_entries(skel->maps.user_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel() 83 err = bpf_map__set_max_entries(skel->maps.kernel_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel()
|
H A D | lookup_and_delete.c | 58 err = bpf_map__set_max_entries(skel->maps.hash_map, MAX_ENTRIES); in setup_prog()
|
/linux/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bloom_filter_map.c | 272 bpf_map__set_max_entries(skel->maps.hashmap, args.nr_entries); in setup_skeleton() 274 bpf_map__set_max_entries(skel->maps.array_map, args.nr_entries); in setup_skeleton() 276 bpf_map__set_max_entries(skel->maps.bloom_map, args.nr_entries); in setup_skeleton()
|
H A D | bench_bpf_hashmap_lookup.c | 166 bpf_map__set_max_entries(ctx.skel->maps.hash_map_bench, args.max_entries); in setup()
|
H A D | bench_htab_mem.c | 183 bpf_map__set_max_entries(map, MAX(8192, 64 * env.nr_cpus)); in htab_mem_setup()
|
H A D | bench_ringbufs.c | 154 bpf_map__set_max_entries(skel->maps.ringbuf, args.ringbuf_sz); in ringbuf_setup_skeleton()
|
/linux/tools/lib/bpf/ |
H A D | libbpf.map | 188 bpf_map__set_max_entries;
|
H A D | libbpf.h | 1061 LIBBPF_API int bpf_map__set_max_entries(struct bpf_map *map, __u32 max_entries);
|
H A D | libbpf.c | 4964 int bpf_map__set_max_entries(struct bpf_map *map, __u32 max_entries) in bpf_map__set_max_entries() function
|
/linux/Documentation/bpf/ |
H A D | map_cpumap.rst | 166 if (bpf_map__set_max_entries(cpu_map, libbpf_num_possible_cpus()) < 0) {
|
/linux/tools/bpf/bpftool/ |
H A D | prog.c | 2463 bpf_map__set_max_entries(profile_obj->maps.events, num_metric * num_cpu); in do_profile() 2464 bpf_map__set_max_entries(profile_obj->maps.fentry_readings, num_metric); in do_profile() 2465 bpf_map__set_max_entries(profile_obj->maps.accum_readings, num_metric); in do_profile() 2466 bpf_map__set_max_entries(profile_obj->maps.counts, 1); in do_profile()
|
/linux/samples/bpf/ |
H A D | map_perf_test_user.c | 415 bpf_map__set_max_entries(map, num_map_entries); in fixup_map()
|
H A D | xdp_sample_user.c | 1220 if (bpf_map__set_max_entries(sample_map[i], sample_map_count[i]) < 0) in sample_setup_maps()
|
/linux/tools/testing/selftests/bpf/ |
H A D | veristat.c | 1241 bpf_map__set_max_entries(map, 1); in fixup_obj()
|