| /linux/lib/ |
| H A D | stackdepot.c | 422 static inline size_t depot_stack_record_size(struct stack_record *s, unsigned int nr_entries) in depot_stack_record_size() argument 424 const size_t used = flex_array_size(s, entries, nr_entries); in depot_stack_record_size() 434 depot_alloc_stack(unsigned long *entries, unsigned int nr_entries, u32 hash, depot_flags_t flags, v… in depot_alloc_stack() argument 443 if (WARN_ON_ONCE(!nr_entries)) in depot_alloc_stack() 447 if (nr_entries > CONFIG_STACKDEPOT_MAX_FRAMES) in depot_alloc_stack() 448 nr_entries = CONFIG_STACKDEPOT_MAX_FRAMES; in depot_alloc_stack() 458 record_size = depot_stack_record_size(stack, nr_entries); in depot_alloc_stack() 469 stack->size = nr_entries; in depot_alloc_stack() 471 memcpy(stack->entries, entries, flex_array_size(stack, entries, nr_entries)); in depot_alloc_stack() 642 unsigned int nr_entries, in stack_depot_save_flags() argument [all …]
|
| /linux/drivers/dma/dw-edma/ |
| H A D | dw-edma-v0-debugfs.c | 101 int nr_entries, struct dentry *dent) in dw_edma_debugfs_create_x32() argument 106 entries = devm_kcalloc(dw->chip->dev, nr_entries, sizeof(*entries), in dw_edma_debugfs_create_x32() 111 for (i = 0; i < nr_entries; i++) { in dw_edma_debugfs_create_x32() 133 int nr_entries; in dw_edma_debugfs_regs_ch() local 135 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_ch() 136 dw_edma_debugfs_create_x32(dw, debugfs_regs, nr_entries, dent); in dw_edma_debugfs_regs_ch() 178 int nr_entries, i; in dw_edma_debugfs_regs_wr() local 183 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_wr() 184 dw_edma_debugfs_create_x32(dw, debugfs_regs, nr_entries, regs_dent); in dw_edma_debugfs_regs_wr() 187 nr_entries = ARRAY_SIZE(debugfs_unroll_regs); in dw_edma_debugfs_regs_wr() [all …]
|
| /linux/drivers/md/persistent-data/ |
| H A D | dm-btree-remove.c | 62 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in node_shift() local 67 BUG_ON(shift > nr_entries); in node_shift() 71 (nr_entries - shift) * sizeof(__le64)); in node_shift() 74 (nr_entries - shift) * value_size); in node_shift() 76 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift() 79 nr_entries * sizeof(__le64)); in node_shift() 82 nr_entries * value_size); in node_shift() 88 uint32_t nr_left = le32_to_cpu(left->header.nr_entries); in node_copy() 131 unsigned int nr_entries = le32_to_cpu(n->header.nr_entries); in delete_at() local 132 unsigned int nr_to_copy = nr_entries - (index + 1); in delete_at() [all …]
|
| H A D | dm-btree.c | 46 int lo = -1, hi = le32_to_cpu(n->header.nr_entries); in bsearch() 77 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in inc_children() local 80 dm_tm_with_runs(tm, value_ptr(n, 0), nr_entries, dm_tm_inc_range); in inc_children() 83 vt->inc(vt->context, value_ptr(n, 0), nr_entries); in inc_children() 90 uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); in insert_at() local 94 if (index > nr_entries || in insert_at() 96 nr_entries >= max_entries) { in insert_at() 104 array_insert(node->keys, sizeof(*node->keys), nr_entries, index, &key_le); in insert_at() 105 array_insert(value_base(node), value_size, nr_entries, index, value); in insert_at() 106 node->header.nr_entries = cpu_to_le32(nr_entries + 1); in insert_at() [all …]
|
| H A D | dm-array.c | 27 __le32 nr_entries; member 114 unsigned int nr_entries = le32_to_cpu(ab->nr_entries); in on_entries() local 116 fn(info->value_type.context, element_at(info, ab, 0), nr_entries); in on_entries() 164 (*ab)->nr_entries = cpu_to_le32(0); in alloc_ablock() 178 uint32_t nr_entries, delta, i; in fill_ablock() local 182 BUG_ON(new_nr < le32_to_cpu(ab->nr_entries)); in fill_ablock() 184 nr_entries = le32_to_cpu(ab->nr_entries); in fill_ablock() 185 delta = new_nr - nr_entries; in fill_ablock() 188 for (i = nr_entries; i < new_nr; i++) in fill_ablock() 190 ab->nr_entries = cpu_to_le32(new_nr); in fill_ablock() [all …]
|
| /linux/kernel/ |
| H A D | stacktrace.c | 24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 32 for (i = 0; i < nr_entries; i++) in stack_trace_print() 48 unsigned int nr_entries, int spaces) in stack_trace_snprint() argument 55 for (i = 0; i < nr_entries && size; i++) { in stack_trace_snprint() 278 return trace.nr_entries; in stack_trace_save() 303 return trace.nr_entries; in stack_trace_save_tsk() 326 return trace.nr_entries; in stack_trace_save_regs() 351 return ret ? ret : trace.nr_entries; in stack_trace_save_tsk_reliable() 371 return trace.nr_entries; in stack_trace_save_user() 392 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries) in filter_irq_stacks() argument [all …]
|
| H A D | backtracetest.c | 44 unsigned int nr_entries; in backtrace_test_saved() local 49 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 0); in backtrace_test_saved() 50 stack_trace_print(entries, nr_entries, 0); in backtrace_test_saved()
|
| /linux/tools/perf/ui/browsers/ |
| H A D | annotate-data.c | 28 int nr_entries; /* # of visible entries: self + descendents */ member 130 bracket->nr_entries = 1; in add_child_entries() 138 entry->nr_entries = 1; in add_child_entries() 148 nr += entry->nr_entries; in count_visible_entries() 165 browser->b.nr_entries = count_visible_entries(browser); in annotated_data_browser__collect_entries() 268 if (uib->nr_entries == 0) in browser__seek() 461 entry->nr_entries = 1; in annotated_data_browser__fold() 470 int nr_entries; in annotated_data_browser__unfold() local 477 nr_entries = 1; /* for self */ in annotated_data_browser__unfold() 482 nr_entries += child->nr_entries; in annotated_data_browser__unfold() [all …]
|
| /linux/arch/x86/kernel/ |
| H A D | e820.c | 81 for (idx = 0; idx < table->nr_entries; idx++) { in _e820__mapped_any() 116 for (idx = 0; idx < e820_table->nr_entries; idx++) { in __e820__mapped_all() 167 u32 idx = table->nr_entries; in __e820__range_add() 182 table->nr_entries++; in __e820__range_add() 210 for (idx = 0; idx < e820_table->nr_entries; idx++) { in e820__print_table() 354 if (table->nr_entries < 2) in e820__update_table() 357 BUG_ON(table->nr_entries > max_nr_entries); in e820__update_table() 360 for (idx = 0; idx < table->nr_entries; idx++) { in e820__update_table() 366 for (idx = 0; idx < 2 * table->nr_entries; idx++) in e820__update_table() 374 for (idx = 0; idx < table->nr_entries; idx++) { in e820__update_table() [all …]
|
| H A D | ldt.c | 81 set_ldt(ldt_slot_va(ldt->slot), ldt->nr_entries); in load_mm_ldt() 83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt() 183 new_ldt->nr_entries = num_entries; in alloc_ldt_struct() 309 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in map_ldt_struct() 361 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in unmap_ldt_struct() 418 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 439 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 440 if (ldt->nr_entries * LDT_ENTRY_SIZE > PAGE_SIZE) in free_ldt_struct() 463 new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); in ldt_dup_context() 470 new_ldt->nr_entries * LDT_ENTRY_SIZE); in ldt_dup_context() [all …]
|
| /linux/mm/kmsan/ |
| H A D | report.c | 89 unsigned int nr_entries, chained_nr_entries, skipnr; in kmsan_print_origin() local 100 nr_entries = stack_depot_fetch(origin, &entries); in kmsan_print_origin() 102 magic = nr_entries ? entries[0] : 0; in kmsan_print_origin() 103 if ((nr_entries == 4) && (magic == KMSAN_ALLOCA_MAGIC_ORIGIN)) { in kmsan_print_origin() 115 if ((nr_entries == 3) && (magic == KMSAN_CHAIN_MAGIC_ORIGIN)) { in kmsan_print_origin() 139 if (nr_entries) { in kmsan_print_origin() 140 skipnr = get_stack_skipnr(entries, nr_entries); in kmsan_print_origin() 141 stack_trace_print(entries + skipnr, nr_entries - skipnr, in kmsan_print_origin()
|
| /linux/tools/testing/selftests/bpf/benchs/ |
| H A D | bench_bpf_hashmap_lookup.c | 24 __u32 nr_entries; member 30 .nr_entries = 500, 91 args.nr_entries = ret; in parse_arg() 121 if (args.nr_entries > args.max_entries) { in validate() 123 args.max_entries, args.nr_entries); in validate() 171 ctx.skel->bss->nr_entries = args.nr_entries; in setup() 172 ctx.skel->bss->nr_loops = args.nr_loops / args.nr_entries; in setup() 188 for (u64 i = 0; i < args.nr_entries; i++) { in setup()
|
| /linux/arch/x86/xen/ |
| H A D | setup.c | 166 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_find_pfn_range() 428 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_foreach_remap_area() 430 if (entry->type == E820_TYPE_RAM || i == xen_e820_table.nr_entries - 1) { in xen_foreach_remap_area() 564 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_ignore_unusable() 582 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_is_e820_reserved() 607 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++, entry++) { in xen_find_free_area() 643 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_e820_swap_entry_with_ram() 652 xen_e820_table.nr_entries; in xen_e820_swap_entry_with_ram() 653 xen_e820_table.nr_entries++; in xen_e820_swap_entry_with_ram() 700 while (mapcnt < xen_e820_table.nr_entries) { in xen_e820_resolve_conflicts() [all …]
|
| /linux/tools/perf/ui/ |
| H A D | browser.c | 108 if (browser->nr_entries == 0) in ui_browser__list_head_seek() 326 if (browser->nr_entries > 1) { in ui_browser__scrollbar_set() 328 (browser->nr_entries - 1)); in ui_browser__scrollbar_set() 358 if (browser->nr_entries == 0 && browser->no_samples_msg) in __ui_browser__refresh() 378 void ui_browser__update_nr_entries(struct ui_browser *browser, u32 nr_entries) in ui_browser__update_nr_entries() argument 380 off_t offset = nr_entries - browser->nr_entries; in ui_browser__update_nr_entries() 382 browser->nr_entries = nr_entries; in ui_browser__update_nr_entries() 434 if (browser->index == browser->nr_entries - 1) in ui_browser__run() 467 if (browser->top_idx + browser->rows > browser->nr_entries - 1) in ui_browser__run() 471 if (browser->index + offset > browser->nr_entries - 1) in ui_browser__run() [all …]
|
| /linux/arch/mips/kernel/ |
| H A D | stacktrace.c | 30 trace->entries[trace->nr_entries++] = addr; in save_raw_context_stack() 31 if (trace->nr_entries >= trace->max_entries) in save_raw_context_stack() 58 trace->entries[trace->nr_entries++] = pc; in save_context_stack() 59 if (trace->nr_entries >= trace->max_entries) in save_context_stack() 83 WARN_ON(trace->nr_entries || !trace->max_entries); in save_stack_trace_tsk()
|
| /linux/tools/perf/trace/beauty/ |
| H A D | ioctl.c | 41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd() 52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd() 63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd() 74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd() 85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd() 98 if (nr < s->nr_entries && s->entries[nr] != NULL) in ioctl__scnprintf_vhost_virtio_cmd() 109 if (nr < strarray__perf_ioctl_cmds.nr_entries && strarray__perf_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_perf_cmd() 120 …if (nr < strarray__usbdevfs_ioctl_cmds.nr_entries && strarray__usbdevfs_ioctl_cmds.entries[nr] != … in ioctl__scnprintf_usbdevfs_cmd()
|
| /linux/drivers/md/ |
| H A D | dm-cache-policy-internal.h | 117 static inline size_t bitset_size_in_bytes(unsigned int nr_entries) in bitset_size_in_bytes() argument 119 return sizeof(unsigned long) * dm_div_up(nr_entries, BITS_PER_LONG); in bitset_size_in_bytes() 122 static inline unsigned long *alloc_bitset(unsigned int nr_entries) in alloc_bitset() argument 124 size_t s = bitset_size_in_bytes(nr_entries); in alloc_bitset() 129 static inline void clear_bitset(void *bitset, unsigned int nr_entries) in clear_bitset() argument 131 size_t s = bitset_size_in_bytes(nr_entries); in clear_bitset()
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | spectrum1_kvdl.c | 122 unsigned int entry_index, nr_entries; in mlxsw_sp1_kvdl_part_alloc() local 124 nr_entries = (info->end_index - info->start_index + 1) / in mlxsw_sp1_kvdl_part_alloc() 126 entry_index = find_first_zero_bit(part->usage, nr_entries); in mlxsw_sp1_kvdl_part_alloc() 127 if (entry_index == nr_entries) in mlxsw_sp1_kvdl_part_alloc() 215 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_init() local 226 nr_entries = div_u64(resource_size, info->alloc_size); in mlxsw_sp1_kvdl_part_init() 227 part = kzalloc_flex(*part, usage, BITS_TO_LONGS(nr_entries)); in mlxsw_sp1_kvdl_part_init() 279 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_occ() local 283 nr_entries = (info->end_index - in mlxsw_sp1_kvdl_part_occ() 286 while ((bit = find_next_bit(part->usage, nr_entries, bit + 1)) in mlxsw_sp1_kvdl_part_occ() [all …]
|
| /linux/arch/sh/kernel/ |
| H A D | stacktrace.c | 33 if (trace->nr_entries < trace->max_entries) in save_stack_address() 34 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 65 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 66 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
| /linux/tools/perf/arch/x86/tests/ |
| H A D | hybrid.c | 29 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_hw_event_with_pmu() 41 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_group_event() 60 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_sw_hw_group_event() 77 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_sw_group_event() 94 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_group_modifier1() 130 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_raw2() 140 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_cache_event() 151 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_pmu() 171 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_group_event_2()
|
| /linux/arch/openrisc/kernel/ |
| H A D | stacktrace.c | 38 if (trace->nr_entries < trace->max_entries) in save_stack_address() 39 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 64 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 65 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
| /linux/include/linux/ |
| H A D | stacktrace.h | 66 void stack_trace_print(const unsigned long *trace, unsigned int nr_entries, 69 unsigned int nr_entries, int spaces); 78 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries); 83 unsigned int nr_entries, max_entries; member
|
| /linux/tools/perf/tests/ |
| H A D | parse-events.c | 112 TEST_ASSERT_EVLIST("wrong number of entries", 1 == evlist->core.nr_entries, evlist); in test__checkevent_tracepoint() 125 TEST_ASSERT_EVLIST("wrong number of entries", evlist->core.nr_entries > 1, evlist); in test__checkevent_tracepoint_multi() 147 TEST_ASSERT_EVLIST("wrong number of entries", 0 != evlist->core.nr_entries, evlist); in test__checkevent_raw() 185 TEST_ASSERT_EVLIST("wrong number of entries", 1 == evlist->core.nr_entries, evlist); in test__checkevent_numeric() 196 TEST_ASSERT_EVLIST("wrong number of entries", 0 != evlist->core.nr_entries, evlist); in test__checkevent_symbolic_name() 210 TEST_ASSERT_EVLIST("wrong number of entries", 0 != evlist->core.nr_entries, evlist); in test__checkevent_symbolic_name_config() 231 TEST_ASSERT_EVLIST("wrong number of entries", 1 == evlist->core.nr_entries, evlist); in test__checkevent_symbolic_alias() 241 TEST_ASSERT_EVLIST("wrong number of entries", 0 != evlist->core.nr_entries, evlist); in test__checkevent_genhw() 254 TEST_ASSERT_EVLIST("wrong number of entries", 1 == evlist->core.nr_entries, evlist); in test__checkevent_breakpoint() 268 TEST_ASSERT_EVLIST("wrong number of entries", 1 == evlist->core.nr_entries, evlis in test__checkevent_breakpoint_x() [all...] |
| H A D | hists_filter.c | 168 hists->nr_entries == 9); in test__hists_filter() 175 hists->nr_entries == hists->nr_non_filtered_entries); in test__hists_filter() 193 hists->nr_entries == 9); in test__hists_filter() 222 hists->nr_entries == 9); in test__hists_filter() 257 hists->nr_entries == 9); in test__hists_filter() 286 hists->nr_entries == 9); in test__hists_filter() 317 hists->nr_entries == 9); in test__hists_filter()
|
| /linux/arch/sparc/kernel/ |
| H A D | stacktrace.c | 58 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 66 if (trace->nr_entries < in __save_stack_trace() 68 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 74 } while (trace->nr_entries < trace->max_entries); in __save_stack_trace()
|