/linux/fs/xfs/libxfs/ |
H A D | xfs_iext_tree.c | 461 int *nr_entries) in xfs_iext_split_node() argument 473 *nr_entries = 0; in xfs_iext_split_node() 489 *nr_entries = nr_move; in xfs_iext_split_node() 491 *nr_entries = nr_keep; in xfs_iext_split_node() 507 int i, pos, nr_entries; in xfs_iext_insert_node() local 516 nr_entries = xfs_iext_node_nr_entries(node, pos); in xfs_iext_insert_node() 518 ASSERT(pos >= nr_entries || xfs_iext_key_cmp(node, pos, offset) != 0); in xfs_iext_insert_node() 519 ASSERT(nr_entries <= KEYS_PER_NODE); in xfs_iext_insert_node() 521 if (nr_entries == KEYS_PER_NODE) in xfs_iext_insert_node() 522 new = xfs_iext_split_node(&node, &pos, &nr_entries); in xfs_iext_insert_node() [all …]
|
/linux/lib/ |
H A D | stackdepot.c | 369 static inline size_t depot_stack_record_size(struct stack_record *s, unsigned int nr_entries) in depot_stack_record_size() argument 371 const size_t used = flex_array_size(s, entries, nr_entries); in depot_stack_record_size() 381 depot_alloc_stack(unsigned long *entries, unsigned int nr_entries, u32 hash, depot_flags_t flags, v… in depot_alloc_stack() argument 389 if (WARN_ON_ONCE(!nr_entries)) in depot_alloc_stack() 393 if (nr_entries > CONFIG_STACKDEPOT_MAX_FRAMES) in depot_alloc_stack() 394 nr_entries = CONFIG_STACKDEPOT_MAX_FRAMES; in depot_alloc_stack() 404 record_size = depot_stack_record_size(stack, nr_entries); in depot_alloc_stack() 415 stack->size = nr_entries; in depot_alloc_stack() 417 memcpy(stack->entries, entries, flex_array_size(stack, entries, nr_entries)); in depot_alloc_stack() 585 unsigned int nr_entries, in stack_depot_save_flags() argument [all …]
|
/linux/drivers/dma/dw-edma/ |
H A D | dw-edma-v0-debugfs.c | 101 int nr_entries, struct dentry *dent) in dw_edma_debugfs_create_x32() argument 106 entries = devm_kcalloc(dw->chip->dev, nr_entries, sizeof(*entries), in dw_edma_debugfs_create_x32() 111 for (i = 0; i < nr_entries; i++) { in dw_edma_debugfs_create_x32() 133 int nr_entries; in dw_edma_debugfs_regs_ch() local 135 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_ch() 136 dw_edma_debugfs_create_x32(dw, debugfs_regs, nr_entries, dent); in dw_edma_debugfs_regs_ch() 178 int nr_entries, i; in dw_edma_debugfs_regs_wr() local 183 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_wr() 184 dw_edma_debugfs_create_x32(dw, debugfs_regs, nr_entries, regs_dent); in dw_edma_debugfs_regs_wr() 187 nr_entries = ARRAY_SIZE(debugfs_unroll_regs); in dw_edma_debugfs_regs_wr() [all …]
|
/linux/drivers/md/persistent-data/ |
H A D | dm-btree-remove.c | 62 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in node_shift() local 67 BUG_ON(shift > nr_entries); in node_shift() 71 (nr_entries - shift) * sizeof(__le64)); in node_shift() 74 (nr_entries - shift) * value_size); in node_shift() 76 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift() 79 nr_entries * sizeof(__le64)); in node_shift() 82 nr_entries * value_size); in node_shift() 88 uint32_t nr_left = le32_to_cpu(left->header.nr_entries); in node_copy() 131 unsigned int nr_entries = le32_to_cpu(n->header.nr_entries); in delete_at() local 132 unsigned int nr_to_copy = nr_entries - (index + 1); in delete_at() [all …]
|
H A D | dm-btree.c | 46 int lo = -1, hi = le32_to_cpu(n->header.nr_entries); in bsearch() 77 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in inc_children() local 80 dm_tm_with_runs(tm, value_ptr(n, 0), nr_entries, dm_tm_inc_range); in inc_children() 83 vt->inc(vt->context, value_ptr(n, 0), nr_entries); in inc_children() 90 uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); in insert_at() local 94 if (index > nr_entries || in insert_at() 96 nr_entries >= max_entries) { in insert_at() 104 array_insert(node->keys, sizeof(*node->keys), nr_entries, index, &key_le); in insert_at() 105 array_insert(value_base(node), value_size, nr_entries, index, value); in insert_at() 106 node->header.nr_entries = cpu_to_le32(nr_entries + 1); in insert_at() [all …]
|
H A D | dm-array.c | 27 __le32 nr_entries; member 114 unsigned int nr_entries = le32_to_cpu(ab->nr_entries); in on_entries() local 116 fn(info->value_type.context, element_at(info, ab, 0), nr_entries); in on_entries() 164 (*ab)->nr_entries = cpu_to_le32(0); in alloc_ablock() 178 uint32_t nr_entries, delta, i; in fill_ablock() local 182 BUG_ON(new_nr < le32_to_cpu(ab->nr_entries)); in fill_ablock() 184 nr_entries = le32_to_cpu(ab->nr_entries); in fill_ablock() 185 delta = new_nr - nr_entries; in fill_ablock() 188 for (i = nr_entries; i < new_nr; i++) in fill_ablock() 190 ab->nr_entries = cpu_to_le32(new_nr); in fill_ablock() [all …]
|
/linux/kernel/ |
H A D | stacktrace.c | 24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 32 for (i = 0; i < nr_entries; i++) in stack_trace_print() 48 unsigned int nr_entries, int spaces) in stack_trace_snprint() argument 55 for (i = 0; i < nr_entries && size; i++) { in stack_trace_snprint() 278 return trace.nr_entries; in stack_trace_save() 303 return trace.nr_entries; in stack_trace_save_tsk() 326 return trace.nr_entries; in stack_trace_save_regs() 351 return ret ? ret : trace.nr_entries; in stack_trace_save_tsk_reliable() 371 return trace.nr_entries; in stack_trace_save_user() 392 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries) in filter_irq_stacks() argument [all …]
|
/linux/arch/x86/kernel/ |
H A D | e820.c | 82 for (i = 0; i < table->nr_entries; i++) { in _e820__mapped_any() 117 for (i = 0; i < e820_table->nr_entries; i++) { in __e820__mapped_all() 168 int x = table->nr_entries; in __e820__range_add() 179 table->nr_entries++; in __e820__range_add() 207 for (i = 0; i < e820_table->nr_entries; i++) { in e820__print_table() 332 if (table->nr_entries < 2) in e820__update_table() 335 BUG_ON(table->nr_entries > max_nr_entries); in e820__update_table() 338 for (i = 0; i < table->nr_entries; i++) { in e820__update_table() 344 for (i = 0; i < 2 * table->nr_entries; i++) in e820__update_table() 352 for (i = 0; i < table->nr_entries; i++) { in e820__update_table() [all …]
|
H A D | ldt.c | 81 set_ldt(ldt_slot_va(ldt->slot), ldt->nr_entries); in load_mm_ldt() 83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt() 183 new_ldt->nr_entries = num_entries; in alloc_ldt_struct() 309 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in map_ldt_struct() 361 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in unmap_ldt_struct() 418 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 439 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 440 if (ldt->nr_entries * LDT_ENTRY_SIZE > PAGE_SIZE) in free_ldt_struct() 463 new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); in ldt_dup_context() 470 new_ldt->nr_entries * LDT_ENTRY_SIZE); in ldt_dup_context() [all …]
|
/linux/tools/perf/ui/browsers/ |
H A D | annotate-data.c | 28 int nr_entries; /* # of visible entries: self + descendents */ member 130 bracket->nr_entries = 1; in add_child_entries() 138 entry->nr_entries = 1; in add_child_entries() 148 nr += entry->nr_entries; in count_visible_entries() 165 browser->b.nr_entries = count_visible_entries(browser); in annotated_data_browser__collect_entries() 268 if (uib->nr_entries == 0) in browser__seek() 461 entry->nr_entries = 1; in annotated_data_browser__fold() 470 int nr_entries; in annotated_data_browser__unfold() local 477 nr_entries = 1; /* for self */ in annotated_data_browser__unfold() 482 nr_entries += child->nr_entries; in annotated_data_browser__unfold() [all …]
|
/linux/tools/perf/util/ |
H A D | syscalltbl.c | 82 int nr_entries = 0, i, j; in syscalltbl__init_native() 87 ++nr_entries; in syscalltbl__init_native() 89 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native() 101 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native() 102 tbl->syscalls.nr_entries = nr_entries; 133 tbl->syscalls.nr_entries, sizeof(*sc), in syscalltbl__id() 143 return idx < tbl->syscalls.nr_entries ? syscalls[idx].id : -1; in syscalltbl__strglobmatch_next() 151 for (i = *idx + 1; i < tbl->syscalls.nr_entries; ++i) { in syscalltbl__strglobmatch_next() 78 int nr_entries = 0, i, j; syscalltbl__init_native() local
|
/linux/mm/kmsan/ |
H A D | report.c | 89 unsigned int nr_entries, chained_nr_entries, skipnr; in kmsan_print_origin() local 100 nr_entries = stack_depot_fetch(origin, &entries); in kmsan_print_origin() 102 magic = nr_entries ? entries[0] : 0; in kmsan_print_origin() 103 if ((nr_entries == 4) && (magic == KMSAN_ALLOCA_MAGIC_ORIGIN)) { in kmsan_print_origin() 115 if ((nr_entries == 3) && (magic == KMSAN_CHAIN_MAGIC_ORIGIN)) { in kmsan_print_origin() 139 if (nr_entries) { in kmsan_print_origin() 140 skipnr = get_stack_skipnr(entries, nr_entries); in kmsan_print_origin() 141 stack_trace_print(entries + skipnr, nr_entries - skipnr, in kmsan_print_origin()
|
/linux/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bpf_hashmap_lookup.c | 24 __u32 nr_entries; member 30 .nr_entries = 500, 91 args.nr_entries = ret; in parse_arg() 121 if (args.nr_entries > args.max_entries) { in validate() 123 args.max_entries, args.nr_entries); in validate() 171 ctx.skel->bss->nr_entries = args.nr_entries; in setup() 172 ctx.skel->bss->nr_loops = args.nr_loops / args.nr_entries; in setup() 188 for (u64 i = 0; i < args.nr_entries; i++) { in setup()
|
/linux/tools/perf/ui/ |
H A D | browser.c | 108 if (browser->nr_entries == 0) in ui_browser__list_head_seek() 318 if (browser->nr_entries > 1) { in ui_browser__scrollbar_set() 320 (browser->nr_entries - 1)); in ui_browser__scrollbar_set() 350 if (browser->nr_entries == 0 && browser->no_samples_msg) in __ui_browser__refresh() 365 * Here we're updating nr_entries _after_ we started browsing, i.e. we have to 370 void ui_browser__update_nr_entries(struct ui_browser *browser, u32 nr_entries) in ui_browser__update_nr_entries() 372 off_t offset = nr_entries - browser->nr_entries; in ui_browser__update_nr_entries() 374 browser->nr_entries = nr_entries; in ui_browser__update_nr_entries() 368 ui_browser__update_nr_entries(struct ui_browser * browser,u32 nr_entries) ui_browser__update_nr_entries() argument [all...] |
/linux/arch/x86/xen/ |
H A D | setup.c | 169 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_find_pfn_range() 431 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_foreach_remap_area() 433 if (entry->type == E820_TYPE_RAM || i == xen_e820_table.nr_entries - 1) { in xen_foreach_remap_area() 567 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_ignore_unusable() 585 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_is_e820_reserved() 610 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++, entry++) { in xen_find_free_area() 646 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_e820_swap_entry_with_ram() 655 xen_e820_table.nr_entries; in xen_e820_swap_entry_with_ram() 656 xen_e820_table.nr_entries++; in xen_e820_swap_entry_with_ram() 703 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_e820_resolve_conflicts() [all …]
|
/linux/arch/mips/kernel/ |
H A D | stacktrace.c | 30 trace->entries[trace->nr_entries++] = addr; in save_raw_context_stack() 31 if (trace->nr_entries >= trace->max_entries) in save_raw_context_stack() 58 trace->entries[trace->nr_entries++] = pc; in save_context_stack() 59 if (trace->nr_entries >= trace->max_entries) in save_context_stack() 83 WARN_ON(trace->nr_entries || !trace->max_entries); in save_stack_trace_tsk()
|
/linux/tools/perf/trace/beauty/ |
H A D | ioctl.c | 41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd() 52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd() 63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd() 74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd() 85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd() 98 if (nr < s->nr_entries && s->entries[nr] != NULL) in ioctl__scnprintf_vhost_virtio_cmd() 109 if (nr < strarray__perf_ioctl_cmds.nr_entries && strarray__perf_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_perf_cmd() 120 …if (nr < strarray__usbdevfs_ioctl_cmds.nr_entries && strarray__usbdevfs_ioctl_cmds.entries[nr] != … in ioctl__scnprintf_usbdevfs_cmd()
|
/linux/drivers/md/ |
H A D | dm-cache-policy-internal.h | 117 static inline size_t bitset_size_in_bytes(unsigned int nr_entries) in bitset_size_in_bytes() argument 119 return sizeof(unsigned long) * dm_div_up(nr_entries, BITS_PER_LONG); in bitset_size_in_bytes() 122 static inline unsigned long *alloc_bitset(unsigned int nr_entries) in alloc_bitset() argument 124 size_t s = bitset_size_in_bytes(nr_entries); in alloc_bitset() 129 static inline void clear_bitset(void *bitset, unsigned int nr_entries) in clear_bitset() argument 131 size_t s = bitset_size_in_bytes(nr_entries); in clear_bitset()
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum1_kvdl.c | 122 unsigned int entry_index, nr_entries; in mlxsw_sp1_kvdl_part_alloc() local 124 nr_entries = (info->end_index - info->start_index + 1) / in mlxsw_sp1_kvdl_part_alloc() 126 entry_index = find_first_zero_bit(part->usage, nr_entries); in mlxsw_sp1_kvdl_part_alloc() 127 if (entry_index == nr_entries) in mlxsw_sp1_kvdl_part_alloc() 215 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_init() local 226 nr_entries = div_u64(resource_size, info->alloc_size); in mlxsw_sp1_kvdl_part_init() 227 part = kzalloc(struct_size(part, usage, BITS_TO_LONGS(nr_entries)), in mlxsw_sp1_kvdl_part_init() 280 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_occ() local 284 nr_entries = (info->end_index - in mlxsw_sp1_kvdl_part_occ() 287 while ((bit = find_next_bit(part->usage, nr_entries, bit + 1)) in mlxsw_sp1_kvdl_part_occ() [all …]
|
/linux/arch/sh/kernel/ |
H A D | stacktrace.c | 33 if (trace->nr_entries < trace->max_entries) in save_stack_address() 34 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 65 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 66 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
/linux/tools/perf/arch/x86/tests/ |
H A D | hybrid.c | 29 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_hw_event_with_pmu() 41 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_group_event() 60 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_sw_hw_group_event() 77 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_sw_group_event() 94 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_group_modifier1() 130 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_raw2() 140 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_cache_event() 151 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_pmu() 171 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_group_event_2()
|
/linux/arch/openrisc/kernel/ |
H A D | stacktrace.c | 38 if (trace->nr_entries < trace->max_entries) in save_stack_address() 39 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 64 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 65 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
/linux/include/linux/ |
H A D | stacktrace.h | 66 void stack_trace_print(const unsigned long *trace, unsigned int nr_entries, 69 unsigned int nr_entries, int spaces); 78 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries); 83 unsigned int nr_entries, max_entries; member
|
/linux/arch/sparc/kernel/ |
H A D | stacktrace.c | 58 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 66 if (trace->nr_entries < in __save_stack_trace() 68 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 74 } while (trace->nr_entries < trace->max_entries); in __save_stack_trace()
|
/linux/tools/perf/arch/x86/util/ |
H A D | iostat.c | 71 int nr_entries; member 103 for (idx = 0; idx < list->nr_entries; idx++) in iio_root_ports_list_free() 117 for (idx = 0; idx < list->nr_entries; idx++) { in iio_root_port_find_by_notation() 132 rp->idx = list->nr_entries++; in iio_root_ports_list_insert() 134 list->nr_entries * sizeof(*list->rps)); in iio_root_ports_list_insert() 283 if (tmp_list->nr_entries == 0) { in iio_root_ports_list_filter() 316 for (idx = 0; idx < list->nr_entries; idx++) { in iostat_event_group() 328 list->nr_entries = 0; in iostat_event_group() 337 if (evlist->core.nr_entries > 0) { in iostat_prepare()
|