| /linux/lib/ |
| H A D | stackdepot.c | 364 struct stack_record *stack; in depot_pop_free_pool() local 382 stack = current_pool + pool_offset; in depot_pop_free_pool() 385 stack->handle.pool_index_plus_1 = pool_index + 1; in depot_pop_free_pool() 386 stack->handle.offset = pool_offset >> DEPOT_STACK_ALIGN; in depot_pop_free_pool() 387 stack->handle.extra = 0; in depot_pop_free_pool() 388 INIT_LIST_HEAD(&stack->hash_list); in depot_pop_free_pool() 392 return stack; in depot_pop_free_pool() 399 struct stack_record *stack; in depot_pop_free() local 412 stack = list_first_entry(&free_stacks, struct stack_record, free_list); in depot_pop_free() 413 if (!poll_state_synchronize_rcu(stack->rcu_state)) in depot_pop_free() [all …]
|
| /linux/drivers/misc/altera-stapl/ |
| H A D | altera.c | 213 long *stack = astate->stack; in altera_execute() local 528 stack[stack_ptr] = stack[stack_ptr - 1]; in altera_execute() 534 swap(stack[stack_ptr - 2], stack[stack_ptr - 1]); in altera_execute() 539 stack[stack_ptr - 1] += stack[stack_ptr]; in altera_execute() 545 stack[stack_ptr - 1] -= stack[stack_ptr]; in altera_execute() 551 stack[stack_ptr - 1] *= stack[stack_ptr]; in altera_execute() 557 stack[stack_ptr - 1] /= stack[stack_ptr]; in altera_execute() 563 stack[stack_ptr - 1] %= stack[stack_ptr]; in altera_execute() 569 stack[stack_ptr - 1] <<= stack[stack_ptr]; in altera_execute() 575 stack[stack_ptr - 1] >>= stack[stack_ptr]; in altera_execute() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_global_func_ctx_args.c | 11 static long stack[256]; variable 19 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in kprobe_typedef_ctx_subprog() 50 return bpf_get_stack((void *)ctx, &stack, sizeof(stack), 0); in kprobe_struct_ctx_subprog() 67 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in kprobe_workaround_ctx_subprog() 83 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in raw_tp_ctx_subprog() 99 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in raw_tp_writable_ctx_subprog() 115 return bpf_get_stack(ctx, &stack, sizeof(stack), 0); in perf_event_ctx_subprog() 130 return bpf_get_stack(ctx, stack, sizeof(stack), 0); in subprog_ctx_tag() 142 return bpf_get_stack(ctx1, stack, sizeof(stack), 0) + in subprog_multi_ctx_tags() 144 bpf_get_stack(ctx2, stack, sizeof(stack), 0); in subprog_multi_ctx_tags()
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | build_id.c | 10 static void print_stack(struct bpf_stack_build_id *stack, int frame_cnt) in print_stack() argument 16 switch (stack[i].status) { in print_stack() 23 printf("%02hhx", (unsigned)stack[i].build_id[j]); in print_stack() 24 printf(" OFFSET = %llx", (unsigned long long)stack[i].offset); in print_stack() 27 printf("IP = %llx", (unsigned long long)stack[i].ip); in print_stack() 30 printf("UNEXPECTED STATUS %d ", stack[i].status); in print_stack() 40 struct bpf_stack_build_id *stack; in subtest_nofault() local 59 stack = skel->bss->stack_nofault; in subtest_nofault() 62 print_stack(stack, frame_cnt); in subtest_nofault() 65 ASSERT_EQ(stack[0].status, BPF_STACK_BUILD_ID_VALID, "build_id_status"); in subtest_nofault() [all …]
|
| /linux/Documentation/mm/ |
| H A D | vmalloced-kernel-stacks.rst | 21 Kernel stack overflows are often hard to debug and make the kernel 25 Virtually mapped kernel stacks with guard pages cause kernel stack 31 causes reliable faults when the stack overflows. The usability of 32 the stack trace after overflow and response to the overflow itself 49 needs to work while the stack points to a virtual address with 51 most likely) needs to ensure that the stack's page table entries 52 are populated before running on a possibly unpopulated stack. 53 - If the stack overflows into a guard page, something reasonable 64 with guard pages. This causes kernel stack overflows to be caught 75 VMAP_STACK is enabled, it is not possible to run DMA on stack [all …]
|
| /linux/mm/ |
| H A D | page_owner.c | 39 struct stack { struct 41 struct stack *next; argument 43 static struct stack dummy_stack; 44 static struct stack failure_stack; 45 static struct stack *stack_list; 53 struct stack *stack; member 178 struct stack *stack; in add_stack_record_to_list() local 184 stack = kmalloc_obj(*stack, gfp_nested_mask(gfp_mask)); in add_stack_record_to_list() 185 if (!stack) { in add_stack_record_to_list() 191 stack->stack_record = stack_record; in add_stack_record_to_list() [all …]
|
| /linux/arch/nios2/kernel/ |
| H A D | traps.c | 60 void show_stack(struct task_struct *task, unsigned long *stack, in show_stack() argument 66 if (!stack) { in show_stack() 68 stack = (unsigned long *)task->thread.ksp; in show_stack() 70 stack = (unsigned long *)&stack; in show_stack() 73 addr = (unsigned long) stack; in show_stack() 76 printk("%sStack from %08lx:", loglvl, (unsigned long)stack); in show_stack() 78 if (stack + 1 > endstack) in show_stack() 82 printk("%s %08lx", loglvl, *stack++); in show_stack() 87 while (stack + 1 <= endstack) { in show_stack() 88 addr = *stack++; in show_stack()
|
| /linux/Documentation/arch/x86/ |
| H A D | kernel-stacks.rst | 14 Like all other architectures, x86_64 has a kernel stack for every 17 zombie. While the thread is in user space the kernel stack is empty 25 * Interrupt stack. IRQ_STACK_SIZE 29 kernel switches from the current task to the interrupt stack. Like 32 of every per thread stack. 34 The interrupt stack is also used when processing a softirq. 36 Switching to the kernel interrupt stack is done by software based on a 41 to automatically switch to a new stack for designated events such as 46 point to dedicated stacks; each stack can be a different size. 50 loads such a descriptor, the hardware automatically sets the new stack [all …]
|
| /linux/mm/kmsan/ |
| H A D | init.c | 149 static void smallstack_push(struct smallstack *stack, struct page *pages) in smallstack_push() argument 151 KMSAN_WARN_ON(stack->index == MAX_BLOCKS); in smallstack_push() 152 stack->items[stack->index] = pages; in smallstack_push() 153 stack->index++; in smallstack_push() 157 static struct page *smallstack_pop(struct smallstack *stack) in smallstack_pop() argument 161 KMSAN_WARN_ON(stack->index == 0); in smallstack_pop() 162 stack->index--; in smallstack_pop() 163 ret = stack->items[stack->index]; in smallstack_pop() 164 stack->items[stack->index] = NULL; in smallstack_pop()
|
| /linux/arch/openrisc/kernel/ |
| H A D | unwinder.c | 60 void unwind_stack(void *data, unsigned long *stack, in unwind_stack() argument 67 while (!kstack_end(stack)) { in unwind_stack() 68 frameinfo = container_of(stack, in unwind_stack() 83 stack++; in unwind_stack() 93 void unwind_stack(void *data, unsigned long *stack, in unwind_stack() argument 98 while (!kstack_end(stack)) { in unwind_stack() 99 addr = *stack++; in unwind_stack()
|
| /linux/include/linux/sched/ |
| H A D | task_stack.h | 23 return task->stack; in task_stack_page() 31 return (unsigned long *)((unsigned long)task->stack + THREAD_SIZE) - 1; in end_of_stack() 33 return task->stack; in end_of_stack() 39 #define task_stack_page(task) ((void *)(task)->stack) 91 void *stack = task_stack_page(current); in object_is_on_stack() local 94 return (obj >= stack) && (obj < (stack + THREAD_SIZE)); in object_is_on_stack()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | pkey_sighandler_tests.c | 80 static inline long clone_raw(unsigned long flags, void *stack, in clone_raw() argument 84 long a2 = (long)stack; in clone_raw() 179 stack_t *stack = ptr; in thread_segv_maperr_ptr() local 187 syscall_raw(SYS_sigaltstack, (long)stack, 0, 0, 0, 0, 0); in thread_segv_maperr_ptr() 286 void *stack; in test_sigsegv_handler_with_different_pkey_for_stack() local 302 stack = mmap(0, STACK_SIZE, PROT_READ | PROT_WRITE, in test_sigsegv_handler_with_different_pkey_for_stack() 305 assert(stack != MAP_FAILED); in test_sigsegv_handler_with_different_pkey_for_stack() 315 sys_mprotect_pkey(stack, STACK_SIZE, PROT_READ | PROT_WRITE, pkey); in test_sigsegv_handler_with_different_pkey_for_stack() 330 stack + STACK_SIZE, in test_sigsegv_handler_with_different_pkey_for_stack() 405 stack_t *stack = ptr; in thread_sigusr2_self() local [all …]
|
| /linux/scripts/kconfig/ |
| H A D | symbol.c | 1089 static void dep_stack_insert(struct dep_stack *stack, struct symbol *sym) in dep_stack_insert() argument 1091 memset(stack, 0, sizeof(*stack)); in dep_stack_insert() 1093 check_top->next = stack; in dep_stack_insert() 1094 stack->prev = check_top; in dep_stack_insert() 1095 stack->sym = sym; in dep_stack_insert() 1096 check_top = stack; in dep_stack_insert() 1113 struct dep_stack *stack; in sym_check_print_recursive() local 1125 for (stack = check_top; stack != NULL; stack = stack->prev) in sym_check_print_recursive() 1126 if (stack->sym == last_sym) in sym_check_print_recursive() 1128 if (!stack) { in sym_check_print_recursive() [all …]
|
| /linux/arch/um/kernel/ |
| H A D | sysrq.c | 29 void show_stack(struct task_struct *task, unsigned long *stack, in show_stack() argument 35 if (!stack) in show_stack() 36 stack = get_stack_pointer(task, segv_regs); in show_stack() 40 if (kstack_end(stack)) in show_stack() 44 pr_cont(" %08lx", READ_ONCE_NOCHECK(*stack)); in show_stack() 45 stack++; in show_stack()
|
| /linux/tools/perf/scripts/python/ |
| H A D | stackcollapse.py | 97 stack = list() 103 stack.append(tidy_function_name(entry['sym']['name'], 108 stack.append(tidy_function_name(param_dict['symbol'], 119 stack.append(comm) 121 stack_string = ';'.join(reversed(stack)) 126 for stack in list: 127 print("%s %d" % (stack, lines[stack]))
|
| /linux/arch/x86/include/asm/ |
| H A D | stacktrace.h | 31 bool in_task_stack(unsigned long *stack, struct task_struct *task, 34 bool in_entry_stack(unsigned long *stack, struct stack_info *info); 36 int get_stack_info(unsigned long *stack, struct task_struct *task, 38 bool get_stack_info_noinstr(unsigned long *stack, struct task_struct *task, 42 bool get_stack_guard_info(unsigned long *stack, struct stack_info *info) in get_stack_guard_info() argument 45 if (get_stack_info_noinstr(stack, current, info)) in get_stack_guard_info() 48 return get_stack_info_noinstr((void *)stack + PAGE_SIZE, current, info); in get_stack_guard_info()
|
| /linux/arch/powerpc/kernel/ |
| H A D | stacktrace.c | 44 unsigned long *stack = (unsigned long *) sp; in arch_stack_walk() local 50 newsp = stack[0]; in arch_stack_walk() 51 ip = stack[STACK_FRAME_LR_SAVE]; in arch_stack_walk() 96 unsigned long *stack = (unsigned long *) sp; in arch_stack_walk_reliable() local 103 newsp = stack[0]; in arch_stack_walk_reliable() 123 stack[STACK_INT_FRAME_MARKER_LONGS] == STACK_FRAME_REGS_MARKER) { in arch_stack_walk_reliable() 128 ip = stack[STACK_FRAME_LR_SAVE]; in arch_stack_walk_reliable() 136 ip = ftrace_graph_ret_addr(task, &graph_idx, ip, stack); in arch_stack_walk_reliable()
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-ocfs2 | 38 * 'o2cb' - The classic o2cb cluster stack that ocfs2 has 62 the cluster stack in use. The contents may change 63 when all filesystems are unmounted and the cluster stack 71 of current ocfs2 cluster stack. This value is set by 72 userspace tools when bringing the cluster stack online. 74 Cluster stack names are 4 characters in length. 76 When the 'o2cb' cluster stack is used, the 'o2cb' cluster 81 cluster stack on a single line. 83 Writing a new stack name to this file changes the current 84 cluster stack unless there are mounted ocfs2 filesystems. [all …]
|
| /linux/arch/um/os-Linux/skas/ |
| H A D | mem.c | 27 struct stub_data *proc_data = (void *)mm_idp->stack; in syscall_stub_dump_error() 59 unsigned long *stack) in check_init_stack() argument 61 if (stack == NULL) { in check_init_stack() 62 stack = (unsigned long *) mm_idp->stack + 2; in check_init_stack() 63 *stack = 0; in check_init_stack() 65 return stack; in check_init_stack() 89 struct stub_data *proc_data = (void *)mm_idp->stack; in do_syscall_stub() 160 struct stub_data *proc_data = (struct stub_data *) mm_idp->stack; in syscall_stub_alloc() 183 struct stub_data *proc_data = (void *) mm_idp->stack; in syscall_stub_get_previous()
|
| /linux/tools/perf/util/ |
| H A D | annotate-data.c | 185 struct type_state_stack *stack, *tmp; in exit_type_state() local 187 list_for_each_entry_safe(stack, tmp, &state->stack_vars, list) { in exit_type_state() 188 list_del(&stack->list); in exit_type_state() 189 free(stack); in exit_type_state() 566 struct type_state_stack *stack; in find_stack_state() local 568 list_for_each_entry(stack, &state->stack_vars, list) { in find_stack_state() 569 if (offset == stack->offset) in find_stack_state() 570 return stack; in find_stack_state() 572 if (stack->compound && stack->offset < offset && in find_stack_state() 573 offset < stack->offset + stack->size) in find_stack_state() [all …]
|
| H A D | thread-stack.c | 89 struct thread_stack_entry *stack; member 126 new_stack = realloc(ts->stack, sz); in thread_stack__grow() 130 ts->stack = new_stack; in thread_stack__grow() 201 if (!ts->stack && in thread_stack__new() 220 if (!ts->stack) in thread__cpu_stack() 251 ts->stack[ts->cnt].trace_end = trace_end; in thread_stack__push() 252 ts->stack[ts->cnt++].ret_addr = ret_addr; in thread_stack__push() 271 if (ts->stack[--i].ret_addr == ret_addr) { in thread_stack__pop() 283 if (ts->stack[--i].trace_end) in thread_stack__pop_trace_end() 295 return ts->stack[ts->cnt - 1].cp->in_kernel; in thread_stack__in_kernel() [all …]
|
| /linux/arch/loongarch/kernel/ |
| H A D | process.c | 174 unsigned long usp = args->stack; in copy_thread() 258 bool in_irq_stack(unsigned long stack, struct stack_info *info) in in_irq_stack() argument 264 if (stack < begin || stack >= end) in in_irq_stack() 279 bool in_task_stack(unsigned long stack, struct task_struct *task, in in_task_stack() argument 285 if (stack < begin || stack >= end) in in_task_stack() 296 int get_stack_info(unsigned long stack, struct task_struct *task, in get_stack_info() argument 301 if (!stack || stack & (SZREG - 1)) in get_stack_info() 304 if (in_task_stack(stack, task, info)) in get_stack_info() 310 if (in_irq_stack(stack, info)) in get_stack_info()
|
| /linux/fs/jfs/ |
| H A D | jfs_btree.h | 108 struct btframe stack[MAXTREEHEIGHT]; member 112 (btstack)->top = (btstack)->stack 115 ( (btstack)->top == &((btstack)->stack[MAXTREEHEIGHT-1])) 126 ( (btstack)->top == (btstack)->stack ? NULL : --(btstack)->top ) 129 ( (btstack)->top == (btstack)->stack ? NULL : (btstack)->top ) 137 (long long)btstack->stack[i].bn, in BT_STACK_DUMP() 138 btstack->stack[i].index); in BT_STACK_DUMP()
|
| /linux/arch/loongarch/ |
| H A D | Kconfig.debug | 5 This determines which method will be used for unwinding kernel stack 6 traces for panics, oopses, bugs, warnings, perf, /proc/<pid>/stack, 12 This option enables the "guess" unwinder for unwinding kernel stack 13 traces. It scans the stack and reports every kernel text address it 23 This option enables the "prologue" unwinder for unwinding kernel stack 24 traces. It unwind the stack frame based on prologue code analyze. Symbol 35 unwinding kernel stack traces. It uses a custom data format which is
|
| /linux/security/apparmor/ |
| H A D | domain.c | 94 bool stack, aa_state_t state) in match_component() argument 99 if (stack) in match_component() 129 struct aa_label *label, bool stack, in label_compound_match() argument 142 state = match_component(profile, tp, stack, state); in label_compound_match() 191 struct aa_label *label, bool stack, in label_components_match() argument 206 state = match_component(profile, tp, stack, start); in label_components_match() 223 state = match_component(profile, tp, stack, start); in label_components_match() 255 bool stack, aa_state_t state, bool inview, u32 request, in label_match() argument 261 error = label_compound_match(profile, label, stack, state, inview, in label_match() 267 return label_components_match(profile, label, stack, state, inview, in label_match() [all …]
|