| /linux/Documentation/arch/x86/ |
| H A D | kernel-stacks.rst | 4 Kernel Stacks 7 Kernel stacks on x86-64 bit 15 active thread. These thread stacks are THREAD_SIZE (4*PAGE_SIZE) big. 16 These stacks contain useful data as long as a thread is alive or a 20 In addition to the per thread stacks, there are specialized stacks 21 associated with each CPU. These stacks are only used while the kernel 23 specialized stacks contain no useful data. The main CPU stacks are: 30 the split thread and interrupt stacks on i386, this gives more room 38 hardware stacks cannot nest without races. 46 point to dedicated stacks; each stack can be a different size. [all …]
|
| H A D | xstate.rst | 18 alternate signal stacks, often using MINSIGSTKSZ which is typically 2KB. 23 However, that means that the size of signal stacks is dynamic, not static,
|
| H A D | index.rst | 16 kernel-stacks
|
| /linux/Documentation/mm/ |
| H A D | vmalloced-kernel-stacks.rst | 15 series that introduced the `Virtually Mapped Kernel Stacks feature 25 Virtually mapped kernel stacks with guard pages cause kernel stack 30 support for virtually mapped stacks with guard pages. This feature 42 Architectures that can support Virtually Mapped Kernel Stacks should 45 - vmalloc space must be large enough to hold many kernel stacks. This 47 - Stacks in vmalloc space need to work reliably. For example, if 61 mapped task stacks. This option depends on HAVE_ARCH_VMAP_STACK. 63 - Enable this if you want the use virtually-mapped kernel stacks 94 - Allocated stacks are cached and later reused by new threads, so memcg 95 accounting is performed manually on assigning/releasing stacks to tasks. [all …]
|
| H A D | page_owner.rst | 27 It can also be used to show all the stacks and their current number of 31 identifier of all the stacks (without stack traces) and their number of 33 can be matched with stacks later (show_handles and show_stacks_handles). 79 cat /sys/kernel/debug/page_owner_stacks/show_stacks > stacks.txt 80 cat stacks.txt
|
| /linux/arch/sh/ |
| H A D | Kconfig.debug | 37 bool "Use 4Kb for kernel stacks instead of 8Kb" 44 will also use IRQ stacks to compensate for the reduced stackspace. 47 bool "Use separate kernel stacks when processing interrupts" 50 If you say Y here the kernel will use separate kernel stacks 52 overflowing the process kernel stacks.
|
| /linux/Documentation/arch/arm64/ |
| H A D | gcs.rst | 32 control stacks with checks to ensure that the new stack is a valid 37 shadow stacks rather than GCS. 61 2. Enabling and disabling Guarded Control Stacks 113 3. Allocation of Guarded Control Stacks 129 * Additional Guarded Control Stacks can be allocated using the 132 * Stacks allocated using map_shadow_stack() can optionally have an end of 140 * Stacks allocated using map_shadow_stack() must have a size which is a
|
| /linux/lib/ |
| H A D | ref_tracker.c | 29 } stacks[]; member 77 stats = kmalloc_flex(*stats, stacks, limit, GFP_NOWAIT); in ref_tracker_get_stats() 89 if (stats->stacks[i].stack_handle == stack) in ref_tracker_get_stats() 94 stats->stacks[i].stack_handle = stack; in ref_tracker_get_stats() 95 stats->stacks[i].count = 0; in ref_tracker_get_stats() 98 ++stats->stacks[i].count; in ref_tracker_get_stats() 164 stack = stats->stacks[i].stack_handle; in __ref_tracker_dir_pr_ostream() 168 dir->class, dir, stats->stacks[i].count, in __ref_tracker_dir_pr_ostream() 170 skipped -= stats->stacks[i].count; in __ref_tracker_dir_pr_ostream()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | irq.h | 38 * Per-cpu stacks for handling critical, debug and machine check 47 * Per-cpu stacks for handling hard and soft interrupts.
|
| /linux/samples/fprobe/ |
| H A D | fprobe_example.c | 44 unsigned long stacks[BACKTRACE_DEPTH]; in show_backtrace() local 47 len = stack_trace_save(stacks, BACKTRACE_DEPTH, 2); in show_backtrace() 48 stack_trace_print(stacks, len, 24); in show_backtrace()
|
| /linux/arch/arm64/kernel/ |
| H A D | stacktrace.c | 304 * Per-cpu stacks are only accessible when unwinding the current task in a 315 * SDEI stacks are only accessible when unwinding the current task in an NMI 337 struct stack_info stacks[] = { in kunwind_stack_walk() local 351 .stacks = stacks, in kunwind_stack_walk() 352 .nr_stacks = ARRAY_SIZE(stacks), in kunwind_stack_walk()
|
| /linux/arch/x86/include/asm/ |
| H A D | cpu_entry_area.h | 35 /* The exception stacks' physical storage. No guard pages required */ 116 * Exception stacks used for IST entries with guard pages.
|
| H A D | irq_stack.h | 136 * switch stacks. If from user mode the task stack is empty. \ 145 * switching stacks. Interrupts are disabled in both \
|
| /linux/arch/powerpc/kernel/ |
| H A D | setup_64.c | 688 * This is used to allocate PACAs and various interrupt stacks that 719 panic("cannot allocate stacks"); in alloc_stack() 730 * Interrupt stacks must be in the first segment since we in irqstack_early_init() 777 * Emergency stacks must be under 256MB, we cannot afford to take in emergency_stack_init() 781 * Since we use these as temporary stacks during secondary CPU in emergency_stack_init() 786 * The IRQ stacks allocated elsewhere in this file are zeroed and in emergency_stack_init() 788 * to have emergency stacks available as early as possible. in emergency_stack_init()
|
| /linux/tools/perf/scripts/python/ |
| H A D | stackcollapse.py | 15 # perf script report stackcollapse > out.stacks-folded 48 help="do not separate stacks according to comm"),
|
| H A D | flamegraph.py | 39 const stacks = [/** @flamegraph_json **/]; 45 .datum(stacks[0]) 222 output_fn = self.args.output or "stacks.json"
|
| /linux/tools/perf/util/bpf_skel/ |
| H A D | off_cpu.bpf.c | 49 } stacks SEC(".maps"); 268 stack_id = bpf_get_stackid(ctx, &stacks, in off_cpu_stat() 281 * If stacks are successfully collected by bpf_get_stackid(), collect them once more in off_cpu_stat()
|
| /linux/arch/arc/ |
| H A D | Kconfig.debug | 4 bool "Use 16Kb for kernel stacks instead of 8Kb"
|
| /linux/security/yama/ |
| H A D | Kconfig | 10 Like capabilities, this security module stacks with other LSMs.
|
| /linux/arch/openrisc/include/asm/ |
| H A D | unwinder.h | 4 * Architecture API for unwinding stacks.
|
| /linux/arch/x86/mm/ |
| H A D | kmsan_shadow.c | 13 * Addresses within the CPU entry area (including e.g. exception stacks) do not
|
| /linux/arch/parisc/include/asm/ |
| H A D | mman.h | 8 /* PARISC cannot allow mdwe as it needs writable stacks */
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | uretprobe_stack.c | 144 * Stacks captured on ENTRY uprobes in test_uretprobe_stack() 170 * Now stacks captured on the way out in EXIT uprobes in test_uretprobe_stack()
|
| /linux/arch/riscv/include/asm/ |
| H A D | irq_stack.h | 21 * stacks need to have the same alignment.
|
| /linux/tools/testing/selftests/x86/ |
| H A D | thunks.S | 27 // Switch stacks
|