| /linux/tools/perf/util/ |
| H A D | thread-stack.c | 96 u64 kernel_start; member 162 ts->kernel_start = machine__kernel_start(machine); in thread_stack__init() 166 ts->kernel_start = 1ULL << 63; in thread_stack__init() 516 static inline u64 callchain_context(u64 ip, u64 kernel_start) in callchain_context() argument 518 return ip < kernel_start ? PERF_CONTEXT_USER : PERF_CONTEXT_KERNEL; in callchain_context() 523 size_t sz, u64 ip, u64 kernel_start) in thread_stack__sample() argument 526 u64 context = callchain_context(ip, kernel_start); in thread_stack__sample() 547 context = callchain_context(ip, kernel_start); in thread_stack__sample() 566 u64 sample_ip, u64 kernel_start) in thread_stack__sample_late() argument 569 u64 sample_context = callchain_context(sample_ip, kernel_start); in thread_stack__sample_late() [all …]
|
| H A D | thread-stack.h | 87 size_t sz, u64 ip, u64 kernel_start); 90 u64 kernel_start); 95 u64 sample_ip, u64 kernel_start);
|
| H A D | machine.h | 48 u64 kernel_start; member 91 if (!machine->kernel_start) in machine__kernel_start() 93 return machine->kernel_start; in machine__kernel_start() 98 u64 kernel_start = machine__kernel_start(machine); in machine__kernel_ip() local 100 return ip >= kernel_start; in machine__kernel_ip()
|
| H A D | db-export.c | 215 u64 kernel_start = machine__kernel_start(machine); in call_path_from_sample() local 269 kernel_start); in call_path_from_sample()
|
| H A D | intel-pt.c | 88 u64 kernel_start; member 683 return ip >= ptq->pt->kernel_start ? in intel_pt_nr_cpumode() 999 } else if (ip >= ptq->pt->kernel_start) { in __intel_pt_pgd_ip() 1246 pt->kernel_start); in intel_pt_add_callchain() 1284 pt->kernel_start); in intel_pt_add_br_stack() 1829 sample->ip, pt->kernel_start); in intel_pt_prep_sample() 2468 pt->kernel_start); in intel_pt_do_synth_pebs_sample() 3127 if (!pt->kernel_start) { in intel_pt_run_decoder() 3128 pt->kernel_start = machine__kernel_start(pt->machine); in intel_pt_run_decoder() 3154 state->from_ip >= pt->kernel_start) { in intel_pt_run_decoder() [all …]
|
| H A D | arm-spe.c | 94 u64 kernel_start; member 280 return ip >= spe->kernel_start ? in arm_spe_cpumode() 1130 if (!spe->kernel_start) in arm_spe_run_decoder() 1131 spe->kernel_start = machine__kernel_start(spe->machine); in arm_spe_run_decoder()
|
| H A D | machine.c | 97 machine->kernel_start = 0; in machine__init() 3147 machine->kernel_start = 1ULL << 63; in machine__get_kernel_start() 3156 machine->kernel_start = map__start(map); in machine__get_kernel_start()
|
| /linux/arch/x86/boot/startup/ |
| H A D | sme.c | 288 unsigned long kernel_start, kernel_end, kernel_len; in sme_encrypt_kernel() local 318 kernel_start = (unsigned long)rip_rel_ptr(_text); in sme_encrypt_kernel() 320 kernel_len = kernel_end - kernel_start; in sme_encrypt_kernel() 354 pgtable_area_len += sme_pgtable_calc(execute_end - kernel_start) * 2; in sme_encrypt_kernel() 419 ppd.paddr = kernel_start; in sme_encrypt_kernel() 420 ppd.vaddr = kernel_start; in sme_encrypt_kernel() 425 ppd.paddr = kernel_start; in sme_encrypt_kernel() 426 ppd.vaddr = kernel_start + decrypted_base; in sme_encrypt_kernel() 457 sme_encrypt_execute(kernel_start, kernel_start + decrypted_base, in sme_encrypt_kernel() 470 ppd.vaddr = kernel_start + decrypted_base; in sme_encrypt_kernel()
|
| /linux/arch/riscv/mm/ |
| H A D | physaddr.c | 24 unsigned long kernel_start = kernel_map.virt_addr; in __phys_addr_symbol() local 25 unsigned long kernel_end = kernel_start + kernel_map.size; in __phys_addr_symbol() 31 VIRTUAL_BUG_ON(x < kernel_start || x > kernel_end); in __phys_addr_symbol()
|
| /linux/arch/powerpc/kexec/ |
| H A D | core.c | 109 unsigned long long kernel_start, kernel_size; in arch_reserve_crashkernel() local 125 kernel_start = __pa(_stext); in arch_reserve_crashkernel() 129 if ((kernel_start + kernel_size > crash_base) && (kernel_start <= crash_end)) { in arch_reserve_crashkernel()
|
| /linux/arch/arm/boot/bootp/ |
| H A D | kernel.S | 2 .globl kernel_start 3 kernel_start: label
|
| H A D | init.S | 58 b kernel_start @ call kernel
|
| /linux/arch/s390/boot/ |
| H A D | vmem.c | 96 static void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 126 kasan_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KASAN_MAP_SHADOW); in kasan_populate_shadow() 191 static inline void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 468 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit) in setup_vmem() argument 537 pgtable_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KERNEL); in setup_vmem() 545 kasan_populate_shadow(kernel_start, kernel_end); in setup_vmem()
|
| /linux/arch/riscv/kernel/ |
| H A D | machine_kexec_file.c | 256 int load_extra_segments(struct kimage *image, unsigned long kernel_start, in load_extra_segments() argument 268 kbuf.buf_min = kernel_start + kernel_len; in load_extra_segments() 325 &kernel_start, in load_extra_segments() 326 sizeof(kernel_start), 0); in load_extra_segments()
|
| /linux/arch/arm64/mm/ |
| H A D | init.c | 461 u64 kernel_start = (u64)_text; in module_init_limits() local 462 u64 kernel_size = kernel_end - kernel_start; in module_init_limits() 477 u64 min = kernel_start; in module_init_limits()
|
| H A D | mmu.c | 1120 phys_addr_t kernel_start = __pa_symbol(_text); in map_mem() local 1153 memblock_mark_nomap(kernel_start, kernel_end - kernel_start); in map_mem() 1178 __map_memblock(pgdp, kernel_start, kernel_end, in map_mem() 1180 memblock_clear_nomap(kernel_start, kernel_end - kernel_start); in map_mem()
|
| /linux/arch/riscv/include/asm/ |
| H A D | kexec.h | 72 int load_extra_segments(struct kimage *image, unsigned long kernel_start,
|