/linux/arch/alpha/mm/ |
H A D | init.c | 117 callback_init(void * kernel_end) in callback_init() argument 156 (((unsigned long)kernel_end + ~PAGE_MASK) & PAGE_MASK); in callback_init() 157 kernel_end = two_pages + 2*PAGE_SIZE; in callback_init() 194 memset(kernel_end, 0, PAGE_SIZE); in callback_init() 196 pmd_set(pmd, (pte_t *)kernel_end); in callback_init() 197 kernel_end += PAGE_SIZE; in callback_init() 208 return kernel_end; in callback_init()
|
/linux/arch/arm/boot/bootp/ |
H A D | kernel.S | 5 .globl kernel_end 6 kernel_end: label
|
/linux/arch/s390/boot/ |
H A D | vmem.c | 97 static void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 127 kasan_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KASAN_MAP_SHADOW); in kasan_populate_shadow() 135 kasan_populate(kernel_end, _REGION1_SIZE, POPULATE_KASAN_ZERO_SHADOW); in kasan_populate_shadow() 192 static inline void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 481 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit) in setup_vmem() argument 550 pgtable_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KERNEL); in setup_vmem() 558 kasan_populate_shadow(kernel_start, kernel_end); in setup_vmem()
|
H A D | startup.c | 370 unsigned long kernel_end, kaslr_len, slots, pos; in setup_kernel_memory_layout() local 376 kernel_end = vmax - pos * THREAD_SIZE; in setup_kernel_memory_layout() 377 kernel_start = round_down(kernel_end - kernel_size, THREAD_SIZE); in setup_kernel_memory_layout()
|
H A D | boot.h | 72 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit);
|
/linux/arch/riscv/mm/ |
H A D | physaddr.c | 25 unsigned long kernel_end = kernel_start + kernel_map.size; in __phys_addr_symbol() local 31 VIRTUAL_BUG_ON(x < kernel_start || x > kernel_end); in __phys_addr_symbol()
|
/linux/arch/arm/kernel/ |
H A D | head-nommu.S | 357 ldr r6, =KERNEL_END 394 /* Setup Background: max(KERNEL_END, _exiprom) - 0xffffffff */ 396 ldr r5, =KERNEL_END 401 ldr r5, =KERNEL_END 415 /* Setup Background: min(_exiprom, KERNEL_END) - max(KERNEL_START, XIP_PHYS_ADDR) */ 417 ldr r6, =KERNEL_END
|
/linux/arch/powerpc/kexec/ |
H A D | core.c | 139 static phys_addr_t kernel_end; variable 147 .value = &kernel_end, 206 kernel_end = cpu_to_be_ulong(__pa(_end)); in kexec_setup()
|
/linux/arch/x86/boot/startup/ |
H A D | sme.c | 288 unsigned long kernel_start, kernel_end, kernel_len; in sme_encrypt_kernel() local 319 kernel_end = ALIGN((unsigned long)rip_rel_ptr(_end), PMD_SIZE); in sme_encrypt_kernel() 320 kernel_len = kernel_end - kernel_start; in sme_encrypt_kernel() 421 ppd.vaddr_end = kernel_end; in sme_encrypt_kernel() 427 ppd.vaddr_end = kernel_end + decrypted_base; in sme_encrypt_kernel() 471 ppd.vaddr_end = kernel_end + decrypted_base; in sme_encrypt_kernel()
|
/linux/arch/arm/mm/ |
H A D | pmsa-v8.c | 255 subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup() 256 subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup() 280 err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
|
H A D | physaddr.c | 54 x > (unsigned long)KERNEL_END); in __phys_addr_symbol()
|
/linux/arch/arm64/mm/ |
H A D | init.c | 460 u64 kernel_end = (u64)_end; in module_init_limits() local 462 u64 kernel_size = kernel_end - kernel_start; in module_init_limits() 473 module_direct_base = kernel_end - SZ_128M; in module_init_limits() 475 module_plt_base = kernel_end - SZ_2G; in module_init_limits() 478 u64 max = kernel_end; in module_init_limits()
|
H A D | mmu.c | 1026 phys_addr_t kernel_end = __pa_symbol(__init_begin); in map_mem() local 1058 memblock_mark_nomap(kernel_start, kernel_end - kernel_start); in map_mem() 1083 __map_memblock(pgdp, kernel_start, kernel_end, in map_mem() 1085 memblock_clear_nomap(kernel_start, kernel_end - kernel_start); in map_mem()
|
H A D | kasan_init.c | 310 kimg_shadow_end = PAGE_ALIGN((u64)kasan_mem_to_shadow(KERNEL_END)); in kasan_init_shadow()
|
/linux/arch/alpha/kernel/ |
H A D | setup.c | 286 setup_memory(void *kernel_end) in setup_memory() argument 346 kernel_size = virt_to_phys(kernel_end) - KERNEL_START_PHYS; in setup_memory() 427 void *kernel_end = _end; /* end of kernel */ in setup_arch() local 466 kernel_end = callback_init(kernel_end); in setup_arch() 608 setup_memory(kernel_end); in setup_arch()
|
/linux/arch/x86/mm/ |
H A D | init.c | 784 unsigned long kernel_end = __pa_symbol(_end); in init_mem_mapping() local 789 * [kernel_end, end) to make memory above the kernel be mapped in init_mem_mapping() 791 * the kernel to map [ISA_END_ADDRESS, kernel_end). in init_mem_mapping() 793 memory_map_bottom_up(kernel_end, end); in init_mem_mapping() 794 memory_map_bottom_up(ISA_END_ADDRESS, kernel_end); in init_mem_mapping()
|
/linux/scripts/gdb/linux/ |
H A D | mm.py | 110 self.KERNEL_END = gdb.parse_and_eval("_end") 241 if x < self.KERNEL_START or x > self.KERNEL_END:
|
/linux/arch/arm64/include/asm/ |
H A D | memory.h | 69 #define KERNEL_END _end macro
|
/linux/arch/m68k/kernel/ |
H A D | head.S | 1277 * Allocated memory so far goes from kernel_end to memory_start that 1298 movel %pc@(L(kernel_end)),%a0 2555 lea %pc@(L(kernel_end)),%a0 3808 L(kernel_end):
|