/linux/arch/alpha/mm/ |
H A D | init.c | 144 callback_init(void * kernel_end) in callback_init() argument 183 (((unsigned long)kernel_end + ~PAGE_MASK) & PAGE_MASK); in callback_init() 184 kernel_end = two_pages + 2*PAGE_SIZE; in callback_init() 221 memset(kernel_end, 0, PAGE_SIZE); in callback_init() 223 pmd_set(pmd, (pte_t *)kernel_end); in callback_init() 224 kernel_end += PAGE_SIZE; in callback_init() 235 return kernel_end; in callback_init()
|
/linux/arch/arm/boot/bootp/ |
H A D | kernel.S | 5 .globl kernel_end 6 kernel_end: label
|
/linux/arch/arm64/mm/ |
H A D | init.c | 492 u64 kernel_end = (u64)_end; in module_init_limits() local 494 u64 kernel_size = kernel_end - kernel_start; in module_init_limits() 505 module_direct_base = kernel_end - SZ_128M; in module_init_limits() 507 module_plt_base = kernel_end - SZ_2G; in module_init_limits() 510 u64 max = kernel_end; in module_init_limits()
|
/linux/arch/riscv/mm/ |
H A D | physaddr.c | 25 unsigned long kernel_end = kernel_start + kernel_map.size; in __phys_addr_symbol() local 31 VIRTUAL_BUG_ON(x < kernel_start || x > kernel_end); in __phys_addr_symbol()
|
/linux/arch/s390/boot/ |
H A D | vmem.c | 95 static void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 125 kasan_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KASAN_MAP_SHADOW); in kasan_populate_shadow() 133 kasan_populate(kernel_end, _REGION1_SIZE, POPULATE_KASAN_ZERO_SHADOW); in kasan_populate_shadow() 190 static inline void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 479 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit) in setup_vmem() argument 545 pgtable_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KERNEL); in setup_vmem() 553 kasan_populate_shadow(kernel_start, kernel_end); in setup_vmem()
|
H A D | startup.c | 309 unsigned long kernel_end, kaslr_len, slots, pos; in setup_kernel_memory_layout() local 315 kernel_end = vmax - pos * THREAD_SIZE; in setup_kernel_memory_layout() 316 kernel_start = round_down(kernel_end - kernel_size, THREAD_SIZE); in setup_kernel_memory_layout()
|
H A D | boot.h | 75 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit);
|
/linux/arch/parisc/mm/ |
H A D | init.c | 358 unsigned long kernel_start, kernel_end; in map_pages() local 363 kernel_end = __pa((unsigned long)&_end); in map_pages() 401 } else if (address < kernel_start || address >= kernel_end) { in map_pages() 455 unsigned long kernel_end = (unsigned long)&_end; in free_initmem() 458 map_pages(init_end, __pa(init_end), kernel_end - init_end, in free_initmem() 475 __flush_tlb_range(0, init_begin, kernel_end); in free_initmem() 459 unsigned long kernel_end = (unsigned long)&_end; free_initmem() local
|
/linux/arch/alpha/kernel/ |
H A D | setup.c | 286 setup_memory(void *kernel_end) in setup_memory() argument 346 kernel_size = virt_to_phys(kernel_end) - KERNEL_START_PHYS; in setup_memory() 427 void *kernel_end = _end; /* end of kernel */ in setup_arch() local 466 kernel_end = callback_init(kernel_end); in setup_arch() 608 setup_memory(kernel_end); in setup_arch()
|
/linux/arch/x86/mm/ |
H A D | init.c | 781 unsigned long kernel_end = __pa_symbol(_end); in init_mem_mapping() local 790 memory_map_bottom_up(kernel_end, end); in init_mem_mapping() 791 memory_map_bottom_up(ISA_END_ADDRESS, kernel_end); in init_mem_mapping()
|
/linux/arch/m68k/kernel/ |
H A D | head.S | 1298 movel %pc@(L(kernel_end)),%a0 2555 lea %pc@(L(kernel_end)),%a0 3807 L(kernel_end):
|
/linux/arch/x86/xen/ |
H A D | mmu_pv.c | 1066 unsigned long kernel_end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in xen_cleanhighmap() local 1075 if (vaddr < (unsigned long) _text || vaddr > kernel_end) in xen_cleanhighmap()
|