/linux/arch/x86/include/asm/ |
H A D | pgtable_32_areas.h | 48 #define MODULES_END VMALLOC_END macro 49 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
|
/linux/arch/powerpc/include/asm/nohash/32/ |
H A D | mmu-8xx.h | 173 #define MODULES_END PAGE_OFFSET macro 175 #define MODULES_VADDR (MODULES_END - MODULES_SIZE)
|
/linux/arch/s390/boot/ |
H A D | startup.c | 32 unsigned long __bootdata_preserved(MODULES_END); 332 MODULES_END = round_down(kernel_start, _SEGMENT_SIZE); in setup_kernel_memory_layout() 333 MODULES_VADDR = MODULES_END - MODULES_LEN; in setup_kernel_memory_layout() 337 boot_debug("modules area: 0x%016lx-0x%016lx\n", MODULES_VADDR, MODULES_END); in setup_kernel_memory_layout()
|
/linux/arch/sparc/mm/ |
H A D | execmem.c | 13 .end = MODULES_END, in execmem_arch_setup()
|
/linux/arch/riscv/kernel/ |
H A D | vmcore_info.c | 18 vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END); in arch_crash_save_vmcoreinfo()
|
/linux/arch/arm64/kernel/ |
H A D | vmcore_info.c | 25 vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END); in arch_crash_save_vmcoreinfo()
|
/linux/arch/arm/mm/ |
H A D | pageattr.c | 66 if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
|
H A D | kasan_init.c | 285 create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END)); in kasan_init()
|
/linux/mm/kmsan/ |
H A D | kmsan.h | 180 return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END); in kmsan_internal_is_module_addr()
|
/linux/arch/nios2/mm/ |
H A D | init.c | 161 .end = MODULES_END, in execmem_arch_setup()
|
/linux/arch/powerpc/include/asm/book3s/32/ |
H A D | pgtable.h | 198 #define MODULES_END ALIGN_DOWN(PAGE_OFFSET, SZ_256M) macro 200 #define MODULES_VADDR (MODULES_END - MODULES_SIZE)
|
/linux/arch/x86/mm/ |
H A D | init.c | 1097 .end = MODULES_END, in execmem_arch_setup() 1104 .end = MODULES_END, in execmem_arch_setup() 1111 .end = MODULES_END, in execmem_arch_setup()
|
/linux/arch/riscv/include/asm/ |
H A D | pgtable.h | 46 #define BPF_JIT_REGION_END (MODULES_END) 57 #define MODULES_END (PFN_ALIGN((unsigned long)&_start)) macro 60 #define MODULES_END VMALLOC_END macro
|
/linux/Documentation/admin-guide/kdump/ |
H A D | vmcoreinfo.rst | 461 MODULES_VADDR|MODULES_END|VMALLOC_START|VMALLOC_END|VMEMMAP_START|VMEMMAP_END 465 MODULES_VADDR ~ MODULES_END-1 : Kernel module space. 582 MODULES_VADDR|MODULES_END|VMALLOC_START|VMALLOC_END|VMEMMAP_START|VMEMMAP_END|KERNEL_LINK_ADDR 587 * MODULES_VADDR ~ MODULES_END : Kernel module space.
|
/linux/arch/riscv/mm/ |
H A D | kasan_init.c | 504 (void *)kasan_mem_to_shadow((void *)MODULES_END)); in kasan_init() 522 kasan_populate(kasan_mem_to_shadow((const void *)MODULES_END), in kasan_init()
|
/linux/arch/loongarch/mm/ |
H A D | init.c | 257 .end = MODULES_END,
|
/linux/scripts/gdb/linux/ |
H A D | mm.py | 95 self.MODULES_END = self.MODULES_VADDR + self.MODULES_VSIZE 102 self.VMALLOC_START = self.MODULES_END
|
/linux/fs/proc/ |
H A D | kcore.c | 712 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { in add_modules_range() 714 MODULES_END - MODULES_VADDR, KCORE_VMALLOC); in add_modules_range()
|
/linux/arch/mips/mm/ |
H A D | fault.c | 87 if (unlikely(address >= MODULES_VADDR && address < MODULES_END)) in __do_page_fault()
|
/linux/arch/nios2/include/asm/ |
H A D | pgtable.h | 31 #define MODULES_END (CONFIG_NIOS2_KERNEL_REGION_BASE - 1) macro
|
/linux/arch/powerpc/mm/ptdump/ |
H A D | ptdump.c | 268 address_markers[i++].start_address = MODULES_END; in populate_markers()
|
/linux/Documentation/arch/arm/ |
H A D | memory.rst | 76 MODULES_VADDR MODULES_END-1 Kernel module space
|
/linux/arch/s390/mm/ |
H A D | dump_pagetables.c | 299 rc |= add_marker(MODULES_VADDR, MODULES_END, "Modules Area"); in pt_dump_init()
|
/linux/arch/arm64/mm/ |
H A D | ptdump.c | 340 { MODULES_END, "Modules end" }, in ptdump_init()
|
/linux/arch/s390/include/asm/ |
H A D | pgtable.h | 94 extern unsigned long MODULES_END; 96 #define MODULES_END MODULES_END macro 104 if (addr > (void *)MODULES_END) in is_module_addr()
|