| /linux/arch/x86/include/asm/ |
| H A D | pgtable_64_types.h | 174 # define MODULES_END _AC(0xffffffffff000000, UL) macro 176 # define MODULES_END _AC(0xfffffffffe000000, UL) macro 178 #define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
| /linux/arch/powerpc/mm/ |
| H A D | mem.c | 390 for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE) in prealloc_execmem_pgtable() 412 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) { in execmem_arch_setup() 415 fallback_end = MODULES_END; in execmem_arch_setup() 420 end = MODULES_END; in execmem_arch_setup()
|
| /linux/arch/sparc/mm/ |
| H A D | execmem.c | 13 .end = MODULES_END, in execmem_arch_setup()
|
| H A D | fault_64.c | 306 (tpc >= MODULES_VADDR && tpc < MODULES_END)) { in do_sparc64_fault()
|
| /linux/arch/riscv/kernel/ |
| H A D | vmcore_info.c | 18 vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END); in arch_crash_save_vmcoreinfo()
|
| /linux/arch/x86/kernel/ |
| H A D | head64.c | 234 MAYBE_BUILD_BUG_ON(!(((MODULES_END - 1) & PGDIR_MASK) == in x86_64_start_kernel() 236 BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END); in x86_64_start_kernel()
|
| /linux/arch/arm/mm/ |
| H A D | pageattr.c | 66 if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
|
| H A D | kasan_init.c | 285 create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END)); in kasan_init()
|
| /linux/mm/kmsan/ |
| H A D | kmsan.h | 179 return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END); in kmsan_internal_is_module_addr()
|
| /linux/arch/powerpc/mm/kasan/ |
| H A D | init_32.c | 121 k_end = (unsigned long)kasan_mem_to_shadow((void *)MODULES_END); in kasan_unmap_early_shadow_vmalloc()
|
| /linux/Documentation/admin-guide/kdump/ |
| H A D | vmcoreinfo.rst | 461 MODULES_VADDR|MODULES_END|VMALLOC_START|VMALLOC_END|VMEMMAP_START|VMEMMAP_END 465 MODULES_VADDR ~ MODULES_END-1 : Kernel module space. 582 MODULES_VADDR|MODULES_END|VMALLOC_START|VMALLOC_END|VMEMMAP_START|VMEMMAP_END|KERNEL_LINK_ADDR 587 * MODULES_VADDR ~ MODULES_END : Kernel module space.
|
| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 504 (void *)kasan_mem_to_shadow((void *)MODULES_END)); in kasan_init() 522 kasan_populate(kasan_mem_to_shadow((const void *)MODULES_END), in kasan_init()
|
| /linux/fs/proc/ |
| H A D | kcore.c | 712 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { in add_modules_range() 714 MODULES_END - MODULES_VADDR, KCORE_VMALLOC); in add_modules_range()
|
| /linux/arch/mips/mm/ |
| H A D | fault.c | 87 if (unlikely(address >= MODULES_VADDR && address < MODULES_END)) in __do_page_fault()
|
| /linux/arch/nios2/include/asm/ |
| H A D | pgtable.h | 31 #define MODULES_END (CONFIG_NIOS2_KERNEL_REGION_BASE - 1) macro
|
| /linux/arch/riscv/include/asm/ |
| H A D | pgtable.h | 50 #define BPF_JIT_REGION_END (MODULES_END) 61 #define MODULES_END (PFN_ALIGN((unsigned long)&_start)) macro 64 #define MODULES_END VMALLOC_END macro
|
| /linux/arch/um/include/asm/ |
| H A D | pgtable.h | 56 #define MODULES_END VMALLOC_END macro
|
| /linux/Documentation/arch/arm/ |
| H A D | memory.rst | 76 MODULES_VADDR MODULES_END-1 Kernel module space
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 266 self.MODULES_END = self.MODULES_VADDR + self.MODULES_VSIZE 273 self.VMALLOC_START = self.MODULES_END
|
| /linux/arch/s390/mm/ |
| H A D | dump_pagetables.c | 336 rc |= add_marker(MODULES_VADDR, MODULES_END, "Modules Area"); in pt_dump_init()
|
| /linux/arch/powerpc/mm/ptdump/ |
| H A D | ptdump.c | 269 address_markers[i++].start_address = MODULES_END; in populate_markers()
|
| /linux/arch/sparc/include/asm/ |
| H A D | pgtable_32.h | 427 #define MODULES_END VMALLOC_END macro
|
| /linux/arch/arm64/mm/ |
| H A D | ptdump.c | 389 { MODULES_END, "Modules end" }, in ptdump_init()
|
| H A D | kasan_init.c | 348 BUILD_BUG_ON(VMALLOC_START != MODULES_END); in kasan_init_shadow()
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_init.c | 526 MODULES_END, GFP_KERNEL, PAGE_KERNEL_ROX, in hyperv_init()
|