| /linux/arch/arm/mm/ |
| H A D | pgd.c | 55 new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR), in pgd_alloc() 56 MODULES_VADDR); in pgd_alloc() 60 new_pud = pud_alloc(mm, new_p4d, MODULES_VADDR); in pgd_alloc() 77 (pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE)) in pgd_alloc()
|
| H A D | pageattr.c | 66 if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
|
| H A D | kasan_init.c | 285 create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END)); in kasan_init()
|
| /linux/arch/powerpc/mm/ |
| H A D | mem.c | 390 for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE) in prealloc_execmem_pgtable() 408 #ifdef MODULES_VADDR in execmem_arch_setup() 412 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) { in execmem_arch_setup() 414 fallback_start = MODULES_VADDR; in execmem_arch_setup() 417 start = MODULES_VADDR; in execmem_arch_setup()
|
| /linux/arch/x86/kernel/ |
| H A D | head64.c | 228 BUILD_BUG_ON(MODULES_VADDR < __START_KERNEL_map); in x86_64_start_kernel() 229 BUILD_BUG_ON(MODULES_VADDR - __START_KERNEL_map < KERNEL_IMAGE_SIZE); in x86_64_start_kernel() 232 BUILD_BUG_ON((MODULES_VADDR & ~PMD_MASK) != 0); in x86_64_start_kernel() 233 BUILD_BUG_ON(!(MODULES_VADDR > __START_KERNEL)); in x86_64_start_kernel()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | kasan.h | 53 #define XKVRANGE_VC_START MODULES_VADDR 54 #define XKVRANGE_VC_SIZE round_up(KFENCE_AREA_END - MODULES_VADDR + 1, PGDIR_SIZE)
|
| /linux/arch/x86/include/asm/ |
| H A D | pgtable_64_types.h | 171 #define MODULES_VADDR (__START_KERNEL_map + KERNEL_IMAGE_SIZE) macro 178 #define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
| /linux/Documentation/arch/arm/ |
| H A D | memory.rst | 76 MODULES_VADDR MODULES_END-1 Kernel module space 80 TASK_SIZE MODULES_VADDR-1 KASAn shadow memory when KASan is in use. 81 The range from MODULES_VADDR to the top
|
| /linux/arch/powerpc/mm/ptdump/ |
| H A D | ptdump.c | 77 #ifdef MODULES_VADDR 267 #ifdef MODULES_VADDR in populate_markers() 268 address_markers[i++].start_address = MODULES_VADDR; in populate_markers()
|
| /linux/arch/powerpc/mm/kasan/ |
| H A D | init_32.c | 119 #ifdef MODULES_VADDR in kasan_unmap_early_shadow_vmalloc() 120 k_start = (unsigned long)kasan_mem_to_shadow((void *)MODULES_VADDR); in kasan_unmap_early_shadow_vmalloc()
|
| /linux/arch/sparc/mm/ |
| H A D | execmem.c | 12 .start = MODULES_VADDR, in execmem_arch_setup()
|
| /linux/arch/riscv/kernel/ |
| H A D | vmcore_info.c | 17 vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR); in arch_crash_save_vmcoreinfo()
|
| /linux/fs/proc/ |
| H A D | kcore.c | 705 #if defined(CONFIG_MODULES) && defined(MODULES_VADDR) 712 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { in add_modules_range() 713 kclist_add(&kcore_modules, (void *)MODULES_VADDR, in add_modules_range() 714 MODULES_END - MODULES_VADDR, KCORE_VMALLOC); in add_modules_range()
|
| /linux/arch/mips/mm/ |
| H A D | fault.c | 86 #ifdef MODULES_VADDR in __do_page_fault() 87 if (unlikely(address >= MODULES_VADDR && address < MODULES_END)) in __do_page_fault()
|
| H A D | init.c | 566 #ifdef MODULES_VADDR 574 .start = MODULES_VADDR, in execmem_arch_setup()
|
| /linux/mm/kmsan/ |
| H A D | kmsan.h | 179 return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END); in kmsan_internal_is_module_addr()
|
| H A D | shadow.c | 66 off = addr64 - MODULES_VADDR; in vmalloc_meta()
|
| /linux/Documentation/admin-guide/kdump/ |
| H A D | vmcoreinfo.rst | 461 MODULES_VADDR|MODULES_END|VMALLOC_START|VMALLOC_END|VMEMMAP_START|VMEMMAP_END 465 MODULES_VADDR ~ MODULES_END-1 : Kernel module space. 582 MODULES_VADDR|MODULES_END|VMALLOC_START|VMALLOC_END|VMEMMAP_START|VMEMMAP_END|KERNEL_LINK_ADDR 587 * MODULES_VADDR ~ MODULES_END : Kernel module space.
|
| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 503 (void *)kasan_mem_to_shadow((void *)MODULES_VADDR), in kasan_init() 523 kasan_mem_to_shadow((const void *)MODULES_VADDR + SZ_2G)); in kasan_init()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | uaccess.h | 507 #ifdef MODULES_VADDR in mask_user_address() 508 const unsigned long border = MODULES_VADDR; in mask_user_address()
|
| /linux/arch/sparc/kernel/ |
| H A D | ktlb.S | 221 sethi %hi(MODULES_VADDR), %g5
|
| /linux/arch/nios2/include/asm/ |
| H A D | pgtable.h | 30 #define MODULES_VADDR (CONFIG_NIOS2_KERNEL_REGION_BASE - SZ_32M) macro
|
| /linux/arch/um/include/asm/ |
| H A D | pgtable.h | 55 #define MODULES_VADDR VMALLOC_START macro
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 265 self.MODULES_VADDR = self._PAGE_END(self.VA_BITS_MIN) 266 self.MODULES_END = self.MODULES_VADDR + self.MODULES_VSIZE
|
| /linux/arch/s390/mm/ |
| H A D | dump_pagetables.c | 336 rc |= add_marker(MODULES_VADDR, MODULES_END, "Modules Area"); in pt_dump_init()
|