| /linux/arch/x86/include/asm/ |
| H A D | pgtable_64_types.h | 133 #define VMALLOC_END VMEMORY_END macro 151 #define VMALLOC_END (VMALLOC_START + VMALLOC_QUARTER_SIZE - 1) macro 167 #define KMSAN_MODULES_SHADOW_START (VMALLOC_END + KMSAN_VMALLOC_ORIGIN_OFFSET + 1)
|
| /linux/arch/powerpc/include/asm/nohash/32/ |
| H A D | pgtable.h | 84 #define IOREMAP_END VMALLOC_END 111 #define VMALLOC_END ALIGN_DOWN(ioremap_bot, PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT) macro 113 #define VMALLOC_END ioremap_bot macro
|
| /linux/arch/s390/boot/ |
| H A D | startup.c | 35 unsigned long __bootdata_preserved(VMALLOC_END); 399 VMALLOC_END = MODULES_VADDR; in setup_kernel_memory_layout() 401 VMALLOC_END -= MODULES_LEN * 2; in setup_kernel_memory_layout() 405 vsize = (VMALLOC_END - FIXMAP_SIZE) / 2; in setup_kernel_memory_layout() 411 VMALLOC_END -= vmalloc_size * 2; in setup_kernel_memory_layout() 413 VMALLOC_START = VMALLOC_END - vmalloc_size; in setup_kernel_memory_layout() 414 boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END); in setup_kernel_memory_layout()
|
| /linux/arch/powerpc/mm/ |
| H A D | mem.c | 420 end = VMALLOC_END; in execmem_arch_setup() 437 .end = VMALLOC_END, in execmem_arch_setup() 443 .end = VMALLOC_END, in execmem_arch_setup()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | pgtable.h | 86 #define VMALLOC_END (PKMAP_BASE - (2 * PAGE_SIZE)) macro 88 #define VMALLOC_END (FIXADDR_START - (2 * PAGE_SIZE)) macro 110 #define VMALLOC_END \ macro 114 #define VMALLOC_END \ macro 120 #define vmemmap ((struct page *)(ALIGN(VMALLOC_END, VMEMMAP_ALIGN)))
|
| /linux/Documentation/arch/xtensa/ |
| H A D | mmu.rst | 84 +------------------+ VMALLOC_END 127 +------------------+ VMALLOC_END 171 +------------------+ VMALLOC_END
|
| /linux/include/linux/ |
| H A D | ioremap.h | 16 #define IOREMAP_END VMALLOC_END
|
| /linux/arch/csky/include/asm/ |
| H A D | memory.h | 14 #define VMALLOC_END (PKMAP_BASE - (PAGE_SIZE * 2)) macro
|
| /linux/arch/riscv/kernel/ |
| H A D | vmcore_info.c | 16 vmcoreinfo_append_str("NUMBER(VMALLOC_END)=0x%lx\n", VMALLOC_END); in arch_crash_save_vmcoreinfo()
|
| /linux/arch/arm/mm/ |
| H A D | iomap.c | 41 (unsigned long)addr < VMALLOC_END) in pci_iounmap()
|
| H A D | kasan_init.c | 246 kasan_mem_to_shadow((void *)VMALLOC_END)); in kasan_init() 248 kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_END), in kasan_init()
|
| H A D | pageattr.c | 67 !range_in_range(start, size, VMALLOC_START, VMALLOC_END)) in change_memory_common()
|
| /linux/arch/parisc/include/asm/ |
| H A D | fixmap.h | 55 #define VMALLOC_END (KERNEL_MAP_END) macro
|
| /linux/arch/arc/mm/ |
| H A D | highmem.c | 67 BUILD_BUG_ON(PAGE_OFFSET < (VMALLOC_END + FIXMAP_SIZE + PKMAP_SIZE)); in kmap_init()
|
| /linux/arch/powerpc/include/asm/book3s/32/ |
| H A D | pgtable.h | 169 #define IOREMAP_END VMALLOC_END 193 #define VMALLOC_END ALIGN_DOWN(ioremap_bot, PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT) macro 195 #define VMALLOC_END ioremap_bot macro
|
| /linux/arch/arm/include/asm/ |
| H A D | pgtable-nommu.h | 63 #define VMALLOC_END 0xffffffffUL macro
|
| /linux/arch/hexagon/include/asm/ |
| H A D | mem-layout.h | 103 #define VMALLOC_END (PKMAP_BASE-PAGE_SIZE*2) macro
|
| /linux/arch/x86/mm/ |
| H A D | kasan_init_64.c | 402 kasan_mem_to_shadow((void *)VMALLOC_END)); in kasan_init() 406 kasan_mem_to_shadow((void *)VMALLOC_END)); in kasan_init() 409 kasan_mem_to_shadow((void *)VMALLOC_END + 1), in kasan_init()
|
| /linux/arch/sparc/kernel/ |
| H A D | ktlb.S | 224 sethi %hi(VMALLOC_END), %g5 225 ldx [%g5 + %lo(VMALLOC_END)], %g5
|
| /linux/arch/arc/include/asm/ |
| H A D | processor.h | 92 #define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE) macro
|
| /linux/Documentation/arch/arm/ |
| H A D | memory.rst | 59 VMALLOC_START VMALLOC_END-1 vmalloc() / ioremap() space. 65 of the high_memory variable, and VMALLOC_END
|
| /linux/scripts/gdb/linux/ |
| H A D | vmalloc.py | 20 return addr >= pg_ops.VMALLOC_START and addr < pg_ops.VMALLOC_END
|
| H A D | mm.py | 51 self.VMALLOC_END = self.VMALLOC_START + (32 * 1024 * 1024 * 1024 * 1024) - 1 53 self.VMALLOC_END = self.VMALLOC_START + (12800 * 1024 * 1024 * 1024 * 1024) - 1 55 self.VMALLOC_END = self.VMALLOC_START + (12800 * 1024 * 1024 * 1024 * 1024) - 1 274 self.VMALLOC_END = self.VMEMMAP_START - 256 * 1024 * 1024
|
| /linux/mm/kmsan/ |
| H A D | kmsan.h | 184 return ((u64)addr >= VMALLOC_START) && ((u64)addr < VMALLOC_END); in kmsan_internal_is_vmalloc_addr()
|
| /linux/kernel/ |
| H A D | scs.c | 46 s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END, in __scs_alloc()
|