| /linux/arch/x86/include/asm/ |
| H A D | pgtable_64_types.h | 116 # define VMALLOC_START vmalloc_base macro 130 #define VMEMORY_END (VMALLOC_START + (VMALLOC_SIZE_TB << 40) - 1) 151 #define VMALLOC_END (VMALLOC_START + VMALLOC_QUARTER_SIZE - 1) 160 #define KMSAN_VMALLOC_SHADOW_START (VMALLOC_START + KMSAN_VMALLOC_SHADOW_OFFSET) 161 #define KMSAN_VMALLOC_ORIGIN_START (VMALLOC_START + KMSAN_VMALLOC_ORIGIN_OFFSET)
|
| /linux/arch/arc/include/asm/ |
| H A D | processor.h | 87 #define VMALLOC_START (PAGE_OFFSET - (CONFIG_ARC_KVADDR_SIZE << 20)) macro 92 #define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE) 94 #define USER_KERNEL_GUTTER (VMALLOC_START - TASK_SIZE)
|
| /linux/arch/powerpc/include/asm/nohash/32/ |
| H A D | pgtable.h | 83 #define IOREMAP_START VMALLOC_START 105 #define VMALLOC_START (((ALIGN((long)high_memory, PPC_PIN_SIZE) + VMALLOC_OFFSET) & ~(VMALLOC_OFFSE… macro 107 #define VMALLOC_START ((((long)high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1))) macro
|
| /linux/arch/xtensa/mm/ |
| H A D | kasan_init.c | 75 BUILD_BUG_ON(VMALLOC_START < KASAN_START_VADDR); in kasan_init() 81 populate(kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init()
|
| /linux/fs/proc/ |
| H A D | kcore.c | 231 if (VMALLOC_START > ent->addr) { in kclist_add_private() 232 if (VMALLOC_START - ent->addr < ent->size) in kclist_add_private() 233 ent->size = VMALLOC_START - ent->addr; in kclist_add_private() 712 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { in add_modules_range() 733 kclist_add(&kcore_vmalloc, (void *)VMALLOC_START, in proc_kcore_init() 734 VMALLOC_END - VMALLOC_START, KCORE_VMALLOC); in proc_kcore_init()
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 49 self.VMALLOC_START = int(gdb.parse_and_eval("vmalloc_base")) 50 if self.VMALLOC_START == 0xffffc90000000000: 51 self.VMALLOC_END = self.VMALLOC_START + (32 * 1024 * 1024 * 1024 * 1024) - 1 52 elif self.VMALLOC_START == 0xffa0000000000000: 53 self.VMALLOC_END = self.VMALLOC_START + (12800 * 1024 * 1024 * 1024 * 1024) - 1 55 self.VMALLOC_END = self.VMALLOC_START + (12800 * 1024 * 1024 * 1024 * 1024) - 1 273 self.VMALLOC_START = self.MODULES_END
|
| H A D | vmalloc.py | 20 return addr >= pg_ops.VMALLOC_START and addr < pg_ops.VMALLOC_END
|
| /linux/arch/powerpc/mm/ |
| H A D | mem.c | 419 start = VMALLOC_START; in execmem_arch_setup() 436 .start = VMALLOC_START, in execmem_arch_setup() 442 .start = VMALLOC_START, in execmem_arch_setup()
|
| /linux/Documentation/arch/xtensa/ |
| H A D | mmu.rst | 83 | VMALLOC area | VMALLOC_START 0xc0000000 128MB - 64KB 126 | VMALLOC area | VMALLOC_START 0xa0000000 128MB - 64KB 170 | VMALLOC area | VMALLOC_START 0x90000000 128MB - 64KB
|
| /linux/arch/s390/boot/ |
| H A D | startup.c | 34 unsigned long __bootdata_preserved(VMALLOC_START); 413 VMALLOC_START = VMALLOC_END - vmalloc_size; in setup_kernel_memory_layout() 414 boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END); in setup_kernel_memory_layout() 416 __memcpy_real_area = round_down(VMALLOC_START - MEMCPY_REAL_SIZE, PAGE_SIZE); in setup_kernel_memory_layout()
|
| /linux/include/linux/ |
| H A D | ioremap.h | 15 #define IOREMAP_START VMALLOC_START
|
| /linux/arch/csky/include/asm/ |
| H A D | memory.h | 13 #define VMALLOC_START (PAGE_OFFSET + LOWMEM_LIMIT + (PAGE_SIZE * 8)) macro
|
| /linux/arch/arm/mm/ |
| H A D | iomap.c | 40 if ((unsigned long)addr >= VMALLOC_START && in pci_iounmap()
|
| H A D | pageattr.c | 67 !range_in_range(start, size, VMALLOC_START, VMALLOC_END)) in change_memory_common()
|
| /linux/arch/parisc/include/asm/ |
| H A D | fixmap.h | 54 #define VMALLOC_START ((unsigned long)parisc_vmalloc_start) macro
|
| /linux/lib/ |
| H A D | test_debug_virtual.c | 27 va = (void *)VMALLOC_START; in test_debug_virtual_init()
|
| /linux/arch/arm/include/asm/ |
| H A D | pgtable-nommu.h | 62 #define VMALLOC_START 0UL macro
|
| /linux/arch/hexagon/include/asm/ |
| H A D | mem-layout.h | 62 #define VMALLOC_START ((unsigned long) __va(high_memory + VMALLOC_OFFSET)) macro
|
| /linux/arch/x86/mm/ |
| H A D | kasan_init_64.c | 392 kasan_mem_to_shadow((void *)VMALLOC_START)); in kasan_init() 401 kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init() 405 kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init()
|
| /linux/Documentation/arch/arm/ |
| H A D | memory.rst | 59 VMALLOC_START VMALLOC_END-1 vmalloc() / ioremap() space. 64 VMALLOC_START is based upon the value
|
| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 495 (void *)kasan_mem_to_shadow((void *)VMALLOC_START)); in kasan_init() 499 (void *)kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init() 506 kasan_populate_early_shadow((void *)kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init()
|
| /linux/mm/kmsan/ |
| H A D | kmsan.h | 184 return ((u64)addr >= VMALLOC_START) && ((u64)addr < VMALLOC_END); in kmsan_internal_is_vmalloc_addr()
|
| /linux/arch/riscv/include/asm/ |
| H A D | pgtable.h | 45 #define VMALLOC_START (PAGE_OFFSET - VMALLOC_SIZE) macro 63 #define MODULES_VADDR VMALLOC_START 87 #define VMEMMAP_END VMALLOC_START 88 #define VMEMMAP_START (VMALLOC_START - VMEMMAP_SIZE) 1252 #define VMALLOC_START _AC(0, UL) macro
|
| /linux/kernel/ |
| H A D | scs.c | 46 s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END, in __scs_alloc()
|
| /linux/mm/ |
| H A D | vmalloc.c | 83 return addr >= VMALLOC_START && addr < VMALLOC_END; in is_vmalloc_addr() 1655 vstart = VMALLOC_START + rnd; in find_vmap_lowest_match_check() 1995 if (vstart != VMALLOC_START || vend != VMALLOC_END || in node_alloc() 2675 addr -= VMALLOC_START & ~(VMAP_BLOCK_SIZE-1); in addr_to_vb_idx() 2714 VMALLOC_START, VMALLOC_END, in new_vmap_block() 3027 BUG_ON(addr < VMALLOC_START); in vm_unmap_ram() 3076 VMALLOC_START, VMALLOC_END, in vm_map_ram() 3168 unsigned long addr = ALIGN(VMALLOC_START, align); in vm_area_register_early() 3269 VMALLOC_START, VMALLOC_END, in get_vm_area() 3278 VMALLOC_START, VMALLOC_END, in get_vm_area_caller() [all …]
|