1 /*- 2 * Copyright (c) 1990 The Regents of the University of California. 3 * All rights reserved. 4 * Copyright (c) 1994 John S. Dyson 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to Berkeley by 8 * William Jolitz. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 3. Neither the name of the University nor the names of its contributors 19 * may be used to endorse or promote products derived from this software 20 * without specific prior written permission. 21 * 22 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 25 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 26 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 32 * SUCH DAMAGE. 33 * 34 * from: @(#)vmparam.h 5.9 (Berkeley) 5/12/91 35 * from: FreeBSD: src/sys/i386/include/vmparam.h,v 1.33 2000/03/30 36 * $FreeBSD$ 37 */ 38 39 #ifdef __arm__ 40 #include <arm/vmparam.h> 41 #else /* !__arm__ */ 42 43 #ifndef _MACHINE_VMPARAM_H_ 44 #define _MACHINE_VMPARAM_H_ 45 46 /* 47 * Virtual memory related constants, all in bytes 48 */ 49 #ifndef MAXTSIZ 50 #define MAXTSIZ (1*1024*1024*1024) /* max text size */ 51 #endif 52 #ifndef DFLDSIZ 53 #define DFLDSIZ (128*1024*1024) /* initial data size limit */ 54 #endif 55 #ifndef MAXDSIZ 56 #define MAXDSIZ (1*1024*1024*1024) /* max data size */ 57 #endif 58 #ifndef DFLSSIZ 59 #define DFLSSIZ (128*1024*1024) /* initial stack size limit */ 60 #endif 61 #ifndef MAXSSIZ 62 #define MAXSSIZ (1*1024*1024*1024) /* max stack size */ 63 #endif 64 #ifndef SGROWSIZ 65 #define SGROWSIZ (128*1024) /* amount to grow stack */ 66 #endif 67 68 /* 69 * The physical address space is sparsely populated. 70 */ 71 #define VM_PHYSSEG_SPARSE 72 73 /* 74 * The number of PHYSSEG entries. 75 */ 76 #define VM_PHYSSEG_MAX 64 77 78 /* 79 * Create two free page pools: VM_FREEPOOL_DEFAULT is the default pool 80 * from which physical pages are allocated and VM_FREEPOOL_DIRECT is 81 * the pool from which physical pages for small UMA objects are 82 * allocated. 83 */ 84 #define VM_NFREEPOOL 2 85 #define VM_FREEPOOL_DEFAULT 0 86 #define VM_FREEPOOL_DIRECT 1 87 88 /* 89 * Create one free page lists: VM_FREELIST_DEFAULT is for all physical 90 * pages. 91 */ 92 #define VM_NFREELIST 1 93 #define VM_FREELIST_DEFAULT 0 94 95 /* 96 * When PAGE_SIZE is 4KB, an allocation size of 16MB is supported in order 97 * to optimize the use of the direct map by UMA. Specifically, a 64-byte 98 * cache line contains at most 8 L2 BLOCK entries, collectively mapping 16MB 99 * of physical memory. By reducing the number of distinct 16MB "pages" that 100 * are used by UMA, the physical memory allocator reduces the likelihood of 101 * both 2MB page TLB misses and cache misses during the page table walk when 102 * a 2MB page TLB miss does occur. 103 */ 104 #define VM_NFREEORDER 13 105 106 /* 107 * Enable superpage reservations: 1 level. 108 */ 109 #ifndef VM_NRESERVLEVEL 110 #define VM_NRESERVLEVEL 1 111 #endif 112 113 /* 114 * Level 0 reservations consist of 512 pages. 115 */ 116 #ifndef VM_LEVEL_0_ORDER 117 #define VM_LEVEL_0_ORDER 9 118 #endif 119 120 /** 121 * Address space layout. 122 * 123 * ARMv8 implements up to a 48 bit virtual address space. The address space is 124 * split into 2 regions at each end of the 64 bit address space, with an 125 * out of range "hole" in the middle. 126 * 127 * We use the full 48 bits for each region, however the kernel may only use 128 * a limited range within this space. 129 * 130 * Upper region: 0xffffffffffffffff Top of virtual memory 131 * 132 * 0xfffffeffffffffff End of DMAP 133 * 0xffffa00000000000 Start of DMAP 134 * 135 * 0xffff009fffffffff End of KASAN shadow map 136 * 0xffff008000000000 Start of KASAN shadow map 137 * 138 * 0xffff007fffffffff End of KVA 139 * 0xffff000000000000 Kernel base address & start of KVA 140 * 141 * Hole: 0xfffeffffffffffff 142 * 0x0001000000000000 143 * 144 * Lower region: 0x0000ffffffffffff End of user address space 145 * 0x0000000000000000 Start of user address space 146 * 147 * We use the upper region for the kernel, and the lower region for userland. 148 * 149 * We define some interesting address constants: 150 * 151 * VM_MIN_ADDRESS and VM_MAX_ADDRESS define the start and end of the entire 152 * 64 bit address space, mostly just for convenience. 153 * 154 * VM_MIN_KERNEL_ADDRESS and VM_MAX_KERNEL_ADDRESS define the start and end of 155 * mappable kernel virtual address space. 156 * 157 * VM_MIN_USER_ADDRESS and VM_MAX_USER_ADDRESS define the start and end of the 158 * user address space. 159 */ 160 #define VM_MIN_ADDRESS (0x0000000000000000UL) 161 #define VM_MAX_ADDRESS (0xffffffffffffffffUL) 162 163 /* 512 GiB of kernel addresses */ 164 #define VM_MIN_KERNEL_ADDRESS (0xffff000000000000UL) 165 #define VM_MAX_KERNEL_ADDRESS (0xffff008000000000UL) 166 167 /* 128 GiB KASAN shadow map */ 168 #define KASAN_MIN_ADDRESS (0xffff008000000000UL) 169 #define KASAN_MAX_ADDRESS (0xffff00a000000000UL) 170 171 /* The address bits that hold a pointer authentication code */ 172 #define PAC_ADDR_MASK (0xff7f000000000000UL) 173 174 /* If true addr is in the kernel address space */ 175 #define ADDR_IS_KERNEL(addr) (((addr) & (1ul << 55)) == (1ul << 55)) 176 /* If true addr is in its canonical form (i.e. no TBI, PAC, etc.) */ 177 #define ADDR_IS_CANONICAL(addr) \ 178 (((addr) & 0xffff000000000000UL) == 0 || \ 179 ((addr) & 0xffff000000000000UL) == 0xffff000000000000UL) 180 #define ADDR_MAKE_CANONICAL(addr) ({ \ 181 __typeof(addr) _tmp_addr = (addr); \ 182 \ 183 _tmp_addr &= ~0xffff000000000000UL; \ 184 if (ADDR_IS_KERNEL(addr)) \ 185 _tmp_addr |= 0xffff000000000000UL; \ 186 \ 187 _tmp_addr; \ 188 }) 189 190 /* 95 TiB maximum for the direct map region */ 191 #define DMAP_MIN_ADDRESS (0xffffa00000000000UL) 192 #define DMAP_MAX_ADDRESS (0xffffff0000000000UL) 193 194 #define DMAP_MIN_PHYSADDR (dmap_phys_base) 195 #define DMAP_MAX_PHYSADDR (dmap_phys_max) 196 197 /* True if pa is in the dmap range */ 198 #define PHYS_IN_DMAP(pa) ((pa) >= DMAP_MIN_PHYSADDR && \ 199 (pa) < DMAP_MAX_PHYSADDR) 200 /* True if va is in the dmap range */ 201 #define VIRT_IN_DMAP(va) ((va) >= DMAP_MIN_ADDRESS && \ 202 (va) < (dmap_max_addr)) 203 204 #define PMAP_HAS_DMAP 1 205 #define PHYS_TO_DMAP(pa) \ 206 ({ \ 207 KASSERT(PHYS_IN_DMAP(pa), \ 208 ("%s: PA out of range, PA: 0x%lx", __func__, \ 209 (vm_paddr_t)(pa))); \ 210 ((pa) - dmap_phys_base) + DMAP_MIN_ADDRESS; \ 211 }) 212 213 #define DMAP_TO_PHYS(va) \ 214 ({ \ 215 KASSERT(VIRT_IN_DMAP(va), \ 216 ("%s: VA out of range, VA: 0x%lx", __func__, \ 217 (vm_offset_t)(va))); \ 218 ((va) - DMAP_MIN_ADDRESS) + dmap_phys_base; \ 219 }) 220 221 #define VM_MIN_USER_ADDRESS (0x0000000000000000UL) 222 #define VM_MAX_USER_ADDRESS (0x0001000000000000UL) 223 224 #define VM_MINUSER_ADDRESS (VM_MIN_USER_ADDRESS) 225 #define VM_MAXUSER_ADDRESS (VM_MAX_USER_ADDRESS) 226 227 #define KERNBASE (VM_MIN_KERNEL_ADDRESS) 228 #define SHAREDPAGE (VM_MAXUSER_ADDRESS - PAGE_SIZE) 229 #define USRSTACK SHAREDPAGE 230 231 /* 232 * How many physical pages per kmem arena virtual page. 233 */ 234 #ifndef VM_KMEM_SIZE_SCALE 235 #define VM_KMEM_SIZE_SCALE (1) 236 #endif 237 238 /* 239 * Optional ceiling (in bytes) on the size of the kmem arena: 60% of the 240 * kernel map. 241 */ 242 #ifndef VM_KMEM_SIZE_MAX 243 #define VM_KMEM_SIZE_MAX ((VM_MAX_KERNEL_ADDRESS - \ 244 VM_MIN_KERNEL_ADDRESS + 1) * 3 / 5) 245 #endif 246 247 /* 248 * Initial pagein size of beginning of executable file. 249 */ 250 #ifndef VM_INITIAL_PAGEIN 251 #define VM_INITIAL_PAGEIN 16 252 #endif 253 254 #if !defined(KASAN) && !defined(KMSAN) 255 #define UMA_MD_SMALL_ALLOC 256 #endif 257 258 #ifndef LOCORE 259 260 extern vm_paddr_t dmap_phys_base; 261 extern vm_paddr_t dmap_phys_max; 262 extern vm_offset_t dmap_max_addr; 263 extern vm_offset_t vm_max_kernel_address; 264 265 #endif 266 267 #define ZERO_REGION_SIZE (64 * 1024) /* 64KB */ 268 269 #define DEVMAP_MAX_VADDR VM_MAX_KERNEL_ADDRESS 270 271 /* 272 * The pmap can create non-transparent large page mappings. 273 */ 274 #define PMAP_HAS_LARGEPAGES 1 275 276 /* 277 * Need a page dump array for minidump. 278 */ 279 #define MINIDUMP_PAGE_TRACKING 1 280 281 #endif /* !_MACHINE_VMPARAM_H_ */ 282 283 #endif /* !__arm__ */ 284