160727d8bSWarner Losh /*- 2796df753SPedro F. Giffuni * SPDX-License-Identifier: (BSD-3-Clause AND MIT-CMU) 351369649SPedro F. Giffuni * 4df8bae1dSRodney W. Grimes * Copyright (c) 1991, 1993 5df8bae1dSRodney W. Grimes * The Regents of the University of California. All rights reserved. 6df8bae1dSRodney W. Grimes * 7df8bae1dSRodney W. Grimes * This code is derived from software contributed to Berkeley by 8df8bae1dSRodney W. Grimes * The Mach Operating System project at Carnegie-Mellon University. 9df8bae1dSRodney W. Grimes * 10df8bae1dSRodney W. Grimes * Redistribution and use in source and binary forms, with or without 11df8bae1dSRodney W. Grimes * modification, are permitted provided that the following conditions 12df8bae1dSRodney W. Grimes * are met: 13df8bae1dSRodney W. Grimes * 1. Redistributions of source code must retain the above copyright 14df8bae1dSRodney W. Grimes * notice, this list of conditions and the following disclaimer. 15df8bae1dSRodney W. Grimes * 2. Redistributions in binary form must reproduce the above copyright 16df8bae1dSRodney W. Grimes * notice, this list of conditions and the following disclaimer in the 17df8bae1dSRodney W. Grimes * documentation and/or other materials provided with the distribution. 18fbbd9655SWarner Losh * 3. Neither the name of the University nor the names of its contributors 19df8bae1dSRodney W. Grimes * may be used to endorse or promote products derived from this software 20df8bae1dSRodney W. Grimes * without specific prior written permission. 21df8bae1dSRodney W. Grimes * 22df8bae1dSRodney W. Grimes * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 23df8bae1dSRodney W. Grimes * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 24df8bae1dSRodney W. Grimes * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 25df8bae1dSRodney W. Grimes * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 26df8bae1dSRodney W. Grimes * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 27df8bae1dSRodney W. Grimes * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 28df8bae1dSRodney W. Grimes * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 29df8bae1dSRodney W. Grimes * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 30df8bae1dSRodney W. Grimes * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 31df8bae1dSRodney W. Grimes * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 32df8bae1dSRodney W. Grimes * SUCH DAMAGE. 33df8bae1dSRodney W. Grimes * 343c4dd356SDavid Greenman * from: @(#)vm_init.c 8.1 (Berkeley) 6/11/93 35df8bae1dSRodney W. Grimes * 36df8bae1dSRodney W. Grimes * 37df8bae1dSRodney W. Grimes * Copyright (c) 1987, 1990 Carnegie-Mellon University. 38df8bae1dSRodney W. Grimes * All rights reserved. 39df8bae1dSRodney W. Grimes * 40df8bae1dSRodney W. Grimes * Authors: Avadis Tevanian, Jr., Michael Wayne Young 41df8bae1dSRodney W. Grimes * 42df8bae1dSRodney W. Grimes * Permission to use, copy, modify and distribute this software and 43df8bae1dSRodney W. Grimes * its documentation is hereby granted, provided that both the copyright 44df8bae1dSRodney W. Grimes * notice and this permission notice appear in all copies of the 45df8bae1dSRodney W. Grimes * software, derivative works or modified versions, and any portions 46df8bae1dSRodney W. Grimes * thereof, and that both notices appear in supporting documentation. 47df8bae1dSRodney W. Grimes * 48df8bae1dSRodney W. Grimes * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" 49df8bae1dSRodney W. Grimes * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND 50df8bae1dSRodney W. Grimes * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. 51df8bae1dSRodney W. Grimes * 52df8bae1dSRodney W. Grimes * Carnegie Mellon requests users of this software to return to 53df8bae1dSRodney W. Grimes * 54df8bae1dSRodney W. Grimes * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU 55df8bae1dSRodney W. Grimes * School of Computer Science 56df8bae1dSRodney W. Grimes * Carnegie Mellon University 57df8bae1dSRodney W. Grimes * Pittsburgh PA 15213-3890 58df8bae1dSRodney W. Grimes * 59df8bae1dSRodney W. Grimes * any improvements or extensions that they make and grant Carnegie the 60df8bae1dSRodney W. Grimes * rights to redistribute these changes. 61df8bae1dSRodney W. Grimes */ 62df8bae1dSRodney W. Grimes 63df8bae1dSRodney W. Grimes /* 64df8bae1dSRodney W. Grimes * Initialize the Virtual Memory subsystem. 65df8bae1dSRodney W. Grimes */ 66df8bae1dSRodney W. Grimes 67874651b1SDavid E. O'Brien #include <sys/cdefs.h> 68874651b1SDavid E. O'Brien __FBSDID("$FreeBSD$"); 69874651b1SDavid E. O'Brien 70df8bae1dSRodney W. Grimes #include <sys/param.h> 712b14f991SJulian Elischer #include <sys/kernel.h> 72fb919e4dSMark Murray #include <sys/lock.h> 73fb919e4dSMark Murray #include <sys/proc.h> 7489f6b863SAttilio Rao #include <sys/rwlock.h> 755df87b21SJeff Roberson #include <sys/malloc.h> 76f5fca0d8SKris Kennaway #include <sys/sysctl.h> 7705f0fdd2SPoul-Henning Kamp #include <sys/systm.h> 78cebde069SMike Silbersack #include <sys/selinfo.h> 7921fae961SJeff Roberson #include <sys/smp.h> 80cebde069SMike Silbersack #include <sys/pipe.h> 81219d632cSMatthew Dillon #include <sys/bio.h> 82219d632cSMatthew Dillon #include <sys/buf.h> 835f518366SJeff Roberson #include <sys/vmem.h> 847a469c8eSJeff Roberson #include <sys/vmmeter.h> 85df8bae1dSRodney W. Grimes 86df8bae1dSRodney W. Grimes #include <vm/vm.h> 87219d632cSMatthew Dillon #include <vm/vm_param.h> 88219d632cSMatthew Dillon #include <vm/vm_kern.h> 89efeaf95aSDavid Greenman #include <vm/vm_object.h> 90df8bae1dSRodney W. Grimes #include <vm/vm_page.h> 917a469c8eSJeff Roberson #include <vm/vm_phys.h> 92e2068d0bSJeff Roberson #include <vm/vm_pagequeue.h> 93efeaf95aSDavid Greenman #include <vm/vm_map.h> 9424a1cce3SDavid Greenman #include <vm/vm_pager.h> 95efeaf95aSDavid Greenman #include <vm/vm_extern.h> 96df8bae1dSRodney W. Grimes 97ae941b1bSGleb Smirnoff extern void uma_startup1(void); 98*f7d35785SGleb Smirnoff extern void uma_startup2(void); 99ae941b1bSGleb Smirnoff extern void vm_radix_reserve_kva(void); 1007a469c8eSJeff Roberson 1017a469c8eSJeff Roberson #if VM_NRESERVLEVEL > 0 1027a469c8eSJeff Roberson #define KVA_QUANTUM (1 << (VM_LEVEL_0_ORDER + PAGE_SHIFT)) 1037a469c8eSJeff Roberson #else 1047a469c8eSJeff Roberson /* On non-superpage architectures want large import sizes. */ 1057a469c8eSJeff Roberson #define KVA_QUANTUM (PAGE_SIZE * 1024) 1067a469c8eSJeff Roberson #endif 1070fca57b8SThomas Moestl long physmem; 1080fca57b8SThomas Moestl 109df8bae1dSRodney W. Grimes /* 1102b14f991SJulian Elischer * System initialization 1112b14f991SJulian Elischer */ 11211caded3SAlfred Perlstein static void vm_mem_init(void *); 113237fdd78SRobert Watson SYSINIT(vm_mem, SI_SUB_VM, SI_ORDER_FIRST, vm_mem_init, NULL); 1142b14f991SJulian Elischer 1152b14f991SJulian Elischer /* 1165df87b21SJeff Roberson * Import kva into the kernel arena. 1175df87b21SJeff Roberson */ 1185df87b21SJeff Roberson static int 1195df87b21SJeff Roberson kva_import(void *unused, vmem_size_t size, int flags, vmem_addr_t *addrp) 1205df87b21SJeff Roberson { 1215df87b21SJeff Roberson vm_offset_t addr; 1225df87b21SJeff Roberson int result; 1235df87b21SJeff Roberson 1247a469c8eSJeff Roberson KASSERT((size % KVA_QUANTUM) == 0, 1257a469c8eSJeff Roberson ("kva_import: Size %jd is not a multiple of %d", 1267a469c8eSJeff Roberson (intmax_t)size, (int)KVA_QUANTUM)); 1275df87b21SJeff Roberson addr = vm_map_min(kernel_map); 128edb572a3SJohn Baldwin result = vm_map_find(kernel_map, NULL, 0, &addr, size, 0, 1295aa60b6fSJohn Baldwin VMFS_SUPER_SPACE, VM_PROT_ALL, VM_PROT_ALL, MAP_NOFAULT); 1305df87b21SJeff Roberson if (result != KERN_SUCCESS) 1315df87b21SJeff Roberson return (ENOMEM); 1325df87b21SJeff Roberson 1335df87b21SJeff Roberson *addrp = addr; 1345df87b21SJeff Roberson 1355df87b21SJeff Roberson return (0); 1365df87b21SJeff Roberson } 1375df87b21SJeff Roberson 1385df87b21SJeff Roberson /* 139df8bae1dSRodney W. Grimes * vm_init initializes the virtual memory system. 140df8bae1dSRodney W. Grimes * This is done only by the first cpu up. 141df8bae1dSRodney W. Grimes * 142df8bae1dSRodney W. Grimes * The start and end address of physical memory is passed in. 143df8bae1dSRodney W. Grimes */ 1442b14f991SJulian Elischer /* ARGSUSED*/ 1452b14f991SJulian Elischer static void 146d841aaa7SBruce Evans vm_mem_init(dummy) 147d841aaa7SBruce Evans void *dummy; 148df8bae1dSRodney W. Grimes { 1497a469c8eSJeff Roberson int domain; 1505df87b21SJeff Roberson 151df8bae1dSRodney W. Grimes /* 1520d94caffSDavid Greenman * Initializes resident memory structures. From here on, all physical 1530d94caffSDavid Greenman * memory is accounted for, and we use only virtual addresses. 154df8bae1dSRodney W. Grimes */ 15526f9a767SRodney W. Grimes vm_set_page_size(); 156889eb0fcSAlan Cox virtual_avail = vm_page_startup(virtual_avail); 157e7841165SDag-Erling Smørgrav 158ae941b1bSGleb Smirnoff #ifdef UMA_MD_SMALL_ALLOC 159ae941b1bSGleb Smirnoff /* Announce page availability to UMA. */ 160ae941b1bSGleb Smirnoff uma_startup1(); 161ae941b1bSGleb Smirnoff #endif 162df8bae1dSRodney W. Grimes /* 163df8bae1dSRodney W. Grimes * Initialize other VM packages 164df8bae1dSRodney W. Grimes */ 1655f518366SJeff Roberson vmem_startup(); 166a316d390SJohn Dyson vm_object_init(); 167df8bae1dSRodney W. Grimes vm_map_startup(); 168df8bae1dSRodney W. Grimes kmem_init(virtual_avail, virtual_end); 1695df87b21SJeff Roberson 1705df87b21SJeff Roberson /* 1715df87b21SJeff Roberson * Initialize the kernel_arena. This can grow on demand. 1725df87b21SJeff Roberson */ 1735df87b21SJeff Roberson vmem_init(kernel_arena, "kernel arena", 0, 0, PAGE_SIZE, 0, 0); 1747a469c8eSJeff Roberson vmem_set_import(kernel_arena, kva_import, NULL, NULL, KVA_QUANTUM); 1757a469c8eSJeff Roberson 1767a469c8eSJeff Roberson for (domain = 0; domain < vm_ndomains; domain++) { 1777a469c8eSJeff Roberson vm_dom[domain].vmd_kernel_arena = vmem_create( 1787a469c8eSJeff Roberson "kernel arena domain", 0, 0, PAGE_SIZE, 0, M_WAITOK); 1797a469c8eSJeff Roberson vmem_set_import(vm_dom[domain].vmd_kernel_arena, 1807a469c8eSJeff Roberson (vmem_import_t *)vmem_alloc, NULL, kernel_arena, 1817a469c8eSJeff Roberson KVA_QUANTUM); 1827a469c8eSJeff Roberson } 1835df87b21SJeff Roberson 184ae941b1bSGleb Smirnoff #ifndef UMA_MD_SMALL_ALLOC 185ae941b1bSGleb Smirnoff /* Set up radix zone to use noobj_alloc. */ 186ae941b1bSGleb Smirnoff vm_radix_reserve_kva(); 187ae941b1bSGleb Smirnoff #endif 188*f7d35785SGleb Smirnoff /* Announce full page availability to UMA. */ 189*f7d35785SGleb Smirnoff uma_startup2(); 1905df87b21SJeff Roberson kmem_init_zero_region(); 191bdb93eb2SAlan Cox pmap_init(); 192df8bae1dSRodney W. Grimes vm_pager_init(); 193df8bae1dSRodney W. Grimes } 194219d632cSMatthew Dillon 195219d632cSMatthew Dillon void 196219d632cSMatthew Dillon vm_ksubmap_init(struct kva_md_info *kmi) 197219d632cSMatthew Dillon { 198219d632cSMatthew Dillon vm_offset_t firstaddr; 199219d632cSMatthew Dillon caddr_t v; 200219d632cSMatthew Dillon vm_size_t size = 0; 201447b3772SPeter Wemm long physmem_est; 202219d632cSMatthew Dillon vm_offset_t minaddr; 203219d632cSMatthew Dillon vm_offset_t maxaddr; 204219d632cSMatthew Dillon 205219d632cSMatthew Dillon /* 206219d632cSMatthew Dillon * Allocate space for system data structures. 207219d632cSMatthew Dillon * The first available kernel virtual address is in "v". 208219d632cSMatthew Dillon * As pages of kernel virtual memory are allocated, "v" is incremented. 209219d632cSMatthew Dillon * As pages of memory are allocated and cleared, 210219d632cSMatthew Dillon * "firstaddr" is incremented. 211219d632cSMatthew Dillon */ 212219d632cSMatthew Dillon 213219d632cSMatthew Dillon /* 214219d632cSMatthew Dillon * Make two passes. The first pass calculates how much memory is 215219d632cSMatthew Dillon * needed and allocates it. The second pass assigns virtual 216219d632cSMatthew Dillon * addresses to the various data structures. 217219d632cSMatthew Dillon */ 218219d632cSMatthew Dillon firstaddr = 0; 219219d632cSMatthew Dillon again: 220219d632cSMatthew Dillon v = (caddr_t)firstaddr; 221219d632cSMatthew Dillon 222219d632cSMatthew Dillon /* 223219d632cSMatthew Dillon * Discount the physical memory larger than the size of kernel_map 224219d632cSMatthew Dillon * to avoid eating up all of KVA space. 225219d632cSMatthew Dillon */ 226447b3772SPeter Wemm physmem_est = lmin(physmem, btoc(kernel_map->max_offset - 227219d632cSMatthew Dillon kernel_map->min_offset)); 228219d632cSMatthew Dillon 229219d632cSMatthew Dillon v = kern_vfs_bio_buffer_alloc(v, physmem_est); 230219d632cSMatthew Dillon 231219d632cSMatthew Dillon /* 232219d632cSMatthew Dillon * End of first pass, size has been calculated so allocate memory 233219d632cSMatthew Dillon */ 234219d632cSMatthew Dillon if (firstaddr == 0) { 235857961d9SRobert Drehmel size = (vm_size_t)v; 236edb2994aSAndrew Gallatin #ifdef VM_FREELIST_DMA32 237edb2994aSAndrew Gallatin /* 238edb2994aSAndrew Gallatin * Try to protect 32-bit DMAable memory from the largest 239edb2994aSAndrew Gallatin * early alloc of wired mem. 240edb2994aSAndrew Gallatin */ 241edb2994aSAndrew Gallatin firstaddr = kmem_alloc_attr(kernel_arena, size, 242edb2994aSAndrew Gallatin M_ZERO | M_NOWAIT, (vm_paddr_t)1 << 32, 243edb2994aSAndrew Gallatin ~(vm_paddr_t)0, VM_MEMATTR_DEFAULT); 244edb2994aSAndrew Gallatin if (firstaddr == 0) 245edb2994aSAndrew Gallatin #endif 246edb2994aSAndrew Gallatin firstaddr = kmem_malloc(kernel_arena, size, 2475df87b21SJeff Roberson M_ZERO | M_WAITOK); 248219d632cSMatthew Dillon if (firstaddr == 0) 249219d632cSMatthew Dillon panic("startup: no room for tables"); 250219d632cSMatthew Dillon goto again; 251219d632cSMatthew Dillon } 252219d632cSMatthew Dillon 253219d632cSMatthew Dillon /* 254219d632cSMatthew Dillon * End of second pass, addresses have been assigned 255219d632cSMatthew Dillon */ 256219d632cSMatthew Dillon if ((vm_size_t)((char *)v - firstaddr) != size) 257219d632cSMatthew Dillon panic("startup: table size inconsistency"); 258219d632cSMatthew Dillon 2595df87b21SJeff Roberson /* 2605df87b21SJeff Roberson * Allocate the clean map to hold all of the paging and I/O virtual 2615df87b21SJeff Roberson * memory. 2625df87b21SJeff Roberson */ 2635f518366SJeff Roberson size = (long)nbuf * BKVASIZE + (long)nswbuf * MAXPHYS + 2645f518366SJeff Roberson (long)bio_transient_maxcnt * MAXPHYS; 2655df87b21SJeff Roberson kmi->clean_sva = firstaddr = kva_alloc(size); 2665df87b21SJeff Roberson kmi->clean_eva = firstaddr + size; 2675f518366SJeff Roberson 2685df87b21SJeff Roberson /* 2695df87b21SJeff Roberson * Allocate the buffer arena. 27021fae961SJeff Roberson * 27121fae961SJeff Roberson * Enable the quantum cache if we have more than 4 cpus. This 27221fae961SJeff Roberson * avoids lock contention at the expense of some fragmentation. 2735df87b21SJeff Roberson */ 2745f518366SJeff Roberson size = (long)nbuf * BKVASIZE; 2755df87b21SJeff Roberson kmi->buffer_sva = firstaddr; 2765f518366SJeff Roberson kmi->buffer_eva = kmi->buffer_sva + size; 2775f518366SJeff Roberson vmem_init(buffer_arena, "buffer arena", kmi->buffer_sva, size, 27821fae961SJeff Roberson PAGE_SIZE, (mp_ncpus > 4) ? BKVASIZE * 8 : 0, 0); 2795df87b21SJeff Roberson firstaddr += size; 2805f518366SJeff Roberson 2815df87b21SJeff Roberson /* 2825df87b21SJeff Roberson * Now swap kva. 2835df87b21SJeff Roberson */ 2845df87b21SJeff Roberson swapbkva = firstaddr; 2855f518366SJeff Roberson size = (long)nswbuf * MAXPHYS; 2865df87b21SJeff Roberson firstaddr += size; 2875f518366SJeff Roberson 2885df87b21SJeff Roberson /* 2895df87b21SJeff Roberson * And optionally transient bio space. 2905df87b21SJeff Roberson */ 2917db07e1cSKonstantin Belousov if (bio_transient_maxcnt != 0) { 2925f518366SJeff Roberson size = (long)bio_transient_maxcnt * MAXPHYS; 2935f518366SJeff Roberson vmem_init(transient_arena, "transient arena", 2945df87b21SJeff Roberson firstaddr, size, PAGE_SIZE, 0, 0); 2955df87b21SJeff Roberson firstaddr += size; 2967db07e1cSKonstantin Belousov } 2975df87b21SJeff Roberson if (firstaddr != kmi->clean_eva) 2985df87b21SJeff Roberson panic("Clean map calculation incorrect"); 2995df87b21SJeff Roberson 3005df87b21SJeff Roberson /* 301ec492b13SMark Johnston * Allocate the pageable submaps. We may cache an exec map entry per 302ec492b13SMark Johnston * CPU, so we therefore need to reserve space for at least ncpu+1 303ec492b13SMark Johnston * entries to avoid deadlock. The exec map is also used by some image 304ec492b13SMark Johnston * activators, so we leave a fixed number of pages for their use. 3055df87b21SJeff Roberson */ 306ec492b13SMark Johnston #ifdef __LP64__ 307ec492b13SMark Johnston exec_map_entries = 8 * mp_ncpus; 308ec492b13SMark Johnston #else 3098d65cba2SMark Johnston exec_map_entries = 2 * mp_ncpus + 4; 310ec492b13SMark Johnston #endif 311ec492b13SMark Johnston exec_map_entry_size = round_page(PATH_MAX + ARG_MAX); 312219d632cSMatthew Dillon exec_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr, 313ec492b13SMark Johnston exec_map_entries * exec_map_entry_size + 64 * PAGE_SIZE, FALSE); 3143202ed75SAlan Cox pipe_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr, maxpipekva, 3153202ed75SAlan Cox FALSE); 316219d632cSMatthew Dillon } 317