xref: /freebsd/sys/i386/i386/vm_machdep.c (revision dca5129987a1ec4da4a2627b0a4e74a4eaebc5c9)
15b81b6b3SRodney W. Grimes /*-
25b81b6b3SRodney W. Grimes  * Copyright (c) 1982, 1986 The Regents of the University of California.
35b81b6b3SRodney W. Grimes  * Copyright (c) 1989, 1990 William Jolitz
41561d038SDavid Greenman  * Copyright (c) 1994 John Dyson
55b81b6b3SRodney W. Grimes  * All rights reserved.
65b81b6b3SRodney W. Grimes  *
75b81b6b3SRodney W. Grimes  * This code is derived from software contributed to Berkeley by
85b81b6b3SRodney W. Grimes  * the Systems Programming Group of the University of Utah Computer
95b81b6b3SRodney W. Grimes  * Science Department, and William Jolitz.
105b81b6b3SRodney W. Grimes  *
115b81b6b3SRodney W. Grimes  * Redistribution and use in source and binary forms, with or without
125b81b6b3SRodney W. Grimes  * modification, are permitted provided that the following conditions
135b81b6b3SRodney W. Grimes  * are met:
145b81b6b3SRodney W. Grimes  * 1. Redistributions of source code must retain the above copyright
155b81b6b3SRodney W. Grimes  *    notice, this list of conditions and the following disclaimer.
165b81b6b3SRodney W. Grimes  * 2. Redistributions in binary form must reproduce the above copyright
175b81b6b3SRodney W. Grimes  *    notice, this list of conditions and the following disclaimer in the
185b81b6b3SRodney W. Grimes  *    documentation and/or other materials provided with the distribution.
195b81b6b3SRodney W. Grimes  * 3. All advertising materials mentioning features or use of this software
205b81b6b3SRodney W. Grimes  *    must display the following acknowledgement:
215b81b6b3SRodney W. Grimes  *	This product includes software developed by the University of
225b81b6b3SRodney W. Grimes  *	California, Berkeley and its contributors.
235b81b6b3SRodney W. Grimes  * 4. Neither the name of the University nor the names of its contributors
245b81b6b3SRodney W. Grimes  *    may be used to endorse or promote products derived from this software
255b81b6b3SRodney W. Grimes  *    without specific prior written permission.
265b81b6b3SRodney W. Grimes  *
275b81b6b3SRodney W. Grimes  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
285b81b6b3SRodney W. Grimes  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
295b81b6b3SRodney W. Grimes  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
305b81b6b3SRodney W. Grimes  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
315b81b6b3SRodney W. Grimes  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
325b81b6b3SRodney W. Grimes  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
335b81b6b3SRodney W. Grimes  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
345b81b6b3SRodney W. Grimes  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
355b81b6b3SRodney W. Grimes  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
365b81b6b3SRodney W. Grimes  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
375b81b6b3SRodney W. Grimes  * SUCH DAMAGE.
385b81b6b3SRodney W. Grimes  *
39960173b9SRodney W. Grimes  *	from: @(#)vm_machdep.c	7.3 (Berkeley) 5/13/91
405b81b6b3SRodney W. Grimes  *	Utah $Hdr: vm_machdep.c 1.16.1.1 89/06/23$
41267173e7SDavid Greenman  *	$Id: vm_machdep.c,v 1.53 1996/01/30 12:54:21 davidg Exp $
425b81b6b3SRodney W. Grimes  */
435b81b6b3SRodney W. Grimes 
44960173b9SRodney W. Grimes #include "npx.h"
458890984dSGarrett Wollman #include "opt_bounce.h"
468890984dSGarrett Wollman 
4726f9a767SRodney W. Grimes #include <sys/param.h>
4826f9a767SRodney W. Grimes #include <sys/systm.h>
4926f9a767SRodney W. Grimes #include <sys/proc.h>
5026f9a767SRodney W. Grimes #include <sys/malloc.h>
5126f9a767SRodney W. Grimes #include <sys/buf.h>
5226f9a767SRodney W. Grimes #include <sys/vnode.h>
53efeaf95aSDavid Greenman #include <sys/vmmeter.h>
545b81b6b3SRodney W. Grimes 
552320728fSRodney W. Grimes #include <machine/clock.h>
5626f9a767SRodney W. Grimes #include <machine/cpu.h>
57b5e8ce9fSBruce Evans #include <machine/md_var.h>
585b81b6b3SRodney W. Grimes 
5926f9a767SRodney W. Grimes #include <vm/vm.h>
60efeaf95aSDavid Greenman #include <vm/vm_param.h>
61efeaf95aSDavid Greenman #include <vm/vm_prot.h>
62efeaf95aSDavid Greenman #include <vm/lock.h>
6326f9a767SRodney W. Grimes #include <vm/vm_kern.h>
6424a1cce3SDavid Greenman #include <vm/vm_page.h>
65efeaf95aSDavid Greenman #include <vm/vm_map.h>
66efeaf95aSDavid Greenman #include <vm/vm_extern.h>
67efeaf95aSDavid Greenman 
68efeaf95aSDavid Greenman #include <sys/user.h>
695b81b6b3SRodney W. Grimes 
702320728fSRodney W. Grimes #include <i386/isa/isa.h>
712320728fSRodney W. Grimes 
729aba88bfSDavid Greenman #ifdef BOUNCE_BUFFERS
7387b91157SPoul-Henning Kamp static vm_offset_t
746c146e28SBruce Evans 		vm_bounce_kva __P((int size, int waitok));
7587b91157SPoul-Henning Kamp static void	vm_bounce_kva_free __P((vm_offset_t addr, vm_offset_t size,
766c146e28SBruce Evans 					int now));
7787b91157SPoul-Henning Kamp static vm_offset_t
786c146e28SBruce Evans 		vm_bounce_page_find __P((int count));
7987b91157SPoul-Henning Kamp static void	vm_bounce_page_free __P((vm_offset_t pa, int count));
806c146e28SBruce Evans 
8187b6de2bSPoul-Henning Kamp static volatile int	kvasfreecnt;
82ae92ea44SDavid Greenman 
83d5e26ef0SDavid Greenman caddr_t		bouncememory;
8487b91157SPoul-Henning Kamp int		bouncepages;
8587b91157SPoul-Henning Kamp static int	bpwait;
8687b91157SPoul-Henning Kamp static vm_offset_t	*bouncepa;
8787b91157SPoul-Henning Kamp static int		bmwait, bmfreeing;
88d5e26ef0SDavid Greenman 
89ed7fcbd0SDavid Greenman #define BITS_IN_UNSIGNED (8*sizeof(unsigned))
9087b91157SPoul-Henning Kamp static int		bounceallocarraysize;
9187b91157SPoul-Henning Kamp static unsigned	*bounceallocarray;
9287b91157SPoul-Henning Kamp static int		bouncefree;
93d5e26ef0SDavid Greenman 
94d5e26ef0SDavid Greenman #define SIXTEENMEG (4096*4096)
95ae92ea44SDavid Greenman #define MAXBKVA 1024
9626f9a767SRodney W. Grimes int		maxbkva = MAXBKVA*NBPG;
97d5e26ef0SDavid Greenman 
98d5e26ef0SDavid Greenman /* special list that can be used at interrupt time for eventual kva free */
9987b91157SPoul-Henning Kamp static struct kvasfree {
100d5e26ef0SDavid Greenman 	vm_offset_t addr;
101d5e26ef0SDavid Greenman 	vm_offset_t size;
102d5e26ef0SDavid Greenman } kvaf[MAXBKVA];
103d5e26ef0SDavid Greenman 
104d5e26ef0SDavid Greenman /*
105d5e26ef0SDavid Greenman  * get bounce buffer pages (count physically contiguous)
106d5e26ef0SDavid Greenman  * (only 1 inplemented now)
107d5e26ef0SDavid Greenman  */
10887b91157SPoul-Henning Kamp static vm_offset_t
109d5e26ef0SDavid Greenman vm_bounce_page_find(count)
110d5e26ef0SDavid Greenman 	int count;
111d5e26ef0SDavid Greenman {
112d5e26ef0SDavid Greenman 	int bit;
113d5e26ef0SDavid Greenman 	int s,i;
114d5e26ef0SDavid Greenman 
115d5e26ef0SDavid Greenman 	if (count != 1)
116d5e26ef0SDavid Greenman 		panic("vm_bounce_page_find -- no support for > 1 page yet!!!");
117d5e26ef0SDavid Greenman 
118d5e26ef0SDavid Greenman 	s = splbio();
119d5e26ef0SDavid Greenman retry:
120d5e26ef0SDavid Greenman 	for (i = 0; i < bounceallocarraysize; i++) {
121d5e26ef0SDavid Greenman 		if (bounceallocarray[i] != 0xffffffff) {
1223fb3086eSPoul-Henning Kamp 			bit = ffs(~bounceallocarray[i]);
1233fb3086eSPoul-Henning Kamp 			if (bit) {
124d5e26ef0SDavid Greenman 				bounceallocarray[i] |= 1 << (bit - 1) ;
125d5e26ef0SDavid Greenman 				bouncefree -= count;
126d5e26ef0SDavid Greenman 				splx(s);
127ae92ea44SDavid Greenman 				return bouncepa[(i * BITS_IN_UNSIGNED + (bit - 1))];
128d5e26ef0SDavid Greenman 			}
129d5e26ef0SDavid Greenman 		}
130d5e26ef0SDavid Greenman 	}
131ed7fcbd0SDavid Greenman 	bpwait = 1;
132d5e26ef0SDavid Greenman 	tsleep((caddr_t) &bounceallocarray, PRIBIO, "bncwai", 0);
133d5e26ef0SDavid Greenman 	goto retry;
134d5e26ef0SDavid Greenman }
135d5e26ef0SDavid Greenman 
13687b91157SPoul-Henning Kamp static void
1371561d038SDavid Greenman vm_bounce_kva_free(addr, size, now)
1381561d038SDavid Greenman 	vm_offset_t addr;
1391561d038SDavid Greenman 	vm_offset_t size;
1401561d038SDavid Greenman 	int now;
1411561d038SDavid Greenman {
1421561d038SDavid Greenman 	int s = splbio();
1431561d038SDavid Greenman 	kvaf[kvasfreecnt].addr = addr;
144ae92ea44SDavid Greenman 	kvaf[kvasfreecnt].size = size;
145ae92ea44SDavid Greenman 	++kvasfreecnt;
1460e195446SDavid Greenman 	if( now) {
1470e195446SDavid Greenman 		/*
1480e195446SDavid Greenman 		 * this will do wakeups
1490e195446SDavid Greenman 		 */
1501561d038SDavid Greenman 		vm_bounce_kva(0,0);
1510e195446SDavid Greenman 	} else {
1520e195446SDavid Greenman 		if (bmwait) {
1530e195446SDavid Greenman 		/*
1540e195446SDavid Greenman 		 * if anyone is waiting on the bounce-map, then wakeup
1550e195446SDavid Greenman 		 */
1561561d038SDavid Greenman 			wakeup((caddr_t) io_map);
1570e195446SDavid Greenman 			bmwait = 0;
1580e195446SDavid Greenman 		}
1590e195446SDavid Greenman 	}
1601561d038SDavid Greenman 	splx(s);
1611561d038SDavid Greenman }
1621561d038SDavid Greenman 
163d5e26ef0SDavid Greenman /*
164d5e26ef0SDavid Greenman  * free count bounce buffer pages
165d5e26ef0SDavid Greenman  */
16687b91157SPoul-Henning Kamp static void
167d5e26ef0SDavid Greenman vm_bounce_page_free(pa, count)
168d5e26ef0SDavid Greenman 	vm_offset_t pa;
169d5e26ef0SDavid Greenman 	int count;
170d5e26ef0SDavid Greenman {
171d5e26ef0SDavid Greenman 	int allocindex;
172d5e26ef0SDavid Greenman 	int index;
173d5e26ef0SDavid Greenman 	int bit;
174d5e26ef0SDavid Greenman 
175d5e26ef0SDavid Greenman 	if (count != 1)
176edf8a815SDavid Greenman 		panic("vm_bounce_page_free -- no support for > 1 page yet!!!");
177d5e26ef0SDavid Greenman 
178ae92ea44SDavid Greenman 	for(index=0;index<bouncepages;index++) {
179ae92ea44SDavid Greenman 		if( pa == bouncepa[index])
180ae92ea44SDavid Greenman 			break;
181ae92ea44SDavid Greenman 	}
182d5e26ef0SDavid Greenman 
183ae92ea44SDavid Greenman 	if( index == bouncepages)
184ae92ea44SDavid Greenman 		panic("vm_bounce_page_free: invalid bounce buffer");
185d5e26ef0SDavid Greenman 
186ed7fcbd0SDavid Greenman 	allocindex = index / BITS_IN_UNSIGNED;
187ed7fcbd0SDavid Greenman 	bit = index % BITS_IN_UNSIGNED;
188d5e26ef0SDavid Greenman 
189d5e26ef0SDavid Greenman 	bounceallocarray[allocindex] &= ~(1 << bit);
190d5e26ef0SDavid Greenman 
191d5e26ef0SDavid Greenman 	bouncefree += count;
192ed7fcbd0SDavid Greenman 	if (bpwait) {
193ed7fcbd0SDavid Greenman 		bpwait = 0;
194d5e26ef0SDavid Greenman 		wakeup((caddr_t) &bounceallocarray);
195d5e26ef0SDavid Greenman 	}
196ed7fcbd0SDavid Greenman }
197d5e26ef0SDavid Greenman 
198d5e26ef0SDavid Greenman /*
199d5e26ef0SDavid Greenman  * allocate count bounce buffer kva pages
200d5e26ef0SDavid Greenman  */
20187b91157SPoul-Henning Kamp static vm_offset_t
202ae92ea44SDavid Greenman vm_bounce_kva(size, waitok)
203ae92ea44SDavid Greenman 	int size;
2041561d038SDavid Greenman 	int waitok;
205d5e26ef0SDavid Greenman {
206d5e26ef0SDavid Greenman 	int i;
2076b4ac811SDavid Greenman 	vm_offset_t kva = 0;
208c564966bSDavid Greenman 	vm_offset_t off;
209d5e26ef0SDavid Greenman 	int s = splbio();
210d5e26ef0SDavid Greenman more:
211ae92ea44SDavid Greenman 	if (!bmfreeing && kvasfreecnt) {
212d5e26ef0SDavid Greenman 		bmfreeing = 1;
213ae92ea44SDavid Greenman 		for (i = 0; i < kvasfreecnt; i++) {
214c564966bSDavid Greenman 			for(off=0;off<kvaf[i].size;off+=NBPG) {
215c564966bSDavid Greenman 				pmap_kremove( kvaf[i].addr + off);
216c564966bSDavid Greenman 			}
2171561d038SDavid Greenman 			kmem_free_wakeup(io_map, kvaf[i].addr,
218d5e26ef0SDavid Greenman 				kvaf[i].size);
219d5e26ef0SDavid Greenman 		}
220d5e26ef0SDavid Greenman 		kvasfreecnt = 0;
221d5e26ef0SDavid Greenman 		bmfreeing = 0;
222ae92ea44SDavid Greenman 		if( bmwait) {
223ae92ea44SDavid Greenman 			bmwait = 0;
224ae92ea44SDavid Greenman 			wakeup( (caddr_t) io_map);
225ae92ea44SDavid Greenman 		}
226d5e26ef0SDavid Greenman 	}
227d5e26ef0SDavid Greenman 
2281561d038SDavid Greenman 	if( size == 0) {
2291561d038SDavid Greenman 		splx(s);
2301561d038SDavid Greenman 		return NULL;
2311561d038SDavid Greenman 	}
2321561d038SDavid Greenman 
233ae92ea44SDavid Greenman 	if ((kva = kmem_alloc_pageable(io_map, size)) == 0) {
2341561d038SDavid Greenman 		if( !waitok) {
2351561d038SDavid Greenman 			splx(s);
2361561d038SDavid Greenman 			return NULL;
2371561d038SDavid Greenman 		}
238d5e26ef0SDavid Greenman 		bmwait = 1;
2391561d038SDavid Greenman 		tsleep((caddr_t) io_map, PRIBIO, "bmwait", 0);
240d5e26ef0SDavid Greenman 		goto more;
241d5e26ef0SDavid Greenman 	}
242d5e26ef0SDavid Greenman 	splx(s);
243d5e26ef0SDavid Greenman 	return kva;
244d5e26ef0SDavid Greenman }
245d5e26ef0SDavid Greenman 
246d5e26ef0SDavid Greenman /*
247ae92ea44SDavid Greenman  * same as vm_bounce_kva -- but really allocate (but takes pages as arg)
2480e195446SDavid Greenman  */
2490e195446SDavid Greenman vm_offset_t
2500e195446SDavid Greenman vm_bounce_kva_alloc(count)
2510e195446SDavid Greenman int count;
2520e195446SDavid Greenman {
2530e195446SDavid Greenman 	int i;
2540e195446SDavid Greenman 	vm_offset_t kva;
2550e195446SDavid Greenman 	vm_offset_t pa;
2560e195446SDavid Greenman 	if( bouncepages == 0) {
2570e195446SDavid Greenman 		kva = (vm_offset_t) malloc(count*NBPG, M_TEMP, M_WAITOK);
2580e195446SDavid Greenman 		return kva;
2590e195446SDavid Greenman 	}
260ae92ea44SDavid Greenman 	kva = vm_bounce_kva(count*NBPG, 1);
2610e195446SDavid Greenman 	for(i=0;i<count;i++) {
2620e195446SDavid Greenman 		pa = vm_bounce_page_find(1);
2630e195446SDavid Greenman 		pmap_kenter(kva + i * NBPG, pa);
2640e195446SDavid Greenman 	}
2650e195446SDavid Greenman 	return kva;
2660e195446SDavid Greenman }
2670e195446SDavid Greenman 
2680e195446SDavid Greenman /*
2690e195446SDavid Greenman  * same as vm_bounce_kva_free -- but really free
2700e195446SDavid Greenman  */
2710e195446SDavid Greenman void
2720e195446SDavid Greenman vm_bounce_kva_alloc_free(kva, count)
2730e195446SDavid Greenman 	vm_offset_t kva;
2740e195446SDavid Greenman 	int count;
2750e195446SDavid Greenman {
2760e195446SDavid Greenman 	int i;
2770e195446SDavid Greenman 	vm_offset_t pa;
2780e195446SDavid Greenman 	if( bouncepages == 0) {
2790e195446SDavid Greenman 		free((caddr_t) kva, M_TEMP);
2800e195446SDavid Greenman 		return;
2810e195446SDavid Greenman 	}
2820e195446SDavid Greenman 	for(i = 0; i < count; i++) {
2830e195446SDavid Greenman 		pa = pmap_kextract(kva + i * NBPG);
2840e195446SDavid Greenman 		vm_bounce_page_free(pa, 1);
2850e195446SDavid Greenman 	}
286ae92ea44SDavid Greenman 	vm_bounce_kva_free(kva, count*NBPG, 0);
2870e195446SDavid Greenman }
2880e195446SDavid Greenman 
2890e195446SDavid Greenman /*
290d5e26ef0SDavid Greenman  * do the things necessary to the struct buf to implement
291d5e26ef0SDavid Greenman  * bounce buffers...  inserted before the disk sort
292d5e26ef0SDavid Greenman  */
293d5e26ef0SDavid Greenman void
294d5e26ef0SDavid Greenman vm_bounce_alloc(bp)
295d5e26ef0SDavid Greenman 	struct buf *bp;
296d5e26ef0SDavid Greenman {
297d5e26ef0SDavid Greenman 	int countvmpg;
298d5e26ef0SDavid Greenman 	vm_offset_t vastart, vaend;
299d5e26ef0SDavid Greenman 	vm_offset_t vapstart, vapend;
300d5e26ef0SDavid Greenman 	vm_offset_t va, kva;
301d5e26ef0SDavid Greenman 	vm_offset_t pa;
302d5e26ef0SDavid Greenman 	int dobounceflag = 0;
303d5e26ef0SDavid Greenman 	int i;
304d5e26ef0SDavid Greenman 
305d5e26ef0SDavid Greenman 	if (bouncepages == 0)
306d5e26ef0SDavid Greenman 		return;
307d5e26ef0SDavid Greenman 
308ae92ea44SDavid Greenman 	if (bp->b_flags & B_BOUNCE) {
309ae92ea44SDavid Greenman 		printf("vm_bounce_alloc: called recursively???\n");
310ae92ea44SDavid Greenman 		return;
3111561d038SDavid Greenman 	}
3121561d038SDavid Greenman 
313ae92ea44SDavid Greenman 	if (bp->b_bufsize < bp->b_bcount) {
3143fb3086eSPoul-Henning Kamp 		printf(
3153fb3086eSPoul-Henning Kamp 		    "vm_bounce_alloc: b_bufsize(0x%lx) < b_bcount(0x%lx) !!\n",
316ae92ea44SDavid Greenman 			bp->b_bufsize, bp->b_bcount);
317ae92ea44SDavid Greenman 		panic("vm_bounce_alloc");
318ae92ea44SDavid Greenman 	}
319ae92ea44SDavid Greenman 
320ae92ea44SDavid Greenman /*
321ae92ea44SDavid Greenman  *  This is not really necessary
322ae92ea44SDavid Greenman  *	if( bp->b_bufsize != bp->b_bcount) {
323ae92ea44SDavid Greenman  *		printf("size: %d, count: %d\n", bp->b_bufsize, bp->b_bcount);
324ae92ea44SDavid Greenman  *	}
325ae92ea44SDavid Greenman  */
326ae92ea44SDavid Greenman 
327ae92ea44SDavid Greenman 
32826f9a767SRodney W. Grimes 	vastart = (vm_offset_t) bp->b_data;
32926f9a767SRodney W. Grimes 	vaend = (vm_offset_t) bp->b_data + bp->b_bufsize;
330d5e26ef0SDavid Greenman 
331bd7e5f99SJohn Dyson 	vapstart = trunc_page(vastart);
332bd7e5f99SJohn Dyson 	vapend = round_page(vaend);
333d5e26ef0SDavid Greenman 	countvmpg = (vapend - vapstart) / NBPG;
334d5e26ef0SDavid Greenman 
335d5e26ef0SDavid Greenman /*
336d5e26ef0SDavid Greenman  * if any page is above 16MB, then go into bounce-buffer mode
337d5e26ef0SDavid Greenman  */
338d5e26ef0SDavid Greenman 	va = vapstart;
339d5e26ef0SDavid Greenman 	for (i = 0; i < countvmpg; i++) {
340ed7fcbd0SDavid Greenman 		pa = pmap_kextract(va);
341d5e26ef0SDavid Greenman 		if (pa >= SIXTEENMEG)
342d5e26ef0SDavid Greenman 			++dobounceflag;
3430d94caffSDavid Greenman 		if( pa == 0)
3440d94caffSDavid Greenman 			panic("vm_bounce_alloc: Unmapped page");
345d5e26ef0SDavid Greenman 		va += NBPG;
346d5e26ef0SDavid Greenman 	}
347d5e26ef0SDavid Greenman 	if (dobounceflag == 0)
348d5e26ef0SDavid Greenman 		return;
349d5e26ef0SDavid Greenman 
350d5e26ef0SDavid Greenman 	if (bouncepages < dobounceflag)
351d5e26ef0SDavid Greenman 		panic("Not enough bounce buffers!!!");
352d5e26ef0SDavid Greenman 
353d5e26ef0SDavid Greenman /*
354d5e26ef0SDavid Greenman  * allocate a replacement kva for b_addr
355d5e26ef0SDavid Greenman  */
3561561d038SDavid Greenman 	kva = vm_bounce_kva(countvmpg*NBPG, 1);
357ae92ea44SDavid Greenman #if 0
358ae92ea44SDavid Greenman 	printf("%s: vapstart: %x, vapend: %x, countvmpg: %d, kva: %x ",
359ae92ea44SDavid Greenman 		(bp->b_flags & B_READ) ? "read":"write",
360ae92ea44SDavid Greenman 			vapstart, vapend, countvmpg, kva);
361ae92ea44SDavid Greenman #endif
362d5e26ef0SDavid Greenman 	va = vapstart;
363d5e26ef0SDavid Greenman 	for (i = 0; i < countvmpg; i++) {
364ed7fcbd0SDavid Greenman 		pa = pmap_kextract(va);
365d5e26ef0SDavid Greenman 		if (pa >= SIXTEENMEG) {
366d5e26ef0SDavid Greenman 			/*
367d5e26ef0SDavid Greenman 			 * allocate a replacement page
368d5e26ef0SDavid Greenman 			 */
369d5e26ef0SDavid Greenman 			vm_offset_t bpa = vm_bounce_page_find(1);
3706b4ac811SDavid Greenman 			pmap_kenter(kva + (NBPG * i), bpa);
371ae92ea44SDavid Greenman #if 0
372ae92ea44SDavid Greenman 			printf("r(%d): (%x,%x,%x) ", i, va, pa, bpa);
373ae92ea44SDavid Greenman #endif
374d5e26ef0SDavid Greenman 			/*
375d5e26ef0SDavid Greenman 			 * if we are writing, the copy the data into the page
376d5e26ef0SDavid Greenman 			 */
3771561d038SDavid Greenman 			if ((bp->b_flags & B_READ) == 0) {
378d5e26ef0SDavid Greenman 				bcopy((caddr_t) va, (caddr_t) kva + (NBPG * i), NBPG);
3791561d038SDavid Greenman 			}
380d5e26ef0SDavid Greenman 		} else {
381d5e26ef0SDavid Greenman 			/*
382d5e26ef0SDavid Greenman 			 * use original page
383d5e26ef0SDavid Greenman 			 */
3846b4ac811SDavid Greenman 			pmap_kenter(kva + (NBPG * i), pa);
385d5e26ef0SDavid Greenman 		}
386d5e26ef0SDavid Greenman 		va += NBPG;
387d5e26ef0SDavid Greenman 	}
388d5e26ef0SDavid Greenman 
389d5e26ef0SDavid Greenman /*
390d5e26ef0SDavid Greenman  * flag the buffer as being bounced
391d5e26ef0SDavid Greenman  */
392d5e26ef0SDavid Greenman 	bp->b_flags |= B_BOUNCE;
393d5e26ef0SDavid Greenman /*
394d5e26ef0SDavid Greenman  * save the original buffer kva
395d5e26ef0SDavid Greenman  */
39626f9a767SRodney W. Grimes 	bp->b_savekva = bp->b_data;
397d5e26ef0SDavid Greenman /*
398d5e26ef0SDavid Greenman  * put our new kva into the buffer (offset by original offset)
399d5e26ef0SDavid Greenman  */
40026f9a767SRodney W. Grimes 	bp->b_data = (caddr_t) (((vm_offset_t) kva) |
401d5e26ef0SDavid Greenman 				((vm_offset_t) bp->b_savekva & (NBPG - 1)));
402ae92ea44SDavid Greenman #if 0
403ae92ea44SDavid Greenman 	printf("b_savekva: %x, newva: %x\n", bp->b_savekva, bp->b_data);
404ae92ea44SDavid Greenman #endif
405d5e26ef0SDavid Greenman 	return;
406d5e26ef0SDavid Greenman }
407d5e26ef0SDavid Greenman 
408d5e26ef0SDavid Greenman /*
409d5e26ef0SDavid Greenman  * hook into biodone to free bounce buffer
410d5e26ef0SDavid Greenman  */
411d5e26ef0SDavid Greenman void
412d5e26ef0SDavid Greenman vm_bounce_free(bp)
413d5e26ef0SDavid Greenman 	struct buf *bp;
414d5e26ef0SDavid Greenman {
415d5e26ef0SDavid Greenman 	int i;
416ae92ea44SDavid Greenman 	vm_offset_t origkva, bouncekva, bouncekvaend;
417d5e26ef0SDavid Greenman 
418d5e26ef0SDavid Greenman /*
419d5e26ef0SDavid Greenman  * if this isn't a bounced buffer, then just return
420d5e26ef0SDavid Greenman  */
421d5e26ef0SDavid Greenman 	if ((bp->b_flags & B_BOUNCE) == 0)
422d5e26ef0SDavid Greenman 		return;
423d5e26ef0SDavid Greenman 
424ae92ea44SDavid Greenman /*
425ae92ea44SDavid Greenman  *  This check is not necessary
426ae92ea44SDavid Greenman  *	if (bp->b_bufsize != bp->b_bcount) {
427ae92ea44SDavid Greenman  *		printf("vm_bounce_free: b_bufsize=%d, b_bcount=%d\n",
428ae92ea44SDavid Greenman  *			bp->b_bufsize, bp->b_bcount);
429ae92ea44SDavid Greenman  *	}
430ae92ea44SDavid Greenman  */
431ae92ea44SDavid Greenman 
432d5e26ef0SDavid Greenman 	origkva = (vm_offset_t) bp->b_savekva;
43326f9a767SRodney W. Grimes 	bouncekva = (vm_offset_t) bp->b_data;
434ae92ea44SDavid Greenman /*
435ae92ea44SDavid Greenman 	printf("free: %d ", bp->b_bufsize);
436ae92ea44SDavid Greenman */
437d5e26ef0SDavid Greenman 
438d5e26ef0SDavid Greenman /*
439d5e26ef0SDavid Greenman  * check every page in the kva space for b_addr
440d5e26ef0SDavid Greenman  */
441ae92ea44SDavid Greenman 	for (i = 0; i < bp->b_bufsize; ) {
442d5e26ef0SDavid Greenman 		vm_offset_t mybouncepa;
443d5e26ef0SDavid Greenman 		vm_offset_t copycount;
444d5e26ef0SDavid Greenman 
445bd7e5f99SJohn Dyson 		copycount = round_page(bouncekva + 1) - bouncekva;
446bd7e5f99SJohn Dyson 		mybouncepa = pmap_kextract(trunc_page(bouncekva));
447d5e26ef0SDavid Greenman 
448d5e26ef0SDavid Greenman /*
449d5e26ef0SDavid Greenman  * if this is a bounced pa, then process as one
450d5e26ef0SDavid Greenman  */
451bd7e5f99SJohn Dyson 		if ( mybouncepa != pmap_kextract( trunc_page( origkva))) {
452ae92ea44SDavid Greenman 			vm_offset_t tocopy = copycount;
453ae92ea44SDavid Greenman 			if (i + tocopy > bp->b_bufsize)
454ae92ea44SDavid Greenman 				tocopy = bp->b_bufsize - i;
455d5e26ef0SDavid Greenman /*
456d5e26ef0SDavid Greenman  * if this is a read, then copy from bounce buffer into original buffer
457d5e26ef0SDavid Greenman  */
458d5e26ef0SDavid Greenman 			if (bp->b_flags & B_READ)
459ae92ea44SDavid Greenman 				bcopy((caddr_t) bouncekva, (caddr_t) origkva, tocopy);
460d5e26ef0SDavid Greenman /*
461d5e26ef0SDavid Greenman  * free the bounce allocation
462d5e26ef0SDavid Greenman  */
463ae92ea44SDavid Greenman 
464ae92ea44SDavid Greenman /*
465ae92ea44SDavid Greenman 			printf("(kva: %x, pa: %x)", bouncekva, mybouncepa);
466ae92ea44SDavid Greenman */
467ae92ea44SDavid Greenman 			vm_bounce_page_free(mybouncepa, 1);
468d5e26ef0SDavid Greenman 		}
469d5e26ef0SDavid Greenman 
470d5e26ef0SDavid Greenman 		origkva += copycount;
471d5e26ef0SDavid Greenman 		bouncekva += copycount;
472ae92ea44SDavid Greenman 		i += copycount;
473d5e26ef0SDavid Greenman 	}
474d5e26ef0SDavid Greenman 
475d5e26ef0SDavid Greenman /*
476ae92ea44SDavid Greenman 	printf("\n");
477ae92ea44SDavid Greenman */
478ae92ea44SDavid Greenman /*
479d5e26ef0SDavid Greenman  * add the old kva into the "to free" list
480d5e26ef0SDavid Greenman  */
481ae92ea44SDavid Greenman 
482bd7e5f99SJohn Dyson 	bouncekva= trunc_page((vm_offset_t) bp->b_data);
483bd7e5f99SJohn Dyson 	bouncekvaend= round_page((vm_offset_t)bp->b_data + bp->b_bufsize);
484ae92ea44SDavid Greenman 
485ae92ea44SDavid Greenman /*
486ae92ea44SDavid Greenman 	printf("freeva: %d\n", (bouncekvaend - bouncekva) / NBPG);
487ae92ea44SDavid Greenman */
488ae92ea44SDavid Greenman 	vm_bounce_kva_free( bouncekva, (bouncekvaend - bouncekva), 0);
48926f9a767SRodney W. Grimes 	bp->b_data = bp->b_savekva;
490d5e26ef0SDavid Greenman 	bp->b_savekva = 0;
491d5e26ef0SDavid Greenman 	bp->b_flags &= ~B_BOUNCE;
492d5e26ef0SDavid Greenman 
493d5e26ef0SDavid Greenman 	return;
494d5e26ef0SDavid Greenman }
495d5e26ef0SDavid Greenman 
496ae92ea44SDavid Greenman 
4975b81b6b3SRodney W. Grimes /*
4981561d038SDavid Greenman  * init the bounce buffer system
4991561d038SDavid Greenman  */
5001561d038SDavid Greenman void
5011561d038SDavid Greenman vm_bounce_init()
5021561d038SDavid Greenman {
503ae92ea44SDavid Greenman 	int i;
5041561d038SDavid Greenman 
5051561d038SDavid Greenman 	kvasfreecnt = 0;
5061561d038SDavid Greenman 
5071561d038SDavid Greenman 	if (bouncepages == 0)
5081561d038SDavid Greenman 		return;
5091561d038SDavid Greenman 
5101561d038SDavid Greenman 	bounceallocarraysize = (bouncepages + BITS_IN_UNSIGNED - 1) / BITS_IN_UNSIGNED;
5111561d038SDavid Greenman 	bounceallocarray = malloc(bounceallocarraysize * sizeof(unsigned), M_TEMP, M_NOWAIT);
5121561d038SDavid Greenman 
5131561d038SDavid Greenman 	if (!bounceallocarray)
514edf8a815SDavid Greenman 		panic("Cannot allocate bounce resource array");
5151561d038SDavid Greenman 
516ae92ea44SDavid Greenman 	bouncepa = malloc(bouncepages * sizeof(vm_offset_t), M_TEMP, M_NOWAIT);
517ae92ea44SDavid Greenman 	if (!bouncepa)
518edf8a815SDavid Greenman 		panic("Cannot allocate physical memory array");
5191561d038SDavid Greenman 
5200d94caffSDavid Greenman 	for(i=0;i<bounceallocarraysize;i++) {
5210d94caffSDavid Greenman 		bounceallocarray[i] = 0xffffffff;
5220d94caffSDavid Greenman 	}
5230d94caffSDavid Greenman 
524ae92ea44SDavid Greenman 	for(i=0;i<bouncepages;i++) {
525ae92ea44SDavid Greenman 		vm_offset_t pa;
526ae92ea44SDavid Greenman 		if( (pa = pmap_kextract((vm_offset_t) bouncememory + i * NBPG)) >= SIXTEENMEG)
527ae92ea44SDavid Greenman 			panic("bounce memory out of range");
528ae92ea44SDavid Greenman 		if( pa == 0)
529ae92ea44SDavid Greenman 			panic("bounce memory not resident");
530ae92ea44SDavid Greenman 		bouncepa[i] = pa;
5310d94caffSDavid Greenman 		bounceallocarray[i/(8*sizeof(int))] &= ~(1<<(i%(8*sizeof(int))));
532ae92ea44SDavid Greenman 	}
5331561d038SDavid Greenman 	bouncefree = bouncepages;
5341561d038SDavid Greenman 
5351561d038SDavid Greenman }
5369aba88bfSDavid Greenman #endif /* BOUNCE_BUFFERS */
5376c146e28SBruce Evans 
538a4f7a4c9SDavid Greenman /*
539a4f7a4c9SDavid Greenman  * quick version of vm_fault
540a4f7a4c9SDavid Greenman  */
541dca51299SJohn Dyson void
542a4f7a4c9SDavid Greenman vm_fault_quick(v, prot)
5436c146e28SBruce Evans 	caddr_t v;
544a4f7a4c9SDavid Greenman 	int prot;
545a4f7a4c9SDavid Greenman {
5461d1b971bSDavid Greenman 	if (prot & VM_PROT_WRITE)
5476c146e28SBruce Evans 		subyte(v, fubyte(v));
548a4f7a4c9SDavid Greenman 	else
5496c146e28SBruce Evans 		fubyte(v);
550a4f7a4c9SDavid Greenman }
551a4f7a4c9SDavid Greenman 
5521561d038SDavid Greenman /*
5535b81b6b3SRodney W. Grimes  * Finish a fork operation, with process p2 nearly set up.
5545b81b6b3SRodney W. Grimes  * Copy and update the kernel stack and pcb, making the child
5555b81b6b3SRodney W. Grimes  * ready to run, and marking it so that it can return differently
5565b81b6b3SRodney W. Grimes  * than the parent.  Returns 1 in the child process, 0 in the parent.
5575b81b6b3SRodney W. Grimes  * We currently double-map the user area so that the stack is at the same
5585b81b6b3SRodney W. Grimes  * address in each process; in the future we will probably relocate
5595b81b6b3SRodney W. Grimes  * the frame pointers on the stack after copying.
5605b81b6b3SRodney W. Grimes  */
561381fe1aaSGarrett Wollman int
5625b81b6b3SRodney W. Grimes cpu_fork(p1, p2)
5635b81b6b3SRodney W. Grimes 	register struct proc *p1, *p2;
5645b81b6b3SRodney W. Grimes {
565dca51299SJohn Dyson 	register struct user *up = p2->p_addr;
566dca51299SJohn Dyson 	int offset;
5675b81b6b3SRodney W. Grimes 
5685b81b6b3SRodney W. Grimes 	/*
5695b81b6b3SRodney W. Grimes 	 * Copy pcb and stack from proc p1 to p2.
5705b81b6b3SRodney W. Grimes 	 * We do this as cheaply as possible, copying only the active
5715b81b6b3SRodney W. Grimes 	 * part of the stack.  The stack and pcb need to agree;
5725b81b6b3SRodney W. Grimes 	 * this is tricky, as the final pcb is constructed by savectx,
5735b81b6b3SRodney W. Grimes 	 * but its frame isn't yet on the stack when the stack is copied.
574dca51299SJohn Dyson 	 * swtch compensates for this when the child eventually runs.
5755b81b6b3SRodney W. Grimes 	 * This should be done differently, with a single call
5765b81b6b3SRodney W. Grimes 	 * that copies and updates the pcb+stack,
5775b81b6b3SRodney W. Grimes 	 * replacing the bcopy and savectx.
5785b81b6b3SRodney W. Grimes 	 */
579dca51299SJohn Dyson 	p2->p_addr->u_pcb = p1->p_addr->u_pcb;
580dca51299SJohn Dyson 	offset = mvesp() - (int)kstack;
5815b81b6b3SRodney W. Grimes 	bcopy((caddr_t)kstack + offset, (caddr_t)p2->p_addr + offset,
5825b81b6b3SRodney W. Grimes 	    (unsigned) ctob(UPAGES) - offset);
58326f9a767SRodney W. Grimes 	p2->p_md.md_regs = p1->p_md.md_regs;
5845b81b6b3SRodney W. Grimes 
585dca51299SJohn Dyson 	pmap_activate(&p2->p_vmspace->vm_pmap, &up->u_pcb);
5865b81b6b3SRodney W. Grimes 
5875b81b6b3SRodney W. Grimes 	/*
588dca51299SJohn Dyson 	 * Return (0) in parent, (1) in child.
5895b81b6b3SRodney W. Grimes 	 */
590dca51299SJohn Dyson 	return (savectx(&up->u_pcb));
5915b81b6b3SRodney W. Grimes }
5925b81b6b3SRodney W. Grimes 
5937c2b54e8SNate Williams void
5945b81b6b3SRodney W. Grimes cpu_exit(p)
5955b81b6b3SRodney W. Grimes 	register struct proc *p;
5965b81b6b3SRodney W. Grimes {
5975b81b6b3SRodney W. Grimes 
598960173b9SRodney W. Grimes #if NNPX > 0
5995b81b6b3SRodney W. Grimes 	npxexit(p);
600960173b9SRodney W. Grimes #endif	/* NNPX */
6011a051896SBruce Evans 	cnt.v_swtch++;
6021a051896SBruce Evans 	cpu_switch(p);
6037c2b54e8SNate Williams 	panic("cpu_exit");
6045b81b6b3SRodney W. Grimes }
6055b81b6b3SRodney W. Grimes 
606381fe1aaSGarrett Wollman void
607bd7e5f99SJohn Dyson cpu_wait(p)
608bd7e5f99SJohn Dyson 	struct proc *p;
609bd7e5f99SJohn Dyson {
6105b81b6b3SRodney W. Grimes 	/* drop per-process resources */
611bd7e5f99SJohn Dyson 	pmap_qremove((vm_offset_t) p->p_addr, UPAGES);
6120d94caffSDavid Greenman 	kmem_free(u_map, (vm_offset_t)p->p_addr, ctob(UPAGES));
6137f8cb368SDavid Greenman 	vmspace_free(p->p_vmspace);
6145b81b6b3SRodney W. Grimes }
6155b81b6b3SRodney W. Grimes 
6165b81b6b3SRodney W. Grimes /*
61726f9a767SRodney W. Grimes  * Dump the machine specific header information at the start of a core dump.
61826f9a767SRodney W. Grimes  */
61926f9a767SRodney W. Grimes int
62026f9a767SRodney W. Grimes cpu_coredump(p, vp, cred)
62126f9a767SRodney W. Grimes 	struct proc *p;
62226f9a767SRodney W. Grimes 	struct vnode *vp;
62326f9a767SRodney W. Grimes 	struct ucred *cred;
62426f9a767SRodney W. Grimes {
62526f9a767SRodney W. Grimes 
62626f9a767SRodney W. Grimes 	return (vn_rdwr(UIO_WRITE, vp, (caddr_t) p->p_addr, ctob(UPAGES),
62726f9a767SRodney W. Grimes 	    (off_t)0, UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *)NULL,
62826f9a767SRodney W. Grimes 	    p));
62926f9a767SRodney W. Grimes }
63026f9a767SRodney W. Grimes 
63187b6de2bSPoul-Henning Kamp #ifdef notyet
63287b91157SPoul-Henning Kamp static void
6335b81b6b3SRodney W. Grimes setredzone(pte, vaddr)
6345b81b6b3SRodney W. Grimes 	u_short *pte;
6355b81b6b3SRodney W. Grimes 	caddr_t vaddr;
6365b81b6b3SRodney W. Grimes {
6375b81b6b3SRodney W. Grimes /* eventually do this by setting up an expand-down stack segment
6385b81b6b3SRodney W. Grimes    for ss0: selector, allowing stack access down to top of u.
6395b81b6b3SRodney W. Grimes    this means though that protection violations need to be handled
6405b81b6b3SRodney W. Grimes    thru a double fault exception that must do an integral task
6415b81b6b3SRodney W. Grimes    switch to a known good context, within which a dump can be
6425b81b6b3SRodney W. Grimes    taken. a sensible scheme might be to save the initial context
6435b81b6b3SRodney W. Grimes    used by sched (that has physical memory mapped 1:1 at bottom)
6445b81b6b3SRodney W. Grimes    and take the dump while still in mapped mode */
6455b81b6b3SRodney W. Grimes }
64687b6de2bSPoul-Henning Kamp #endif
64726f9a767SRodney W. Grimes 
64826f9a767SRodney W. Grimes /*
6495b81b6b3SRodney W. Grimes  * Convert kernel VA to physical address
6505b81b6b3SRodney W. Grimes  */
651aaf08d94SGarrett Wollman u_long
6527f8cb368SDavid Greenman kvtop(void *addr)
6535b81b6b3SRodney W. Grimes {
6545b81b6b3SRodney W. Grimes 	vm_offset_t va;
6555b81b6b3SRodney W. Grimes 
656ed7fcbd0SDavid Greenman 	va = pmap_kextract((vm_offset_t)addr);
6575b81b6b3SRodney W. Grimes 	if (va == 0)
6585b81b6b3SRodney W. Grimes 		panic("kvtop: zero page frame");
6597f8cb368SDavid Greenman 	return((int)va);
6605b81b6b3SRodney W. Grimes }
6615b81b6b3SRodney W. Grimes 
6625b81b6b3SRodney W. Grimes /*
663ac322158SDavid Greenman  * Map an IO request into kernel virtual address space.
6645b81b6b3SRodney W. Grimes  *
665ac322158SDavid Greenman  * All requests are (re)mapped into kernel VA space.
666ac322158SDavid Greenman  * Notice that we use b_bufsize for the size of the buffer
667ac322158SDavid Greenman  * to be mapped.  b_bcount might be modified by the driver.
6685b81b6b3SRodney W. Grimes  */
669381fe1aaSGarrett Wollman void
6705b81b6b3SRodney W. Grimes vmapbuf(bp)
6715b81b6b3SRodney W. Grimes 	register struct buf *bp;
6725b81b6b3SRodney W. Grimes {
6735b81b6b3SRodney W. Grimes 	register int npf;
6745b81b6b3SRodney W. Grimes 	register caddr_t addr;
6755b81b6b3SRodney W. Grimes 	int off;
6765b81b6b3SRodney W. Grimes 	vm_offset_t kva;
67707658526SPoul-Henning Kamp 	vm_offset_t pa;
6785b81b6b3SRodney W. Grimes 
67926f9a767SRodney W. Grimes 	if ((bp->b_flags & B_PHYS) == 0)
6805b81b6b3SRodney W. Grimes 		panic("vmapbuf");
68126f9a767SRodney W. Grimes 
682ae92ea44SDavid Greenman 	/*
683ae92ea44SDavid Greenman 	 * this is the kva that is to be used for
684ae92ea44SDavid Greenman 	 * the temporary kernel mapping
685ae92ea44SDavid Greenman 	 */
686ae92ea44SDavid Greenman 	kva = (vm_offset_t) bp->b_saveaddr;
687ae92ea44SDavid Greenman 
68826f9a767SRodney W. Grimes 	for (addr = (caddr_t)trunc_page(bp->b_data);
68926f9a767SRodney W. Grimes 		addr < bp->b_data + bp->b_bufsize;
69026f9a767SRodney W. Grimes 		addr += PAGE_SIZE) {
69126f9a767SRodney W. Grimes 
69226f9a767SRodney W. Grimes /*
69326f9a767SRodney W. Grimes  * do the vm_fault if needed, do the copy-on-write thing when
69426f9a767SRodney W. Grimes  * reading stuff off device into memory.
69526f9a767SRodney W. Grimes  */
69626f9a767SRodney W. Grimes 		vm_fault_quick(addr,
69726f9a767SRodney W. Grimes 			(bp->b_flags&B_READ)?(VM_PROT_READ|VM_PROT_WRITE):VM_PROT_READ);
698c564966bSDavid Greenman 		pa = pmap_kextract((vm_offset_t) addr);
699641a27e6SJohn Dyson 		if (pa == 0)
700641a27e6SJohn Dyson 			panic("vmapbuf: page not present");
70126f9a767SRodney W. Grimes /*
70226f9a767SRodney W. Grimes  * hold the data page
70326f9a767SRodney W. Grimes  */
704641a27e6SJohn Dyson #ifdef DIAGNOSTIC
705641a27e6SJohn Dyson 		if( VM_PAGE_TO_PHYS(PHYS_TO_VM_PAGE(pa)) != pa)
706641a27e6SJohn Dyson 			panic("vmapbuf: confused PHYS_TO_VM_PAGE mapping");
707641a27e6SJohn Dyson #endif
70826f9a767SRodney W. Grimes 		vm_page_hold(PHYS_TO_VM_PAGE(pa));
70926f9a767SRodney W. Grimes 	}
71026f9a767SRodney W. Grimes 
711ae92ea44SDavid Greenman 	addr = bp->b_saveaddr = bp->b_data;
7125b81b6b3SRodney W. Grimes 	off = (int)addr & PGOFSET;
713ac322158SDavid Greenman 	npf = btoc(round_page(bp->b_bufsize + off));
714ae92ea44SDavid Greenman 	bp->b_data = (caddr_t) (kva + off);
7155b81b6b3SRodney W. Grimes 	while (npf--) {
716c564966bSDavid Greenman 		pa = pmap_kextract((vm_offset_t)addr);
7175b81b6b3SRodney W. Grimes 		if (pa == 0)
7185b81b6b3SRodney W. Grimes 			panic("vmapbuf: null page frame");
7196b4ac811SDavid Greenman 		pmap_kenter(kva, trunc_page(pa));
7205b81b6b3SRodney W. Grimes 		addr += PAGE_SIZE;
7215b81b6b3SRodney W. Grimes 		kva += PAGE_SIZE;
7225b81b6b3SRodney W. Grimes 	}
7235b81b6b3SRodney W. Grimes }
7245b81b6b3SRodney W. Grimes 
7255b81b6b3SRodney W. Grimes /*
7265b81b6b3SRodney W. Grimes  * Free the io map PTEs associated with this IO operation.
7275b81b6b3SRodney W. Grimes  * We also invalidate the TLB entries and restore the original b_addr.
7285b81b6b3SRodney W. Grimes  */
729381fe1aaSGarrett Wollman void
7305b81b6b3SRodney W. Grimes vunmapbuf(bp)
7315b81b6b3SRodney W. Grimes 	register struct buf *bp;
7325b81b6b3SRodney W. Grimes {
733c564966bSDavid Greenman 	register caddr_t addr;
73407658526SPoul-Henning Kamp 	vm_offset_t pa;
7355b81b6b3SRodney W. Grimes 
7365b81b6b3SRodney W. Grimes 	if ((bp->b_flags & B_PHYS) == 0)
7375b81b6b3SRodney W. Grimes 		panic("vunmapbuf");
738c564966bSDavid Greenman 
739c564966bSDavid Greenman 	for (addr = (caddr_t)trunc_page((vm_offset_t) bp->b_data);
740c564966bSDavid Greenman 		addr < bp->b_data + bp->b_bufsize;
741c564966bSDavid Greenman 		addr += NBPG)
742c564966bSDavid Greenman 		pmap_kremove((vm_offset_t) addr);
743c564966bSDavid Greenman 
744ae92ea44SDavid Greenman 	bp->b_data = bp->b_saveaddr;
7455b81b6b3SRodney W. Grimes 	bp->b_saveaddr = NULL;
74626f9a767SRodney W. Grimes 
74726f9a767SRodney W. Grimes /*
74826f9a767SRodney W. Grimes  * unhold the pde, and data pages
74926f9a767SRodney W. Grimes  */
750c564966bSDavid Greenman 	for (addr = (caddr_t)trunc_page((vm_offset_t) bp->b_data);
75126f9a767SRodney W. Grimes 		addr < bp->b_data + bp->b_bufsize;
75226f9a767SRodney W. Grimes 		addr += NBPG) {
75326f9a767SRodney W. Grimes 	/*
75426f9a767SRodney W. Grimes 	 * release the data page
75526f9a767SRodney W. Grimes 	 */
756c564966bSDavid Greenman 		pa = pmap_kextract((vm_offset_t) addr);
75726f9a767SRodney W. Grimes 		vm_page_unhold(PHYS_TO_VM_PAGE(pa));
75826f9a767SRodney W. Grimes 	}
7595b81b6b3SRodney W. Grimes }
7605b81b6b3SRodney W. Grimes 
7615b81b6b3SRodney W. Grimes /*
7625b81b6b3SRodney W. Grimes  * Force reset the processor by invalidating the entire address space!
7635b81b6b3SRodney W. Grimes  */
7647f8cb368SDavid Greenman void
7655b81b6b3SRodney W. Grimes cpu_reset() {
7665b81b6b3SRodney W. Grimes 
7672320728fSRodney W. Grimes 	/*
7682320728fSRodney W. Grimes 	 * Attempt to do a CPU reset via the keyboard controller,
7692320728fSRodney W. Grimes 	 * do not turn of the GateA20, as any machine that fails
7702320728fSRodney W. Grimes 	 * to do the reset here would then end up in no man's land.
7712320728fSRodney W. Grimes 	 */
7725eb46edfSDavid Greenman 
7735eb46edfSDavid Greenman #ifndef BROKEN_KEYBOARD_RESET
7742320728fSRodney W. Grimes 	outb(IO_KBD + 4, 0xFE);
7752320728fSRodney W. Grimes 	DELAY(500000);	/* wait 0.5 sec to see if that did it */
7762320728fSRodney W. Grimes 	printf("Keyboard reset did not work, attempting CPU shutdown\n");
7772320728fSRodney W. Grimes 	DELAY(1000000);	/* wait 1 sec for printf to complete */
7785eb46edfSDavid Greenman #endif
7792320728fSRodney W. Grimes 
7805b81b6b3SRodney W. Grimes 	/* force a shutdown by unmapping entire address space ! */
7815b81b6b3SRodney W. Grimes 	bzero((caddr_t) PTD, NBPG);
7825b81b6b3SRodney W. Grimes 
7835b81b6b3SRodney W. Grimes 	/* "good night, sweet prince .... <THUNK!>" */
7842c7a40c7SDavid Greenman 	pmap_update();
7855b81b6b3SRodney W. Grimes 	/* NOTREACHED */
7867f8cb368SDavid Greenman 	while(1);
7875b81b6b3SRodney W. Grimes }
788b9d60b3fSDavid Greenman 
789b9d60b3fSDavid Greenman /*
790b9d60b3fSDavid Greenman  * Grow the user stack to allow for 'sp'. This version grows the stack in
79129360eb0SDavid Greenman  *	chunks of SGROWSIZ.
792b9d60b3fSDavid Greenman  */
793b9d60b3fSDavid Greenman int
794b9d60b3fSDavid Greenman grow(p, sp)
795b9d60b3fSDavid Greenman 	struct proc *p;
79626f9a767SRodney W. Grimes 	u_int sp;
797b9d60b3fSDavid Greenman {
798b9d60b3fSDavid Greenman 	unsigned int nss;
799b9d60b3fSDavid Greenman 	caddr_t v;
800b9d60b3fSDavid Greenman 	struct vmspace *vm = p->p_vmspace;
801b9d60b3fSDavid Greenman 
802b9d60b3fSDavid Greenman 	if ((caddr_t)sp <= vm->vm_maxsaddr || (unsigned)sp >= (unsigned)USRSTACK)
803b9d60b3fSDavid Greenman 	    return (1);
804b9d60b3fSDavid Greenman 
805b9d60b3fSDavid Greenman 	nss = roundup(USRSTACK - (unsigned)sp, PAGE_SIZE);
806b9d60b3fSDavid Greenman 
807b9d60b3fSDavid Greenman 	if (nss > p->p_rlimit[RLIMIT_STACK].rlim_cur)
808b9d60b3fSDavid Greenman 		return (0);
809b9d60b3fSDavid Greenman 
810b9d60b3fSDavid Greenman 	if (vm->vm_ssize && roundup(vm->vm_ssize << PAGE_SHIFT,
81129360eb0SDavid Greenman 	    SGROWSIZ) < nss) {
812b9d60b3fSDavid Greenman 		int grow_amount;
813b9d60b3fSDavid Greenman 		/*
814b9d60b3fSDavid Greenman 		 * If necessary, grow the VM that the stack occupies
815b9d60b3fSDavid Greenman 		 * to allow for the rlimit. This allows us to not have
816b9d60b3fSDavid Greenman 		 * to allocate all of the VM up-front in execve (which
817b9d60b3fSDavid Greenman 		 * is expensive).
818b9d60b3fSDavid Greenman 		 * Grow the VM by the amount requested rounded up to
81929360eb0SDavid Greenman 		 * the nearest SGROWSIZ to provide for some hysteresis.
820b9d60b3fSDavid Greenman 		 */
82129360eb0SDavid Greenman 		grow_amount = roundup((nss - (vm->vm_ssize << PAGE_SHIFT)), SGROWSIZ);
822b9d60b3fSDavid Greenman 		v = (char *)USRSTACK - roundup(vm->vm_ssize << PAGE_SHIFT,
82329360eb0SDavid Greenman 		    SGROWSIZ) - grow_amount;
824b9d60b3fSDavid Greenman 		/*
82529360eb0SDavid Greenman 		 * If there isn't enough room to extend by SGROWSIZ, then
826b9d60b3fSDavid Greenman 		 * just extend to the maximum size
827b9d60b3fSDavid Greenman 		 */
828b9d60b3fSDavid Greenman 		if (v < vm->vm_maxsaddr) {
829b9d60b3fSDavid Greenman 			v = vm->vm_maxsaddr;
830b9d60b3fSDavid Greenman 			grow_amount = MAXSSIZ - (vm->vm_ssize << PAGE_SHIFT);
831b9d60b3fSDavid Greenman 		}
83268940ac1SDavid Greenman 		if ((grow_amount == 0) || (vm_map_find(&vm->vm_map, NULL, 0, (vm_offset_t *)&v,
833bd7e5f99SJohn Dyson 		    grow_amount, FALSE, VM_PROT_ALL, VM_PROT_ALL, 0) != KERN_SUCCESS)) {
834b9d60b3fSDavid Greenman 			return (0);
835b9d60b3fSDavid Greenman 		}
836b9d60b3fSDavid Greenman 		vm->vm_ssize += grow_amount >> PAGE_SHIFT;
837b9d60b3fSDavid Greenman 	}
838b9d60b3fSDavid Greenman 
839b9d60b3fSDavid Greenman 	return (1);
840b9d60b3fSDavid Greenman }
841a29b63cbSJohn Dyson 
842a29b63cbSJohn Dyson /*
843a29b63cbSJohn Dyson  * prototype routine to implement the pre-zeroed page mechanism
844a29b63cbSJohn Dyson  * this routine is called from the idle loop.
845a29b63cbSJohn Dyson  */
846a29b63cbSJohn Dyson int
847a29b63cbSJohn Dyson vm_page_zero_idle() {
848a29b63cbSJohn Dyson 	vm_page_t m;
849a29b63cbSJohn Dyson 	if ((cnt.v_free_count > cnt.v_interrupt_free_min) &&
850a29b63cbSJohn Dyson 		(m = vm_page_queue_free.tqh_first)) {
851a29b63cbSJohn Dyson 		TAILQ_REMOVE(&vm_page_queue_free, m, pageq);
852a29b63cbSJohn Dyson 		enable_intr();
853a29b63cbSJohn Dyson 		pmap_zero_page(VM_PAGE_TO_PHYS(m));
854a29b63cbSJohn Dyson 		disable_intr();
85523e56ee8SDavid Greenman 		TAILQ_INSERT_HEAD(&vm_page_queue_zero, m, pageq);
856bd7e5f99SJohn Dyson 		m->queue = PQ_ZERO;
857a316d390SJohn Dyson 		++vm_page_zero_count;
858a29b63cbSJohn Dyson 		return 1;
859a29b63cbSJohn Dyson 	}
860a29b63cbSJohn Dyson 	return 0;
861a29b63cbSJohn Dyson }
862