xref: /freebsd/sys/i386/i386/vm_machdep.c (revision c564966bf099002b949a16f8da162e1028b21215)
15b81b6b3SRodney W. Grimes /*-
25b81b6b3SRodney W. Grimes  * Copyright (c) 1982, 1986 The Regents of the University of California.
35b81b6b3SRodney W. Grimes  * Copyright (c) 1989, 1990 William Jolitz
41561d038SDavid Greenman  * Copyright (c) 1994 John Dyson
55b81b6b3SRodney W. Grimes  * All rights reserved.
65b81b6b3SRodney W. Grimes  *
75b81b6b3SRodney W. Grimes  * This code is derived from software contributed to Berkeley by
85b81b6b3SRodney W. Grimes  * the Systems Programming Group of the University of Utah Computer
95b81b6b3SRodney W. Grimes  * Science Department, and William Jolitz.
105b81b6b3SRodney W. Grimes  *
115b81b6b3SRodney W. Grimes  * Redistribution and use in source and binary forms, with or without
125b81b6b3SRodney W. Grimes  * modification, are permitted provided that the following conditions
135b81b6b3SRodney W. Grimes  * are met:
145b81b6b3SRodney W. Grimes  * 1. Redistributions of source code must retain the above copyright
155b81b6b3SRodney W. Grimes  *    notice, this list of conditions and the following disclaimer.
165b81b6b3SRodney W. Grimes  * 2. Redistributions in binary form must reproduce the above copyright
175b81b6b3SRodney W. Grimes  *    notice, this list of conditions and the following disclaimer in the
185b81b6b3SRodney W. Grimes  *    documentation and/or other materials provided with the distribution.
195b81b6b3SRodney W. Grimes  * 3. All advertising materials mentioning features or use of this software
205b81b6b3SRodney W. Grimes  *    must display the following acknowledgement:
215b81b6b3SRodney W. Grimes  *	This product includes software developed by the University of
225b81b6b3SRodney W. Grimes  *	California, Berkeley and its contributors.
235b81b6b3SRodney W. Grimes  * 4. Neither the name of the University nor the names of its contributors
245b81b6b3SRodney W. Grimes  *    may be used to endorse or promote products derived from this software
255b81b6b3SRodney W. Grimes  *    without specific prior written permission.
265b81b6b3SRodney W. Grimes  *
275b81b6b3SRodney W. Grimes  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
285b81b6b3SRodney W. Grimes  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
295b81b6b3SRodney W. Grimes  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
305b81b6b3SRodney W. Grimes  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
315b81b6b3SRodney W. Grimes  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
325b81b6b3SRodney W. Grimes  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
335b81b6b3SRodney W. Grimes  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
345b81b6b3SRodney W. Grimes  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
355b81b6b3SRodney W. Grimes  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
365b81b6b3SRodney W. Grimes  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
375b81b6b3SRodney W. Grimes  * SUCH DAMAGE.
385b81b6b3SRodney W. Grimes  *
39960173b9SRodney W. Grimes  *	from: @(#)vm_machdep.c	7.3 (Berkeley) 5/13/91
405b81b6b3SRodney W. Grimes  *	Utah $Hdr: vm_machdep.c 1.16.1.1 89/06/23$
41c564966bSDavid Greenman  *	$Id: vm_machdep.c,v 1.24 1994/08/06 10:25:37 davidg Exp $
425b81b6b3SRodney W. Grimes  */
435b81b6b3SRodney W. Grimes 
44960173b9SRodney W. Grimes #include "npx.h"
4526f9a767SRodney W. Grimes #include <sys/param.h>
4626f9a767SRodney W. Grimes #include <sys/systm.h>
4726f9a767SRodney W. Grimes #include <sys/proc.h>
4826f9a767SRodney W. Grimes #include <sys/malloc.h>
4926f9a767SRodney W. Grimes #include <sys/buf.h>
5026f9a767SRodney W. Grimes #include <sys/vnode.h>
5126f9a767SRodney W. Grimes #include <sys/user.h>
525b81b6b3SRodney W. Grimes 
5326f9a767SRodney W. Grimes #include <machine/cpu.h>
545b81b6b3SRodney W. Grimes 
5526f9a767SRodney W. Grimes #include <vm/vm.h>
5626f9a767SRodney W. Grimes #include <vm/vm_kern.h>
575b81b6b3SRodney W. Grimes 
58ae92ea44SDavid Greenman #ifndef NOBOUNCE
59ae92ea44SDavid Greenman vm_map_t	io_map;
60ae92ea44SDavid Greenman volatile int	kvasfreecnt;
61ae92ea44SDavid Greenman 
621561d038SDavid Greenman 
63d5e26ef0SDavid Greenman caddr_t		bouncememory;
64ed7fcbd0SDavid Greenman int		bouncepages, bpwait;
65ae92ea44SDavid Greenman vm_offset_t	*bouncepa;
66d5e26ef0SDavid Greenman int		bmwait, bmfreeing;
67d5e26ef0SDavid Greenman 
68ed7fcbd0SDavid Greenman #define BITS_IN_UNSIGNED (8*sizeof(unsigned))
69d5e26ef0SDavid Greenman int		bounceallocarraysize;
70d5e26ef0SDavid Greenman unsigned	*bounceallocarray;
71d5e26ef0SDavid Greenman int		bouncefree;
72d5e26ef0SDavid Greenman 
73d5e26ef0SDavid Greenman #define SIXTEENMEG (4096*4096)
74ae92ea44SDavid Greenman #define MAXBKVA 1024
7526f9a767SRodney W. Grimes int		maxbkva = MAXBKVA*NBPG;
76d5e26ef0SDavid Greenman 
77d5e26ef0SDavid Greenman /* special list that can be used at interrupt time for eventual kva free */
78d5e26ef0SDavid Greenman struct kvasfree {
79d5e26ef0SDavid Greenman 	vm_offset_t addr;
80d5e26ef0SDavid Greenman 	vm_offset_t size;
81d5e26ef0SDavid Greenman } kvaf[MAXBKVA];
82d5e26ef0SDavid Greenman 
83d5e26ef0SDavid Greenman 
841561d038SDavid Greenman vm_offset_t vm_bounce_kva();
85d5e26ef0SDavid Greenman /*
86d5e26ef0SDavid Greenman  * get bounce buffer pages (count physically contiguous)
87d5e26ef0SDavid Greenman  * (only 1 inplemented now)
88d5e26ef0SDavid Greenman  */
89d5e26ef0SDavid Greenman vm_offset_t
90d5e26ef0SDavid Greenman vm_bounce_page_find(count)
91d5e26ef0SDavid Greenman 	int count;
92d5e26ef0SDavid Greenman {
93d5e26ef0SDavid Greenman 	int bit;
94d5e26ef0SDavid Greenman 	int s,i;
95d5e26ef0SDavid Greenman 
96d5e26ef0SDavid Greenman 	if (count != 1)
97d5e26ef0SDavid Greenman 		panic("vm_bounce_page_find -- no support for > 1 page yet!!!");
98d5e26ef0SDavid Greenman 
99d5e26ef0SDavid Greenman 	s = splbio();
100d5e26ef0SDavid Greenman retry:
101d5e26ef0SDavid Greenman 	for (i = 0; i < bounceallocarraysize; i++) {
102d5e26ef0SDavid Greenman 		if (bounceallocarray[i] != 0xffffffff) {
103d5e26ef0SDavid Greenman 			if (bit = ffs(~bounceallocarray[i])) {
104d5e26ef0SDavid Greenman 				bounceallocarray[i] |= 1 << (bit - 1) ;
105d5e26ef0SDavid Greenman 				bouncefree -= count;
106d5e26ef0SDavid Greenman 				splx(s);
107ae92ea44SDavid Greenman 				return bouncepa[(i * BITS_IN_UNSIGNED + (bit - 1))];
108d5e26ef0SDavid Greenman 			}
109d5e26ef0SDavid Greenman 		}
110d5e26ef0SDavid Greenman 	}
111ed7fcbd0SDavid Greenman 	bpwait = 1;
112d5e26ef0SDavid Greenman 	tsleep((caddr_t) &bounceallocarray, PRIBIO, "bncwai", 0);
113d5e26ef0SDavid Greenman 	goto retry;
114d5e26ef0SDavid Greenman }
115d5e26ef0SDavid Greenman 
1161561d038SDavid Greenman void
1171561d038SDavid Greenman vm_bounce_kva_free(addr, size, now)
1181561d038SDavid Greenman 	vm_offset_t addr;
1191561d038SDavid Greenman 	vm_offset_t size;
1201561d038SDavid Greenman 	int now;
1211561d038SDavid Greenman {
1221561d038SDavid Greenman 	int s = splbio();
1231561d038SDavid Greenman 	kvaf[kvasfreecnt].addr = addr;
124ae92ea44SDavid Greenman 	kvaf[kvasfreecnt].size = size;
125ae92ea44SDavid Greenman 	++kvasfreecnt;
1260e195446SDavid Greenman 	if( now) {
1270e195446SDavid Greenman 		/*
1280e195446SDavid Greenman 		 * this will do wakeups
1290e195446SDavid Greenman 		 */
1301561d038SDavid Greenman 		vm_bounce_kva(0,0);
1310e195446SDavid Greenman 	} else {
1320e195446SDavid Greenman 		if (bmwait) {
1330e195446SDavid Greenman 		/*
1340e195446SDavid Greenman 		 * if anyone is waiting on the bounce-map, then wakeup
1350e195446SDavid Greenman 		 */
1361561d038SDavid Greenman 			wakeup((caddr_t) io_map);
1370e195446SDavid Greenman 			bmwait = 0;
1380e195446SDavid Greenman 		}
1390e195446SDavid Greenman 	}
1401561d038SDavid Greenman 	splx(s);
1411561d038SDavid Greenman }
1421561d038SDavid Greenman 
143d5e26ef0SDavid Greenman /*
144d5e26ef0SDavid Greenman  * free count bounce buffer pages
145d5e26ef0SDavid Greenman  */
146d5e26ef0SDavid Greenman void
147d5e26ef0SDavid Greenman vm_bounce_page_free(pa, count)
148d5e26ef0SDavid Greenman 	vm_offset_t pa;
149d5e26ef0SDavid Greenman 	int count;
150d5e26ef0SDavid Greenman {
151d5e26ef0SDavid Greenman 	int allocindex;
152d5e26ef0SDavid Greenman 	int index;
153d5e26ef0SDavid Greenman 	int bit;
154d5e26ef0SDavid Greenman 
155d5e26ef0SDavid Greenman 	if (count != 1)
156d5e26ef0SDavid Greenman 		panic("vm_bounce_page_free -- no support for > 1 page yet!!!\n");
157d5e26ef0SDavid Greenman 
158ae92ea44SDavid Greenman 	for(index=0;index<bouncepages;index++) {
159ae92ea44SDavid Greenman 		if( pa == bouncepa[index])
160ae92ea44SDavid Greenman 			break;
161ae92ea44SDavid Greenman 	}
162d5e26ef0SDavid Greenman 
163ae92ea44SDavid Greenman 	if( index == bouncepages)
164ae92ea44SDavid Greenman 		panic("vm_bounce_page_free: invalid bounce buffer");
165d5e26ef0SDavid Greenman 
166ed7fcbd0SDavid Greenman 	allocindex = index / BITS_IN_UNSIGNED;
167ed7fcbd0SDavid Greenman 	bit = index % BITS_IN_UNSIGNED;
168d5e26ef0SDavid Greenman 
169d5e26ef0SDavid Greenman 	bounceallocarray[allocindex] &= ~(1 << bit);
170d5e26ef0SDavid Greenman 
171d5e26ef0SDavid Greenman 	bouncefree += count;
172ed7fcbd0SDavid Greenman 	if (bpwait) {
173ed7fcbd0SDavid Greenman 		bpwait = 0;
174d5e26ef0SDavid Greenman 		wakeup((caddr_t) &bounceallocarray);
175d5e26ef0SDavid Greenman 	}
176ed7fcbd0SDavid Greenman }
177d5e26ef0SDavid Greenman 
178d5e26ef0SDavid Greenman /*
179d5e26ef0SDavid Greenman  * allocate count bounce buffer kva pages
180d5e26ef0SDavid Greenman  */
181d5e26ef0SDavid Greenman vm_offset_t
182ae92ea44SDavid Greenman vm_bounce_kva(size, waitok)
183ae92ea44SDavid Greenman 	int size;
1841561d038SDavid Greenman 	int waitok;
185d5e26ef0SDavid Greenman {
186d5e26ef0SDavid Greenman 	int i;
187d5e26ef0SDavid Greenman 	int startfree;
1886b4ac811SDavid Greenman 	vm_offset_t kva = 0;
189c564966bSDavid Greenman 	vm_offset_t off;
190d5e26ef0SDavid Greenman 	int s = splbio();
191d5e26ef0SDavid Greenman more:
192ae92ea44SDavid Greenman 	if (!bmfreeing && kvasfreecnt) {
193d5e26ef0SDavid Greenman 		bmfreeing = 1;
194ae92ea44SDavid Greenman 		for (i = 0; i < kvasfreecnt; i++) {
195c564966bSDavid Greenman 			for(off=0;off<kvaf[i].size;off+=NBPG) {
196c564966bSDavid Greenman 				pmap_kremove( kvaf[i].addr + off);
197c564966bSDavid Greenman 			}
1981561d038SDavid Greenman 			kmem_free_wakeup(io_map, kvaf[i].addr,
199d5e26ef0SDavid Greenman 				kvaf[i].size);
200d5e26ef0SDavid Greenman 		}
201d5e26ef0SDavid Greenman 		kvasfreecnt = 0;
202d5e26ef0SDavid Greenman 		bmfreeing = 0;
203ae92ea44SDavid Greenman 		if( bmwait) {
204ae92ea44SDavid Greenman 			bmwait = 0;
205ae92ea44SDavid Greenman 			wakeup( (caddr_t) io_map);
206ae92ea44SDavid Greenman 		}
207d5e26ef0SDavid Greenman 	}
208d5e26ef0SDavid Greenman 
2091561d038SDavid Greenman 	if( size == 0) {
2101561d038SDavid Greenman 		splx(s);
2111561d038SDavid Greenman 		return NULL;
2121561d038SDavid Greenman 	}
2131561d038SDavid Greenman 
214ae92ea44SDavid Greenman 	if ((kva = kmem_alloc_pageable(io_map, size)) == 0) {
2151561d038SDavid Greenman 		if( !waitok) {
2161561d038SDavid Greenman 			splx(s);
2171561d038SDavid Greenman 			return NULL;
2181561d038SDavid Greenman 		}
219d5e26ef0SDavid Greenman 		bmwait = 1;
2201561d038SDavid Greenman 		tsleep((caddr_t) io_map, PRIBIO, "bmwait", 0);
221d5e26ef0SDavid Greenman 		goto more;
222d5e26ef0SDavid Greenman 	}
223d5e26ef0SDavid Greenman 	splx(s);
224d5e26ef0SDavid Greenman 	return kva;
225d5e26ef0SDavid Greenman }
226d5e26ef0SDavid Greenman 
227d5e26ef0SDavid Greenman /*
228ae92ea44SDavid Greenman  * same as vm_bounce_kva -- but really allocate (but takes pages as arg)
2290e195446SDavid Greenman  */
2300e195446SDavid Greenman vm_offset_t
2310e195446SDavid Greenman vm_bounce_kva_alloc(count)
2320e195446SDavid Greenman int count;
2330e195446SDavid Greenman {
2340e195446SDavid Greenman 	int i;
2350e195446SDavid Greenman 	vm_offset_t kva;
2360e195446SDavid Greenman 	vm_offset_t pa;
2370e195446SDavid Greenman 	if( bouncepages == 0) {
2380e195446SDavid Greenman 		kva = (vm_offset_t) malloc(count*NBPG, M_TEMP, M_WAITOK);
2390e195446SDavid Greenman 		return kva;
2400e195446SDavid Greenman 	}
241ae92ea44SDavid Greenman 	kva = vm_bounce_kva(count*NBPG, 1);
2420e195446SDavid Greenman 	for(i=0;i<count;i++) {
2430e195446SDavid Greenman 		pa = vm_bounce_page_find(1);
2440e195446SDavid Greenman 		pmap_kenter(kva + i * NBPG, pa);
2450e195446SDavid Greenman 	}
24626f9a767SRodney W. Grimes 	pmap_update();
2470e195446SDavid Greenman 	return kva;
2480e195446SDavid Greenman }
2490e195446SDavid Greenman 
2500e195446SDavid Greenman /*
2510e195446SDavid Greenman  * same as vm_bounce_kva_free -- but really free
2520e195446SDavid Greenman  */
2530e195446SDavid Greenman void
2540e195446SDavid Greenman vm_bounce_kva_alloc_free(kva, count)
2550e195446SDavid Greenman 	vm_offset_t kva;
2560e195446SDavid Greenman 	int count;
2570e195446SDavid Greenman {
2580e195446SDavid Greenman 	int i;
2590e195446SDavid Greenman 	vm_offset_t pa;
2600e195446SDavid Greenman 	if( bouncepages == 0) {
2610e195446SDavid Greenman 		free((caddr_t) kva, M_TEMP);
2620e195446SDavid Greenman 		return;
2630e195446SDavid Greenman 	}
2640e195446SDavid Greenman 	for(i = 0; i < count; i++) {
2650e195446SDavid Greenman 		pa = pmap_kextract(kva + i * NBPG);
2660e195446SDavid Greenman 		vm_bounce_page_free(pa, 1);
2670e195446SDavid Greenman 	}
268ae92ea44SDavid Greenman 	vm_bounce_kva_free(kva, count*NBPG, 0);
2690e195446SDavid Greenman }
2700e195446SDavid Greenman 
2710e195446SDavid Greenman /*
272d5e26ef0SDavid Greenman  * do the things necessary to the struct buf to implement
273d5e26ef0SDavid Greenman  * bounce buffers...  inserted before the disk sort
274d5e26ef0SDavid Greenman  */
275d5e26ef0SDavid Greenman void
276d5e26ef0SDavid Greenman vm_bounce_alloc(bp)
277d5e26ef0SDavid Greenman 	struct buf *bp;
278d5e26ef0SDavid Greenman {
279d5e26ef0SDavid Greenman 	int countvmpg;
280d5e26ef0SDavid Greenman 	vm_offset_t vastart, vaend;
281d5e26ef0SDavid Greenman 	vm_offset_t vapstart, vapend;
282d5e26ef0SDavid Greenman 	vm_offset_t va, kva;
283d5e26ef0SDavid Greenman 	vm_offset_t pa;
284d5e26ef0SDavid Greenman 	int dobounceflag = 0;
285d5e26ef0SDavid Greenman 	int bounceindex;
286d5e26ef0SDavid Greenman 	int i;
287d5e26ef0SDavid Greenman 	int s;
288d5e26ef0SDavid Greenman 
289d5e26ef0SDavid Greenman 	if (bouncepages == 0)
290d5e26ef0SDavid Greenman 		return;
291d5e26ef0SDavid Greenman 
292ae92ea44SDavid Greenman 	if (bp->b_flags & B_BOUNCE) {
293ae92ea44SDavid Greenman 		printf("vm_bounce_alloc: called recursively???\n");
294ae92ea44SDavid Greenman 		return;
2951561d038SDavid Greenman 	}
2961561d038SDavid Greenman 
297ae92ea44SDavid Greenman 	if (bp->b_bufsize < bp->b_bcount) {
298ae92ea44SDavid Greenman 		printf("vm_bounce_alloc: b_bufsize(0x%x) < b_bcount(0x%x) !!!!\n",
299ae92ea44SDavid Greenman 			bp->b_bufsize, bp->b_bcount);
300ae92ea44SDavid Greenman 		panic("vm_bounce_alloc");
301ae92ea44SDavid Greenman 	}
302ae92ea44SDavid Greenman 
303ae92ea44SDavid Greenman /*
304ae92ea44SDavid Greenman  *  This is not really necessary
305ae92ea44SDavid Greenman  *	if( bp->b_bufsize != bp->b_bcount) {
306ae92ea44SDavid Greenman  *		printf("size: %d, count: %d\n", bp->b_bufsize, bp->b_bcount);
307ae92ea44SDavid Greenman  *	}
308ae92ea44SDavid Greenman  */
309ae92ea44SDavid Greenman 
310ae92ea44SDavid Greenman 
31126f9a767SRodney W. Grimes 	vastart = (vm_offset_t) bp->b_data;
31226f9a767SRodney W. Grimes 	vaend = (vm_offset_t) bp->b_data + bp->b_bufsize;
313d5e26ef0SDavid Greenman 
314d5e26ef0SDavid Greenman 	vapstart = i386_trunc_page(vastart);
315d5e26ef0SDavid Greenman 	vapend = i386_round_page(vaend);
316d5e26ef0SDavid Greenman 	countvmpg = (vapend - vapstart) / NBPG;
317d5e26ef0SDavid Greenman 
318d5e26ef0SDavid Greenman /*
319d5e26ef0SDavid Greenman  * if any page is above 16MB, then go into bounce-buffer mode
320d5e26ef0SDavid Greenman  */
321d5e26ef0SDavid Greenman 	va = vapstart;
322d5e26ef0SDavid Greenman 	for (i = 0; i < countvmpg; i++) {
323ed7fcbd0SDavid Greenman 		pa = pmap_kextract(va);
324d5e26ef0SDavid Greenman 		if (pa >= SIXTEENMEG)
325d5e26ef0SDavid Greenman 			++dobounceflag;
326d5e26ef0SDavid Greenman 		va += NBPG;
327d5e26ef0SDavid Greenman 	}
328d5e26ef0SDavid Greenman 	if (dobounceflag == 0)
329d5e26ef0SDavid Greenman 		return;
330d5e26ef0SDavid Greenman 
331d5e26ef0SDavid Greenman 	if (bouncepages < dobounceflag)
332d5e26ef0SDavid Greenman 		panic("Not enough bounce buffers!!!");
333d5e26ef0SDavid Greenman 
334d5e26ef0SDavid Greenman /*
335d5e26ef0SDavid Greenman  * allocate a replacement kva for b_addr
336d5e26ef0SDavid Greenman  */
3371561d038SDavid Greenman 	kva = vm_bounce_kva(countvmpg*NBPG, 1);
338ae92ea44SDavid Greenman #if 0
339ae92ea44SDavid Greenman 	printf("%s: vapstart: %x, vapend: %x, countvmpg: %d, kva: %x ",
340ae92ea44SDavid Greenman 		(bp->b_flags & B_READ) ? "read":"write",
341ae92ea44SDavid Greenman 			vapstart, vapend, countvmpg, kva);
342ae92ea44SDavid Greenman #endif
343d5e26ef0SDavid Greenman 	va = vapstart;
344d5e26ef0SDavid Greenman 	for (i = 0; i < countvmpg; i++) {
345ed7fcbd0SDavid Greenman 		pa = pmap_kextract(va);
346d5e26ef0SDavid Greenman 		if (pa >= SIXTEENMEG) {
347d5e26ef0SDavid Greenman 			/*
348d5e26ef0SDavid Greenman 			 * allocate a replacement page
349d5e26ef0SDavid Greenman 			 */
350d5e26ef0SDavid Greenman 			vm_offset_t bpa = vm_bounce_page_find(1);
3516b4ac811SDavid Greenman 			pmap_kenter(kva + (NBPG * i), bpa);
352ae92ea44SDavid Greenman #if 0
353ae92ea44SDavid Greenman 			printf("r(%d): (%x,%x,%x) ", i, va, pa, bpa);
354ae92ea44SDavid Greenman #endif
355d5e26ef0SDavid Greenman 			/*
356d5e26ef0SDavid Greenman 			 * if we are writing, the copy the data into the page
357d5e26ef0SDavid Greenman 			 */
3581561d038SDavid Greenman 			if ((bp->b_flags & B_READ) == 0) {
3591561d038SDavid Greenman 				pmap_update();
360d5e26ef0SDavid Greenman 				bcopy((caddr_t) va, (caddr_t) kva + (NBPG * i), NBPG);
3611561d038SDavid Greenman 			}
362d5e26ef0SDavid Greenman 		} else {
363d5e26ef0SDavid Greenman 			/*
364d5e26ef0SDavid Greenman 			 * use original page
365d5e26ef0SDavid Greenman 			 */
3666b4ac811SDavid Greenman 			pmap_kenter(kva + (NBPG * i), pa);
367d5e26ef0SDavid Greenman 		}
368d5e26ef0SDavid Greenman 		va += NBPG;
369d5e26ef0SDavid Greenman 	}
3706b4ac811SDavid Greenman 	pmap_update();
371d5e26ef0SDavid Greenman 
372d5e26ef0SDavid Greenman /*
373d5e26ef0SDavid Greenman  * flag the buffer as being bounced
374d5e26ef0SDavid Greenman  */
375d5e26ef0SDavid Greenman 	bp->b_flags |= B_BOUNCE;
376d5e26ef0SDavid Greenman /*
377d5e26ef0SDavid Greenman  * save the original buffer kva
378d5e26ef0SDavid Greenman  */
37926f9a767SRodney W. Grimes 	bp->b_savekva = bp->b_data;
380d5e26ef0SDavid Greenman /*
381d5e26ef0SDavid Greenman  * put our new kva into the buffer (offset by original offset)
382d5e26ef0SDavid Greenman  */
38326f9a767SRodney W. Grimes 	bp->b_data = (caddr_t) (((vm_offset_t) kva) |
384d5e26ef0SDavid Greenman 				((vm_offset_t) bp->b_savekva & (NBPG - 1)));
385ae92ea44SDavid Greenman #if 0
386ae92ea44SDavid Greenman 	printf("b_savekva: %x, newva: %x\n", bp->b_savekva, bp->b_data);
387ae92ea44SDavid Greenman #endif
388d5e26ef0SDavid Greenman 	return;
389d5e26ef0SDavid Greenman }
390d5e26ef0SDavid Greenman 
391d5e26ef0SDavid Greenman /*
392d5e26ef0SDavid Greenman  * hook into biodone to free bounce buffer
393d5e26ef0SDavid Greenman  */
394d5e26ef0SDavid Greenman void
395d5e26ef0SDavid Greenman vm_bounce_free(bp)
396d5e26ef0SDavid Greenman 	struct buf *bp;
397d5e26ef0SDavid Greenman {
398d5e26ef0SDavid Greenman 	int i;
399ae92ea44SDavid Greenman 	vm_offset_t origkva, bouncekva, bouncekvaend;
400d5e26ef0SDavid Greenman 	int countvmpg;
401d5e26ef0SDavid Greenman 	int s;
402d5e26ef0SDavid Greenman 
403d5e26ef0SDavid Greenman /*
404d5e26ef0SDavid Greenman  * if this isn't a bounced buffer, then just return
405d5e26ef0SDavid Greenman  */
406d5e26ef0SDavid Greenman 	if ((bp->b_flags & B_BOUNCE) == 0)
407d5e26ef0SDavid Greenman 		return;
408d5e26ef0SDavid Greenman 
409ae92ea44SDavid Greenman /*
410ae92ea44SDavid Greenman  *  This check is not necessary
411ae92ea44SDavid Greenman  *	if (bp->b_bufsize != bp->b_bcount) {
412ae92ea44SDavid Greenman  *		printf("vm_bounce_free: b_bufsize=%d, b_bcount=%d\n",
413ae92ea44SDavid Greenman  *			bp->b_bufsize, bp->b_bcount);
414ae92ea44SDavid Greenman  *	}
415ae92ea44SDavid Greenman  */
416ae92ea44SDavid Greenman 
417d5e26ef0SDavid Greenman 	origkva = (vm_offset_t) bp->b_savekva;
41826f9a767SRodney W. Grimes 	bouncekva = (vm_offset_t) bp->b_data;
419ae92ea44SDavid Greenman /*
420ae92ea44SDavid Greenman 	printf("free: %d ", bp->b_bufsize);
421ae92ea44SDavid Greenman */
422d5e26ef0SDavid Greenman 
423d5e26ef0SDavid Greenman /*
424d5e26ef0SDavid Greenman  * check every page in the kva space for b_addr
425d5e26ef0SDavid Greenman  */
426ae92ea44SDavid Greenman 	for (i = 0; i < bp->b_bufsize; ) {
427d5e26ef0SDavid Greenman 		vm_offset_t mybouncepa;
428d5e26ef0SDavid Greenman 		vm_offset_t copycount;
429d5e26ef0SDavid Greenman 
430d5e26ef0SDavid Greenman 		copycount = i386_round_page(bouncekva + 1) - bouncekva;
431ed7fcbd0SDavid Greenman 		mybouncepa = pmap_kextract(i386_trunc_page(bouncekva));
432d5e26ef0SDavid Greenman 
433d5e26ef0SDavid Greenman /*
434d5e26ef0SDavid Greenman  * if this is a bounced pa, then process as one
435d5e26ef0SDavid Greenman  */
436ae92ea44SDavid Greenman 		if ( mybouncepa != pmap_kextract( i386_trunc_page( origkva))) {
437ae92ea44SDavid Greenman 			vm_offset_t tocopy = copycount;
438ae92ea44SDavid Greenman 			if (i + tocopy > bp->b_bufsize)
439ae92ea44SDavid Greenman 				tocopy = bp->b_bufsize - i;
440d5e26ef0SDavid Greenman /*
441d5e26ef0SDavid Greenman  * if this is a read, then copy from bounce buffer into original buffer
442d5e26ef0SDavid Greenman  */
443d5e26ef0SDavid Greenman 			if (bp->b_flags & B_READ)
444ae92ea44SDavid Greenman 				bcopy((caddr_t) bouncekva, (caddr_t) origkva, tocopy);
445d5e26ef0SDavid Greenman /*
446d5e26ef0SDavid Greenman  * free the bounce allocation
447d5e26ef0SDavid Greenman  */
448ae92ea44SDavid Greenman 
449ae92ea44SDavid Greenman /*
450ae92ea44SDavid Greenman 			printf("(kva: %x, pa: %x)", bouncekva, mybouncepa);
451ae92ea44SDavid Greenman */
452ae92ea44SDavid Greenman 			vm_bounce_page_free(mybouncepa, 1);
453d5e26ef0SDavid Greenman 		}
454d5e26ef0SDavid Greenman 
455d5e26ef0SDavid Greenman 		origkva += copycount;
456d5e26ef0SDavid Greenman 		bouncekva += copycount;
457ae92ea44SDavid Greenman 		i += copycount;
458d5e26ef0SDavid Greenman 	}
459d5e26ef0SDavid Greenman 
460d5e26ef0SDavid Greenman /*
461ae92ea44SDavid Greenman 	printf("\n");
462ae92ea44SDavid Greenman */
463ae92ea44SDavid Greenman /*
464d5e26ef0SDavid Greenman  * add the old kva into the "to free" list
465d5e26ef0SDavid Greenman  */
466ae92ea44SDavid Greenman 
46726f9a767SRodney W. Grimes 	bouncekva= i386_trunc_page((vm_offset_t) bp->b_data);
468ae92ea44SDavid Greenman 	bouncekvaend= i386_round_page((vm_offset_t)bp->b_data + bp->b_bufsize);
469ae92ea44SDavid Greenman 
470ae92ea44SDavid Greenman /*
471ae92ea44SDavid Greenman 	printf("freeva: %d\n", (bouncekvaend - bouncekva) / NBPG);
472ae92ea44SDavid Greenman */
473ae92ea44SDavid Greenman 	vm_bounce_kva_free( bouncekva, (bouncekvaend - bouncekva), 0);
47426f9a767SRodney W. Grimes 	bp->b_data = bp->b_savekva;
475d5e26ef0SDavid Greenman 	bp->b_savekva = 0;
476d5e26ef0SDavid Greenman 	bp->b_flags &= ~B_BOUNCE;
477d5e26ef0SDavid Greenman 
478d5e26ef0SDavid Greenman 	return;
479d5e26ef0SDavid Greenman }
480d5e26ef0SDavid Greenman 
481ae92ea44SDavid Greenman 
4825b81b6b3SRodney W. Grimes /*
4831561d038SDavid Greenman  * init the bounce buffer system
4841561d038SDavid Greenman  */
4851561d038SDavid Greenman void
4861561d038SDavid Greenman vm_bounce_init()
4871561d038SDavid Greenman {
4881561d038SDavid Greenman 	vm_offset_t minaddr, maxaddr;
489ae92ea44SDavid Greenman 	int i;
4901561d038SDavid Greenman 
4911561d038SDavid Greenman 	kvasfreecnt = 0;
4921561d038SDavid Greenman 
4931561d038SDavid Greenman 	if (bouncepages == 0)
4941561d038SDavid Greenman 		return;
4951561d038SDavid Greenman 
4961561d038SDavid Greenman 	bounceallocarraysize = (bouncepages + BITS_IN_UNSIGNED - 1) / BITS_IN_UNSIGNED;
4971561d038SDavid Greenman 	bounceallocarray = malloc(bounceallocarraysize * sizeof(unsigned), M_TEMP, M_NOWAIT);
4981561d038SDavid Greenman 
4991561d038SDavid Greenman 	if (!bounceallocarray)
5001561d038SDavid Greenman 		panic("Cannot allocate bounce resource array\n");
5011561d038SDavid Greenman 
502ae92ea44SDavid Greenman 	bzero(bounceallocarray, bounceallocarraysize * sizeof(unsigned));
503ae92ea44SDavid Greenman 	bouncepa = malloc(bouncepages * sizeof(vm_offset_t), M_TEMP, M_NOWAIT);
504ae92ea44SDavid Greenman 	if (!bouncepa)
505ae92ea44SDavid Greenman 		panic("Cannot allocate physical memory array\n");
5061561d038SDavid Greenman 
507ae92ea44SDavid Greenman 	for(i=0;i<bouncepages;i++) {
508ae92ea44SDavid Greenman 		vm_offset_t pa;
509ae92ea44SDavid Greenman 		if( (pa = pmap_kextract((vm_offset_t) bouncememory + i * NBPG)) >= SIXTEENMEG)
510ae92ea44SDavid Greenman 			panic("bounce memory out of range");
511ae92ea44SDavid Greenman 		if( pa == 0)
512ae92ea44SDavid Greenman 			panic("bounce memory not resident");
513ae92ea44SDavid Greenman 		bouncepa[i] = pa;
514ae92ea44SDavid Greenman 	}
5151561d038SDavid Greenman 	bouncefree = bouncepages;
5161561d038SDavid Greenman 
5171561d038SDavid Greenman }
518ae92ea44SDavid Greenman #endif /* NOBOUNCE */
519a4f7a4c9SDavid Greenman /*
520a4f7a4c9SDavid Greenman  * quick version of vm_fault
521a4f7a4c9SDavid Greenman  */
522a4f7a4c9SDavid Greenman 
523a4f7a4c9SDavid Greenman void
524a4f7a4c9SDavid Greenman vm_fault_quick( v, prot)
525a4f7a4c9SDavid Greenman 	vm_offset_t v;
526a4f7a4c9SDavid Greenman 	int prot;
527a4f7a4c9SDavid Greenman {
528a4f7a4c9SDavid Greenman 	if( (cpu_class == CPUCLASS_386) &&
529a4f7a4c9SDavid Greenman 		(prot & VM_PROT_WRITE))
530a4f7a4c9SDavid Greenman 		vm_fault(&curproc->p_vmspace->vm_map, v,
531a4f7a4c9SDavid Greenman 			VM_PROT_READ|VM_PROT_WRITE, FALSE);
532a4f7a4c9SDavid Greenman 	else if( prot & VM_PROT_WRITE)
533a4f7a4c9SDavid Greenman 		*(volatile char *)v += 0;
534a4f7a4c9SDavid Greenman 	else
535a4f7a4c9SDavid Greenman 		*(volatile char *)v;
536a4f7a4c9SDavid Greenman }
537a4f7a4c9SDavid Greenman 
5381561d038SDavid Greenman 
5391561d038SDavid Greenman /*
5405b81b6b3SRodney W. Grimes  * Finish a fork operation, with process p2 nearly set up.
5415b81b6b3SRodney W. Grimes  * Copy and update the kernel stack and pcb, making the child
5425b81b6b3SRodney W. Grimes  * ready to run, and marking it so that it can return differently
5435b81b6b3SRodney W. Grimes  * than the parent.  Returns 1 in the child process, 0 in the parent.
5445b81b6b3SRodney W. Grimes  * We currently double-map the user area so that the stack is at the same
5455b81b6b3SRodney W. Grimes  * address in each process; in the future we will probably relocate
5465b81b6b3SRodney W. Grimes  * the frame pointers on the stack after copying.
5475b81b6b3SRodney W. Grimes  */
548381fe1aaSGarrett Wollman int
5495b81b6b3SRodney W. Grimes cpu_fork(p1, p2)
5505b81b6b3SRodney W. Grimes 	register struct proc *p1, *p2;
5515b81b6b3SRodney W. Grimes {
5525b81b6b3SRodney W. Grimes 	register struct user *up = p2->p_addr;
5535b81b6b3SRodney W. Grimes 	int foo, offset, addr, i;
5545b81b6b3SRodney W. Grimes 	extern char kstack[];
5555b81b6b3SRodney W. Grimes 	extern int mvesp();
5565b81b6b3SRodney W. Grimes 
5575b81b6b3SRodney W. Grimes 	/*
5585b81b6b3SRodney W. Grimes 	 * Copy pcb and stack from proc p1 to p2.
5595b81b6b3SRodney W. Grimes 	 * We do this as cheaply as possible, copying only the active
5605b81b6b3SRodney W. Grimes 	 * part of the stack.  The stack and pcb need to agree;
5615b81b6b3SRodney W. Grimes 	 * this is tricky, as the final pcb is constructed by savectx,
5625b81b6b3SRodney W. Grimes 	 * but its frame isn't yet on the stack when the stack is copied.
5635b81b6b3SRodney W. Grimes 	 * swtch compensates for this when the child eventually runs.
5645b81b6b3SRodney W. Grimes 	 * This should be done differently, with a single call
5655b81b6b3SRodney W. Grimes 	 * that copies and updates the pcb+stack,
5665b81b6b3SRodney W. Grimes 	 * replacing the bcopy and savectx.
5675b81b6b3SRodney W. Grimes 	 */
5685b81b6b3SRodney W. Grimes 	p2->p_addr->u_pcb = p1->p_addr->u_pcb;
5695b81b6b3SRodney W. Grimes 	offset = mvesp() - (int)kstack;
5705b81b6b3SRodney W. Grimes 	bcopy((caddr_t)kstack + offset, (caddr_t)p2->p_addr + offset,
5715b81b6b3SRodney W. Grimes 	    (unsigned) ctob(UPAGES) - offset);
57226f9a767SRodney W. Grimes 	p2->p_md.md_regs = p1->p_md.md_regs;
5735b81b6b3SRodney W. Grimes 
5745b81b6b3SRodney W. Grimes 	pmap_activate(&p2->p_vmspace->vm_pmap, &up->u_pcb);
5755b81b6b3SRodney W. Grimes 
5765b81b6b3SRodney W. Grimes 	/*
5775b81b6b3SRodney W. Grimes 	 *
5785b81b6b3SRodney W. Grimes 	 * Arrange for a non-local goto when the new process
5795b81b6b3SRodney W. Grimes 	 * is started, to resume here, returning nonzero from setjmp.
5805b81b6b3SRodney W. Grimes 	 */
5815b81b6b3SRodney W. Grimes 	if (savectx(up, 1)) {
5825b81b6b3SRodney W. Grimes 		/*
5835b81b6b3SRodney W. Grimes 		 * Return 1 in child.
5845b81b6b3SRodney W. Grimes 		 */
5855b81b6b3SRodney W. Grimes 		return (1);
5865b81b6b3SRodney W. Grimes 	}
5875b81b6b3SRodney W. Grimes 	return (0);
5885b81b6b3SRodney W. Grimes }
5895b81b6b3SRodney W. Grimes 
5907c2b54e8SNate Williams void
5915b81b6b3SRodney W. Grimes cpu_exit(p)
5925b81b6b3SRodney W. Grimes 	register struct proc *p;
5935b81b6b3SRodney W. Grimes {
5945b81b6b3SRodney W. Grimes 
595960173b9SRodney W. Grimes #if NNPX > 0
5965b81b6b3SRodney W. Grimes 	npxexit(p);
597960173b9SRodney W. Grimes #endif	/* NNPX */
59826f9a767SRodney W. Grimes 	curproc = p;
59926f9a767SRodney W. Grimes 	mi_switch();
6007c2b54e8SNate Williams 	/*
6017c2b54e8SNate Williams 	 * This is to shutup the compiler, and if swtch() failed I suppose
6027c2b54e8SNate Williams 	 * this would be a good thing.  This keeps gcc happy because panic
6037c2b54e8SNate Williams 	 * is a volatile void function as well.
6047c2b54e8SNate Williams 	 */
6057c2b54e8SNate Williams 	panic("cpu_exit");
6065b81b6b3SRodney W. Grimes }
6075b81b6b3SRodney W. Grimes 
608381fe1aaSGarrett Wollman void
6097f8cb368SDavid Greenman cpu_wait(p) struct proc *p; {
6107f8cb368SDavid Greenman /*	extern vm_map_t upages_map; */
6117f8cb368SDavid Greenman 	extern char kstack[];
6125b81b6b3SRodney W. Grimes 
6135b81b6b3SRodney W. Grimes 	/* drop per-process resources */
6147f8cb368SDavid Greenman  	pmap_remove(vm_map_pmap(kernel_map), (vm_offset_t) p->p_addr,
6157f8cb368SDavid Greenman 		((vm_offset_t) p->p_addr) + ctob(UPAGES));
6165b81b6b3SRodney W. Grimes 	kmem_free(kernel_map, (vm_offset_t)p->p_addr, ctob(UPAGES));
6177f8cb368SDavid Greenman 	vmspace_free(p->p_vmspace);
6185b81b6b3SRodney W. Grimes }
6195b81b6b3SRodney W. Grimes 
6205b81b6b3SRodney W. Grimes /*
62126f9a767SRodney W. Grimes  * Dump the machine specific header information at the start of a core dump.
62226f9a767SRodney W. Grimes  */
62326f9a767SRodney W. Grimes int
62426f9a767SRodney W. Grimes cpu_coredump(p, vp, cred)
62526f9a767SRodney W. Grimes 	struct proc *p;
62626f9a767SRodney W. Grimes 	struct vnode *vp;
62726f9a767SRodney W. Grimes 	struct ucred *cred;
62826f9a767SRodney W. Grimes {
62926f9a767SRodney W. Grimes 
63026f9a767SRodney W. Grimes 	return (vn_rdwr(UIO_WRITE, vp, (caddr_t) p->p_addr, ctob(UPAGES),
63126f9a767SRodney W. Grimes 	    (off_t)0, UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *)NULL,
63226f9a767SRodney W. Grimes 	    p));
63326f9a767SRodney W. Grimes }
63426f9a767SRodney W. Grimes 
63526f9a767SRodney W. Grimes /*
6365b81b6b3SRodney W. Grimes  * Set a red zone in the kernel stack after the u. area.
6375b81b6b3SRodney W. Grimes  */
638381fe1aaSGarrett Wollman void
6395b81b6b3SRodney W. Grimes setredzone(pte, vaddr)
6405b81b6b3SRodney W. Grimes 	u_short *pte;
6415b81b6b3SRodney W. Grimes 	caddr_t vaddr;
6425b81b6b3SRodney W. Grimes {
6435b81b6b3SRodney W. Grimes /* eventually do this by setting up an expand-down stack segment
6445b81b6b3SRodney W. Grimes    for ss0: selector, allowing stack access down to top of u.
6455b81b6b3SRodney W. Grimes    this means though that protection violations need to be handled
6465b81b6b3SRodney W. Grimes    thru a double fault exception that must do an integral task
6475b81b6b3SRodney W. Grimes    switch to a known good context, within which a dump can be
6485b81b6b3SRodney W. Grimes    taken. a sensible scheme might be to save the initial context
6495b81b6b3SRodney W. Grimes    used by sched (that has physical memory mapped 1:1 at bottom)
6505b81b6b3SRodney W. Grimes    and take the dump while still in mapped mode */
6515b81b6b3SRodney W. Grimes }
6525b81b6b3SRodney W. Grimes 
6535b81b6b3SRodney W. Grimes /*
65426f9a767SRodney W. Grimes  * Move pages from one kernel virtual address to another.
65526f9a767SRodney W. Grimes  * Both addresses are assumed to reside in the Sysmap,
65626f9a767SRodney W. Grimes  * and size must be a multiple of CLSIZE.
65726f9a767SRodney W. Grimes  */
65826f9a767SRodney W. Grimes 
65926f9a767SRodney W. Grimes void
66026f9a767SRodney W. Grimes pagemove(from, to, size)
66126f9a767SRodney W. Grimes 	register caddr_t from, to;
66226f9a767SRodney W. Grimes 	int size;
66326f9a767SRodney W. Grimes {
66426f9a767SRodney W. Grimes 	register vm_offset_t pa;
66526f9a767SRodney W. Grimes 
66626f9a767SRodney W. Grimes 	if (size & CLOFSET)
66726f9a767SRodney W. Grimes 		panic("pagemove");
66826f9a767SRodney W. Grimes 	while (size > 0) {
66926f9a767SRodney W. Grimes 		pa = pmap_kextract((vm_offset_t)from);
67026f9a767SRodney W. Grimes 		if (pa == 0)
67126f9a767SRodney W. Grimes 			panic("pagemove 2");
67226f9a767SRodney W. Grimes 		if (pmap_kextract((vm_offset_t)to) != 0)
67326f9a767SRodney W. Grimes 			panic("pagemove 3");
674c564966bSDavid Greenman 		pmap_kremove((vm_offset_t)from);
67526f9a767SRodney W. Grimes 		pmap_kenter((vm_offset_t)to, pa);
67626f9a767SRodney W. Grimes 		from += PAGE_SIZE;
67726f9a767SRodney W. Grimes 		to += PAGE_SIZE;
67826f9a767SRodney W. Grimes 		size -= PAGE_SIZE;
67926f9a767SRodney W. Grimes 	}
68026f9a767SRodney W. Grimes 	pmap_update();
68126f9a767SRodney W. Grimes }
68226f9a767SRodney W. Grimes 
68326f9a767SRodney W. Grimes /*
6845b81b6b3SRodney W. Grimes  * Convert kernel VA to physical address
6855b81b6b3SRodney W. Grimes  */
686aaf08d94SGarrett Wollman u_long
6877f8cb368SDavid Greenman kvtop(void *addr)
6885b81b6b3SRodney W. Grimes {
6895b81b6b3SRodney W. Grimes 	vm_offset_t va;
6905b81b6b3SRodney W. Grimes 
691ed7fcbd0SDavid Greenman 	va = pmap_kextract((vm_offset_t)addr);
6925b81b6b3SRodney W. Grimes 	if (va == 0)
6935b81b6b3SRodney W. Grimes 		panic("kvtop: zero page frame");
6947f8cb368SDavid Greenman 	return((int)va);
6955b81b6b3SRodney W. Grimes }
6965b81b6b3SRodney W. Grimes 
6975b81b6b3SRodney W. Grimes /*
698ac322158SDavid Greenman  * Map an IO request into kernel virtual address space.
6995b81b6b3SRodney W. Grimes  *
700ac322158SDavid Greenman  * All requests are (re)mapped into kernel VA space.
701ac322158SDavid Greenman  * Notice that we use b_bufsize for the size of the buffer
702ac322158SDavid Greenman  * to be mapped.  b_bcount might be modified by the driver.
7035b81b6b3SRodney W. Grimes  */
704381fe1aaSGarrett Wollman void
7055b81b6b3SRodney W. Grimes vmapbuf(bp)
7065b81b6b3SRodney W. Grimes 	register struct buf *bp;
7075b81b6b3SRodney W. Grimes {
7085b81b6b3SRodney W. Grimes 	register int npf;
7095b81b6b3SRodney W. Grimes 	register caddr_t addr;
7105b81b6b3SRodney W. Grimes 	int off;
7115b81b6b3SRodney W. Grimes 	vm_offset_t kva;
71226f9a767SRodney W. Grimes 	vm_offset_t pa, lastv, v;
7135b81b6b3SRodney W. Grimes 
71426f9a767SRodney W. Grimes 	if ((bp->b_flags & B_PHYS) == 0)
7155b81b6b3SRodney W. Grimes 		panic("vmapbuf");
71626f9a767SRodney W. Grimes 
717ae92ea44SDavid Greenman 	/*
718ae92ea44SDavid Greenman 	 * this is the kva that is to be used for
719ae92ea44SDavid Greenman 	 * the temporary kernel mapping
720ae92ea44SDavid Greenman 	 */
721ae92ea44SDavid Greenman 	kva = (vm_offset_t) bp->b_saveaddr;
722ae92ea44SDavid Greenman 
72326f9a767SRodney W. Grimes 	lastv = 0;
72426f9a767SRodney W. Grimes 	for (addr = (caddr_t)trunc_page(bp->b_data);
72526f9a767SRodney W. Grimes 		addr < bp->b_data + bp->b_bufsize;
72626f9a767SRodney W. Grimes 		addr += PAGE_SIZE) {
72726f9a767SRodney W. Grimes 
72826f9a767SRodney W. Grimes /*
72926f9a767SRodney W. Grimes  * make sure that the pde is valid and held
73026f9a767SRodney W. Grimes  */
73126f9a767SRodney W. Grimes 		v = trunc_page(((vm_offset_t)vtopte(addr)));
73226f9a767SRodney W. Grimes 		if (v != lastv) {
73326f9a767SRodney W. Grimes 			vm_fault_quick(v, VM_PROT_READ);
734c564966bSDavid Greenman 			pa = pmap_kextract( v);
73526f9a767SRodney W. Grimes 			vm_page_hold(PHYS_TO_VM_PAGE(pa));
73626f9a767SRodney W. Grimes 			lastv = v;
73726f9a767SRodney W. Grimes 		}
73826f9a767SRodney W. Grimes 
73926f9a767SRodney W. Grimes /*
74026f9a767SRodney W. Grimes  * do the vm_fault if needed, do the copy-on-write thing when
74126f9a767SRodney W. Grimes  * reading stuff off device into memory.
74226f9a767SRodney W. Grimes  */
74326f9a767SRodney W. Grimes 		vm_fault_quick(addr,
74426f9a767SRodney W. Grimes 			(bp->b_flags&B_READ)?(VM_PROT_READ|VM_PROT_WRITE):VM_PROT_READ);
745c564966bSDavid Greenman 		pa = pmap_kextract((vm_offset_t) addr);
74626f9a767SRodney W. Grimes /*
74726f9a767SRodney W. Grimes  * hold the data page
74826f9a767SRodney W. Grimes  */
74926f9a767SRodney W. Grimes 		vm_page_hold(PHYS_TO_VM_PAGE(pa));
75026f9a767SRodney W. Grimes 	}
75126f9a767SRodney W. Grimes 
752ae92ea44SDavid Greenman 	addr = bp->b_saveaddr = bp->b_data;
7535b81b6b3SRodney W. Grimes 	off = (int)addr & PGOFSET;
754ac322158SDavid Greenman 	npf = btoc(round_page(bp->b_bufsize + off));
755ae92ea44SDavid Greenman 	bp->b_data = (caddr_t) (kva + off);
7565b81b6b3SRodney W. Grimes 	while (npf--) {
757c564966bSDavid Greenman 		pa = pmap_kextract((vm_offset_t)addr);
7585b81b6b3SRodney W. Grimes 		if (pa == 0)
7595b81b6b3SRodney W. Grimes 			panic("vmapbuf: null page frame");
7606b4ac811SDavid Greenman 		pmap_kenter(kva, trunc_page(pa));
7615b81b6b3SRodney W. Grimes 		addr += PAGE_SIZE;
7625b81b6b3SRodney W. Grimes 		kva += PAGE_SIZE;
7635b81b6b3SRodney W. Grimes 	}
7646b4ac811SDavid Greenman 	pmap_update();
7655b81b6b3SRodney W. Grimes }
7665b81b6b3SRodney W. Grimes 
7675b81b6b3SRodney W. Grimes /*
7685b81b6b3SRodney W. Grimes  * Free the io map PTEs associated with this IO operation.
7695b81b6b3SRodney W. Grimes  * We also invalidate the TLB entries and restore the original b_addr.
7705b81b6b3SRodney W. Grimes  */
771381fe1aaSGarrett Wollman void
7725b81b6b3SRodney W. Grimes vunmapbuf(bp)
7735b81b6b3SRodney W. Grimes 	register struct buf *bp;
7745b81b6b3SRodney W. Grimes {
775c564966bSDavid Greenman 	register caddr_t addr;
77626f9a767SRodney W. Grimes 	vm_offset_t kva,va,v,lastv,pa;
7775b81b6b3SRodney W. Grimes 
7785b81b6b3SRodney W. Grimes 	if ((bp->b_flags & B_PHYS) == 0)
7795b81b6b3SRodney W. Grimes 		panic("vunmapbuf");
780c564966bSDavid Greenman 
781c564966bSDavid Greenman 	for (addr = (caddr_t)trunc_page((vm_offset_t) bp->b_data);
782c564966bSDavid Greenman 		addr < bp->b_data + bp->b_bufsize;
783c564966bSDavid Greenman 		addr += NBPG)
784c564966bSDavid Greenman 		pmap_kremove((vm_offset_t) addr);
785c564966bSDavid Greenman 
786ae92ea44SDavid Greenman 	bp->b_data = bp->b_saveaddr;
7875b81b6b3SRodney W. Grimes 	bp->b_saveaddr = NULL;
78826f9a767SRodney W. Grimes 
78926f9a767SRodney W. Grimes /*
79026f9a767SRodney W. Grimes  * unhold the pde, and data pages
79126f9a767SRodney W. Grimes  */
79226f9a767SRodney W. Grimes 	lastv = 0;
793c564966bSDavid Greenman 	for (addr = (caddr_t)trunc_page((vm_offset_t) bp->b_data);
79426f9a767SRodney W. Grimes 		addr < bp->b_data + bp->b_bufsize;
79526f9a767SRodney W. Grimes 		addr += NBPG) {
79626f9a767SRodney W. Grimes 
79726f9a767SRodney W. Grimes 	/*
79826f9a767SRodney W. Grimes 	 * release the data page
79926f9a767SRodney W. Grimes 	 */
800c564966bSDavid Greenman 		pa = pmap_kextract((vm_offset_t) addr);
80126f9a767SRodney W. Grimes 		vm_page_unhold(PHYS_TO_VM_PAGE(pa));
80226f9a767SRodney W. Grimes 
80326f9a767SRodney W. Grimes 	/*
80426f9a767SRodney W. Grimes 	 * and unhold the page table
80526f9a767SRodney W. Grimes 	 */
80626f9a767SRodney W. Grimes 		v = trunc_page(((vm_offset_t)vtopte(addr)));
80726f9a767SRodney W. Grimes 		if (v != lastv) {
808c564966bSDavid Greenman 			pa = pmap_kextract(v);
80926f9a767SRodney W. Grimes 			vm_page_unhold(PHYS_TO_VM_PAGE(pa));
81026f9a767SRodney W. Grimes 			lastv = v;
81126f9a767SRodney W. Grimes 		}
81226f9a767SRodney W. Grimes 	}
8135b81b6b3SRodney W. Grimes }
8145b81b6b3SRodney W. Grimes 
8155b81b6b3SRodney W. Grimes /*
8165b81b6b3SRodney W. Grimes  * Force reset the processor by invalidating the entire address space!
8175b81b6b3SRodney W. Grimes  */
8187f8cb368SDavid Greenman void
8195b81b6b3SRodney W. Grimes cpu_reset() {
8205b81b6b3SRodney W. Grimes 
8215b81b6b3SRodney W. Grimes 	/* force a shutdown by unmapping entire address space ! */
8225b81b6b3SRodney W. Grimes 	bzero((caddr_t) PTD, NBPG);
8235b81b6b3SRodney W. Grimes 
8245b81b6b3SRodney W. Grimes 	/* "good night, sweet prince .... <THUNK!>" */
8255b81b6b3SRodney W. Grimes 	tlbflush();
8265b81b6b3SRodney W. Grimes 	/* NOTREACHED */
8277f8cb368SDavid Greenman 	while(1);
8285b81b6b3SRodney W. Grimes }
829b9d60b3fSDavid Greenman 
830b9d60b3fSDavid Greenman /*
831b9d60b3fSDavid Greenman  * Grow the user stack to allow for 'sp'. This version grows the stack in
83229360eb0SDavid Greenman  *	chunks of SGROWSIZ.
833b9d60b3fSDavid Greenman  */
834b9d60b3fSDavid Greenman int
835b9d60b3fSDavid Greenman grow(p, sp)
836b9d60b3fSDavid Greenman 	struct proc *p;
83726f9a767SRodney W. Grimes 	u_int sp;
838b9d60b3fSDavid Greenman {
839b9d60b3fSDavid Greenman 	unsigned int nss;
840b9d60b3fSDavid Greenman 	caddr_t v;
841b9d60b3fSDavid Greenman 	struct vmspace *vm = p->p_vmspace;
842b9d60b3fSDavid Greenman 
843b9d60b3fSDavid Greenman 	if ((caddr_t)sp <= vm->vm_maxsaddr || (unsigned)sp >= (unsigned)USRSTACK)
844b9d60b3fSDavid Greenman 	    return (1);
845b9d60b3fSDavid Greenman 
846b9d60b3fSDavid Greenman 	nss = roundup(USRSTACK - (unsigned)sp, PAGE_SIZE);
847b9d60b3fSDavid Greenman 
848b9d60b3fSDavid Greenman 	if (nss > p->p_rlimit[RLIMIT_STACK].rlim_cur)
849b9d60b3fSDavid Greenman 		return (0);
850b9d60b3fSDavid Greenman 
851b9d60b3fSDavid Greenman 	if (vm->vm_ssize && roundup(vm->vm_ssize << PAGE_SHIFT,
85229360eb0SDavid Greenman 	    SGROWSIZ) < nss) {
853b9d60b3fSDavid Greenman 		int grow_amount;
854b9d60b3fSDavid Greenman 		/*
855b9d60b3fSDavid Greenman 		 * If necessary, grow the VM that the stack occupies
856b9d60b3fSDavid Greenman 		 * to allow for the rlimit. This allows us to not have
857b9d60b3fSDavid Greenman 		 * to allocate all of the VM up-front in execve (which
858b9d60b3fSDavid Greenman 		 * is expensive).
859b9d60b3fSDavid Greenman 		 * Grow the VM by the amount requested rounded up to
86029360eb0SDavid Greenman 		 * the nearest SGROWSIZ to provide for some hysteresis.
861b9d60b3fSDavid Greenman 		 */
86229360eb0SDavid Greenman 		grow_amount = roundup((nss - (vm->vm_ssize << PAGE_SHIFT)), SGROWSIZ);
863b9d60b3fSDavid Greenman 		v = (char *)USRSTACK - roundup(vm->vm_ssize << PAGE_SHIFT,
86429360eb0SDavid Greenman 		    SGROWSIZ) - grow_amount;
865b9d60b3fSDavid Greenman 		/*
86629360eb0SDavid Greenman 		 * If there isn't enough room to extend by SGROWSIZ, then
867b9d60b3fSDavid Greenman 		 * just extend to the maximum size
868b9d60b3fSDavid Greenman 		 */
869b9d60b3fSDavid Greenman 		if (v < vm->vm_maxsaddr) {
870b9d60b3fSDavid Greenman 			v = vm->vm_maxsaddr;
871b9d60b3fSDavid Greenman 			grow_amount = MAXSSIZ - (vm->vm_ssize << PAGE_SHIFT);
872b9d60b3fSDavid Greenman 		}
873b9d60b3fSDavid Greenman 		if (vm_allocate(&vm->vm_map, (vm_offset_t *)&v,
874b9d60b3fSDavid Greenman 		    grow_amount, FALSE) != KERN_SUCCESS) {
875b9d60b3fSDavid Greenman 			return (0);
876b9d60b3fSDavid Greenman 		}
877b9d60b3fSDavid Greenman 		vm->vm_ssize += grow_amount >> PAGE_SHIFT;
878b9d60b3fSDavid Greenman 	}
879b9d60b3fSDavid Greenman 
880b9d60b3fSDavid Greenman 	return (1);
881b9d60b3fSDavid Greenman }
882