17c478bd9Sstevel@tonic-gate /*
27c478bd9Sstevel@tonic-gate * CDDL HEADER START
37c478bd9Sstevel@tonic-gate *
47c478bd9Sstevel@tonic-gate * The contents of this file are subject to the terms of the
57a1b360bScth * Common Development and Distribution License (the "License").
67a1b360bScth * You may not use this file except in compliance with the License.
77c478bd9Sstevel@tonic-gate *
87c478bd9Sstevel@tonic-gate * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
97c478bd9Sstevel@tonic-gate * or http://www.opensolaris.org/os/licensing.
107c478bd9Sstevel@tonic-gate * See the License for the specific language governing permissions
117c478bd9Sstevel@tonic-gate * and limitations under the License.
127c478bd9Sstevel@tonic-gate *
137c478bd9Sstevel@tonic-gate * When distributing Covered Code, include this CDDL HEADER in each
147c478bd9Sstevel@tonic-gate * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
157c478bd9Sstevel@tonic-gate * If applicable, add the following below this CDDL HEADER, with the
167c478bd9Sstevel@tonic-gate * fields enclosed by brackets "[]" replaced with your own identifying
177c478bd9Sstevel@tonic-gate * information: Portions Copyright [yyyy] [name of copyright owner]
187c478bd9Sstevel@tonic-gate *
197c478bd9Sstevel@tonic-gate * CDDL HEADER END
207c478bd9Sstevel@tonic-gate */
217c478bd9Sstevel@tonic-gate /*
22*36945f79Smrj * Copyright 2007 Sun Microsystems, Inc. All rights reserved.
237c478bd9Sstevel@tonic-gate * Use is subject to license terms.
247c478bd9Sstevel@tonic-gate */
257c478bd9Sstevel@tonic-gate
267c478bd9Sstevel@tonic-gate #pragma ident "%Z%%M% %I% %E% SMI"
277c478bd9Sstevel@tonic-gate
287c478bd9Sstevel@tonic-gate #include <sys/types.h>
297c478bd9Sstevel@tonic-gate #include <sys/sysmacros.h>
307c478bd9Sstevel@tonic-gate #include <sys/systm.h>
317c478bd9Sstevel@tonic-gate #include <sys/mman.h>
327c478bd9Sstevel@tonic-gate #include <sys/buf.h>
337c478bd9Sstevel@tonic-gate #include <sys/vmem.h>
347c478bd9Sstevel@tonic-gate #include <sys/cmn_err.h>
357c478bd9Sstevel@tonic-gate #include <sys/debug.h>
367c478bd9Sstevel@tonic-gate #include <sys/machparam.h>
377c478bd9Sstevel@tonic-gate #include <vm/page.h>
387c478bd9Sstevel@tonic-gate #include <vm/seg_kmem.h>
392d3bf78eScth #include <vm/seg_kpm.h>
407c478bd9Sstevel@tonic-gate
417c478bd9Sstevel@tonic-gate #ifdef __sparc
427c478bd9Sstevel@tonic-gate #include <sys/cpu_module.h>
437c478bd9Sstevel@tonic-gate #define BP_FLUSH(addr, size) flush_instr_mem((void *)addr, size);
447c478bd9Sstevel@tonic-gate #else
457c478bd9Sstevel@tonic-gate #define BP_FLUSH(addr, size)
467c478bd9Sstevel@tonic-gate #endif
477c478bd9Sstevel@tonic-gate
48*36945f79Smrj int bp_force_copy = 0;
49*36945f79Smrj typedef enum {
50*36945f79Smrj BP_COPYIN = 0,
51*36945f79Smrj BP_COPYOUT = 1
52*36945f79Smrj } bp_copydir_t;
53*36945f79Smrj static int bp_copy_common(bp_copydir_t dir, struct buf *bp, void *driverbuf,
54*36945f79Smrj offset_t offset, size_t size);
55*36945f79Smrj
567c478bd9Sstevel@tonic-gate static vmem_t *bp_map_arena;
577c478bd9Sstevel@tonic-gate static size_t bp_align;
587c478bd9Sstevel@tonic-gate static uint_t bp_devload_flags = PROT_READ | PROT_WRITE | HAT_NOSYNC;
597c478bd9Sstevel@tonic-gate int bp_max_cache = 1 << 17; /* 128K default; tunable */
602d3bf78eScth int bp_mapin_kpm_enable = 1; /* enable default; tunable */
617c478bd9Sstevel@tonic-gate
627c478bd9Sstevel@tonic-gate static void *
bp_vmem_alloc(vmem_t * vmp,size_t size,int vmflag)637c478bd9Sstevel@tonic-gate bp_vmem_alloc(vmem_t *vmp, size_t size, int vmflag)
647c478bd9Sstevel@tonic-gate {
657c478bd9Sstevel@tonic-gate return (vmem_xalloc(vmp, size, bp_align, 0, 0, NULL, NULL, vmflag));
667c478bd9Sstevel@tonic-gate }
677c478bd9Sstevel@tonic-gate
687c478bd9Sstevel@tonic-gate void
bp_init(size_t align,uint_t devload_flags)697c478bd9Sstevel@tonic-gate bp_init(size_t align, uint_t devload_flags)
707c478bd9Sstevel@tonic-gate {
717c478bd9Sstevel@tonic-gate bp_align = MAX(align, PAGESIZE);
727c478bd9Sstevel@tonic-gate bp_devload_flags |= devload_flags;
737c478bd9Sstevel@tonic-gate
747c478bd9Sstevel@tonic-gate if (bp_align <= bp_max_cache)
757c478bd9Sstevel@tonic-gate bp_map_arena = vmem_create("bp_map", NULL, 0, bp_align,
767c478bd9Sstevel@tonic-gate bp_vmem_alloc, vmem_free, heap_arena,
777c478bd9Sstevel@tonic-gate MIN(8 * bp_align, bp_max_cache), VM_SLEEP);
787c478bd9Sstevel@tonic-gate }
797c478bd9Sstevel@tonic-gate
807c478bd9Sstevel@tonic-gate /*
817c478bd9Sstevel@tonic-gate * common routine so can be called with/without VM_SLEEP
827c478bd9Sstevel@tonic-gate */
837c478bd9Sstevel@tonic-gate void *
bp_mapin_common(struct buf * bp,int flag)847c478bd9Sstevel@tonic-gate bp_mapin_common(struct buf *bp, int flag)
857c478bd9Sstevel@tonic-gate {
867c478bd9Sstevel@tonic-gate struct as *as;
877c478bd9Sstevel@tonic-gate pfn_t pfnum;
882d3bf78eScth page_t *pp;
892d3bf78eScth page_t **pplist;
907c478bd9Sstevel@tonic-gate caddr_t kaddr;
912d3bf78eScth caddr_t addr;
922d3bf78eScth uintptr_t off;
932d3bf78eScth size_t size;
942d3bf78eScth pgcnt_t npages;
957c478bd9Sstevel@tonic-gate int color;
967c478bd9Sstevel@tonic-gate
977c478bd9Sstevel@tonic-gate /* return if already mapped in, no pageio/physio, or physio to kas */
987c478bd9Sstevel@tonic-gate if ((bp->b_flags & B_REMAPPED) ||
997c478bd9Sstevel@tonic-gate !(bp->b_flags & (B_PAGEIO | B_PHYS)) ||
1007c478bd9Sstevel@tonic-gate (((bp->b_flags & (B_PAGEIO | B_PHYS)) == B_PHYS) &&
1017c478bd9Sstevel@tonic-gate ((bp->b_proc == NULL) || (bp->b_proc->p_as == &kas))))
1027c478bd9Sstevel@tonic-gate return (bp->b_un.b_addr);
1037c478bd9Sstevel@tonic-gate
1047c478bd9Sstevel@tonic-gate ASSERT((bp->b_flags & (B_PAGEIO | B_PHYS)) != (B_PAGEIO | B_PHYS));
1057c478bd9Sstevel@tonic-gate
1062d3bf78eScth addr = (caddr_t)bp->b_un.b_addr;
1072d3bf78eScth off = (uintptr_t)addr & PAGEOFFSET;
1082d3bf78eScth size = P2ROUNDUP(bp->b_bcount + off, PAGESIZE);
1092d3bf78eScth npages = btop(size);
1102d3bf78eScth
1112d3bf78eScth /* Fastpath single page IO to locked memory by using kpm. */
1122d3bf78eScth if ((bp->b_flags & (B_SHADOW | B_PAGEIO)) && (npages == 1) &&
1132d3bf78eScth kpm_enable && bp_mapin_kpm_enable) {
1142d3bf78eScth if (bp->b_flags & B_SHADOW)
1152d3bf78eScth pp = *bp->b_shadow;
1162d3bf78eScth else
1172d3bf78eScth pp = bp->b_pages;
1182d3bf78eScth kaddr = hat_kpm_mapin(pp, NULL);
1192d3bf78eScth bp->b_un.b_addr = kaddr + off;
1202d3bf78eScth bp->b_flags |= B_REMAPPED;
1212d3bf78eScth return (bp->b_un.b_addr);
1222d3bf78eScth }
1232d3bf78eScth
1247c478bd9Sstevel@tonic-gate /*
1257c478bd9Sstevel@tonic-gate * Allocate kernel virtual space for remapping.
1267c478bd9Sstevel@tonic-gate */
1277c478bd9Sstevel@tonic-gate color = bp_color(bp);
1287c478bd9Sstevel@tonic-gate ASSERT(color < bp_align);
1297c478bd9Sstevel@tonic-gate
1307c478bd9Sstevel@tonic-gate if (bp_map_arena != NULL) {
1317c478bd9Sstevel@tonic-gate kaddr = (caddr_t)vmem_alloc(bp_map_arena,
1327c478bd9Sstevel@tonic-gate P2ROUNDUP(color + size, bp_align), flag);
1337c478bd9Sstevel@tonic-gate if (kaddr == NULL)
1347c478bd9Sstevel@tonic-gate return (NULL);
1357c478bd9Sstevel@tonic-gate kaddr += color;
1367c478bd9Sstevel@tonic-gate } else {
1377c478bd9Sstevel@tonic-gate kaddr = vmem_xalloc(heap_arena, size, bp_align, color,
1387c478bd9Sstevel@tonic-gate 0, NULL, NULL, flag);
1397c478bd9Sstevel@tonic-gate if (kaddr == NULL)
1407c478bd9Sstevel@tonic-gate return (NULL);
1417c478bd9Sstevel@tonic-gate }
1427c478bd9Sstevel@tonic-gate
1437c478bd9Sstevel@tonic-gate ASSERT(P2PHASE((uintptr_t)kaddr, bp_align) == color);
1447c478bd9Sstevel@tonic-gate
1457c478bd9Sstevel@tonic-gate /*
1467c478bd9Sstevel@tonic-gate * Map bp into the virtual space we just allocated.
1477c478bd9Sstevel@tonic-gate */
1487c478bd9Sstevel@tonic-gate if (bp->b_flags & B_PAGEIO) {
1497c478bd9Sstevel@tonic-gate pp = bp->b_pages;
1502d3bf78eScth pplist = NULL;
1517c478bd9Sstevel@tonic-gate } else if (bp->b_flags & B_SHADOW) {
1522d3bf78eScth pp = NULL;
1537c478bd9Sstevel@tonic-gate pplist = bp->b_shadow;
1547c478bd9Sstevel@tonic-gate } else {
1552d3bf78eScth pp = NULL;
1562d3bf78eScth pplist = NULL;
1577c478bd9Sstevel@tonic-gate if (bp->b_proc == NULL || (as = bp->b_proc->p_as) == NULL)
1587c478bd9Sstevel@tonic-gate as = &kas;
1597c478bd9Sstevel@tonic-gate }
1607c478bd9Sstevel@tonic-gate
1617c478bd9Sstevel@tonic-gate bp->b_flags |= B_REMAPPED;
1627c478bd9Sstevel@tonic-gate bp->b_un.b_addr = kaddr + off;
1637c478bd9Sstevel@tonic-gate
1647c478bd9Sstevel@tonic-gate while (npages-- != 0) {
1657c478bd9Sstevel@tonic-gate if (pp) {
1667c478bd9Sstevel@tonic-gate pfnum = pp->p_pagenum;
1677c478bd9Sstevel@tonic-gate pp = pp->p_next;
1687c478bd9Sstevel@tonic-gate } else if (pplist == NULL) {
1692d3bf78eScth pfnum = hat_getpfnum(as->a_hat,
1702d3bf78eScth (caddr_t)((uintptr_t)addr & MMU_PAGEMASK));
1712d3bf78eScth if (pfnum == PFN_INVALID)
1727c478bd9Sstevel@tonic-gate panic("bp_mapin_common: hat_getpfnum for"
1737c478bd9Sstevel@tonic-gate " addr %p failed\n", (void *)addr);
1747c478bd9Sstevel@tonic-gate addr += PAGESIZE;
1757c478bd9Sstevel@tonic-gate } else {
1767c478bd9Sstevel@tonic-gate pfnum = (*pplist)->p_pagenum;
1777c478bd9Sstevel@tonic-gate pplist++;
1787c478bd9Sstevel@tonic-gate }
1797c478bd9Sstevel@tonic-gate
1807c478bd9Sstevel@tonic-gate hat_devload(kas.a_hat, kaddr, PAGESIZE, pfnum,
1817c478bd9Sstevel@tonic-gate bp_devload_flags, HAT_LOAD_LOCK);
1827c478bd9Sstevel@tonic-gate
1837c478bd9Sstevel@tonic-gate kaddr += PAGESIZE;
1847c478bd9Sstevel@tonic-gate }
1857c478bd9Sstevel@tonic-gate return (bp->b_un.b_addr);
1867c478bd9Sstevel@tonic-gate }
1877c478bd9Sstevel@tonic-gate
1887c478bd9Sstevel@tonic-gate /*
1897c478bd9Sstevel@tonic-gate * Convert bp for pageio/physio to a kernel addressable location.
1907c478bd9Sstevel@tonic-gate */
1917c478bd9Sstevel@tonic-gate void
bp_mapin(struct buf * bp)1927c478bd9Sstevel@tonic-gate bp_mapin(struct buf *bp)
1937c478bd9Sstevel@tonic-gate {
1947c478bd9Sstevel@tonic-gate (void) bp_mapin_common(bp, VM_SLEEP);
1957c478bd9Sstevel@tonic-gate }
1967c478bd9Sstevel@tonic-gate
1977c478bd9Sstevel@tonic-gate /*
1987c478bd9Sstevel@tonic-gate * Release all the resources associated with a previous bp_mapin() call.
1997c478bd9Sstevel@tonic-gate */
2007c478bd9Sstevel@tonic-gate void
bp_mapout(struct buf * bp)2017c478bd9Sstevel@tonic-gate bp_mapout(struct buf *bp)
2027c478bd9Sstevel@tonic-gate {
2032d3bf78eScth caddr_t addr;
2042d3bf78eScth uintptr_t off;
2052d3bf78eScth uintptr_t base;
2062d3bf78eScth uintptr_t color;
2072d3bf78eScth size_t size;
2082d3bf78eScth pgcnt_t npages;
2092d3bf78eScth page_t *pp;
2102d3bf78eScth
2112d3bf78eScth if ((bp->b_flags & B_REMAPPED) == 0)
2122d3bf78eScth return;
2132d3bf78eScth
2142d3bf78eScth addr = bp->b_un.b_addr;
2152d3bf78eScth off = (uintptr_t)addr & PAGEOFFSET;
2162d3bf78eScth size = P2ROUNDUP(bp->b_bcount + off, PAGESIZE);
2172d3bf78eScth npages = btop(size);
2182d3bf78eScth
2197c478bd9Sstevel@tonic-gate bp->b_un.b_addr = (caddr_t)off; /* debugging aid */
2202d3bf78eScth
2212d3bf78eScth if ((bp->b_flags & (B_SHADOW | B_PAGEIO)) && (npages == 1) &&
2222d3bf78eScth kpm_enable && bp_mapin_kpm_enable) {
2232d3bf78eScth if (bp->b_flags & B_SHADOW)
2242d3bf78eScth pp = *bp->b_shadow;
2252d3bf78eScth else
2262d3bf78eScth pp = bp->b_pages;
2277a1b360bScth addr = (caddr_t)((uintptr_t)addr & MMU_PAGEMASK);
2282d3bf78eScth hat_kpm_mapout(pp, NULL, addr);
2292d3bf78eScth bp->b_flags &= ~B_REMAPPED;
2302d3bf78eScth return;
2312d3bf78eScth }
2322d3bf78eScth
2332d3bf78eScth base = (uintptr_t)addr & MMU_PAGEMASK;
2347c478bd9Sstevel@tonic-gate BP_FLUSH(base, size);
2357c478bd9Sstevel@tonic-gate hat_unload(kas.a_hat, (void *)base, size,
2367c478bd9Sstevel@tonic-gate HAT_UNLOAD_NOSYNC | HAT_UNLOAD_UNLOCK);
2372d3bf78eScth if (bp_map_arena != NULL) {
2382d3bf78eScth color = P2PHASE(base, bp_align);
2397c478bd9Sstevel@tonic-gate vmem_free(bp_map_arena, (void *)(base - color),
2407c478bd9Sstevel@tonic-gate P2ROUNDUP(color + size, bp_align));
2412d3bf78eScth } else
2427c478bd9Sstevel@tonic-gate vmem_free(heap_arena, (void *)base, size);
2437c478bd9Sstevel@tonic-gate bp->b_flags &= ~B_REMAPPED;
2447c478bd9Sstevel@tonic-gate }
245*36945f79Smrj
246*36945f79Smrj /*
247*36945f79Smrj * copy data from a KVA into a buf_t which may not be mapped in. offset
248*36945f79Smrj * is relative to the buf_t only.
249*36945f79Smrj */
250*36945f79Smrj int
bp_copyout(void * driverbuf,struct buf * bp,offset_t offset,size_t size)251*36945f79Smrj bp_copyout(void *driverbuf, struct buf *bp, offset_t offset, size_t size)
252*36945f79Smrj {
253*36945f79Smrj return (bp_copy_common(BP_COPYOUT, bp, driverbuf, offset, size));
254*36945f79Smrj }
255*36945f79Smrj
256*36945f79Smrj /*
257*36945f79Smrj * copy data from a buf_t which may not be mapped in, into a KVA.. offset
258*36945f79Smrj * is relative to the buf_t only.
259*36945f79Smrj */
260*36945f79Smrj int
bp_copyin(struct buf * bp,void * driverbuf,offset_t offset,size_t size)261*36945f79Smrj bp_copyin(struct buf *bp, void *driverbuf, offset_t offset, size_t size)
262*36945f79Smrj {
263*36945f79Smrj return (bp_copy_common(BP_COPYIN, bp, driverbuf, offset, size));
264*36945f79Smrj }
265*36945f79Smrj
266*36945f79Smrj
267*36945f79Smrj #define BP_COPY(dir, driverbuf, baddr, sz) \
268*36945f79Smrj (dir == BP_COPYIN) ? \
269*36945f79Smrj bcopy(baddr, driverbuf, sz) : bcopy(driverbuf, baddr, sz)
270*36945f79Smrj
271*36945f79Smrj static int
bp_copy_common(bp_copydir_t dir,struct buf * bp,void * driverbuf,offset_t offset,size_t size)272*36945f79Smrj bp_copy_common(bp_copydir_t dir, struct buf *bp, void *driverbuf,
273*36945f79Smrj offset_t offset, size_t size)
274*36945f79Smrj {
275*36945f79Smrj page_t **pplist;
276*36945f79Smrj uintptr_t poff;
277*36945f79Smrj uintptr_t voff;
278*36945f79Smrj struct as *as;
279*36945f79Smrj caddr_t kaddr;
280*36945f79Smrj caddr_t addr;
281*36945f79Smrj page_t *page;
282*36945f79Smrj size_t psize;
283*36945f79Smrj page_t *pp;
284*36945f79Smrj pfn_t pfn;
285*36945f79Smrj
286*36945f79Smrj
287*36945f79Smrj ASSERT((offset + size) <= bp->b_bcount);
288*36945f79Smrj
289*36945f79Smrj /* if the buf_t already has a KVA, just do a bcopy */
290*36945f79Smrj if (!(bp->b_flags & (B_PHYS | B_PAGEIO))) {
291*36945f79Smrj BP_COPY(dir, driverbuf, bp->b_un.b_addr + offset, size);
292*36945f79Smrj return (0);
293*36945f79Smrj }
294*36945f79Smrj
295*36945f79Smrj /* if we don't have kpm enabled, we need to do the slow path */
296*36945f79Smrj if (!kpm_enable || bp_force_copy) {
297*36945f79Smrj bp_mapin(bp);
298*36945f79Smrj BP_COPY(dir, driverbuf, bp->b_un.b_addr + offset, size);
299*36945f79Smrj bp_mapout(bp);
300*36945f79Smrj return (0);
301*36945f79Smrj }
302*36945f79Smrj
303*36945f79Smrj /*
304*36945f79Smrj * kpm is enabled, and we need to map in the buf_t for the copy
305*36945f79Smrj */
306*36945f79Smrj
307*36945f79Smrj /* setup pp, plist, and make sure 'as' is right */
308*36945f79Smrj if (bp->b_flags & B_PAGEIO) {
309*36945f79Smrj pp = bp->b_pages;
310*36945f79Smrj pplist = NULL;
311*36945f79Smrj } else if (bp->b_flags & B_SHADOW) {
312*36945f79Smrj pp = NULL;
313*36945f79Smrj pplist = bp->b_shadow;
314*36945f79Smrj } else {
315*36945f79Smrj pp = NULL;
316*36945f79Smrj pplist = NULL;
317*36945f79Smrj if (bp->b_proc == NULL || (as = bp->b_proc->p_as) == NULL) {
318*36945f79Smrj as = &kas;
319*36945f79Smrj }
320*36945f79Smrj }
321*36945f79Smrj
322*36945f79Smrj /*
323*36945f79Smrj * locals for the address, the offset into the first page, and the
324*36945f79Smrj * size of the first page we are going to copy.
325*36945f79Smrj */
326*36945f79Smrj addr = (caddr_t)bp->b_un.b_addr;
327*36945f79Smrj poff = (uintptr_t)addr & PAGEOFFSET;
328*36945f79Smrj psize = MIN(PAGESIZE - poff, size);
329*36945f79Smrj
330*36945f79Smrj /*
331*36945f79Smrj * we always start with a 0 offset into the driverbuf provided. The
332*36945f79Smrj * offset passed in only applies to the buf_t.
333*36945f79Smrj */
334*36945f79Smrj voff = 0;
335*36945f79Smrj
336*36945f79Smrj /* Loop until we've copied al the data */
337*36945f79Smrj while (size > 0) {
338*36945f79Smrj
339*36945f79Smrj /*
340*36945f79Smrj * for a pp or pplist, get the pfn, then go to the next page_t
341*36945f79Smrj * for the next time around the loop.
342*36945f79Smrj */
343*36945f79Smrj if (pp) {
344*36945f79Smrj page = pp;
345*36945f79Smrj pp = pp->p_next;
346*36945f79Smrj } else if (pplist != NULL) {
347*36945f79Smrj page = (*pplist);
348*36945f79Smrj pplist++;
349*36945f79Smrj
350*36945f79Smrj /*
351*36945f79Smrj * We have a user VA. If we are going to copy this page, (e.g.
352*36945f79Smrj * the offset into the buf_t where we start to copy is
353*36945f79Smrj * within this page), get the pfn. Don't waste the cycles
354*36945f79Smrj * getting the pfn if we're not copying this page.
355*36945f79Smrj */
356*36945f79Smrj } else if (offset < psize) {
357*36945f79Smrj pfn = hat_getpfnum(as->a_hat,
358*36945f79Smrj (caddr_t)((uintptr_t)addr & PAGEMASK));
359*36945f79Smrj if (pfn == PFN_INVALID) {
360*36945f79Smrj return (-1);
361*36945f79Smrj }
362*36945f79Smrj page = page_numtopp_nolock(pfn);
363*36945f79Smrj addr += psize - offset;
364*36945f79Smrj } else {
365*36945f79Smrj addr += psize;
366*36945f79Smrj }
367*36945f79Smrj
368*36945f79Smrj /*
369*36945f79Smrj * if we have an initial offset into the buf_t passed in,
370*36945f79Smrj * and it falls within the current page, account for it in
371*36945f79Smrj * the page size (how much we will copy) and the offset into the
372*36945f79Smrj * page (where we'll start copying from).
373*36945f79Smrj */
374*36945f79Smrj if ((offset > 0) && (offset < psize)) {
375*36945f79Smrj psize -= offset;
376*36945f79Smrj poff += offset;
377*36945f79Smrj offset = 0;
378*36945f79Smrj
379*36945f79Smrj /*
380*36945f79Smrj * if we have an initial offset into the buf_t passed in,
381*36945f79Smrj * and it's not within the current page, skip this page.
382*36945f79Smrj * We don't have to worry about the first page offset and size
383*36945f79Smrj * anymore. psize will normally be PAGESIZE now unless we are
384*36945f79Smrj * on the last page.
385*36945f79Smrj */
386*36945f79Smrj } else if (offset >= psize) {
387*36945f79Smrj offset -= psize;
388*36945f79Smrj psize = MIN(PAGESIZE, size);
389*36945f79Smrj poff = 0;
390*36945f79Smrj continue;
391*36945f79Smrj }
392*36945f79Smrj
393*36945f79Smrj /*
394*36945f79Smrj * get a kpm mapping to the page, them copy in/out of the
395*36945f79Smrj * page. update size left and offset into the driverbuf passed
396*36945f79Smrj * in for the next time around the loop.
397*36945f79Smrj */
398*36945f79Smrj kaddr = hat_kpm_mapin(page, NULL) + poff;
399*36945f79Smrj BP_COPY(dir, (void *)((uintptr_t)driverbuf + voff), kaddr,
400*36945f79Smrj psize);
401*36945f79Smrj hat_kpm_mapout(page, NULL, kaddr - poff);
402*36945f79Smrj
403*36945f79Smrj size -= psize;
404*36945f79Smrj voff += psize;
405*36945f79Smrj
406*36945f79Smrj poff = 0;
407*36945f79Smrj psize = MIN(PAGESIZE, size);
408*36945f79Smrj }
409*36945f79Smrj
410*36945f79Smrj return (0);
411*36945f79Smrj }
412