xref: /titanic_51/usr/src/uts/sun4/vm/sfmmu.c (revision 1e2e7a75ddb1eedcefa449ce98fd5862749b72ee)
17c478bd9Sstevel@tonic-gate /*
27c478bd9Sstevel@tonic-gate  * CDDL HEADER START
37c478bd9Sstevel@tonic-gate  *
47c478bd9Sstevel@tonic-gate  * The contents of this file are subject to the terms of the
5*1e2e7a75Shuah  * Common Development and Distribution License (the "License").
6*1e2e7a75Shuah  * You may not use this file except in compliance with the License.
77c478bd9Sstevel@tonic-gate  *
87c478bd9Sstevel@tonic-gate  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
97c478bd9Sstevel@tonic-gate  * or http://www.opensolaris.org/os/licensing.
107c478bd9Sstevel@tonic-gate  * See the License for the specific language governing permissions
117c478bd9Sstevel@tonic-gate  * and limitations under the License.
127c478bd9Sstevel@tonic-gate  *
137c478bd9Sstevel@tonic-gate  * When distributing Covered Code, include this CDDL HEADER in each
147c478bd9Sstevel@tonic-gate  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
157c478bd9Sstevel@tonic-gate  * If applicable, add the following below this CDDL HEADER, with the
167c478bd9Sstevel@tonic-gate  * fields enclosed by brackets "[]" replaced with your own identifying
177c478bd9Sstevel@tonic-gate  * information: Portions Copyright [yyyy] [name of copyright owner]
187c478bd9Sstevel@tonic-gate  *
197c478bd9Sstevel@tonic-gate  * CDDL HEADER END
207c478bd9Sstevel@tonic-gate  */
217c478bd9Sstevel@tonic-gate /*
22*1e2e7a75Shuah  * Copyright 2006 Sun Microsystems, Inc.  All rights reserved.
237c478bd9Sstevel@tonic-gate  * Use is subject to license terms.
247c478bd9Sstevel@tonic-gate  */
257c478bd9Sstevel@tonic-gate 
267c478bd9Sstevel@tonic-gate #pragma ident	"%Z%%M%	%I%	%E% SMI"
277c478bd9Sstevel@tonic-gate 
287c478bd9Sstevel@tonic-gate #include <sys/types.h>
297c478bd9Sstevel@tonic-gate #include <vm/hat.h>
307c478bd9Sstevel@tonic-gate #include <vm/hat_sfmmu.h>
317c478bd9Sstevel@tonic-gate #include <vm/page.h>
327c478bd9Sstevel@tonic-gate #include <sys/pte.h>
337c478bd9Sstevel@tonic-gate #include <sys/systm.h>
347c478bd9Sstevel@tonic-gate #include <sys/mman.h>
357c478bd9Sstevel@tonic-gate #include <sys/sysmacros.h>
367c478bd9Sstevel@tonic-gate #include <sys/machparam.h>
377c478bd9Sstevel@tonic-gate #include <sys/vtrace.h>
387c478bd9Sstevel@tonic-gate #include <sys/kmem.h>
397c478bd9Sstevel@tonic-gate #include <sys/mmu.h>
407c478bd9Sstevel@tonic-gate #include <sys/cmn_err.h>
417c478bd9Sstevel@tonic-gate #include <sys/cpu.h>
427c478bd9Sstevel@tonic-gate #include <sys/cpuvar.h>
437c478bd9Sstevel@tonic-gate #include <sys/debug.h>
447c478bd9Sstevel@tonic-gate #include <sys/lgrp.h>
457c478bd9Sstevel@tonic-gate #include <sys/archsystm.h>
467c478bd9Sstevel@tonic-gate #include <sys/machsystm.h>
477c478bd9Sstevel@tonic-gate #include <sys/vmsystm.h>
487c478bd9Sstevel@tonic-gate #include <sys/bitmap.h>
497c478bd9Sstevel@tonic-gate #include <vm/as.h>
507c478bd9Sstevel@tonic-gate #include <vm/seg.h>
517c478bd9Sstevel@tonic-gate #include <vm/seg_kmem.h>
527c478bd9Sstevel@tonic-gate #include <vm/seg_kp.h>
537c478bd9Sstevel@tonic-gate #include <vm/seg_kpm.h>
547c478bd9Sstevel@tonic-gate #include <vm/rm.h>
557c478bd9Sstevel@tonic-gate #include <vm/vm_dep.h>
567c478bd9Sstevel@tonic-gate #include <sys/t_lock.h>
577c478bd9Sstevel@tonic-gate #include <sys/vm_machparam.h>
587c478bd9Sstevel@tonic-gate #include <sys/promif.h>
597c478bd9Sstevel@tonic-gate #include <sys/prom_isa.h>
607c478bd9Sstevel@tonic-gate #include <sys/prom_plat.h>
617c478bd9Sstevel@tonic-gate #include <sys/prom_debug.h>
627c478bd9Sstevel@tonic-gate #include <sys/privregs.h>
637c478bd9Sstevel@tonic-gate #include <sys/bootconf.h>
647c478bd9Sstevel@tonic-gate #include <sys/memlist.h>
657c478bd9Sstevel@tonic-gate #include <sys/memlist_plat.h>
667c478bd9Sstevel@tonic-gate #include <sys/cpu_module.h>
677c478bd9Sstevel@tonic-gate #include <sys/reboot.h>
687c478bd9Sstevel@tonic-gate #include <sys/kdi.h>
697c478bd9Sstevel@tonic-gate 
707c478bd9Sstevel@tonic-gate /*
717c478bd9Sstevel@tonic-gate  * Static routines
727c478bd9Sstevel@tonic-gate  */
737c478bd9Sstevel@tonic-gate static void	sfmmu_map_prom_mappings(struct translation *, size_t);
747c478bd9Sstevel@tonic-gate static struct translation *read_prom_mappings(size_t *);
757c478bd9Sstevel@tonic-gate static void	sfmmu_reloc_trap_handler(void *, void *, size_t);
767c478bd9Sstevel@tonic-gate 
777c478bd9Sstevel@tonic-gate /*
787c478bd9Sstevel@tonic-gate  * External routines
797c478bd9Sstevel@tonic-gate  */
807c478bd9Sstevel@tonic-gate extern void sfmmu_remap_kernel(void);
817c478bd9Sstevel@tonic-gate extern void sfmmu_patch_utsb(void);
827c478bd9Sstevel@tonic-gate 
837c478bd9Sstevel@tonic-gate /*
847c478bd9Sstevel@tonic-gate  * Global Data:
857c478bd9Sstevel@tonic-gate  */
867c478bd9Sstevel@tonic-gate extern caddr_t	textva, datava;
877c478bd9Sstevel@tonic-gate extern tte_t	ktext_tte, kdata_tte;	/* ttes for kernel text and data */
887c478bd9Sstevel@tonic-gate extern int	enable_bigktsb;
897c478bd9Sstevel@tonic-gate 
907c478bd9Sstevel@tonic-gate uint64_t memsegspa = (uintptr_t)MSEG_NULLPTR_PA; /* memsegs physical linkage */
917c478bd9Sstevel@tonic-gate uint64_t memseg_phash[N_MEM_SLOTS];	/* use physical memseg addresses */
927c478bd9Sstevel@tonic-gate 
937c478bd9Sstevel@tonic-gate int	sfmmu_kern_mapped = 0;
947c478bd9Sstevel@tonic-gate 
957c478bd9Sstevel@tonic-gate /*
967c478bd9Sstevel@tonic-gate  * DMMU primary context register for the kernel context. Machine specific code
977c478bd9Sstevel@tonic-gate  * inserts correct page size codes when necessary
987c478bd9Sstevel@tonic-gate  */
997c478bd9Sstevel@tonic-gate uint64_t kcontextreg = KCONTEXT;
1007c478bd9Sstevel@tonic-gate 
1017c478bd9Sstevel@tonic-gate /* Extern Global Data */
1027c478bd9Sstevel@tonic-gate 
1037c478bd9Sstevel@tonic-gate extern int page_relocate_ready;
1047c478bd9Sstevel@tonic-gate 
1057c478bd9Sstevel@tonic-gate /*
1067c478bd9Sstevel@tonic-gate  * Controls the logic which enables the use of the
1077c478bd9Sstevel@tonic-gate  * QUAD_LDD_PHYS ASI for TSB accesses.
1087c478bd9Sstevel@tonic-gate  */
1097c478bd9Sstevel@tonic-gate extern int	ktsb_phys;
1107c478bd9Sstevel@tonic-gate 
1117c478bd9Sstevel@tonic-gate /*
1127c478bd9Sstevel@tonic-gate  * Global Routines called from within:
1137c478bd9Sstevel@tonic-gate  *	usr/src/uts/sun4u
1147c478bd9Sstevel@tonic-gate  *	usr/src/uts/sfmmu
1157c478bd9Sstevel@tonic-gate  *	usr/src/uts/sun
1167c478bd9Sstevel@tonic-gate  */
1177c478bd9Sstevel@tonic-gate 
1187c478bd9Sstevel@tonic-gate pfn_t
1197c478bd9Sstevel@tonic-gate va_to_pfn(void *vaddr)
1207c478bd9Sstevel@tonic-gate {
1217c478bd9Sstevel@tonic-gate 	u_longlong_t physaddr;
1227c478bd9Sstevel@tonic-gate 	int mode, valid;
1237c478bd9Sstevel@tonic-gate 
1247c478bd9Sstevel@tonic-gate 	if (tba_taken_over)
1257c478bd9Sstevel@tonic-gate 		return (hat_getpfnum(kas.a_hat, (caddr_t)vaddr));
1267c478bd9Sstevel@tonic-gate 
1277c478bd9Sstevel@tonic-gate 	if ((prom_translate_virt(vaddr, &valid, &physaddr, &mode) != -1) &&
1287c478bd9Sstevel@tonic-gate 	    (valid == -1)) {
1297c478bd9Sstevel@tonic-gate 		return ((pfn_t)(physaddr >> MMU_PAGESHIFT));
1307c478bd9Sstevel@tonic-gate 	}
1317c478bd9Sstevel@tonic-gate 	return (PFN_INVALID);
1327c478bd9Sstevel@tonic-gate }
1337c478bd9Sstevel@tonic-gate 
1347c478bd9Sstevel@tonic-gate uint64_t
1357c478bd9Sstevel@tonic-gate va_to_pa(void *vaddr)
1367c478bd9Sstevel@tonic-gate {
1377c478bd9Sstevel@tonic-gate 	pfn_t pfn;
1387c478bd9Sstevel@tonic-gate 
1397c478bd9Sstevel@tonic-gate 	if ((pfn = va_to_pfn(vaddr)) == PFN_INVALID)
1407c478bd9Sstevel@tonic-gate 		return ((uint64_t)-1);
1417c478bd9Sstevel@tonic-gate 	return (((uint64_t)pfn << MMU_PAGESHIFT) |
1427c478bd9Sstevel@tonic-gate 		((uint64_t)vaddr & MMU_PAGEOFFSET));
1437c478bd9Sstevel@tonic-gate }
1447c478bd9Sstevel@tonic-gate 
1457c478bd9Sstevel@tonic-gate void
1467c478bd9Sstevel@tonic-gate hat_kern_setup(void)
1477c478bd9Sstevel@tonic-gate {
1487c478bd9Sstevel@tonic-gate 	struct translation *trans_root;
1497c478bd9Sstevel@tonic-gate 	size_t ntrans_root;
1507c478bd9Sstevel@tonic-gate 	extern void startup_fixup_physavail(void);
1517c478bd9Sstevel@tonic-gate 
1527c478bd9Sstevel@tonic-gate 	/*
1537c478bd9Sstevel@tonic-gate 	 * These are the steps we take to take over the mmu from the prom.
1547c478bd9Sstevel@tonic-gate 	 *
1557c478bd9Sstevel@tonic-gate 	 * (1)	Read the prom's mappings through the translation property.
1567c478bd9Sstevel@tonic-gate 	 * (2)	Remap the kernel text and kernel data with 2 locked 4MB ttes.
1577c478bd9Sstevel@tonic-gate 	 *	Create the the hmeblks for these 2 ttes at this time.
1587c478bd9Sstevel@tonic-gate 	 * (3)	Create hat structures for all other prom mappings.  Since the
1597c478bd9Sstevel@tonic-gate 	 *	kernel text and data hme_blks have already been created we
1607c478bd9Sstevel@tonic-gate 	 *	skip the equivalent prom's mappings.
1617c478bd9Sstevel@tonic-gate 	 * (4)	Initialize the tsb and its corresponding hardware regs.
1627c478bd9Sstevel@tonic-gate 	 * (5)	Take over the trap table (currently in startup).
1637c478bd9Sstevel@tonic-gate 	 * (6)	Up to this point it is possible the prom required some of its
1647c478bd9Sstevel@tonic-gate 	 *	locked tte's.  Now that we own the trap table we remove them.
1657c478bd9Sstevel@tonic-gate 	 */
1667c478bd9Sstevel@tonic-gate 
1677c478bd9Sstevel@tonic-gate 	ktsb_pbase = va_to_pa(ktsb_base);
1687c478bd9Sstevel@tonic-gate 	ktsb4m_pbase = va_to_pa(ktsb4m_base);
1697c478bd9Sstevel@tonic-gate 	PRM_DEBUG(ktsb_pbase);
1707c478bd9Sstevel@tonic-gate 	PRM_DEBUG(ktsb4m_pbase);
1717c478bd9Sstevel@tonic-gate 
1727c478bd9Sstevel@tonic-gate 	sfmmu_setup_4lp();
1737c478bd9Sstevel@tonic-gate 	sfmmu_patch_ktsb();
1747c478bd9Sstevel@tonic-gate 	sfmmu_patch_utsb();
1757c478bd9Sstevel@tonic-gate 	sfmmu_patch_mmu_asi(ktsb_phys);
1767c478bd9Sstevel@tonic-gate 
1777c478bd9Sstevel@tonic-gate 	sfmmu_init_tsbs();
1787c478bd9Sstevel@tonic-gate 
1797c478bd9Sstevel@tonic-gate 	if (kpm_enable) {
1807c478bd9Sstevel@tonic-gate 		sfmmu_kpm_patch_tlbm();
1817c478bd9Sstevel@tonic-gate 		if (kpm_smallpages == 0) {
1827c478bd9Sstevel@tonic-gate 			sfmmu_kpm_patch_tsbm();
1837c478bd9Sstevel@tonic-gate 		}
1847c478bd9Sstevel@tonic-gate 	}
1857c478bd9Sstevel@tonic-gate 
1867c478bd9Sstevel@tonic-gate 	/*
1877c478bd9Sstevel@tonic-gate 	 * The 8K-indexed kernel TSB space is used to hold
1887c478bd9Sstevel@tonic-gate 	 * translations below...
1897c478bd9Sstevel@tonic-gate 	 */
1907c478bd9Sstevel@tonic-gate 	trans_root = read_prom_mappings(&ntrans_root);
1917c478bd9Sstevel@tonic-gate 	sfmmu_remap_kernel();
1927c478bd9Sstevel@tonic-gate 	startup_fixup_physavail();
1937c478bd9Sstevel@tonic-gate 	mmu_init_kernel_pgsz(kas.a_hat);
1947c478bd9Sstevel@tonic-gate 	sfmmu_map_prom_mappings(trans_root, ntrans_root);
1957c478bd9Sstevel@tonic-gate 
1967c478bd9Sstevel@tonic-gate 	/*
1977c478bd9Sstevel@tonic-gate 	 * We invalidate 8K kernel TSB because we used it in
1987c478bd9Sstevel@tonic-gate 	 * sfmmu_map_prom_mappings()
1997c478bd9Sstevel@tonic-gate 	 */
2007c478bd9Sstevel@tonic-gate 	sfmmu_inv_tsb(ktsb_base, ktsb_sz);
2017c478bd9Sstevel@tonic-gate 	sfmmu_inv_tsb(ktsb4m_base, ktsb4m_sz);
2027c478bd9Sstevel@tonic-gate 
2037c478bd9Sstevel@tonic-gate 	sfmmu_init_ktsbinfo();
2047c478bd9Sstevel@tonic-gate 
2057c478bd9Sstevel@tonic-gate 
2067c478bd9Sstevel@tonic-gate 	sfmmu_kern_mapped = 1;
2077c478bd9Sstevel@tonic-gate 
2087c478bd9Sstevel@tonic-gate 	/*
2097c478bd9Sstevel@tonic-gate 	 * hments have been created for mapped pages, and thus we're ready
2107c478bd9Sstevel@tonic-gate 	 * for kmdb to start using its own trap table.  It walks the hments
2117c478bd9Sstevel@tonic-gate 	 * to resolve TLB misses, and can't be used until they're ready.
2127c478bd9Sstevel@tonic-gate 	 */
2137c478bd9Sstevel@tonic-gate 	if (boothowto & RB_DEBUG)
2147c478bd9Sstevel@tonic-gate 		kdi_dvec_vmready();
2157c478bd9Sstevel@tonic-gate }
2167c478bd9Sstevel@tonic-gate 
2177c478bd9Sstevel@tonic-gate /*
2187c478bd9Sstevel@tonic-gate  * Macro used below to convert the prom's 32-bit high and low fields into
2197c478bd9Sstevel@tonic-gate  * a value appropriate for the 64-bit kernel.
2207c478bd9Sstevel@tonic-gate  */
2217c478bd9Sstevel@tonic-gate 
2227c478bd9Sstevel@tonic-gate #define	COMBINE(hi, lo) (((uint64_t)(uint32_t)(hi) << 32) | (uint32_t)(lo))
2237c478bd9Sstevel@tonic-gate 
2247c478bd9Sstevel@tonic-gate /*
2257c478bd9Sstevel@tonic-gate  * This function traverses the prom mapping list and creates equivalent
2267c478bd9Sstevel@tonic-gate  * mappings in the sfmmu mapping hash.
2277c478bd9Sstevel@tonic-gate  */
2287c478bd9Sstevel@tonic-gate static void
2297c478bd9Sstevel@tonic-gate sfmmu_map_prom_mappings(struct translation *trans_root, size_t ntrans_root)
2307c478bd9Sstevel@tonic-gate {
2317c478bd9Sstevel@tonic-gate 	struct translation *promt;
2327c478bd9Sstevel@tonic-gate 	tte_t	tte, oldtte, *ttep;
2337c478bd9Sstevel@tonic-gate 	pfn_t	pfn, oldpfn, basepfn;
2347c478bd9Sstevel@tonic-gate 	caddr_t vaddr;
2357c478bd9Sstevel@tonic-gate 	size_t	size, offset;
2367c478bd9Sstevel@tonic-gate 	unsigned long i;
2377c478bd9Sstevel@tonic-gate 	uint_t	attr;
2387c478bd9Sstevel@tonic-gate 	page_t *pp;
2397c478bd9Sstevel@tonic-gate 	extern struct memlist *virt_avail;
2407c478bd9Sstevel@tonic-gate 
2417c478bd9Sstevel@tonic-gate 	ttep = &tte;
2427c478bd9Sstevel@tonic-gate 	for (i = 0, promt = trans_root; i < ntrans_root; i++, promt++) {
2437c478bd9Sstevel@tonic-gate 		ASSERT(promt->tte_hi != 0);
2447c478bd9Sstevel@tonic-gate 		ASSERT32(promt->virt_hi == 0 && promt->size_hi == 0);
2457c478bd9Sstevel@tonic-gate 
2467c478bd9Sstevel@tonic-gate 		/*
2477c478bd9Sstevel@tonic-gate 		 * hack until we get rid of map-for-unix
2487c478bd9Sstevel@tonic-gate 		 */
2497c478bd9Sstevel@tonic-gate 		if (COMBINE(promt->virt_hi, promt->virt_lo) < KERNELBASE)
2507c478bd9Sstevel@tonic-gate 			continue;
2517c478bd9Sstevel@tonic-gate 
2527c478bd9Sstevel@tonic-gate 		ttep->tte_inthi = promt->tte_hi;
2537c478bd9Sstevel@tonic-gate 		ttep->tte_intlo = promt->tte_lo;
2547c478bd9Sstevel@tonic-gate 		attr = PROC_DATA | HAT_NOSYNC;
2557c478bd9Sstevel@tonic-gate #if defined(TTE_IS_GLOBAL)
2567c478bd9Sstevel@tonic-gate 		if (TTE_IS_GLOBAL(ttep)) {
2577c478bd9Sstevel@tonic-gate 			/*
2587c478bd9Sstevel@tonic-gate 			 * The prom better not use global translations
2597c478bd9Sstevel@tonic-gate 			 * because a user process might use the same
2607c478bd9Sstevel@tonic-gate 			 * virtual addresses
2617c478bd9Sstevel@tonic-gate 			 */
2627c478bd9Sstevel@tonic-gate 			cmn_err(CE_PANIC, "map_prom: global translation");
2637c478bd9Sstevel@tonic-gate 			TTE_SET_LOFLAGS(ttep, TTE_GLB_INT, 0);
2647c478bd9Sstevel@tonic-gate 		}
2657c478bd9Sstevel@tonic-gate #endif
2667c478bd9Sstevel@tonic-gate 		if (TTE_IS_LOCKED(ttep)) {
2677c478bd9Sstevel@tonic-gate 			/* clear the lock bits */
2687c478bd9Sstevel@tonic-gate 			TTE_CLR_LOCKED(ttep);
2697c478bd9Sstevel@tonic-gate 		}
2707c478bd9Sstevel@tonic-gate 		attr |= (TTE_IS_VCACHEABLE(ttep)) ? 0 : SFMMU_UNCACHEVTTE;
2717c478bd9Sstevel@tonic-gate 		attr |= (TTE_IS_PCACHEABLE(ttep)) ? 0 : SFMMU_UNCACHEPTTE;
2727c478bd9Sstevel@tonic-gate 		attr |= (TTE_IS_SIDEFFECT(ttep)) ? SFMMU_SIDEFFECT : 0;
2737c478bd9Sstevel@tonic-gate 		attr |= (TTE_IS_IE(ttep)) ? HAT_STRUCTURE_LE : 0;
2747c478bd9Sstevel@tonic-gate 
2757c478bd9Sstevel@tonic-gate 		size = COMBINE(promt->size_hi, promt->size_lo);
2767c478bd9Sstevel@tonic-gate 		offset = 0;
2777c478bd9Sstevel@tonic-gate 		basepfn = TTE_TO_PFN((caddr_t)COMBINE(promt->virt_hi,
2787c478bd9Sstevel@tonic-gate 		    promt->virt_lo), ttep);
2797c478bd9Sstevel@tonic-gate 		while (size) {
2807c478bd9Sstevel@tonic-gate 			vaddr = (caddr_t)(COMBINE(promt->virt_hi,
2817c478bd9Sstevel@tonic-gate 			    promt->virt_lo) + offset);
2827c478bd9Sstevel@tonic-gate 
2837c478bd9Sstevel@tonic-gate 			/*
2847c478bd9Sstevel@tonic-gate 			 * make sure address is not in virt-avail list
2857c478bd9Sstevel@tonic-gate 			 */
2867c478bd9Sstevel@tonic-gate 			if (address_in_memlist(virt_avail, (uint64_t)vaddr,
2877c478bd9Sstevel@tonic-gate 			    size)) {
2887c478bd9Sstevel@tonic-gate 				cmn_err(CE_PANIC, "map_prom: inconsistent "
2897c478bd9Sstevel@tonic-gate 				    "translation/avail lists");
2907c478bd9Sstevel@tonic-gate 			}
2917c478bd9Sstevel@tonic-gate 
2927c478bd9Sstevel@tonic-gate 			pfn = basepfn + mmu_btop(offset);
2937c478bd9Sstevel@tonic-gate 			if (pf_is_memory(pfn)) {
2947c478bd9Sstevel@tonic-gate 				if (attr & SFMMU_UNCACHEPTTE) {
2957c478bd9Sstevel@tonic-gate 					cmn_err(CE_PANIC, "map_prom: "
2967c478bd9Sstevel@tonic-gate 					    "uncached prom memory page");
2977c478bd9Sstevel@tonic-gate 				}
2987c478bd9Sstevel@tonic-gate 			} else {
2997c478bd9Sstevel@tonic-gate 				if (!(attr & SFMMU_SIDEFFECT)) {
3007c478bd9Sstevel@tonic-gate 					cmn_err(CE_PANIC, "map_prom: prom "
3017c478bd9Sstevel@tonic-gate 					    "i/o page without side-effect");
3027c478bd9Sstevel@tonic-gate 				}
3037c478bd9Sstevel@tonic-gate 			}
3047c478bd9Sstevel@tonic-gate 			oldpfn = sfmmu_vatopfn(vaddr, KHATID, &oldtte);
3057c478bd9Sstevel@tonic-gate 			ASSERT(oldpfn != PFN_SUSPENDED);
3067c478bd9Sstevel@tonic-gate 			ASSERT(page_relocate_ready == 0);
3077c478bd9Sstevel@tonic-gate 
3087c478bd9Sstevel@tonic-gate 			if (oldpfn != PFN_INVALID) {
3097c478bd9Sstevel@tonic-gate 				/*
3107c478bd9Sstevel@tonic-gate 				 * mapping already exists.
3117c478bd9Sstevel@tonic-gate 				 * Verify they are equal
3127c478bd9Sstevel@tonic-gate 				 */
3137c478bd9Sstevel@tonic-gate 				if (pfn != oldpfn) {
3147c478bd9Sstevel@tonic-gate 					cmn_err(CE_PANIC, "map_prom: mapping "
3157c478bd9Sstevel@tonic-gate 					    "conflict (va=0x%p pfn=%p, "
3167c478bd9Sstevel@tonic-gate 					    "oldpfn=%p)",
3177c478bd9Sstevel@tonic-gate 					    (void *)vaddr, (void *)pfn,
3187c478bd9Sstevel@tonic-gate 					    (void *)oldpfn);
3197c478bd9Sstevel@tonic-gate 				}
3207c478bd9Sstevel@tonic-gate 				size -= MMU_PAGESIZE;
3217c478bd9Sstevel@tonic-gate 				offset += MMU_PAGESIZE;
3227c478bd9Sstevel@tonic-gate 				continue;
3237c478bd9Sstevel@tonic-gate 			}
3247c478bd9Sstevel@tonic-gate 
3257c478bd9Sstevel@tonic-gate 			pp = page_numtopp_nolock(pfn);
3267c478bd9Sstevel@tonic-gate 			if ((pp != NULL) && PP_ISFREE((page_t *)pp)) {
3277c478bd9Sstevel@tonic-gate 				cmn_err(CE_PANIC, "map_prom: "
3287c478bd9Sstevel@tonic-gate 				    "prom-mapped page (va 0x%p, pfn 0x%p) "
3297c478bd9Sstevel@tonic-gate 				    "on free list", (void *)vaddr, (void *)pfn);
3307c478bd9Sstevel@tonic-gate 			}
3317c478bd9Sstevel@tonic-gate 
3327c478bd9Sstevel@tonic-gate 			sfmmu_memtte(ttep, pfn, attr, TTE8K);
3337c478bd9Sstevel@tonic-gate 			sfmmu_tteload(kas.a_hat, ttep, vaddr, pp,
3347c478bd9Sstevel@tonic-gate 			    HAT_LOAD_LOCK | SFMMU_NO_TSBLOAD);
3357c478bd9Sstevel@tonic-gate 			size -= MMU_PAGESIZE;
3367c478bd9Sstevel@tonic-gate 			offset += MMU_PAGESIZE;
3377c478bd9Sstevel@tonic-gate 		}
3387c478bd9Sstevel@tonic-gate 	}
3397c478bd9Sstevel@tonic-gate }
3407c478bd9Sstevel@tonic-gate 
3417c478bd9Sstevel@tonic-gate #undef COMBINE	/* local to previous routine */
3427c478bd9Sstevel@tonic-gate 
3437c478bd9Sstevel@tonic-gate /*
3447c478bd9Sstevel@tonic-gate  * This routine reads in the "translations" property in to a buffer and
3457c478bd9Sstevel@tonic-gate  * returns a pointer to this buffer and the number of translations.
3467c478bd9Sstevel@tonic-gate  */
3477c478bd9Sstevel@tonic-gate static struct translation *
3487c478bd9Sstevel@tonic-gate read_prom_mappings(size_t *ntransrootp)
3497c478bd9Sstevel@tonic-gate {
3507c478bd9Sstevel@tonic-gate 	char *prop = "translations";
3517c478bd9Sstevel@tonic-gate 	size_t translen;
352fa9e4066Sahrens 	pnode_t node;
3537c478bd9Sstevel@tonic-gate 	struct translation *transroot;
3547c478bd9Sstevel@tonic-gate 
3557c478bd9Sstevel@tonic-gate 	/*
3567c478bd9Sstevel@tonic-gate 	 * the "translations" property is associated with the mmu node
3577c478bd9Sstevel@tonic-gate 	 */
358fa9e4066Sahrens 	node = (pnode_t)prom_getphandle(prom_mmu_ihandle());
3597c478bd9Sstevel@tonic-gate 
3607c478bd9Sstevel@tonic-gate 	/*
3617c478bd9Sstevel@tonic-gate 	 * We use the TSB space to read in the prom mappings.  This space
3627c478bd9Sstevel@tonic-gate 	 * is currently not being used because we haven't taken over the
3637c478bd9Sstevel@tonic-gate 	 * trap table yet.  It should be big enough to hold the mappings.
3647c478bd9Sstevel@tonic-gate 	 */
3657c478bd9Sstevel@tonic-gate 	if ((translen = prom_getproplen(node, prop)) == -1)
3667c478bd9Sstevel@tonic-gate 		cmn_err(CE_PANIC, "no translations property");
3677c478bd9Sstevel@tonic-gate 	*ntransrootp = translen / sizeof (*transroot);
3687c478bd9Sstevel@tonic-gate 	translen = roundup(translen, MMU_PAGESIZE);
3697c478bd9Sstevel@tonic-gate 	PRM_DEBUG(translen);
3707c478bd9Sstevel@tonic-gate 	if (translen > TSB_BYTES(ktsb_szcode))
3717c478bd9Sstevel@tonic-gate 		cmn_err(CE_PANIC, "not enough space for translations");
3727c478bd9Sstevel@tonic-gate 
3737c478bd9Sstevel@tonic-gate 	transroot = (struct translation *)ktsb_base;
3747c478bd9Sstevel@tonic-gate 	ASSERT(transroot);
3757c478bd9Sstevel@tonic-gate 	if (prom_getprop(node, prop, (caddr_t)transroot) == -1) {
3767c478bd9Sstevel@tonic-gate 		cmn_err(CE_PANIC, "translations getprop failed");
3777c478bd9Sstevel@tonic-gate 	}
3787c478bd9Sstevel@tonic-gate 	return (transroot);
3797c478bd9Sstevel@tonic-gate }
3807c478bd9Sstevel@tonic-gate 
3817c478bd9Sstevel@tonic-gate /*
3827c478bd9Sstevel@tonic-gate  * Init routine of the nucleus data memory allocator.
3837c478bd9Sstevel@tonic-gate  *
3847c478bd9Sstevel@tonic-gate  * The nucleus data memory allocator is organized in ecache_alignsize'd
3857c478bd9Sstevel@tonic-gate  * memory chunks. Memory allocated by ndata_alloc() will never be freed.
3867c478bd9Sstevel@tonic-gate  *
3877c478bd9Sstevel@tonic-gate  * The ndata argument is used as header of the ndata freelist.
3887c478bd9Sstevel@tonic-gate  * Other freelist nodes are placed in the nucleus memory itself
3897c478bd9Sstevel@tonic-gate  * at the beginning of a free memory chunk. Therefore a freelist
3907c478bd9Sstevel@tonic-gate  * node (struct memlist) must fit into the smallest allocatable
3917c478bd9Sstevel@tonic-gate  * memory chunk (ecache_alignsize bytes).
3927c478bd9Sstevel@tonic-gate  *
3937c478bd9Sstevel@tonic-gate  * The memory interval [base, end] passed to ndata_alloc_init() must be
3947c478bd9Sstevel@tonic-gate  * bzero'd to allow the allocator to return bzero'd memory easily.
3957c478bd9Sstevel@tonic-gate  */
3967c478bd9Sstevel@tonic-gate void
3977c478bd9Sstevel@tonic-gate ndata_alloc_init(struct memlist *ndata, uintptr_t base, uintptr_t end)
3987c478bd9Sstevel@tonic-gate {
3997c478bd9Sstevel@tonic-gate 	ASSERT(sizeof (struct memlist) <= ecache_alignsize);
4007c478bd9Sstevel@tonic-gate 
4017c478bd9Sstevel@tonic-gate 	base = roundup(base, ecache_alignsize);
4027c478bd9Sstevel@tonic-gate 	end = end - end % ecache_alignsize;
4037c478bd9Sstevel@tonic-gate 
4047c478bd9Sstevel@tonic-gate 	ASSERT(base < end);
4057c478bd9Sstevel@tonic-gate 
4067c478bd9Sstevel@tonic-gate 	ndata->address = base;
4077c478bd9Sstevel@tonic-gate 	ndata->size = end - base;
4087c478bd9Sstevel@tonic-gate 	ndata->next = NULL;
4097c478bd9Sstevel@tonic-gate 	ndata->prev = NULL;
4107c478bd9Sstevel@tonic-gate }
4117c478bd9Sstevel@tonic-gate 
4127c478bd9Sstevel@tonic-gate /*
4137c478bd9Sstevel@tonic-gate  * Deliver the size of the largest free memory chunk.
4147c478bd9Sstevel@tonic-gate  */
4157c478bd9Sstevel@tonic-gate size_t
4167c478bd9Sstevel@tonic-gate ndata_maxsize(struct memlist *ndata)
4177c478bd9Sstevel@tonic-gate {
4187c478bd9Sstevel@tonic-gate 	size_t chunksize = ndata->size;
4197c478bd9Sstevel@tonic-gate 
4207c478bd9Sstevel@tonic-gate 	while ((ndata = ndata->next) != NULL) {
4217c478bd9Sstevel@tonic-gate 		if (chunksize < ndata->size)
4227c478bd9Sstevel@tonic-gate 			chunksize = ndata->size;
4237c478bd9Sstevel@tonic-gate 	}
4247c478bd9Sstevel@tonic-gate 
4257c478bd9Sstevel@tonic-gate 	return (chunksize);
4267c478bd9Sstevel@tonic-gate }
4277c478bd9Sstevel@tonic-gate 
4287c478bd9Sstevel@tonic-gate /*
4297c478bd9Sstevel@tonic-gate  * This is a special function to figure out if the memory chunk needed
4307c478bd9Sstevel@tonic-gate  * for the page structs can fit in the nucleus or not. If it fits the
4317c478bd9Sstevel@tonic-gate  * function calculates and returns the possible remaining ndata size
4327c478bd9Sstevel@tonic-gate  * in the last element if the size needed for page structs would be
4337c478bd9Sstevel@tonic-gate  * allocated from the nucleus.
4347c478bd9Sstevel@tonic-gate  */
4357c478bd9Sstevel@tonic-gate size_t
4367c478bd9Sstevel@tonic-gate ndata_spare(struct memlist *ndata, size_t wanted, size_t alignment)
4377c478bd9Sstevel@tonic-gate {
4387c478bd9Sstevel@tonic-gate 	struct memlist *frlist;
4397c478bd9Sstevel@tonic-gate 	uintptr_t base;
4407c478bd9Sstevel@tonic-gate 	uintptr_t end;
4417c478bd9Sstevel@tonic-gate 
4427c478bd9Sstevel@tonic-gate 	for (frlist = ndata; frlist != NULL; frlist = frlist->next) {
4437c478bd9Sstevel@tonic-gate 		base = roundup(frlist->address, alignment);
4447c478bd9Sstevel@tonic-gate 		end = roundup(base + wanted, ecache_alignsize);
4457c478bd9Sstevel@tonic-gate 
4467c478bd9Sstevel@tonic-gate 		if (end <= frlist->address + frlist->size) {
4477c478bd9Sstevel@tonic-gate 			if (frlist->next == NULL)
4487c478bd9Sstevel@tonic-gate 				return (frlist->address + frlist->size - end);
4497c478bd9Sstevel@tonic-gate 
4507c478bd9Sstevel@tonic-gate 			while (frlist->next != NULL)
4517c478bd9Sstevel@tonic-gate 				frlist = frlist->next;
4527c478bd9Sstevel@tonic-gate 
4537c478bd9Sstevel@tonic-gate 			return (frlist->size);
4547c478bd9Sstevel@tonic-gate 		}
4557c478bd9Sstevel@tonic-gate 	}
4567c478bd9Sstevel@tonic-gate 
4577c478bd9Sstevel@tonic-gate 	return (0);
4587c478bd9Sstevel@tonic-gate }
4597c478bd9Sstevel@tonic-gate 
4607c478bd9Sstevel@tonic-gate /*
4617c478bd9Sstevel@tonic-gate  * Allocate the last properly aligned memory chunk.
4627c478bd9Sstevel@tonic-gate  * This function is called when no more large nucleus memory chunks
4637c478bd9Sstevel@tonic-gate  * will be allocated.  The remaining free nucleus memory at the end
4647c478bd9Sstevel@tonic-gate  * of the nucleus can be added to the phys_avail list.
4657c478bd9Sstevel@tonic-gate  */
4667c478bd9Sstevel@tonic-gate void *
4677c478bd9Sstevel@tonic-gate ndata_extra_base(struct memlist *ndata, size_t alignment)
4687c478bd9Sstevel@tonic-gate {
4697c478bd9Sstevel@tonic-gate 	uintptr_t base;
4707c478bd9Sstevel@tonic-gate 	size_t wasteage = 0;
4717c478bd9Sstevel@tonic-gate #ifdef	DEBUG
4727c478bd9Sstevel@tonic-gate 	static int called = 0;
4737c478bd9Sstevel@tonic-gate 
4747c478bd9Sstevel@tonic-gate 	if (called++ > 0)
4757c478bd9Sstevel@tonic-gate 		cmn_err(CE_PANIC, "ndata_extra_base() called more than once");
4767c478bd9Sstevel@tonic-gate #endif /* DEBUG */
4777c478bd9Sstevel@tonic-gate 
4787c478bd9Sstevel@tonic-gate 	/*
4797c478bd9Sstevel@tonic-gate 	 * The alignment needs to be a multiple of ecache_alignsize.
4807c478bd9Sstevel@tonic-gate 	 */
4817c478bd9Sstevel@tonic-gate 	ASSERT((alignment % ecache_alignsize) ==  0);
4827c478bd9Sstevel@tonic-gate 
4837c478bd9Sstevel@tonic-gate 	while (ndata->next != NULL) {
4847c478bd9Sstevel@tonic-gate 		wasteage += ndata->size;
4857c478bd9Sstevel@tonic-gate 		ndata = ndata->next;
4867c478bd9Sstevel@tonic-gate 	}
4877c478bd9Sstevel@tonic-gate 
4887c478bd9Sstevel@tonic-gate 	base = roundup(ndata->address, alignment);
4897c478bd9Sstevel@tonic-gate 
4907c478bd9Sstevel@tonic-gate 	if (base >= ndata->address + ndata->size)
4917c478bd9Sstevel@tonic-gate 		return (NULL);
4927c478bd9Sstevel@tonic-gate 
4937c478bd9Sstevel@tonic-gate 	if (base == ndata->address) {
4947c478bd9Sstevel@tonic-gate 		if (ndata->prev != NULL)
4957c478bd9Sstevel@tonic-gate 			ndata->prev->next = NULL;
4967c478bd9Sstevel@tonic-gate 		else
4977c478bd9Sstevel@tonic-gate 			ndata->size = 0;
4987c478bd9Sstevel@tonic-gate 
4997c478bd9Sstevel@tonic-gate 		bzero((void *)base, sizeof (struct memlist));
5007c478bd9Sstevel@tonic-gate 
5017c478bd9Sstevel@tonic-gate 	} else {
5027c478bd9Sstevel@tonic-gate 		ndata->size = base - ndata->address;
5037c478bd9Sstevel@tonic-gate 		wasteage += ndata->size;
5047c478bd9Sstevel@tonic-gate 	}
5057c478bd9Sstevel@tonic-gate 	PRM_DEBUG(wasteage);
5067c478bd9Sstevel@tonic-gate 
5077c478bd9Sstevel@tonic-gate 	return ((void *)base);
5087c478bd9Sstevel@tonic-gate }
5097c478bd9Sstevel@tonic-gate 
5107c478bd9Sstevel@tonic-gate /*
5117c478bd9Sstevel@tonic-gate  * Select the best matching buffer, avoid memory fragmentation.
5127c478bd9Sstevel@tonic-gate  */
5137c478bd9Sstevel@tonic-gate static struct memlist *
5147c478bd9Sstevel@tonic-gate ndata_select_chunk(struct memlist *ndata, size_t wanted, size_t alignment)
5157c478bd9Sstevel@tonic-gate {
5167c478bd9Sstevel@tonic-gate 	struct memlist *fnd_below = NULL;
5177c478bd9Sstevel@tonic-gate 	struct memlist *fnd_above = NULL;
5187c478bd9Sstevel@tonic-gate 	struct memlist *fnd_unused = NULL;
5197c478bd9Sstevel@tonic-gate 	struct memlist *frlist;
5207c478bd9Sstevel@tonic-gate 	uintptr_t base;
5217c478bd9Sstevel@tonic-gate 	uintptr_t end;
5227c478bd9Sstevel@tonic-gate 	size_t below;
5237c478bd9Sstevel@tonic-gate 	size_t above;
5247c478bd9Sstevel@tonic-gate 	size_t unused;
5257c478bd9Sstevel@tonic-gate 	size_t best_below = ULONG_MAX;
5267c478bd9Sstevel@tonic-gate 	size_t best_above = ULONG_MAX;
5277c478bd9Sstevel@tonic-gate 	size_t best_unused = ULONG_MAX;
5287c478bd9Sstevel@tonic-gate 
5297c478bd9Sstevel@tonic-gate 	ASSERT(ndata != NULL);
5307c478bd9Sstevel@tonic-gate 
5317c478bd9Sstevel@tonic-gate 	/*
5327c478bd9Sstevel@tonic-gate 	 * Look for the best matching buffer, avoid memory fragmentation.
5337c478bd9Sstevel@tonic-gate 	 * The following strategy is used, try to find
5347c478bd9Sstevel@tonic-gate 	 *   1. an exact fitting buffer
5357c478bd9Sstevel@tonic-gate 	 *   2. avoid wasting any space below the buffer, take first
5367c478bd9Sstevel@tonic-gate 	 *	fitting buffer
5377c478bd9Sstevel@tonic-gate 	 *   3. avoid wasting any space above the buffer, take first
5387c478bd9Sstevel@tonic-gate 	 *	fitting buffer
5397c478bd9Sstevel@tonic-gate 	 *   4. avoid wasting space, take first fitting buffer
5407c478bd9Sstevel@tonic-gate 	 *   5. take the last buffer in chain
5417c478bd9Sstevel@tonic-gate 	 */
5427c478bd9Sstevel@tonic-gate 	for (frlist = ndata; frlist != NULL; frlist = frlist->next) {
5437c478bd9Sstevel@tonic-gate 		base = roundup(frlist->address, alignment);
5447c478bd9Sstevel@tonic-gate 		end = roundup(base + wanted, ecache_alignsize);
5457c478bd9Sstevel@tonic-gate 
5467c478bd9Sstevel@tonic-gate 		if (end > frlist->address + frlist->size)
5477c478bd9Sstevel@tonic-gate 			continue;
5487c478bd9Sstevel@tonic-gate 
5497c478bd9Sstevel@tonic-gate 		below = (base - frlist->address) / ecache_alignsize;
5507c478bd9Sstevel@tonic-gate 		above = (frlist->address + frlist->size - end) /
5517c478bd9Sstevel@tonic-gate 		    ecache_alignsize;
5527c478bd9Sstevel@tonic-gate 		unused = below + above;
5537c478bd9Sstevel@tonic-gate 
5547c478bd9Sstevel@tonic-gate 		if (unused == 0)
5557c478bd9Sstevel@tonic-gate 			return (frlist);
5567c478bd9Sstevel@tonic-gate 
5577c478bd9Sstevel@tonic-gate 		if (frlist->next == NULL)
5587c478bd9Sstevel@tonic-gate 			break;
5597c478bd9Sstevel@tonic-gate 
5607c478bd9Sstevel@tonic-gate 		if (below < best_below) {
5617c478bd9Sstevel@tonic-gate 			best_below = below;
5627c478bd9Sstevel@tonic-gate 			fnd_below = frlist;
5637c478bd9Sstevel@tonic-gate 		}
5647c478bd9Sstevel@tonic-gate 
5657c478bd9Sstevel@tonic-gate 		if (above < best_above) {
5667c478bd9Sstevel@tonic-gate 			best_above = above;
5677c478bd9Sstevel@tonic-gate 			fnd_above = frlist;
5687c478bd9Sstevel@tonic-gate 		}
5697c478bd9Sstevel@tonic-gate 
5707c478bd9Sstevel@tonic-gate 		if (unused < best_unused) {
5717c478bd9Sstevel@tonic-gate 			best_unused = unused;
5727c478bd9Sstevel@tonic-gate 			fnd_unused = frlist;
5737c478bd9Sstevel@tonic-gate 		}
5747c478bd9Sstevel@tonic-gate 	}
5757c478bd9Sstevel@tonic-gate 
5767c478bd9Sstevel@tonic-gate 	if (best_below == 0)
5777c478bd9Sstevel@tonic-gate 		return (fnd_below);
5787c478bd9Sstevel@tonic-gate 	if (best_above == 0)
5797c478bd9Sstevel@tonic-gate 		return (fnd_above);
5807c478bd9Sstevel@tonic-gate 	if (best_unused < ULONG_MAX)
5817c478bd9Sstevel@tonic-gate 		return (fnd_unused);
5827c478bd9Sstevel@tonic-gate 
5837c478bd9Sstevel@tonic-gate 	return (frlist);
5847c478bd9Sstevel@tonic-gate }
5857c478bd9Sstevel@tonic-gate 
5867c478bd9Sstevel@tonic-gate /*
5877c478bd9Sstevel@tonic-gate  * Nucleus data memory allocator.
5887c478bd9Sstevel@tonic-gate  * The granularity of the allocator is ecache_alignsize.
5897c478bd9Sstevel@tonic-gate  * See also comment for ndata_alloc_init().
5907c478bd9Sstevel@tonic-gate  */
5917c478bd9Sstevel@tonic-gate void *
5927c478bd9Sstevel@tonic-gate ndata_alloc(struct memlist *ndata, size_t wanted, size_t alignment)
5937c478bd9Sstevel@tonic-gate {
5947c478bd9Sstevel@tonic-gate 	struct memlist *found;
5957c478bd9Sstevel@tonic-gate 	struct memlist *fnd_above;
5967c478bd9Sstevel@tonic-gate 	uintptr_t base;
5977c478bd9Sstevel@tonic-gate 	uintptr_t end;
5987c478bd9Sstevel@tonic-gate 	size_t below;
5997c478bd9Sstevel@tonic-gate 	size_t above;
6007c478bd9Sstevel@tonic-gate 
6017c478bd9Sstevel@tonic-gate 	/*
6027c478bd9Sstevel@tonic-gate 	 * Look for the best matching buffer, avoid memory fragmentation.
6037c478bd9Sstevel@tonic-gate 	 */
6047c478bd9Sstevel@tonic-gate 	if ((found = ndata_select_chunk(ndata, wanted, alignment)) == NULL)
6057c478bd9Sstevel@tonic-gate 		return (NULL);
6067c478bd9Sstevel@tonic-gate 
6077c478bd9Sstevel@tonic-gate 	/*
6087c478bd9Sstevel@tonic-gate 	 * Allocate the nucleus data buffer.
6097c478bd9Sstevel@tonic-gate 	 */
6107c478bd9Sstevel@tonic-gate 	base = roundup(found->address, alignment);
6117c478bd9Sstevel@tonic-gate 	end = roundup(base + wanted, ecache_alignsize);
6127c478bd9Sstevel@tonic-gate 	ASSERT(end <= found->address + found->size);
6137c478bd9Sstevel@tonic-gate 
6147c478bd9Sstevel@tonic-gate 	below = base - found->address;
6157c478bd9Sstevel@tonic-gate 	above = found->address + found->size - end;
6167c478bd9Sstevel@tonic-gate 	ASSERT(above == 0 || (above % ecache_alignsize) == 0);
6177c478bd9Sstevel@tonic-gate 
6187c478bd9Sstevel@tonic-gate 	if (below >= ecache_alignsize) {
6197c478bd9Sstevel@tonic-gate 		/*
6207c478bd9Sstevel@tonic-gate 		 * There is free memory below the allocated memory chunk.
6217c478bd9Sstevel@tonic-gate 		 */
6227c478bd9Sstevel@tonic-gate 		found->size = below - below % ecache_alignsize;
6237c478bd9Sstevel@tonic-gate 
6247c478bd9Sstevel@tonic-gate 		if (above) {
6257c478bd9Sstevel@tonic-gate 			fnd_above = (struct memlist *)end;
6267c478bd9Sstevel@tonic-gate 			fnd_above->address = end;
6277c478bd9Sstevel@tonic-gate 			fnd_above->size = above;
6287c478bd9Sstevel@tonic-gate 
6297c478bd9Sstevel@tonic-gate 			if ((fnd_above->next = found->next) != NULL)
6307c478bd9Sstevel@tonic-gate 				found->next->prev = fnd_above;
6317c478bd9Sstevel@tonic-gate 			fnd_above->prev = found;
6327c478bd9Sstevel@tonic-gate 			found->next = fnd_above;
6337c478bd9Sstevel@tonic-gate 		}
6347c478bd9Sstevel@tonic-gate 
6357c478bd9Sstevel@tonic-gate 		return ((void *)base);
6367c478bd9Sstevel@tonic-gate 	}
6377c478bd9Sstevel@tonic-gate 
6387c478bd9Sstevel@tonic-gate 	if (found->prev == NULL) {
6397c478bd9Sstevel@tonic-gate 		/*
6407c478bd9Sstevel@tonic-gate 		 * The first chunk (ndata) is selected.
6417c478bd9Sstevel@tonic-gate 		 */
6427c478bd9Sstevel@tonic-gate 		ASSERT(found == ndata);
6437c478bd9Sstevel@tonic-gate 		if (above) {
6447c478bd9Sstevel@tonic-gate 			found->address = end;
6457c478bd9Sstevel@tonic-gate 			found->size = above;
6467c478bd9Sstevel@tonic-gate 		} else if (found->next != NULL) {
6477c478bd9Sstevel@tonic-gate 			found->address = found->next->address;
6487c478bd9Sstevel@tonic-gate 			found->size = found->next->size;
6497c478bd9Sstevel@tonic-gate 			if ((found->next = found->next->next) != NULL)
6507c478bd9Sstevel@tonic-gate 				found->next->prev = found;
6517c478bd9Sstevel@tonic-gate 
6527c478bd9Sstevel@tonic-gate 			bzero((void *)found->address, sizeof (struct memlist));
6537c478bd9Sstevel@tonic-gate 		} else {
6547c478bd9Sstevel@tonic-gate 			found->address = end;
6557c478bd9Sstevel@tonic-gate 			found->size = 0;
6567c478bd9Sstevel@tonic-gate 		}
6577c478bd9Sstevel@tonic-gate 
6587c478bd9Sstevel@tonic-gate 		return ((void *)base);
6597c478bd9Sstevel@tonic-gate 	}
6607c478bd9Sstevel@tonic-gate 
6617c478bd9Sstevel@tonic-gate 	/*
6627c478bd9Sstevel@tonic-gate 	 * Not the first chunk.
6637c478bd9Sstevel@tonic-gate 	 */
6647c478bd9Sstevel@tonic-gate 	if (above) {
6657c478bd9Sstevel@tonic-gate 		fnd_above = (struct memlist *)end;
6667c478bd9Sstevel@tonic-gate 		fnd_above->address = end;
6677c478bd9Sstevel@tonic-gate 		fnd_above->size = above;
6687c478bd9Sstevel@tonic-gate 
6697c478bd9Sstevel@tonic-gate 		if ((fnd_above->next = found->next) != NULL)
6707c478bd9Sstevel@tonic-gate 			fnd_above->next->prev = fnd_above;
6717c478bd9Sstevel@tonic-gate 		fnd_above->prev = found->prev;
6727c478bd9Sstevel@tonic-gate 		found->prev->next = fnd_above;
6737c478bd9Sstevel@tonic-gate 
6747c478bd9Sstevel@tonic-gate 	} else {
6757c478bd9Sstevel@tonic-gate 		if ((found->prev->next = found->next) != NULL)
6767c478bd9Sstevel@tonic-gate 			found->next->prev = found->prev;
6777c478bd9Sstevel@tonic-gate 	}
6787c478bd9Sstevel@tonic-gate 
6797c478bd9Sstevel@tonic-gate 	bzero((void *)found->address, sizeof (struct memlist));
6807c478bd9Sstevel@tonic-gate 
6817c478bd9Sstevel@tonic-gate 	return ((void *)base);
6827c478bd9Sstevel@tonic-gate }
6837c478bd9Sstevel@tonic-gate 
6847c478bd9Sstevel@tonic-gate /*
6857c478bd9Sstevel@tonic-gate  * Size the kernel TSBs based upon the amount of physical
6867c478bd9Sstevel@tonic-gate  * memory in the system.
6877c478bd9Sstevel@tonic-gate  */
6887c478bd9Sstevel@tonic-gate static void
6897c478bd9Sstevel@tonic-gate calc_tsb_sizes(pgcnt_t npages)
6907c478bd9Sstevel@tonic-gate {
6917c478bd9Sstevel@tonic-gate 	PRM_DEBUG(npages);
6927c478bd9Sstevel@tonic-gate 
6937c478bd9Sstevel@tonic-gate 	if (npages <= TSB_FREEMEM_MIN) {
6947c478bd9Sstevel@tonic-gate 		ktsb_szcode = TSB_128K_SZCODE;
6957c478bd9Sstevel@tonic-gate 		enable_bigktsb = 0;
6967c478bd9Sstevel@tonic-gate 	} else if (npages <= TSB_FREEMEM_LARGE / 2) {
6977c478bd9Sstevel@tonic-gate 		ktsb_szcode = TSB_256K_SZCODE;
6987c478bd9Sstevel@tonic-gate 		enable_bigktsb = 0;
6997c478bd9Sstevel@tonic-gate 	} else if (npages <= TSB_FREEMEM_LARGE) {
7007c478bd9Sstevel@tonic-gate 		ktsb_szcode = TSB_512K_SZCODE;
7017c478bd9Sstevel@tonic-gate 		enable_bigktsb = 0;
7027c478bd9Sstevel@tonic-gate 	} else if (npages <= TSB_FREEMEM_LARGE * 2 ||
7037c478bd9Sstevel@tonic-gate 	    enable_bigktsb == 0) {
7047c478bd9Sstevel@tonic-gate 		ktsb_szcode = TSB_1M_SZCODE;
7057c478bd9Sstevel@tonic-gate 		enable_bigktsb = 0;
7067c478bd9Sstevel@tonic-gate 	} else {
7077c478bd9Sstevel@tonic-gate 		ktsb_szcode = highbit(npages - 1);
7087c478bd9Sstevel@tonic-gate 		ktsb_szcode -= TSB_START_SIZE;
7097c478bd9Sstevel@tonic-gate 		ktsb_szcode = MAX(ktsb_szcode, MIN_BIGKTSB_SZCODE);
7107c478bd9Sstevel@tonic-gate 		ktsb_szcode = MIN(ktsb_szcode, MAX_BIGKTSB_SZCODE);
7117c478bd9Sstevel@tonic-gate 	}
7127c478bd9Sstevel@tonic-gate 
7137c478bd9Sstevel@tonic-gate 	/*
7147c478bd9Sstevel@tonic-gate 	 * We choose the TSB to hold kernel 4M mappings to have twice
7157c478bd9Sstevel@tonic-gate 	 * the reach as the primary kernel TSB since this TSB will
7167c478bd9Sstevel@tonic-gate 	 * potentially (currently) be shared by both mappings to all of
7177c478bd9Sstevel@tonic-gate 	 * physical memory plus user TSBs.  Since the current
7187c478bd9Sstevel@tonic-gate 	 * limit on primary kernel TSB size is 16MB this will top out
7197c478bd9Sstevel@tonic-gate 	 * at 64K which we can certainly afford.
7207c478bd9Sstevel@tonic-gate 	 */
7217c478bd9Sstevel@tonic-gate 	ktsb4m_szcode = ktsb_szcode - (MMU_PAGESHIFT4M - MMU_PAGESHIFT) + 1;
7227c478bd9Sstevel@tonic-gate 	if (ktsb4m_szcode < TSB_MIN_SZCODE)
7237c478bd9Sstevel@tonic-gate 		ktsb4m_szcode = TSB_MIN_SZCODE;
7247c478bd9Sstevel@tonic-gate 
7257c478bd9Sstevel@tonic-gate 	ktsb_sz = TSB_BYTES(ktsb_szcode);	/* kernel 8K tsb size */
7267c478bd9Sstevel@tonic-gate 	ktsb4m_sz = TSB_BYTES(ktsb4m_szcode);	/* kernel 4M tsb size */
7277c478bd9Sstevel@tonic-gate }
7287c478bd9Sstevel@tonic-gate 
7297c478bd9Sstevel@tonic-gate /*
7307c478bd9Sstevel@tonic-gate  * Allocate kernel TSBs from nucleus data memory.
7317c478bd9Sstevel@tonic-gate  * The function return 0 on success and -1 on failure.
7327c478bd9Sstevel@tonic-gate  */
7337c478bd9Sstevel@tonic-gate int
7347c478bd9Sstevel@tonic-gate ndata_alloc_tsbs(struct memlist *ndata, pgcnt_t npages)
7357c478bd9Sstevel@tonic-gate {
7367c478bd9Sstevel@tonic-gate 	/*
7377c478bd9Sstevel@tonic-gate 	 * Size the kernel TSBs based upon the amount of physical
7387c478bd9Sstevel@tonic-gate 	 * memory in the system.
7397c478bd9Sstevel@tonic-gate 	 */
7407c478bd9Sstevel@tonic-gate 	calc_tsb_sizes(npages);
7417c478bd9Sstevel@tonic-gate 
7427c478bd9Sstevel@tonic-gate 	/*
7437c478bd9Sstevel@tonic-gate 	 * Allocate the 8K kernel TSB if it belongs inside the nucleus.
7447c478bd9Sstevel@tonic-gate 	 */
7457c478bd9Sstevel@tonic-gate 	if (enable_bigktsb == 0) {
7467c478bd9Sstevel@tonic-gate 		if ((ktsb_base = ndata_alloc(ndata, ktsb_sz, ktsb_sz)) == NULL)
7477c478bd9Sstevel@tonic-gate 			return (-1);
7487c478bd9Sstevel@tonic-gate 		ASSERT(!((uintptr_t)ktsb_base & (ktsb_sz - 1)));
7497c478bd9Sstevel@tonic-gate 
7507c478bd9Sstevel@tonic-gate 		PRM_DEBUG(ktsb_base);
7517c478bd9Sstevel@tonic-gate 		PRM_DEBUG(ktsb_sz);
7527c478bd9Sstevel@tonic-gate 		PRM_DEBUG(ktsb_szcode);
7537c478bd9Sstevel@tonic-gate 	}
7547c478bd9Sstevel@tonic-gate 
7557c478bd9Sstevel@tonic-gate 	/*
7567c478bd9Sstevel@tonic-gate 	 * Next, allocate 4M kernel TSB from the nucleus since it's small.
7577c478bd9Sstevel@tonic-gate 	 */
7587c478bd9Sstevel@tonic-gate 	if ((ktsb4m_base = ndata_alloc(ndata, ktsb4m_sz, ktsb4m_sz)) == NULL)
7597c478bd9Sstevel@tonic-gate 		return (-1);
7607c478bd9Sstevel@tonic-gate 	ASSERT(!((uintptr_t)ktsb4m_base & (ktsb4m_sz - 1)));
7617c478bd9Sstevel@tonic-gate 
7627c478bd9Sstevel@tonic-gate 	PRM_DEBUG(ktsb4m_base);
7637c478bd9Sstevel@tonic-gate 	PRM_DEBUG(ktsb4m_sz);
7647c478bd9Sstevel@tonic-gate 	PRM_DEBUG(ktsb4m_szcode);
7657c478bd9Sstevel@tonic-gate 
7667c478bd9Sstevel@tonic-gate 	return (0);
7677c478bd9Sstevel@tonic-gate }
7687c478bd9Sstevel@tonic-gate 
7697c478bd9Sstevel@tonic-gate /*
7707c478bd9Sstevel@tonic-gate  * Allocate hat structs from the nucleus data memory.
7717c478bd9Sstevel@tonic-gate  */
7727c478bd9Sstevel@tonic-gate int
7737c478bd9Sstevel@tonic-gate ndata_alloc_hat(struct memlist *ndata, pgcnt_t npages, pgcnt_t kpm_npages)
7747c478bd9Sstevel@tonic-gate {
7757c478bd9Sstevel@tonic-gate 	size_t	mml_alloc_sz;
7767c478bd9Sstevel@tonic-gate 	size_t	cb_alloc_sz;
7777c478bd9Sstevel@tonic-gate 	int	max_nucuhme_buckets = MAX_NUCUHME_BUCKETS;
7787c478bd9Sstevel@tonic-gate 	int	max_nuckhme_buckets = MAX_NUCKHME_BUCKETS;
7797c478bd9Sstevel@tonic-gate 	ulong_t hme_buckets;
7807c478bd9Sstevel@tonic-gate 
7817c478bd9Sstevel@tonic-gate 	if (enable_bigktsb) {
7827c478bd9Sstevel@tonic-gate 		ASSERT((max_nucuhme_buckets + max_nuckhme_buckets) *
7837c478bd9Sstevel@tonic-gate 		    sizeof (struct hmehash_bucket) <=
7847c478bd9Sstevel@tonic-gate 			TSB_BYTES(TSB_1M_SZCODE));
7857c478bd9Sstevel@tonic-gate 
7867c478bd9Sstevel@tonic-gate 		max_nucuhme_buckets *= 2;
7877c478bd9Sstevel@tonic-gate 		max_nuckhme_buckets *= 2;
7887c478bd9Sstevel@tonic-gate 	}
7897c478bd9Sstevel@tonic-gate 
7907c478bd9Sstevel@tonic-gate 	/*
7917c478bd9Sstevel@tonic-gate 	 * The number of buckets in the hme hash tables
7927c478bd9Sstevel@tonic-gate 	 * is a power of 2 such that the average hash chain length is
7937c478bd9Sstevel@tonic-gate 	 * HMENT_HASHAVELEN.  The number of buckets for the user hash is
7947c478bd9Sstevel@tonic-gate 	 * a function of physical memory and a predefined overmapping factor.
7957c478bd9Sstevel@tonic-gate 	 * The number of buckets for the kernel hash is a function of
7967c478bd9Sstevel@tonic-gate 	 * physical memory only.
7977c478bd9Sstevel@tonic-gate 	 */
7987c478bd9Sstevel@tonic-gate 	hme_buckets = (npages * HMEHASH_FACTOR) /
7997c478bd9Sstevel@tonic-gate 		(HMENT_HASHAVELEN * (HMEBLK_SPAN(TTE8K) >> MMU_PAGESHIFT));
8007c478bd9Sstevel@tonic-gate 
8017c478bd9Sstevel@tonic-gate 	uhmehash_num = (int)MIN(hme_buckets, MAX_UHME_BUCKETS);
8027c478bd9Sstevel@tonic-gate 
8037c478bd9Sstevel@tonic-gate 	if (uhmehash_num > USER_BUCKETS_THRESHOLD) {
8047c478bd9Sstevel@tonic-gate 		/*
8057c478bd9Sstevel@tonic-gate 		 * if uhmehash_num is not power of 2 round it down to the
8067c478bd9Sstevel@tonic-gate 		 *  next power of 2.
8077c478bd9Sstevel@tonic-gate 		 */
8087c478bd9Sstevel@tonic-gate 		uint_t align = 1 << (highbit(uhmehash_num - 1) - 1);
8097c478bd9Sstevel@tonic-gate 		uhmehash_num = P2ALIGN(uhmehash_num, align);
8107c478bd9Sstevel@tonic-gate 	} else
8117c478bd9Sstevel@tonic-gate 		uhmehash_num = 1 << highbit(uhmehash_num - 1);
8127c478bd9Sstevel@tonic-gate 
8137c478bd9Sstevel@tonic-gate 	hme_buckets = npages / (HMEBLK_SPAN(TTE8K) >> MMU_PAGESHIFT);
8147c478bd9Sstevel@tonic-gate 	khmehash_num = (int)MIN(hme_buckets, MAX_KHME_BUCKETS);
8157c478bd9Sstevel@tonic-gate 	khmehash_num = 1 << highbit(khmehash_num - 1);
8167c478bd9Sstevel@tonic-gate 	khmehash_num = MAX(khmehash_num, MIN_KHME_BUCKETS);
8177c478bd9Sstevel@tonic-gate 
8187c478bd9Sstevel@tonic-gate 	if ((khmehash_num > max_nuckhme_buckets) ||
8197c478bd9Sstevel@tonic-gate 		(uhmehash_num > max_nucuhme_buckets)) {
8207c478bd9Sstevel@tonic-gate 		khme_hash = NULL;
8217c478bd9Sstevel@tonic-gate 		uhme_hash = NULL;
8227c478bd9Sstevel@tonic-gate 	} else {
8237c478bd9Sstevel@tonic-gate 		size_t hmehash_sz = (uhmehash_num + khmehash_num) *
8247c478bd9Sstevel@tonic-gate 		    sizeof (struct hmehash_bucket);
8257c478bd9Sstevel@tonic-gate 
8267c478bd9Sstevel@tonic-gate 		if ((khme_hash = ndata_alloc(ndata, hmehash_sz,
8277c478bd9Sstevel@tonic-gate 		    ecache_alignsize)) != NULL)
8287c478bd9Sstevel@tonic-gate 			uhme_hash = &khme_hash[khmehash_num];
8297c478bd9Sstevel@tonic-gate 		else
8307c478bd9Sstevel@tonic-gate 			uhme_hash = NULL;
8317c478bd9Sstevel@tonic-gate 
8327c478bd9Sstevel@tonic-gate 		PRM_DEBUG(hmehash_sz);
8337c478bd9Sstevel@tonic-gate 	}
8347c478bd9Sstevel@tonic-gate 
8357c478bd9Sstevel@tonic-gate 	PRM_DEBUG(khme_hash);
8367c478bd9Sstevel@tonic-gate 	PRM_DEBUG(khmehash_num);
8377c478bd9Sstevel@tonic-gate 	PRM_DEBUG(uhme_hash);
8387c478bd9Sstevel@tonic-gate 	PRM_DEBUG(uhmehash_num);
8397c478bd9Sstevel@tonic-gate 
8407c478bd9Sstevel@tonic-gate 	/*
8417c478bd9Sstevel@tonic-gate 	 * For the page mapping list mutex array we allocate one mutex
8427c478bd9Sstevel@tonic-gate 	 * for every 128 pages (1 MB) with a minimum of 64 entries and
8437c478bd9Sstevel@tonic-gate 	 * a maximum of 8K entries. For the initial computation npages
8447c478bd9Sstevel@tonic-gate 	 * is rounded up (ie. 1 << highbit(npages * 1.5 / 128))
8457c478bd9Sstevel@tonic-gate 	 *
8467c478bd9Sstevel@tonic-gate 	 * mml_shift is roughly log2(mml_table_sz) + 3 for MLIST_HASH
8477c478bd9Sstevel@tonic-gate 	 *
8487c478bd9Sstevel@tonic-gate 	 * It is not required that this be allocated from the nucleus,
8497c478bd9Sstevel@tonic-gate 	 * but it is desirable.  So we first allocate from the nucleus
8507c478bd9Sstevel@tonic-gate 	 * everything that must be there.  Having done so, if mml_table
8517c478bd9Sstevel@tonic-gate 	 * will fit within what remains of the nucleus then it will be
8527c478bd9Sstevel@tonic-gate 	 * allocated here.  If not, set mml_table to NULL, which will cause
8537c478bd9Sstevel@tonic-gate 	 * startup_memlist() to BOP_ALLOC() space for it after our return...
8547c478bd9Sstevel@tonic-gate 	 */
8557c478bd9Sstevel@tonic-gate 	mml_table_sz = 1 << highbit((npages * 3) / 256);
8567c478bd9Sstevel@tonic-gate 	if (mml_table_sz < 64)
8577c478bd9Sstevel@tonic-gate 		mml_table_sz = 64;
8587c478bd9Sstevel@tonic-gate 	else if (mml_table_sz > 8192)
8597c478bd9Sstevel@tonic-gate 		mml_table_sz = 8192;
8607c478bd9Sstevel@tonic-gate 	mml_shift = highbit(mml_table_sz) + 3;
8617c478bd9Sstevel@tonic-gate 
8627c478bd9Sstevel@tonic-gate 	PRM_DEBUG(mml_table_sz);
8637c478bd9Sstevel@tonic-gate 	PRM_DEBUG(mml_shift);
8647c478bd9Sstevel@tonic-gate 
8657c478bd9Sstevel@tonic-gate 	mml_alloc_sz = mml_table_sz * sizeof (kmutex_t);
8667c478bd9Sstevel@tonic-gate 
8677c478bd9Sstevel@tonic-gate 	mml_table = ndata_alloc(ndata, mml_alloc_sz, ecache_alignsize);
8687c478bd9Sstevel@tonic-gate 
8697c478bd9Sstevel@tonic-gate 	PRM_DEBUG(mml_table);
8707c478bd9Sstevel@tonic-gate 
8717c478bd9Sstevel@tonic-gate 	cb_alloc_sz = sfmmu_max_cb_id * sizeof (struct sfmmu_callback);
8727c478bd9Sstevel@tonic-gate 	PRM_DEBUG(cb_alloc_sz);
8737c478bd9Sstevel@tonic-gate 	sfmmu_cb_table = ndata_alloc(ndata, cb_alloc_sz, ecache_alignsize);
8747c478bd9Sstevel@tonic-gate 	PRM_DEBUG(sfmmu_cb_table);
8757c478bd9Sstevel@tonic-gate 
8767c478bd9Sstevel@tonic-gate 	/*
8777c478bd9Sstevel@tonic-gate 	 * For the kpm_page mutex array we allocate one mutex every 16
8787c478bd9Sstevel@tonic-gate 	 * kpm pages (64MB). In smallpage mode we allocate one mutex
8797c478bd9Sstevel@tonic-gate 	 * every 8K pages. The minimum is set to 64 entries and the
8807c478bd9Sstevel@tonic-gate 	 * maximum to 8K entries.
8817c478bd9Sstevel@tonic-gate 	 *
8827c478bd9Sstevel@tonic-gate 	 * It is not required that this be allocated from the nucleus,
8837c478bd9Sstevel@tonic-gate 	 * but it is desirable.  So we first allocate from the nucleus
8847c478bd9Sstevel@tonic-gate 	 * everything that must be there.  Having done so, if kpmp_table
8857c478bd9Sstevel@tonic-gate 	 * or kpmp_stable will fit within what remains of the nucleus
8867c478bd9Sstevel@tonic-gate 	 * then it will be allocated here.  If not, startup_memlist()
8877c478bd9Sstevel@tonic-gate 	 * will use BOP_ALLOC() space for it after our return...
8887c478bd9Sstevel@tonic-gate 	 */
8897c478bd9Sstevel@tonic-gate 	if (kpm_enable) {
8907c478bd9Sstevel@tonic-gate 		size_t	kpmp_alloc_sz;
8917c478bd9Sstevel@tonic-gate 
8927c478bd9Sstevel@tonic-gate 		if (kpm_smallpages == 0) {
8937c478bd9Sstevel@tonic-gate 			kpmp_shift = highbit(sizeof (kpm_page_t)) - 1;
8947c478bd9Sstevel@tonic-gate 			kpmp_table_sz = 1 << highbit(kpm_npages / 16);
8957c478bd9Sstevel@tonic-gate 			kpmp_table_sz = (kpmp_table_sz < 64) ? 64 :
8967c478bd9Sstevel@tonic-gate 			    ((kpmp_table_sz > 8192) ? 8192 : kpmp_table_sz);
8977c478bd9Sstevel@tonic-gate 			kpmp_alloc_sz = kpmp_table_sz * sizeof (kpm_hlk_t);
8987c478bd9Sstevel@tonic-gate 
8997c478bd9Sstevel@tonic-gate 			kpmp_table = ndata_alloc(ndata, kpmp_alloc_sz,
9007c478bd9Sstevel@tonic-gate 			    ecache_alignsize);
9017c478bd9Sstevel@tonic-gate 
9027c478bd9Sstevel@tonic-gate 			PRM_DEBUG(kpmp_table);
9037c478bd9Sstevel@tonic-gate 			PRM_DEBUG(kpmp_table_sz);
9047c478bd9Sstevel@tonic-gate 
9057c478bd9Sstevel@tonic-gate 			kpmp_stable_sz = 0;
9067c478bd9Sstevel@tonic-gate 			kpmp_stable = NULL;
9077c478bd9Sstevel@tonic-gate 		} else {
9087c478bd9Sstevel@tonic-gate 			ASSERT(kpm_pgsz == PAGESIZE);
9097c478bd9Sstevel@tonic-gate 			kpmp_shift = highbit(sizeof (kpm_shlk_t)) + 1;
9107c478bd9Sstevel@tonic-gate 			kpmp_stable_sz = 1 << highbit(kpm_npages / 8192);
9117c478bd9Sstevel@tonic-gate 			kpmp_stable_sz = (kpmp_stable_sz < 64) ? 64 :
9127c478bd9Sstevel@tonic-gate 			    ((kpmp_stable_sz > 8192) ? 8192 : kpmp_stable_sz);
9137c478bd9Sstevel@tonic-gate 			kpmp_alloc_sz = kpmp_stable_sz * sizeof (kpm_shlk_t);
9147c478bd9Sstevel@tonic-gate 
9157c478bd9Sstevel@tonic-gate 			kpmp_stable = ndata_alloc(ndata, kpmp_alloc_sz,
9167c478bd9Sstevel@tonic-gate 			    ecache_alignsize);
9177c478bd9Sstevel@tonic-gate 
9187c478bd9Sstevel@tonic-gate 			PRM_DEBUG(kpmp_stable);
9197c478bd9Sstevel@tonic-gate 			PRM_DEBUG(kpmp_stable_sz);
9207c478bd9Sstevel@tonic-gate 
9217c478bd9Sstevel@tonic-gate 			kpmp_table_sz = 0;
9227c478bd9Sstevel@tonic-gate 			kpmp_table = NULL;
9237c478bd9Sstevel@tonic-gate 		}
9247c478bd9Sstevel@tonic-gate 		PRM_DEBUG(kpmp_shift);
9257c478bd9Sstevel@tonic-gate 	}
9267c478bd9Sstevel@tonic-gate 
9277c478bd9Sstevel@tonic-gate 	return (0);
9287c478bd9Sstevel@tonic-gate }
9297c478bd9Sstevel@tonic-gate 
9307c478bd9Sstevel@tonic-gate caddr_t
9317c478bd9Sstevel@tonic-gate alloc_hme_buckets(caddr_t base, int pagesize)
9327c478bd9Sstevel@tonic-gate {
9337c478bd9Sstevel@tonic-gate 	size_t hmehash_sz = (uhmehash_num + khmehash_num) *
9347c478bd9Sstevel@tonic-gate 	sizeof (struct hmehash_bucket);
9357c478bd9Sstevel@tonic-gate 
9367c478bd9Sstevel@tonic-gate 	ASSERT(khme_hash == NULL);
9377c478bd9Sstevel@tonic-gate 	ASSERT(uhme_hash == NULL);
9387c478bd9Sstevel@tonic-gate 
9397c478bd9Sstevel@tonic-gate 	/* If no pagesize specified, use default MMU pagesize */
9407c478bd9Sstevel@tonic-gate 	if (!pagesize)
9417c478bd9Sstevel@tonic-gate 		pagesize = MMU_PAGESIZE;
9427c478bd9Sstevel@tonic-gate 
9437c478bd9Sstevel@tonic-gate 	/*
9447c478bd9Sstevel@tonic-gate 	 * If we start aligned and ask for a multiple of a pagesize, and OBP
9457c478bd9Sstevel@tonic-gate 	 * supports large pages, we will then use mappings of the largest size
9467c478bd9Sstevel@tonic-gate 	 * possible for the BOP_ALLOC, possibly saving us tens of thousands of
9477c478bd9Sstevel@tonic-gate 	 * TLB miss-induced traversals of the TSBs and/or the HME hashes...
9487c478bd9Sstevel@tonic-gate 	 */
9497c478bd9Sstevel@tonic-gate 	base = (caddr_t)roundup((uintptr_t)base, pagesize);
9507c478bd9Sstevel@tonic-gate 	hmehash_sz = roundup(hmehash_sz, pagesize);
9517c478bd9Sstevel@tonic-gate 
9527c478bd9Sstevel@tonic-gate 	khme_hash = (struct hmehash_bucket *)BOP_ALLOC(bootops, base,
9537c478bd9Sstevel@tonic-gate 		hmehash_sz, pagesize);
9547c478bd9Sstevel@tonic-gate 
9557c478bd9Sstevel@tonic-gate 	if ((caddr_t)khme_hash != base)
9567c478bd9Sstevel@tonic-gate 		cmn_err(CE_PANIC, "Cannot bop_alloc hme hash buckets.");
9577c478bd9Sstevel@tonic-gate 
9587c478bd9Sstevel@tonic-gate 	uhme_hash = (struct hmehash_bucket *)((caddr_t)khme_hash +
9597c478bd9Sstevel@tonic-gate 		khmehash_num * sizeof (struct hmehash_bucket));
9607c478bd9Sstevel@tonic-gate 	base += hmehash_sz;
9617c478bd9Sstevel@tonic-gate 	return (base);
9627c478bd9Sstevel@tonic-gate }
9637c478bd9Sstevel@tonic-gate 
9647c478bd9Sstevel@tonic-gate /*
9657c478bd9Sstevel@tonic-gate  * This function bop allocs the kernel TSB.
9667c478bd9Sstevel@tonic-gate  */
9677c478bd9Sstevel@tonic-gate caddr_t
9687c478bd9Sstevel@tonic-gate sfmmu_ktsb_alloc(caddr_t tsbbase)
9697c478bd9Sstevel@tonic-gate {
9707c478bd9Sstevel@tonic-gate 	caddr_t vaddr;
9717c478bd9Sstevel@tonic-gate 
9727c478bd9Sstevel@tonic-gate 	if (enable_bigktsb) {
9737c478bd9Sstevel@tonic-gate 		ktsb_base = (caddr_t)roundup((uintptr_t)tsbbase, ktsb_sz);
9747c478bd9Sstevel@tonic-gate 		vaddr = (caddr_t)BOP_ALLOC(bootops, ktsb_base, ktsb_sz,
9757c478bd9Sstevel@tonic-gate 		    ktsb_sz);
9767c478bd9Sstevel@tonic-gate 		if (vaddr != ktsb_base)
9777c478bd9Sstevel@tonic-gate 			cmn_err(CE_PANIC, "sfmmu_ktsb_alloc: can't alloc"
9787c478bd9Sstevel@tonic-gate 			    " bigktsb");
9797c478bd9Sstevel@tonic-gate 		ktsb_base = vaddr;
9807c478bd9Sstevel@tonic-gate 		tsbbase = ktsb_base + ktsb_sz;
9817c478bd9Sstevel@tonic-gate 		PRM_DEBUG(ktsb_base);
9827c478bd9Sstevel@tonic-gate 		PRM_DEBUG(tsbbase);
9837c478bd9Sstevel@tonic-gate 	}
9847c478bd9Sstevel@tonic-gate 	return (tsbbase);
9857c478bd9Sstevel@tonic-gate }
9867c478bd9Sstevel@tonic-gate 
9877c478bd9Sstevel@tonic-gate /*
9887c478bd9Sstevel@tonic-gate  * Moves code assembled outside of the trap table into the trap
9897c478bd9Sstevel@tonic-gate  * table taking care to relocate relative branches to code outside
9907c478bd9Sstevel@tonic-gate  * of the trap handler.
9917c478bd9Sstevel@tonic-gate  */
9927c478bd9Sstevel@tonic-gate static void
9937c478bd9Sstevel@tonic-gate sfmmu_reloc_trap_handler(void *tablep, void *start, size_t count)
9947c478bd9Sstevel@tonic-gate {
9957c478bd9Sstevel@tonic-gate 	size_t i;
9967c478bd9Sstevel@tonic-gate 	uint32_t *src;
9977c478bd9Sstevel@tonic-gate 	uint32_t *dst;
9987c478bd9Sstevel@tonic-gate 	uint32_t inst;
9997c478bd9Sstevel@tonic-gate 	int op, op2;
10007c478bd9Sstevel@tonic-gate 	int32_t offset;
10017c478bd9Sstevel@tonic-gate 	int disp;
10027c478bd9Sstevel@tonic-gate 
10037c478bd9Sstevel@tonic-gate 	src = start;
10047c478bd9Sstevel@tonic-gate 	dst = tablep;
10057c478bd9Sstevel@tonic-gate 	offset = src - dst;
10067c478bd9Sstevel@tonic-gate 	for (src = start, i = 0; i < count; i++, src++, dst++) {
10077c478bd9Sstevel@tonic-gate 		inst = *dst = *src;
10087c478bd9Sstevel@tonic-gate 		op = (inst >> 30) & 0x2;
10097c478bd9Sstevel@tonic-gate 		if (op == 1) {
10107c478bd9Sstevel@tonic-gate 			/* call */
10117c478bd9Sstevel@tonic-gate 			disp = ((int32_t)inst << 2) >> 2; /* sign-extend */
10127c478bd9Sstevel@tonic-gate 			if (disp + i >= 0 && disp + i < count)
10137c478bd9Sstevel@tonic-gate 				continue;
10147c478bd9Sstevel@tonic-gate 			disp += offset;
10157c478bd9Sstevel@tonic-gate 			inst = 0x40000000u | (disp & 0x3fffffffu);
10167c478bd9Sstevel@tonic-gate 			*dst = inst;
10177c478bd9Sstevel@tonic-gate 		} else if (op == 0) {
10187c478bd9Sstevel@tonic-gate 			/* branch or sethi */
10197c478bd9Sstevel@tonic-gate 			op2 = (inst >> 22) & 0x7;
10207c478bd9Sstevel@tonic-gate 
10217c478bd9Sstevel@tonic-gate 			switch (op2) {
10227c478bd9Sstevel@tonic-gate 			case 0x3: /* BPr */
10237c478bd9Sstevel@tonic-gate 				disp = (((inst >> 20) & 0x3) << 14) |
10247c478bd9Sstevel@tonic-gate 				    (inst & 0x3fff);
10257c478bd9Sstevel@tonic-gate 				disp = (disp << 16) >> 16; /* sign-extend */
10267c478bd9Sstevel@tonic-gate 				if (disp + i >= 0 && disp + i < count)
10277c478bd9Sstevel@tonic-gate 					continue;
10287c478bd9Sstevel@tonic-gate 				disp += offset;
10297c478bd9Sstevel@tonic-gate 				if (((disp << 16) >> 16) != disp)
10307c478bd9Sstevel@tonic-gate 					cmn_err(CE_PANIC, "bad reloc");
10317c478bd9Sstevel@tonic-gate 				inst &= ~0x303fff;
10327c478bd9Sstevel@tonic-gate 				inst |= (disp & 0x3fff);
10337c478bd9Sstevel@tonic-gate 				inst |= (disp & 0xc000) << 6;
10347c478bd9Sstevel@tonic-gate 				break;
10357c478bd9Sstevel@tonic-gate 
10367c478bd9Sstevel@tonic-gate 			case 0x2: /* Bicc */
10377c478bd9Sstevel@tonic-gate 				disp = ((int32_t)inst << 10) >> 10;
10387c478bd9Sstevel@tonic-gate 				if (disp + i >= 0 && disp + i < count)
10397c478bd9Sstevel@tonic-gate 					continue;
10407c478bd9Sstevel@tonic-gate 				disp += offset;
10417c478bd9Sstevel@tonic-gate 				if (((disp << 10) >> 10) != disp)
10427c478bd9Sstevel@tonic-gate 					cmn_err(CE_PANIC, "bad reloc");
10437c478bd9Sstevel@tonic-gate 				inst &= ~0x3fffff;
10447c478bd9Sstevel@tonic-gate 				inst |= (disp & 0x3fffff);
10457c478bd9Sstevel@tonic-gate 				break;
10467c478bd9Sstevel@tonic-gate 
10477c478bd9Sstevel@tonic-gate 			case 0x1: /* Bpcc */
10487c478bd9Sstevel@tonic-gate 				disp = ((int32_t)inst << 13) >> 13;
10497c478bd9Sstevel@tonic-gate 				if (disp + i >= 0 && disp + i < count)
10507c478bd9Sstevel@tonic-gate 					continue;
10517c478bd9Sstevel@tonic-gate 				disp += offset;
10527c478bd9Sstevel@tonic-gate 				if (((disp << 13) >> 13) != disp)
10537c478bd9Sstevel@tonic-gate 					cmn_err(CE_PANIC, "bad reloc");
10547c478bd9Sstevel@tonic-gate 				inst &= ~0x7ffff;
10557c478bd9Sstevel@tonic-gate 				inst |= (disp & 0x7ffffu);
10567c478bd9Sstevel@tonic-gate 				break;
10577c478bd9Sstevel@tonic-gate 			}
10587c478bd9Sstevel@tonic-gate 			*dst = inst;
10597c478bd9Sstevel@tonic-gate 		}
10607c478bd9Sstevel@tonic-gate 	}
10617c478bd9Sstevel@tonic-gate 	flush_instr_mem(tablep, count * sizeof (uint32_t));
10627c478bd9Sstevel@tonic-gate }
10637c478bd9Sstevel@tonic-gate 
10647c478bd9Sstevel@tonic-gate /*
10657c478bd9Sstevel@tonic-gate  * Routine to allocate a large page to use in the TSB caches.
10667c478bd9Sstevel@tonic-gate  */
10677c478bd9Sstevel@tonic-gate /*ARGSUSED*/
10687c478bd9Sstevel@tonic-gate static page_t *
10697c478bd9Sstevel@tonic-gate sfmmu_tsb_page_create(void *addr, size_t size, int vmflag, void *arg)
10707c478bd9Sstevel@tonic-gate {
10717c478bd9Sstevel@tonic-gate 	int pgflags;
10727c478bd9Sstevel@tonic-gate 
10737c478bd9Sstevel@tonic-gate 	pgflags = PG_EXCL;
10747c478bd9Sstevel@tonic-gate 	if ((vmflag & VM_NOSLEEP) == 0)
10757c478bd9Sstevel@tonic-gate 		pgflags |= PG_WAIT;
10767c478bd9Sstevel@tonic-gate 	if (vmflag & VM_PANIC)
10777c478bd9Sstevel@tonic-gate 		pgflags |= PG_PANIC;
10787c478bd9Sstevel@tonic-gate 	if (vmflag & VM_PUSHPAGE)
10797c478bd9Sstevel@tonic-gate 		pgflags |= PG_PUSHPAGE;
10807c478bd9Sstevel@tonic-gate 
10817c478bd9Sstevel@tonic-gate 	return (page_create_va_large(&kvp, (u_offset_t)(uintptr_t)addr, size,
10827c478bd9Sstevel@tonic-gate 	    pgflags, &kvseg, addr, arg));
10837c478bd9Sstevel@tonic-gate }
10847c478bd9Sstevel@tonic-gate 
10857c478bd9Sstevel@tonic-gate /*
10867c478bd9Sstevel@tonic-gate  * Allocate a large page to back the virtual address range
10877c478bd9Sstevel@tonic-gate  * [addr, addr + size).  If addr is NULL, allocate the virtual address
10887c478bd9Sstevel@tonic-gate  * space as well.
10897c478bd9Sstevel@tonic-gate  */
10907c478bd9Sstevel@tonic-gate static void *
10917c478bd9Sstevel@tonic-gate sfmmu_tsb_xalloc(vmem_t *vmp, void *inaddr, size_t size, int vmflag,
10927c478bd9Sstevel@tonic-gate     uint_t attr, page_t *(*page_create_func)(void *, size_t, int, void *),
10937c478bd9Sstevel@tonic-gate     void *pcarg)
10947c478bd9Sstevel@tonic-gate {
10957c478bd9Sstevel@tonic-gate 	page_t *ppl;
10967c478bd9Sstevel@tonic-gate 	page_t *rootpp;
10977c478bd9Sstevel@tonic-gate 	caddr_t addr = inaddr;
10987c478bd9Sstevel@tonic-gate 	pgcnt_t npages = btopr(size);
10997c478bd9Sstevel@tonic-gate 	page_t **ppa;
11007c478bd9Sstevel@tonic-gate 	int i = 0;
11017c478bd9Sstevel@tonic-gate 
11027c478bd9Sstevel@tonic-gate 	/*
11037c478bd9Sstevel@tonic-gate 	 * Assuming that only TSBs will call this with size > PAGESIZE
11047c478bd9Sstevel@tonic-gate 	 * There is no reason why this couldn't be expanded to 8k pages as
11057c478bd9Sstevel@tonic-gate 	 * well, or other page sizes in the future .... but for now, we
11067c478bd9Sstevel@tonic-gate 	 * only support fixed sized page requests.
11077c478bd9Sstevel@tonic-gate 	 */
11087c478bd9Sstevel@tonic-gate 	if ((inaddr == NULL) && ((addr = vmem_xalloc(vmp, size, size, 0, 0,
11097c478bd9Sstevel@tonic-gate 	    NULL, NULL, vmflag)) == NULL))
11107c478bd9Sstevel@tonic-gate 		return (NULL);
11117c478bd9Sstevel@tonic-gate 
11127c478bd9Sstevel@tonic-gate 	/* If we ever don't want TSB slab-sized pages, this will panic */
11137c478bd9Sstevel@tonic-gate 	ASSERT(((uintptr_t)addr & (tsb_slab_size - 1)) == 0);
11147c478bd9Sstevel@tonic-gate 
11157c478bd9Sstevel@tonic-gate 	if (page_resv(npages, vmflag & VM_KMFLAGS) == 0) {
11167c478bd9Sstevel@tonic-gate 		if (inaddr == NULL)
11177c478bd9Sstevel@tonic-gate 			vmem_xfree(vmp, addr, size);
11187c478bd9Sstevel@tonic-gate 		return (NULL);
11197c478bd9Sstevel@tonic-gate 	}
11207c478bd9Sstevel@tonic-gate 
11217c478bd9Sstevel@tonic-gate 	ppl = page_create_func(addr, size, vmflag, pcarg);
11227c478bd9Sstevel@tonic-gate 	if (ppl == NULL) {
11237c478bd9Sstevel@tonic-gate 		if (inaddr == NULL)
11247c478bd9Sstevel@tonic-gate 			vmem_xfree(vmp, addr, size);
11257c478bd9Sstevel@tonic-gate 		page_unresv(npages);
11267c478bd9Sstevel@tonic-gate 		return (NULL);
11277c478bd9Sstevel@tonic-gate 	}
11287c478bd9Sstevel@tonic-gate 
11297c478bd9Sstevel@tonic-gate 	rootpp = ppl;
11307c478bd9Sstevel@tonic-gate 	ppa = kmem_zalloc(npages * sizeof (page_t *), KM_SLEEP);
11317c478bd9Sstevel@tonic-gate 	while (ppl != NULL) {
11327c478bd9Sstevel@tonic-gate 		page_t *pp = ppl;
11337c478bd9Sstevel@tonic-gate 		ppa[i++] = pp;
11347c478bd9Sstevel@tonic-gate 		page_sub(&ppl, pp);
11357c478bd9Sstevel@tonic-gate 		ASSERT(page_iolock_assert(pp));
11367c478bd9Sstevel@tonic-gate 		page_io_unlock(pp);
11377c478bd9Sstevel@tonic-gate 	}
11387c478bd9Sstevel@tonic-gate 
11397c478bd9Sstevel@tonic-gate 	/*
11407c478bd9Sstevel@tonic-gate 	 * Load the locked entry.  It's OK to preload the entry into
11417c478bd9Sstevel@tonic-gate 	 * the TSB since we now support large mappings in the kernel TSB.
11427c478bd9Sstevel@tonic-gate 	 */
11437c478bd9Sstevel@tonic-gate 	hat_memload_array(kas.a_hat, (caddr_t)rootpp->p_offset, size,
11447c478bd9Sstevel@tonic-gate 	    ppa, (PROT_ALL & ~PROT_USER) | HAT_NOSYNC | attr, HAT_LOAD_LOCK);
11457c478bd9Sstevel@tonic-gate 
11467c478bd9Sstevel@tonic-gate 	for (--i; i >= 0; --i) {
11477c478bd9Sstevel@tonic-gate 		(void) page_pp_lock(ppa[i], 0, 1);
11487c478bd9Sstevel@tonic-gate 		page_unlock(ppa[i]);
11497c478bd9Sstevel@tonic-gate 	}
11507c478bd9Sstevel@tonic-gate 
11517c478bd9Sstevel@tonic-gate 	kmem_free(ppa, npages * sizeof (page_t *));
11527c478bd9Sstevel@tonic-gate 	return (addr);
11537c478bd9Sstevel@tonic-gate }
11547c478bd9Sstevel@tonic-gate 
11557c478bd9Sstevel@tonic-gate /* Called to import new spans into the TSB vmem arenas */
11567c478bd9Sstevel@tonic-gate void *
11577c478bd9Sstevel@tonic-gate sfmmu_tsb_segkmem_alloc(vmem_t *vmp, size_t size, int vmflag)
11587c478bd9Sstevel@tonic-gate {
11597c478bd9Sstevel@tonic-gate 	lgrp_id_t lgrpid = LGRP_NONE;
11607c478bd9Sstevel@tonic-gate 
11617c478bd9Sstevel@tonic-gate 	if (tsb_lgrp_affinity) {
11627c478bd9Sstevel@tonic-gate 		/*
11637c478bd9Sstevel@tonic-gate 		 * Search for the vmp->lgrpid mapping by brute force;
11647c478bd9Sstevel@tonic-gate 		 * some day vmp will have an lgrp, until then we have
11657c478bd9Sstevel@tonic-gate 		 * to do this the hard way.
11667c478bd9Sstevel@tonic-gate 		 */
11677c478bd9Sstevel@tonic-gate 		for (lgrpid = 0; lgrpid < NLGRPS_MAX &&
11687c478bd9Sstevel@tonic-gate 		    vmp != kmem_tsb_default_arena[lgrpid]; lgrpid++);
11697c478bd9Sstevel@tonic-gate 		if (lgrpid == NLGRPS_MAX)
11707c478bd9Sstevel@tonic-gate 			lgrpid = LGRP_NONE;
11717c478bd9Sstevel@tonic-gate 	}
11727c478bd9Sstevel@tonic-gate 
11737c478bd9Sstevel@tonic-gate 	return (sfmmu_tsb_xalloc(vmp, NULL, size, vmflag, 0,
11747c478bd9Sstevel@tonic-gate 	    sfmmu_tsb_page_create, lgrpid != LGRP_NONE? &lgrpid : NULL));
11757c478bd9Sstevel@tonic-gate }
11767c478bd9Sstevel@tonic-gate 
11777c478bd9Sstevel@tonic-gate /* Called to free spans from the TSB vmem arenas */
11787c478bd9Sstevel@tonic-gate void
11797c478bd9Sstevel@tonic-gate sfmmu_tsb_segkmem_free(vmem_t *vmp, void *inaddr, size_t size)
11807c478bd9Sstevel@tonic-gate {
11817c478bd9Sstevel@tonic-gate 	page_t *pp;
11827c478bd9Sstevel@tonic-gate 	caddr_t addr = inaddr;
11837c478bd9Sstevel@tonic-gate 	caddr_t eaddr;
11847c478bd9Sstevel@tonic-gate 	pgcnt_t npages = btopr(size);
11857c478bd9Sstevel@tonic-gate 	pgcnt_t pgs_left = npages;
11867c478bd9Sstevel@tonic-gate 	page_t *rootpp = NULL;
11877c478bd9Sstevel@tonic-gate 
11887c478bd9Sstevel@tonic-gate 	ASSERT(((uintptr_t)addr & (tsb_slab_size - 1)) == 0);
11897c478bd9Sstevel@tonic-gate 
11907c478bd9Sstevel@tonic-gate 	hat_unload(kas.a_hat, addr, size, HAT_UNLOAD_UNLOCK);
11917c478bd9Sstevel@tonic-gate 
11927c478bd9Sstevel@tonic-gate 	for (eaddr = addr + size; addr < eaddr; addr += PAGESIZE) {
11937c478bd9Sstevel@tonic-gate 		pp = page_lookup(&kvp, (u_offset_t)(uintptr_t)addr, SE_EXCL);
11947c478bd9Sstevel@tonic-gate 		if (pp == NULL)
11957c478bd9Sstevel@tonic-gate 			panic("sfmmu_tsb_segkmem_free: page not found");
11967c478bd9Sstevel@tonic-gate 
11977c478bd9Sstevel@tonic-gate 		ASSERT(PAGE_EXCL(pp));
11987c478bd9Sstevel@tonic-gate 		page_pp_unlock(pp, 0, 1);
11997c478bd9Sstevel@tonic-gate 
12007c478bd9Sstevel@tonic-gate 		if (rootpp == NULL)
12017c478bd9Sstevel@tonic-gate 			rootpp = pp;
12027c478bd9Sstevel@tonic-gate 		if (--pgs_left == 0) {
12037c478bd9Sstevel@tonic-gate 			/*
12047c478bd9Sstevel@tonic-gate 			 * similar logic to segspt_free_pages, but we know we
12057c478bd9Sstevel@tonic-gate 			 * have one large page.
12067c478bd9Sstevel@tonic-gate 			 */
12077c478bd9Sstevel@tonic-gate 			page_destroy_pages(rootpp);
12087c478bd9Sstevel@tonic-gate 		}
12097c478bd9Sstevel@tonic-gate 	}
12107c478bd9Sstevel@tonic-gate 	page_unresv(npages);
12117c478bd9Sstevel@tonic-gate 
12127c478bd9Sstevel@tonic-gate 	if (vmp != NULL)
12137c478bd9Sstevel@tonic-gate 		vmem_xfree(vmp, inaddr, size);
12147c478bd9Sstevel@tonic-gate }
1215