xref: /titanic_41/usr/src/uts/i86pc/dboot/dboot_startkern.c (revision cde2885fdf538266ee2a3b08dee2d5075ce8fa2b)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 
28 #include <sys/types.h>
29 #include <sys/machparam.h>
30 #include <sys/x86_archext.h>
31 #include <sys/systm.h>
32 #include <sys/mach_mmu.h>
33 #include <sys/multiboot.h>
34 
35 #if defined(__xpv)
36 
37 #include <sys/hypervisor.h>
38 uintptr_t xen_virt_start;
39 pfn_t *mfn_to_pfn_mapping;
40 
41 #else /* !__xpv */
42 
43 extern multiboot_header_t mb_header;
44 extern int have_cpuid(void);
45 
46 #endif /* !__xpv */
47 
48 #include <sys/inttypes.h>
49 #include <sys/bootinfo.h>
50 #include <sys/mach_mmu.h>
51 #include <sys/boot_console.h>
52 
53 #include "dboot_asm.h"
54 #include "dboot_printf.h"
55 #include "dboot_xboot.h"
56 #include "dboot_elfload.h"
57 
58 /*
59  * This file contains code that runs to transition us from either a multiboot
60  * compliant loader (32 bit non-paging) or a XPV domain loader to
61  * regular kernel execution. Its task is to setup the kernel memory image
62  * and page tables.
63  *
64  * The code executes as:
65  *	- 32 bits under GRUB (for 32 or 64 bit Solaris)
66  * 	- a 32 bit program for the 32-bit PV hypervisor
67  *	- a 64 bit program for the 64-bit PV hypervisor (at least for now)
68  *
69  * Under the PV hypervisor, we must create mappings for any memory beyond the
70  * initial start of day allocation (such as the kernel itself).
71  *
72  * When on the metal, the mapping between maddr_t and paddr_t is 1:1.
73  * Since we are running in real mode, so all such memory is accessible.
74  */
75 
76 /*
77  * Standard bits used in PTE (page level) and PTP (internal levels)
78  */
79 x86pte_t ptp_bits = PT_VALID | PT_REF | PT_WRITABLE | PT_USER;
80 x86pte_t pte_bits = PT_VALID | PT_REF | PT_WRITABLE | PT_MOD | PT_NOCONSIST;
81 
82 /*
83  * This is the target addresses (physical) where the kernel text and data
84  * nucleus pages will be unpacked. On the hypervisor this is actually a
85  * virtual address.
86  */
87 paddr_t ktext_phys;
88 uint32_t ksize = 2 * FOUR_MEG;	/* kernel nucleus is 8Meg */
89 
90 static uint64_t target_kernel_text;	/* value to use for KERNEL_TEXT */
91 
92 /*
93  * The stack is setup in assembler before entering startup_kernel()
94  */
95 char stack_space[STACK_SIZE];
96 
97 /*
98  * Used to track physical memory allocation
99  */
100 static paddr_t next_avail_addr = 0;
101 
102 #if defined(__xpv)
103 /*
104  * Additional information needed for hypervisor memory allocation.
105  * Only memory up to scratch_end is mapped by page tables.
106  * mfn_base is the start of the hypervisor virtual image. It's ONE_GIG, so
107  * to derive a pfn from a pointer, you subtract mfn_base.
108  */
109 
110 static paddr_t scratch_end = 0;	/* we can't write all of mem here */
111 static paddr_t mfn_base;		/* addr corresponding to mfn_list[0] */
112 start_info_t *xen_info;
113 
114 #else	/* __xpv */
115 
116 /*
117  * If on the metal, then we have a multiboot loader.
118  */
119 multiboot_info_t *mb_info;
120 
121 #endif	/* __xpv */
122 
123 /*
124  * This contains information passed to the kernel
125  */
126 struct xboot_info boot_info[2];	/* extra space to fix alignement for amd64 */
127 struct xboot_info *bi;
128 
129 /*
130  * Page table and memory stuff.
131  */
132 static paddr_t max_mem;			/* maximum memory address */
133 
134 /*
135  * Information about processor MMU
136  */
137 int amd64_support = 0;
138 int largepage_support = 0;
139 int pae_support = 0;
140 int pge_support = 0;
141 int NX_support = 0;
142 
143 /*
144  * Low 32 bits of kernel entry address passed back to assembler.
145  * When running a 64 bit kernel, the high 32 bits are 0xffffffff.
146  */
147 uint32_t entry_addr_low;
148 
149 /*
150  * Memlists for the kernel. We shouldn't need a lot of these.
151  */
152 #define	MAX_MEMLIST (50)
153 struct boot_memlist memlists[MAX_MEMLIST];
154 uint_t memlists_used = 0;
155 struct boot_memlist pcimemlists[MAX_MEMLIST];
156 uint_t pcimemlists_used = 0;
157 
158 #define	MAX_MODULES (10)
159 struct boot_modules modules[MAX_MODULES];
160 uint_t modules_used = 0;
161 
162 /*
163  * Debugging macros
164  */
165 uint_t prom_debug = 0;
166 uint_t map_debug = 0;
167 
168 /*
169  * Either hypervisor-specific or grub-specific code builds the initial
170  * memlists. This code does the sort/merge/link for final use.
171  */
172 static void
173 sort_physinstall(void)
174 {
175 	int i;
176 #if !defined(__xpv)
177 	int j;
178 	struct boot_memlist tmp;
179 
180 	/*
181 	 * Now sort the memlists, in case they weren't in order.
182 	 * Yeah, this is a bubble sort; small, simple and easy to get right.
183 	 */
184 	DBG_MSG("Sorting phys-installed list\n");
185 	for (j = memlists_used - 1; j > 0; --j) {
186 		for (i = 0; i < j; ++i) {
187 			if (memlists[i].addr < memlists[i + 1].addr)
188 				continue;
189 			tmp = memlists[i];
190 			memlists[i] = memlists[i + 1];
191 			memlists[i + 1] = tmp;
192 		}
193 	}
194 
195 	/*
196 	 * Merge any memlists that don't have holes between them.
197 	 */
198 	for (i = 0; i <= memlists_used - 1; ++i) {
199 		if (memlists[i].addr + memlists[i].size != memlists[i + 1].addr)
200 			continue;
201 
202 		if (prom_debug)
203 			dboot_printf(
204 			    "merging mem segs %" PRIx64 "...%" PRIx64
205 			    " w/ %" PRIx64 "...%" PRIx64 "\n",
206 			    memlists[i].addr,
207 			    memlists[i].addr + memlists[i].size,
208 			    memlists[i + 1].addr,
209 			    memlists[i + 1].addr + memlists[i + 1].size);
210 
211 		memlists[i].size += memlists[i + 1].size;
212 		for (j = i + 1; j < memlists_used - 1; ++j)
213 			memlists[j] = memlists[j + 1];
214 		--memlists_used;
215 		DBG(memlists_used);
216 		--i;	/* after merging we need to reexamine, so do this */
217 	}
218 #endif	/* __xpv */
219 
220 	if (prom_debug) {
221 		dboot_printf("\nFinal memlists:\n");
222 		for (i = 0; i < memlists_used; ++i) {
223 			dboot_printf("\t%d: addr=%" PRIx64 " size=%"
224 			    PRIx64 "\n", i, memlists[i].addr, memlists[i].size);
225 		}
226 	}
227 
228 	/*
229 	 * link together the memlists with native size pointers
230 	 */
231 	memlists[0].next = 0;
232 	memlists[0].prev = 0;
233 	for (i = 1; i < memlists_used; ++i) {
234 		memlists[i].prev = (native_ptr_t)(uintptr_t)(memlists + i - 1);
235 		memlists[i].next = 0;
236 		memlists[i - 1].next = (native_ptr_t)(uintptr_t)(memlists + i);
237 	}
238 	bi->bi_phys_install = (native_ptr_t)memlists;
239 	DBG(bi->bi_phys_install);
240 }
241 
242 #if defined(__xpv)
243 
244 /*
245  * halt on the hypervisor after a delay to drain console output
246  */
247 void
248 dboot_halt(void)
249 {
250 	uint_t i = 10000;
251 
252 	while (--i)
253 		HYPERVISOR_yield();
254 	HYPERVISOR_shutdown(SHUTDOWN_poweroff);
255 }
256 
257 /*
258  * From a machine address, find the corresponding pseudo-physical address.
259  * Pseudo-physical address are contiguous and run from mfn_base in each VM.
260  * Machine addresses are the real underlying hardware addresses.
261  * These are needed for page table entries. Note that this routine is
262  * poorly protected. A bad value of "ma" will cause a page fault.
263  */
264 paddr_t
265 ma_to_pa(maddr_t ma)
266 {
267 	ulong_t pgoff = ma & MMU_PAGEOFFSET;
268 	ulong_t pfn = mfn_to_pfn_mapping[mmu_btop(ma)];
269 	paddr_t pa;
270 
271 	if (pfn >= xen_info->nr_pages)
272 		return (-(paddr_t)1);
273 	pa = mfn_base + mmu_ptob((paddr_t)pfn) + pgoff;
274 #ifdef DEBUG
275 	if (ma != pa_to_ma(pa))
276 		dboot_printf("ma_to_pa(%" PRIx64 ") got %" PRIx64 ", "
277 		    "pa_to_ma() says %" PRIx64 "\n", ma, pa, pa_to_ma(pa));
278 #endif
279 	return (pa);
280 }
281 
282 /*
283  * From a pseudo-physical address, find the corresponding machine address.
284  */
285 maddr_t
286 pa_to_ma(paddr_t pa)
287 {
288 	pfn_t pfn;
289 	ulong_t mfn;
290 
291 	pfn = mmu_btop(pa - mfn_base);
292 	if (pa < mfn_base || pfn >= xen_info->nr_pages)
293 		dboot_panic("pa_to_ma(): illegal address 0x%lx", (ulong_t)pa);
294 	mfn = ((ulong_t *)xen_info->mfn_list)[pfn];
295 #ifdef DEBUG
296 	if (mfn_to_pfn_mapping[mfn] != pfn)
297 		dboot_printf("pa_to_ma(pfn=%lx) got %lx ma_to_pa() says %lx\n",
298 		    pfn, mfn, mfn_to_pfn_mapping[mfn]);
299 #endif
300 	return (mfn_to_ma(mfn) | (pa & MMU_PAGEOFFSET));
301 }
302 
303 #endif	/* __xpv */
304 
305 x86pte_t
306 get_pteval(paddr_t table, uint_t index)
307 {
308 	if (pae_support)
309 		return (((x86pte_t *)(uintptr_t)table)[index]);
310 	return (((x86pte32_t *)(uintptr_t)table)[index]);
311 }
312 
313 /*ARGSUSED*/
314 void
315 set_pteval(paddr_t table, uint_t index, uint_t level, x86pte_t pteval)
316 {
317 #ifdef __xpv
318 	mmu_update_t t;
319 	maddr_t mtable = pa_to_ma(table);
320 	int retcnt;
321 
322 	t.ptr = (mtable + index * pte_size) | MMU_NORMAL_PT_UPDATE;
323 	t.val = pteval;
324 	if (HYPERVISOR_mmu_update(&t, 1, &retcnt, DOMID_SELF) || retcnt != 1)
325 		dboot_panic("HYPERVISOR_mmu_update() failed");
326 #else /* __xpv */
327 	uintptr_t tab_addr = (uintptr_t)table;
328 
329 	if (pae_support)
330 		((x86pte_t *)tab_addr)[index] = pteval;
331 	else
332 		((x86pte32_t *)tab_addr)[index] = (x86pte32_t)pteval;
333 	if (level == top_level && level == 2)
334 		reload_cr3();
335 #endif /* __xpv */
336 }
337 
338 paddr_t
339 make_ptable(x86pte_t *pteval, uint_t level)
340 {
341 	paddr_t new_table = (paddr_t)(uintptr_t)mem_alloc(MMU_PAGESIZE);
342 
343 	if (level == top_level && level == 2)
344 		*pteval = pa_to_ma((uintptr_t)new_table) | PT_VALID;
345 	else
346 		*pteval = pa_to_ma((uintptr_t)new_table) | ptp_bits;
347 
348 #ifdef __xpv
349 	/* Remove write permission to the new page table. */
350 	if (HYPERVISOR_update_va_mapping(new_table,
351 	    *pteval & ~(x86pte_t)PT_WRITABLE, UVMF_INVLPG | UVMF_LOCAL))
352 		dboot_panic("HYP_update_va_mapping error");
353 #endif
354 
355 	if (map_debug)
356 		dboot_printf("new page table lvl=%d paddr=0x%lx ptp=0x%"
357 		    PRIx64 "\n", level, (ulong_t)new_table, *pteval);
358 	return (new_table);
359 }
360 
361 x86pte_t *
362 map_pte(paddr_t table, uint_t index)
363 {
364 	return ((x86pte_t *)(uintptr_t)(table + index * pte_size));
365 }
366 
367 /*
368  * dump out the contents of page tables...
369  */
370 static void
371 dump_tables(void)
372 {
373 	uint_t save_index[4];	/* for recursion */
374 	char *save_table[4];	/* for recursion */
375 	uint_t	l;
376 	uint64_t va;
377 	uint64_t pgsize;
378 	int index;
379 	int i;
380 	x86pte_t pteval;
381 	char *table;
382 	static char *tablist = "\t\t\t";
383 	char *tabs = tablist + 3 - top_level;
384 	uint_t pa, pa1;
385 #if !defined(__xpv)
386 #define	maddr_t paddr_t
387 #endif /* !__xpv */
388 
389 	dboot_printf("Finished pagetables:\n");
390 	table = (char *)(uintptr_t)top_page_table;
391 	l = top_level;
392 	va = 0;
393 	for (index = 0; index < ptes_per_table; ++index) {
394 		pgsize = 1ull << shift_amt[l];
395 		if (pae_support)
396 			pteval = ((x86pte_t *)table)[index];
397 		else
398 			pteval = ((x86pte32_t *)table)[index];
399 		if (pteval == 0)
400 			goto next_entry;
401 
402 		dboot_printf("%s %p[0x%x] = %" PRIx64 ", va=%" PRIx64,
403 		    tabs + l, table, index, (uint64_t)pteval, va);
404 		pa = ma_to_pa(pteval & MMU_PAGEMASK);
405 		dboot_printf(" physaddr=%x\n", pa);
406 
407 		/*
408 		 * Don't try to walk hypervisor private pagetables
409 		 */
410 		if ((l > 1 || (l == 1 && (pteval & PT_PAGESIZE) == 0))) {
411 			save_table[l] = table;
412 			save_index[l] = index;
413 			--l;
414 			index = -1;
415 			table = (char *)(uintptr_t)
416 			    ma_to_pa(pteval & MMU_PAGEMASK);
417 			goto recursion;
418 		}
419 
420 		/*
421 		 * shorten dump for consecutive mappings
422 		 */
423 		for (i = 1; index + i < ptes_per_table; ++i) {
424 			if (pae_support)
425 				pteval = ((x86pte_t *)table)[index + i];
426 			else
427 				pteval = ((x86pte32_t *)table)[index + i];
428 			if (pteval == 0)
429 				break;
430 			pa1 = ma_to_pa(pteval & MMU_PAGEMASK);
431 			if (pa1 != pa + i * pgsize)
432 				break;
433 		}
434 		if (i > 2) {
435 			dboot_printf("%s...\n", tabs + l);
436 			va += pgsize * (i - 2);
437 			index += i - 2;
438 		}
439 next_entry:
440 		va += pgsize;
441 		if (l == 3 && index == 256)	/* VA hole */
442 			va = 0xffff800000000000ull;
443 recursion:
444 		;
445 	}
446 	if (l < top_level) {
447 		++l;
448 		index = save_index[l];
449 		table = save_table[l];
450 		goto recursion;
451 	}
452 }
453 
454 /*
455  * Add a mapping for the machine page at the given virtual address.
456  */
457 static void
458 map_ma_at_va(maddr_t ma, native_ptr_t va, uint_t level)
459 {
460 	x86pte_t *ptep;
461 	x86pte_t pteval;
462 
463 	pteval = ma | pte_bits;
464 	if (level > 0)
465 		pteval |= PT_PAGESIZE;
466 	if (va >= target_kernel_text && pge_support)
467 		pteval |= PT_GLOBAL;
468 
469 	if (map_debug && ma != va)
470 		dboot_printf("mapping ma=0x%" PRIx64 " va=0x%" PRIx64
471 		    " pte=0x%" PRIx64 " l=%d\n",
472 		    (uint64_t)ma, (uint64_t)va, pteval, level);
473 
474 #if defined(__xpv)
475 	/*
476 	 * see if we can avoid find_pte() on the hypervisor
477 	 */
478 	if (HYPERVISOR_update_va_mapping(va, pteval,
479 	    UVMF_INVLPG | UVMF_LOCAL) == 0)
480 		return;
481 #endif
482 
483 	/*
484 	 * Find the pte that will map this address. This creates any
485 	 * missing intermediate level page tables
486 	 */
487 	ptep = find_pte(va, NULL, level, 0);
488 
489 	/*
490 	 * When paravirtualized, we must use hypervisor calls to modify the
491 	 * PTE, since paging is active. On real hardware we just write to
492 	 * the pagetables which aren't in use yet.
493 	 */
494 #if defined(__xpv)
495 	ptep = ptep;	/* shut lint up */
496 	if (HYPERVISOR_update_va_mapping(va, pteval, UVMF_INVLPG | UVMF_LOCAL))
497 		dboot_panic("mmu_update failed-map_pa_at_va va=0x%" PRIx64
498 		    " l=%d ma=0x%" PRIx64 ", pte=0x%" PRIx64 "",
499 		    (uint64_t)va, level, (uint64_t)ma, pteval);
500 #else
501 	if (va < 1024 * 1024)
502 		pteval |= PT_NOCACHE;		/* for video RAM */
503 	if (pae_support)
504 		*ptep = pteval;
505 	else
506 		*((x86pte32_t *)ptep) = (x86pte32_t)pteval;
507 #endif
508 }
509 
510 /*
511  * Add a mapping for the physical page at the given virtual address.
512  */
513 static void
514 map_pa_at_va(paddr_t pa, native_ptr_t va, uint_t level)
515 {
516 	map_ma_at_va(pa_to_ma(pa), va, level);
517 }
518 
519 /*
520  * This is called to remove start..end from the
521  * possible range of PCI addresses.
522  */
523 const uint64_t pci_lo_limit = 0x00100000ul;
524 const uint64_t pci_hi_limit = 0xfff00000ul;
525 static void
526 exclude_from_pci(uint64_t start, uint64_t end)
527 {
528 	int i;
529 	int j;
530 	struct boot_memlist *ml;
531 
532 	for (i = 0; i < pcimemlists_used; ++i) {
533 		ml = &pcimemlists[i];
534 
535 		/* delete the entire range? */
536 		if (start <= ml->addr && ml->addr + ml->size <= end) {
537 			--pcimemlists_used;
538 			for (j = i; j < pcimemlists_used; ++j)
539 				pcimemlists[j] = pcimemlists[j + 1];
540 			--i;	/* to revisit the new one at this index */
541 		}
542 
543 		/* split a range? */
544 		else if (ml->addr < start && end < ml->addr + ml->size) {
545 
546 			++pcimemlists_used;
547 			if (pcimemlists_used > MAX_MEMLIST)
548 				dboot_panic("too many pcimemlists");
549 
550 			for (j = pcimemlists_used - 1; j > i; --j)
551 				pcimemlists[j] = pcimemlists[j - 1];
552 			ml->size = start - ml->addr;
553 
554 			++ml;
555 			ml->size = (ml->addr + ml->size) - end;
556 			ml->addr = end;
557 			++i;	/* skip on to next one */
558 		}
559 
560 		/* cut memory off the start? */
561 		else if (ml->addr < end && end < ml->addr + ml->size) {
562 			ml->size -= end - ml->addr;
563 			ml->addr = end;
564 		}
565 
566 		/* cut memory off the end? */
567 		else if (ml->addr <= start && start < ml->addr + ml->size) {
568 			ml->size = start - ml->addr;
569 		}
570 	}
571 }
572 
573 /*
574  * Xen strips the size field out of the mb_memory_map_t, see struct e820entry
575  * definition in Xen source.
576  */
577 #ifdef __xpv
578 typedef struct {
579 	uint32_t	base_addr_low;
580 	uint32_t	base_addr_high;
581 	uint32_t	length_low;
582 	uint32_t	length_high;
583 	uint32_t	type;
584 } mmap_t;
585 #else
586 typedef mb_memory_map_t mmap_t;
587 #endif
588 
589 static void
590 build_pcimemlists(mmap_t *mem, int num)
591 {
592 	mmap_t *mmap;
593 	uint64_t page_offset = MMU_PAGEOFFSET;	/* needs to be 64 bits */
594 	uint64_t start;
595 	uint64_t end;
596 	int i;
597 
598 	/*
599 	 * initialize
600 	 */
601 	pcimemlists[0].addr = pci_lo_limit;
602 	pcimemlists[0].size = pci_hi_limit - pci_lo_limit;
603 	pcimemlists_used = 1;
604 
605 	/*
606 	 * Fill in PCI memlists.
607 	 */
608 	for (mmap = mem, i = 0; i < num; ++i, ++mmap) {
609 		start = ((uint64_t)mmap->base_addr_high << 32) +
610 		    mmap->base_addr_low;
611 		end = start + ((uint64_t)mmap->length_high << 32) +
612 		    mmap->length_low;
613 
614 		if (prom_debug)
615 			dboot_printf("\ttype: %d %" PRIx64 "..%"
616 			    PRIx64 "\n", mmap->type, start, end);
617 
618 		/*
619 		 * page align start and end
620 		 */
621 		start = (start + page_offset) & ~page_offset;
622 		end &= ~page_offset;
623 		if (end <= start)
624 			continue;
625 
626 		exclude_from_pci(start, end);
627 	}
628 
629 	/*
630 	 * Finish off the pcimemlist
631 	 */
632 	if (prom_debug) {
633 		for (i = 0; i < pcimemlists_used; ++i) {
634 			dboot_printf("pcimemlist entry 0x%" PRIx64 "..0x%"
635 			    PRIx64 "\n", pcimemlists[i].addr,
636 			    pcimemlists[i].addr + pcimemlists[i].size);
637 		}
638 	}
639 	pcimemlists[0].next = 0;
640 	pcimemlists[0].prev = 0;
641 	for (i = 1; i < pcimemlists_used; ++i) {
642 		pcimemlists[i].prev =
643 		    (native_ptr_t)(uintptr_t)(pcimemlists + i - 1);
644 		pcimemlists[i].next = 0;
645 		pcimemlists[i - 1].next =
646 		    (native_ptr_t)(uintptr_t)(pcimemlists + i);
647 	}
648 	bi->bi_pcimem = (native_ptr_t)pcimemlists;
649 	DBG(bi->bi_pcimem);
650 }
651 
652 #if defined(__xpv)
653 /*
654  * Initialize memory allocator stuff from hypervisor-supplied start info.
655  *
656  * There is 512KB of scratch area after the boot stack page.
657  * We'll use that for everything except the kernel nucleus pages which are too
658  * big to fit there and are allocated last anyway.
659  */
660 #define	MAXMAPS	100
661 static mmap_t map_buffer[MAXMAPS];
662 static void
663 init_mem_alloc(void)
664 {
665 	int	local;	/* variables needed to find start region */
666 	paddr_t	scratch_start;
667 	xen_memory_map_t map;
668 
669 	DBG_MSG("Entered init_mem_alloc()\n");
670 
671 	/*
672 	 * Free memory follows the stack. There's at least 512KB of scratch
673 	 * space, rounded up to at least 2Mb alignment.  That should be enough
674 	 * for the page tables we'll need to build.  The nucleus memory is
675 	 * allocated last and will be outside the addressible range.  We'll
676 	 * switch to new page tables before we unpack the kernel
677 	 */
678 	scratch_start = RNDUP((paddr_t)(uintptr_t)&local, MMU_PAGESIZE);
679 	DBG(scratch_start);
680 	scratch_end = RNDUP((paddr_t)scratch_start + 512 * 1024, TWO_MEG);
681 	DBG(scratch_end);
682 
683 	/*
684 	 * For paranoia, leave some space between hypervisor data and ours.
685 	 * Use 500 instead of 512.
686 	 */
687 	next_avail_addr = scratch_end - 500 * 1024;
688 	DBG(next_avail_addr);
689 
690 	/*
691 	 * The domain builder gives us at most 1 module
692 	 */
693 	DBG(xen_info->mod_len);
694 	if (xen_info->mod_len > 0) {
695 		DBG(xen_info->mod_start);
696 		modules[0].bm_addr = xen_info->mod_start;
697 		modules[0].bm_size = xen_info->mod_len;
698 		bi->bi_module_cnt = 1;
699 		bi->bi_modules = (native_ptr_t)modules;
700 	} else {
701 		bi->bi_module_cnt = 0;
702 		bi->bi_modules = NULL;
703 	}
704 	DBG(bi->bi_module_cnt);
705 	DBG(bi->bi_modules);
706 
707 	DBG(xen_info->mfn_list);
708 	DBG(xen_info->nr_pages);
709 	max_mem = (paddr_t)xen_info->nr_pages << MMU_PAGESHIFT;
710 	DBG(max_mem);
711 
712 	/*
713 	 * Using pseudo-physical addresses, so only 1 memlist element
714 	 */
715 	memlists[0].addr = 0;
716 	DBG(memlists[0].addr);
717 	memlists[0].size = max_mem;
718 	DBG(memlists[0].size);
719 	memlists_used = 1;
720 	DBG(memlists_used);
721 
722 	/*
723 	 * finish building physinstall list
724 	 */
725 	sort_physinstall();
726 
727 	if (DOMAIN_IS_INITDOMAIN(xen_info)) {
728 		/*
729 		 * build PCI Memory list
730 		 */
731 		map.nr_entries = MAXMAPS;
732 		/*LINTED: constant in conditional context*/
733 		set_xen_guest_handle(map.buffer, map_buffer);
734 		if (HYPERVISOR_memory_op(XENMEM_machine_memory_map, &map) != 0)
735 			dboot_panic("getting XENMEM_machine_memory_map failed");
736 		build_pcimemlists(map_buffer, map.nr_entries);
737 	}
738 }
739 
740 #else	/* !__xpv */
741 
742 /*
743  * During memory allocation, find the highest address not used yet.
744  */
745 static void
746 check_higher(paddr_t a)
747 {
748 	if (a < next_avail_addr)
749 		return;
750 	next_avail_addr = RNDUP(a + 1, MMU_PAGESIZE);
751 	DBG(next_avail_addr);
752 }
753 
754 /*
755  * Walk through the module information finding the last used address.
756  * The first available address will become the top level page table.
757  *
758  * We then build the phys_install memlist from the multiboot information.
759  */
760 static void
761 init_mem_alloc(void)
762 {
763 	mb_memory_map_t *mmap;
764 	mb_module_t *mod;
765 	uint64_t start;
766 	uint64_t end;
767 	uint64_t page_offset = MMU_PAGEOFFSET;	/* needs to be 64 bits */
768 	extern char _end[];
769 	int i;
770 
771 	DBG_MSG("Entered init_mem_alloc()\n");
772 	DBG((uintptr_t)mb_info);
773 
774 	/*
775 	 * search the modules to find the last used address
776 	 * we'll build the module list while we're walking through here
777 	 */
778 	DBG_MSG("\nFinding Modules\n");
779 	check_higher((paddr_t)&_end);
780 	for (mod = (mb_module_t *)(mb_info->mods_addr), i = 0;
781 	    i < mb_info->mods_count;
782 	    ++mod, ++i) {
783 		if (prom_debug) {
784 			dboot_printf("\tmodule #%d: %s at: 0x%lx, len 0x%lx\n",
785 			    i, (char *)(mod->mod_name),
786 			    (ulong_t)mod->mod_start, (ulong_t)mod->mod_end);
787 		}
788 		modules[i].bm_addr = mod->mod_start;
789 		modules[i].bm_size = mod->mod_end;
790 
791 		check_higher(mod->mod_end);
792 	}
793 	bi->bi_modules = (native_ptr_t)modules;
794 	DBG(bi->bi_modules);
795 	bi->bi_module_cnt = mb_info->mods_count;
796 	DBG(bi->bi_module_cnt);
797 
798 	/*
799 	 * Walk through the memory map from multiboot and build our memlist
800 	 * structures. Note these will have native format pointers.
801 	 */
802 	DBG_MSG("\nFinding Memory Map\n");
803 	DBG(mb_info->flags);
804 	max_mem = 0;
805 	if (mb_info->flags & 0x40) {
806 		int cnt = 0;
807 
808 		DBG(mb_info->mmap_addr);
809 		DBG(mb_info->mmap_length);
810 		check_higher(mb_info->mmap_addr + mb_info->mmap_length);
811 
812 		for (mmap = (mb_memory_map_t *)mb_info->mmap_addr;
813 		    (uint32_t)mmap < mb_info->mmap_addr + mb_info->mmap_length;
814 		    mmap = (mb_memory_map_t *)((uint32_t)mmap + mmap->size
815 		    + sizeof (mmap->size))) {
816 			++cnt;
817 			start = ((uint64_t)mmap->base_addr_high << 32) +
818 			    mmap->base_addr_low;
819 			end = start + ((uint64_t)mmap->length_high << 32) +
820 			    mmap->length_low;
821 
822 			if (prom_debug)
823 				dboot_printf("\ttype: %d %" PRIx64 "..%"
824 				    PRIx64 "\n", mmap->type, start, end);
825 
826 			/*
827 			 * page align start and end
828 			 */
829 			start = (start + page_offset) & ~page_offset;
830 			end &= ~page_offset;
831 			if (end <= start)
832 				continue;
833 
834 			/*
835 			 * only type 1 is usable RAM
836 			 */
837 			if (mmap->type != 1)
838 				continue;
839 
840 			if (end > max_mem)
841 				max_mem = end;
842 
843 			memlists[memlists_used].addr = start;
844 			memlists[memlists_used].size = end - start;
845 			++memlists_used;
846 			if (memlists_used > MAX_MEMLIST)
847 				dboot_panic("too many memlists");
848 		}
849 		build_pcimemlists((mb_memory_map_t *)mb_info->mmap_addr, cnt);
850 	} else if (mb_info->flags & 0x01) {
851 		DBG(mb_info->mem_lower);
852 		memlists[memlists_used].addr = 0;
853 		memlists[memlists_used].size = mb_info->mem_lower * 1024;
854 		++memlists_used;
855 		DBG(mb_info->mem_upper);
856 		memlists[memlists_used].addr = 1024 * 1024;
857 		memlists[memlists_used].size = mb_info->mem_upper * 1024;
858 		++memlists_used;
859 
860 		/*
861 		 * Old platform - assume I/O space at the end of memory.
862 		 */
863 		pcimemlists[0].addr =
864 		    (mb_info->mem_upper * 1024) + (1024 * 1024);
865 		pcimemlists[0].size = pci_hi_limit - pcimemlists[0].addr;
866 		pcimemlists[0].next = 0;
867 		pcimemlists[0].prev = 0;
868 		bi->bi_pcimem = (native_ptr_t)pcimemlists;
869 		DBG(bi->bi_pcimem);
870 	} else {
871 		dboot_panic("No memory info from boot loader!!!");
872 	}
873 
874 	check_higher(bi->bi_cmdline);
875 
876 	/*
877 	 * finish processing the physinstall list
878 	 */
879 	sort_physinstall();
880 }
881 #endif /* !__xpv */
882 
883 /*
884  * Simple memory allocator, allocates aligned physical memory.
885  * Note that startup_kernel() only allocates memory, never frees.
886  * Memory usage just grows in an upward direction.
887  */
888 static void *
889 do_mem_alloc(uint32_t size, uint32_t align)
890 {
891 	uint_t i;
892 	uint64_t best;
893 	uint64_t start;
894 	uint64_t end;
895 
896 	/*
897 	 * make sure size is a multiple of pagesize
898 	 */
899 	size = RNDUP(size, MMU_PAGESIZE);
900 	next_avail_addr = RNDUP(next_avail_addr, align);
901 
902 	/*
903 	 * XXPV fixme joe
904 	 *
905 	 * a really large bootarchive that causes you to run out of memory
906 	 * may cause this to blow up
907 	 */
908 	/* LINTED E_UNEXPECTED_UINT_PROMOTION */
909 	best = (uint64_t)-size;
910 	for (i = 0; i < memlists_used; ++i) {
911 		start = memlists[i].addr;
912 #if defined(__xpv)
913 		start += mfn_base;
914 #endif
915 		end = start + memlists[i].size;
916 
917 		/*
918 		 * did we find the desired address?
919 		 */
920 		if (start <= next_avail_addr && next_avail_addr + size <= end) {
921 			best = next_avail_addr;
922 			goto done;
923 		}
924 
925 		/*
926 		 * if not is this address the best so far?
927 		 */
928 		if (start > next_avail_addr && start < best &&
929 		    RNDUP(start, align) + size <= end)
930 			best = RNDUP(start, align);
931 	}
932 
933 	/*
934 	 * We didn't find exactly the address we wanted, due to going off the
935 	 * end of a memory region. Return the best found memory address.
936 	 */
937 done:
938 	next_avail_addr = best + size;
939 #if defined(__xpv)
940 	if (next_avail_addr > scratch_end)
941 		dboot_panic("Out of mem next_avail: 0x%lx, scratch_end: "
942 		    "0x%lx", (ulong_t)next_avail_addr,
943 		    (ulong_t)scratch_end);
944 #endif
945 	(void) memset((void *)(uintptr_t)best, 0, size);
946 	return ((void *)(uintptr_t)best);
947 }
948 
949 void *
950 mem_alloc(uint32_t size)
951 {
952 	return (do_mem_alloc(size, MMU_PAGESIZE));
953 }
954 
955 
956 /*
957  * Build page tables to map all of memory used so far as well as the kernel.
958  */
959 static void
960 build_page_tables(void)
961 {
962 	uint32_t psize;
963 	uint32_t level;
964 	uint32_t off;
965 	uint64_t start;
966 #if !defined(__xpv)
967 	uint32_t i;
968 	uint64_t end;
969 #endif	/* __xpv */
970 
971 	/*
972 	 * If we're on metal, we need to create the top level pagetable.
973 	 */
974 #if defined(__xpv)
975 	top_page_table = (paddr_t)(uintptr_t)xen_info->pt_base;
976 #else /* __xpv */
977 	top_page_table = (paddr_t)(uintptr_t)mem_alloc(MMU_PAGESIZE);
978 #endif /* __xpv */
979 	DBG((uintptr_t)top_page_table);
980 
981 	/*
982 	 * Determine if we'll use large mappings for kernel, then map it.
983 	 */
984 	if (largepage_support) {
985 		psize = lpagesize;
986 		level = 1;
987 	} else {
988 		psize = MMU_PAGESIZE;
989 		level = 0;
990 	}
991 
992 	DBG_MSG("Mapping kernel\n");
993 	DBG(ktext_phys);
994 	DBG(target_kernel_text);
995 	DBG(ksize);
996 	DBG(psize);
997 	for (off = 0; off < ksize; off += psize)
998 		map_pa_at_va(ktext_phys + off, target_kernel_text + off, level);
999 
1000 	/*
1001 	 * The kernel will need a 1 page window to work with page tables
1002 	 */
1003 	bi->bi_pt_window = (uintptr_t)mem_alloc(MMU_PAGESIZE);
1004 	DBG(bi->bi_pt_window);
1005 	bi->bi_pte_to_pt_window =
1006 	    (uintptr_t)find_pte(bi->bi_pt_window, NULL, 0, 0);
1007 	DBG(bi->bi_pte_to_pt_window);
1008 
1009 #if defined(__xpv)
1010 	if (!DOMAIN_IS_INITDOMAIN(xen_info)) {
1011 		/* If this is a domU we're done. */
1012 		DBG_MSG("\nPage tables constructed\n");
1013 		return;
1014 	}
1015 #endif /* __xpv */
1016 
1017 	/*
1018 	 * We need 1:1 mappings for the lower 1M of memory to access
1019 	 * BIOS tables used by a couple of drivers during boot.
1020 	 *
1021 	 * The following code works because our simple memory allocator
1022 	 * only grows usage in an upwards direction.
1023 	 *
1024 	 * Note that by this point in boot some mappings for low memory
1025 	 * may already exist because we've already accessed device in low
1026 	 * memory.  (Specifically the video frame buffer and keyboard
1027 	 * status ports.)  If we're booting on raw hardware then GRUB
1028 	 * created these mappings for us.  If we're booting under a
1029 	 * hypervisor then we went ahead and remapped these devices into
1030 	 * memory allocated within dboot itself.
1031 	 */
1032 	if (map_debug)
1033 		dboot_printf("1:1 map pa=0..1Meg\n");
1034 	for (start = 0; start < 1024 * 1024; start += MMU_PAGESIZE) {
1035 #if defined(__xpv)
1036 		map_ma_at_va(start, start, 0);
1037 #else /* __xpv */
1038 		map_pa_at_va(start, start, 0);
1039 #endif /* __xpv */
1040 	}
1041 
1042 #if !defined(__xpv)
1043 	for (i = 0; i < memlists_used; ++i) {
1044 		start = memlists[i].addr;
1045 
1046 		end = start + memlists[i].size;
1047 
1048 		if (map_debug)
1049 			dboot_printf("1:1 map pa=%" PRIx64 "..%" PRIx64 "\n",
1050 			    start, end);
1051 		while (start < end && start < next_avail_addr) {
1052 			map_pa_at_va(start, start, 0);
1053 			start += MMU_PAGESIZE;
1054 		}
1055 	}
1056 #endif /* !__xpv */
1057 
1058 	DBG_MSG("\nPage tables constructed\n");
1059 }
1060 
1061 #define	NO_MULTIBOOT	\
1062 "multiboot is no longer used to boot the Solaris Operating System.\n\
1063 The grub entry should be changed to:\n\
1064 kernel$ /platform/i86pc/kernel/$ISADIR/unix\n\
1065 module$ /platform/i86pc/$ISADIR/boot_archive\n\
1066 See http://www.sun.com/msg/SUNOS-8000-AK for details.\n"
1067 
1068 /*
1069  * startup_kernel has a pretty simple job. It builds pagetables which reflect
1070  * 1:1 mappings for all memory in use. It then also adds mappings for
1071  * the kernel nucleus at virtual address of target_kernel_text using large page
1072  * mappings. The page table pages are also accessible at 1:1 mapped
1073  * virtual addresses.
1074  */
1075 /*ARGSUSED*/
1076 void
1077 startup_kernel(void)
1078 {
1079 	char *cmdline;
1080 	uintptr_t addr;
1081 #if defined(__xpv)
1082 	physdev_set_iopl_t set_iopl;
1083 #endif /* __xpv */
1084 
1085 	/*
1086 	 * At this point we are executing in a 32 bit real mode.
1087 	 */
1088 #if defined(__xpv)
1089 	cmdline = (char *)xen_info->cmd_line;
1090 #else /* __xpv */
1091 	cmdline = (char *)mb_info->cmdline;
1092 #endif /* __xpv */
1093 
1094 	prom_debug = (strstr(cmdline, "prom_debug") != NULL);
1095 	map_debug = (strstr(cmdline, "map_debug") != NULL);
1096 
1097 #if defined(__xpv)
1098 	/*
1099 	 * For dom0, before we initialize the console subsystem we'll
1100 	 * need to enable io operations, so set I/O priveldge level to 1.
1101 	 */
1102 	if (DOMAIN_IS_INITDOMAIN(xen_info)) {
1103 		set_iopl.iopl = 1;
1104 		(void) HYPERVISOR_physdev_op(PHYSDEVOP_set_iopl, &set_iopl);
1105 	}
1106 #endif /* __xpv */
1107 
1108 	bcons_init(cmdline);
1109 	DBG_MSG("\n\nSolaris prekernel set: ");
1110 	DBG_MSG(cmdline);
1111 	DBG_MSG("\n");
1112 
1113 	if (strstr(cmdline, "multiboot") != NULL) {
1114 		dboot_panic(NO_MULTIBOOT);
1115 	}
1116 
1117 	/*
1118 	 * boot info must be 16 byte aligned for 64 bit kernel ABI
1119 	 */
1120 	addr = (uintptr_t)boot_info;
1121 	addr = (addr + 0xf) & ~0xf;
1122 	bi = (struct xboot_info *)addr;
1123 	DBG((uintptr_t)bi);
1124 	bi->bi_cmdline = (native_ptr_t)(uintptr_t)cmdline;
1125 
1126 	/*
1127 	 * Need correct target_kernel_text value
1128 	 */
1129 #if defined(_BOOT_TARGET_amd64)
1130 	target_kernel_text = KERNEL_TEXT_amd64;
1131 #elif defined(__xpv)
1132 	target_kernel_text = KERNEL_TEXT_i386_xpv;
1133 #else
1134 	target_kernel_text = KERNEL_TEXT_i386;
1135 #endif
1136 	DBG(target_kernel_text);
1137 
1138 #if defined(__xpv)
1139 
1140 	/*
1141 	 * XXPV	Derive this stuff from CPUID / what the hypervisor has enabled
1142 	 */
1143 
1144 #if defined(_BOOT_TARGET_amd64)
1145 	/*
1146 	 * 64-bit hypervisor.
1147 	 */
1148 	amd64_support = 1;
1149 	pae_support = 1;
1150 
1151 #else	/* _BOOT_TARGET_amd64 */
1152 
1153 	/*
1154 	 * See if we are running on a PAE Hypervisor
1155 	 */
1156 	{
1157 		xen_capabilities_info_t caps;
1158 
1159 		if (HYPERVISOR_xen_version(XENVER_capabilities, &caps) != 0)
1160 			dboot_panic("HYPERVISOR_xen_version(caps) failed");
1161 		caps[sizeof (caps) - 1] = 0;
1162 		if (prom_debug)
1163 			dboot_printf("xen capabilities %s\n", caps);
1164 		if (strstr(caps, "x86_32p") != NULL)
1165 			pae_support = 1;
1166 	}
1167 
1168 #endif	/* _BOOT_TARGET_amd64 */
1169 	{
1170 		xen_platform_parameters_t p;
1171 
1172 		if (HYPERVISOR_xen_version(XENVER_platform_parameters, &p) != 0)
1173 			dboot_panic("HYPERVISOR_xen_version(parms) failed");
1174 		DBG(p.virt_start);
1175 		mfn_to_pfn_mapping = (pfn_t *)(xen_virt_start = p.virt_start);
1176 	}
1177 
1178 	/*
1179 	 * The hypervisor loads stuff starting at 1Gig
1180 	 */
1181 	mfn_base = ONE_GIG;
1182 	DBG(mfn_base);
1183 
1184 	/*
1185 	 * enable writable page table mode for the hypervisor
1186 	 */
1187 	if (HYPERVISOR_vm_assist(VMASST_CMD_enable,
1188 	    VMASST_TYPE_writable_pagetables) < 0)
1189 		dboot_panic("HYPERVISOR_vm_assist(writable_pagetables) failed");
1190 
1191 	/*
1192 	 * check for NX support
1193 	 */
1194 	if (pae_support) {
1195 		uint32_t eax = 0x80000000;
1196 		uint32_t edx = get_cpuid_edx(&eax);
1197 
1198 		if (eax >= 0x80000001) {
1199 			eax = 0x80000001;
1200 			edx = get_cpuid_edx(&eax);
1201 			if (edx & CPUID_AMD_EDX_NX)
1202 				NX_support = 1;
1203 		}
1204 	}
1205 
1206 #if !defined(_BOOT_TARGET_amd64)
1207 
1208 	/*
1209 	 * The 32-bit hypervisor uses segmentation to protect itself from
1210 	 * guests. This means when a guest attempts to install a flat 4GB
1211 	 * code or data descriptor the 32-bit hypervisor will protect itself
1212 	 * by silently shrinking the segment such that if the guest attempts
1213 	 * any access where the hypervisor lives a #gp fault is generated.
1214 	 * The problem is that some applications expect a full 4GB flat
1215 	 * segment for their current thread pointer and will use negative
1216 	 * offset segment wrap around to access data. TLS support in linux
1217 	 * brand is one example of this.
1218 	 *
1219 	 * The 32-bit hypervisor can catch the #gp fault in these cases
1220 	 * and emulate the access without passing the #gp fault to the guest
1221 	 * but only if VMASST_TYPE_4gb_segments is explicitly turned on.
1222 	 * Seems like this should have been the default.
1223 	 * Either way, we want the hypervisor -- and not Solaris -- to deal
1224 	 * to deal with emulating these accesses.
1225 	 */
1226 	if (HYPERVISOR_vm_assist(VMASST_CMD_enable,
1227 	    VMASST_TYPE_4gb_segments) < 0)
1228 		dboot_panic("HYPERVISOR_vm_assist(4gb_segments) failed");
1229 #endif	/* !_BOOT_TARGET_amd64 */
1230 
1231 #else	/* __xpv */
1232 
1233 	/*
1234 	 * use cpuid to enable MMU features
1235 	 */
1236 	if (have_cpuid()) {
1237 		uint32_t eax, edx;
1238 
1239 		eax = 1;
1240 		edx = get_cpuid_edx(&eax);
1241 		if (edx & CPUID_INTC_EDX_PSE)
1242 			largepage_support = 1;
1243 		if (edx & CPUID_INTC_EDX_PGE)
1244 			pge_support = 1;
1245 		if (edx & CPUID_INTC_EDX_PAE)
1246 			pae_support = 1;
1247 
1248 		eax = 0x80000000;
1249 		edx = get_cpuid_edx(&eax);
1250 		if (eax >= 0x80000001) {
1251 			eax = 0x80000001;
1252 			edx = get_cpuid_edx(&eax);
1253 			if (edx & CPUID_AMD_EDX_LM)
1254 				amd64_support = 1;
1255 			if (edx & CPUID_AMD_EDX_NX)
1256 				NX_support = 1;
1257 		}
1258 	} else {
1259 		dboot_printf("cpuid not supported\n");
1260 	}
1261 #endif /* __xpv */
1262 
1263 
1264 #if defined(_BOOT_TARGET_amd64)
1265 	if (amd64_support == 0)
1266 		dboot_panic("long mode not supported, rebooting");
1267 	else if (pae_support == 0)
1268 		dboot_panic("long mode, but no PAE; rebooting");
1269 #else
1270 	/*
1271 	 * Allow the command line to over-ride use of PAE for 32 bit.
1272 	 */
1273 	if (strstr(cmdline, "disablePAE=true") != NULL) {
1274 		pae_support = 0;
1275 		NX_support = 0;
1276 		amd64_support = 0;
1277 	}
1278 #endif
1279 
1280 	/*
1281 	 * initialize the simple memory allocator
1282 	 */
1283 	init_mem_alloc();
1284 
1285 #if !defined(__xpv) && !defined(_BOOT_TARGET_amd64)
1286 	/*
1287 	 * disable PAE on 32 bit h/w w/o NX and < 4Gig of memory
1288 	 */
1289 	if (max_mem < FOUR_GIG && NX_support == 0)
1290 		pae_support = 0;
1291 #endif
1292 
1293 	/*
1294 	 * configure mmu information
1295 	 */
1296 	if (pae_support) {
1297 		shift_amt = shift_amt_pae;
1298 		ptes_per_table = 512;
1299 		pte_size = 8;
1300 		lpagesize = TWO_MEG;
1301 #if defined(_BOOT_TARGET_amd64)
1302 		top_level = 3;
1303 #else
1304 		top_level = 2;
1305 #endif
1306 	} else {
1307 		pae_support = 0;
1308 		NX_support = 0;
1309 		shift_amt = shift_amt_nopae;
1310 		ptes_per_table = 1024;
1311 		pte_size = 4;
1312 		lpagesize = FOUR_MEG;
1313 		top_level = 1;
1314 	}
1315 
1316 	DBG(pge_support);
1317 	DBG(NX_support);
1318 	DBG(largepage_support);
1319 	DBG(amd64_support);
1320 	DBG(top_level);
1321 	DBG(pte_size);
1322 	DBG(ptes_per_table);
1323 	DBG(lpagesize);
1324 
1325 #if defined(__xpv)
1326 	ktext_phys = ONE_GIG;		/* from UNIX Mapfile */
1327 #else
1328 	ktext_phys = FOUR_MEG;		/* from UNIX Mapfile */
1329 #endif
1330 
1331 #if !defined(__xpv) && defined(_BOOT_TARGET_amd64)
1332 	/*
1333 	 * For grub, copy kernel bits from the ELF64 file to final place.
1334 	 */
1335 	DBG_MSG("\nAllocating nucleus pages.\n");
1336 	ktext_phys = (uintptr_t)do_mem_alloc(ksize, FOUR_MEG);
1337 	if (ktext_phys == 0)
1338 		dboot_panic("failed to allocate aligned kernel memory");
1339 	if (dboot_elfload64(mb_header.load_addr) != 0)
1340 		dboot_panic("failed to parse kernel ELF image, rebooting");
1341 #endif
1342 
1343 	DBG(ktext_phys);
1344 
1345 	/*
1346 	 * Allocate page tables.
1347 	 */
1348 	build_page_tables();
1349 
1350 	/*
1351 	 * return to assembly code to switch to running kernel
1352 	 */
1353 	entry_addr_low = (uint32_t)target_kernel_text;
1354 	DBG(entry_addr_low);
1355 	bi->bi_use_largepage = largepage_support;
1356 	bi->bi_use_pae = pae_support;
1357 	bi->bi_use_pge = pge_support;
1358 	bi->bi_use_nx = NX_support;
1359 
1360 #if defined(__xpv)
1361 
1362 	bi->bi_next_paddr = next_avail_addr - mfn_base;
1363 	DBG(bi->bi_next_paddr);
1364 	bi->bi_next_vaddr = (native_ptr_t)next_avail_addr;
1365 	DBG(bi->bi_next_vaddr);
1366 
1367 	/*
1368 	 * unmap unused pages in start area to make them available for DMA
1369 	 */
1370 	while (next_avail_addr < scratch_end) {
1371 		(void) HYPERVISOR_update_va_mapping(next_avail_addr,
1372 		    0, UVMF_INVLPG | UVMF_LOCAL);
1373 		next_avail_addr += MMU_PAGESIZE;
1374 	}
1375 
1376 	bi->bi_xen_start_info = (uintptr_t)xen_info;
1377 	DBG((uintptr_t)HYPERVISOR_shared_info);
1378 	bi->bi_shared_info = (native_ptr_t)HYPERVISOR_shared_info;
1379 	bi->bi_top_page_table = (uintptr_t)top_page_table - mfn_base;
1380 
1381 #else /* __xpv */
1382 
1383 	bi->bi_next_paddr = next_avail_addr;
1384 	DBG(bi->bi_next_paddr);
1385 	bi->bi_next_vaddr = (uintptr_t)next_avail_addr;
1386 	DBG(bi->bi_next_vaddr);
1387 	bi->bi_mb_info = (uintptr_t)mb_info;
1388 	bi->bi_top_page_table = (uintptr_t)top_page_table;
1389 
1390 #endif /* __xpv */
1391 
1392 	bi->bi_kseg_size = FOUR_MEG;
1393 	DBG(bi->bi_kseg_size);
1394 
1395 #ifndef __xpv
1396 	if (prom_debug)
1397 		dump_tables();
1398 #endif
1399 
1400 	DBG_MSG("\n\n*** DBOOT DONE -- back to asm to jump to kernel\n\n");
1401 }
1402