1 /* $NetBSD: arm32_machdep.c,v 1.44 2004/03/24 15:34:47 atatat Exp $ */ 2 3 /*- 4 * SPDX-License-Identifier: BSD-4-Clause 5 * 6 * Copyright (c) 2004 Olivier Houchard 7 * Copyright (c) 1994-1998 Mark Brinicombe. 8 * Copyright (c) 1994 Brini. 9 * All rights reserved. 10 * 11 * This code is derived from software written for Brini by Mark Brinicombe 12 * 13 * Redistribution and use in source and binary forms, with or without 14 * modification, are permitted provided that the following conditions 15 * are met: 16 * 1. Redistributions of source code must retain the above copyright 17 * notice, this list of conditions and the following disclaimer. 18 * 2. Redistributions in binary form must reproduce the above copyright 19 * notice, this list of conditions and the following disclaimer in the 20 * documentation and/or other materials provided with the distribution. 21 * 3. All advertising materials mentioning features or use of this software 22 * must display the following acknowledgement: 23 * This product includes software developed by Mark Brinicombe 24 * for the NetBSD Project. 25 * 4. The name of the company nor the name of the author may be used to 26 * endorse or promote products derived from this software without specific 27 * prior written permission. 28 * 29 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED 30 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF 31 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 32 * IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, 33 * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES 34 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR 35 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 36 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 37 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 38 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 39 * SUCH DAMAGE. 40 * 41 * Machine dependent functions for kernel setup 42 * 43 * Created : 17/09/94 44 * Updated : 18/04/01 updated for new wscons 45 */ 46 47 #include "opt_ddb.h" 48 #include "opt_kstack_pages.h" 49 #include "opt_platform.h" 50 #include "opt_sched.h" 51 52 #include <sys/param.h> 53 #include <sys/buf.h> 54 #include <sys/bus.h> 55 #include <sys/cons.h> 56 #include <sys/cpu.h> 57 #include <sys/devmap.h> 58 #include <sys/efi.h> 59 #include <sys/efi_map.h> 60 #include <sys/imgact.h> 61 #include <sys/kdb.h> 62 #include <sys/kernel.h> 63 #include <sys/ktr.h> 64 #include <sys/linker.h> 65 #include <sys/msgbuf.h> 66 #include <sys/physmem.h> 67 #include <sys/reboot.h> 68 #include <sys/rwlock.h> 69 #include <sys/sched.h> 70 #include <sys/syscallsubr.h> 71 #include <sys/sysent.h> 72 #include <sys/sysproto.h> 73 #include <sys/vmmeter.h> 74 75 #include <vm/vm_object.h> 76 #include <vm/vm_page.h> 77 #include <vm/vm_pager.h> 78 79 #include <machine/asm.h> 80 #include <machine/debug_monitor.h> 81 #include <machine/machdep.h> 82 #include <machine/metadata.h> 83 #include <machine/pcb.h> 84 #include <machine/platform.h> 85 #include <machine/sysarch.h> 86 #include <machine/undefined.h> 87 #include <machine/vfp.h> 88 #include <machine/vmparam.h> 89 90 #ifdef FDT 91 #include <dev/fdt/fdt_common.h> 92 #include <machine/ofw_machdep.h> 93 #endif 94 95 #ifdef DEBUG 96 #define debugf(fmt, args...) printf(fmt, ##args) 97 #else 98 #define debugf(fmt, args...) 99 #endif 100 101 #if defined(COMPAT_FREEBSD4) || defined(COMPAT_FREEBSD5) || \ 102 defined(COMPAT_FREEBSD6) || defined(COMPAT_FREEBSD7) || \ 103 defined(COMPAT_FREEBSD9) 104 #error FreeBSD/arm doesn't provide compatibility with releases prior to 10 105 #endif 106 107 108 #if __ARM_ARCH < 7 109 #error FreeBSD requires ARMv7 or later 110 #endif 111 112 struct pcpu __pcpu[MAXCPU]; 113 struct pcpu *pcpup = &__pcpu[0]; 114 115 static struct trapframe proc0_tf; 116 uint32_t cpu_reset_address = 0; 117 int cold = 1; 118 vm_offset_t vector_page; 119 120 /* The address at which the kernel was loaded. Set early in initarm(). */ 121 vm_paddr_t arm_physmem_kernaddr; 122 123 extern int *end; 124 125 #ifdef FDT 126 vm_paddr_t pmap_pa; 127 vm_offset_t systempage; 128 vm_offset_t irqstack; 129 vm_offset_t undstack; 130 vm_offset_t abtstack; 131 #endif /* FDT */ 132 133 #ifdef PLATFORM 134 static delay_func *delay_impl; 135 static void *delay_arg; 136 #endif 137 138 #if defined(SOCDEV_PA) 139 #if !defined(SOCDEV_VA) 140 #error SOCDEV_PA defined, but not SOCDEV_VA 141 #endif 142 uintptr_t socdev_va = SOCDEV_VA; 143 #endif 144 145 146 struct kva_md_info kmi; 147 /* 148 * arm32_vector_init: 149 * 150 * Initialize the vector page, and select whether or not to 151 * relocate the vectors. 152 * 153 * NOTE: We expect the vector page to be mapped at its expected 154 * destination. 155 */ 156 157 extern unsigned int page0[], page0_data[]; 158 void arm_vector_init(vm_offset_t va,int which)159 arm_vector_init(vm_offset_t va, int which) 160 { 161 unsigned int *vectors = (int *) va; 162 unsigned int *vectors_data = vectors + (page0_data - page0); 163 int vec; 164 165 /* 166 * Loop through the vectors we're taking over, and copy the 167 * vector's insn and data word. 168 */ 169 for (vec = 0; vec < ARM_NVEC; vec++) { 170 if ((which & (1 << vec)) == 0) { 171 /* Don't want to take over this vector. */ 172 continue; 173 } 174 vectors[vec] = page0[vec]; 175 vectors_data[vec] = page0_data[vec]; 176 } 177 178 /* Now sync the vectors. */ 179 icache_sync(va, (ARM_NVEC * 2) * sizeof(u_int)); 180 181 vector_page = va; 182 } 183 184 static void cpu_startup(void * dummy)185 cpu_startup(void *dummy) 186 { 187 struct pcb *pcb = thread0.td_pcb; 188 const unsigned int mbyte = 1024 * 1024; 189 190 identify_arm_cpu(); 191 192 vm_ksubmap_init(&kmi); 193 194 /* 195 * Display the RAM layout. 196 */ 197 printf("real memory = %ju (%ju MB)\n", 198 (uintmax_t)arm32_ptob(realmem), 199 (uintmax_t)arm32_ptob(realmem) / mbyte); 200 printf("avail memory = %ju (%ju MB)\n", 201 (uintmax_t)arm32_ptob(vm_free_count()), 202 (uintmax_t)arm32_ptob(vm_free_count()) / mbyte); 203 if (bootverbose) { 204 physmem_print_tables(); 205 devmap_print_table(); 206 } 207 208 bufinit(); 209 vm_pager_bufferinit(); 210 pcb->pcb_regs.sf_sp = (u_int)thread0.td_kstack + 211 USPACE_SVC_STACK_TOP; 212 pmap_set_pcb_pagedir(kernel_pmap, pcb); 213 } 214 215 SYSINIT(cpu, SI_SUB_CPU, SI_ORDER_FIRST, cpu_startup, NULL); 216 217 /* 218 * Flush the D-cache for non-DMA I/O so that the I-cache can 219 * be made coherent later. 220 */ 221 void cpu_flush_dcache(void * ptr,size_t len)222 cpu_flush_dcache(void *ptr, size_t len) 223 { 224 225 dcache_wb_poc((vm_offset_t)ptr, (vm_paddr_t)vtophys(ptr), len); 226 } 227 228 /* Get current clock frequency for the given cpu id. */ 229 int cpu_est_clockrate(int cpu_id,uint64_t * rate)230 cpu_est_clockrate(int cpu_id, uint64_t *rate) 231 { 232 struct pcpu *pc; 233 234 pc = pcpu_find(cpu_id); 235 if (pc == NULL || rate == NULL) 236 return (EINVAL); 237 238 if (pc->pc_clock == 0) 239 return (EOPNOTSUPP); 240 241 *rate = pc->pc_clock; 242 243 return (0); 244 } 245 246 void cpu_idle(int busy)247 cpu_idle(int busy) 248 { 249 250 CTR2(KTR_SPARE2, "cpu_idle(%d) at %d", busy, curcpu); 251 spinlock_enter(); 252 if (!busy) 253 cpu_idleclock(); 254 if (!sched_runnable()) 255 cpu_sleep(0); 256 if (!busy) 257 cpu_activeclock(); 258 spinlock_exit(); 259 CTR2(KTR_SPARE2, "cpu_idle(%d) at %d done", busy, curcpu); 260 } 261 262 int cpu_idle_wakeup(int cpu)263 cpu_idle_wakeup(int cpu) 264 { 265 266 return (0); 267 } 268 269 void cpu_initclocks(void)270 cpu_initclocks(void) 271 { 272 273 #ifdef SMP 274 if (PCPU_GET(cpuid) == 0) 275 cpu_initclocks_bsp(); 276 else 277 cpu_initclocks_ap(); 278 #else 279 cpu_initclocks_bsp(); 280 #endif 281 } 282 283 #ifdef PLATFORM 284 void arm_set_delay(delay_func * impl,void * arg)285 arm_set_delay(delay_func *impl, void *arg) 286 { 287 288 KASSERT(impl != NULL, ("No DELAY implementation")); 289 delay_impl = impl; 290 delay_arg = arg; 291 } 292 293 void DELAY(int usec)294 DELAY(int usec) 295 { 296 297 TSENTER(); 298 delay_impl(usec, delay_arg); 299 TSEXIT(); 300 } 301 #endif 302 303 void cpu_pcpu_init(struct pcpu * pcpu,int cpuid,size_t size)304 cpu_pcpu_init(struct pcpu *pcpu, int cpuid, size_t size) 305 { 306 307 pcpu->pc_mpidr = 0xffffffff; 308 } 309 310 void spinlock_enter(void)311 spinlock_enter(void) 312 { 313 struct thread *td; 314 register_t cspr; 315 316 td = curthread; 317 if (td->td_md.md_spinlock_count == 0) { 318 cspr = disable_interrupts(PSR_I); 319 td->td_md.md_spinlock_count = 1; 320 td->td_md.md_saved_cspr = cspr; 321 critical_enter(); 322 } else 323 td->td_md.md_spinlock_count++; 324 } 325 326 void spinlock_exit(void)327 spinlock_exit(void) 328 { 329 struct thread *td; 330 register_t cspr; 331 332 td = curthread; 333 cspr = td->td_md.md_saved_cspr; 334 td->td_md.md_spinlock_count--; 335 if (td->td_md.md_spinlock_count == 0) { 336 critical_exit(); 337 restore_interrupts(cspr); 338 } 339 } 340 341 /* 342 * Construct a PCB from a trapframe. This is called from kdb_trap() where 343 * we want to start a backtrace from the function that caused us to enter 344 * the debugger. We have the context in the trapframe, but base the trace 345 * on the PCB. The PCB doesn't have to be perfect, as long as it contains 346 * enough for a backtrace. 347 */ 348 void makectx(struct trapframe * tf,struct pcb * pcb)349 makectx(struct trapframe *tf, struct pcb *pcb) 350 { 351 pcb->pcb_regs.sf_r4 = tf->tf_r4; 352 pcb->pcb_regs.sf_r5 = tf->tf_r5; 353 pcb->pcb_regs.sf_r6 = tf->tf_r6; 354 pcb->pcb_regs.sf_r7 = tf->tf_r7; 355 pcb->pcb_regs.sf_r8 = tf->tf_r8; 356 pcb->pcb_regs.sf_r9 = tf->tf_r9; 357 pcb->pcb_regs.sf_r10 = tf->tf_r10; 358 pcb->pcb_regs.sf_r11 = tf->tf_r11; 359 pcb->pcb_regs.sf_r12 = tf->tf_r12; 360 pcb->pcb_regs.sf_pc = tf->tf_pc; 361 pcb->pcb_regs.sf_lr = tf->tf_usr_lr; 362 pcb->pcb_regs.sf_sp = tf->tf_usr_sp; 363 } 364 365 void pcpu0_init(void)366 pcpu0_init(void) 367 { 368 set_curthread(&thread0); 369 pcpu_init(pcpup, 0, sizeof(struct pcpu)); 370 pcpup->pc_mpidr = cp15_mpidr_get() & 0xFFFFFF; 371 PCPU_SET(curthread, &thread0); 372 } 373 374 /* 375 * Initialize proc0 376 */ 377 void init_proc0(vm_offset_t kstack)378 init_proc0(vm_offset_t kstack) 379 { 380 proc_linkup0(&proc0, &thread0); 381 thread0.td_kstack = kstack; 382 thread0.td_kstack_pages = kstack_pages; 383 thread0.td_pcb = (struct pcb *)(thread0.td_kstack + 384 thread0.td_kstack_pages * PAGE_SIZE) - 1; 385 thread0.td_pcb->pcb_flags = 0; 386 thread0.td_pcb->pcb_fpflags = 0; 387 thread0.td_pcb->pcb_vfpcpu = -1; 388 thread0.td_pcb->pcb_vfpstate.fpscr = VFPSCR_DN; 389 thread0.td_pcb->pcb_vfpsaved = &thread0.td_pcb->pcb_vfpstate; 390 thread0.td_frame = &proc0_tf; 391 pcpup->pc_curpcb = thread0.td_pcb; 392 } 393 394 void set_stackptrs(int cpu)395 set_stackptrs(int cpu) 396 { 397 398 set_stackptr(PSR_IRQ32_MODE, 399 irqstack + ((IRQ_STACK_SIZE * PAGE_SIZE) * (cpu + 1))); 400 set_stackptr(PSR_ABT32_MODE, 401 abtstack + ((ABT_STACK_SIZE * PAGE_SIZE) * (cpu + 1))); 402 set_stackptr(PSR_UND32_MODE, 403 undstack + ((UND_STACK_SIZE * PAGE_SIZE) * (cpu + 1))); 404 } 405 406 static void arm_kdb_init(void)407 arm_kdb_init(void) 408 { 409 410 kdb_init(); 411 #ifdef KDB 412 if (boothowto & RB_KDB) 413 kdb_enter(KDB_WHY_BOOTFLAGS, "Boot flags requested debugger"); 414 #endif 415 } 416 417 #ifdef FDT 418 static void fdt_physmem_hardware_region_cb(const struct mem_region * mr,void * arg __unused)419 fdt_physmem_hardware_region_cb(const struct mem_region *mr, void *arg __unused) 420 { 421 physmem_hardware_region(mr->mr_start, mr->mr_size); 422 } 423 424 static void fdt_physmem_exclude_region_cb(const struct mem_region * mr,void * arg __unused)425 fdt_physmem_exclude_region_cb(const struct mem_region *mr, void *arg __unused) 426 { 427 physmem_exclude_region(mr->mr_start, mr->mr_size, 428 EXFLAG_NODUMP | EXFLAG_NOALLOC); 429 } 430 431 void * initarm(struct arm_boot_params * abp)432 initarm(struct arm_boot_params *abp) 433 { 434 vm_paddr_t lastaddr; 435 vm_offset_t dtbp, kernelstack, dpcpu; 436 char *env; 437 int err_devmap; 438 phandle_t root; 439 char dts_version[255]; 440 #ifdef EFI 441 struct efi_map_header *efihdr; 442 #endif 443 444 /* get last allocated physical address */ 445 arm_physmem_kernaddr = abp->abp_physaddr; 446 lastaddr = parse_boot_param(abp) - KERNVIRTADDR + arm_physmem_kernaddr; 447 448 set_cpufuncs(); 449 cpuinfo_init(); 450 451 /* 452 * Find the dtb passed in by the boot loader. 453 */ 454 dtbp = MD_FETCH(preload_kmdp, MODINFOMD_DTBP, vm_offset_t); 455 #if defined(FDT_DTB_STATIC) 456 /* 457 * In case the device tree blob was not retrieved (from metadata) try 458 * to use the statically embedded one. 459 */ 460 if (dtbp == (vm_offset_t)NULL) 461 dtbp = (vm_offset_t)&fdt_static_dtb; 462 #endif 463 464 if (OF_install(OFW_FDT, 0) == FALSE) 465 panic("Cannot install FDT"); 466 467 if (OF_init((void *)dtbp) != 0) 468 panic("OF_init failed with the found device tree"); 469 470 #if defined(LINUX_BOOT_ABI) 471 arm_parse_fdt_bootargs(); 472 #endif 473 474 #ifdef EFI 475 efihdr = (struct efi_map_header *)preload_search_info(preload_kmdp, 476 MODINFO_METADATA | MODINFOMD_EFI_MAP); 477 if (efihdr != NULL) { 478 efi_map_add_entries(efihdr); 479 efi_map_exclude_entries(efihdr); 480 } else 481 #endif 482 { 483 /* Grab physical memory regions information from device tree. */ 484 if (fdt_foreach_mem_region(fdt_physmem_hardware_region_cb, 485 NULL) != 0) 486 panic("Cannot get physical memory regions"); 487 488 /* Grab reserved memory regions information from device tree. */ 489 fdt_foreach_reserved_region(fdt_physmem_exclude_region_cb, 490 NULL); 491 } 492 493 /* 494 * Set TEX remapping registers. 495 * Setup kernel page tables and switch to kernel L1 page table. 496 */ 497 pmap_set_tex(); 498 pmap_bootstrap_prepare(lastaddr); 499 500 /* 501 * If EARLY_PRINTF support is enabled, we need to re-establish the 502 * mapping after pmap_bootstrap_prepare() switches to new page tables. 503 * Note that we can only do the remapping if the VA is outside the 504 * kernel, now that we have real virtual (not VA=PA) mappings in effect. 505 * Early printf does not work between the time pmap_set_tex() does 506 * cp15_prrr_set() and this code remaps the VA. 507 */ 508 #if defined(EARLY_PRINTF) && defined(SOCDEV_PA) && defined(SOCDEV_VA) && SOCDEV_VA < KERNBASE 509 pmap_preboot_map_attr(SOCDEV_PA, SOCDEV_VA, 1024 * 1024, 510 VM_PROT_READ | VM_PROT_WRITE, VM_MEMATTR_DEVICE); 511 #endif 512 513 /* 514 * Now that proper page tables are installed, call cpu_setup() to enable 515 * instruction and data caches and other chip-specific features. 516 */ 517 cpu_setup(); 518 519 /* Platform-specific initialisation */ 520 platform_probe_and_attach(); 521 pcpu0_init(); 522 523 /* Do basic tuning, hz etc */ 524 init_param1(); 525 526 /* 527 * Allocate a page for the system page mapped to 0xffff0000 528 * This page will just contain the system vectors and can be 529 * shared by all processes. 530 */ 531 systempage = pmap_preboot_get_pages(1); 532 533 /* Map the vector page. */ 534 pmap_preboot_map_pages(systempage, ARM_VECTORS_HIGH, 1); 535 if (virtual_end >= ARM_VECTORS_HIGH) 536 virtual_end = ARM_VECTORS_HIGH - 1; 537 538 /* Allocate dynamic per-cpu area. */ 539 dpcpu = pmap_preboot_get_vpages(DPCPU_SIZE / PAGE_SIZE); 540 dpcpu_init((void *)dpcpu, 0); 541 542 /* Allocate stacks for all modes */ 543 irqstack = pmap_preboot_get_vpages(IRQ_STACK_SIZE * MAXCPU); 544 abtstack = pmap_preboot_get_vpages(ABT_STACK_SIZE * MAXCPU); 545 undstack = pmap_preboot_get_vpages(UND_STACK_SIZE * MAXCPU ); 546 kernelstack = pmap_preboot_get_vpages(kstack_pages); 547 548 /* Allocate message buffer. */ 549 msgbufp = (void *)pmap_preboot_get_vpages( 550 round_page(msgbufsize) / PAGE_SIZE); 551 552 /* 553 * Pages were allocated during the secondary bootstrap for the 554 * stacks for different CPU modes. 555 * We must now set the r13 registers in the different CPU modes to 556 * point to these stacks. 557 * Since the ARM stacks use STMFD etc. we must set r13 to the top end 558 * of the stack memory. 559 */ 560 set_stackptrs(0); 561 mutex_init(); 562 563 /* Establish static device mappings. */ 564 err_devmap = platform_devmap_init(); 565 devmap_bootstrap(); 566 vm_max_kernel_address = platform_lastaddr(); 567 568 /* 569 * Only after the SOC registers block is mapped we can perform device 570 * tree fixups, as they may attempt to read parameters from hardware. 571 */ 572 OF_interpret("perform-fixup", 0); 573 platform_gpio_init(); 574 cninit(); 575 576 /* 577 * If we made a mapping for EARLY_PRINTF after pmap_bootstrap_prepare(), 578 * undo it now that the normal console printf works. 579 */ 580 #if defined(EARLY_PRINTF) && defined(SOCDEV_PA) && defined(SOCDEV_VA) && SOCDEV_VA < KERNBASE 581 pmap_kremove(SOCDEV_VA); 582 #endif 583 584 debugf("initarm: console initialized\n"); 585 debugf(" arg1 kmdp = 0x%08x\n", (uint32_t)preload_kmdp); 586 debugf(" boothowto = 0x%08x\n", boothowto); 587 debugf(" dtbp = 0x%08x\n", (uint32_t)dtbp); 588 debugf(" lastaddr1: 0x%08x\n", lastaddr); 589 arm_print_kenv(); 590 591 env = kern_getenv("kernelname"); 592 if (env != NULL) 593 strlcpy(kernelname, env, sizeof(kernelname)); 594 595 if (err_devmap != 0) 596 printf("WARNING: could not fully configure devmap, error=%d\n", 597 err_devmap); 598 599 platform_late_init(); 600 601 root = OF_finddevice("/"); 602 if (OF_getprop(root, "freebsd,dts-version", dts_version, sizeof(dts_version)) > 0) { 603 if (strcmp(LINUX_DTS_VERSION, dts_version) != 0) 604 printf("WARNING: DTB version is %s while kernel expects %s, " 605 "please update the DTB in the ESP\n", 606 dts_version, 607 LINUX_DTS_VERSION); 608 } else { 609 printf("WARNING: Cannot find freebsd,dts-version property, " 610 "cannot check DTB compliance\n"); 611 } 612 613 /* 614 * We must now clean the cache again.... 615 * Cleaning may be done by reading new data to displace any 616 * dirty data in the cache. This will have happened in cpu_setttb() 617 * but since we are boot strapping the addresses used for the read 618 * may have just been remapped and thus the cache could be out 619 * of sync. A re-clean after the switch will cure this. 620 * After booting there are no gross relocations of the kernel thus 621 * this problem will not occur after initarm(). 622 */ 623 /* Set stack for exception handlers */ 624 undefined_init(); 625 init_proc0(kernelstack); 626 arm_vector_init(ARM_VECTORS_HIGH, ARM_VEC_ALL); 627 enable_interrupts(PSR_A); 628 pmap_bootstrap(0); 629 630 /* Exclude the kernel (and all the things we allocated which immediately 631 * follow the kernel) from the VM allocation pool but not from crash 632 * dumps. virtual_avail is a global variable which tracks the kva we've 633 * "allocated" while setting up pmaps. 634 * 635 * Prepare the list of physical memory available to the vm subsystem. 636 */ 637 physmem_exclude_region(abp->abp_physaddr, 638 pmap_preboot_get_pages(0) - abp->abp_physaddr, EXFLAG_NOALLOC); 639 physmem_init_kernel_globals(); 640 641 init_param2(physmem); 642 /* Init message buffer. */ 643 msgbufinit(msgbufp, msgbufsize); 644 dbg_monitor_init(); 645 arm_kdb_init(); 646 /* Apply possible BP hardening. */ 647 cpuinfo_init_bp_hardening(); 648 649 #ifdef EFI 650 if (boothowto & RB_VERBOSE) { 651 if (efihdr != NULL) 652 efi_map_print_entries(efihdr); 653 } 654 #endif 655 656 return ((void *)STACKALIGN(thread0.td_pcb)); 657 658 } 659 #endif /* FDT */ 660