1 /* 2 * linux/arch/arm/kernel/setup.c 3 * 4 * Copyright (C) 1995-2001 Russell King 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License version 2 as 8 * published by the Free Software Foundation. 9 */ 10 #include <linux/config.h> 11 #include <linux/module.h> 12 #include <linux/kernel.h> 13 #include <linux/stddef.h> 14 #include <linux/ioport.h> 15 #include <linux/delay.h> 16 #include <linux/utsname.h> 17 #include <linux/initrd.h> 18 #include <linux/console.h> 19 #include <linux/bootmem.h> 20 #include <linux/seq_file.h> 21 #include <linux/tty.h> 22 #include <linux/init.h> 23 #include <linux/root_dev.h> 24 #include <linux/cpu.h> 25 #include <linux/interrupt.h> 26 27 #include <asm/cpu.h> 28 #include <asm/elf.h> 29 #include <asm/hardware.h> 30 #include <asm/io.h> 31 #include <asm/procinfo.h> 32 #include <asm/setup.h> 33 #include <asm/mach-types.h> 34 #include <asm/cacheflush.h> 35 #include <asm/tlbflush.h> 36 37 #include <asm/mach/arch.h> 38 #include <asm/mach/irq.h> 39 #include <asm/mach/time.h> 40 41 #ifndef MEM_SIZE 42 #define MEM_SIZE (16*1024*1024) 43 #endif 44 45 #if defined(CONFIG_FPE_NWFPE) || defined(CONFIG_FPE_FASTFPE) 46 char fpe_type[8]; 47 48 static int __init fpe_setup(char *line) 49 { 50 memcpy(fpe_type, line, 8); 51 return 1; 52 } 53 54 __setup("fpe=", fpe_setup); 55 #endif 56 57 extern unsigned int mem_fclk_21285; 58 extern void paging_init(struct meminfo *, struct machine_desc *desc); 59 extern void convert_to_tag_list(struct tag *tags); 60 extern void squash_mem_tags(struct tag *tag); 61 extern void reboot_setup(char *str); 62 extern int root_mountflags; 63 extern void _stext, _text, _etext, __data_start, _edata, _end; 64 65 unsigned int processor_id; 66 unsigned int __machine_arch_type; 67 EXPORT_SYMBOL(__machine_arch_type); 68 69 unsigned int system_rev; 70 EXPORT_SYMBOL(system_rev); 71 72 unsigned int system_serial_low; 73 EXPORT_SYMBOL(system_serial_low); 74 75 unsigned int system_serial_high; 76 EXPORT_SYMBOL(system_serial_high); 77 78 unsigned int elf_hwcap; 79 EXPORT_SYMBOL(elf_hwcap); 80 81 82 #ifdef MULTI_CPU 83 struct processor processor; 84 #endif 85 #ifdef MULTI_TLB 86 struct cpu_tlb_fns cpu_tlb; 87 #endif 88 #ifdef MULTI_USER 89 struct cpu_user_fns cpu_user; 90 #endif 91 #ifdef MULTI_CACHE 92 struct cpu_cache_fns cpu_cache; 93 #endif 94 95 struct stack { 96 u32 irq[3]; 97 u32 abt[3]; 98 u32 und[3]; 99 } ____cacheline_aligned; 100 101 static struct stack stacks[NR_CPUS]; 102 103 char elf_platform[ELF_PLATFORM_SIZE]; 104 EXPORT_SYMBOL(elf_platform); 105 106 unsigned long phys_initrd_start __initdata = 0; 107 unsigned long phys_initrd_size __initdata = 0; 108 109 static struct meminfo meminfo __initdata = { 0, }; 110 static const char *cpu_name; 111 static const char *machine_name; 112 static char command_line[COMMAND_LINE_SIZE]; 113 114 static char default_command_line[COMMAND_LINE_SIZE] __initdata = CONFIG_CMDLINE; 115 static union { char c[4]; unsigned long l; } endian_test __initdata = { { 'l', '?', '?', 'b' } }; 116 #define ENDIANNESS ((char)endian_test.l) 117 118 DEFINE_PER_CPU(struct cpuinfo_arm, cpu_data); 119 120 /* 121 * Standard memory resources 122 */ 123 static struct resource mem_res[] = { 124 { "Video RAM", 0, 0, IORESOURCE_MEM }, 125 { "Kernel text", 0, 0, IORESOURCE_MEM }, 126 { "Kernel data", 0, 0, IORESOURCE_MEM } 127 }; 128 129 #define video_ram mem_res[0] 130 #define kernel_code mem_res[1] 131 #define kernel_data mem_res[2] 132 133 static struct resource io_res[] = { 134 { "reserved", 0x3bc, 0x3be, IORESOURCE_IO | IORESOURCE_BUSY }, 135 { "reserved", 0x378, 0x37f, IORESOURCE_IO | IORESOURCE_BUSY }, 136 { "reserved", 0x278, 0x27f, IORESOURCE_IO | IORESOURCE_BUSY } 137 }; 138 139 #define lp0 io_res[0] 140 #define lp1 io_res[1] 141 #define lp2 io_res[2] 142 143 static const char *cache_types[16] = { 144 "write-through", 145 "write-back", 146 "write-back", 147 "undefined 3", 148 "undefined 4", 149 "undefined 5", 150 "write-back", 151 "write-back", 152 "undefined 8", 153 "undefined 9", 154 "undefined 10", 155 "undefined 11", 156 "undefined 12", 157 "undefined 13", 158 "write-back", 159 "undefined 15", 160 }; 161 162 static const char *cache_clean[16] = { 163 "not required", 164 "read-block", 165 "cp15 c7 ops", 166 "undefined 3", 167 "undefined 4", 168 "undefined 5", 169 "cp15 c7 ops", 170 "cp15 c7 ops", 171 "undefined 8", 172 "undefined 9", 173 "undefined 10", 174 "undefined 11", 175 "undefined 12", 176 "undefined 13", 177 "cp15 c7 ops", 178 "undefined 15", 179 }; 180 181 static const char *cache_lockdown[16] = { 182 "not supported", 183 "not supported", 184 "not supported", 185 "undefined 3", 186 "undefined 4", 187 "undefined 5", 188 "format A", 189 "format B", 190 "undefined 8", 191 "undefined 9", 192 "undefined 10", 193 "undefined 11", 194 "undefined 12", 195 "undefined 13", 196 "format C", 197 "undefined 15", 198 }; 199 200 static const char *proc_arch[] = { 201 "undefined/unknown", 202 "3", 203 "4", 204 "4T", 205 "5", 206 "5T", 207 "5TE", 208 "5TEJ", 209 "6TEJ", 210 "?(10)", 211 "?(11)", 212 "?(12)", 213 "?(13)", 214 "?(14)", 215 "?(15)", 216 "?(16)", 217 "?(17)", 218 }; 219 220 #define CACHE_TYPE(x) (((x) >> 25) & 15) 221 #define CACHE_S(x) ((x) & (1 << 24)) 222 #define CACHE_DSIZE(x) (((x) >> 12) & 4095) /* only if S=1 */ 223 #define CACHE_ISIZE(x) ((x) & 4095) 224 225 #define CACHE_SIZE(y) (((y) >> 6) & 7) 226 #define CACHE_ASSOC(y) (((y) >> 3) & 7) 227 #define CACHE_M(y) ((y) & (1 << 2)) 228 #define CACHE_LINE(y) ((y) & 3) 229 230 static inline void dump_cache(const char *prefix, int cpu, unsigned int cache) 231 { 232 unsigned int mult = 2 + (CACHE_M(cache) ? 1 : 0); 233 234 printk("CPU%u: %s: %d bytes, associativity %d, %d byte lines, %d sets\n", 235 cpu, prefix, 236 mult << (8 + CACHE_SIZE(cache)), 237 (mult << CACHE_ASSOC(cache)) >> 1, 238 8 << CACHE_LINE(cache), 239 1 << (6 + CACHE_SIZE(cache) - CACHE_ASSOC(cache) - 240 CACHE_LINE(cache))); 241 } 242 243 static void __init dump_cpu_info(int cpu) 244 { 245 unsigned int info = read_cpuid(CPUID_CACHETYPE); 246 247 if (info != processor_id) { 248 printk("CPU%u: D %s %s cache\n", cpu, cache_is_vivt() ? "VIVT" : "VIPT", 249 cache_types[CACHE_TYPE(info)]); 250 if (CACHE_S(info)) { 251 dump_cache("I cache", cpu, CACHE_ISIZE(info)); 252 dump_cache("D cache", cpu, CACHE_DSIZE(info)); 253 } else { 254 dump_cache("cache", cpu, CACHE_ISIZE(info)); 255 } 256 } 257 } 258 259 int cpu_architecture(void) 260 { 261 int cpu_arch; 262 263 if ((processor_id & 0x0000f000) == 0) { 264 cpu_arch = CPU_ARCH_UNKNOWN; 265 } else if ((processor_id & 0x0000f000) == 0x00007000) { 266 cpu_arch = (processor_id & (1 << 23)) ? CPU_ARCH_ARMv4T : CPU_ARCH_ARMv3; 267 } else { 268 cpu_arch = (processor_id >> 16) & 7; 269 if (cpu_arch) 270 cpu_arch += CPU_ARCH_ARMv3; 271 } 272 273 return cpu_arch; 274 } 275 276 /* 277 * These functions re-use the assembly code in head.S, which 278 * already provide the required functionality. 279 */ 280 extern struct proc_info_list *lookup_processor_type(void); 281 extern struct machine_desc *lookup_machine_type(unsigned int); 282 283 static void __init setup_processor(void) 284 { 285 struct proc_info_list *list; 286 287 /* 288 * locate processor in the list of supported processor 289 * types. The linker builds this table for us from the 290 * entries in arch/arm/mm/proc-*.S 291 */ 292 list = lookup_processor_type(); 293 if (!list) { 294 printk("CPU configuration botched (ID %08x), unable " 295 "to continue.\n", processor_id); 296 while (1); 297 } 298 299 cpu_name = list->cpu_name; 300 301 #ifdef MULTI_CPU 302 processor = *list->proc; 303 #endif 304 #ifdef MULTI_TLB 305 cpu_tlb = *list->tlb; 306 #endif 307 #ifdef MULTI_USER 308 cpu_user = *list->user; 309 #endif 310 #ifdef MULTI_CACHE 311 cpu_cache = *list->cache; 312 #endif 313 314 printk("CPU: %s [%08x] revision %d (ARMv%s)\n", 315 cpu_name, processor_id, (int)processor_id & 15, 316 proc_arch[cpu_architecture()]); 317 318 sprintf(system_utsname.machine, "%s%c", list->arch_name, ENDIANNESS); 319 sprintf(elf_platform, "%s%c", list->elf_name, ENDIANNESS); 320 elf_hwcap = list->elf_hwcap; 321 322 cpu_proc_init(); 323 } 324 325 /* 326 * cpu_init - initialise one CPU. 327 * 328 * cpu_init dumps the cache information, initialises SMP specific 329 * information, and sets up the per-CPU stacks. 330 */ 331 void cpu_init(void) 332 { 333 unsigned int cpu = smp_processor_id(); 334 struct stack *stk = &stacks[cpu]; 335 336 if (cpu >= NR_CPUS) { 337 printk(KERN_CRIT "CPU%u: bad primary CPU number\n", cpu); 338 BUG(); 339 } 340 341 if (system_state == SYSTEM_BOOTING) 342 dump_cpu_info(cpu); 343 344 /* 345 * setup stacks for re-entrant exception handlers 346 */ 347 __asm__ ( 348 "msr cpsr_c, %1\n\t" 349 "add sp, %0, %2\n\t" 350 "msr cpsr_c, %3\n\t" 351 "add sp, %0, %4\n\t" 352 "msr cpsr_c, %5\n\t" 353 "add sp, %0, %6\n\t" 354 "msr cpsr_c, %7" 355 : 356 : "r" (stk), 357 "I" (PSR_F_BIT | PSR_I_BIT | IRQ_MODE), 358 "I" (offsetof(struct stack, irq[0])), 359 "I" (PSR_F_BIT | PSR_I_BIT | ABT_MODE), 360 "I" (offsetof(struct stack, abt[0])), 361 "I" (PSR_F_BIT | PSR_I_BIT | UND_MODE), 362 "I" (offsetof(struct stack, und[0])), 363 "I" (PSR_F_BIT | PSR_I_BIT | SVC_MODE) 364 : "r14"); 365 } 366 367 static struct machine_desc * __init setup_machine(unsigned int nr) 368 { 369 struct machine_desc *list; 370 371 /* 372 * locate machine in the list of supported machines. 373 */ 374 list = lookup_machine_type(nr); 375 if (!list) { 376 printk("Machine configuration botched (nr %d), unable " 377 "to continue.\n", nr); 378 while (1); 379 } 380 381 printk("Machine: %s\n", list->name); 382 383 return list; 384 } 385 386 static void __init early_initrd(char **p) 387 { 388 unsigned long start, size; 389 390 start = memparse(*p, p); 391 if (**p == ',') { 392 size = memparse((*p) + 1, p); 393 394 phys_initrd_start = start; 395 phys_initrd_size = size; 396 } 397 } 398 __early_param("initrd=", early_initrd); 399 400 static void __init add_memory(unsigned long start, unsigned long size) 401 { 402 /* 403 * Ensure that start/size are aligned to a page boundary. 404 * Size is appropriately rounded down, start is rounded up. 405 */ 406 size -= start & ~PAGE_MASK; 407 408 meminfo.bank[meminfo.nr_banks].start = PAGE_ALIGN(start); 409 meminfo.bank[meminfo.nr_banks].size = size & PAGE_MASK; 410 meminfo.bank[meminfo.nr_banks].node = PHYS_TO_NID(start); 411 meminfo.nr_banks += 1; 412 } 413 414 /* 415 * Pick out the memory size. We look for mem=size@start, 416 * where start and size are "size[KkMm]" 417 */ 418 static void __init early_mem(char **p) 419 { 420 static int usermem __initdata = 0; 421 unsigned long size, start; 422 423 /* 424 * If the user specifies memory size, we 425 * blow away any automatically generated 426 * size. 427 */ 428 if (usermem == 0) { 429 usermem = 1; 430 meminfo.nr_banks = 0; 431 } 432 433 start = PHYS_OFFSET; 434 size = memparse(*p, p); 435 if (**p == '@') 436 start = memparse(*p + 1, p); 437 438 add_memory(start, size); 439 } 440 __early_param("mem=", early_mem); 441 442 /* 443 * Initial parsing of the command line. 444 */ 445 static void __init parse_cmdline(char **cmdline_p, char *from) 446 { 447 char c = ' ', *to = command_line; 448 int len = 0; 449 450 for (;;) { 451 if (c == ' ') { 452 extern struct early_params __early_begin, __early_end; 453 struct early_params *p; 454 455 for (p = &__early_begin; p < &__early_end; p++) { 456 int len = strlen(p->arg); 457 458 if (memcmp(from, p->arg, len) == 0) { 459 if (to != command_line) 460 to -= 1; 461 from += len; 462 p->fn(&from); 463 464 while (*from != ' ' && *from != '\0') 465 from++; 466 break; 467 } 468 } 469 } 470 c = *from++; 471 if (!c) 472 break; 473 if (COMMAND_LINE_SIZE <= ++len) 474 break; 475 *to++ = c; 476 } 477 *to = '\0'; 478 *cmdline_p = command_line; 479 } 480 481 static void __init 482 setup_ramdisk(int doload, int prompt, int image_start, unsigned int rd_sz) 483 { 484 #ifdef CONFIG_BLK_DEV_RAM 485 extern int rd_size, rd_image_start, rd_prompt, rd_doload; 486 487 rd_image_start = image_start; 488 rd_prompt = prompt; 489 rd_doload = doload; 490 491 if (rd_sz) 492 rd_size = rd_sz; 493 #endif 494 } 495 496 static void __init 497 request_standard_resources(struct meminfo *mi, struct machine_desc *mdesc) 498 { 499 struct resource *res; 500 int i; 501 502 kernel_code.start = virt_to_phys(&_text); 503 kernel_code.end = virt_to_phys(&_etext - 1); 504 kernel_data.start = virt_to_phys(&__data_start); 505 kernel_data.end = virt_to_phys(&_end - 1); 506 507 for (i = 0; i < mi->nr_banks; i++) { 508 unsigned long virt_start, virt_end; 509 510 if (mi->bank[i].size == 0) 511 continue; 512 513 virt_start = __phys_to_virt(mi->bank[i].start); 514 virt_end = virt_start + mi->bank[i].size - 1; 515 516 res = alloc_bootmem_low(sizeof(*res)); 517 res->name = "System RAM"; 518 res->start = __virt_to_phys(virt_start); 519 res->end = __virt_to_phys(virt_end); 520 res->flags = IORESOURCE_MEM | IORESOURCE_BUSY; 521 522 request_resource(&iomem_resource, res); 523 524 if (kernel_code.start >= res->start && 525 kernel_code.end <= res->end) 526 request_resource(res, &kernel_code); 527 if (kernel_data.start >= res->start && 528 kernel_data.end <= res->end) 529 request_resource(res, &kernel_data); 530 } 531 532 if (mdesc->video_start) { 533 video_ram.start = mdesc->video_start; 534 video_ram.end = mdesc->video_end; 535 request_resource(&iomem_resource, &video_ram); 536 } 537 538 /* 539 * Some machines don't have the possibility of ever 540 * possessing lp0, lp1 or lp2 541 */ 542 if (mdesc->reserve_lp0) 543 request_resource(&ioport_resource, &lp0); 544 if (mdesc->reserve_lp1) 545 request_resource(&ioport_resource, &lp1); 546 if (mdesc->reserve_lp2) 547 request_resource(&ioport_resource, &lp2); 548 } 549 550 /* 551 * Tag parsing. 552 * 553 * This is the new way of passing data to the kernel at boot time. Rather 554 * than passing a fixed inflexible structure to the kernel, we pass a list 555 * of variable-sized tags to the kernel. The first tag must be a ATAG_CORE 556 * tag for the list to be recognised (to distinguish the tagged list from 557 * a param_struct). The list is terminated with a zero-length tag (this tag 558 * is not parsed in any way). 559 */ 560 static int __init parse_tag_core(const struct tag *tag) 561 { 562 if (tag->hdr.size > 2) { 563 if ((tag->u.core.flags & 1) == 0) 564 root_mountflags &= ~MS_RDONLY; 565 ROOT_DEV = old_decode_dev(tag->u.core.rootdev); 566 } 567 return 0; 568 } 569 570 __tagtable(ATAG_CORE, parse_tag_core); 571 572 static int __init parse_tag_mem32(const struct tag *tag) 573 { 574 if (meminfo.nr_banks >= NR_BANKS) { 575 printk(KERN_WARNING 576 "Ignoring memory bank 0x%08x size %dKB\n", 577 tag->u.mem.start, tag->u.mem.size / 1024); 578 return -EINVAL; 579 } 580 add_memory(tag->u.mem.start, tag->u.mem.size); 581 return 0; 582 } 583 584 __tagtable(ATAG_MEM, parse_tag_mem32); 585 586 #if defined(CONFIG_VGA_CONSOLE) || defined(CONFIG_DUMMY_CONSOLE) 587 struct screen_info screen_info = { 588 .orig_video_lines = 30, 589 .orig_video_cols = 80, 590 .orig_video_mode = 0, 591 .orig_video_ega_bx = 0, 592 .orig_video_isVGA = 1, 593 .orig_video_points = 8 594 }; 595 596 static int __init parse_tag_videotext(const struct tag *tag) 597 { 598 screen_info.orig_x = tag->u.videotext.x; 599 screen_info.orig_y = tag->u.videotext.y; 600 screen_info.orig_video_page = tag->u.videotext.video_page; 601 screen_info.orig_video_mode = tag->u.videotext.video_mode; 602 screen_info.orig_video_cols = tag->u.videotext.video_cols; 603 screen_info.orig_video_ega_bx = tag->u.videotext.video_ega_bx; 604 screen_info.orig_video_lines = tag->u.videotext.video_lines; 605 screen_info.orig_video_isVGA = tag->u.videotext.video_isvga; 606 screen_info.orig_video_points = tag->u.videotext.video_points; 607 return 0; 608 } 609 610 __tagtable(ATAG_VIDEOTEXT, parse_tag_videotext); 611 #endif 612 613 static int __init parse_tag_ramdisk(const struct tag *tag) 614 { 615 setup_ramdisk((tag->u.ramdisk.flags & 1) == 0, 616 (tag->u.ramdisk.flags & 2) == 0, 617 tag->u.ramdisk.start, tag->u.ramdisk.size); 618 return 0; 619 } 620 621 __tagtable(ATAG_RAMDISK, parse_tag_ramdisk); 622 623 static int __init parse_tag_initrd(const struct tag *tag) 624 { 625 printk(KERN_WARNING "ATAG_INITRD is deprecated; " 626 "please update your bootloader.\n"); 627 phys_initrd_start = __virt_to_phys(tag->u.initrd.start); 628 phys_initrd_size = tag->u.initrd.size; 629 return 0; 630 } 631 632 __tagtable(ATAG_INITRD, parse_tag_initrd); 633 634 static int __init parse_tag_initrd2(const struct tag *tag) 635 { 636 phys_initrd_start = tag->u.initrd.start; 637 phys_initrd_size = tag->u.initrd.size; 638 return 0; 639 } 640 641 __tagtable(ATAG_INITRD2, parse_tag_initrd2); 642 643 static int __init parse_tag_serialnr(const struct tag *tag) 644 { 645 system_serial_low = tag->u.serialnr.low; 646 system_serial_high = tag->u.serialnr.high; 647 return 0; 648 } 649 650 __tagtable(ATAG_SERIAL, parse_tag_serialnr); 651 652 static int __init parse_tag_revision(const struct tag *tag) 653 { 654 system_rev = tag->u.revision.rev; 655 return 0; 656 } 657 658 __tagtable(ATAG_REVISION, parse_tag_revision); 659 660 static int __init parse_tag_cmdline(const struct tag *tag) 661 { 662 strlcpy(default_command_line, tag->u.cmdline.cmdline, COMMAND_LINE_SIZE); 663 return 0; 664 } 665 666 __tagtable(ATAG_CMDLINE, parse_tag_cmdline); 667 668 /* 669 * Scan the tag table for this tag, and call its parse function. 670 * The tag table is built by the linker from all the __tagtable 671 * declarations. 672 */ 673 static int __init parse_tag(const struct tag *tag) 674 { 675 extern struct tagtable __tagtable_begin, __tagtable_end; 676 struct tagtable *t; 677 678 for (t = &__tagtable_begin; t < &__tagtable_end; t++) 679 if (tag->hdr.tag == t->tag) { 680 t->parse(tag); 681 break; 682 } 683 684 return t < &__tagtable_end; 685 } 686 687 /* 688 * Parse all tags in the list, checking both the global and architecture 689 * specific tag tables. 690 */ 691 static void __init parse_tags(const struct tag *t) 692 { 693 for (; t->hdr.size; t = tag_next(t)) 694 if (!parse_tag(t)) 695 printk(KERN_WARNING 696 "Ignoring unrecognised tag 0x%08x\n", 697 t->hdr.tag); 698 } 699 700 /* 701 * This holds our defaults. 702 */ 703 static struct init_tags { 704 struct tag_header hdr1; 705 struct tag_core core; 706 struct tag_header hdr2; 707 struct tag_mem32 mem; 708 struct tag_header hdr3; 709 } init_tags __initdata = { 710 { tag_size(tag_core), ATAG_CORE }, 711 { 1, PAGE_SIZE, 0xff }, 712 { tag_size(tag_mem32), ATAG_MEM }, 713 { MEM_SIZE, PHYS_OFFSET }, 714 { 0, ATAG_NONE } 715 }; 716 717 static void (*init_machine)(void) __initdata; 718 719 static int __init customize_machine(void) 720 { 721 /* customizes platform devices, or adds new ones */ 722 if (init_machine) 723 init_machine(); 724 return 0; 725 } 726 arch_initcall(customize_machine); 727 728 void __init setup_arch(char **cmdline_p) 729 { 730 struct tag *tags = (struct tag *)&init_tags; 731 struct machine_desc *mdesc; 732 char *from = default_command_line; 733 734 setup_processor(); 735 mdesc = setup_machine(machine_arch_type); 736 machine_name = mdesc->name; 737 738 if (mdesc->soft_reboot) 739 reboot_setup("s"); 740 741 if (mdesc->boot_params) 742 tags = phys_to_virt(mdesc->boot_params); 743 744 /* 745 * If we have the old style parameters, convert them to 746 * a tag list. 747 */ 748 if (tags->hdr.tag != ATAG_CORE) 749 convert_to_tag_list(tags); 750 if (tags->hdr.tag != ATAG_CORE) 751 tags = (struct tag *)&init_tags; 752 753 if (mdesc->fixup) 754 mdesc->fixup(mdesc, tags, &from, &meminfo); 755 756 if (tags->hdr.tag == ATAG_CORE) { 757 if (meminfo.nr_banks != 0) 758 squash_mem_tags(tags); 759 parse_tags(tags); 760 } 761 762 init_mm.start_code = (unsigned long) &_text; 763 init_mm.end_code = (unsigned long) &_etext; 764 init_mm.end_data = (unsigned long) &_edata; 765 init_mm.brk = (unsigned long) &_end; 766 767 memcpy(saved_command_line, from, COMMAND_LINE_SIZE); 768 saved_command_line[COMMAND_LINE_SIZE-1] = '\0'; 769 parse_cmdline(cmdline_p, from); 770 paging_init(&meminfo, mdesc); 771 request_standard_resources(&meminfo, mdesc); 772 773 cpu_init(); 774 775 /* 776 * Set up various architecture-specific pointers 777 */ 778 init_arch_irq = mdesc->init_irq; 779 system_timer = mdesc->timer; 780 init_machine = mdesc->init_machine; 781 782 #ifdef CONFIG_VT 783 #if defined(CONFIG_VGA_CONSOLE) 784 conswitchp = &vga_con; 785 #elif defined(CONFIG_DUMMY_CONSOLE) 786 conswitchp = &dummy_con; 787 #endif 788 #endif 789 } 790 791 792 static int __init topology_init(void) 793 { 794 int cpu; 795 796 for_each_cpu(cpu) 797 register_cpu(&per_cpu(cpu_data, cpu).cpu, cpu, NULL); 798 799 return 0; 800 } 801 802 subsys_initcall(topology_init); 803 804 static const char *hwcap_str[] = { 805 "swp", 806 "half", 807 "thumb", 808 "26bit", 809 "fastmult", 810 "fpa", 811 "vfp", 812 "edsp", 813 "java", 814 NULL 815 }; 816 817 static void 818 c_show_cache(struct seq_file *m, const char *type, unsigned int cache) 819 { 820 unsigned int mult = 2 + (CACHE_M(cache) ? 1 : 0); 821 822 seq_printf(m, "%s size\t\t: %d\n" 823 "%s assoc\t\t: %d\n" 824 "%s line length\t: %d\n" 825 "%s sets\t\t: %d\n", 826 type, mult << (8 + CACHE_SIZE(cache)), 827 type, (mult << CACHE_ASSOC(cache)) >> 1, 828 type, 8 << CACHE_LINE(cache), 829 type, 1 << (6 + CACHE_SIZE(cache) - CACHE_ASSOC(cache) - 830 CACHE_LINE(cache))); 831 } 832 833 static int c_show(struct seq_file *m, void *v) 834 { 835 int i; 836 837 seq_printf(m, "Processor\t: %s rev %d (%s)\n", 838 cpu_name, (int)processor_id & 15, elf_platform); 839 840 #if defined(CONFIG_SMP) 841 for_each_online_cpu(i) { 842 /* 843 * glibc reads /proc/cpuinfo to determine the number of 844 * online processors, looking for lines beginning with 845 * "processor". Give glibc what it expects. 846 */ 847 seq_printf(m, "processor\t: %d\n", i); 848 seq_printf(m, "BogoMIPS\t: %lu.%02lu\n\n", 849 per_cpu(cpu_data, i).loops_per_jiffy / (500000UL/HZ), 850 (per_cpu(cpu_data, i).loops_per_jiffy / (5000UL/HZ)) % 100); 851 } 852 #else /* CONFIG_SMP */ 853 seq_printf(m, "BogoMIPS\t: %lu.%02lu\n", 854 loops_per_jiffy / (500000/HZ), 855 (loops_per_jiffy / (5000/HZ)) % 100); 856 #endif 857 858 /* dump out the processor features */ 859 seq_puts(m, "Features\t: "); 860 861 for (i = 0; hwcap_str[i]; i++) 862 if (elf_hwcap & (1 << i)) 863 seq_printf(m, "%s ", hwcap_str[i]); 864 865 seq_printf(m, "\nCPU implementer\t: 0x%02x\n", processor_id >> 24); 866 seq_printf(m, "CPU architecture: %s\n", proc_arch[cpu_architecture()]); 867 868 if ((processor_id & 0x0000f000) == 0x00000000) { 869 /* pre-ARM7 */ 870 seq_printf(m, "CPU part\t\t: %07x\n", processor_id >> 4); 871 } else { 872 if ((processor_id & 0x0000f000) == 0x00007000) { 873 /* ARM7 */ 874 seq_printf(m, "CPU variant\t: 0x%02x\n", 875 (processor_id >> 16) & 127); 876 } else { 877 /* post-ARM7 */ 878 seq_printf(m, "CPU variant\t: 0x%x\n", 879 (processor_id >> 20) & 15); 880 } 881 seq_printf(m, "CPU part\t: 0x%03x\n", 882 (processor_id >> 4) & 0xfff); 883 } 884 seq_printf(m, "CPU revision\t: %d\n", processor_id & 15); 885 886 { 887 unsigned int cache_info = read_cpuid(CPUID_CACHETYPE); 888 if (cache_info != processor_id) { 889 seq_printf(m, "Cache type\t: %s\n" 890 "Cache clean\t: %s\n" 891 "Cache lockdown\t: %s\n" 892 "Cache format\t: %s\n", 893 cache_types[CACHE_TYPE(cache_info)], 894 cache_clean[CACHE_TYPE(cache_info)], 895 cache_lockdown[CACHE_TYPE(cache_info)], 896 CACHE_S(cache_info) ? "Harvard" : "Unified"); 897 898 if (CACHE_S(cache_info)) { 899 c_show_cache(m, "I", CACHE_ISIZE(cache_info)); 900 c_show_cache(m, "D", CACHE_DSIZE(cache_info)); 901 } else { 902 c_show_cache(m, "Cache", CACHE_ISIZE(cache_info)); 903 } 904 } 905 } 906 907 seq_puts(m, "\n"); 908 909 seq_printf(m, "Hardware\t: %s\n", machine_name); 910 seq_printf(m, "Revision\t: %04x\n", system_rev); 911 seq_printf(m, "Serial\t\t: %08x%08x\n", 912 system_serial_high, system_serial_low); 913 914 return 0; 915 } 916 917 static void *c_start(struct seq_file *m, loff_t *pos) 918 { 919 return *pos < 1 ? (void *)1 : NULL; 920 } 921 922 static void *c_next(struct seq_file *m, void *v, loff_t *pos) 923 { 924 ++*pos; 925 return NULL; 926 } 927 928 static void c_stop(struct seq_file *m, void *v) 929 { 930 } 931 932 struct seq_operations cpuinfo_op = { 933 .start = c_start, 934 .next = c_next, 935 .stop = c_stop, 936 .show = c_show 937 }; 938