1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Intel Multiprocessor Specification 1.1 and 1.4 4 * compliant MP-table parsing routines. 5 * 6 * (c) 1995 Alan Cox, Building #3 <alan@lxorguk.ukuu.org.uk> 7 * (c) 1998, 1999, 2000, 2009 Ingo Molnar <mingo@redhat.com> 8 * (c) 2008 Alexey Starikovskiy <astarikovskiy@suse.de> 9 */ 10 11 #include <linux/mm.h> 12 #include <linux/init.h> 13 #include <linux/delay.h> 14 #include <linux/memblock.h> 15 #include <linux/kernel_stat.h> 16 #include <linux/mc146818rtc.h> 17 #include <linux/bitops.h> 18 #include <linux/acpi.h> 19 #include <linux/smp.h> 20 #include <linux/pci.h> 21 22 #include <asm/io_apic.h> 23 #include <asm/acpi.h> 24 #include <asm/irqdomain.h> 25 #include <asm/mtrr.h> 26 #include <asm/mpspec.h> 27 #include <asm/io_apic.h> 28 #include <asm/proto.h> 29 #include <asm/bios_ebda.h> 30 #include <asm/e820/api.h> 31 #include <asm/setup.h> 32 #include <asm/smp.h> 33 34 #include <asm/apic.h> 35 /* 36 * Checksum an MP configuration block. 37 */ 38 39 static int __init mpf_checksum(unsigned char *mp, int len) 40 { 41 int sum = 0; 42 43 while (len--) 44 sum += *mp++; 45 46 return sum & 0xFF; 47 } 48 49 int __init default_mpc_apic_id(struct mpc_cpu *m) 50 { 51 return m->apicid; 52 } 53 54 static void __init MP_processor_info(struct mpc_cpu *m) 55 { 56 int apicid; 57 char *bootup_cpu = ""; 58 59 if (!(m->cpuflag & CPU_ENABLED)) { 60 disabled_cpus++; 61 return; 62 } 63 64 apicid = x86_init.mpparse.mpc_apic_id(m); 65 66 if (m->cpuflag & CPU_BOOTPROCESSOR) { 67 bootup_cpu = " (Bootup-CPU)"; 68 boot_cpu_physical_apicid = m->apicid; 69 } 70 71 pr_info("Processor #%d%s\n", m->apicid, bootup_cpu); 72 generic_processor_info(apicid, m->apicver); 73 } 74 75 #ifdef CONFIG_X86_IO_APIC 76 void __init default_mpc_oem_bus_info(struct mpc_bus *m, char *str) 77 { 78 memcpy(str, m->bustype, 6); 79 str[6] = 0; 80 apic_printk(APIC_VERBOSE, "Bus #%d is %s\n", m->busid, str); 81 } 82 83 static void __init MP_bus_info(struct mpc_bus *m) 84 { 85 char str[7]; 86 87 x86_init.mpparse.mpc_oem_bus_info(m, str); 88 89 #if MAX_MP_BUSSES < 256 90 if (m->busid >= MAX_MP_BUSSES) { 91 pr_warn("MP table busid value (%d) for bustype %s is too large, max. supported is %d\n", 92 m->busid, str, MAX_MP_BUSSES - 1); 93 return; 94 } 95 #endif 96 97 set_bit(m->busid, mp_bus_not_pci); 98 if (strncmp(str, BUSTYPE_ISA, sizeof(BUSTYPE_ISA) - 1) == 0) { 99 #ifdef CONFIG_EISA 100 mp_bus_id_to_type[m->busid] = MP_BUS_ISA; 101 #endif 102 } else if (strncmp(str, BUSTYPE_PCI, sizeof(BUSTYPE_PCI) - 1) == 0) { 103 if (x86_init.mpparse.mpc_oem_pci_bus) 104 x86_init.mpparse.mpc_oem_pci_bus(m); 105 106 clear_bit(m->busid, mp_bus_not_pci); 107 #ifdef CONFIG_EISA 108 mp_bus_id_to_type[m->busid] = MP_BUS_PCI; 109 } else if (strncmp(str, BUSTYPE_EISA, sizeof(BUSTYPE_EISA) - 1) == 0) { 110 mp_bus_id_to_type[m->busid] = MP_BUS_EISA; 111 #endif 112 } else 113 pr_warn("Unknown bustype %s - ignoring\n", str); 114 } 115 116 static void __init MP_ioapic_info(struct mpc_ioapic *m) 117 { 118 struct ioapic_domain_cfg cfg = { 119 .type = IOAPIC_DOMAIN_LEGACY, 120 .ops = &mp_ioapic_irqdomain_ops, 121 }; 122 123 if (m->flags & MPC_APIC_USABLE) 124 mp_register_ioapic(m->apicid, m->apicaddr, gsi_top, &cfg); 125 } 126 127 static void __init print_mp_irq_info(struct mpc_intsrc *mp_irq) 128 { 129 apic_printk(APIC_VERBOSE, 130 "Int: type %d, pol %d, trig %d, bus %02x, IRQ %02x, APIC ID %x, APIC INT %02x\n", 131 mp_irq->irqtype, mp_irq->irqflag & 3, 132 (mp_irq->irqflag >> 2) & 3, mp_irq->srcbus, 133 mp_irq->srcbusirq, mp_irq->dstapic, mp_irq->dstirq); 134 } 135 136 #else /* CONFIG_X86_IO_APIC */ 137 static inline void __init MP_bus_info(struct mpc_bus *m) {} 138 static inline void __init MP_ioapic_info(struct mpc_ioapic *m) {} 139 #endif /* CONFIG_X86_IO_APIC */ 140 141 static void __init MP_lintsrc_info(struct mpc_lintsrc *m) 142 { 143 apic_printk(APIC_VERBOSE, 144 "Lint: type %d, pol %d, trig %d, bus %02x, IRQ %02x, APIC ID %x, APIC LINT %02x\n", 145 m->irqtype, m->irqflag & 3, (m->irqflag >> 2) & 3, m->srcbusid, 146 m->srcbusirq, m->destapic, m->destapiclint); 147 } 148 149 /* 150 * Read/parse the MPC 151 */ 152 static int __init smp_check_mpc(struct mpc_table *mpc, char *oem, char *str) 153 { 154 155 if (memcmp(mpc->signature, MPC_SIGNATURE, 4)) { 156 pr_err("MPTABLE: bad signature [%c%c%c%c]!\n", 157 mpc->signature[0], mpc->signature[1], 158 mpc->signature[2], mpc->signature[3]); 159 return 0; 160 } 161 if (mpf_checksum((unsigned char *)mpc, mpc->length)) { 162 pr_err("MPTABLE: checksum error!\n"); 163 return 0; 164 } 165 if (mpc->spec != 0x01 && mpc->spec != 0x04) { 166 pr_err("MPTABLE: bad table version (%d)!!\n", mpc->spec); 167 return 0; 168 } 169 if (!mpc->lapic) { 170 pr_err("MPTABLE: null local APIC address!\n"); 171 return 0; 172 } 173 memcpy(oem, mpc->oem, 8); 174 oem[8] = 0; 175 pr_info("MPTABLE: OEM ID: %s\n", oem); 176 177 memcpy(str, mpc->productid, 12); 178 str[12] = 0; 179 180 pr_info("MPTABLE: Product ID: %s\n", str); 181 182 pr_info("MPTABLE: APIC at: 0x%X\n", mpc->lapic); 183 184 return 1; 185 } 186 187 static void skip_entry(unsigned char **ptr, int *count, int size) 188 { 189 *ptr += size; 190 *count += size; 191 } 192 193 static void __init smp_dump_mptable(struct mpc_table *mpc, unsigned char *mpt) 194 { 195 pr_err("Your mptable is wrong, contact your HW vendor!\n"); 196 pr_cont("type %x\n", *mpt); 197 print_hex_dump(KERN_ERR, " ", DUMP_PREFIX_ADDRESS, 16, 198 1, mpc, mpc->length, 1); 199 } 200 201 void __init default_smp_read_mpc_oem(struct mpc_table *mpc) { } 202 203 static int __init smp_read_mpc(struct mpc_table *mpc, unsigned early) 204 { 205 char str[16]; 206 char oem[10]; 207 208 int count = sizeof(*mpc); 209 unsigned char *mpt = ((unsigned char *)mpc) + count; 210 211 if (!smp_check_mpc(mpc, oem, str)) 212 return 0; 213 214 /* Initialize the lapic mapping */ 215 if (!acpi_lapic) 216 register_lapic_address(mpc->lapic); 217 218 if (early) 219 return 1; 220 221 if (mpc->oemptr) 222 x86_init.mpparse.smp_read_mpc_oem(mpc); 223 224 /* 225 * Now process the configuration blocks. 226 */ 227 x86_init.mpparse.mpc_record(0); 228 229 while (count < mpc->length) { 230 switch (*mpt) { 231 case MP_PROCESSOR: 232 /* ACPI may have already provided this data */ 233 if (!acpi_lapic) 234 MP_processor_info((struct mpc_cpu *)mpt); 235 skip_entry(&mpt, &count, sizeof(struct mpc_cpu)); 236 break; 237 case MP_BUS: 238 MP_bus_info((struct mpc_bus *)mpt); 239 skip_entry(&mpt, &count, sizeof(struct mpc_bus)); 240 break; 241 case MP_IOAPIC: 242 MP_ioapic_info((struct mpc_ioapic *)mpt); 243 skip_entry(&mpt, &count, sizeof(struct mpc_ioapic)); 244 break; 245 case MP_INTSRC: 246 mp_save_irq((struct mpc_intsrc *)mpt); 247 skip_entry(&mpt, &count, sizeof(struct mpc_intsrc)); 248 break; 249 case MP_LINTSRC: 250 MP_lintsrc_info((struct mpc_lintsrc *)mpt); 251 skip_entry(&mpt, &count, sizeof(struct mpc_lintsrc)); 252 break; 253 default: 254 /* wrong mptable */ 255 smp_dump_mptable(mpc, mpt); 256 count = mpc->length; 257 break; 258 } 259 x86_init.mpparse.mpc_record(1); 260 } 261 262 if (!num_processors) 263 pr_err("MPTABLE: no processors registered!\n"); 264 return num_processors; 265 } 266 267 #ifdef CONFIG_X86_IO_APIC 268 269 static int __init ELCR_trigger(unsigned int irq) 270 { 271 unsigned int port; 272 273 port = 0x4d0 + (irq >> 3); 274 return (inb(port) >> (irq & 7)) & 1; 275 } 276 277 static void __init construct_default_ioirq_mptable(int mpc_default_type) 278 { 279 struct mpc_intsrc intsrc; 280 int i; 281 int ELCR_fallback = 0; 282 283 intsrc.type = MP_INTSRC; 284 intsrc.irqflag = MP_IRQTRIG_DEFAULT | MP_IRQPOL_DEFAULT; 285 intsrc.srcbus = 0; 286 intsrc.dstapic = mpc_ioapic_id(0); 287 288 intsrc.irqtype = mp_INT; 289 290 /* 291 * If true, we have an ISA/PCI system with no IRQ entries 292 * in the MP table. To prevent the PCI interrupts from being set up 293 * incorrectly, we try to use the ELCR. The sanity check to see if 294 * there is good ELCR data is very simple - IRQ0, 1, 2 and 13 can 295 * never be level sensitive, so we simply see if the ELCR agrees. 296 * If it does, we assume it's valid. 297 */ 298 if (mpc_default_type == 5) { 299 pr_info("ISA/PCI bus type with no IRQ information... falling back to ELCR\n"); 300 301 if (ELCR_trigger(0) || ELCR_trigger(1) || ELCR_trigger(2) || 302 ELCR_trigger(13)) 303 pr_err("ELCR contains invalid data... not using ELCR\n"); 304 else { 305 pr_info("Using ELCR to identify PCI interrupts\n"); 306 ELCR_fallback = 1; 307 } 308 } 309 310 for (i = 0; i < 16; i++) { 311 switch (mpc_default_type) { 312 case 2: 313 if (i == 0 || i == 13) 314 continue; /* IRQ0 & IRQ13 not connected */ 315 /* fall through */ 316 default: 317 if (i == 2) 318 continue; /* IRQ2 is never connected */ 319 } 320 321 if (ELCR_fallback) { 322 /* 323 * If the ELCR indicates a level-sensitive interrupt, we 324 * copy that information over to the MP table in the 325 * irqflag field (level sensitive, active high polarity). 326 */ 327 if (ELCR_trigger(i)) { 328 intsrc.irqflag = MP_IRQTRIG_LEVEL | 329 MP_IRQPOL_ACTIVE_HIGH; 330 } else { 331 intsrc.irqflag = MP_IRQTRIG_DEFAULT | 332 MP_IRQPOL_DEFAULT; 333 } 334 } 335 336 intsrc.srcbusirq = i; 337 intsrc.dstirq = i ? i : 2; /* IRQ0 to INTIN2 */ 338 mp_save_irq(&intsrc); 339 } 340 341 intsrc.irqtype = mp_ExtINT; 342 intsrc.srcbusirq = 0; 343 intsrc.dstirq = 0; /* 8259A to INTIN0 */ 344 mp_save_irq(&intsrc); 345 } 346 347 348 static void __init construct_ioapic_table(int mpc_default_type) 349 { 350 struct mpc_ioapic ioapic; 351 struct mpc_bus bus; 352 353 bus.type = MP_BUS; 354 bus.busid = 0; 355 switch (mpc_default_type) { 356 default: 357 pr_err("???\nUnknown standard configuration %d\n", 358 mpc_default_type); 359 /* fall through */ 360 case 1: 361 case 5: 362 memcpy(bus.bustype, "ISA ", 6); 363 break; 364 case 2: 365 case 6: 366 case 3: 367 memcpy(bus.bustype, "EISA ", 6); 368 break; 369 } 370 MP_bus_info(&bus); 371 if (mpc_default_type > 4) { 372 bus.busid = 1; 373 memcpy(bus.bustype, "PCI ", 6); 374 MP_bus_info(&bus); 375 } 376 377 ioapic.type = MP_IOAPIC; 378 ioapic.apicid = 2; 379 ioapic.apicver = mpc_default_type > 4 ? 0x10 : 0x01; 380 ioapic.flags = MPC_APIC_USABLE; 381 ioapic.apicaddr = IO_APIC_DEFAULT_PHYS_BASE; 382 MP_ioapic_info(&ioapic); 383 384 /* 385 * We set up most of the low 16 IO-APIC pins according to MPS rules. 386 */ 387 construct_default_ioirq_mptable(mpc_default_type); 388 } 389 #else 390 static inline void __init construct_ioapic_table(int mpc_default_type) { } 391 #endif 392 393 static inline void __init construct_default_ISA_mptable(int mpc_default_type) 394 { 395 struct mpc_cpu processor; 396 struct mpc_lintsrc lintsrc; 397 int linttypes[2] = { mp_ExtINT, mp_NMI }; 398 int i; 399 400 /* 401 * local APIC has default address 402 */ 403 mp_lapic_addr = APIC_DEFAULT_PHYS_BASE; 404 405 /* 406 * 2 CPUs, numbered 0 & 1. 407 */ 408 processor.type = MP_PROCESSOR; 409 /* Either an integrated APIC or a discrete 82489DX. */ 410 processor.apicver = mpc_default_type > 4 ? 0x10 : 0x01; 411 processor.cpuflag = CPU_ENABLED; 412 processor.cpufeature = (boot_cpu_data.x86 << 8) | 413 (boot_cpu_data.x86_model << 4) | boot_cpu_data.x86_stepping; 414 processor.featureflag = boot_cpu_data.x86_capability[CPUID_1_EDX]; 415 processor.reserved[0] = 0; 416 processor.reserved[1] = 0; 417 for (i = 0; i < 2; i++) { 418 processor.apicid = i; 419 MP_processor_info(&processor); 420 } 421 422 construct_ioapic_table(mpc_default_type); 423 424 lintsrc.type = MP_LINTSRC; 425 lintsrc.irqflag = MP_IRQTRIG_DEFAULT | MP_IRQPOL_DEFAULT; 426 lintsrc.srcbusid = 0; 427 lintsrc.srcbusirq = 0; 428 lintsrc.destapic = MP_APIC_ALL; 429 for (i = 0; i < 2; i++) { 430 lintsrc.irqtype = linttypes[i]; 431 lintsrc.destapiclint = i; 432 MP_lintsrc_info(&lintsrc); 433 } 434 } 435 436 static unsigned long mpf_base; 437 static bool mpf_found; 438 439 static unsigned long __init get_mpc_size(unsigned long physptr) 440 { 441 struct mpc_table *mpc; 442 unsigned long size; 443 444 mpc = early_memremap(physptr, PAGE_SIZE); 445 size = mpc->length; 446 early_memunmap(mpc, PAGE_SIZE); 447 apic_printk(APIC_VERBOSE, " mpc: %lx-%lx\n", physptr, physptr + size); 448 449 return size; 450 } 451 452 static int __init check_physptr(struct mpf_intel *mpf, unsigned int early) 453 { 454 struct mpc_table *mpc; 455 unsigned long size; 456 457 size = get_mpc_size(mpf->physptr); 458 mpc = early_memremap(mpf->physptr, size); 459 460 /* 461 * Read the physical hardware table. Anything here will 462 * override the defaults. 463 */ 464 if (!smp_read_mpc(mpc, early)) { 465 #ifdef CONFIG_X86_LOCAL_APIC 466 smp_found_config = 0; 467 #endif 468 pr_err("BIOS bug, MP table errors detected!...\n"); 469 pr_cont("... disabling SMP support. (tell your hw vendor)\n"); 470 early_memunmap(mpc, size); 471 return -1; 472 } 473 early_memunmap(mpc, size); 474 475 if (early) 476 return -1; 477 478 #ifdef CONFIG_X86_IO_APIC 479 /* 480 * If there are no explicit MP IRQ entries, then we are 481 * broken. We set up most of the low 16 IO-APIC pins to 482 * ISA defaults and hope it will work. 483 */ 484 if (!mp_irq_entries) { 485 struct mpc_bus bus; 486 487 pr_err("BIOS bug, no explicit IRQ entries, using default mptable. (tell your hw vendor)\n"); 488 489 bus.type = MP_BUS; 490 bus.busid = 0; 491 memcpy(bus.bustype, "ISA ", 6); 492 MP_bus_info(&bus); 493 494 construct_default_ioirq_mptable(0); 495 } 496 #endif 497 498 return 0; 499 } 500 501 /* 502 * Scan the memory blocks for an SMP configuration block. 503 */ 504 void __init default_get_smp_config(unsigned int early) 505 { 506 struct mpf_intel *mpf; 507 508 if (!smp_found_config) 509 return; 510 511 if (!mpf_found) 512 return; 513 514 if (acpi_lapic && early) 515 return; 516 517 /* 518 * MPS doesn't support hyperthreading, aka only have 519 * thread 0 apic id in MPS table 520 */ 521 if (acpi_lapic && acpi_ioapic) 522 return; 523 524 mpf = early_memremap(mpf_base, sizeof(*mpf)); 525 if (!mpf) { 526 pr_err("MPTABLE: error mapping MP table\n"); 527 return; 528 } 529 530 pr_info("Intel MultiProcessor Specification v1.%d\n", 531 mpf->specification); 532 #if defined(CONFIG_X86_LOCAL_APIC) && defined(CONFIG_X86_32) 533 if (mpf->feature2 & (1 << 7)) { 534 pr_info(" IMCR and PIC compatibility mode.\n"); 535 pic_mode = 1; 536 } else { 537 pr_info(" Virtual Wire compatibility mode.\n"); 538 pic_mode = 0; 539 } 540 #endif 541 /* 542 * Now see if we need to read further. 543 */ 544 if (mpf->feature1) { 545 if (early) { 546 /* 547 * local APIC has default address 548 */ 549 mp_lapic_addr = APIC_DEFAULT_PHYS_BASE; 550 goto out; 551 } 552 553 pr_info("Default MP configuration #%d\n", mpf->feature1); 554 construct_default_ISA_mptable(mpf->feature1); 555 556 } else if (mpf->physptr) { 557 if (check_physptr(mpf, early)) 558 goto out; 559 } else 560 BUG(); 561 562 if (!early) 563 pr_info("Processors: %d\n", num_processors); 564 /* 565 * Only use the first configuration found. 566 */ 567 out: 568 early_memunmap(mpf, sizeof(*mpf)); 569 } 570 571 static void __init smp_reserve_memory(struct mpf_intel *mpf) 572 { 573 memblock_reserve(mpf->physptr, get_mpc_size(mpf->physptr)); 574 } 575 576 static int __init smp_scan_config(unsigned long base, unsigned long length) 577 { 578 unsigned int *bp; 579 struct mpf_intel *mpf; 580 int ret = 0; 581 582 apic_printk(APIC_VERBOSE, "Scan for SMP in [mem %#010lx-%#010lx]\n", 583 base, base + length - 1); 584 BUILD_BUG_ON(sizeof(*mpf) != 16); 585 586 while (length > 0) { 587 bp = early_memremap(base, length); 588 mpf = (struct mpf_intel *)bp; 589 if ((*bp == SMP_MAGIC_IDENT) && 590 (mpf->length == 1) && 591 !mpf_checksum((unsigned char *)bp, 16) && 592 ((mpf->specification == 1) 593 || (mpf->specification == 4))) { 594 #ifdef CONFIG_X86_LOCAL_APIC 595 smp_found_config = 1; 596 #endif 597 mpf_base = base; 598 mpf_found = true; 599 600 pr_info("found SMP MP-table at [mem %#010lx-%#010lx]\n", 601 base, base + sizeof(*mpf) - 1); 602 603 memblock_reserve(base, sizeof(*mpf)); 604 if (mpf->physptr) 605 smp_reserve_memory(mpf); 606 607 ret = 1; 608 } 609 early_memunmap(bp, length); 610 611 if (ret) 612 break; 613 614 base += 16; 615 length -= 16; 616 } 617 return ret; 618 } 619 620 void __init default_find_smp_config(void) 621 { 622 unsigned int address; 623 624 /* 625 * FIXME: Linux assumes you have 640K of base ram.. 626 * this continues the error... 627 * 628 * 1) Scan the bottom 1K for a signature 629 * 2) Scan the top 1K of base RAM 630 * 3) Scan the 64K of bios 631 */ 632 if (smp_scan_config(0x0, 0x400) || 633 smp_scan_config(639 * 0x400, 0x400) || 634 smp_scan_config(0xF0000, 0x10000)) 635 return; 636 /* 637 * If it is an SMP machine we should know now, unless the 638 * configuration is in an EISA bus machine with an 639 * extended bios data area. 640 * 641 * there is a real-mode segmented pointer pointing to the 642 * 4K EBDA area at 0x40E, calculate and scan it here. 643 * 644 * NOTE! There are Linux loaders that will corrupt the EBDA 645 * area, and as such this kind of SMP config may be less 646 * trustworthy, simply because the SMP table may have been 647 * stomped on during early boot. These loaders are buggy and 648 * should be fixed. 649 * 650 * MP1.4 SPEC states to only scan first 1K of 4K EBDA. 651 */ 652 653 address = get_bios_ebda(); 654 if (address) 655 smp_scan_config(address, 0x400); 656 } 657 658 #ifdef CONFIG_X86_IO_APIC 659 static u8 __initdata irq_used[MAX_IRQ_SOURCES]; 660 661 static int __init get_MP_intsrc_index(struct mpc_intsrc *m) 662 { 663 int i; 664 665 if (m->irqtype != mp_INT) 666 return 0; 667 668 if (m->irqflag != (MP_IRQTRIG_LEVEL | MP_IRQPOL_ACTIVE_LOW)) 669 return 0; 670 671 /* not legacy */ 672 673 for (i = 0; i < mp_irq_entries; i++) { 674 if (mp_irqs[i].irqtype != mp_INT) 675 continue; 676 677 if (mp_irqs[i].irqflag != (MP_IRQTRIG_LEVEL | 678 MP_IRQPOL_ACTIVE_LOW)) 679 continue; 680 681 if (mp_irqs[i].srcbus != m->srcbus) 682 continue; 683 if (mp_irqs[i].srcbusirq != m->srcbusirq) 684 continue; 685 if (irq_used[i]) { 686 /* already claimed */ 687 return -2; 688 } 689 irq_used[i] = 1; 690 return i; 691 } 692 693 /* not found */ 694 return -1; 695 } 696 697 #define SPARE_SLOT_NUM 20 698 699 static struct mpc_intsrc __initdata *m_spare[SPARE_SLOT_NUM]; 700 701 static void __init check_irq_src(struct mpc_intsrc *m, int *nr_m_spare) 702 { 703 int i; 704 705 apic_printk(APIC_VERBOSE, "OLD "); 706 print_mp_irq_info(m); 707 708 i = get_MP_intsrc_index(m); 709 if (i > 0) { 710 memcpy(m, &mp_irqs[i], sizeof(*m)); 711 apic_printk(APIC_VERBOSE, "NEW "); 712 print_mp_irq_info(&mp_irqs[i]); 713 return; 714 } 715 if (!i) { 716 /* legacy, do nothing */ 717 return; 718 } 719 if (*nr_m_spare < SPARE_SLOT_NUM) { 720 /* 721 * not found (-1), or duplicated (-2) are invalid entries, 722 * we need to use the slot later 723 */ 724 m_spare[*nr_m_spare] = m; 725 *nr_m_spare += 1; 726 } 727 } 728 729 static int __init 730 check_slot(unsigned long mpc_new_phys, unsigned long mpc_new_length, int count) 731 { 732 if (!mpc_new_phys || count <= mpc_new_length) { 733 WARN(1, "update_mptable: No spare slots (length: %x)\n", count); 734 return -1; 735 } 736 737 return 0; 738 } 739 #else /* CONFIG_X86_IO_APIC */ 740 static 741 inline void __init check_irq_src(struct mpc_intsrc *m, int *nr_m_spare) {} 742 #endif /* CONFIG_X86_IO_APIC */ 743 744 static int __init replace_intsrc_all(struct mpc_table *mpc, 745 unsigned long mpc_new_phys, 746 unsigned long mpc_new_length) 747 { 748 #ifdef CONFIG_X86_IO_APIC 749 int i; 750 #endif 751 int count = sizeof(*mpc); 752 int nr_m_spare = 0; 753 unsigned char *mpt = ((unsigned char *)mpc) + count; 754 755 pr_info("mpc_length %x\n", mpc->length); 756 while (count < mpc->length) { 757 switch (*mpt) { 758 case MP_PROCESSOR: 759 skip_entry(&mpt, &count, sizeof(struct mpc_cpu)); 760 break; 761 case MP_BUS: 762 skip_entry(&mpt, &count, sizeof(struct mpc_bus)); 763 break; 764 case MP_IOAPIC: 765 skip_entry(&mpt, &count, sizeof(struct mpc_ioapic)); 766 break; 767 case MP_INTSRC: 768 check_irq_src((struct mpc_intsrc *)mpt, &nr_m_spare); 769 skip_entry(&mpt, &count, sizeof(struct mpc_intsrc)); 770 break; 771 case MP_LINTSRC: 772 skip_entry(&mpt, &count, sizeof(struct mpc_lintsrc)); 773 break; 774 default: 775 /* wrong mptable */ 776 smp_dump_mptable(mpc, mpt); 777 goto out; 778 } 779 } 780 781 #ifdef CONFIG_X86_IO_APIC 782 for (i = 0; i < mp_irq_entries; i++) { 783 if (irq_used[i]) 784 continue; 785 786 if (mp_irqs[i].irqtype != mp_INT) 787 continue; 788 789 if (mp_irqs[i].irqflag != (MP_IRQTRIG_LEVEL | 790 MP_IRQPOL_ACTIVE_LOW)) 791 continue; 792 793 if (nr_m_spare > 0) { 794 apic_printk(APIC_VERBOSE, "*NEW* found\n"); 795 nr_m_spare--; 796 memcpy(m_spare[nr_m_spare], &mp_irqs[i], sizeof(mp_irqs[i])); 797 m_spare[nr_m_spare] = NULL; 798 } else { 799 struct mpc_intsrc *m = (struct mpc_intsrc *)mpt; 800 count += sizeof(struct mpc_intsrc); 801 if (check_slot(mpc_new_phys, mpc_new_length, count) < 0) 802 goto out; 803 memcpy(m, &mp_irqs[i], sizeof(*m)); 804 mpc->length = count; 805 mpt += sizeof(struct mpc_intsrc); 806 } 807 print_mp_irq_info(&mp_irqs[i]); 808 } 809 #endif 810 out: 811 /* update checksum */ 812 mpc->checksum = 0; 813 mpc->checksum -= mpf_checksum((unsigned char *)mpc, mpc->length); 814 815 return 0; 816 } 817 818 int enable_update_mptable; 819 820 static int __init update_mptable_setup(char *str) 821 { 822 enable_update_mptable = 1; 823 #ifdef CONFIG_PCI 824 pci_routeirq = 1; 825 #endif 826 return 0; 827 } 828 early_param("update_mptable", update_mptable_setup); 829 830 static unsigned long __initdata mpc_new_phys; 831 static unsigned long mpc_new_length __initdata = 4096; 832 833 /* alloc_mptable or alloc_mptable=4k */ 834 static int __initdata alloc_mptable; 835 static int __init parse_alloc_mptable_opt(char *p) 836 { 837 enable_update_mptable = 1; 838 #ifdef CONFIG_PCI 839 pci_routeirq = 1; 840 #endif 841 alloc_mptable = 1; 842 if (!p) 843 return 0; 844 mpc_new_length = memparse(p, &p); 845 return 0; 846 } 847 early_param("alloc_mptable", parse_alloc_mptable_opt); 848 849 void __init e820__memblock_alloc_reserved_mpc_new(void) 850 { 851 if (enable_update_mptable && alloc_mptable) 852 mpc_new_phys = e820__memblock_alloc_reserved(mpc_new_length, 4); 853 } 854 855 static int __init update_mp_table(void) 856 { 857 char str[16]; 858 char oem[10]; 859 struct mpf_intel *mpf; 860 struct mpc_table *mpc, *mpc_new; 861 unsigned long size; 862 863 if (!enable_update_mptable) 864 return 0; 865 866 if (!mpf_found) 867 return 0; 868 869 mpf = early_memremap(mpf_base, sizeof(*mpf)); 870 if (!mpf) { 871 pr_err("MPTABLE: mpf early_memremap() failed\n"); 872 return 0; 873 } 874 875 /* 876 * Now see if we need to go further. 877 */ 878 if (mpf->feature1) 879 goto do_unmap_mpf; 880 881 if (!mpf->physptr) 882 goto do_unmap_mpf; 883 884 size = get_mpc_size(mpf->physptr); 885 mpc = early_memremap(mpf->physptr, size); 886 if (!mpc) { 887 pr_err("MPTABLE: mpc early_memremap() failed\n"); 888 goto do_unmap_mpf; 889 } 890 891 if (!smp_check_mpc(mpc, oem, str)) 892 goto do_unmap_mpc; 893 894 pr_info("mpf: %llx\n", (u64)mpf_base); 895 pr_info("physptr: %x\n", mpf->physptr); 896 897 if (mpc_new_phys && mpc->length > mpc_new_length) { 898 mpc_new_phys = 0; 899 pr_info("mpc_new_length is %ld, please use alloc_mptable=8k\n", 900 mpc_new_length); 901 } 902 903 if (!mpc_new_phys) { 904 unsigned char old, new; 905 /* check if we can change the position */ 906 mpc->checksum = 0; 907 old = mpf_checksum((unsigned char *)mpc, mpc->length); 908 mpc->checksum = 0xff; 909 new = mpf_checksum((unsigned char *)mpc, mpc->length); 910 if (old == new) { 911 pr_info("mpc is readonly, please try alloc_mptable instead\n"); 912 goto do_unmap_mpc; 913 } 914 pr_info("use in-position replacing\n"); 915 } else { 916 mpc_new = early_memremap(mpc_new_phys, mpc_new_length); 917 if (!mpc_new) { 918 pr_err("MPTABLE: new mpc early_memremap() failed\n"); 919 goto do_unmap_mpc; 920 } 921 mpf->physptr = mpc_new_phys; 922 memcpy(mpc_new, mpc, mpc->length); 923 early_memunmap(mpc, size); 924 mpc = mpc_new; 925 size = mpc_new_length; 926 /* check if we can modify that */ 927 if (mpc_new_phys - mpf->physptr) { 928 struct mpf_intel *mpf_new; 929 /* steal 16 bytes from [0, 1k) */ 930 mpf_new = early_memremap(0x400 - 16, sizeof(*mpf_new)); 931 if (!mpf_new) { 932 pr_err("MPTABLE: new mpf early_memremap() failed\n"); 933 goto do_unmap_mpc; 934 } 935 pr_info("mpf new: %x\n", 0x400 - 16); 936 memcpy(mpf_new, mpf, 16); 937 early_memunmap(mpf, sizeof(*mpf)); 938 mpf = mpf_new; 939 mpf->physptr = mpc_new_phys; 940 } 941 mpf->checksum = 0; 942 mpf->checksum -= mpf_checksum((unsigned char *)mpf, 16); 943 pr_info("physptr new: %x\n", mpf->physptr); 944 } 945 946 /* 947 * only replace the one with mp_INT and 948 * MP_IRQ_TRIGGER_LEVEL|MP_IRQ_POLARITY_LOW, 949 * already in mp_irqs , stored by ... and mp_config_acpi_gsi, 950 * may need pci=routeirq for all coverage 951 */ 952 replace_intsrc_all(mpc, mpc_new_phys, mpc_new_length); 953 954 do_unmap_mpc: 955 early_memunmap(mpc, size); 956 957 do_unmap_mpf: 958 early_memunmap(mpf, sizeof(*mpf)); 959 960 return 0; 961 } 962 963 late_initcall(update_mp_table); 964