1 /* 2 * inventory.c 3 * 4 * This program is free software; you can redistribute it and/or 5 * modify it under the terms of the GNU General Public License 6 * as published by the Free Software Foundation; either version 7 * 2 of the License, or (at your option) any later version. 8 * 9 * Copyright (c) 1999 The Puffin Group (David Kennedy and Alex deVries) 10 * Copyright (c) 2001 Matthew Wilcox for Hewlett-Packard 11 * 12 * These are the routines to discover what hardware exists in this box. 13 * This task is complicated by there being 3 different ways of 14 * performing an inventory, depending largely on the age of the box. 15 * The recommended way to do this is to check to see whether the machine 16 * is a `Snake' first, then try System Map, then try PAT. We try System 17 * Map before checking for a Snake -- this probably doesn't cause any 18 * problems, but... 19 */ 20 21 #include <linux/types.h> 22 #include <linux/kernel.h> 23 #include <linux/init.h> 24 #include <linux/slab.h> 25 #include <linux/mm.h> 26 #include <asm/hardware.h> 27 #include <asm/io.h> 28 #include <asm/mmzone.h> 29 #include <asm/pdc.h> 30 #include <asm/pdcpat.h> 31 #include <asm/processor.h> 32 #include <asm/page.h> 33 #include <asm/parisc-device.h> 34 35 /* 36 ** Debug options 37 ** DEBUG_PAT Dump details which PDC PAT provides about ranges/devices. 38 */ 39 #undef DEBUG_PAT 40 41 int pdc_type __read_mostly = PDC_TYPE_ILLEGAL; 42 43 void __init setup_pdc(void) 44 { 45 long status; 46 unsigned int bus_id; 47 struct pdc_system_map_mod_info module_result; 48 struct pdc_module_path module_path; 49 struct pdc_model model; 50 #ifdef CONFIG_64BIT 51 struct pdc_pat_cell_num cell_info; 52 #endif 53 54 /* Determine the pdc "type" used on this machine */ 55 56 printk(KERN_INFO "Determining PDC firmware type: "); 57 58 status = pdc_system_map_find_mods(&module_result, &module_path, 0); 59 if (status == PDC_OK) { 60 pdc_type = PDC_TYPE_SYSTEM_MAP; 61 printk("System Map.\n"); 62 return; 63 } 64 65 /* 66 * If the machine doesn't support PDC_SYSTEM_MAP then either it 67 * is a pdc pat box, or it is an older box. All 64 bit capable 68 * machines are either pdc pat boxes or they support PDC_SYSTEM_MAP. 69 */ 70 71 /* 72 * TODO: We should test for 64 bit capability and give a 73 * clearer message. 74 */ 75 76 #ifdef CONFIG_64BIT 77 status = pdc_pat_cell_get_number(&cell_info); 78 if (status == PDC_OK) { 79 pdc_type = PDC_TYPE_PAT; 80 printk("64 bit PAT.\n"); 81 return; 82 } 83 #endif 84 85 /* Check the CPU's bus ID. There's probably a better test. */ 86 87 status = pdc_model_info(&model); 88 89 bus_id = (model.hversion >> (4 + 7)) & 0x1f; 90 91 switch (bus_id) { 92 case 0x4: /* 720, 730, 750, 735, 755 */ 93 case 0x6: /* 705, 710 */ 94 case 0x7: /* 715, 725 */ 95 case 0x8: /* 745, 747, 742 */ 96 case 0xA: /* 712 and similar */ 97 case 0xC: /* 715/64, at least */ 98 99 pdc_type = PDC_TYPE_SNAKE; 100 printk("Snake.\n"); 101 return; 102 103 default: /* Everything else */ 104 105 printk("Unsupported.\n"); 106 panic("If this is a 64-bit machine, please try a 64-bit kernel.\n"); 107 } 108 } 109 110 #define PDC_PAGE_ADJ_SHIFT (PAGE_SHIFT - 12) /* pdc pages are always 4k */ 111 112 static void __init 113 set_pmem_entry(physmem_range_t *pmem_ptr, unsigned long start, 114 unsigned long pages4k) 115 { 116 /* Rather than aligning and potentially throwing away 117 * memory, we'll assume that any ranges are already 118 * nicely aligned with any reasonable page size, and 119 * panic if they are not (it's more likely that the 120 * pdc info is bad in this case). 121 */ 122 123 if (unlikely( ((start & (PAGE_SIZE - 1)) != 0) 124 || ((pages4k & ((1UL << PDC_PAGE_ADJ_SHIFT) - 1)) != 0) )) { 125 126 panic("Memory range doesn't align with page size!\n"); 127 } 128 129 pmem_ptr->start_pfn = (start >> PAGE_SHIFT); 130 pmem_ptr->pages = (pages4k >> PDC_PAGE_ADJ_SHIFT); 131 } 132 133 static void __init pagezero_memconfig(void) 134 { 135 unsigned long npages; 136 137 /* Use the 32 bit information from page zero to create a single 138 * entry in the pmem_ranges[] table. 139 * 140 * We currently don't support machines with contiguous memory 141 * >= 4 Gb, who report that memory using 64 bit only fields 142 * on page zero. It's not worth doing until it can be tested, 143 * and it is not clear we can support those machines for other 144 * reasons. 145 * 146 * If that support is done in the future, this is where it 147 * should be done. 148 */ 149 150 npages = (PAGE_ALIGN(PAGE0->imm_max_mem) >> PAGE_SHIFT); 151 set_pmem_entry(pmem_ranges,0UL,npages); 152 npmem_ranges = 1; 153 } 154 155 #ifdef CONFIG_64BIT 156 157 /* All of the PDC PAT specific code is 64-bit only */ 158 159 /* 160 ** The module object is filled via PDC_PAT_CELL[Return Cell Module]. 161 ** If a module is found, register module will get the IODC bytes via 162 ** pdc_iodc_read() using the PA view of conf_base_addr for the hpa parameter. 163 ** 164 ** The IO view can be used by PDC_PAT_CELL[Return Cell Module] 165 ** only for SBAs and LBAs. This view will cause an invalid 166 ** argument error for all other cell module types. 167 ** 168 */ 169 170 static int __init 171 pat_query_module(ulong pcell_loc, ulong mod_index) 172 { 173 pdc_pat_cell_mod_maddr_block_t *pa_pdc_cell; 174 unsigned long bytecnt; 175 unsigned long temp; /* 64-bit scratch value */ 176 long status; /* PDC return value status */ 177 struct parisc_device *dev; 178 179 pa_pdc_cell = kmalloc(sizeof (*pa_pdc_cell), GFP_KERNEL); 180 if (!pa_pdc_cell) 181 panic("couldn't allocate memory for PDC_PAT_CELL!"); 182 183 /* return cell module (PA or Processor view) */ 184 status = pdc_pat_cell_module(&bytecnt, pcell_loc, mod_index, 185 PA_VIEW, pa_pdc_cell); 186 187 if (status != PDC_OK) { 188 /* no more cell modules or error */ 189 kfree(pa_pdc_cell); 190 return status; 191 } 192 193 temp = pa_pdc_cell->cba; 194 dev = alloc_pa_dev(PAT_GET_CBA(temp), &(pa_pdc_cell->mod_path)); 195 if (!dev) { 196 kfree(pa_pdc_cell); 197 return PDC_OK; 198 } 199 200 /* alloc_pa_dev sets dev->hpa */ 201 202 /* 203 ** save parameters in the parisc_device 204 ** (The idea being the device driver will call pdc_pat_cell_module() 205 ** and store the results in its own data structure.) 206 */ 207 dev->pcell_loc = pcell_loc; 208 dev->mod_index = mod_index; 209 210 /* save generic info returned from the call */ 211 /* REVISIT: who is the consumer of this? not sure yet... */ 212 dev->mod_info = pa_pdc_cell->mod_info; /* pass to PAT_GET_ENTITY() */ 213 dev->pmod_loc = pa_pdc_cell->mod_location; 214 dev->mod0 = pa_pdc_cell->mod[0]; 215 216 register_parisc_device(dev); /* advertise device */ 217 218 #ifdef DEBUG_PAT 219 pdc_pat_cell_mod_maddr_block_t io_pdc_cell; 220 /* dump what we see so far... */ 221 switch (PAT_GET_ENTITY(dev->mod_info)) { 222 unsigned long i; 223 224 case PAT_ENTITY_PROC: 225 printk(KERN_DEBUG "PAT_ENTITY_PROC: id_eid 0x%lx\n", 226 pa_pdc_cell->mod[0]); 227 break; 228 229 case PAT_ENTITY_MEM: 230 printk(KERN_DEBUG 231 "PAT_ENTITY_MEM: amount 0x%lx min_gni_base 0x%lx min_gni_len 0x%lx\n", 232 pa_pdc_cell->mod[0], pa_pdc_cell->mod[1], 233 pa_pdc_cell->mod[2]); 234 break; 235 case PAT_ENTITY_CA: 236 printk(KERN_DEBUG "PAT_ENTITY_CA: %ld\n", pcell_loc); 237 break; 238 239 case PAT_ENTITY_PBC: 240 printk(KERN_DEBUG "PAT_ENTITY_PBC: "); 241 goto print_ranges; 242 243 case PAT_ENTITY_SBA: 244 printk(KERN_DEBUG "PAT_ENTITY_SBA: "); 245 goto print_ranges; 246 247 case PAT_ENTITY_LBA: 248 printk(KERN_DEBUG "PAT_ENTITY_LBA: "); 249 250 print_ranges: 251 pdc_pat_cell_module(&bytecnt, pcell_loc, mod_index, 252 IO_VIEW, &io_pdc_cell); 253 printk(KERN_DEBUG "ranges %ld\n", pa_pdc_cell->mod[1]); 254 for (i = 0; i < pa_pdc_cell->mod[1]; i++) { 255 printk(KERN_DEBUG 256 " PA_VIEW %ld: 0x%016lx 0x%016lx 0x%016lx\n", 257 i, pa_pdc_cell->mod[2 + i * 3], /* type */ 258 pa_pdc_cell->mod[3 + i * 3], /* start */ 259 pa_pdc_cell->mod[4 + i * 3]); /* finish (ie end) */ 260 printk(KERN_DEBUG 261 " IO_VIEW %ld: 0x%016lx 0x%016lx 0x%016lx\n", 262 i, io_pdc_cell->mod[2 + i * 3], /* type */ 263 io_pdc_cell->mod[3 + i * 3], /* start */ 264 io_pdc_cell->mod[4 + i * 3]); /* finish (ie end) */ 265 } 266 printk(KERN_DEBUG "\n"); 267 break; 268 } 269 #endif /* DEBUG_PAT */ 270 271 kfree(pa_pdc_cell); 272 273 return PDC_OK; 274 } 275 276 277 /* pat pdc can return information about a variety of different 278 * types of memory (e.g. firmware,i/o, etc) but we only care about 279 * the usable physical ram right now. Since the firmware specific 280 * information is allocated on the stack, we'll be generous, in 281 * case there is a lot of other information we don't care about. 282 */ 283 284 #define PAT_MAX_RANGES (4 * MAX_PHYSMEM_RANGES) 285 286 static void __init pat_memconfig(void) 287 { 288 unsigned long actual_len; 289 struct pdc_pat_pd_addr_map_entry mem_table[PAT_MAX_RANGES+1]; 290 struct pdc_pat_pd_addr_map_entry *mtbl_ptr; 291 physmem_range_t *pmem_ptr; 292 long status; 293 int entries; 294 unsigned long length; 295 int i; 296 297 length = (PAT_MAX_RANGES + 1) * sizeof(struct pdc_pat_pd_addr_map_entry); 298 299 status = pdc_pat_pd_get_addr_map(&actual_len, mem_table, length, 0L); 300 301 if ((status != PDC_OK) 302 || ((actual_len % sizeof(struct pdc_pat_pd_addr_map_entry)) != 0)) { 303 304 /* The above pdc call shouldn't fail, but, just in 305 * case, just use the PAGE0 info. 306 */ 307 308 printk("\n\n\n"); 309 printk(KERN_WARNING "WARNING! Could not get full memory configuration. " 310 "All memory may not be used!\n\n\n"); 311 pagezero_memconfig(); 312 return; 313 } 314 315 entries = actual_len / sizeof(struct pdc_pat_pd_addr_map_entry); 316 317 if (entries > PAT_MAX_RANGES) { 318 printk(KERN_WARNING "This Machine has more memory ranges than we support!\n"); 319 printk(KERN_WARNING "Some memory may not be used!\n"); 320 } 321 322 /* Copy information into the firmware independent pmem_ranges 323 * array, skipping types we don't care about. Notice we said 324 * "may" above. We'll use all the entries that were returned. 325 */ 326 327 npmem_ranges = 0; 328 mtbl_ptr = mem_table; 329 pmem_ptr = pmem_ranges; /* Global firmware independent table */ 330 for (i = 0; i < entries; i++,mtbl_ptr++) { 331 if ( (mtbl_ptr->entry_type != PAT_MEMORY_DESCRIPTOR) 332 || (mtbl_ptr->memory_type != PAT_MEMTYPE_MEMORY) 333 || (mtbl_ptr->pages == 0) 334 || ( (mtbl_ptr->memory_usage != PAT_MEMUSE_GENERAL) 335 && (mtbl_ptr->memory_usage != PAT_MEMUSE_GI) 336 && (mtbl_ptr->memory_usage != PAT_MEMUSE_GNI) ) ) { 337 338 continue; 339 } 340 341 if (npmem_ranges == MAX_PHYSMEM_RANGES) { 342 printk(KERN_WARNING "This Machine has more memory ranges than we support!\n"); 343 printk(KERN_WARNING "Some memory will not be used!\n"); 344 break; 345 } 346 347 set_pmem_entry(pmem_ptr++,mtbl_ptr->paddr,mtbl_ptr->pages); 348 npmem_ranges++; 349 } 350 } 351 352 static int __init pat_inventory(void) 353 { 354 int status; 355 ulong mod_index = 0; 356 struct pdc_pat_cell_num cell_info; 357 358 /* 359 ** Note: Prelude (and it's successors: Lclass, A400/500) only 360 ** implement PDC_PAT_CELL sub-options 0 and 2. 361 */ 362 status = pdc_pat_cell_get_number(&cell_info); 363 if (status != PDC_OK) { 364 return 0; 365 } 366 367 #ifdef DEBUG_PAT 368 printk(KERN_DEBUG "CELL_GET_NUMBER: 0x%lx 0x%lx\n", cell_info.cell_num, 369 cell_info.cell_loc); 370 #endif 371 372 while (PDC_OK == pat_query_module(cell_info.cell_loc, mod_index)) { 373 mod_index++; 374 } 375 376 return mod_index; 377 } 378 379 /* We only look for extended memory ranges on a 64 bit capable box */ 380 static void __init sprockets_memconfig(void) 381 { 382 struct pdc_memory_table_raddr r_addr; 383 struct pdc_memory_table mem_table[MAX_PHYSMEM_RANGES]; 384 struct pdc_memory_table *mtbl_ptr; 385 physmem_range_t *pmem_ptr; 386 long status; 387 int entries; 388 int i; 389 390 status = pdc_mem_mem_table(&r_addr,mem_table, 391 (unsigned long)MAX_PHYSMEM_RANGES); 392 393 if (status != PDC_OK) { 394 395 /* The above pdc call only works on boxes with sprockets 396 * firmware (newer B,C,J class). Other non PAT PDC machines 397 * do support more than 3.75 Gb of memory, but we don't 398 * support them yet. 399 */ 400 401 pagezero_memconfig(); 402 return; 403 } 404 405 if (r_addr.entries_total > MAX_PHYSMEM_RANGES) { 406 printk(KERN_WARNING "This Machine has more memory ranges than we support!\n"); 407 printk(KERN_WARNING "Some memory will not be used!\n"); 408 } 409 410 entries = (int)r_addr.entries_returned; 411 412 npmem_ranges = 0; 413 mtbl_ptr = mem_table; 414 pmem_ptr = pmem_ranges; /* Global firmware independent table */ 415 for (i = 0; i < entries; i++,mtbl_ptr++) { 416 set_pmem_entry(pmem_ptr++,mtbl_ptr->paddr,mtbl_ptr->pages); 417 npmem_ranges++; 418 } 419 } 420 421 #else /* !CONFIG_64BIT */ 422 423 #define pat_inventory() do { } while (0) 424 #define pat_memconfig() do { } while (0) 425 #define sprockets_memconfig() pagezero_memconfig() 426 427 #endif /* !CONFIG_64BIT */ 428 429 430 #ifndef CONFIG_PA20 431 432 /* Code to support Snake machines (7[2350], 7[235]5, 715/Scorpio) */ 433 434 static struct parisc_device * __init 435 legacy_create_device(struct pdc_memory_map *r_addr, 436 struct pdc_module_path *module_path) 437 { 438 struct parisc_device *dev; 439 int status = pdc_mem_map_hpa(r_addr, module_path); 440 if (status != PDC_OK) 441 return NULL; 442 443 dev = alloc_pa_dev(r_addr->hpa, &module_path->path); 444 if (dev == NULL) 445 return NULL; 446 447 register_parisc_device(dev); 448 return dev; 449 } 450 451 /** 452 * snake_inventory 453 * 454 * Before PDC_SYSTEM_MAP was invented, the PDC_MEM_MAP call was used. 455 * To use it, we initialise the mod_path.bc to 0xff and try all values of 456 * mod to get the HPA for the top-level devices. Bus adapters may have 457 * sub-devices which are discovered by setting bc[5] to 0 and bc[4] to the 458 * module, then trying all possible functions. 459 */ 460 static void __init snake_inventory(void) 461 { 462 int mod; 463 for (mod = 0; mod < 16; mod++) { 464 struct parisc_device *dev; 465 struct pdc_module_path module_path; 466 struct pdc_memory_map r_addr; 467 unsigned int func; 468 469 memset(module_path.path.bc, 0xff, 6); 470 module_path.path.mod = mod; 471 dev = legacy_create_device(&r_addr, &module_path); 472 if ((!dev) || (dev->id.hw_type != HPHW_BA)) 473 continue; 474 475 memset(module_path.path.bc, 0xff, 4); 476 module_path.path.bc[4] = mod; 477 478 for (func = 0; func < 16; func++) { 479 module_path.path.bc[5] = 0; 480 module_path.path.mod = func; 481 legacy_create_device(&r_addr, &module_path); 482 } 483 } 484 } 485 486 #else /* CONFIG_PA20 */ 487 #define snake_inventory() do { } while (0) 488 #endif /* CONFIG_PA20 */ 489 490 /* Common 32/64 bit based code goes here */ 491 492 /** 493 * add_system_map_addresses - Add additional addresses to the parisc device. 494 * @dev: The parisc device. 495 * @num_addrs: Then number of addresses to add; 496 * @module_instance: The system_map module instance. 497 * 498 * This function adds any additional addresses reported by the system_map 499 * firmware to the parisc device. 500 */ 501 static void __init 502 add_system_map_addresses(struct parisc_device *dev, int num_addrs, 503 int module_instance) 504 { 505 int i; 506 long status; 507 struct pdc_system_map_addr_info addr_result; 508 509 dev->addr = kmalloc(num_addrs * sizeof(unsigned long), GFP_KERNEL); 510 if(!dev->addr) { 511 printk(KERN_ERR "%s %s(): memory allocation failure\n", 512 __FILE__, __func__); 513 return; 514 } 515 516 for(i = 1; i <= num_addrs; ++i) { 517 status = pdc_system_map_find_addrs(&addr_result, 518 module_instance, i); 519 if(PDC_OK == status) { 520 dev->addr[dev->num_addrs] = (unsigned long)addr_result.mod_addr; 521 dev->num_addrs++; 522 } else { 523 printk(KERN_WARNING 524 "Bad PDC_FIND_ADDRESS status return (%ld) for index %d\n", 525 status, i); 526 } 527 } 528 } 529 530 /** 531 * system_map_inventory - Retrieve firmware devices via SYSTEM_MAP. 532 * 533 * This function attempts to retrieve and register all the devices firmware 534 * knows about via the SYSTEM_MAP PDC call. 535 */ 536 static void __init system_map_inventory(void) 537 { 538 int i; 539 long status = PDC_OK; 540 541 for (i = 0; i < 256; i++) { 542 struct parisc_device *dev; 543 struct pdc_system_map_mod_info module_result; 544 struct pdc_module_path module_path; 545 546 status = pdc_system_map_find_mods(&module_result, 547 &module_path, i); 548 if ((status == PDC_BAD_PROC) || (status == PDC_NE_MOD)) 549 break; 550 if (status != PDC_OK) 551 continue; 552 553 dev = alloc_pa_dev(module_result.mod_addr, &module_path.path); 554 if (!dev) 555 continue; 556 557 register_parisc_device(dev); 558 559 /* if available, get the additional addresses for a module */ 560 if (!module_result.add_addrs) 561 continue; 562 563 add_system_map_addresses(dev, module_result.add_addrs, i); 564 } 565 566 walk_central_bus(); 567 return; 568 } 569 570 void __init do_memory_inventory(void) 571 { 572 switch (pdc_type) { 573 574 case PDC_TYPE_PAT: 575 pat_memconfig(); 576 break; 577 578 case PDC_TYPE_SYSTEM_MAP: 579 sprockets_memconfig(); 580 break; 581 582 case PDC_TYPE_SNAKE: 583 pagezero_memconfig(); 584 return; 585 586 default: 587 panic("Unknown PDC type!\n"); 588 } 589 590 if (npmem_ranges == 0 || pmem_ranges[0].start_pfn != 0) { 591 printk(KERN_WARNING "Bad memory configuration returned!\n"); 592 printk(KERN_WARNING "Some memory may not be used!\n"); 593 pagezero_memconfig(); 594 } 595 } 596 597 void __init do_device_inventory(void) 598 { 599 printk(KERN_INFO "Searching for devices...\n"); 600 601 init_parisc_bus(); 602 603 switch (pdc_type) { 604 605 case PDC_TYPE_PAT: 606 pat_inventory(); 607 break; 608 609 case PDC_TYPE_SYSTEM_MAP: 610 system_map_inventory(); 611 break; 612 613 case PDC_TYPE_SNAKE: 614 snake_inventory(); 615 break; 616 617 default: 618 panic("Unknown PDC type!\n"); 619 } 620 printk(KERN_INFO "Found devices:\n"); 621 print_parisc_devices(); 622 } 623