1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> 4 * (C) Copyright 2007 Novell Inc. 5 */ 6 7 #include <linux/pci.h> 8 #include <linux/module.h> 9 #include <linux/init.h> 10 #include <linux/device.h> 11 #include <linux/mempolicy.h> 12 #include <linux/string.h> 13 #include <linux/slab.h> 14 #include <linux/sched.h> 15 #include <linux/sched/isolation.h> 16 #include <linux/cpu.h> 17 #include <linux/pm_runtime.h> 18 #include <linux/suspend.h> 19 #include <linux/kexec.h> 20 #include <linux/of_device.h> 21 #include <linux/acpi.h> 22 #include <linux/dma-map-ops.h> 23 #include "pci.h" 24 #include "pcie/portdrv.h" 25 26 struct pci_dynid { 27 struct list_head node; 28 struct pci_device_id id; 29 }; 30 31 /** 32 * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices 33 * @drv: target pci driver 34 * @vendor: PCI vendor ID 35 * @device: PCI device ID 36 * @subvendor: PCI subvendor ID 37 * @subdevice: PCI subdevice ID 38 * @class: PCI class 39 * @class_mask: PCI class mask 40 * @driver_data: private driver data 41 * 42 * Adds a new dynamic pci device ID to this driver and causes the 43 * driver to probe for all devices again. @drv must have been 44 * registered prior to calling this function. 45 * 46 * CONTEXT: 47 * Does GFP_KERNEL allocation. 48 * 49 * RETURNS: 50 * 0 on success, -errno on failure. 51 */ 52 int pci_add_dynid(struct pci_driver *drv, 53 unsigned int vendor, unsigned int device, 54 unsigned int subvendor, unsigned int subdevice, 55 unsigned int class, unsigned int class_mask, 56 unsigned long driver_data) 57 { 58 struct pci_dynid *dynid; 59 60 dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); 61 if (!dynid) 62 return -ENOMEM; 63 64 dynid->id.vendor = vendor; 65 dynid->id.device = device; 66 dynid->id.subvendor = subvendor; 67 dynid->id.subdevice = subdevice; 68 dynid->id.class = class; 69 dynid->id.class_mask = class_mask; 70 dynid->id.driver_data = driver_data; 71 72 spin_lock(&drv->dynids.lock); 73 list_add_tail(&dynid->node, &drv->dynids.list); 74 spin_unlock(&drv->dynids.lock); 75 76 return driver_attach(&drv->driver); 77 } 78 EXPORT_SYMBOL_GPL(pci_add_dynid); 79 80 static void pci_free_dynids(struct pci_driver *drv) 81 { 82 struct pci_dynid *dynid, *n; 83 84 spin_lock(&drv->dynids.lock); 85 list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { 86 list_del(&dynid->node); 87 kfree(dynid); 88 } 89 spin_unlock(&drv->dynids.lock); 90 } 91 92 /** 93 * pci_match_id - See if a PCI device matches a given pci_id table 94 * @ids: array of PCI device ID structures to search in 95 * @dev: the PCI device structure to match against. 96 * 97 * Used by a driver to check whether a PCI device is in its list of 98 * supported devices. Returns the matching pci_device_id structure or 99 * %NULL if there is no match. 100 * 101 * Deprecated; don't use this as it will not catch any dynamic IDs 102 * that a driver might want to check for. 103 */ 104 const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, 105 struct pci_dev *dev) 106 { 107 if (ids) { 108 while (ids->vendor || ids->subvendor || ids->class_mask) { 109 if (pci_match_one_device(ids, dev)) 110 return ids; 111 ids++; 112 } 113 } 114 return NULL; 115 } 116 EXPORT_SYMBOL(pci_match_id); 117 118 static const struct pci_device_id pci_device_id_any = { 119 .vendor = PCI_ANY_ID, 120 .device = PCI_ANY_ID, 121 .subvendor = PCI_ANY_ID, 122 .subdevice = PCI_ANY_ID, 123 }; 124 125 /** 126 * pci_match_device - See if a device matches a driver's list of IDs 127 * @drv: the PCI driver to match against 128 * @dev: the PCI device structure to match against 129 * 130 * Used by a driver to check whether a PCI device is in its list of 131 * supported devices or in the dynids list, which may have been augmented 132 * via the sysfs "new_id" file. Returns the matching pci_device_id 133 * structure or %NULL if there is no match. 134 */ 135 static const struct pci_device_id *pci_match_device(struct pci_driver *drv, 136 struct pci_dev *dev) 137 { 138 struct pci_dynid *dynid; 139 const struct pci_device_id *found_id = NULL, *ids; 140 141 /* When driver_override is set, only bind to the matching driver */ 142 if (dev->driver_override && strcmp(dev->driver_override, drv->name)) 143 return NULL; 144 145 /* Look at the dynamic ids first, before the static ones */ 146 spin_lock(&drv->dynids.lock); 147 list_for_each_entry(dynid, &drv->dynids.list, node) { 148 if (pci_match_one_device(&dynid->id, dev)) { 149 found_id = &dynid->id; 150 break; 151 } 152 } 153 spin_unlock(&drv->dynids.lock); 154 155 if (found_id) 156 return found_id; 157 158 for (ids = drv->id_table; (found_id = pci_match_id(ids, dev)); 159 ids = found_id + 1) { 160 /* 161 * The match table is split based on driver_override. 162 * In case override_only was set, enforce driver_override 163 * matching. 164 */ 165 if (found_id->override_only) { 166 if (dev->driver_override) 167 return found_id; 168 } else { 169 return found_id; 170 } 171 } 172 173 /* driver_override will always match, send a dummy id */ 174 if (dev->driver_override) 175 return &pci_device_id_any; 176 return NULL; 177 } 178 179 /** 180 * new_id_store - sysfs frontend to pci_add_dynid() 181 * @driver: target device driver 182 * @buf: buffer for scanning device ID data 183 * @count: input size 184 * 185 * Allow PCI IDs to be added to an existing driver via sysfs. 186 */ 187 static ssize_t new_id_store(struct device_driver *driver, const char *buf, 188 size_t count) 189 { 190 struct pci_driver *pdrv = to_pci_driver(driver); 191 const struct pci_device_id *ids = pdrv->id_table; 192 u32 vendor, device, subvendor = PCI_ANY_ID, 193 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 194 unsigned long driver_data = 0; 195 int fields = 0; 196 int retval = 0; 197 198 fields = sscanf(buf, "%x %x %x %x %x %x %lx", 199 &vendor, &device, &subvendor, &subdevice, 200 &class, &class_mask, &driver_data); 201 if (fields < 2) 202 return -EINVAL; 203 204 if (fields != 7) { 205 struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); 206 if (!pdev) 207 return -ENOMEM; 208 209 pdev->vendor = vendor; 210 pdev->device = device; 211 pdev->subsystem_vendor = subvendor; 212 pdev->subsystem_device = subdevice; 213 pdev->class = class; 214 215 if (pci_match_device(pdrv, pdev)) 216 retval = -EEXIST; 217 218 kfree(pdev); 219 220 if (retval) 221 return retval; 222 } 223 224 /* Only accept driver_data values that match an existing id_table 225 entry */ 226 if (ids) { 227 retval = -EINVAL; 228 while (ids->vendor || ids->subvendor || ids->class_mask) { 229 if (driver_data == ids->driver_data) { 230 retval = 0; 231 break; 232 } 233 ids++; 234 } 235 if (retval) /* No match */ 236 return retval; 237 } 238 239 retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, 240 class, class_mask, driver_data); 241 if (retval) 242 return retval; 243 return count; 244 } 245 static DRIVER_ATTR_WO(new_id); 246 247 /** 248 * remove_id_store - remove a PCI device ID from this driver 249 * @driver: target device driver 250 * @buf: buffer for scanning device ID data 251 * @count: input size 252 * 253 * Removes a dynamic pci device ID to this driver. 254 */ 255 static ssize_t remove_id_store(struct device_driver *driver, const char *buf, 256 size_t count) 257 { 258 struct pci_dynid *dynid, *n; 259 struct pci_driver *pdrv = to_pci_driver(driver); 260 u32 vendor, device, subvendor = PCI_ANY_ID, 261 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 262 int fields = 0; 263 size_t retval = -ENODEV; 264 265 fields = sscanf(buf, "%x %x %x %x %x %x", 266 &vendor, &device, &subvendor, &subdevice, 267 &class, &class_mask); 268 if (fields < 2) 269 return -EINVAL; 270 271 spin_lock(&pdrv->dynids.lock); 272 list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { 273 struct pci_device_id *id = &dynid->id; 274 if ((id->vendor == vendor) && 275 (id->device == device) && 276 (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && 277 (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && 278 !((id->class ^ class) & class_mask)) { 279 list_del(&dynid->node); 280 kfree(dynid); 281 retval = count; 282 break; 283 } 284 } 285 spin_unlock(&pdrv->dynids.lock); 286 287 return retval; 288 } 289 static DRIVER_ATTR_WO(remove_id); 290 291 static struct attribute *pci_drv_attrs[] = { 292 &driver_attr_new_id.attr, 293 &driver_attr_remove_id.attr, 294 NULL, 295 }; 296 ATTRIBUTE_GROUPS(pci_drv); 297 298 struct drv_dev_and_id { 299 struct pci_driver *drv; 300 struct pci_dev *dev; 301 const struct pci_device_id *id; 302 }; 303 304 static long local_pci_probe(void *_ddi) 305 { 306 struct drv_dev_and_id *ddi = _ddi; 307 struct pci_dev *pci_dev = ddi->dev; 308 struct pci_driver *pci_drv = ddi->drv; 309 struct device *dev = &pci_dev->dev; 310 int rc; 311 312 /* 313 * Unbound PCI devices are always put in D0, regardless of 314 * runtime PM status. During probe, the device is set to 315 * active and the usage count is incremented. If the driver 316 * supports runtime PM, it should call pm_runtime_put_noidle(), 317 * or any other runtime PM helper function decrementing the usage 318 * count, in its probe routine and pm_runtime_get_noresume() in 319 * its remove routine. 320 */ 321 pm_runtime_get_sync(dev); 322 pci_dev->driver = pci_drv; 323 rc = pci_drv->probe(pci_dev, ddi->id); 324 if (!rc) 325 return rc; 326 if (rc < 0) { 327 pci_dev->driver = NULL; 328 pm_runtime_put_sync(dev); 329 return rc; 330 } 331 /* 332 * Probe function should return < 0 for failure, 0 for success 333 * Treat values > 0 as success, but warn. 334 */ 335 pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", 336 rc); 337 return 0; 338 } 339 340 static bool pci_physfn_is_probed(struct pci_dev *dev) 341 { 342 #ifdef CONFIG_PCI_IOV 343 return dev->is_virtfn && dev->physfn->is_probed; 344 #else 345 return false; 346 #endif 347 } 348 349 static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, 350 const struct pci_device_id *id) 351 { 352 int error, node, cpu; 353 struct drv_dev_and_id ddi = { drv, dev, id }; 354 355 /* 356 * Execute driver initialization on node where the device is 357 * attached. This way the driver likely allocates its local memory 358 * on the right node. 359 */ 360 node = dev_to_node(&dev->dev); 361 dev->is_probed = 1; 362 363 cpu_hotplug_disable(); 364 365 /* 366 * Prevent nesting work_on_cpu() for the case where a Virtual Function 367 * device is probed from work_on_cpu() of the Physical device. 368 */ 369 if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || 370 pci_physfn_is_probed(dev)) { 371 cpu = nr_cpu_ids; 372 } else { 373 cpumask_var_t wq_domain_mask; 374 375 if (!zalloc_cpumask_var(&wq_domain_mask, GFP_KERNEL)) { 376 error = -ENOMEM; 377 goto out; 378 } 379 cpumask_and(wq_domain_mask, 380 housekeeping_cpumask(HK_TYPE_WQ), 381 housekeeping_cpumask(HK_TYPE_DOMAIN)); 382 383 cpu = cpumask_any_and(cpumask_of_node(node), 384 wq_domain_mask); 385 free_cpumask_var(wq_domain_mask); 386 } 387 388 if (cpu < nr_cpu_ids) 389 error = work_on_cpu(cpu, local_pci_probe, &ddi); 390 else 391 error = local_pci_probe(&ddi); 392 out: 393 dev->is_probed = 0; 394 cpu_hotplug_enable(); 395 return error; 396 } 397 398 /** 399 * __pci_device_probe - check if a driver wants to claim a specific PCI device 400 * @drv: driver to call to check if it wants the PCI device 401 * @pci_dev: PCI device being probed 402 * 403 * returns 0 on success, else error. 404 * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. 405 */ 406 static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) 407 { 408 const struct pci_device_id *id; 409 int error = 0; 410 411 if (drv->probe) { 412 error = -ENODEV; 413 414 id = pci_match_device(drv, pci_dev); 415 if (id) 416 error = pci_call_probe(drv, pci_dev, id); 417 } 418 return error; 419 } 420 421 int __weak pcibios_alloc_irq(struct pci_dev *dev) 422 { 423 return 0; 424 } 425 426 void __weak pcibios_free_irq(struct pci_dev *dev) 427 { 428 } 429 430 #ifdef CONFIG_PCI_IOV 431 static inline bool pci_device_can_probe(struct pci_dev *pdev) 432 { 433 return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || 434 pdev->driver_override); 435 } 436 #else 437 static inline bool pci_device_can_probe(struct pci_dev *pdev) 438 { 439 return true; 440 } 441 #endif 442 443 static int pci_device_probe(struct device *dev) 444 { 445 int error; 446 struct pci_dev *pci_dev = to_pci_dev(dev); 447 struct pci_driver *drv = to_pci_driver(dev->driver); 448 449 if (!pci_device_can_probe(pci_dev)) 450 return -ENODEV; 451 452 pci_assign_irq(pci_dev); 453 454 error = pcibios_alloc_irq(pci_dev); 455 if (error < 0) 456 return error; 457 458 pci_dev_get(pci_dev); 459 error = __pci_device_probe(drv, pci_dev); 460 if (error) { 461 pcibios_free_irq(pci_dev); 462 pci_dev_put(pci_dev); 463 } 464 465 return error; 466 } 467 468 static void pci_device_remove(struct device *dev) 469 { 470 struct pci_dev *pci_dev = to_pci_dev(dev); 471 struct pci_driver *drv = pci_dev->driver; 472 473 if (drv->remove) { 474 pm_runtime_get_sync(dev); 475 drv->remove(pci_dev); 476 pm_runtime_put_noidle(dev); 477 } 478 pcibios_free_irq(pci_dev); 479 pci_dev->driver = NULL; 480 pci_iov_remove(pci_dev); 481 482 /* Undo the runtime PM settings in local_pci_probe() */ 483 pm_runtime_put_sync(dev); 484 485 /* 486 * If the device is still on, set the power state as "unknown", 487 * since it might change by the next time we load the driver. 488 */ 489 if (pci_dev->current_state == PCI_D0) 490 pci_dev->current_state = PCI_UNKNOWN; 491 492 /* 493 * We would love to complain here if pci_dev->is_enabled is set, that 494 * the driver should have called pci_disable_device(), but the 495 * unfortunate fact is there are too many odd BIOS and bridge setups 496 * that don't like drivers doing that all of the time. 497 * Oh well, we can dream of sane hardware when we sleep, no matter how 498 * horrible the crap we have to deal with is when we are awake... 499 */ 500 501 pci_dev_put(pci_dev); 502 } 503 504 static void pci_device_shutdown(struct device *dev) 505 { 506 struct pci_dev *pci_dev = to_pci_dev(dev); 507 struct pci_driver *drv = pci_dev->driver; 508 509 pm_runtime_resume(dev); 510 511 if (drv && drv->shutdown) 512 drv->shutdown(pci_dev); 513 514 /* 515 * If this is a kexec reboot, turn off Bus Master bit on the 516 * device to tell it to not continue to do DMA. Don't touch 517 * devices in D3cold or unknown states. 518 * If it is not a kexec reboot, firmware will hit the PCI 519 * devices with big hammer and stop their DMA any way. 520 */ 521 if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) 522 pci_clear_master(pci_dev); 523 } 524 525 #ifdef CONFIG_PM_SLEEP 526 527 /* Auxiliary functions used for system resume */ 528 529 /** 530 * pci_restore_standard_config - restore standard config registers of PCI device 531 * @pci_dev: PCI device to handle 532 */ 533 static int pci_restore_standard_config(struct pci_dev *pci_dev) 534 { 535 pci_update_current_state(pci_dev, PCI_UNKNOWN); 536 537 if (pci_dev->current_state != PCI_D0) { 538 int error = pci_set_power_state(pci_dev, PCI_D0); 539 if (error) 540 return error; 541 } 542 543 pci_restore_state(pci_dev); 544 pci_pme_restore(pci_dev); 545 return 0; 546 } 547 #endif /* CONFIG_PM_SLEEP */ 548 549 #ifdef CONFIG_PM 550 551 /* Auxiliary functions used for system resume and run-time resume */ 552 553 static void pci_pm_default_resume(struct pci_dev *pci_dev) 554 { 555 pci_fixup_device(pci_fixup_resume, pci_dev); 556 pci_enable_wake(pci_dev, PCI_D0, false); 557 } 558 559 static void pci_pm_power_up_and_verify_state(struct pci_dev *pci_dev) 560 { 561 pci_power_up(pci_dev); 562 pci_update_current_state(pci_dev, PCI_D0); 563 } 564 565 static void pci_pm_default_resume_early(struct pci_dev *pci_dev) 566 { 567 pci_pm_power_up_and_verify_state(pci_dev); 568 pci_restore_state(pci_dev); 569 pci_pme_restore(pci_dev); 570 } 571 572 static void pci_pm_bridge_power_up_actions(struct pci_dev *pci_dev) 573 { 574 pci_bridge_wait_for_secondary_bus(pci_dev); 575 /* 576 * When powering on a bridge from D3cold, the whole hierarchy may be 577 * powered on into D0uninitialized state, resume them to give them a 578 * chance to suspend again 579 */ 580 pci_resume_bus(pci_dev->subordinate); 581 } 582 583 #endif /* CONFIG_PM */ 584 585 #ifdef CONFIG_PM_SLEEP 586 587 /* 588 * Default "suspend" method for devices that have no driver provided suspend, 589 * or not even a driver at all (second part). 590 */ 591 static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) 592 { 593 /* 594 * mark its power state as "unknown", since we don't know if 595 * e.g. the BIOS will change its device state when we suspend. 596 */ 597 if (pci_dev->current_state == PCI_D0) 598 pci_dev->current_state = PCI_UNKNOWN; 599 } 600 601 /* 602 * Default "resume" method for devices that have no driver provided resume, 603 * or not even a driver at all (second part). 604 */ 605 static int pci_pm_reenable_device(struct pci_dev *pci_dev) 606 { 607 int retval; 608 609 /* if the device was enabled before suspend, re-enable */ 610 retval = pci_reenable_device(pci_dev); 611 /* 612 * if the device was busmaster before the suspend, make it busmaster 613 * again 614 */ 615 if (pci_dev->is_busmaster) 616 pci_set_master(pci_dev); 617 618 return retval; 619 } 620 621 static int pci_legacy_suspend(struct device *dev, pm_message_t state) 622 { 623 struct pci_dev *pci_dev = to_pci_dev(dev); 624 struct pci_driver *drv = pci_dev->driver; 625 626 if (drv && drv->suspend) { 627 pci_power_t prev = pci_dev->current_state; 628 int error; 629 630 error = drv->suspend(pci_dev, state); 631 suspend_report_result(dev, drv->suspend, error); 632 if (error) 633 return error; 634 635 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 636 && pci_dev->current_state != PCI_UNKNOWN) { 637 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 638 "PCI PM: Device state not saved by %pS\n", 639 drv->suspend); 640 } 641 } 642 643 pci_fixup_device(pci_fixup_suspend, pci_dev); 644 645 return 0; 646 } 647 648 static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) 649 { 650 struct pci_dev *pci_dev = to_pci_dev(dev); 651 652 if (!pci_dev->state_saved) 653 pci_save_state(pci_dev); 654 655 pci_pm_set_unknown_state(pci_dev); 656 657 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 658 659 return 0; 660 } 661 662 static int pci_legacy_resume(struct device *dev) 663 { 664 struct pci_dev *pci_dev = to_pci_dev(dev); 665 struct pci_driver *drv = pci_dev->driver; 666 667 pci_fixup_device(pci_fixup_resume, pci_dev); 668 669 return drv && drv->resume ? 670 drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); 671 } 672 673 /* Auxiliary functions used by the new power management framework */ 674 675 static void pci_pm_default_suspend(struct pci_dev *pci_dev) 676 { 677 /* Disable non-bridge devices without PM support */ 678 if (!pci_has_subordinate(pci_dev)) 679 pci_disable_enabled_device(pci_dev); 680 } 681 682 static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) 683 { 684 struct pci_driver *drv = pci_dev->driver; 685 bool ret = drv && (drv->suspend || drv->resume); 686 687 /* 688 * Legacy PM support is used by default, so warn if the new framework is 689 * supported as well. Drivers are supposed to support either the 690 * former, or the latter, but not both at the same time. 691 */ 692 pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", 693 pci_dev->vendor, pci_dev->device); 694 695 return ret; 696 } 697 698 /* New power management framework */ 699 700 static int pci_pm_prepare(struct device *dev) 701 { 702 struct pci_dev *pci_dev = to_pci_dev(dev); 703 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 704 705 if (pm && pm->prepare) { 706 int error = pm->prepare(dev); 707 if (error < 0) 708 return error; 709 710 if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) 711 return 0; 712 } 713 if (pci_dev_need_resume(pci_dev)) 714 return 0; 715 716 /* 717 * The PME setting needs to be adjusted here in case the direct-complete 718 * optimization is used with respect to this device. 719 */ 720 pci_dev_adjust_pme(pci_dev); 721 return 1; 722 } 723 724 static void pci_pm_complete(struct device *dev) 725 { 726 struct pci_dev *pci_dev = to_pci_dev(dev); 727 728 pci_dev_complete_resume(pci_dev); 729 pm_generic_complete(dev); 730 731 /* Resume device if platform firmware has put it in reset-power-on */ 732 if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { 733 pci_power_t pre_sleep_state = pci_dev->current_state; 734 735 pci_refresh_power_state(pci_dev); 736 /* 737 * On platforms with ACPI this check may also trigger for 738 * devices sharing power resources if one of those power 739 * resources has been activated as a result of a change of the 740 * power state of another device sharing it. However, in that 741 * case it is also better to resume the device, in general. 742 */ 743 if (pci_dev->current_state < pre_sleep_state) 744 pm_request_resume(dev); 745 } 746 } 747 748 #else /* !CONFIG_PM_SLEEP */ 749 750 #define pci_pm_prepare NULL 751 #define pci_pm_complete NULL 752 753 #endif /* !CONFIG_PM_SLEEP */ 754 755 #ifdef CONFIG_SUSPEND 756 static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) 757 { 758 /* 759 * Some BIOSes forget to clear Root PME Status bits after system 760 * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. 761 * Clear those bits now just in case (shouldn't hurt). 762 */ 763 if (pci_is_pcie(pci_dev) && 764 (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || 765 pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) 766 pcie_clear_root_pme_status(pci_dev); 767 } 768 769 static int pci_pm_suspend(struct device *dev) 770 { 771 struct pci_dev *pci_dev = to_pci_dev(dev); 772 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 773 774 pci_dev->skip_bus_pm = false; 775 776 if (pci_has_legacy_pm_support(pci_dev)) 777 return pci_legacy_suspend(dev, PMSG_SUSPEND); 778 779 if (!pm) { 780 pci_pm_default_suspend(pci_dev); 781 return 0; 782 } 783 784 /* 785 * PCI devices suspended at run time may need to be resumed at this 786 * point, because in general it may be necessary to reconfigure them for 787 * system suspend. Namely, if the device is expected to wake up the 788 * system from the sleep state, it may have to be reconfigured for this 789 * purpose, or if the device is not expected to wake up the system from 790 * the sleep state, it should be prevented from signaling wakeup events 791 * going forward. 792 * 793 * Also if the driver of the device does not indicate that its system 794 * suspend callbacks can cope with runtime-suspended devices, it is 795 * better to resume the device from runtime suspend here. 796 */ 797 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 798 pci_dev_need_resume(pci_dev)) { 799 pm_runtime_resume(dev); 800 pci_dev->state_saved = false; 801 } else { 802 pci_dev_adjust_pme(pci_dev); 803 } 804 805 if (pm->suspend) { 806 pci_power_t prev = pci_dev->current_state; 807 int error; 808 809 error = pm->suspend(dev); 810 suspend_report_result(dev, pm->suspend, error); 811 if (error) 812 return error; 813 814 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 815 && pci_dev->current_state != PCI_UNKNOWN) { 816 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 817 "PCI PM: State of device not saved by %pS\n", 818 pm->suspend); 819 } 820 } 821 822 return 0; 823 } 824 825 static int pci_pm_suspend_late(struct device *dev) 826 { 827 if (dev_pm_skip_suspend(dev)) 828 return 0; 829 830 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 831 832 return pm_generic_suspend_late(dev); 833 } 834 835 static int pci_pm_suspend_noirq(struct device *dev) 836 { 837 struct pci_dev *pci_dev = to_pci_dev(dev); 838 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 839 840 if (dev_pm_skip_suspend(dev)) 841 return 0; 842 843 if (pci_has_legacy_pm_support(pci_dev)) 844 return pci_legacy_suspend_late(dev, PMSG_SUSPEND); 845 846 if (!pm) { 847 pci_save_state(pci_dev); 848 goto Fixup; 849 } 850 851 if (pm->suspend_noirq) { 852 pci_power_t prev = pci_dev->current_state; 853 int error; 854 855 error = pm->suspend_noirq(dev); 856 suspend_report_result(dev, pm->suspend_noirq, error); 857 if (error) 858 return error; 859 860 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 861 && pci_dev->current_state != PCI_UNKNOWN) { 862 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 863 "PCI PM: State of device not saved by %pS\n", 864 pm->suspend_noirq); 865 goto Fixup; 866 } 867 } 868 869 if (pci_dev->skip_bus_pm) { 870 /* 871 * Either the device is a bridge with a child in D0 below it, or 872 * the function is running for the second time in a row without 873 * going through full resume, which is possible only during 874 * suspend-to-idle in a spurious wakeup case. The device should 875 * be in D0 at this point, but if it is a bridge, it may be 876 * necessary to save its state. 877 */ 878 if (!pci_dev->state_saved) 879 pci_save_state(pci_dev); 880 } else if (!pci_dev->state_saved) { 881 pci_save_state(pci_dev); 882 if (pci_power_manageable(pci_dev)) 883 pci_prepare_to_sleep(pci_dev); 884 } 885 886 pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", 887 pci_power_name(pci_dev->current_state)); 888 889 if (pci_dev->current_state == PCI_D0) { 890 pci_dev->skip_bus_pm = true; 891 /* 892 * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any 893 * downstream device is in D0, so avoid changing the power state 894 * of the parent bridge by setting the skip_bus_pm flag for it. 895 */ 896 if (pci_dev->bus->self) 897 pci_dev->bus->self->skip_bus_pm = true; 898 } 899 900 if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { 901 pci_dbg(pci_dev, "PCI PM: Skipped\n"); 902 goto Fixup; 903 } 904 905 pci_pm_set_unknown_state(pci_dev); 906 907 /* 908 * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's 909 * PCI COMMAND register isn't 0, the BIOS assumes that the controller 910 * hasn't been quiesced and tries to turn it off. If the controller 911 * is already in D3, this can hang or cause memory corruption. 912 * 913 * Since the value of the COMMAND register doesn't matter once the 914 * device has been suspended, we can safely set it to 0 here. 915 */ 916 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 917 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 918 919 Fixup: 920 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 921 922 /* 923 * If the target system sleep state is suspend-to-idle, it is sufficient 924 * to check whether or not the device's wakeup settings are good for 925 * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause 926 * pci_pm_complete() to take care of fixing up the device's state 927 * anyway, if need be. 928 */ 929 if (device_can_wakeup(dev) && !device_may_wakeup(dev)) 930 dev->power.may_skip_resume = false; 931 932 return 0; 933 } 934 935 static int pci_pm_resume_noirq(struct device *dev) 936 { 937 struct pci_dev *pci_dev = to_pci_dev(dev); 938 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 939 pci_power_t prev_state = pci_dev->current_state; 940 bool skip_bus_pm = pci_dev->skip_bus_pm; 941 942 if (dev_pm_skip_resume(dev)) 943 return 0; 944 945 /* 946 * In the suspend-to-idle case, devices left in D0 during suspend will 947 * stay in D0, so it is not necessary to restore or update their 948 * configuration here and attempting to put them into D0 again is 949 * pointless, so avoid doing that. 950 */ 951 if (!(skip_bus_pm && pm_suspend_no_platform())) 952 pci_pm_default_resume_early(pci_dev); 953 954 pci_fixup_device(pci_fixup_resume_early, pci_dev); 955 pcie_pme_root_status_cleanup(pci_dev); 956 957 if (!skip_bus_pm && prev_state == PCI_D3cold) 958 pci_pm_bridge_power_up_actions(pci_dev); 959 960 if (pci_has_legacy_pm_support(pci_dev)) 961 return 0; 962 963 if (pm && pm->resume_noirq) 964 return pm->resume_noirq(dev); 965 966 return 0; 967 } 968 969 static int pci_pm_resume_early(struct device *dev) 970 { 971 if (dev_pm_skip_resume(dev)) 972 return 0; 973 974 return pm_generic_resume_early(dev); 975 } 976 977 static int pci_pm_resume(struct device *dev) 978 { 979 struct pci_dev *pci_dev = to_pci_dev(dev); 980 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 981 982 /* 983 * This is necessary for the suspend error path in which resume is 984 * called without restoring the standard config registers of the device. 985 */ 986 if (pci_dev->state_saved) 987 pci_restore_standard_config(pci_dev); 988 989 if (pci_has_legacy_pm_support(pci_dev)) 990 return pci_legacy_resume(dev); 991 992 pci_pm_default_resume(pci_dev); 993 994 if (pm) { 995 if (pm->resume) 996 return pm->resume(dev); 997 } else { 998 pci_pm_reenable_device(pci_dev); 999 } 1000 1001 return 0; 1002 } 1003 1004 #else /* !CONFIG_SUSPEND */ 1005 1006 #define pci_pm_suspend NULL 1007 #define pci_pm_suspend_late NULL 1008 #define pci_pm_suspend_noirq NULL 1009 #define pci_pm_resume NULL 1010 #define pci_pm_resume_early NULL 1011 #define pci_pm_resume_noirq NULL 1012 1013 #endif /* !CONFIG_SUSPEND */ 1014 1015 #ifdef CONFIG_HIBERNATE_CALLBACKS 1016 1017 static int pci_pm_freeze(struct device *dev) 1018 { 1019 struct pci_dev *pci_dev = to_pci_dev(dev); 1020 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1021 1022 if (pci_has_legacy_pm_support(pci_dev)) 1023 return pci_legacy_suspend(dev, PMSG_FREEZE); 1024 1025 if (!pm) { 1026 pci_pm_default_suspend(pci_dev); 1027 return 0; 1028 } 1029 1030 /* 1031 * Resume all runtime-suspended devices before creating a snapshot 1032 * image of system memory, because the restore kernel generally cannot 1033 * be expected to always handle them consistently and they need to be 1034 * put into the runtime-active metastate during system resume anyway, 1035 * so it is better to ensure that the state saved in the image will be 1036 * always consistent with that. 1037 */ 1038 pm_runtime_resume(dev); 1039 pci_dev->state_saved = false; 1040 1041 if (pm->freeze) { 1042 int error; 1043 1044 error = pm->freeze(dev); 1045 suspend_report_result(dev, pm->freeze, error); 1046 if (error) 1047 return error; 1048 } 1049 1050 return 0; 1051 } 1052 1053 static int pci_pm_freeze_noirq(struct device *dev) 1054 { 1055 struct pci_dev *pci_dev = to_pci_dev(dev); 1056 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1057 1058 if (pci_has_legacy_pm_support(pci_dev)) 1059 return pci_legacy_suspend_late(dev, PMSG_FREEZE); 1060 1061 if (pm && pm->freeze_noirq) { 1062 int error; 1063 1064 error = pm->freeze_noirq(dev); 1065 suspend_report_result(dev, pm->freeze_noirq, error); 1066 if (error) 1067 return error; 1068 } 1069 1070 if (!pci_dev->state_saved) 1071 pci_save_state(pci_dev); 1072 1073 pci_pm_set_unknown_state(pci_dev); 1074 1075 return 0; 1076 } 1077 1078 static int pci_pm_thaw_noirq(struct device *dev) 1079 { 1080 struct pci_dev *pci_dev = to_pci_dev(dev); 1081 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1082 1083 /* 1084 * The pm->thaw_noirq() callback assumes the device has been 1085 * returned to D0 and its config state has been restored. 1086 * 1087 * In addition, pci_restore_state() restores MSI-X state in MMIO 1088 * space, which requires the device to be in D0, so return it to D0 1089 * in case the driver's "freeze" callbacks put it into a low-power 1090 * state. 1091 */ 1092 pci_pm_power_up_and_verify_state(pci_dev); 1093 pci_restore_state(pci_dev); 1094 1095 if (pci_has_legacy_pm_support(pci_dev)) 1096 return 0; 1097 1098 if (pm && pm->thaw_noirq) 1099 return pm->thaw_noirq(dev); 1100 1101 return 0; 1102 } 1103 1104 static int pci_pm_thaw(struct device *dev) 1105 { 1106 struct pci_dev *pci_dev = to_pci_dev(dev); 1107 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1108 int error = 0; 1109 1110 if (pci_has_legacy_pm_support(pci_dev)) 1111 return pci_legacy_resume(dev); 1112 1113 if (pm) { 1114 if (pm->thaw) 1115 error = pm->thaw(dev); 1116 } else { 1117 pci_pm_reenable_device(pci_dev); 1118 } 1119 1120 pci_dev->state_saved = false; 1121 1122 return error; 1123 } 1124 1125 static int pci_pm_poweroff(struct device *dev) 1126 { 1127 struct pci_dev *pci_dev = to_pci_dev(dev); 1128 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1129 1130 if (pci_has_legacy_pm_support(pci_dev)) 1131 return pci_legacy_suspend(dev, PMSG_HIBERNATE); 1132 1133 if (!pm) { 1134 pci_pm_default_suspend(pci_dev); 1135 return 0; 1136 } 1137 1138 /* The reason to do that is the same as in pci_pm_suspend(). */ 1139 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 1140 pci_dev_need_resume(pci_dev)) { 1141 pm_runtime_resume(dev); 1142 pci_dev->state_saved = false; 1143 } else { 1144 pci_dev_adjust_pme(pci_dev); 1145 } 1146 1147 if (pm->poweroff) { 1148 int error; 1149 1150 error = pm->poweroff(dev); 1151 suspend_report_result(dev, pm->poweroff, error); 1152 if (error) 1153 return error; 1154 } 1155 1156 return 0; 1157 } 1158 1159 static int pci_pm_poweroff_late(struct device *dev) 1160 { 1161 if (dev_pm_skip_suspend(dev)) 1162 return 0; 1163 1164 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 1165 1166 return pm_generic_poweroff_late(dev); 1167 } 1168 1169 static int pci_pm_poweroff_noirq(struct device *dev) 1170 { 1171 struct pci_dev *pci_dev = to_pci_dev(dev); 1172 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1173 1174 if (dev_pm_skip_suspend(dev)) 1175 return 0; 1176 1177 if (pci_has_legacy_pm_support(pci_dev)) 1178 return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); 1179 1180 if (!pm) { 1181 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1182 return 0; 1183 } 1184 1185 if (pm->poweroff_noirq) { 1186 int error; 1187 1188 error = pm->poweroff_noirq(dev); 1189 suspend_report_result(dev, pm->poweroff_noirq, error); 1190 if (error) 1191 return error; 1192 } 1193 1194 if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) 1195 pci_prepare_to_sleep(pci_dev); 1196 1197 /* 1198 * The reason for doing this here is the same as for the analogous code 1199 * in pci_pm_suspend_noirq(). 1200 */ 1201 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 1202 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 1203 1204 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1205 1206 return 0; 1207 } 1208 1209 static int pci_pm_restore_noirq(struct device *dev) 1210 { 1211 struct pci_dev *pci_dev = to_pci_dev(dev); 1212 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1213 1214 pci_pm_default_resume_early(pci_dev); 1215 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1216 1217 if (pci_has_legacy_pm_support(pci_dev)) 1218 return 0; 1219 1220 if (pm && pm->restore_noirq) 1221 return pm->restore_noirq(dev); 1222 1223 return 0; 1224 } 1225 1226 static int pci_pm_restore(struct device *dev) 1227 { 1228 struct pci_dev *pci_dev = to_pci_dev(dev); 1229 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1230 1231 /* 1232 * This is necessary for the hibernation error path in which restore is 1233 * called without restoring the standard config registers of the device. 1234 */ 1235 if (pci_dev->state_saved) 1236 pci_restore_standard_config(pci_dev); 1237 1238 if (pci_has_legacy_pm_support(pci_dev)) 1239 return pci_legacy_resume(dev); 1240 1241 pci_pm_default_resume(pci_dev); 1242 1243 if (pm) { 1244 if (pm->restore) 1245 return pm->restore(dev); 1246 } else { 1247 pci_pm_reenable_device(pci_dev); 1248 } 1249 1250 return 0; 1251 } 1252 1253 #else /* !CONFIG_HIBERNATE_CALLBACKS */ 1254 1255 #define pci_pm_freeze NULL 1256 #define pci_pm_freeze_noirq NULL 1257 #define pci_pm_thaw NULL 1258 #define pci_pm_thaw_noirq NULL 1259 #define pci_pm_poweroff NULL 1260 #define pci_pm_poweroff_late NULL 1261 #define pci_pm_poweroff_noirq NULL 1262 #define pci_pm_restore NULL 1263 #define pci_pm_restore_noirq NULL 1264 1265 #endif /* !CONFIG_HIBERNATE_CALLBACKS */ 1266 1267 #ifdef CONFIG_PM 1268 1269 static int pci_pm_runtime_suspend(struct device *dev) 1270 { 1271 struct pci_dev *pci_dev = to_pci_dev(dev); 1272 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1273 pci_power_t prev = pci_dev->current_state; 1274 int error; 1275 1276 /* 1277 * If pci_dev->driver is not set (unbound), we leave the device in D0, 1278 * but it may go to D3cold when the bridge above it runtime suspends. 1279 * Save its config space in case that happens. 1280 */ 1281 if (!pci_dev->driver) { 1282 pci_save_state(pci_dev); 1283 return 0; 1284 } 1285 1286 pci_dev->state_saved = false; 1287 if (pm && pm->runtime_suspend) { 1288 error = pm->runtime_suspend(dev); 1289 /* 1290 * -EBUSY and -EAGAIN is used to request the runtime PM core 1291 * to schedule a new suspend, so log the event only with debug 1292 * log level. 1293 */ 1294 if (error == -EBUSY || error == -EAGAIN) { 1295 pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", 1296 pm->runtime_suspend, error); 1297 return error; 1298 } else if (error) { 1299 pci_err(pci_dev, "can't suspend (%ps returned %d)\n", 1300 pm->runtime_suspend, error); 1301 return error; 1302 } 1303 } 1304 1305 pci_fixup_device(pci_fixup_suspend, pci_dev); 1306 1307 if (pm && pm->runtime_suspend 1308 && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 1309 && pci_dev->current_state != PCI_UNKNOWN) { 1310 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 1311 "PCI PM: State of device not saved by %pS\n", 1312 pm->runtime_suspend); 1313 return 0; 1314 } 1315 1316 if (!pci_dev->state_saved) { 1317 pci_save_state(pci_dev); 1318 pci_finish_runtime_suspend(pci_dev); 1319 } 1320 1321 return 0; 1322 } 1323 1324 static int pci_pm_runtime_resume(struct device *dev) 1325 { 1326 struct pci_dev *pci_dev = to_pci_dev(dev); 1327 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1328 pci_power_t prev_state = pci_dev->current_state; 1329 int error = 0; 1330 1331 /* 1332 * Restoring config space is necessary even if the device is not bound 1333 * to a driver because although we left it in D0, it may have gone to 1334 * D3cold when the bridge above it runtime suspended. 1335 */ 1336 pci_pm_default_resume_early(pci_dev); 1337 1338 if (!pci_dev->driver) 1339 return 0; 1340 1341 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1342 pci_pm_default_resume(pci_dev); 1343 1344 if (prev_state == PCI_D3cold) 1345 pci_pm_bridge_power_up_actions(pci_dev); 1346 1347 if (pm && pm->runtime_resume) 1348 error = pm->runtime_resume(dev); 1349 1350 return error; 1351 } 1352 1353 static int pci_pm_runtime_idle(struct device *dev) 1354 { 1355 struct pci_dev *pci_dev = to_pci_dev(dev); 1356 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1357 1358 /* 1359 * If pci_dev->driver is not set (unbound), the device should 1360 * always remain in D0 regardless of the runtime PM status 1361 */ 1362 if (!pci_dev->driver) 1363 return 0; 1364 1365 if (!pm) 1366 return -ENOSYS; 1367 1368 if (pm->runtime_idle) 1369 return pm->runtime_idle(dev); 1370 1371 return 0; 1372 } 1373 1374 static const struct dev_pm_ops pci_dev_pm_ops = { 1375 .prepare = pci_pm_prepare, 1376 .complete = pci_pm_complete, 1377 .suspend = pci_pm_suspend, 1378 .suspend_late = pci_pm_suspend_late, 1379 .resume = pci_pm_resume, 1380 .resume_early = pci_pm_resume_early, 1381 .freeze = pci_pm_freeze, 1382 .thaw = pci_pm_thaw, 1383 .poweroff = pci_pm_poweroff, 1384 .poweroff_late = pci_pm_poweroff_late, 1385 .restore = pci_pm_restore, 1386 .suspend_noirq = pci_pm_suspend_noirq, 1387 .resume_noirq = pci_pm_resume_noirq, 1388 .freeze_noirq = pci_pm_freeze_noirq, 1389 .thaw_noirq = pci_pm_thaw_noirq, 1390 .poweroff_noirq = pci_pm_poweroff_noirq, 1391 .restore_noirq = pci_pm_restore_noirq, 1392 .runtime_suspend = pci_pm_runtime_suspend, 1393 .runtime_resume = pci_pm_runtime_resume, 1394 .runtime_idle = pci_pm_runtime_idle, 1395 }; 1396 1397 #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) 1398 1399 #else /* !CONFIG_PM */ 1400 1401 #define pci_pm_runtime_suspend NULL 1402 #define pci_pm_runtime_resume NULL 1403 #define pci_pm_runtime_idle NULL 1404 1405 #define PCI_PM_OPS_PTR NULL 1406 1407 #endif /* !CONFIG_PM */ 1408 1409 /** 1410 * __pci_register_driver - register a new pci driver 1411 * @drv: the driver structure to register 1412 * @owner: owner module of drv 1413 * @mod_name: module name string 1414 * 1415 * Adds the driver structure to the list of registered drivers. 1416 * Returns a negative value on error, otherwise 0. 1417 * If no error occurred, the driver remains registered even if 1418 * no device was claimed during registration. 1419 */ 1420 int __pci_register_driver(struct pci_driver *drv, struct module *owner, 1421 const char *mod_name) 1422 { 1423 /* initialize common driver fields */ 1424 drv->driver.name = drv->name; 1425 drv->driver.bus = &pci_bus_type; 1426 drv->driver.owner = owner; 1427 drv->driver.mod_name = mod_name; 1428 drv->driver.groups = drv->groups; 1429 drv->driver.dev_groups = drv->dev_groups; 1430 1431 spin_lock_init(&drv->dynids.lock); 1432 INIT_LIST_HEAD(&drv->dynids.list); 1433 1434 /* register with core */ 1435 return driver_register(&drv->driver); 1436 } 1437 EXPORT_SYMBOL(__pci_register_driver); 1438 1439 /** 1440 * pci_unregister_driver - unregister a pci driver 1441 * @drv: the driver structure to unregister 1442 * 1443 * Deletes the driver structure from the list of registered PCI drivers, 1444 * gives it a chance to clean up by calling its remove() function for 1445 * each device it was responsible for, and marks those devices as 1446 * driverless. 1447 */ 1448 1449 void pci_unregister_driver(struct pci_driver *drv) 1450 { 1451 driver_unregister(&drv->driver); 1452 pci_free_dynids(drv); 1453 } 1454 EXPORT_SYMBOL(pci_unregister_driver); 1455 1456 static struct pci_driver pci_compat_driver = { 1457 .name = "compat" 1458 }; 1459 1460 /** 1461 * pci_dev_driver - get the pci_driver of a device 1462 * @dev: the device to query 1463 * 1464 * Returns the appropriate pci_driver structure or %NULL if there is no 1465 * registered driver for the device. 1466 */ 1467 struct pci_driver *pci_dev_driver(const struct pci_dev *dev) 1468 { 1469 if (dev->driver) 1470 return dev->driver; 1471 else { 1472 int i; 1473 for (i = 0; i <= PCI_ROM_RESOURCE; i++) 1474 if (dev->resource[i].flags & IORESOURCE_BUSY) 1475 return &pci_compat_driver; 1476 } 1477 return NULL; 1478 } 1479 EXPORT_SYMBOL(pci_dev_driver); 1480 1481 /** 1482 * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure 1483 * @dev: the PCI device structure to match against 1484 * @drv: the device driver to search for matching PCI device id structures 1485 * 1486 * Used by a driver to check whether a PCI device present in the 1487 * system is in its list of supported devices. Returns the matching 1488 * pci_device_id structure or %NULL if there is no match. 1489 */ 1490 static int pci_bus_match(struct device *dev, struct device_driver *drv) 1491 { 1492 struct pci_dev *pci_dev = to_pci_dev(dev); 1493 struct pci_driver *pci_drv; 1494 const struct pci_device_id *found_id; 1495 1496 if (!pci_dev->match_driver) 1497 return 0; 1498 1499 pci_drv = to_pci_driver(drv); 1500 found_id = pci_match_device(pci_drv, pci_dev); 1501 if (found_id) 1502 return 1; 1503 1504 return 0; 1505 } 1506 1507 /** 1508 * pci_dev_get - increments the reference count of the pci device structure 1509 * @dev: the device being referenced 1510 * 1511 * Each live reference to a device should be refcounted. 1512 * 1513 * Drivers for PCI devices should normally record such references in 1514 * their probe() methods, when they bind to a device, and release 1515 * them by calling pci_dev_put(), in their disconnect() methods. 1516 * 1517 * A pointer to the device with the incremented reference counter is returned. 1518 */ 1519 struct pci_dev *pci_dev_get(struct pci_dev *dev) 1520 { 1521 if (dev) 1522 get_device(&dev->dev); 1523 return dev; 1524 } 1525 EXPORT_SYMBOL(pci_dev_get); 1526 1527 /** 1528 * pci_dev_put - release a use of the pci device structure 1529 * @dev: device that's been disconnected 1530 * 1531 * Must be called when a user of a device is finished with it. When the last 1532 * user of the device calls this function, the memory of the device is freed. 1533 */ 1534 void pci_dev_put(struct pci_dev *dev) 1535 { 1536 if (dev) 1537 put_device(&dev->dev); 1538 } 1539 EXPORT_SYMBOL(pci_dev_put); 1540 1541 static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) 1542 { 1543 struct pci_dev *pdev; 1544 1545 if (!dev) 1546 return -ENODEV; 1547 1548 pdev = to_pci_dev(dev); 1549 1550 if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) 1551 return -ENOMEM; 1552 1553 if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) 1554 return -ENOMEM; 1555 1556 if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, 1557 pdev->subsystem_device)) 1558 return -ENOMEM; 1559 1560 if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) 1561 return -ENOMEM; 1562 1563 if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", 1564 pdev->vendor, pdev->device, 1565 pdev->subsystem_vendor, pdev->subsystem_device, 1566 (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), 1567 (u8)(pdev->class))) 1568 return -ENOMEM; 1569 1570 return 0; 1571 } 1572 1573 #if defined(CONFIG_PCIEAER) || defined(CONFIG_EEH) 1574 /** 1575 * pci_uevent_ers - emit a uevent during recovery path of PCI device 1576 * @pdev: PCI device undergoing error recovery 1577 * @err_type: type of error event 1578 */ 1579 void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) 1580 { 1581 int idx = 0; 1582 char *envp[3]; 1583 1584 switch (err_type) { 1585 case PCI_ERS_RESULT_NONE: 1586 case PCI_ERS_RESULT_CAN_RECOVER: 1587 envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; 1588 envp[idx++] = "DEVICE_ONLINE=0"; 1589 break; 1590 case PCI_ERS_RESULT_RECOVERED: 1591 envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; 1592 envp[idx++] = "DEVICE_ONLINE=1"; 1593 break; 1594 case PCI_ERS_RESULT_DISCONNECT: 1595 envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; 1596 envp[idx++] = "DEVICE_ONLINE=0"; 1597 break; 1598 default: 1599 break; 1600 } 1601 1602 if (idx > 0) { 1603 envp[idx++] = NULL; 1604 kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); 1605 } 1606 } 1607 #endif 1608 1609 static int pci_bus_num_vf(struct device *dev) 1610 { 1611 return pci_num_vf(to_pci_dev(dev)); 1612 } 1613 1614 /** 1615 * pci_dma_configure - Setup DMA configuration 1616 * @dev: ptr to dev structure 1617 * 1618 * Function to update PCI devices's DMA configuration using the same 1619 * info from the OF node or ACPI node of host bridge's parent (if any). 1620 */ 1621 static int pci_dma_configure(struct device *dev) 1622 { 1623 struct device *bridge; 1624 int ret = 0; 1625 1626 bridge = pci_get_host_bridge_device(to_pci_dev(dev)); 1627 1628 if (IS_ENABLED(CONFIG_OF) && bridge->parent && 1629 bridge->parent->of_node) { 1630 ret = of_dma_configure(dev, bridge->parent->of_node, true); 1631 } else if (has_acpi_companion(bridge)) { 1632 struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); 1633 1634 ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); 1635 } 1636 1637 pci_put_host_bridge_device(bridge); 1638 return ret; 1639 } 1640 1641 struct bus_type pci_bus_type = { 1642 .name = "pci", 1643 .match = pci_bus_match, 1644 .uevent = pci_uevent, 1645 .probe = pci_device_probe, 1646 .remove = pci_device_remove, 1647 .shutdown = pci_device_shutdown, 1648 .dev_groups = pci_dev_groups, 1649 .bus_groups = pci_bus_groups, 1650 .drv_groups = pci_drv_groups, 1651 .pm = PCI_PM_OPS_PTR, 1652 .num_vf = pci_bus_num_vf, 1653 .dma_configure = pci_dma_configure, 1654 }; 1655 EXPORT_SYMBOL(pci_bus_type); 1656 1657 #ifdef CONFIG_PCIEPORTBUS 1658 static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) 1659 { 1660 struct pcie_device *pciedev; 1661 struct pcie_port_service_driver *driver; 1662 1663 if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) 1664 return 0; 1665 1666 pciedev = to_pcie_device(dev); 1667 driver = to_service_driver(drv); 1668 1669 if (driver->service != pciedev->service) 1670 return 0; 1671 1672 if (driver->port_type != PCIE_ANY_PORT && 1673 driver->port_type != pci_pcie_type(pciedev->port)) 1674 return 0; 1675 1676 return 1; 1677 } 1678 1679 struct bus_type pcie_port_bus_type = { 1680 .name = "pci_express", 1681 .match = pcie_port_bus_match, 1682 }; 1683 EXPORT_SYMBOL_GPL(pcie_port_bus_type); 1684 #endif 1685 1686 static int __init pci_driver_init(void) 1687 { 1688 int ret; 1689 1690 ret = bus_register(&pci_bus_type); 1691 if (ret) 1692 return ret; 1693 1694 #ifdef CONFIG_PCIEPORTBUS 1695 ret = bus_register(&pcie_port_bus_type); 1696 if (ret) 1697 return ret; 1698 #endif 1699 dma_debug_add_bus(&pci_bus_type); 1700 return 0; 1701 } 1702 postcore_initcall(pci_driver_init); 1703