1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> 4 * (C) Copyright 2007 Novell Inc. 5 */ 6 7 #include <linux/pci.h> 8 #include <linux/module.h> 9 #include <linux/init.h> 10 #include <linux/device.h> 11 #include <linux/mempolicy.h> 12 #include <linux/string.h> 13 #include <linux/slab.h> 14 #include <linux/sched.h> 15 #include <linux/sched/isolation.h> 16 #include <linux/cpu.h> 17 #include <linux/pm_runtime.h> 18 #include <linux/suspend.h> 19 #include <linux/kexec.h> 20 #include <linux/of_device.h> 21 #include <linux/acpi.h> 22 #include <linux/dma-map-ops.h> 23 #include "pci.h" 24 #include "pcie/portdrv.h" 25 26 struct pci_dynid { 27 struct list_head node; 28 struct pci_device_id id; 29 }; 30 31 /** 32 * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices 33 * @drv: target pci driver 34 * @vendor: PCI vendor ID 35 * @device: PCI device ID 36 * @subvendor: PCI subvendor ID 37 * @subdevice: PCI subdevice ID 38 * @class: PCI class 39 * @class_mask: PCI class mask 40 * @driver_data: private driver data 41 * 42 * Adds a new dynamic pci device ID to this driver and causes the 43 * driver to probe for all devices again. @drv must have been 44 * registered prior to calling this function. 45 * 46 * CONTEXT: 47 * Does GFP_KERNEL allocation. 48 * 49 * RETURNS: 50 * 0 on success, -errno on failure. 51 */ 52 int pci_add_dynid(struct pci_driver *drv, 53 unsigned int vendor, unsigned int device, 54 unsigned int subvendor, unsigned int subdevice, 55 unsigned int class, unsigned int class_mask, 56 unsigned long driver_data) 57 { 58 struct pci_dynid *dynid; 59 60 dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); 61 if (!dynid) 62 return -ENOMEM; 63 64 dynid->id.vendor = vendor; 65 dynid->id.device = device; 66 dynid->id.subvendor = subvendor; 67 dynid->id.subdevice = subdevice; 68 dynid->id.class = class; 69 dynid->id.class_mask = class_mask; 70 dynid->id.driver_data = driver_data; 71 72 spin_lock(&drv->dynids.lock); 73 list_add_tail(&dynid->node, &drv->dynids.list); 74 spin_unlock(&drv->dynids.lock); 75 76 return driver_attach(&drv->driver); 77 } 78 EXPORT_SYMBOL_GPL(pci_add_dynid); 79 80 static void pci_free_dynids(struct pci_driver *drv) 81 { 82 struct pci_dynid *dynid, *n; 83 84 spin_lock(&drv->dynids.lock); 85 list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { 86 list_del(&dynid->node); 87 kfree(dynid); 88 } 89 spin_unlock(&drv->dynids.lock); 90 } 91 92 /** 93 * pci_match_id - See if a PCI device matches a given pci_id table 94 * @ids: array of PCI device ID structures to search in 95 * @dev: the PCI device structure to match against. 96 * 97 * Used by a driver to check whether a PCI device is in its list of 98 * supported devices. Returns the matching pci_device_id structure or 99 * %NULL if there is no match. 100 * 101 * Deprecated; don't use this as it will not catch any dynamic IDs 102 * that a driver might want to check for. 103 */ 104 const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, 105 struct pci_dev *dev) 106 { 107 if (ids) { 108 while (ids->vendor || ids->subvendor || ids->class_mask) { 109 if (pci_match_one_device(ids, dev)) 110 return ids; 111 ids++; 112 } 113 } 114 return NULL; 115 } 116 EXPORT_SYMBOL(pci_match_id); 117 118 static const struct pci_device_id pci_device_id_any = { 119 .vendor = PCI_ANY_ID, 120 .device = PCI_ANY_ID, 121 .subvendor = PCI_ANY_ID, 122 .subdevice = PCI_ANY_ID, 123 }; 124 125 /** 126 * pci_match_device - See if a device matches a driver's list of IDs 127 * @drv: the PCI driver to match against 128 * @dev: the PCI device structure to match against 129 * 130 * Used by a driver to check whether a PCI device is in its list of 131 * supported devices or in the dynids list, which may have been augmented 132 * via the sysfs "new_id" file. Returns the matching pci_device_id 133 * structure or %NULL if there is no match. 134 */ 135 static const struct pci_device_id *pci_match_device(struct pci_driver *drv, 136 struct pci_dev *dev) 137 { 138 struct pci_dynid *dynid; 139 const struct pci_device_id *found_id = NULL; 140 141 /* When driver_override is set, only bind to the matching driver */ 142 if (dev->driver_override && strcmp(dev->driver_override, drv->name)) 143 return NULL; 144 145 /* Look at the dynamic ids first, before the static ones */ 146 spin_lock(&drv->dynids.lock); 147 list_for_each_entry(dynid, &drv->dynids.list, node) { 148 if (pci_match_one_device(&dynid->id, dev)) { 149 found_id = &dynid->id; 150 break; 151 } 152 } 153 spin_unlock(&drv->dynids.lock); 154 155 if (!found_id) 156 found_id = pci_match_id(drv->id_table, dev); 157 158 /* driver_override will always match, send a dummy id */ 159 if (!found_id && dev->driver_override) 160 found_id = &pci_device_id_any; 161 162 return found_id; 163 } 164 165 /** 166 * new_id_store - sysfs frontend to pci_add_dynid() 167 * @driver: target device driver 168 * @buf: buffer for scanning device ID data 169 * @count: input size 170 * 171 * Allow PCI IDs to be added to an existing driver via sysfs. 172 */ 173 static ssize_t new_id_store(struct device_driver *driver, const char *buf, 174 size_t count) 175 { 176 struct pci_driver *pdrv = to_pci_driver(driver); 177 const struct pci_device_id *ids = pdrv->id_table; 178 u32 vendor, device, subvendor = PCI_ANY_ID, 179 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 180 unsigned long driver_data = 0; 181 int fields = 0; 182 int retval = 0; 183 184 fields = sscanf(buf, "%x %x %x %x %x %x %lx", 185 &vendor, &device, &subvendor, &subdevice, 186 &class, &class_mask, &driver_data); 187 if (fields < 2) 188 return -EINVAL; 189 190 if (fields != 7) { 191 struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); 192 if (!pdev) 193 return -ENOMEM; 194 195 pdev->vendor = vendor; 196 pdev->device = device; 197 pdev->subsystem_vendor = subvendor; 198 pdev->subsystem_device = subdevice; 199 pdev->class = class; 200 201 if (pci_match_device(pdrv, pdev)) 202 retval = -EEXIST; 203 204 kfree(pdev); 205 206 if (retval) 207 return retval; 208 } 209 210 /* Only accept driver_data values that match an existing id_table 211 entry */ 212 if (ids) { 213 retval = -EINVAL; 214 while (ids->vendor || ids->subvendor || ids->class_mask) { 215 if (driver_data == ids->driver_data) { 216 retval = 0; 217 break; 218 } 219 ids++; 220 } 221 if (retval) /* No match */ 222 return retval; 223 } 224 225 retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, 226 class, class_mask, driver_data); 227 if (retval) 228 return retval; 229 return count; 230 } 231 static DRIVER_ATTR_WO(new_id); 232 233 /** 234 * remove_id_store - remove a PCI device ID from this driver 235 * @driver: target device driver 236 * @buf: buffer for scanning device ID data 237 * @count: input size 238 * 239 * Removes a dynamic pci device ID to this driver. 240 */ 241 static ssize_t remove_id_store(struct device_driver *driver, const char *buf, 242 size_t count) 243 { 244 struct pci_dynid *dynid, *n; 245 struct pci_driver *pdrv = to_pci_driver(driver); 246 u32 vendor, device, subvendor = PCI_ANY_ID, 247 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 248 int fields = 0; 249 size_t retval = -ENODEV; 250 251 fields = sscanf(buf, "%x %x %x %x %x %x", 252 &vendor, &device, &subvendor, &subdevice, 253 &class, &class_mask); 254 if (fields < 2) 255 return -EINVAL; 256 257 spin_lock(&pdrv->dynids.lock); 258 list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { 259 struct pci_device_id *id = &dynid->id; 260 if ((id->vendor == vendor) && 261 (id->device == device) && 262 (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && 263 (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && 264 !((id->class ^ class) & class_mask)) { 265 list_del(&dynid->node); 266 kfree(dynid); 267 retval = count; 268 break; 269 } 270 } 271 spin_unlock(&pdrv->dynids.lock); 272 273 return retval; 274 } 275 static DRIVER_ATTR_WO(remove_id); 276 277 static struct attribute *pci_drv_attrs[] = { 278 &driver_attr_new_id.attr, 279 &driver_attr_remove_id.attr, 280 NULL, 281 }; 282 ATTRIBUTE_GROUPS(pci_drv); 283 284 struct drv_dev_and_id { 285 struct pci_driver *drv; 286 struct pci_dev *dev; 287 const struct pci_device_id *id; 288 }; 289 290 static long local_pci_probe(void *_ddi) 291 { 292 struct drv_dev_and_id *ddi = _ddi; 293 struct pci_dev *pci_dev = ddi->dev; 294 struct pci_driver *pci_drv = ddi->drv; 295 struct device *dev = &pci_dev->dev; 296 int rc; 297 298 /* 299 * Unbound PCI devices are always put in D0, regardless of 300 * runtime PM status. During probe, the device is set to 301 * active and the usage count is incremented. If the driver 302 * supports runtime PM, it should call pm_runtime_put_noidle(), 303 * or any other runtime PM helper function decrementing the usage 304 * count, in its probe routine and pm_runtime_get_noresume() in 305 * its remove routine. 306 */ 307 pm_runtime_get_sync(dev); 308 pci_dev->driver = pci_drv; 309 rc = pci_drv->probe(pci_dev, ddi->id); 310 if (!rc) 311 return rc; 312 if (rc < 0) { 313 pci_dev->driver = NULL; 314 pm_runtime_put_sync(dev); 315 return rc; 316 } 317 /* 318 * Probe function should return < 0 for failure, 0 for success 319 * Treat values > 0 as success, but warn. 320 */ 321 pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", 322 rc); 323 return 0; 324 } 325 326 static bool pci_physfn_is_probed(struct pci_dev *dev) 327 { 328 #ifdef CONFIG_PCI_IOV 329 return dev->is_virtfn && dev->physfn->is_probed; 330 #else 331 return false; 332 #endif 333 } 334 335 static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, 336 const struct pci_device_id *id) 337 { 338 int error, node, cpu; 339 int hk_flags = HK_FLAG_DOMAIN | HK_FLAG_WQ; 340 struct drv_dev_and_id ddi = { drv, dev, id }; 341 342 /* 343 * Execute driver initialization on node where the device is 344 * attached. This way the driver likely allocates its local memory 345 * on the right node. 346 */ 347 node = dev_to_node(&dev->dev); 348 dev->is_probed = 1; 349 350 cpu_hotplug_disable(); 351 352 /* 353 * Prevent nesting work_on_cpu() for the case where a Virtual Function 354 * device is probed from work_on_cpu() of the Physical device. 355 */ 356 if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || 357 pci_physfn_is_probed(dev)) 358 cpu = nr_cpu_ids; 359 else 360 cpu = cpumask_any_and(cpumask_of_node(node), 361 housekeeping_cpumask(hk_flags)); 362 363 if (cpu < nr_cpu_ids) 364 error = work_on_cpu(cpu, local_pci_probe, &ddi); 365 else 366 error = local_pci_probe(&ddi); 367 368 dev->is_probed = 0; 369 cpu_hotplug_enable(); 370 return error; 371 } 372 373 /** 374 * __pci_device_probe - check if a driver wants to claim a specific PCI device 375 * @drv: driver to call to check if it wants the PCI device 376 * @pci_dev: PCI device being probed 377 * 378 * returns 0 on success, else error. 379 * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. 380 */ 381 static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) 382 { 383 const struct pci_device_id *id; 384 int error = 0; 385 386 if (!pci_dev->driver && drv->probe) { 387 error = -ENODEV; 388 389 id = pci_match_device(drv, pci_dev); 390 if (id) 391 error = pci_call_probe(drv, pci_dev, id); 392 } 393 return error; 394 } 395 396 int __weak pcibios_alloc_irq(struct pci_dev *dev) 397 { 398 return 0; 399 } 400 401 void __weak pcibios_free_irq(struct pci_dev *dev) 402 { 403 } 404 405 #ifdef CONFIG_PCI_IOV 406 static inline bool pci_device_can_probe(struct pci_dev *pdev) 407 { 408 return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || 409 pdev->driver_override); 410 } 411 #else 412 static inline bool pci_device_can_probe(struct pci_dev *pdev) 413 { 414 return true; 415 } 416 #endif 417 418 static int pci_device_probe(struct device *dev) 419 { 420 int error; 421 struct pci_dev *pci_dev = to_pci_dev(dev); 422 struct pci_driver *drv = to_pci_driver(dev->driver); 423 424 if (!pci_device_can_probe(pci_dev)) 425 return -ENODEV; 426 427 pci_assign_irq(pci_dev); 428 429 error = pcibios_alloc_irq(pci_dev); 430 if (error < 0) 431 return error; 432 433 pci_dev_get(pci_dev); 434 error = __pci_device_probe(drv, pci_dev); 435 if (error) { 436 pcibios_free_irq(pci_dev); 437 pci_dev_put(pci_dev); 438 } 439 440 return error; 441 } 442 443 static void pci_device_remove(struct device *dev) 444 { 445 struct pci_dev *pci_dev = to_pci_dev(dev); 446 struct pci_driver *drv = pci_dev->driver; 447 448 if (drv) { 449 if (drv->remove) { 450 pm_runtime_get_sync(dev); 451 drv->remove(pci_dev); 452 pm_runtime_put_noidle(dev); 453 } 454 pcibios_free_irq(pci_dev); 455 pci_dev->driver = NULL; 456 pci_iov_remove(pci_dev); 457 } 458 459 /* Undo the runtime PM settings in local_pci_probe() */ 460 pm_runtime_put_sync(dev); 461 462 /* 463 * If the device is still on, set the power state as "unknown", 464 * since it might change by the next time we load the driver. 465 */ 466 if (pci_dev->current_state == PCI_D0) 467 pci_dev->current_state = PCI_UNKNOWN; 468 469 /* 470 * We would love to complain here if pci_dev->is_enabled is set, that 471 * the driver should have called pci_disable_device(), but the 472 * unfortunate fact is there are too many odd BIOS and bridge setups 473 * that don't like drivers doing that all of the time. 474 * Oh well, we can dream of sane hardware when we sleep, no matter how 475 * horrible the crap we have to deal with is when we are awake... 476 */ 477 478 pci_dev_put(pci_dev); 479 } 480 481 static void pci_device_shutdown(struct device *dev) 482 { 483 struct pci_dev *pci_dev = to_pci_dev(dev); 484 struct pci_driver *drv = pci_dev->driver; 485 486 pm_runtime_resume(dev); 487 488 if (drv && drv->shutdown) 489 drv->shutdown(pci_dev); 490 491 /* 492 * If this is a kexec reboot, turn off Bus Master bit on the 493 * device to tell it to not continue to do DMA. Don't touch 494 * devices in D3cold or unknown states. 495 * If it is not a kexec reboot, firmware will hit the PCI 496 * devices with big hammer and stop their DMA any way. 497 */ 498 if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) 499 pci_clear_master(pci_dev); 500 } 501 502 #ifdef CONFIG_PM 503 504 /* Auxiliary functions used for system resume and run-time resume. */ 505 506 /** 507 * pci_restore_standard_config - restore standard config registers of PCI device 508 * @pci_dev: PCI device to handle 509 */ 510 static int pci_restore_standard_config(struct pci_dev *pci_dev) 511 { 512 pci_update_current_state(pci_dev, PCI_UNKNOWN); 513 514 if (pci_dev->current_state != PCI_D0) { 515 int error = pci_set_power_state(pci_dev, PCI_D0); 516 if (error) 517 return error; 518 } 519 520 pci_restore_state(pci_dev); 521 pci_pme_restore(pci_dev); 522 return 0; 523 } 524 525 static void pci_pm_default_resume(struct pci_dev *pci_dev) 526 { 527 pci_fixup_device(pci_fixup_resume, pci_dev); 528 pci_enable_wake(pci_dev, PCI_D0, false); 529 } 530 531 #endif 532 533 #ifdef CONFIG_PM_SLEEP 534 535 static void pci_pm_default_resume_early(struct pci_dev *pci_dev) 536 { 537 pci_power_up(pci_dev); 538 pci_update_current_state(pci_dev, PCI_D0); 539 pci_restore_state(pci_dev); 540 pci_pme_restore(pci_dev); 541 } 542 543 /* 544 * Default "suspend" method for devices that have no driver provided suspend, 545 * or not even a driver at all (second part). 546 */ 547 static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) 548 { 549 /* 550 * mark its power state as "unknown", since we don't know if 551 * e.g. the BIOS will change its device state when we suspend. 552 */ 553 if (pci_dev->current_state == PCI_D0) 554 pci_dev->current_state = PCI_UNKNOWN; 555 } 556 557 /* 558 * Default "resume" method for devices that have no driver provided resume, 559 * or not even a driver at all (second part). 560 */ 561 static int pci_pm_reenable_device(struct pci_dev *pci_dev) 562 { 563 int retval; 564 565 /* if the device was enabled before suspend, reenable */ 566 retval = pci_reenable_device(pci_dev); 567 /* 568 * if the device was busmaster before the suspend, make it busmaster 569 * again 570 */ 571 if (pci_dev->is_busmaster) 572 pci_set_master(pci_dev); 573 574 return retval; 575 } 576 577 static int pci_legacy_suspend(struct device *dev, pm_message_t state) 578 { 579 struct pci_dev *pci_dev = to_pci_dev(dev); 580 struct pci_driver *drv = pci_dev->driver; 581 582 if (drv && drv->suspend) { 583 pci_power_t prev = pci_dev->current_state; 584 int error; 585 586 error = drv->suspend(pci_dev, state); 587 suspend_report_result(drv->suspend, error); 588 if (error) 589 return error; 590 591 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 592 && pci_dev->current_state != PCI_UNKNOWN) { 593 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 594 "PCI PM: Device state not saved by %pS\n", 595 drv->suspend); 596 } 597 } 598 599 pci_fixup_device(pci_fixup_suspend, pci_dev); 600 601 return 0; 602 } 603 604 static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) 605 { 606 struct pci_dev *pci_dev = to_pci_dev(dev); 607 608 if (!pci_dev->state_saved) 609 pci_save_state(pci_dev); 610 611 pci_pm_set_unknown_state(pci_dev); 612 613 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 614 615 return 0; 616 } 617 618 static int pci_legacy_resume(struct device *dev) 619 { 620 struct pci_dev *pci_dev = to_pci_dev(dev); 621 struct pci_driver *drv = pci_dev->driver; 622 623 pci_fixup_device(pci_fixup_resume, pci_dev); 624 625 return drv && drv->resume ? 626 drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); 627 } 628 629 /* Auxiliary functions used by the new power management framework */ 630 631 static void pci_pm_default_suspend(struct pci_dev *pci_dev) 632 { 633 /* Disable non-bridge devices without PM support */ 634 if (!pci_has_subordinate(pci_dev)) 635 pci_disable_enabled_device(pci_dev); 636 } 637 638 static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) 639 { 640 struct pci_driver *drv = pci_dev->driver; 641 bool ret = drv && (drv->suspend || drv->resume); 642 643 /* 644 * Legacy PM support is used by default, so warn if the new framework is 645 * supported as well. Drivers are supposed to support either the 646 * former, or the latter, but not both at the same time. 647 */ 648 pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", 649 pci_dev->vendor, pci_dev->device); 650 651 return ret; 652 } 653 654 /* New power management framework */ 655 656 static int pci_pm_prepare(struct device *dev) 657 { 658 struct pci_dev *pci_dev = to_pci_dev(dev); 659 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 660 661 if (pm && pm->prepare) { 662 int error = pm->prepare(dev); 663 if (error < 0) 664 return error; 665 666 if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) 667 return 0; 668 } 669 if (pci_dev_need_resume(pci_dev)) 670 return 0; 671 672 /* 673 * The PME setting needs to be adjusted here in case the direct-complete 674 * optimization is used with respect to this device. 675 */ 676 pci_dev_adjust_pme(pci_dev); 677 return 1; 678 } 679 680 static void pci_pm_complete(struct device *dev) 681 { 682 struct pci_dev *pci_dev = to_pci_dev(dev); 683 684 pci_dev_complete_resume(pci_dev); 685 pm_generic_complete(dev); 686 687 /* Resume device if platform firmware has put it in reset-power-on */ 688 if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { 689 pci_power_t pre_sleep_state = pci_dev->current_state; 690 691 pci_refresh_power_state(pci_dev); 692 /* 693 * On platforms with ACPI this check may also trigger for 694 * devices sharing power resources if one of those power 695 * resources has been activated as a result of a change of the 696 * power state of another device sharing it. However, in that 697 * case it is also better to resume the device, in general. 698 */ 699 if (pci_dev->current_state < pre_sleep_state) 700 pm_request_resume(dev); 701 } 702 } 703 704 #else /* !CONFIG_PM_SLEEP */ 705 706 #define pci_pm_prepare NULL 707 #define pci_pm_complete NULL 708 709 #endif /* !CONFIG_PM_SLEEP */ 710 711 #ifdef CONFIG_SUSPEND 712 static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) 713 { 714 /* 715 * Some BIOSes forget to clear Root PME Status bits after system 716 * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. 717 * Clear those bits now just in case (shouldn't hurt). 718 */ 719 if (pci_is_pcie(pci_dev) && 720 (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || 721 pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) 722 pcie_clear_root_pme_status(pci_dev); 723 } 724 725 static int pci_pm_suspend(struct device *dev) 726 { 727 struct pci_dev *pci_dev = to_pci_dev(dev); 728 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 729 730 pci_dev->skip_bus_pm = false; 731 732 if (pci_has_legacy_pm_support(pci_dev)) 733 return pci_legacy_suspend(dev, PMSG_SUSPEND); 734 735 if (!pm) { 736 pci_pm_default_suspend(pci_dev); 737 return 0; 738 } 739 740 /* 741 * PCI devices suspended at run time may need to be resumed at this 742 * point, because in general it may be necessary to reconfigure them for 743 * system suspend. Namely, if the device is expected to wake up the 744 * system from the sleep state, it may have to be reconfigured for this 745 * purpose, or if the device is not expected to wake up the system from 746 * the sleep state, it should be prevented from signaling wakeup events 747 * going forward. 748 * 749 * Also if the driver of the device does not indicate that its system 750 * suspend callbacks can cope with runtime-suspended devices, it is 751 * better to resume the device from runtime suspend here. 752 */ 753 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 754 pci_dev_need_resume(pci_dev)) { 755 pm_runtime_resume(dev); 756 pci_dev->state_saved = false; 757 } else { 758 pci_dev_adjust_pme(pci_dev); 759 } 760 761 if (pm->suspend) { 762 pci_power_t prev = pci_dev->current_state; 763 int error; 764 765 error = pm->suspend(dev); 766 suspend_report_result(pm->suspend, error); 767 if (error) 768 return error; 769 770 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 771 && pci_dev->current_state != PCI_UNKNOWN) { 772 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 773 "PCI PM: State of device not saved by %pS\n", 774 pm->suspend); 775 } 776 } 777 778 return 0; 779 } 780 781 static int pci_pm_suspend_late(struct device *dev) 782 { 783 if (dev_pm_skip_suspend(dev)) 784 return 0; 785 786 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 787 788 return pm_generic_suspend_late(dev); 789 } 790 791 static int pci_pm_suspend_noirq(struct device *dev) 792 { 793 struct pci_dev *pci_dev = to_pci_dev(dev); 794 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 795 796 if (dev_pm_skip_suspend(dev)) 797 return 0; 798 799 if (pci_has_legacy_pm_support(pci_dev)) 800 return pci_legacy_suspend_late(dev, PMSG_SUSPEND); 801 802 if (!pm) { 803 pci_save_state(pci_dev); 804 goto Fixup; 805 } 806 807 if (pm->suspend_noirq) { 808 pci_power_t prev = pci_dev->current_state; 809 int error; 810 811 error = pm->suspend_noirq(dev); 812 suspend_report_result(pm->suspend_noirq, error); 813 if (error) 814 return error; 815 816 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 817 && pci_dev->current_state != PCI_UNKNOWN) { 818 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 819 "PCI PM: State of device not saved by %pS\n", 820 pm->suspend_noirq); 821 goto Fixup; 822 } 823 } 824 825 if (pci_dev->skip_bus_pm) { 826 /* 827 * Either the device is a bridge with a child in D0 below it, or 828 * the function is running for the second time in a row without 829 * going through full resume, which is possible only during 830 * suspend-to-idle in a spurious wakeup case. The device should 831 * be in D0 at this point, but if it is a bridge, it may be 832 * necessary to save its state. 833 */ 834 if (!pci_dev->state_saved) 835 pci_save_state(pci_dev); 836 } else if (!pci_dev->state_saved) { 837 pci_save_state(pci_dev); 838 if (pci_power_manageable(pci_dev)) 839 pci_prepare_to_sleep(pci_dev); 840 } 841 842 pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", 843 pci_power_name(pci_dev->current_state)); 844 845 if (pci_dev->current_state == PCI_D0) { 846 pci_dev->skip_bus_pm = true; 847 /* 848 * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any 849 * downstream device is in D0, so avoid changing the power state 850 * of the parent bridge by setting the skip_bus_pm flag for it. 851 */ 852 if (pci_dev->bus->self) 853 pci_dev->bus->self->skip_bus_pm = true; 854 } 855 856 if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { 857 pci_dbg(pci_dev, "PCI PM: Skipped\n"); 858 goto Fixup; 859 } 860 861 pci_pm_set_unknown_state(pci_dev); 862 863 /* 864 * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's 865 * PCI COMMAND register isn't 0, the BIOS assumes that the controller 866 * hasn't been quiesced and tries to turn it off. If the controller 867 * is already in D3, this can hang or cause memory corruption. 868 * 869 * Since the value of the COMMAND register doesn't matter once the 870 * device has been suspended, we can safely set it to 0 here. 871 */ 872 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 873 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 874 875 Fixup: 876 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 877 878 /* 879 * If the target system sleep state is suspend-to-idle, it is sufficient 880 * to check whether or not the device's wakeup settings are good for 881 * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause 882 * pci_pm_complete() to take care of fixing up the device's state 883 * anyway, if need be. 884 */ 885 if (device_can_wakeup(dev) && !device_may_wakeup(dev)) 886 dev->power.may_skip_resume = false; 887 888 return 0; 889 } 890 891 static int pci_pm_resume_noirq(struct device *dev) 892 { 893 struct pci_dev *pci_dev = to_pci_dev(dev); 894 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 895 pci_power_t prev_state = pci_dev->current_state; 896 bool skip_bus_pm = pci_dev->skip_bus_pm; 897 898 if (dev_pm_skip_resume(dev)) 899 return 0; 900 901 /* 902 * In the suspend-to-idle case, devices left in D0 during suspend will 903 * stay in D0, so it is not necessary to restore or update their 904 * configuration here and attempting to put them into D0 again is 905 * pointless, so avoid doing that. 906 */ 907 if (!(skip_bus_pm && pm_suspend_no_platform())) 908 pci_pm_default_resume_early(pci_dev); 909 910 pci_fixup_device(pci_fixup_resume_early, pci_dev); 911 pcie_pme_root_status_cleanup(pci_dev); 912 913 if (!skip_bus_pm && prev_state == PCI_D3cold) 914 pci_bridge_wait_for_secondary_bus(pci_dev); 915 916 if (pci_has_legacy_pm_support(pci_dev)) 917 return 0; 918 919 if (pm && pm->resume_noirq) 920 return pm->resume_noirq(dev); 921 922 return 0; 923 } 924 925 static int pci_pm_resume_early(struct device *dev) 926 { 927 if (dev_pm_skip_resume(dev)) 928 return 0; 929 930 return pm_generic_resume_early(dev); 931 } 932 933 static int pci_pm_resume(struct device *dev) 934 { 935 struct pci_dev *pci_dev = to_pci_dev(dev); 936 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 937 938 /* 939 * This is necessary for the suspend error path in which resume is 940 * called without restoring the standard config registers of the device. 941 */ 942 if (pci_dev->state_saved) 943 pci_restore_standard_config(pci_dev); 944 945 if (pci_has_legacy_pm_support(pci_dev)) 946 return pci_legacy_resume(dev); 947 948 pci_pm_default_resume(pci_dev); 949 950 if (pm) { 951 if (pm->resume) 952 return pm->resume(dev); 953 } else { 954 pci_pm_reenable_device(pci_dev); 955 } 956 957 return 0; 958 } 959 960 #else /* !CONFIG_SUSPEND */ 961 962 #define pci_pm_suspend NULL 963 #define pci_pm_suspend_late NULL 964 #define pci_pm_suspend_noirq NULL 965 #define pci_pm_resume NULL 966 #define pci_pm_resume_early NULL 967 #define pci_pm_resume_noirq NULL 968 969 #endif /* !CONFIG_SUSPEND */ 970 971 #ifdef CONFIG_HIBERNATE_CALLBACKS 972 973 static int pci_pm_freeze(struct device *dev) 974 { 975 struct pci_dev *pci_dev = to_pci_dev(dev); 976 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 977 978 if (pci_has_legacy_pm_support(pci_dev)) 979 return pci_legacy_suspend(dev, PMSG_FREEZE); 980 981 if (!pm) { 982 pci_pm_default_suspend(pci_dev); 983 return 0; 984 } 985 986 /* 987 * Resume all runtime-suspended devices before creating a snapshot 988 * image of system memory, because the restore kernel generally cannot 989 * be expected to always handle them consistently and they need to be 990 * put into the runtime-active metastate during system resume anyway, 991 * so it is better to ensure that the state saved in the image will be 992 * always consistent with that. 993 */ 994 pm_runtime_resume(dev); 995 pci_dev->state_saved = false; 996 997 if (pm->freeze) { 998 int error; 999 1000 error = pm->freeze(dev); 1001 suspend_report_result(pm->freeze, error); 1002 if (error) 1003 return error; 1004 } 1005 1006 return 0; 1007 } 1008 1009 static int pci_pm_freeze_noirq(struct device *dev) 1010 { 1011 struct pci_dev *pci_dev = to_pci_dev(dev); 1012 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1013 1014 if (pci_has_legacy_pm_support(pci_dev)) 1015 return pci_legacy_suspend_late(dev, PMSG_FREEZE); 1016 1017 if (pm && pm->freeze_noirq) { 1018 int error; 1019 1020 error = pm->freeze_noirq(dev); 1021 suspend_report_result(pm->freeze_noirq, error); 1022 if (error) 1023 return error; 1024 } 1025 1026 if (!pci_dev->state_saved) 1027 pci_save_state(pci_dev); 1028 1029 pci_pm_set_unknown_state(pci_dev); 1030 1031 return 0; 1032 } 1033 1034 static int pci_pm_thaw_noirq(struct device *dev) 1035 { 1036 struct pci_dev *pci_dev = to_pci_dev(dev); 1037 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1038 1039 /* 1040 * The pm->thaw_noirq() callback assumes the device has been 1041 * returned to D0 and its config state has been restored. 1042 * 1043 * In addition, pci_restore_state() restores MSI-X state in MMIO 1044 * space, which requires the device to be in D0, so return it to D0 1045 * in case the driver's "freeze" callbacks put it into a low-power 1046 * state. 1047 */ 1048 pci_set_power_state(pci_dev, PCI_D0); 1049 pci_restore_state(pci_dev); 1050 1051 if (pci_has_legacy_pm_support(pci_dev)) 1052 return 0; 1053 1054 if (pm && pm->thaw_noirq) 1055 return pm->thaw_noirq(dev); 1056 1057 return 0; 1058 } 1059 1060 static int pci_pm_thaw(struct device *dev) 1061 { 1062 struct pci_dev *pci_dev = to_pci_dev(dev); 1063 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1064 int error = 0; 1065 1066 if (pci_has_legacy_pm_support(pci_dev)) 1067 return pci_legacy_resume(dev); 1068 1069 if (pm) { 1070 if (pm->thaw) 1071 error = pm->thaw(dev); 1072 } else { 1073 pci_pm_reenable_device(pci_dev); 1074 } 1075 1076 pci_dev->state_saved = false; 1077 1078 return error; 1079 } 1080 1081 static int pci_pm_poweroff(struct device *dev) 1082 { 1083 struct pci_dev *pci_dev = to_pci_dev(dev); 1084 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1085 1086 if (pci_has_legacy_pm_support(pci_dev)) 1087 return pci_legacy_suspend(dev, PMSG_HIBERNATE); 1088 1089 if (!pm) { 1090 pci_pm_default_suspend(pci_dev); 1091 return 0; 1092 } 1093 1094 /* The reason to do that is the same as in pci_pm_suspend(). */ 1095 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 1096 pci_dev_need_resume(pci_dev)) { 1097 pm_runtime_resume(dev); 1098 pci_dev->state_saved = false; 1099 } else { 1100 pci_dev_adjust_pme(pci_dev); 1101 } 1102 1103 if (pm->poweroff) { 1104 int error; 1105 1106 error = pm->poweroff(dev); 1107 suspend_report_result(pm->poweroff, error); 1108 if (error) 1109 return error; 1110 } 1111 1112 return 0; 1113 } 1114 1115 static int pci_pm_poweroff_late(struct device *dev) 1116 { 1117 if (dev_pm_skip_suspend(dev)) 1118 return 0; 1119 1120 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 1121 1122 return pm_generic_poweroff_late(dev); 1123 } 1124 1125 static int pci_pm_poweroff_noirq(struct device *dev) 1126 { 1127 struct pci_dev *pci_dev = to_pci_dev(dev); 1128 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1129 1130 if (dev_pm_skip_suspend(dev)) 1131 return 0; 1132 1133 if (pci_has_legacy_pm_support(pci_dev)) 1134 return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); 1135 1136 if (!pm) { 1137 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1138 return 0; 1139 } 1140 1141 if (pm->poweroff_noirq) { 1142 int error; 1143 1144 error = pm->poweroff_noirq(dev); 1145 suspend_report_result(pm->poweroff_noirq, error); 1146 if (error) 1147 return error; 1148 } 1149 1150 if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) 1151 pci_prepare_to_sleep(pci_dev); 1152 1153 /* 1154 * The reason for doing this here is the same as for the analogous code 1155 * in pci_pm_suspend_noirq(). 1156 */ 1157 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 1158 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 1159 1160 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1161 1162 return 0; 1163 } 1164 1165 static int pci_pm_restore_noirq(struct device *dev) 1166 { 1167 struct pci_dev *pci_dev = to_pci_dev(dev); 1168 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1169 1170 pci_pm_default_resume_early(pci_dev); 1171 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1172 1173 if (pci_has_legacy_pm_support(pci_dev)) 1174 return 0; 1175 1176 if (pm && pm->restore_noirq) 1177 return pm->restore_noirq(dev); 1178 1179 return 0; 1180 } 1181 1182 static int pci_pm_restore(struct device *dev) 1183 { 1184 struct pci_dev *pci_dev = to_pci_dev(dev); 1185 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1186 1187 /* 1188 * This is necessary for the hibernation error path in which restore is 1189 * called without restoring the standard config registers of the device. 1190 */ 1191 if (pci_dev->state_saved) 1192 pci_restore_standard_config(pci_dev); 1193 1194 if (pci_has_legacy_pm_support(pci_dev)) 1195 return pci_legacy_resume(dev); 1196 1197 pci_pm_default_resume(pci_dev); 1198 1199 if (pm) { 1200 if (pm->restore) 1201 return pm->restore(dev); 1202 } else { 1203 pci_pm_reenable_device(pci_dev); 1204 } 1205 1206 return 0; 1207 } 1208 1209 #else /* !CONFIG_HIBERNATE_CALLBACKS */ 1210 1211 #define pci_pm_freeze NULL 1212 #define pci_pm_freeze_noirq NULL 1213 #define pci_pm_thaw NULL 1214 #define pci_pm_thaw_noirq NULL 1215 #define pci_pm_poweroff NULL 1216 #define pci_pm_poweroff_late NULL 1217 #define pci_pm_poweroff_noirq NULL 1218 #define pci_pm_restore NULL 1219 #define pci_pm_restore_noirq NULL 1220 1221 #endif /* !CONFIG_HIBERNATE_CALLBACKS */ 1222 1223 #ifdef CONFIG_PM 1224 1225 static int pci_pm_runtime_suspend(struct device *dev) 1226 { 1227 struct pci_dev *pci_dev = to_pci_dev(dev); 1228 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1229 pci_power_t prev = pci_dev->current_state; 1230 int error; 1231 1232 /* 1233 * If pci_dev->driver is not set (unbound), we leave the device in D0, 1234 * but it may go to D3cold when the bridge above it runtime suspends. 1235 * Save its config space in case that happens. 1236 */ 1237 if (!pci_dev->driver) { 1238 pci_save_state(pci_dev); 1239 return 0; 1240 } 1241 1242 pci_dev->state_saved = false; 1243 if (pm && pm->runtime_suspend) { 1244 error = pm->runtime_suspend(dev); 1245 /* 1246 * -EBUSY and -EAGAIN is used to request the runtime PM core 1247 * to schedule a new suspend, so log the event only with debug 1248 * log level. 1249 */ 1250 if (error == -EBUSY || error == -EAGAIN) { 1251 pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", 1252 pm->runtime_suspend, error); 1253 return error; 1254 } else if (error) { 1255 pci_err(pci_dev, "can't suspend (%ps returned %d)\n", 1256 pm->runtime_suspend, error); 1257 return error; 1258 } 1259 } 1260 1261 pci_fixup_device(pci_fixup_suspend, pci_dev); 1262 1263 if (pm && pm->runtime_suspend 1264 && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 1265 && pci_dev->current_state != PCI_UNKNOWN) { 1266 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 1267 "PCI PM: State of device not saved by %pS\n", 1268 pm->runtime_suspend); 1269 return 0; 1270 } 1271 1272 if (!pci_dev->state_saved) { 1273 pci_save_state(pci_dev); 1274 pci_finish_runtime_suspend(pci_dev); 1275 } 1276 1277 return 0; 1278 } 1279 1280 static int pci_pm_runtime_resume(struct device *dev) 1281 { 1282 struct pci_dev *pci_dev = to_pci_dev(dev); 1283 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1284 pci_power_t prev_state = pci_dev->current_state; 1285 int error = 0; 1286 1287 /* 1288 * Restoring config space is necessary even if the device is not bound 1289 * to a driver because although we left it in D0, it may have gone to 1290 * D3cold when the bridge above it runtime suspended. 1291 */ 1292 pci_restore_standard_config(pci_dev); 1293 1294 if (!pci_dev->driver) 1295 return 0; 1296 1297 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1298 pci_pm_default_resume(pci_dev); 1299 1300 if (prev_state == PCI_D3cold) 1301 pci_bridge_wait_for_secondary_bus(pci_dev); 1302 1303 if (pm && pm->runtime_resume) 1304 error = pm->runtime_resume(dev); 1305 1306 pci_dev->runtime_d3cold = false; 1307 1308 return error; 1309 } 1310 1311 static int pci_pm_runtime_idle(struct device *dev) 1312 { 1313 struct pci_dev *pci_dev = to_pci_dev(dev); 1314 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1315 1316 /* 1317 * If pci_dev->driver is not set (unbound), the device should 1318 * always remain in D0 regardless of the runtime PM status 1319 */ 1320 if (!pci_dev->driver) 1321 return 0; 1322 1323 if (!pm) 1324 return -ENOSYS; 1325 1326 if (pm->runtime_idle) 1327 return pm->runtime_idle(dev); 1328 1329 return 0; 1330 } 1331 1332 static const struct dev_pm_ops pci_dev_pm_ops = { 1333 .prepare = pci_pm_prepare, 1334 .complete = pci_pm_complete, 1335 .suspend = pci_pm_suspend, 1336 .suspend_late = pci_pm_suspend_late, 1337 .resume = pci_pm_resume, 1338 .resume_early = pci_pm_resume_early, 1339 .freeze = pci_pm_freeze, 1340 .thaw = pci_pm_thaw, 1341 .poweroff = pci_pm_poweroff, 1342 .poweroff_late = pci_pm_poweroff_late, 1343 .restore = pci_pm_restore, 1344 .suspend_noirq = pci_pm_suspend_noirq, 1345 .resume_noirq = pci_pm_resume_noirq, 1346 .freeze_noirq = pci_pm_freeze_noirq, 1347 .thaw_noirq = pci_pm_thaw_noirq, 1348 .poweroff_noirq = pci_pm_poweroff_noirq, 1349 .restore_noirq = pci_pm_restore_noirq, 1350 .runtime_suspend = pci_pm_runtime_suspend, 1351 .runtime_resume = pci_pm_runtime_resume, 1352 .runtime_idle = pci_pm_runtime_idle, 1353 }; 1354 1355 #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) 1356 1357 #else /* !CONFIG_PM */ 1358 1359 #define pci_pm_runtime_suspend NULL 1360 #define pci_pm_runtime_resume NULL 1361 #define pci_pm_runtime_idle NULL 1362 1363 #define PCI_PM_OPS_PTR NULL 1364 1365 #endif /* !CONFIG_PM */ 1366 1367 /** 1368 * __pci_register_driver - register a new pci driver 1369 * @drv: the driver structure to register 1370 * @owner: owner module of drv 1371 * @mod_name: module name string 1372 * 1373 * Adds the driver structure to the list of registered drivers. 1374 * Returns a negative value on error, otherwise 0. 1375 * If no error occurred, the driver remains registered even if 1376 * no device was claimed during registration. 1377 */ 1378 int __pci_register_driver(struct pci_driver *drv, struct module *owner, 1379 const char *mod_name) 1380 { 1381 /* initialize common driver fields */ 1382 drv->driver.name = drv->name; 1383 drv->driver.bus = &pci_bus_type; 1384 drv->driver.owner = owner; 1385 drv->driver.mod_name = mod_name; 1386 drv->driver.groups = drv->groups; 1387 drv->driver.dev_groups = drv->dev_groups; 1388 1389 spin_lock_init(&drv->dynids.lock); 1390 INIT_LIST_HEAD(&drv->dynids.list); 1391 1392 /* register with core */ 1393 return driver_register(&drv->driver); 1394 } 1395 EXPORT_SYMBOL(__pci_register_driver); 1396 1397 /** 1398 * pci_unregister_driver - unregister a pci driver 1399 * @drv: the driver structure to unregister 1400 * 1401 * Deletes the driver structure from the list of registered PCI drivers, 1402 * gives it a chance to clean up by calling its remove() function for 1403 * each device it was responsible for, and marks those devices as 1404 * driverless. 1405 */ 1406 1407 void pci_unregister_driver(struct pci_driver *drv) 1408 { 1409 driver_unregister(&drv->driver); 1410 pci_free_dynids(drv); 1411 } 1412 EXPORT_SYMBOL(pci_unregister_driver); 1413 1414 static struct pci_driver pci_compat_driver = { 1415 .name = "compat" 1416 }; 1417 1418 /** 1419 * pci_dev_driver - get the pci_driver of a device 1420 * @dev: the device to query 1421 * 1422 * Returns the appropriate pci_driver structure or %NULL if there is no 1423 * registered driver for the device. 1424 */ 1425 struct pci_driver *pci_dev_driver(const struct pci_dev *dev) 1426 { 1427 if (dev->driver) 1428 return dev->driver; 1429 else { 1430 int i; 1431 for (i = 0; i <= PCI_ROM_RESOURCE; i++) 1432 if (dev->resource[i].flags & IORESOURCE_BUSY) 1433 return &pci_compat_driver; 1434 } 1435 return NULL; 1436 } 1437 EXPORT_SYMBOL(pci_dev_driver); 1438 1439 /** 1440 * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure 1441 * @dev: the PCI device structure to match against 1442 * @drv: the device driver to search for matching PCI device id structures 1443 * 1444 * Used by a driver to check whether a PCI device present in the 1445 * system is in its list of supported devices. Returns the matching 1446 * pci_device_id structure or %NULL if there is no match. 1447 */ 1448 static int pci_bus_match(struct device *dev, struct device_driver *drv) 1449 { 1450 struct pci_dev *pci_dev = to_pci_dev(dev); 1451 struct pci_driver *pci_drv; 1452 const struct pci_device_id *found_id; 1453 1454 if (!pci_dev->match_driver) 1455 return 0; 1456 1457 pci_drv = to_pci_driver(drv); 1458 found_id = pci_match_device(pci_drv, pci_dev); 1459 if (found_id) 1460 return 1; 1461 1462 return 0; 1463 } 1464 1465 /** 1466 * pci_dev_get - increments the reference count of the pci device structure 1467 * @dev: the device being referenced 1468 * 1469 * Each live reference to a device should be refcounted. 1470 * 1471 * Drivers for PCI devices should normally record such references in 1472 * their probe() methods, when they bind to a device, and release 1473 * them by calling pci_dev_put(), in their disconnect() methods. 1474 * 1475 * A pointer to the device with the incremented reference counter is returned. 1476 */ 1477 struct pci_dev *pci_dev_get(struct pci_dev *dev) 1478 { 1479 if (dev) 1480 get_device(&dev->dev); 1481 return dev; 1482 } 1483 EXPORT_SYMBOL(pci_dev_get); 1484 1485 /** 1486 * pci_dev_put - release a use of the pci device structure 1487 * @dev: device that's been disconnected 1488 * 1489 * Must be called when a user of a device is finished with it. When the last 1490 * user of the device calls this function, the memory of the device is freed. 1491 */ 1492 void pci_dev_put(struct pci_dev *dev) 1493 { 1494 if (dev) 1495 put_device(&dev->dev); 1496 } 1497 EXPORT_SYMBOL(pci_dev_put); 1498 1499 static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) 1500 { 1501 struct pci_dev *pdev; 1502 1503 if (!dev) 1504 return -ENODEV; 1505 1506 pdev = to_pci_dev(dev); 1507 1508 if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) 1509 return -ENOMEM; 1510 1511 if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) 1512 return -ENOMEM; 1513 1514 if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, 1515 pdev->subsystem_device)) 1516 return -ENOMEM; 1517 1518 if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) 1519 return -ENOMEM; 1520 1521 if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", 1522 pdev->vendor, pdev->device, 1523 pdev->subsystem_vendor, pdev->subsystem_device, 1524 (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), 1525 (u8)(pdev->class))) 1526 return -ENOMEM; 1527 1528 return 0; 1529 } 1530 1531 #if defined(CONFIG_PCIEPORTBUS) || defined(CONFIG_EEH) 1532 /** 1533 * pci_uevent_ers - emit a uevent during recovery path of PCI device 1534 * @pdev: PCI device undergoing error recovery 1535 * @err_type: type of error event 1536 */ 1537 void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) 1538 { 1539 int idx = 0; 1540 char *envp[3]; 1541 1542 switch (err_type) { 1543 case PCI_ERS_RESULT_NONE: 1544 case PCI_ERS_RESULT_CAN_RECOVER: 1545 envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; 1546 envp[idx++] = "DEVICE_ONLINE=0"; 1547 break; 1548 case PCI_ERS_RESULT_RECOVERED: 1549 envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; 1550 envp[idx++] = "DEVICE_ONLINE=1"; 1551 break; 1552 case PCI_ERS_RESULT_DISCONNECT: 1553 envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; 1554 envp[idx++] = "DEVICE_ONLINE=0"; 1555 break; 1556 default: 1557 break; 1558 } 1559 1560 if (idx > 0) { 1561 envp[idx++] = NULL; 1562 kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); 1563 } 1564 } 1565 #endif 1566 1567 static int pci_bus_num_vf(struct device *dev) 1568 { 1569 return pci_num_vf(to_pci_dev(dev)); 1570 } 1571 1572 /** 1573 * pci_dma_configure - Setup DMA configuration 1574 * @dev: ptr to dev structure 1575 * 1576 * Function to update PCI devices's DMA configuration using the same 1577 * info from the OF node or ACPI node of host bridge's parent (if any). 1578 */ 1579 static int pci_dma_configure(struct device *dev) 1580 { 1581 struct device *bridge; 1582 int ret = 0; 1583 1584 bridge = pci_get_host_bridge_device(to_pci_dev(dev)); 1585 1586 if (IS_ENABLED(CONFIG_OF) && bridge->parent && 1587 bridge->parent->of_node) { 1588 ret = of_dma_configure(dev, bridge->parent->of_node, true); 1589 } else if (has_acpi_companion(bridge)) { 1590 struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); 1591 1592 ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); 1593 } 1594 1595 pci_put_host_bridge_device(bridge); 1596 return ret; 1597 } 1598 1599 struct bus_type pci_bus_type = { 1600 .name = "pci", 1601 .match = pci_bus_match, 1602 .uevent = pci_uevent, 1603 .probe = pci_device_probe, 1604 .remove = pci_device_remove, 1605 .shutdown = pci_device_shutdown, 1606 .dev_groups = pci_dev_groups, 1607 .bus_groups = pci_bus_groups, 1608 .drv_groups = pci_drv_groups, 1609 .pm = PCI_PM_OPS_PTR, 1610 .num_vf = pci_bus_num_vf, 1611 .dma_configure = pci_dma_configure, 1612 }; 1613 EXPORT_SYMBOL(pci_bus_type); 1614 1615 #ifdef CONFIG_PCIEPORTBUS 1616 static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) 1617 { 1618 struct pcie_device *pciedev; 1619 struct pcie_port_service_driver *driver; 1620 1621 if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) 1622 return 0; 1623 1624 pciedev = to_pcie_device(dev); 1625 driver = to_service_driver(drv); 1626 1627 if (driver->service != pciedev->service) 1628 return 0; 1629 1630 if (driver->port_type != PCIE_ANY_PORT && 1631 driver->port_type != pci_pcie_type(pciedev->port)) 1632 return 0; 1633 1634 return 1; 1635 } 1636 1637 struct bus_type pcie_port_bus_type = { 1638 .name = "pci_express", 1639 .match = pcie_port_bus_match, 1640 }; 1641 EXPORT_SYMBOL_GPL(pcie_port_bus_type); 1642 #endif 1643 1644 static int __init pci_driver_init(void) 1645 { 1646 int ret; 1647 1648 ret = bus_register(&pci_bus_type); 1649 if (ret) 1650 return ret; 1651 1652 #ifdef CONFIG_PCIEPORTBUS 1653 ret = bus_register(&pcie_port_bus_type); 1654 if (ret) 1655 return ret; 1656 #endif 1657 dma_debug_add_bus(&pci_bus_type); 1658 return 0; 1659 } 1660 postcore_initcall(pci_driver_init); 1661