1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> 4 * (C) Copyright 2007 Novell Inc. 5 */ 6 7 #include <linux/pci.h> 8 #include <linux/module.h> 9 #include <linux/init.h> 10 #include <linux/device.h> 11 #include <linux/mempolicy.h> 12 #include <linux/string.h> 13 #include <linux/slab.h> 14 #include <linux/sched.h> 15 #include <linux/sched/isolation.h> 16 #include <linux/cpu.h> 17 #include <linux/pm_runtime.h> 18 #include <linux/suspend.h> 19 #include <linux/kexec.h> 20 #include <linux/of_device.h> 21 #include <linux/acpi.h> 22 #include <linux/dma-map-ops.h> 23 #include "pci.h" 24 #include "pcie/portdrv.h" 25 26 struct pci_dynid { 27 struct list_head node; 28 struct pci_device_id id; 29 }; 30 31 /** 32 * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices 33 * @drv: target pci driver 34 * @vendor: PCI vendor ID 35 * @device: PCI device ID 36 * @subvendor: PCI subvendor ID 37 * @subdevice: PCI subdevice ID 38 * @class: PCI class 39 * @class_mask: PCI class mask 40 * @driver_data: private driver data 41 * 42 * Adds a new dynamic pci device ID to this driver and causes the 43 * driver to probe for all devices again. @drv must have been 44 * registered prior to calling this function. 45 * 46 * CONTEXT: 47 * Does GFP_KERNEL allocation. 48 * 49 * RETURNS: 50 * 0 on success, -errno on failure. 51 */ 52 int pci_add_dynid(struct pci_driver *drv, 53 unsigned int vendor, unsigned int device, 54 unsigned int subvendor, unsigned int subdevice, 55 unsigned int class, unsigned int class_mask, 56 unsigned long driver_data) 57 { 58 struct pci_dynid *dynid; 59 60 dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); 61 if (!dynid) 62 return -ENOMEM; 63 64 dynid->id.vendor = vendor; 65 dynid->id.device = device; 66 dynid->id.subvendor = subvendor; 67 dynid->id.subdevice = subdevice; 68 dynid->id.class = class; 69 dynid->id.class_mask = class_mask; 70 dynid->id.driver_data = driver_data; 71 72 spin_lock(&drv->dynids.lock); 73 list_add_tail(&dynid->node, &drv->dynids.list); 74 spin_unlock(&drv->dynids.lock); 75 76 return driver_attach(&drv->driver); 77 } 78 EXPORT_SYMBOL_GPL(pci_add_dynid); 79 80 static void pci_free_dynids(struct pci_driver *drv) 81 { 82 struct pci_dynid *dynid, *n; 83 84 spin_lock(&drv->dynids.lock); 85 list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { 86 list_del(&dynid->node); 87 kfree(dynid); 88 } 89 spin_unlock(&drv->dynids.lock); 90 } 91 92 /** 93 * pci_match_id - See if a PCI device matches a given pci_id table 94 * @ids: array of PCI device ID structures to search in 95 * @dev: the PCI device structure to match against. 96 * 97 * Used by a driver to check whether a PCI device is in its list of 98 * supported devices. Returns the matching pci_device_id structure or 99 * %NULL if there is no match. 100 * 101 * Deprecated; don't use this as it will not catch any dynamic IDs 102 * that a driver might want to check for. 103 */ 104 const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, 105 struct pci_dev *dev) 106 { 107 if (ids) { 108 while (ids->vendor || ids->subvendor || ids->class_mask) { 109 if (pci_match_one_device(ids, dev)) 110 return ids; 111 ids++; 112 } 113 } 114 return NULL; 115 } 116 EXPORT_SYMBOL(pci_match_id); 117 118 static const struct pci_device_id pci_device_id_any = { 119 .vendor = PCI_ANY_ID, 120 .device = PCI_ANY_ID, 121 .subvendor = PCI_ANY_ID, 122 .subdevice = PCI_ANY_ID, 123 }; 124 125 /** 126 * pci_match_device - See if a device matches a driver's list of IDs 127 * @drv: the PCI driver to match against 128 * @dev: the PCI device structure to match against 129 * 130 * Used by a driver to check whether a PCI device is in its list of 131 * supported devices or in the dynids list, which may have been augmented 132 * via the sysfs "new_id" file. Returns the matching pci_device_id 133 * structure or %NULL if there is no match. 134 */ 135 static const struct pci_device_id *pci_match_device(struct pci_driver *drv, 136 struct pci_dev *dev) 137 { 138 struct pci_dynid *dynid; 139 const struct pci_device_id *found_id = NULL, *ids; 140 141 /* When driver_override is set, only bind to the matching driver */ 142 if (dev->driver_override && strcmp(dev->driver_override, drv->name)) 143 return NULL; 144 145 /* Look at the dynamic ids first, before the static ones */ 146 spin_lock(&drv->dynids.lock); 147 list_for_each_entry(dynid, &drv->dynids.list, node) { 148 if (pci_match_one_device(&dynid->id, dev)) { 149 found_id = &dynid->id; 150 break; 151 } 152 } 153 spin_unlock(&drv->dynids.lock); 154 155 if (found_id) 156 return found_id; 157 158 for (ids = drv->id_table; (found_id = pci_match_id(ids, dev)); 159 ids = found_id + 1) { 160 /* 161 * The match table is split based on driver_override. 162 * In case override_only was set, enforce driver_override 163 * matching. 164 */ 165 if (found_id->override_only) { 166 if (dev->driver_override) 167 return found_id; 168 } else { 169 return found_id; 170 } 171 } 172 173 /* driver_override will always match, send a dummy id */ 174 if (dev->driver_override) 175 return &pci_device_id_any; 176 return NULL; 177 } 178 179 /** 180 * new_id_store - sysfs frontend to pci_add_dynid() 181 * @driver: target device driver 182 * @buf: buffer for scanning device ID data 183 * @count: input size 184 * 185 * Allow PCI IDs to be added to an existing driver via sysfs. 186 */ 187 static ssize_t new_id_store(struct device_driver *driver, const char *buf, 188 size_t count) 189 { 190 struct pci_driver *pdrv = to_pci_driver(driver); 191 const struct pci_device_id *ids = pdrv->id_table; 192 u32 vendor, device, subvendor = PCI_ANY_ID, 193 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 194 unsigned long driver_data = 0; 195 int fields = 0; 196 int retval = 0; 197 198 fields = sscanf(buf, "%x %x %x %x %x %x %lx", 199 &vendor, &device, &subvendor, &subdevice, 200 &class, &class_mask, &driver_data); 201 if (fields < 2) 202 return -EINVAL; 203 204 if (fields != 7) { 205 struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); 206 if (!pdev) 207 return -ENOMEM; 208 209 pdev->vendor = vendor; 210 pdev->device = device; 211 pdev->subsystem_vendor = subvendor; 212 pdev->subsystem_device = subdevice; 213 pdev->class = class; 214 215 if (pci_match_device(pdrv, pdev)) 216 retval = -EEXIST; 217 218 kfree(pdev); 219 220 if (retval) 221 return retval; 222 } 223 224 /* Only accept driver_data values that match an existing id_table 225 entry */ 226 if (ids) { 227 retval = -EINVAL; 228 while (ids->vendor || ids->subvendor || ids->class_mask) { 229 if (driver_data == ids->driver_data) { 230 retval = 0; 231 break; 232 } 233 ids++; 234 } 235 if (retval) /* No match */ 236 return retval; 237 } 238 239 retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, 240 class, class_mask, driver_data); 241 if (retval) 242 return retval; 243 return count; 244 } 245 static DRIVER_ATTR_WO(new_id); 246 247 /** 248 * remove_id_store - remove a PCI device ID from this driver 249 * @driver: target device driver 250 * @buf: buffer for scanning device ID data 251 * @count: input size 252 * 253 * Removes a dynamic pci device ID to this driver. 254 */ 255 static ssize_t remove_id_store(struct device_driver *driver, const char *buf, 256 size_t count) 257 { 258 struct pci_dynid *dynid, *n; 259 struct pci_driver *pdrv = to_pci_driver(driver); 260 u32 vendor, device, subvendor = PCI_ANY_ID, 261 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 262 int fields = 0; 263 size_t retval = -ENODEV; 264 265 fields = sscanf(buf, "%x %x %x %x %x %x", 266 &vendor, &device, &subvendor, &subdevice, 267 &class, &class_mask); 268 if (fields < 2) 269 return -EINVAL; 270 271 spin_lock(&pdrv->dynids.lock); 272 list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { 273 struct pci_device_id *id = &dynid->id; 274 if ((id->vendor == vendor) && 275 (id->device == device) && 276 (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && 277 (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && 278 !((id->class ^ class) & class_mask)) { 279 list_del(&dynid->node); 280 kfree(dynid); 281 retval = count; 282 break; 283 } 284 } 285 spin_unlock(&pdrv->dynids.lock); 286 287 return retval; 288 } 289 static DRIVER_ATTR_WO(remove_id); 290 291 static struct attribute *pci_drv_attrs[] = { 292 &driver_attr_new_id.attr, 293 &driver_attr_remove_id.attr, 294 NULL, 295 }; 296 ATTRIBUTE_GROUPS(pci_drv); 297 298 struct drv_dev_and_id { 299 struct pci_driver *drv; 300 struct pci_dev *dev; 301 const struct pci_device_id *id; 302 }; 303 304 static long local_pci_probe(void *_ddi) 305 { 306 struct drv_dev_and_id *ddi = _ddi; 307 struct pci_dev *pci_dev = ddi->dev; 308 struct pci_driver *pci_drv = ddi->drv; 309 struct device *dev = &pci_dev->dev; 310 int rc; 311 312 /* 313 * Unbound PCI devices are always put in D0, regardless of 314 * runtime PM status. During probe, the device is set to 315 * active and the usage count is incremented. If the driver 316 * supports runtime PM, it should call pm_runtime_put_noidle(), 317 * or any other runtime PM helper function decrementing the usage 318 * count, in its probe routine and pm_runtime_get_noresume() in 319 * its remove routine. 320 */ 321 pm_runtime_get_sync(dev); 322 rc = pci_drv->probe(pci_dev, ddi->id); 323 if (!rc) 324 return rc; 325 if (rc < 0) { 326 pm_runtime_put_sync(dev); 327 return rc; 328 } 329 /* 330 * Probe function should return < 0 for failure, 0 for success 331 * Treat values > 0 as success, but warn. 332 */ 333 pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", 334 rc); 335 return 0; 336 } 337 338 static bool pci_physfn_is_probed(struct pci_dev *dev) 339 { 340 #ifdef CONFIG_PCI_IOV 341 return dev->is_virtfn && dev->physfn->is_probed; 342 #else 343 return false; 344 #endif 345 } 346 347 static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, 348 const struct pci_device_id *id) 349 { 350 int error, node, cpu; 351 int hk_flags = HK_FLAG_DOMAIN | HK_FLAG_WQ; 352 struct drv_dev_and_id ddi = { drv, dev, id }; 353 354 /* 355 * Execute driver initialization on node where the device is 356 * attached. This way the driver likely allocates its local memory 357 * on the right node. 358 */ 359 node = dev_to_node(&dev->dev); 360 dev->is_probed = 1; 361 362 cpu_hotplug_disable(); 363 364 /* 365 * Prevent nesting work_on_cpu() for the case where a Virtual Function 366 * device is probed from work_on_cpu() of the Physical device. 367 */ 368 if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || 369 pci_physfn_is_probed(dev)) 370 cpu = nr_cpu_ids; 371 else 372 cpu = cpumask_any_and(cpumask_of_node(node), 373 housekeeping_cpumask(hk_flags)); 374 375 if (cpu < nr_cpu_ids) 376 error = work_on_cpu(cpu, local_pci_probe, &ddi); 377 else 378 error = local_pci_probe(&ddi); 379 380 dev->is_probed = 0; 381 cpu_hotplug_enable(); 382 return error; 383 } 384 385 /** 386 * __pci_device_probe - check if a driver wants to claim a specific PCI device 387 * @drv: driver to call to check if it wants the PCI device 388 * @pci_dev: PCI device being probed 389 * 390 * returns 0 on success, else error. 391 */ 392 static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) 393 { 394 const struct pci_device_id *id; 395 int error = 0; 396 397 if (drv->probe) { 398 error = -ENODEV; 399 400 id = pci_match_device(drv, pci_dev); 401 if (id) 402 error = pci_call_probe(drv, pci_dev, id); 403 } 404 return error; 405 } 406 407 int __weak pcibios_alloc_irq(struct pci_dev *dev) 408 { 409 return 0; 410 } 411 412 void __weak pcibios_free_irq(struct pci_dev *dev) 413 { 414 } 415 416 #ifdef CONFIG_PCI_IOV 417 static inline bool pci_device_can_probe(struct pci_dev *pdev) 418 { 419 return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || 420 pdev->driver_override); 421 } 422 #else 423 static inline bool pci_device_can_probe(struct pci_dev *pdev) 424 { 425 return true; 426 } 427 #endif 428 429 static int pci_device_probe(struct device *dev) 430 { 431 int error; 432 struct pci_dev *pci_dev = to_pci_dev(dev); 433 struct pci_driver *drv = to_pci_driver(dev->driver); 434 435 if (!pci_device_can_probe(pci_dev)) 436 return -ENODEV; 437 438 pci_assign_irq(pci_dev); 439 440 error = pcibios_alloc_irq(pci_dev); 441 if (error < 0) 442 return error; 443 444 pci_dev_get(pci_dev); 445 error = __pci_device_probe(drv, pci_dev); 446 if (error) { 447 pcibios_free_irq(pci_dev); 448 pci_dev_put(pci_dev); 449 } 450 451 return error; 452 } 453 454 static void pci_device_remove(struct device *dev) 455 { 456 struct pci_dev *pci_dev = to_pci_dev(dev); 457 struct pci_driver *drv = to_pci_driver(dev->driver); 458 459 if (drv->remove) { 460 pm_runtime_get_sync(dev); 461 drv->remove(pci_dev); 462 pm_runtime_put_noidle(dev); 463 } 464 pcibios_free_irq(pci_dev); 465 pci_iov_remove(pci_dev); 466 467 /* Undo the runtime PM settings in local_pci_probe() */ 468 pm_runtime_put_sync(dev); 469 470 /* 471 * If the device is still on, set the power state as "unknown", 472 * since it might change by the next time we load the driver. 473 */ 474 if (pci_dev->current_state == PCI_D0) 475 pci_dev->current_state = PCI_UNKNOWN; 476 477 /* 478 * We would love to complain here if pci_dev->is_enabled is set, that 479 * the driver should have called pci_disable_device(), but the 480 * unfortunate fact is there are too many odd BIOS and bridge setups 481 * that don't like drivers doing that all of the time. 482 * Oh well, we can dream of sane hardware when we sleep, no matter how 483 * horrible the crap we have to deal with is when we are awake... 484 */ 485 486 pci_dev_put(pci_dev); 487 } 488 489 static void pci_device_shutdown(struct device *dev) 490 { 491 struct pci_dev *pci_dev = to_pci_dev(dev); 492 struct pci_driver *drv = to_pci_driver(dev->driver); 493 494 pm_runtime_resume(dev); 495 496 if (drv && drv->shutdown) 497 drv->shutdown(pci_dev); 498 499 /* 500 * If this is a kexec reboot, turn off Bus Master bit on the 501 * device to tell it to not continue to do DMA. Don't touch 502 * devices in D3cold or unknown states. 503 * If it is not a kexec reboot, firmware will hit the PCI 504 * devices with big hammer and stop their DMA any way. 505 */ 506 if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) 507 pci_clear_master(pci_dev); 508 } 509 510 #ifdef CONFIG_PM 511 512 /* Auxiliary functions used for system resume and run-time resume. */ 513 514 /** 515 * pci_restore_standard_config - restore standard config registers of PCI device 516 * @pci_dev: PCI device to handle 517 */ 518 static int pci_restore_standard_config(struct pci_dev *pci_dev) 519 { 520 pci_update_current_state(pci_dev, PCI_UNKNOWN); 521 522 if (pci_dev->current_state != PCI_D0) { 523 int error = pci_set_power_state(pci_dev, PCI_D0); 524 if (error) 525 return error; 526 } 527 528 pci_restore_state(pci_dev); 529 pci_pme_restore(pci_dev); 530 return 0; 531 } 532 533 static void pci_pm_default_resume(struct pci_dev *pci_dev) 534 { 535 pci_fixup_device(pci_fixup_resume, pci_dev); 536 pci_enable_wake(pci_dev, PCI_D0, false); 537 } 538 539 #endif 540 541 #ifdef CONFIG_PM_SLEEP 542 543 static void pci_pm_default_resume_early(struct pci_dev *pci_dev) 544 { 545 pci_power_up(pci_dev); 546 pci_update_current_state(pci_dev, PCI_D0); 547 pci_restore_state(pci_dev); 548 pci_pme_restore(pci_dev); 549 } 550 551 /* 552 * Default "suspend" method for devices that have no driver provided suspend, 553 * or not even a driver at all (second part). 554 */ 555 static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) 556 { 557 /* 558 * mark its power state as "unknown", since we don't know if 559 * e.g. the BIOS will change its device state when we suspend. 560 */ 561 if (pci_dev->current_state == PCI_D0) 562 pci_dev->current_state = PCI_UNKNOWN; 563 } 564 565 /* 566 * Default "resume" method for devices that have no driver provided resume, 567 * or not even a driver at all (second part). 568 */ 569 static int pci_pm_reenable_device(struct pci_dev *pci_dev) 570 { 571 int retval; 572 573 /* if the device was enabled before suspend, re-enable */ 574 retval = pci_reenable_device(pci_dev); 575 /* 576 * if the device was busmaster before the suspend, make it busmaster 577 * again 578 */ 579 if (pci_dev->is_busmaster) 580 pci_set_master(pci_dev); 581 582 return retval; 583 } 584 585 static int pci_legacy_suspend(struct device *dev, pm_message_t state) 586 { 587 struct pci_dev *pci_dev = to_pci_dev(dev); 588 struct pci_driver *drv = to_pci_driver(dev->driver); 589 590 if (drv && drv->suspend) { 591 pci_power_t prev = pci_dev->current_state; 592 int error; 593 594 error = drv->suspend(pci_dev, state); 595 suspend_report_result(drv->suspend, error); 596 if (error) 597 return error; 598 599 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 600 && pci_dev->current_state != PCI_UNKNOWN) { 601 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 602 "PCI PM: Device state not saved by %pS\n", 603 drv->suspend); 604 } 605 } 606 607 pci_fixup_device(pci_fixup_suspend, pci_dev); 608 609 return 0; 610 } 611 612 static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) 613 { 614 struct pci_dev *pci_dev = to_pci_dev(dev); 615 616 if (!pci_dev->state_saved) 617 pci_save_state(pci_dev); 618 619 pci_pm_set_unknown_state(pci_dev); 620 621 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 622 623 return 0; 624 } 625 626 static int pci_legacy_resume(struct device *dev) 627 { 628 struct pci_dev *pci_dev = to_pci_dev(dev); 629 struct pci_driver *drv = to_pci_driver(dev->driver); 630 631 pci_fixup_device(pci_fixup_resume, pci_dev); 632 633 return drv && drv->resume ? 634 drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); 635 } 636 637 /* Auxiliary functions used by the new power management framework */ 638 639 static void pci_pm_default_suspend(struct pci_dev *pci_dev) 640 { 641 /* Disable non-bridge devices without PM support */ 642 if (!pci_has_subordinate(pci_dev)) 643 pci_disable_enabled_device(pci_dev); 644 } 645 646 static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) 647 { 648 struct pci_driver *drv = to_pci_driver(pci_dev->dev.driver); 649 bool ret = drv && (drv->suspend || drv->resume); 650 651 /* 652 * Legacy PM support is used by default, so warn if the new framework is 653 * supported as well. Drivers are supposed to support either the 654 * former, or the latter, but not both at the same time. 655 */ 656 pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", 657 pci_dev->vendor, pci_dev->device); 658 659 return ret; 660 } 661 662 /* New power management framework */ 663 664 static int pci_pm_prepare(struct device *dev) 665 { 666 struct pci_dev *pci_dev = to_pci_dev(dev); 667 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 668 669 if (pm && pm->prepare) { 670 int error = pm->prepare(dev); 671 if (error < 0) 672 return error; 673 674 if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) 675 return 0; 676 } 677 if (pci_dev_need_resume(pci_dev)) 678 return 0; 679 680 /* 681 * The PME setting needs to be adjusted here in case the direct-complete 682 * optimization is used with respect to this device. 683 */ 684 pci_dev_adjust_pme(pci_dev); 685 return 1; 686 } 687 688 static void pci_pm_complete(struct device *dev) 689 { 690 struct pci_dev *pci_dev = to_pci_dev(dev); 691 692 pci_dev_complete_resume(pci_dev); 693 pm_generic_complete(dev); 694 695 /* Resume device if platform firmware has put it in reset-power-on */ 696 if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { 697 pci_power_t pre_sleep_state = pci_dev->current_state; 698 699 pci_refresh_power_state(pci_dev); 700 /* 701 * On platforms with ACPI this check may also trigger for 702 * devices sharing power resources if one of those power 703 * resources has been activated as a result of a change of the 704 * power state of another device sharing it. However, in that 705 * case it is also better to resume the device, in general. 706 */ 707 if (pci_dev->current_state < pre_sleep_state) 708 pm_request_resume(dev); 709 } 710 } 711 712 #else /* !CONFIG_PM_SLEEP */ 713 714 #define pci_pm_prepare NULL 715 #define pci_pm_complete NULL 716 717 #endif /* !CONFIG_PM_SLEEP */ 718 719 #ifdef CONFIG_SUSPEND 720 static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) 721 { 722 /* 723 * Some BIOSes forget to clear Root PME Status bits after system 724 * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. 725 * Clear those bits now just in case (shouldn't hurt). 726 */ 727 if (pci_is_pcie(pci_dev) && 728 (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || 729 pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) 730 pcie_clear_root_pme_status(pci_dev); 731 } 732 733 static int pci_pm_suspend(struct device *dev) 734 { 735 struct pci_dev *pci_dev = to_pci_dev(dev); 736 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 737 738 pci_dev->skip_bus_pm = false; 739 740 if (pci_has_legacy_pm_support(pci_dev)) 741 return pci_legacy_suspend(dev, PMSG_SUSPEND); 742 743 if (!pm) { 744 pci_pm_default_suspend(pci_dev); 745 return 0; 746 } 747 748 /* 749 * PCI devices suspended at run time may need to be resumed at this 750 * point, because in general it may be necessary to reconfigure them for 751 * system suspend. Namely, if the device is expected to wake up the 752 * system from the sleep state, it may have to be reconfigured for this 753 * purpose, or if the device is not expected to wake up the system from 754 * the sleep state, it should be prevented from signaling wakeup events 755 * going forward. 756 * 757 * Also if the driver of the device does not indicate that its system 758 * suspend callbacks can cope with runtime-suspended devices, it is 759 * better to resume the device from runtime suspend here. 760 */ 761 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 762 pci_dev_need_resume(pci_dev)) { 763 pm_runtime_resume(dev); 764 pci_dev->state_saved = false; 765 } else { 766 pci_dev_adjust_pme(pci_dev); 767 } 768 769 if (pm->suspend) { 770 pci_power_t prev = pci_dev->current_state; 771 int error; 772 773 error = pm->suspend(dev); 774 suspend_report_result(pm->suspend, error); 775 if (error) 776 return error; 777 778 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 779 && pci_dev->current_state != PCI_UNKNOWN) { 780 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 781 "PCI PM: State of device not saved by %pS\n", 782 pm->suspend); 783 } 784 } 785 786 return 0; 787 } 788 789 static int pci_pm_suspend_late(struct device *dev) 790 { 791 if (dev_pm_skip_suspend(dev)) 792 return 0; 793 794 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 795 796 return pm_generic_suspend_late(dev); 797 } 798 799 static int pci_pm_suspend_noirq(struct device *dev) 800 { 801 struct pci_dev *pci_dev = to_pci_dev(dev); 802 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 803 804 if (dev_pm_skip_suspend(dev)) 805 return 0; 806 807 if (pci_has_legacy_pm_support(pci_dev)) 808 return pci_legacy_suspend_late(dev, PMSG_SUSPEND); 809 810 if (!pm) { 811 pci_save_state(pci_dev); 812 goto Fixup; 813 } 814 815 if (pm->suspend_noirq) { 816 pci_power_t prev = pci_dev->current_state; 817 int error; 818 819 error = pm->suspend_noirq(dev); 820 suspend_report_result(pm->suspend_noirq, error); 821 if (error) 822 return error; 823 824 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 825 && pci_dev->current_state != PCI_UNKNOWN) { 826 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 827 "PCI PM: State of device not saved by %pS\n", 828 pm->suspend_noirq); 829 goto Fixup; 830 } 831 } 832 833 if (pci_dev->skip_bus_pm) { 834 /* 835 * Either the device is a bridge with a child in D0 below it, or 836 * the function is running for the second time in a row without 837 * going through full resume, which is possible only during 838 * suspend-to-idle in a spurious wakeup case. The device should 839 * be in D0 at this point, but if it is a bridge, it may be 840 * necessary to save its state. 841 */ 842 if (!pci_dev->state_saved) 843 pci_save_state(pci_dev); 844 } else if (!pci_dev->state_saved) { 845 pci_save_state(pci_dev); 846 if (pci_power_manageable(pci_dev)) 847 pci_prepare_to_sleep(pci_dev); 848 } 849 850 pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", 851 pci_power_name(pci_dev->current_state)); 852 853 if (pci_dev->current_state == PCI_D0) { 854 pci_dev->skip_bus_pm = true; 855 /* 856 * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any 857 * downstream device is in D0, so avoid changing the power state 858 * of the parent bridge by setting the skip_bus_pm flag for it. 859 */ 860 if (pci_dev->bus->self) 861 pci_dev->bus->self->skip_bus_pm = true; 862 } 863 864 if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { 865 pci_dbg(pci_dev, "PCI PM: Skipped\n"); 866 goto Fixup; 867 } 868 869 pci_pm_set_unknown_state(pci_dev); 870 871 /* 872 * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's 873 * PCI COMMAND register isn't 0, the BIOS assumes that the controller 874 * hasn't been quiesced and tries to turn it off. If the controller 875 * is already in D3, this can hang or cause memory corruption. 876 * 877 * Since the value of the COMMAND register doesn't matter once the 878 * device has been suspended, we can safely set it to 0 here. 879 */ 880 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 881 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 882 883 Fixup: 884 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 885 886 /* 887 * If the target system sleep state is suspend-to-idle, it is sufficient 888 * to check whether or not the device's wakeup settings are good for 889 * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause 890 * pci_pm_complete() to take care of fixing up the device's state 891 * anyway, if need be. 892 */ 893 if (device_can_wakeup(dev) && !device_may_wakeup(dev)) 894 dev->power.may_skip_resume = false; 895 896 return 0; 897 } 898 899 static int pci_pm_resume_noirq(struct device *dev) 900 { 901 struct pci_dev *pci_dev = to_pci_dev(dev); 902 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 903 pci_power_t prev_state = pci_dev->current_state; 904 bool skip_bus_pm = pci_dev->skip_bus_pm; 905 906 if (dev_pm_skip_resume(dev)) 907 return 0; 908 909 /* 910 * In the suspend-to-idle case, devices left in D0 during suspend will 911 * stay in D0, so it is not necessary to restore or update their 912 * configuration here and attempting to put them into D0 again is 913 * pointless, so avoid doing that. 914 */ 915 if (!(skip_bus_pm && pm_suspend_no_platform())) 916 pci_pm_default_resume_early(pci_dev); 917 918 pci_fixup_device(pci_fixup_resume_early, pci_dev); 919 pcie_pme_root_status_cleanup(pci_dev); 920 921 if (!skip_bus_pm && prev_state == PCI_D3cold) 922 pci_bridge_wait_for_secondary_bus(pci_dev); 923 924 if (pci_has_legacy_pm_support(pci_dev)) 925 return 0; 926 927 if (pm && pm->resume_noirq) 928 return pm->resume_noirq(dev); 929 930 return 0; 931 } 932 933 static int pci_pm_resume_early(struct device *dev) 934 { 935 if (dev_pm_skip_resume(dev)) 936 return 0; 937 938 return pm_generic_resume_early(dev); 939 } 940 941 static int pci_pm_resume(struct device *dev) 942 { 943 struct pci_dev *pci_dev = to_pci_dev(dev); 944 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 945 946 /* 947 * This is necessary for the suspend error path in which resume is 948 * called without restoring the standard config registers of the device. 949 */ 950 if (pci_dev->state_saved) 951 pci_restore_standard_config(pci_dev); 952 953 if (pci_has_legacy_pm_support(pci_dev)) 954 return pci_legacy_resume(dev); 955 956 pci_pm_default_resume(pci_dev); 957 958 if (pm) { 959 if (pm->resume) 960 return pm->resume(dev); 961 } else { 962 pci_pm_reenable_device(pci_dev); 963 } 964 965 return 0; 966 } 967 968 #else /* !CONFIG_SUSPEND */ 969 970 #define pci_pm_suspend NULL 971 #define pci_pm_suspend_late NULL 972 #define pci_pm_suspend_noirq NULL 973 #define pci_pm_resume NULL 974 #define pci_pm_resume_early NULL 975 #define pci_pm_resume_noirq NULL 976 977 #endif /* !CONFIG_SUSPEND */ 978 979 #ifdef CONFIG_HIBERNATE_CALLBACKS 980 981 static int pci_pm_freeze(struct device *dev) 982 { 983 struct pci_dev *pci_dev = to_pci_dev(dev); 984 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 985 986 if (pci_has_legacy_pm_support(pci_dev)) 987 return pci_legacy_suspend(dev, PMSG_FREEZE); 988 989 if (!pm) { 990 pci_pm_default_suspend(pci_dev); 991 return 0; 992 } 993 994 /* 995 * Resume all runtime-suspended devices before creating a snapshot 996 * image of system memory, because the restore kernel generally cannot 997 * be expected to always handle them consistently and they need to be 998 * put into the runtime-active metastate during system resume anyway, 999 * so it is better to ensure that the state saved in the image will be 1000 * always consistent with that. 1001 */ 1002 pm_runtime_resume(dev); 1003 pci_dev->state_saved = false; 1004 1005 if (pm->freeze) { 1006 int error; 1007 1008 error = pm->freeze(dev); 1009 suspend_report_result(pm->freeze, error); 1010 if (error) 1011 return error; 1012 } 1013 1014 return 0; 1015 } 1016 1017 static int pci_pm_freeze_noirq(struct device *dev) 1018 { 1019 struct pci_dev *pci_dev = to_pci_dev(dev); 1020 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1021 1022 if (pci_has_legacy_pm_support(pci_dev)) 1023 return pci_legacy_suspend_late(dev, PMSG_FREEZE); 1024 1025 if (pm && pm->freeze_noirq) { 1026 int error; 1027 1028 error = pm->freeze_noirq(dev); 1029 suspend_report_result(pm->freeze_noirq, error); 1030 if (error) 1031 return error; 1032 } 1033 1034 if (!pci_dev->state_saved) 1035 pci_save_state(pci_dev); 1036 1037 pci_pm_set_unknown_state(pci_dev); 1038 1039 return 0; 1040 } 1041 1042 static int pci_pm_thaw_noirq(struct device *dev) 1043 { 1044 struct pci_dev *pci_dev = to_pci_dev(dev); 1045 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1046 1047 /* 1048 * The pm->thaw_noirq() callback assumes the device has been 1049 * returned to D0 and its config state has been restored. 1050 * 1051 * In addition, pci_restore_state() restores MSI-X state in MMIO 1052 * space, which requires the device to be in D0, so return it to D0 1053 * in case the driver's "freeze" callbacks put it into a low-power 1054 * state. 1055 */ 1056 pci_set_power_state(pci_dev, PCI_D0); 1057 pci_restore_state(pci_dev); 1058 1059 if (pci_has_legacy_pm_support(pci_dev)) 1060 return 0; 1061 1062 if (pm && pm->thaw_noirq) 1063 return pm->thaw_noirq(dev); 1064 1065 return 0; 1066 } 1067 1068 static int pci_pm_thaw(struct device *dev) 1069 { 1070 struct pci_dev *pci_dev = to_pci_dev(dev); 1071 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1072 int error = 0; 1073 1074 if (pci_has_legacy_pm_support(pci_dev)) 1075 return pci_legacy_resume(dev); 1076 1077 if (pm) { 1078 if (pm->thaw) 1079 error = pm->thaw(dev); 1080 } else { 1081 pci_pm_reenable_device(pci_dev); 1082 } 1083 1084 pci_dev->state_saved = false; 1085 1086 return error; 1087 } 1088 1089 static int pci_pm_poweroff(struct device *dev) 1090 { 1091 struct pci_dev *pci_dev = to_pci_dev(dev); 1092 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1093 1094 if (pci_has_legacy_pm_support(pci_dev)) 1095 return pci_legacy_suspend(dev, PMSG_HIBERNATE); 1096 1097 if (!pm) { 1098 pci_pm_default_suspend(pci_dev); 1099 return 0; 1100 } 1101 1102 /* The reason to do that is the same as in pci_pm_suspend(). */ 1103 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 1104 pci_dev_need_resume(pci_dev)) { 1105 pm_runtime_resume(dev); 1106 pci_dev->state_saved = false; 1107 } else { 1108 pci_dev_adjust_pme(pci_dev); 1109 } 1110 1111 if (pm->poweroff) { 1112 int error; 1113 1114 error = pm->poweroff(dev); 1115 suspend_report_result(pm->poweroff, error); 1116 if (error) 1117 return error; 1118 } 1119 1120 return 0; 1121 } 1122 1123 static int pci_pm_poweroff_late(struct device *dev) 1124 { 1125 if (dev_pm_skip_suspend(dev)) 1126 return 0; 1127 1128 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 1129 1130 return pm_generic_poweroff_late(dev); 1131 } 1132 1133 static int pci_pm_poweroff_noirq(struct device *dev) 1134 { 1135 struct pci_dev *pci_dev = to_pci_dev(dev); 1136 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1137 1138 if (dev_pm_skip_suspend(dev)) 1139 return 0; 1140 1141 if (pci_has_legacy_pm_support(pci_dev)) 1142 return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); 1143 1144 if (!pm) { 1145 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1146 return 0; 1147 } 1148 1149 if (pm->poweroff_noirq) { 1150 int error; 1151 1152 error = pm->poweroff_noirq(dev); 1153 suspend_report_result(pm->poweroff_noirq, error); 1154 if (error) 1155 return error; 1156 } 1157 1158 if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) 1159 pci_prepare_to_sleep(pci_dev); 1160 1161 /* 1162 * The reason for doing this here is the same as for the analogous code 1163 * in pci_pm_suspend_noirq(). 1164 */ 1165 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 1166 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 1167 1168 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1169 1170 return 0; 1171 } 1172 1173 static int pci_pm_restore_noirq(struct device *dev) 1174 { 1175 struct pci_dev *pci_dev = to_pci_dev(dev); 1176 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1177 1178 pci_pm_default_resume_early(pci_dev); 1179 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1180 1181 if (pci_has_legacy_pm_support(pci_dev)) 1182 return 0; 1183 1184 if (pm && pm->restore_noirq) 1185 return pm->restore_noirq(dev); 1186 1187 return 0; 1188 } 1189 1190 static int pci_pm_restore(struct device *dev) 1191 { 1192 struct pci_dev *pci_dev = to_pci_dev(dev); 1193 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1194 1195 /* 1196 * This is necessary for the hibernation error path in which restore is 1197 * called without restoring the standard config registers of the device. 1198 */ 1199 if (pci_dev->state_saved) 1200 pci_restore_standard_config(pci_dev); 1201 1202 if (pci_has_legacy_pm_support(pci_dev)) 1203 return pci_legacy_resume(dev); 1204 1205 pci_pm_default_resume(pci_dev); 1206 1207 if (pm) { 1208 if (pm->restore) 1209 return pm->restore(dev); 1210 } else { 1211 pci_pm_reenable_device(pci_dev); 1212 } 1213 1214 return 0; 1215 } 1216 1217 #else /* !CONFIG_HIBERNATE_CALLBACKS */ 1218 1219 #define pci_pm_freeze NULL 1220 #define pci_pm_freeze_noirq NULL 1221 #define pci_pm_thaw NULL 1222 #define pci_pm_thaw_noirq NULL 1223 #define pci_pm_poweroff NULL 1224 #define pci_pm_poweroff_late NULL 1225 #define pci_pm_poweroff_noirq NULL 1226 #define pci_pm_restore NULL 1227 #define pci_pm_restore_noirq NULL 1228 1229 #endif /* !CONFIG_HIBERNATE_CALLBACKS */ 1230 1231 #ifdef CONFIG_PM 1232 1233 static int pci_pm_runtime_suspend(struct device *dev) 1234 { 1235 struct pci_dev *pci_dev = to_pci_dev(dev); 1236 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1237 pci_power_t prev = pci_dev->current_state; 1238 int error; 1239 1240 /* 1241 * If the device has no driver, we leave it in D0, but it may go to 1242 * D3cold when the bridge above it runtime suspends. Save its 1243 * config space in case that happens. 1244 */ 1245 if (!to_pci_driver(dev->driver)) { 1246 pci_save_state(pci_dev); 1247 return 0; 1248 } 1249 1250 pci_dev->state_saved = false; 1251 if (pm && pm->runtime_suspend) { 1252 error = pm->runtime_suspend(dev); 1253 /* 1254 * -EBUSY and -EAGAIN is used to request the runtime PM core 1255 * to schedule a new suspend, so log the event only with debug 1256 * log level. 1257 */ 1258 if (error == -EBUSY || error == -EAGAIN) { 1259 pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", 1260 pm->runtime_suspend, error); 1261 return error; 1262 } else if (error) { 1263 pci_err(pci_dev, "can't suspend (%ps returned %d)\n", 1264 pm->runtime_suspend, error); 1265 return error; 1266 } 1267 } 1268 1269 pci_fixup_device(pci_fixup_suspend, pci_dev); 1270 1271 if (pm && pm->runtime_suspend 1272 && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 1273 && pci_dev->current_state != PCI_UNKNOWN) { 1274 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 1275 "PCI PM: State of device not saved by %pS\n", 1276 pm->runtime_suspend); 1277 return 0; 1278 } 1279 1280 if (!pci_dev->state_saved) { 1281 pci_save_state(pci_dev); 1282 pci_finish_runtime_suspend(pci_dev); 1283 } 1284 1285 return 0; 1286 } 1287 1288 static int pci_pm_runtime_resume(struct device *dev) 1289 { 1290 struct pci_dev *pci_dev = to_pci_dev(dev); 1291 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1292 pci_power_t prev_state = pci_dev->current_state; 1293 int error = 0; 1294 1295 /* 1296 * Restoring config space is necessary even if the device is not bound 1297 * to a driver because although we left it in D0, it may have gone to 1298 * D3cold when the bridge above it runtime suspended. 1299 */ 1300 pci_restore_standard_config(pci_dev); 1301 1302 if (!to_pci_driver(dev->driver)) 1303 return 0; 1304 1305 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1306 pci_pm_default_resume(pci_dev); 1307 1308 if (prev_state == PCI_D3cold) 1309 pci_bridge_wait_for_secondary_bus(pci_dev); 1310 1311 if (pm && pm->runtime_resume) 1312 error = pm->runtime_resume(dev); 1313 1314 pci_dev->runtime_d3cold = false; 1315 1316 return error; 1317 } 1318 1319 static int pci_pm_runtime_idle(struct device *dev) 1320 { 1321 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1322 1323 /* 1324 * If the device has no driver, it should always remain in D0 1325 * regardless of the runtime PM status 1326 */ 1327 if (!to_pci_driver(dev->driver)) 1328 return 0; 1329 1330 if (!pm) 1331 return -ENOSYS; 1332 1333 if (pm->runtime_idle) 1334 return pm->runtime_idle(dev); 1335 1336 return 0; 1337 } 1338 1339 static const struct dev_pm_ops pci_dev_pm_ops = { 1340 .prepare = pci_pm_prepare, 1341 .complete = pci_pm_complete, 1342 .suspend = pci_pm_suspend, 1343 .suspend_late = pci_pm_suspend_late, 1344 .resume = pci_pm_resume, 1345 .resume_early = pci_pm_resume_early, 1346 .freeze = pci_pm_freeze, 1347 .thaw = pci_pm_thaw, 1348 .poweroff = pci_pm_poweroff, 1349 .poweroff_late = pci_pm_poweroff_late, 1350 .restore = pci_pm_restore, 1351 .suspend_noirq = pci_pm_suspend_noirq, 1352 .resume_noirq = pci_pm_resume_noirq, 1353 .freeze_noirq = pci_pm_freeze_noirq, 1354 .thaw_noirq = pci_pm_thaw_noirq, 1355 .poweroff_noirq = pci_pm_poweroff_noirq, 1356 .restore_noirq = pci_pm_restore_noirq, 1357 .runtime_suspend = pci_pm_runtime_suspend, 1358 .runtime_resume = pci_pm_runtime_resume, 1359 .runtime_idle = pci_pm_runtime_idle, 1360 }; 1361 1362 #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) 1363 1364 #else /* !CONFIG_PM */ 1365 1366 #define pci_pm_runtime_suspend NULL 1367 #define pci_pm_runtime_resume NULL 1368 #define pci_pm_runtime_idle NULL 1369 1370 #define PCI_PM_OPS_PTR NULL 1371 1372 #endif /* !CONFIG_PM */ 1373 1374 /** 1375 * __pci_register_driver - register a new pci driver 1376 * @drv: the driver structure to register 1377 * @owner: owner module of drv 1378 * @mod_name: module name string 1379 * 1380 * Adds the driver structure to the list of registered drivers. 1381 * Returns a negative value on error, otherwise 0. 1382 * If no error occurred, the driver remains registered even if 1383 * no device was claimed during registration. 1384 */ 1385 int __pci_register_driver(struct pci_driver *drv, struct module *owner, 1386 const char *mod_name) 1387 { 1388 /* initialize common driver fields */ 1389 drv->driver.name = drv->name; 1390 drv->driver.bus = &pci_bus_type; 1391 drv->driver.owner = owner; 1392 drv->driver.mod_name = mod_name; 1393 drv->driver.groups = drv->groups; 1394 drv->driver.dev_groups = drv->dev_groups; 1395 1396 spin_lock_init(&drv->dynids.lock); 1397 INIT_LIST_HEAD(&drv->dynids.list); 1398 1399 /* register with core */ 1400 return driver_register(&drv->driver); 1401 } 1402 EXPORT_SYMBOL(__pci_register_driver); 1403 1404 /** 1405 * pci_unregister_driver - unregister a pci driver 1406 * @drv: the driver structure to unregister 1407 * 1408 * Deletes the driver structure from the list of registered PCI drivers, 1409 * gives it a chance to clean up by calling its remove() function for 1410 * each device it was responsible for, and marks those devices as 1411 * driverless. 1412 */ 1413 1414 void pci_unregister_driver(struct pci_driver *drv) 1415 { 1416 driver_unregister(&drv->driver); 1417 pci_free_dynids(drv); 1418 } 1419 EXPORT_SYMBOL(pci_unregister_driver); 1420 1421 static struct pci_driver pci_compat_driver = { 1422 .name = "compat" 1423 }; 1424 1425 /** 1426 * pci_dev_driver - get the pci_driver of a device 1427 * @dev: the device to query 1428 * 1429 * Returns the appropriate pci_driver structure or %NULL if there is no 1430 * registered driver for the device. 1431 */ 1432 struct pci_driver *pci_dev_driver(const struct pci_dev *dev) 1433 { 1434 struct pci_driver *drv = to_pci_driver(dev->dev.driver); 1435 1436 if (drv) 1437 return drv; 1438 else { 1439 int i; 1440 for (i = 0; i <= PCI_ROM_RESOURCE; i++) 1441 if (dev->resource[i].flags & IORESOURCE_BUSY) 1442 return &pci_compat_driver; 1443 } 1444 return NULL; 1445 } 1446 EXPORT_SYMBOL(pci_dev_driver); 1447 1448 /** 1449 * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure 1450 * @dev: the PCI device structure to match against 1451 * @drv: the device driver to search for matching PCI device id structures 1452 * 1453 * Used by a driver to check whether a PCI device present in the 1454 * system is in its list of supported devices. Returns the matching 1455 * pci_device_id structure or %NULL if there is no match. 1456 */ 1457 static int pci_bus_match(struct device *dev, struct device_driver *drv) 1458 { 1459 struct pci_dev *pci_dev = to_pci_dev(dev); 1460 struct pci_driver *pci_drv; 1461 const struct pci_device_id *found_id; 1462 1463 if (!pci_dev->match_driver) 1464 return 0; 1465 1466 pci_drv = to_pci_driver(drv); 1467 found_id = pci_match_device(pci_drv, pci_dev); 1468 if (found_id) 1469 return 1; 1470 1471 return 0; 1472 } 1473 1474 /** 1475 * pci_dev_get - increments the reference count of the pci device structure 1476 * @dev: the device being referenced 1477 * 1478 * Each live reference to a device should be refcounted. 1479 * 1480 * Drivers for PCI devices should normally record such references in 1481 * their probe() methods, when they bind to a device, and release 1482 * them by calling pci_dev_put(), in their disconnect() methods. 1483 * 1484 * A pointer to the device with the incremented reference counter is returned. 1485 */ 1486 struct pci_dev *pci_dev_get(struct pci_dev *dev) 1487 { 1488 if (dev) 1489 get_device(&dev->dev); 1490 return dev; 1491 } 1492 EXPORT_SYMBOL(pci_dev_get); 1493 1494 /** 1495 * pci_dev_put - release a use of the pci device structure 1496 * @dev: device that's been disconnected 1497 * 1498 * Must be called when a user of a device is finished with it. When the last 1499 * user of the device calls this function, the memory of the device is freed. 1500 */ 1501 void pci_dev_put(struct pci_dev *dev) 1502 { 1503 if (dev) 1504 put_device(&dev->dev); 1505 } 1506 EXPORT_SYMBOL(pci_dev_put); 1507 1508 static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) 1509 { 1510 struct pci_dev *pdev; 1511 1512 if (!dev) 1513 return -ENODEV; 1514 1515 pdev = to_pci_dev(dev); 1516 1517 if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) 1518 return -ENOMEM; 1519 1520 if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) 1521 return -ENOMEM; 1522 1523 if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, 1524 pdev->subsystem_device)) 1525 return -ENOMEM; 1526 1527 if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) 1528 return -ENOMEM; 1529 1530 if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", 1531 pdev->vendor, pdev->device, 1532 pdev->subsystem_vendor, pdev->subsystem_device, 1533 (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), 1534 (u8)(pdev->class))) 1535 return -ENOMEM; 1536 1537 return 0; 1538 } 1539 1540 #if defined(CONFIG_PCIEAER) || defined(CONFIG_EEH) 1541 /** 1542 * pci_uevent_ers - emit a uevent during recovery path of PCI device 1543 * @pdev: PCI device undergoing error recovery 1544 * @err_type: type of error event 1545 */ 1546 void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) 1547 { 1548 int idx = 0; 1549 char *envp[3]; 1550 1551 switch (err_type) { 1552 case PCI_ERS_RESULT_NONE: 1553 case PCI_ERS_RESULT_CAN_RECOVER: 1554 envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; 1555 envp[idx++] = "DEVICE_ONLINE=0"; 1556 break; 1557 case PCI_ERS_RESULT_RECOVERED: 1558 envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; 1559 envp[idx++] = "DEVICE_ONLINE=1"; 1560 break; 1561 case PCI_ERS_RESULT_DISCONNECT: 1562 envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; 1563 envp[idx++] = "DEVICE_ONLINE=0"; 1564 break; 1565 default: 1566 break; 1567 } 1568 1569 if (idx > 0) { 1570 envp[idx++] = NULL; 1571 kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); 1572 } 1573 } 1574 #endif 1575 1576 static int pci_bus_num_vf(struct device *dev) 1577 { 1578 return pci_num_vf(to_pci_dev(dev)); 1579 } 1580 1581 /** 1582 * pci_dma_configure - Setup DMA configuration 1583 * @dev: ptr to dev structure 1584 * 1585 * Function to update PCI devices's DMA configuration using the same 1586 * info from the OF node or ACPI node of host bridge's parent (if any). 1587 */ 1588 static int pci_dma_configure(struct device *dev) 1589 { 1590 struct device *bridge; 1591 int ret = 0; 1592 1593 bridge = pci_get_host_bridge_device(to_pci_dev(dev)); 1594 1595 if (IS_ENABLED(CONFIG_OF) && bridge->parent && 1596 bridge->parent->of_node) { 1597 ret = of_dma_configure(dev, bridge->parent->of_node, true); 1598 } else if (has_acpi_companion(bridge)) { 1599 struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); 1600 1601 ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); 1602 } 1603 1604 pci_put_host_bridge_device(bridge); 1605 return ret; 1606 } 1607 1608 struct bus_type pci_bus_type = { 1609 .name = "pci", 1610 .match = pci_bus_match, 1611 .uevent = pci_uevent, 1612 .probe = pci_device_probe, 1613 .remove = pci_device_remove, 1614 .shutdown = pci_device_shutdown, 1615 .dev_groups = pci_dev_groups, 1616 .bus_groups = pci_bus_groups, 1617 .drv_groups = pci_drv_groups, 1618 .pm = PCI_PM_OPS_PTR, 1619 .num_vf = pci_bus_num_vf, 1620 .dma_configure = pci_dma_configure, 1621 }; 1622 EXPORT_SYMBOL(pci_bus_type); 1623 1624 #ifdef CONFIG_PCIEPORTBUS 1625 static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) 1626 { 1627 struct pcie_device *pciedev; 1628 struct pcie_port_service_driver *driver; 1629 1630 if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) 1631 return 0; 1632 1633 pciedev = to_pcie_device(dev); 1634 driver = to_service_driver(drv); 1635 1636 if (driver->service != pciedev->service) 1637 return 0; 1638 1639 if (driver->port_type != PCIE_ANY_PORT && 1640 driver->port_type != pci_pcie_type(pciedev->port)) 1641 return 0; 1642 1643 return 1; 1644 } 1645 1646 struct bus_type pcie_port_bus_type = { 1647 .name = "pci_express", 1648 .match = pcie_port_bus_match, 1649 }; 1650 EXPORT_SYMBOL_GPL(pcie_port_bus_type); 1651 #endif 1652 1653 static int __init pci_driver_init(void) 1654 { 1655 int ret; 1656 1657 ret = bus_register(&pci_bus_type); 1658 if (ret) 1659 return ret; 1660 1661 #ifdef CONFIG_PCIEPORTBUS 1662 ret = bus_register(&pcie_port_bus_type); 1663 if (ret) 1664 return ret; 1665 #endif 1666 dma_debug_add_bus(&pci_bus_type); 1667 return 0; 1668 } 1669 postcore_initcall(pci_driver_init); 1670