1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> 4 * (C) Copyright 2007 Novell Inc. 5 */ 6 7 #include <linux/pci.h> 8 #include <linux/module.h> 9 #include <linux/init.h> 10 #include <linux/device.h> 11 #include <linux/mempolicy.h> 12 #include <linux/string.h> 13 #include <linux/slab.h> 14 #include <linux/sched.h> 15 #include <linux/sched/isolation.h> 16 #include <linux/cpu.h> 17 #include <linux/pm_runtime.h> 18 #include <linux/suspend.h> 19 #include <linux/kexec.h> 20 #include <linux/of_device.h> 21 #include <linux/acpi.h> 22 #include <linux/dma-map-ops.h> 23 #include "pci.h" 24 #include "pcie/portdrv.h" 25 26 struct pci_dynid { 27 struct list_head node; 28 struct pci_device_id id; 29 }; 30 31 /** 32 * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices 33 * @drv: target pci driver 34 * @vendor: PCI vendor ID 35 * @device: PCI device ID 36 * @subvendor: PCI subvendor ID 37 * @subdevice: PCI subdevice ID 38 * @class: PCI class 39 * @class_mask: PCI class mask 40 * @driver_data: private driver data 41 * 42 * Adds a new dynamic pci device ID to this driver and causes the 43 * driver to probe for all devices again. @drv must have been 44 * registered prior to calling this function. 45 * 46 * CONTEXT: 47 * Does GFP_KERNEL allocation. 48 * 49 * RETURNS: 50 * 0 on success, -errno on failure. 51 */ 52 int pci_add_dynid(struct pci_driver *drv, 53 unsigned int vendor, unsigned int device, 54 unsigned int subvendor, unsigned int subdevice, 55 unsigned int class, unsigned int class_mask, 56 unsigned long driver_data) 57 { 58 struct pci_dynid *dynid; 59 60 dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); 61 if (!dynid) 62 return -ENOMEM; 63 64 dynid->id.vendor = vendor; 65 dynid->id.device = device; 66 dynid->id.subvendor = subvendor; 67 dynid->id.subdevice = subdevice; 68 dynid->id.class = class; 69 dynid->id.class_mask = class_mask; 70 dynid->id.driver_data = driver_data; 71 72 spin_lock(&drv->dynids.lock); 73 list_add_tail(&dynid->node, &drv->dynids.list); 74 spin_unlock(&drv->dynids.lock); 75 76 return driver_attach(&drv->driver); 77 } 78 EXPORT_SYMBOL_GPL(pci_add_dynid); 79 80 static void pci_free_dynids(struct pci_driver *drv) 81 { 82 struct pci_dynid *dynid, *n; 83 84 spin_lock(&drv->dynids.lock); 85 list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { 86 list_del(&dynid->node); 87 kfree(dynid); 88 } 89 spin_unlock(&drv->dynids.lock); 90 } 91 92 /** 93 * pci_match_id - See if a PCI device matches a given pci_id table 94 * @ids: array of PCI device ID structures to search in 95 * @dev: the PCI device structure to match against. 96 * 97 * Used by a driver to check whether a PCI device is in its list of 98 * supported devices. Returns the matching pci_device_id structure or 99 * %NULL if there is no match. 100 * 101 * Deprecated; don't use this as it will not catch any dynamic IDs 102 * that a driver might want to check for. 103 */ 104 const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, 105 struct pci_dev *dev) 106 { 107 if (ids) { 108 while (ids->vendor || ids->subvendor || ids->class_mask) { 109 if (pci_match_one_device(ids, dev)) 110 return ids; 111 ids++; 112 } 113 } 114 return NULL; 115 } 116 EXPORT_SYMBOL(pci_match_id); 117 118 static const struct pci_device_id pci_device_id_any = { 119 .vendor = PCI_ANY_ID, 120 .device = PCI_ANY_ID, 121 .subvendor = PCI_ANY_ID, 122 .subdevice = PCI_ANY_ID, 123 }; 124 125 /** 126 * pci_match_device - See if a device matches a driver's list of IDs 127 * @drv: the PCI driver to match against 128 * @dev: the PCI device structure to match against 129 * 130 * Used by a driver to check whether a PCI device is in its list of 131 * supported devices or in the dynids list, which may have been augmented 132 * via the sysfs "new_id" file. Returns the matching pci_device_id 133 * structure or %NULL if there is no match. 134 */ 135 static const struct pci_device_id *pci_match_device(struct pci_driver *drv, 136 struct pci_dev *dev) 137 { 138 struct pci_dynid *dynid; 139 const struct pci_device_id *found_id = NULL, *ids; 140 141 /* When driver_override is set, only bind to the matching driver */ 142 if (dev->driver_override && strcmp(dev->driver_override, drv->name)) 143 return NULL; 144 145 /* Look at the dynamic ids first, before the static ones */ 146 spin_lock(&drv->dynids.lock); 147 list_for_each_entry(dynid, &drv->dynids.list, node) { 148 if (pci_match_one_device(&dynid->id, dev)) { 149 found_id = &dynid->id; 150 break; 151 } 152 } 153 spin_unlock(&drv->dynids.lock); 154 155 if (found_id) 156 return found_id; 157 158 for (ids = drv->id_table; (found_id = pci_match_id(ids, dev)); 159 ids = found_id + 1) { 160 /* 161 * The match table is split based on driver_override. 162 * In case override_only was set, enforce driver_override 163 * matching. 164 */ 165 if (found_id->override_only) { 166 if (dev->driver_override) 167 return found_id; 168 } else { 169 return found_id; 170 } 171 } 172 173 /* driver_override will always match, send a dummy id */ 174 if (dev->driver_override) 175 return &pci_device_id_any; 176 return NULL; 177 } 178 179 /** 180 * new_id_store - sysfs frontend to pci_add_dynid() 181 * @driver: target device driver 182 * @buf: buffer for scanning device ID data 183 * @count: input size 184 * 185 * Allow PCI IDs to be added to an existing driver via sysfs. 186 */ 187 static ssize_t new_id_store(struct device_driver *driver, const char *buf, 188 size_t count) 189 { 190 struct pci_driver *pdrv = to_pci_driver(driver); 191 const struct pci_device_id *ids = pdrv->id_table; 192 u32 vendor, device, subvendor = PCI_ANY_ID, 193 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 194 unsigned long driver_data = 0; 195 int fields = 0; 196 int retval = 0; 197 198 fields = sscanf(buf, "%x %x %x %x %x %x %lx", 199 &vendor, &device, &subvendor, &subdevice, 200 &class, &class_mask, &driver_data); 201 if (fields < 2) 202 return -EINVAL; 203 204 if (fields != 7) { 205 struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); 206 if (!pdev) 207 return -ENOMEM; 208 209 pdev->vendor = vendor; 210 pdev->device = device; 211 pdev->subsystem_vendor = subvendor; 212 pdev->subsystem_device = subdevice; 213 pdev->class = class; 214 215 if (pci_match_device(pdrv, pdev)) 216 retval = -EEXIST; 217 218 kfree(pdev); 219 220 if (retval) 221 return retval; 222 } 223 224 /* Only accept driver_data values that match an existing id_table 225 entry */ 226 if (ids) { 227 retval = -EINVAL; 228 while (ids->vendor || ids->subvendor || ids->class_mask) { 229 if (driver_data == ids->driver_data) { 230 retval = 0; 231 break; 232 } 233 ids++; 234 } 235 if (retval) /* No match */ 236 return retval; 237 } 238 239 retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, 240 class, class_mask, driver_data); 241 if (retval) 242 return retval; 243 return count; 244 } 245 static DRIVER_ATTR_WO(new_id); 246 247 /** 248 * remove_id_store - remove a PCI device ID from this driver 249 * @driver: target device driver 250 * @buf: buffer for scanning device ID data 251 * @count: input size 252 * 253 * Removes a dynamic pci device ID to this driver. 254 */ 255 static ssize_t remove_id_store(struct device_driver *driver, const char *buf, 256 size_t count) 257 { 258 struct pci_dynid *dynid, *n; 259 struct pci_driver *pdrv = to_pci_driver(driver); 260 u32 vendor, device, subvendor = PCI_ANY_ID, 261 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 262 int fields = 0; 263 size_t retval = -ENODEV; 264 265 fields = sscanf(buf, "%x %x %x %x %x %x", 266 &vendor, &device, &subvendor, &subdevice, 267 &class, &class_mask); 268 if (fields < 2) 269 return -EINVAL; 270 271 spin_lock(&pdrv->dynids.lock); 272 list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { 273 struct pci_device_id *id = &dynid->id; 274 if ((id->vendor == vendor) && 275 (id->device == device) && 276 (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && 277 (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && 278 !((id->class ^ class) & class_mask)) { 279 list_del(&dynid->node); 280 kfree(dynid); 281 retval = count; 282 break; 283 } 284 } 285 spin_unlock(&pdrv->dynids.lock); 286 287 return retval; 288 } 289 static DRIVER_ATTR_WO(remove_id); 290 291 static struct attribute *pci_drv_attrs[] = { 292 &driver_attr_new_id.attr, 293 &driver_attr_remove_id.attr, 294 NULL, 295 }; 296 ATTRIBUTE_GROUPS(pci_drv); 297 298 struct drv_dev_and_id { 299 struct pci_driver *drv; 300 struct pci_dev *dev; 301 const struct pci_device_id *id; 302 }; 303 304 static long local_pci_probe(void *_ddi) 305 { 306 struct drv_dev_and_id *ddi = _ddi; 307 struct pci_dev *pci_dev = ddi->dev; 308 struct pci_driver *pci_drv = ddi->drv; 309 struct device *dev = &pci_dev->dev; 310 int rc; 311 312 /* 313 * Unbound PCI devices are always put in D0, regardless of 314 * runtime PM status. During probe, the device is set to 315 * active and the usage count is incremented. If the driver 316 * supports runtime PM, it should call pm_runtime_put_noidle(), 317 * or any other runtime PM helper function decrementing the usage 318 * count, in its probe routine and pm_runtime_get_noresume() in 319 * its remove routine. 320 */ 321 pm_runtime_get_sync(dev); 322 pci_dev->driver = pci_drv; 323 rc = pci_drv->probe(pci_dev, ddi->id); 324 if (!rc) 325 return rc; 326 if (rc < 0) { 327 pci_dev->driver = NULL; 328 pm_runtime_put_sync(dev); 329 return rc; 330 } 331 /* 332 * Probe function should return < 0 for failure, 0 for success 333 * Treat values > 0 as success, but warn. 334 */ 335 pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", 336 rc); 337 return 0; 338 } 339 340 static bool pci_physfn_is_probed(struct pci_dev *dev) 341 { 342 #ifdef CONFIG_PCI_IOV 343 return dev->is_virtfn && dev->physfn->is_probed; 344 #else 345 return false; 346 #endif 347 } 348 349 static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, 350 const struct pci_device_id *id) 351 { 352 int error, node, cpu; 353 int hk_flags = HK_FLAG_DOMAIN | HK_FLAG_WQ; 354 struct drv_dev_and_id ddi = { drv, dev, id }; 355 356 /* 357 * Execute driver initialization on node where the device is 358 * attached. This way the driver likely allocates its local memory 359 * on the right node. 360 */ 361 node = dev_to_node(&dev->dev); 362 dev->is_probed = 1; 363 364 cpu_hotplug_disable(); 365 366 /* 367 * Prevent nesting work_on_cpu() for the case where a Virtual Function 368 * device is probed from work_on_cpu() of the Physical device. 369 */ 370 if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || 371 pci_physfn_is_probed(dev)) 372 cpu = nr_cpu_ids; 373 else 374 cpu = cpumask_any_and(cpumask_of_node(node), 375 housekeeping_cpumask(hk_flags)); 376 377 if (cpu < nr_cpu_ids) 378 error = work_on_cpu(cpu, local_pci_probe, &ddi); 379 else 380 error = local_pci_probe(&ddi); 381 382 dev->is_probed = 0; 383 cpu_hotplug_enable(); 384 return error; 385 } 386 387 /** 388 * __pci_device_probe - check if a driver wants to claim a specific PCI device 389 * @drv: driver to call to check if it wants the PCI device 390 * @pci_dev: PCI device being probed 391 * 392 * returns 0 on success, else error. 393 * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. 394 */ 395 static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) 396 { 397 const struct pci_device_id *id; 398 int error = 0; 399 400 if (!pci_dev->driver && drv->probe) { 401 error = -ENODEV; 402 403 id = pci_match_device(drv, pci_dev); 404 if (id) 405 error = pci_call_probe(drv, pci_dev, id); 406 } 407 return error; 408 } 409 410 int __weak pcibios_alloc_irq(struct pci_dev *dev) 411 { 412 return 0; 413 } 414 415 void __weak pcibios_free_irq(struct pci_dev *dev) 416 { 417 } 418 419 #ifdef CONFIG_PCI_IOV 420 static inline bool pci_device_can_probe(struct pci_dev *pdev) 421 { 422 return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || 423 pdev->driver_override); 424 } 425 #else 426 static inline bool pci_device_can_probe(struct pci_dev *pdev) 427 { 428 return true; 429 } 430 #endif 431 432 static int pci_device_probe(struct device *dev) 433 { 434 int error; 435 struct pci_dev *pci_dev = to_pci_dev(dev); 436 struct pci_driver *drv = to_pci_driver(dev->driver); 437 438 if (!pci_device_can_probe(pci_dev)) 439 return -ENODEV; 440 441 pci_assign_irq(pci_dev); 442 443 error = pcibios_alloc_irq(pci_dev); 444 if (error < 0) 445 return error; 446 447 pci_dev_get(pci_dev); 448 error = __pci_device_probe(drv, pci_dev); 449 if (error) { 450 pcibios_free_irq(pci_dev); 451 pci_dev_put(pci_dev); 452 } 453 454 return error; 455 } 456 457 static int pci_device_remove(struct device *dev) 458 { 459 struct pci_dev *pci_dev = to_pci_dev(dev); 460 struct pci_driver *drv = pci_dev->driver; 461 462 if (drv) { 463 if (drv->remove) { 464 pm_runtime_get_sync(dev); 465 drv->remove(pci_dev); 466 pm_runtime_put_noidle(dev); 467 } 468 pcibios_free_irq(pci_dev); 469 pci_dev->driver = NULL; 470 pci_iov_remove(pci_dev); 471 } 472 473 /* Undo the runtime PM settings in local_pci_probe() */ 474 pm_runtime_put_sync(dev); 475 476 /* 477 * If the device is still on, set the power state as "unknown", 478 * since it might change by the next time we load the driver. 479 */ 480 if (pci_dev->current_state == PCI_D0) 481 pci_dev->current_state = PCI_UNKNOWN; 482 483 /* 484 * We would love to complain here if pci_dev->is_enabled is set, that 485 * the driver should have called pci_disable_device(), but the 486 * unfortunate fact is there are too many odd BIOS and bridge setups 487 * that don't like drivers doing that all of the time. 488 * Oh well, we can dream of sane hardware when we sleep, no matter how 489 * horrible the crap we have to deal with is when we are awake... 490 */ 491 492 pci_dev_put(pci_dev); 493 return 0; 494 } 495 496 static void pci_device_shutdown(struct device *dev) 497 { 498 struct pci_dev *pci_dev = to_pci_dev(dev); 499 struct pci_driver *drv = pci_dev->driver; 500 501 pm_runtime_resume(dev); 502 503 if (drv && drv->shutdown) 504 drv->shutdown(pci_dev); 505 506 /* 507 * If this is a kexec reboot, turn off Bus Master bit on the 508 * device to tell it to not continue to do DMA. Don't touch 509 * devices in D3cold or unknown states. 510 * If it is not a kexec reboot, firmware will hit the PCI 511 * devices with big hammer and stop their DMA any way. 512 */ 513 if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) 514 pci_clear_master(pci_dev); 515 } 516 517 #ifdef CONFIG_PM 518 519 /* Auxiliary functions used for system resume and run-time resume. */ 520 521 /** 522 * pci_restore_standard_config - restore standard config registers of PCI device 523 * @pci_dev: PCI device to handle 524 */ 525 static int pci_restore_standard_config(struct pci_dev *pci_dev) 526 { 527 pci_update_current_state(pci_dev, PCI_UNKNOWN); 528 529 if (pci_dev->current_state != PCI_D0) { 530 int error = pci_set_power_state(pci_dev, PCI_D0); 531 if (error) 532 return error; 533 } 534 535 pci_restore_state(pci_dev); 536 pci_pme_restore(pci_dev); 537 return 0; 538 } 539 540 static void pci_pm_default_resume(struct pci_dev *pci_dev) 541 { 542 pci_fixup_device(pci_fixup_resume, pci_dev); 543 pci_enable_wake(pci_dev, PCI_D0, false); 544 } 545 546 #endif 547 548 #ifdef CONFIG_PM_SLEEP 549 550 static void pci_pm_default_resume_early(struct pci_dev *pci_dev) 551 { 552 pci_power_up(pci_dev); 553 pci_update_current_state(pci_dev, PCI_D0); 554 pci_restore_state(pci_dev); 555 pci_pme_restore(pci_dev); 556 } 557 558 /* 559 * Default "suspend" method for devices that have no driver provided suspend, 560 * or not even a driver at all (second part). 561 */ 562 static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) 563 { 564 /* 565 * mark its power state as "unknown", since we don't know if 566 * e.g. the BIOS will change its device state when we suspend. 567 */ 568 if (pci_dev->current_state == PCI_D0) 569 pci_dev->current_state = PCI_UNKNOWN; 570 } 571 572 /* 573 * Default "resume" method for devices that have no driver provided resume, 574 * or not even a driver at all (second part). 575 */ 576 static int pci_pm_reenable_device(struct pci_dev *pci_dev) 577 { 578 int retval; 579 580 /* if the device was enabled before suspend, reenable */ 581 retval = pci_reenable_device(pci_dev); 582 /* 583 * if the device was busmaster before the suspend, make it busmaster 584 * again 585 */ 586 if (pci_dev->is_busmaster) 587 pci_set_master(pci_dev); 588 589 return retval; 590 } 591 592 static int pci_legacy_suspend(struct device *dev, pm_message_t state) 593 { 594 struct pci_dev *pci_dev = to_pci_dev(dev); 595 struct pci_driver *drv = pci_dev->driver; 596 597 if (drv && drv->suspend) { 598 pci_power_t prev = pci_dev->current_state; 599 int error; 600 601 error = drv->suspend(pci_dev, state); 602 suspend_report_result(drv->suspend, error); 603 if (error) 604 return error; 605 606 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 607 && pci_dev->current_state != PCI_UNKNOWN) { 608 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 609 "PCI PM: Device state not saved by %pS\n", 610 drv->suspend); 611 } 612 } 613 614 pci_fixup_device(pci_fixup_suspend, pci_dev); 615 616 return 0; 617 } 618 619 static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) 620 { 621 struct pci_dev *pci_dev = to_pci_dev(dev); 622 623 if (!pci_dev->state_saved) 624 pci_save_state(pci_dev); 625 626 pci_pm_set_unknown_state(pci_dev); 627 628 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 629 630 return 0; 631 } 632 633 static int pci_legacy_resume(struct device *dev) 634 { 635 struct pci_dev *pci_dev = to_pci_dev(dev); 636 struct pci_driver *drv = pci_dev->driver; 637 638 pci_fixup_device(pci_fixup_resume, pci_dev); 639 640 return drv && drv->resume ? 641 drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); 642 } 643 644 /* Auxiliary functions used by the new power management framework */ 645 646 static void pci_pm_default_suspend(struct pci_dev *pci_dev) 647 { 648 /* Disable non-bridge devices without PM support */ 649 if (!pci_has_subordinate(pci_dev)) 650 pci_disable_enabled_device(pci_dev); 651 } 652 653 static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) 654 { 655 struct pci_driver *drv = pci_dev->driver; 656 bool ret = drv && (drv->suspend || drv->resume); 657 658 /* 659 * Legacy PM support is used by default, so warn if the new framework is 660 * supported as well. Drivers are supposed to support either the 661 * former, or the latter, but not both at the same time. 662 */ 663 pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", 664 pci_dev->vendor, pci_dev->device); 665 666 return ret; 667 } 668 669 /* New power management framework */ 670 671 static int pci_pm_prepare(struct device *dev) 672 { 673 struct pci_dev *pci_dev = to_pci_dev(dev); 674 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 675 676 if (pm && pm->prepare) { 677 int error = pm->prepare(dev); 678 if (error < 0) 679 return error; 680 681 if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) 682 return 0; 683 } 684 if (pci_dev_need_resume(pci_dev)) 685 return 0; 686 687 /* 688 * The PME setting needs to be adjusted here in case the direct-complete 689 * optimization is used with respect to this device. 690 */ 691 pci_dev_adjust_pme(pci_dev); 692 return 1; 693 } 694 695 static void pci_pm_complete(struct device *dev) 696 { 697 struct pci_dev *pci_dev = to_pci_dev(dev); 698 699 pci_dev_complete_resume(pci_dev); 700 pm_generic_complete(dev); 701 702 /* Resume device if platform firmware has put it in reset-power-on */ 703 if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { 704 pci_power_t pre_sleep_state = pci_dev->current_state; 705 706 pci_refresh_power_state(pci_dev); 707 /* 708 * On platforms with ACPI this check may also trigger for 709 * devices sharing power resources if one of those power 710 * resources has been activated as a result of a change of the 711 * power state of another device sharing it. However, in that 712 * case it is also better to resume the device, in general. 713 */ 714 if (pci_dev->current_state < pre_sleep_state) 715 pm_request_resume(dev); 716 } 717 } 718 719 #else /* !CONFIG_PM_SLEEP */ 720 721 #define pci_pm_prepare NULL 722 #define pci_pm_complete NULL 723 724 #endif /* !CONFIG_PM_SLEEP */ 725 726 #ifdef CONFIG_SUSPEND 727 static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) 728 { 729 /* 730 * Some BIOSes forget to clear Root PME Status bits after system 731 * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. 732 * Clear those bits now just in case (shouldn't hurt). 733 */ 734 if (pci_is_pcie(pci_dev) && 735 (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || 736 pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) 737 pcie_clear_root_pme_status(pci_dev); 738 } 739 740 static int pci_pm_suspend(struct device *dev) 741 { 742 struct pci_dev *pci_dev = to_pci_dev(dev); 743 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 744 745 pci_dev->skip_bus_pm = false; 746 747 if (pci_has_legacy_pm_support(pci_dev)) 748 return pci_legacy_suspend(dev, PMSG_SUSPEND); 749 750 if (!pm) { 751 pci_pm_default_suspend(pci_dev); 752 return 0; 753 } 754 755 /* 756 * PCI devices suspended at run time may need to be resumed at this 757 * point, because in general it may be necessary to reconfigure them for 758 * system suspend. Namely, if the device is expected to wake up the 759 * system from the sleep state, it may have to be reconfigured for this 760 * purpose, or if the device is not expected to wake up the system from 761 * the sleep state, it should be prevented from signaling wakeup events 762 * going forward. 763 * 764 * Also if the driver of the device does not indicate that its system 765 * suspend callbacks can cope with runtime-suspended devices, it is 766 * better to resume the device from runtime suspend here. 767 */ 768 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 769 pci_dev_need_resume(pci_dev)) { 770 pm_runtime_resume(dev); 771 pci_dev->state_saved = false; 772 } else { 773 pci_dev_adjust_pme(pci_dev); 774 } 775 776 if (pm->suspend) { 777 pci_power_t prev = pci_dev->current_state; 778 int error; 779 780 error = pm->suspend(dev); 781 suspend_report_result(pm->suspend, error); 782 if (error) 783 return error; 784 785 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 786 && pci_dev->current_state != PCI_UNKNOWN) { 787 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 788 "PCI PM: State of device not saved by %pS\n", 789 pm->suspend); 790 } 791 } 792 793 return 0; 794 } 795 796 static int pci_pm_suspend_late(struct device *dev) 797 { 798 if (dev_pm_skip_suspend(dev)) 799 return 0; 800 801 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 802 803 return pm_generic_suspend_late(dev); 804 } 805 806 static int pci_pm_suspend_noirq(struct device *dev) 807 { 808 struct pci_dev *pci_dev = to_pci_dev(dev); 809 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 810 811 if (dev_pm_skip_suspend(dev)) 812 return 0; 813 814 if (pci_has_legacy_pm_support(pci_dev)) 815 return pci_legacy_suspend_late(dev, PMSG_SUSPEND); 816 817 if (!pm) { 818 pci_save_state(pci_dev); 819 goto Fixup; 820 } 821 822 if (pm->suspend_noirq) { 823 pci_power_t prev = pci_dev->current_state; 824 int error; 825 826 error = pm->suspend_noirq(dev); 827 suspend_report_result(pm->suspend_noirq, error); 828 if (error) 829 return error; 830 831 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 832 && pci_dev->current_state != PCI_UNKNOWN) { 833 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 834 "PCI PM: State of device not saved by %pS\n", 835 pm->suspend_noirq); 836 goto Fixup; 837 } 838 } 839 840 if (pci_dev->skip_bus_pm) { 841 /* 842 * Either the device is a bridge with a child in D0 below it, or 843 * the function is running for the second time in a row without 844 * going through full resume, which is possible only during 845 * suspend-to-idle in a spurious wakeup case. The device should 846 * be in D0 at this point, but if it is a bridge, it may be 847 * necessary to save its state. 848 */ 849 if (!pci_dev->state_saved) 850 pci_save_state(pci_dev); 851 } else if (!pci_dev->state_saved) { 852 pci_save_state(pci_dev); 853 if (pci_power_manageable(pci_dev)) 854 pci_prepare_to_sleep(pci_dev); 855 } 856 857 pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", 858 pci_power_name(pci_dev->current_state)); 859 860 if (pci_dev->current_state == PCI_D0) { 861 pci_dev->skip_bus_pm = true; 862 /* 863 * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any 864 * downstream device is in D0, so avoid changing the power state 865 * of the parent bridge by setting the skip_bus_pm flag for it. 866 */ 867 if (pci_dev->bus->self) 868 pci_dev->bus->self->skip_bus_pm = true; 869 } 870 871 if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { 872 pci_dbg(pci_dev, "PCI PM: Skipped\n"); 873 goto Fixup; 874 } 875 876 pci_pm_set_unknown_state(pci_dev); 877 878 /* 879 * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's 880 * PCI COMMAND register isn't 0, the BIOS assumes that the controller 881 * hasn't been quiesced and tries to turn it off. If the controller 882 * is already in D3, this can hang or cause memory corruption. 883 * 884 * Since the value of the COMMAND register doesn't matter once the 885 * device has been suspended, we can safely set it to 0 here. 886 */ 887 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 888 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 889 890 Fixup: 891 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 892 893 /* 894 * If the target system sleep state is suspend-to-idle, it is sufficient 895 * to check whether or not the device's wakeup settings are good for 896 * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause 897 * pci_pm_complete() to take care of fixing up the device's state 898 * anyway, if need be. 899 */ 900 if (device_can_wakeup(dev) && !device_may_wakeup(dev)) 901 dev->power.may_skip_resume = false; 902 903 return 0; 904 } 905 906 static int pci_pm_resume_noirq(struct device *dev) 907 { 908 struct pci_dev *pci_dev = to_pci_dev(dev); 909 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 910 pci_power_t prev_state = pci_dev->current_state; 911 bool skip_bus_pm = pci_dev->skip_bus_pm; 912 913 if (dev_pm_skip_resume(dev)) 914 return 0; 915 916 /* 917 * In the suspend-to-idle case, devices left in D0 during suspend will 918 * stay in D0, so it is not necessary to restore or update their 919 * configuration here and attempting to put them into D0 again is 920 * pointless, so avoid doing that. 921 */ 922 if (!(skip_bus_pm && pm_suspend_no_platform())) 923 pci_pm_default_resume_early(pci_dev); 924 925 pci_fixup_device(pci_fixup_resume_early, pci_dev); 926 pcie_pme_root_status_cleanup(pci_dev); 927 928 if (!skip_bus_pm && prev_state == PCI_D3cold) 929 pci_bridge_wait_for_secondary_bus(pci_dev); 930 931 if (pci_has_legacy_pm_support(pci_dev)) 932 return 0; 933 934 if (pm && pm->resume_noirq) 935 return pm->resume_noirq(dev); 936 937 return 0; 938 } 939 940 static int pci_pm_resume_early(struct device *dev) 941 { 942 if (dev_pm_skip_resume(dev)) 943 return 0; 944 945 return pm_generic_resume_early(dev); 946 } 947 948 static int pci_pm_resume(struct device *dev) 949 { 950 struct pci_dev *pci_dev = to_pci_dev(dev); 951 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 952 953 /* 954 * This is necessary for the suspend error path in which resume is 955 * called without restoring the standard config registers of the device. 956 */ 957 if (pci_dev->state_saved) 958 pci_restore_standard_config(pci_dev); 959 960 if (pci_has_legacy_pm_support(pci_dev)) 961 return pci_legacy_resume(dev); 962 963 pci_pm_default_resume(pci_dev); 964 965 if (pm) { 966 if (pm->resume) 967 return pm->resume(dev); 968 } else { 969 pci_pm_reenable_device(pci_dev); 970 } 971 972 return 0; 973 } 974 975 #else /* !CONFIG_SUSPEND */ 976 977 #define pci_pm_suspend NULL 978 #define pci_pm_suspend_late NULL 979 #define pci_pm_suspend_noirq NULL 980 #define pci_pm_resume NULL 981 #define pci_pm_resume_early NULL 982 #define pci_pm_resume_noirq NULL 983 984 #endif /* !CONFIG_SUSPEND */ 985 986 #ifdef CONFIG_HIBERNATE_CALLBACKS 987 988 static int pci_pm_freeze(struct device *dev) 989 { 990 struct pci_dev *pci_dev = to_pci_dev(dev); 991 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 992 993 if (pci_has_legacy_pm_support(pci_dev)) 994 return pci_legacy_suspend(dev, PMSG_FREEZE); 995 996 if (!pm) { 997 pci_pm_default_suspend(pci_dev); 998 return 0; 999 } 1000 1001 /* 1002 * Resume all runtime-suspended devices before creating a snapshot 1003 * image of system memory, because the restore kernel generally cannot 1004 * be expected to always handle them consistently and they need to be 1005 * put into the runtime-active metastate during system resume anyway, 1006 * so it is better to ensure that the state saved in the image will be 1007 * always consistent with that. 1008 */ 1009 pm_runtime_resume(dev); 1010 pci_dev->state_saved = false; 1011 1012 if (pm->freeze) { 1013 int error; 1014 1015 error = pm->freeze(dev); 1016 suspend_report_result(pm->freeze, error); 1017 if (error) 1018 return error; 1019 } 1020 1021 return 0; 1022 } 1023 1024 static int pci_pm_freeze_noirq(struct device *dev) 1025 { 1026 struct pci_dev *pci_dev = to_pci_dev(dev); 1027 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1028 1029 if (pci_has_legacy_pm_support(pci_dev)) 1030 return pci_legacy_suspend_late(dev, PMSG_FREEZE); 1031 1032 if (pm && pm->freeze_noirq) { 1033 int error; 1034 1035 error = pm->freeze_noirq(dev); 1036 suspend_report_result(pm->freeze_noirq, error); 1037 if (error) 1038 return error; 1039 } 1040 1041 if (!pci_dev->state_saved) 1042 pci_save_state(pci_dev); 1043 1044 pci_pm_set_unknown_state(pci_dev); 1045 1046 return 0; 1047 } 1048 1049 static int pci_pm_thaw_noirq(struct device *dev) 1050 { 1051 struct pci_dev *pci_dev = to_pci_dev(dev); 1052 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1053 1054 /* 1055 * The pm->thaw_noirq() callback assumes the device has been 1056 * returned to D0 and its config state has been restored. 1057 * 1058 * In addition, pci_restore_state() restores MSI-X state in MMIO 1059 * space, which requires the device to be in D0, so return it to D0 1060 * in case the driver's "freeze" callbacks put it into a low-power 1061 * state. 1062 */ 1063 pci_set_power_state(pci_dev, PCI_D0); 1064 pci_restore_state(pci_dev); 1065 1066 if (pci_has_legacy_pm_support(pci_dev)) 1067 return 0; 1068 1069 if (pm && pm->thaw_noirq) 1070 return pm->thaw_noirq(dev); 1071 1072 return 0; 1073 } 1074 1075 static int pci_pm_thaw(struct device *dev) 1076 { 1077 struct pci_dev *pci_dev = to_pci_dev(dev); 1078 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1079 int error = 0; 1080 1081 if (pci_has_legacy_pm_support(pci_dev)) 1082 return pci_legacy_resume(dev); 1083 1084 if (pm) { 1085 if (pm->thaw) 1086 error = pm->thaw(dev); 1087 } else { 1088 pci_pm_reenable_device(pci_dev); 1089 } 1090 1091 pci_dev->state_saved = false; 1092 1093 return error; 1094 } 1095 1096 static int pci_pm_poweroff(struct device *dev) 1097 { 1098 struct pci_dev *pci_dev = to_pci_dev(dev); 1099 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1100 1101 if (pci_has_legacy_pm_support(pci_dev)) 1102 return pci_legacy_suspend(dev, PMSG_HIBERNATE); 1103 1104 if (!pm) { 1105 pci_pm_default_suspend(pci_dev); 1106 return 0; 1107 } 1108 1109 /* The reason to do that is the same as in pci_pm_suspend(). */ 1110 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 1111 pci_dev_need_resume(pci_dev)) { 1112 pm_runtime_resume(dev); 1113 pci_dev->state_saved = false; 1114 } else { 1115 pci_dev_adjust_pme(pci_dev); 1116 } 1117 1118 if (pm->poweroff) { 1119 int error; 1120 1121 error = pm->poweroff(dev); 1122 suspend_report_result(pm->poweroff, error); 1123 if (error) 1124 return error; 1125 } 1126 1127 return 0; 1128 } 1129 1130 static int pci_pm_poweroff_late(struct device *dev) 1131 { 1132 if (dev_pm_skip_suspend(dev)) 1133 return 0; 1134 1135 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 1136 1137 return pm_generic_poweroff_late(dev); 1138 } 1139 1140 static int pci_pm_poweroff_noirq(struct device *dev) 1141 { 1142 struct pci_dev *pci_dev = to_pci_dev(dev); 1143 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1144 1145 if (dev_pm_skip_suspend(dev)) 1146 return 0; 1147 1148 if (pci_has_legacy_pm_support(pci_dev)) 1149 return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); 1150 1151 if (!pm) { 1152 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1153 return 0; 1154 } 1155 1156 if (pm->poweroff_noirq) { 1157 int error; 1158 1159 error = pm->poweroff_noirq(dev); 1160 suspend_report_result(pm->poweroff_noirq, error); 1161 if (error) 1162 return error; 1163 } 1164 1165 if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) 1166 pci_prepare_to_sleep(pci_dev); 1167 1168 /* 1169 * The reason for doing this here is the same as for the analogous code 1170 * in pci_pm_suspend_noirq(). 1171 */ 1172 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 1173 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 1174 1175 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1176 1177 return 0; 1178 } 1179 1180 static int pci_pm_restore_noirq(struct device *dev) 1181 { 1182 struct pci_dev *pci_dev = to_pci_dev(dev); 1183 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1184 1185 pci_pm_default_resume_early(pci_dev); 1186 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1187 1188 if (pci_has_legacy_pm_support(pci_dev)) 1189 return 0; 1190 1191 if (pm && pm->restore_noirq) 1192 return pm->restore_noirq(dev); 1193 1194 return 0; 1195 } 1196 1197 static int pci_pm_restore(struct device *dev) 1198 { 1199 struct pci_dev *pci_dev = to_pci_dev(dev); 1200 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1201 1202 /* 1203 * This is necessary for the hibernation error path in which restore is 1204 * called without restoring the standard config registers of the device. 1205 */ 1206 if (pci_dev->state_saved) 1207 pci_restore_standard_config(pci_dev); 1208 1209 if (pci_has_legacy_pm_support(pci_dev)) 1210 return pci_legacy_resume(dev); 1211 1212 pci_pm_default_resume(pci_dev); 1213 1214 if (pm) { 1215 if (pm->restore) 1216 return pm->restore(dev); 1217 } else { 1218 pci_pm_reenable_device(pci_dev); 1219 } 1220 1221 return 0; 1222 } 1223 1224 #else /* !CONFIG_HIBERNATE_CALLBACKS */ 1225 1226 #define pci_pm_freeze NULL 1227 #define pci_pm_freeze_noirq NULL 1228 #define pci_pm_thaw NULL 1229 #define pci_pm_thaw_noirq NULL 1230 #define pci_pm_poweroff NULL 1231 #define pci_pm_poweroff_late NULL 1232 #define pci_pm_poweroff_noirq NULL 1233 #define pci_pm_restore NULL 1234 #define pci_pm_restore_noirq NULL 1235 1236 #endif /* !CONFIG_HIBERNATE_CALLBACKS */ 1237 1238 #ifdef CONFIG_PM 1239 1240 static int pci_pm_runtime_suspend(struct device *dev) 1241 { 1242 struct pci_dev *pci_dev = to_pci_dev(dev); 1243 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1244 pci_power_t prev = pci_dev->current_state; 1245 int error; 1246 1247 /* 1248 * If pci_dev->driver is not set (unbound), we leave the device in D0, 1249 * but it may go to D3cold when the bridge above it runtime suspends. 1250 * Save its config space in case that happens. 1251 */ 1252 if (!pci_dev->driver) { 1253 pci_save_state(pci_dev); 1254 return 0; 1255 } 1256 1257 pci_dev->state_saved = false; 1258 if (pm && pm->runtime_suspend) { 1259 error = pm->runtime_suspend(dev); 1260 /* 1261 * -EBUSY and -EAGAIN is used to request the runtime PM core 1262 * to schedule a new suspend, so log the event only with debug 1263 * log level. 1264 */ 1265 if (error == -EBUSY || error == -EAGAIN) { 1266 pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", 1267 pm->runtime_suspend, error); 1268 return error; 1269 } else if (error) { 1270 pci_err(pci_dev, "can't suspend (%ps returned %d)\n", 1271 pm->runtime_suspend, error); 1272 return error; 1273 } 1274 } 1275 1276 pci_fixup_device(pci_fixup_suspend, pci_dev); 1277 1278 if (pm && pm->runtime_suspend 1279 && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 1280 && pci_dev->current_state != PCI_UNKNOWN) { 1281 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 1282 "PCI PM: State of device not saved by %pS\n", 1283 pm->runtime_suspend); 1284 return 0; 1285 } 1286 1287 if (!pci_dev->state_saved) { 1288 pci_save_state(pci_dev); 1289 pci_finish_runtime_suspend(pci_dev); 1290 } 1291 1292 return 0; 1293 } 1294 1295 static int pci_pm_runtime_resume(struct device *dev) 1296 { 1297 struct pci_dev *pci_dev = to_pci_dev(dev); 1298 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1299 pci_power_t prev_state = pci_dev->current_state; 1300 int error = 0; 1301 1302 /* 1303 * Restoring config space is necessary even if the device is not bound 1304 * to a driver because although we left it in D0, it may have gone to 1305 * D3cold when the bridge above it runtime suspended. 1306 */ 1307 pci_restore_standard_config(pci_dev); 1308 1309 if (!pci_dev->driver) 1310 return 0; 1311 1312 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1313 pci_pm_default_resume(pci_dev); 1314 1315 if (prev_state == PCI_D3cold) 1316 pci_bridge_wait_for_secondary_bus(pci_dev); 1317 1318 if (pm && pm->runtime_resume) 1319 error = pm->runtime_resume(dev); 1320 1321 pci_dev->runtime_d3cold = false; 1322 1323 return error; 1324 } 1325 1326 static int pci_pm_runtime_idle(struct device *dev) 1327 { 1328 struct pci_dev *pci_dev = to_pci_dev(dev); 1329 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1330 1331 /* 1332 * If pci_dev->driver is not set (unbound), the device should 1333 * always remain in D0 regardless of the runtime PM status 1334 */ 1335 if (!pci_dev->driver) 1336 return 0; 1337 1338 if (!pm) 1339 return -ENOSYS; 1340 1341 if (pm->runtime_idle) 1342 return pm->runtime_idle(dev); 1343 1344 return 0; 1345 } 1346 1347 static const struct dev_pm_ops pci_dev_pm_ops = { 1348 .prepare = pci_pm_prepare, 1349 .complete = pci_pm_complete, 1350 .suspend = pci_pm_suspend, 1351 .suspend_late = pci_pm_suspend_late, 1352 .resume = pci_pm_resume, 1353 .resume_early = pci_pm_resume_early, 1354 .freeze = pci_pm_freeze, 1355 .thaw = pci_pm_thaw, 1356 .poweroff = pci_pm_poweroff, 1357 .poweroff_late = pci_pm_poweroff_late, 1358 .restore = pci_pm_restore, 1359 .suspend_noirq = pci_pm_suspend_noirq, 1360 .resume_noirq = pci_pm_resume_noirq, 1361 .freeze_noirq = pci_pm_freeze_noirq, 1362 .thaw_noirq = pci_pm_thaw_noirq, 1363 .poweroff_noirq = pci_pm_poweroff_noirq, 1364 .restore_noirq = pci_pm_restore_noirq, 1365 .runtime_suspend = pci_pm_runtime_suspend, 1366 .runtime_resume = pci_pm_runtime_resume, 1367 .runtime_idle = pci_pm_runtime_idle, 1368 }; 1369 1370 #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) 1371 1372 #else /* !CONFIG_PM */ 1373 1374 #define pci_pm_runtime_suspend NULL 1375 #define pci_pm_runtime_resume NULL 1376 #define pci_pm_runtime_idle NULL 1377 1378 #define PCI_PM_OPS_PTR NULL 1379 1380 #endif /* !CONFIG_PM */ 1381 1382 /** 1383 * __pci_register_driver - register a new pci driver 1384 * @drv: the driver structure to register 1385 * @owner: owner module of drv 1386 * @mod_name: module name string 1387 * 1388 * Adds the driver structure to the list of registered drivers. 1389 * Returns a negative value on error, otherwise 0. 1390 * If no error occurred, the driver remains registered even if 1391 * no device was claimed during registration. 1392 */ 1393 int __pci_register_driver(struct pci_driver *drv, struct module *owner, 1394 const char *mod_name) 1395 { 1396 /* initialize common driver fields */ 1397 drv->driver.name = drv->name; 1398 drv->driver.bus = &pci_bus_type; 1399 drv->driver.owner = owner; 1400 drv->driver.mod_name = mod_name; 1401 drv->driver.groups = drv->groups; 1402 drv->driver.dev_groups = drv->dev_groups; 1403 1404 spin_lock_init(&drv->dynids.lock); 1405 INIT_LIST_HEAD(&drv->dynids.list); 1406 1407 /* register with core */ 1408 return driver_register(&drv->driver); 1409 } 1410 EXPORT_SYMBOL(__pci_register_driver); 1411 1412 /** 1413 * pci_unregister_driver - unregister a pci driver 1414 * @drv: the driver structure to unregister 1415 * 1416 * Deletes the driver structure from the list of registered PCI drivers, 1417 * gives it a chance to clean up by calling its remove() function for 1418 * each device it was responsible for, and marks those devices as 1419 * driverless. 1420 */ 1421 1422 void pci_unregister_driver(struct pci_driver *drv) 1423 { 1424 driver_unregister(&drv->driver); 1425 pci_free_dynids(drv); 1426 } 1427 EXPORT_SYMBOL(pci_unregister_driver); 1428 1429 static struct pci_driver pci_compat_driver = { 1430 .name = "compat" 1431 }; 1432 1433 /** 1434 * pci_dev_driver - get the pci_driver of a device 1435 * @dev: the device to query 1436 * 1437 * Returns the appropriate pci_driver structure or %NULL if there is no 1438 * registered driver for the device. 1439 */ 1440 struct pci_driver *pci_dev_driver(const struct pci_dev *dev) 1441 { 1442 if (dev->driver) 1443 return dev->driver; 1444 else { 1445 int i; 1446 for (i = 0; i <= PCI_ROM_RESOURCE; i++) 1447 if (dev->resource[i].flags & IORESOURCE_BUSY) 1448 return &pci_compat_driver; 1449 } 1450 return NULL; 1451 } 1452 EXPORT_SYMBOL(pci_dev_driver); 1453 1454 /** 1455 * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure 1456 * @dev: the PCI device structure to match against 1457 * @drv: the device driver to search for matching PCI device id structures 1458 * 1459 * Used by a driver to check whether a PCI device present in the 1460 * system is in its list of supported devices. Returns the matching 1461 * pci_device_id structure or %NULL if there is no match. 1462 */ 1463 static int pci_bus_match(struct device *dev, struct device_driver *drv) 1464 { 1465 struct pci_dev *pci_dev = to_pci_dev(dev); 1466 struct pci_driver *pci_drv; 1467 const struct pci_device_id *found_id; 1468 1469 if (!pci_dev->match_driver) 1470 return 0; 1471 1472 pci_drv = to_pci_driver(drv); 1473 found_id = pci_match_device(pci_drv, pci_dev); 1474 if (found_id) 1475 return 1; 1476 1477 return 0; 1478 } 1479 1480 /** 1481 * pci_dev_get - increments the reference count of the pci device structure 1482 * @dev: the device being referenced 1483 * 1484 * Each live reference to a device should be refcounted. 1485 * 1486 * Drivers for PCI devices should normally record such references in 1487 * their probe() methods, when they bind to a device, and release 1488 * them by calling pci_dev_put(), in their disconnect() methods. 1489 * 1490 * A pointer to the device with the incremented reference counter is returned. 1491 */ 1492 struct pci_dev *pci_dev_get(struct pci_dev *dev) 1493 { 1494 if (dev) 1495 get_device(&dev->dev); 1496 return dev; 1497 } 1498 EXPORT_SYMBOL(pci_dev_get); 1499 1500 /** 1501 * pci_dev_put - release a use of the pci device structure 1502 * @dev: device that's been disconnected 1503 * 1504 * Must be called when a user of a device is finished with it. When the last 1505 * user of the device calls this function, the memory of the device is freed. 1506 */ 1507 void pci_dev_put(struct pci_dev *dev) 1508 { 1509 if (dev) 1510 put_device(&dev->dev); 1511 } 1512 EXPORT_SYMBOL(pci_dev_put); 1513 1514 static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) 1515 { 1516 struct pci_dev *pdev; 1517 1518 if (!dev) 1519 return -ENODEV; 1520 1521 pdev = to_pci_dev(dev); 1522 1523 if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) 1524 return -ENOMEM; 1525 1526 if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) 1527 return -ENOMEM; 1528 1529 if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, 1530 pdev->subsystem_device)) 1531 return -ENOMEM; 1532 1533 if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) 1534 return -ENOMEM; 1535 1536 if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", 1537 pdev->vendor, pdev->device, 1538 pdev->subsystem_vendor, pdev->subsystem_device, 1539 (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), 1540 (u8)(pdev->class))) 1541 return -ENOMEM; 1542 1543 return 0; 1544 } 1545 1546 #if defined(CONFIG_PCIEPORTBUS) || defined(CONFIG_EEH) 1547 /** 1548 * pci_uevent_ers - emit a uevent during recovery path of PCI device 1549 * @pdev: PCI device undergoing error recovery 1550 * @err_type: type of error event 1551 */ 1552 void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) 1553 { 1554 int idx = 0; 1555 char *envp[3]; 1556 1557 switch (err_type) { 1558 case PCI_ERS_RESULT_NONE: 1559 case PCI_ERS_RESULT_CAN_RECOVER: 1560 envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; 1561 envp[idx++] = "DEVICE_ONLINE=0"; 1562 break; 1563 case PCI_ERS_RESULT_RECOVERED: 1564 envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; 1565 envp[idx++] = "DEVICE_ONLINE=1"; 1566 break; 1567 case PCI_ERS_RESULT_DISCONNECT: 1568 envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; 1569 envp[idx++] = "DEVICE_ONLINE=0"; 1570 break; 1571 default: 1572 break; 1573 } 1574 1575 if (idx > 0) { 1576 envp[idx++] = NULL; 1577 kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); 1578 } 1579 } 1580 #endif 1581 1582 static int pci_bus_num_vf(struct device *dev) 1583 { 1584 return pci_num_vf(to_pci_dev(dev)); 1585 } 1586 1587 /** 1588 * pci_dma_configure - Setup DMA configuration 1589 * @dev: ptr to dev structure 1590 * 1591 * Function to update PCI devices's DMA configuration using the same 1592 * info from the OF node or ACPI node of host bridge's parent (if any). 1593 */ 1594 static int pci_dma_configure(struct device *dev) 1595 { 1596 struct device *bridge; 1597 int ret = 0; 1598 1599 bridge = pci_get_host_bridge_device(to_pci_dev(dev)); 1600 1601 if (IS_ENABLED(CONFIG_OF) && bridge->parent && 1602 bridge->parent->of_node) { 1603 ret = of_dma_configure(dev, bridge->parent->of_node, true); 1604 } else if (has_acpi_companion(bridge)) { 1605 struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); 1606 1607 ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); 1608 } 1609 1610 pci_put_host_bridge_device(bridge); 1611 return ret; 1612 } 1613 1614 struct bus_type pci_bus_type = { 1615 .name = "pci", 1616 .match = pci_bus_match, 1617 .uevent = pci_uevent, 1618 .probe = pci_device_probe, 1619 .remove = pci_device_remove, 1620 .shutdown = pci_device_shutdown, 1621 .dev_groups = pci_dev_groups, 1622 .bus_groups = pci_bus_groups, 1623 .drv_groups = pci_drv_groups, 1624 .pm = PCI_PM_OPS_PTR, 1625 .num_vf = pci_bus_num_vf, 1626 .dma_configure = pci_dma_configure, 1627 }; 1628 EXPORT_SYMBOL(pci_bus_type); 1629 1630 #ifdef CONFIG_PCIEPORTBUS 1631 static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) 1632 { 1633 struct pcie_device *pciedev; 1634 struct pcie_port_service_driver *driver; 1635 1636 if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) 1637 return 0; 1638 1639 pciedev = to_pcie_device(dev); 1640 driver = to_service_driver(drv); 1641 1642 if (driver->service != pciedev->service) 1643 return 0; 1644 1645 if (driver->port_type != PCIE_ANY_PORT && 1646 driver->port_type != pci_pcie_type(pciedev->port)) 1647 return 0; 1648 1649 return 1; 1650 } 1651 1652 struct bus_type pcie_port_bus_type = { 1653 .name = "pci_express", 1654 .match = pcie_port_bus_match, 1655 }; 1656 EXPORT_SYMBOL_GPL(pcie_port_bus_type); 1657 #endif 1658 1659 static int __init pci_driver_init(void) 1660 { 1661 int ret; 1662 1663 ret = bus_register(&pci_bus_type); 1664 if (ret) 1665 return ret; 1666 1667 #ifdef CONFIG_PCIEPORTBUS 1668 ret = bus_register(&pcie_port_bus_type); 1669 if (ret) 1670 return ret; 1671 #endif 1672 dma_debug_add_bus(&pci_bus_type); 1673 return 0; 1674 } 1675 postcore_initcall(pci_driver_init); 1676