1 /* 2 * Copyright 2008 Advanced Micro Devices, Inc. 3 * Copyright 2008 Red Hat Inc. 4 * Copyright 2009 Jerome Glisse. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the "Software"), 8 * to deal in the Software without restriction, including without limitation 9 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 10 * and/or sell copies of the Software, and to permit persons to whom the 11 * Software is furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 22 * OTHER DEALINGS IN THE SOFTWARE. 23 * 24 * Authors: Dave Airlie 25 * Alex Deucher 26 * Jerome Glisse 27 */ 28 29 #include <linux/aperture.h> 30 #include <linux/power_supply.h> 31 #include <linux/kthread.h> 32 #include <linux/module.h> 33 #include <linux/console.h> 34 #include <linux/slab.h> 35 #include <linux/iommu.h> 36 #include <linux/pci.h> 37 #include <linux/pci-p2pdma.h> 38 #include <linux/apple-gmux.h> 39 #include <linux/nospec.h> 40 41 #include <drm/drm_atomic_helper.h> 42 #include <drm/drm_client_event.h> 43 #include <drm/drm_crtc_helper.h> 44 #include <drm/drm_probe_helper.h> 45 #include <drm/amdgpu_drm.h> 46 #include <linux/device.h> 47 #include <linux/vgaarb.h> 48 #include <linux/vga_switcheroo.h> 49 #include <linux/efi.h> 50 #include "amdgpu.h" 51 #include "amdgpu_trace.h" 52 #include "amdgpu_i2c.h" 53 #include "atom.h" 54 #include "amdgpu_atombios.h" 55 #include "amdgpu_atomfirmware.h" 56 #include "amd_pcie.h" 57 #ifdef CONFIG_DRM_AMDGPU_SI 58 #include "si.h" 59 #endif 60 #ifdef CONFIG_DRM_AMDGPU_CIK 61 #include "cik.h" 62 #endif 63 #include "vi.h" 64 #include "soc15.h" 65 #include "nv.h" 66 #include "bif/bif_4_1_d.h" 67 #include <linux/firmware.h> 68 #include "amdgpu_vf_error.h" 69 70 #include "amdgpu_amdkfd.h" 71 #include "amdgpu_pm.h" 72 73 #include "amdgpu_xgmi.h" 74 #include "amdgpu_ras.h" 75 #include "amdgpu_ras_mgr.h" 76 #include "amdgpu_pmu.h" 77 #include "amdgpu_fru_eeprom.h" 78 #include "amdgpu_reset.h" 79 #include "amdgpu_virt.h" 80 #include "amdgpu_dev_coredump.h" 81 82 #include <linux/suspend.h> 83 #include <drm/task_barrier.h> 84 #include <linux/pm_runtime.h> 85 86 #include <drm/drm_drv.h> 87 88 #if IS_ENABLED(CONFIG_X86) 89 #include <asm/intel-family.h> 90 #include <asm/cpu_device_id.h> 91 #endif 92 93 MODULE_FIRMWARE("amdgpu/vega10_gpu_info.bin"); 94 MODULE_FIRMWARE("amdgpu/vega12_gpu_info.bin"); 95 MODULE_FIRMWARE("amdgpu/raven_gpu_info.bin"); 96 MODULE_FIRMWARE("amdgpu/picasso_gpu_info.bin"); 97 MODULE_FIRMWARE("amdgpu/raven2_gpu_info.bin"); 98 MODULE_FIRMWARE("amdgpu/arcturus_gpu_info.bin"); 99 MODULE_FIRMWARE("amdgpu/navi12_gpu_info.bin"); 100 MODULE_FIRMWARE("amdgpu/cyan_skillfish_gpu_info.bin"); 101 102 #define AMDGPU_RESUME_MS 2000 103 #define AMDGPU_MAX_RETRY_LIMIT 2 104 #define AMDGPU_RETRY_SRIOV_RESET(r) ((r) == -EBUSY || (r) == -ETIMEDOUT || (r) == -EINVAL) 105 #define AMDGPU_PCIE_INDEX_FALLBACK (0x38 >> 2) 106 #define AMDGPU_PCIE_INDEX_HI_FALLBACK (0x44 >> 2) 107 #define AMDGPU_PCIE_DATA_FALLBACK (0x3C >> 2) 108 109 #define AMDGPU_VBIOS_SKIP (1U << 0) 110 #define AMDGPU_VBIOS_OPTIONAL (1U << 1) 111 112 static const struct drm_driver amdgpu_kms_driver; 113 114 const char *amdgpu_asic_name[] = { 115 "TAHITI", 116 "PITCAIRN", 117 "VERDE", 118 "OLAND", 119 "HAINAN", 120 "BONAIRE", 121 "KAVERI", 122 "KABINI", 123 "HAWAII", 124 "MULLINS", 125 "TOPAZ", 126 "TONGA", 127 "FIJI", 128 "CARRIZO", 129 "STONEY", 130 "POLARIS10", 131 "POLARIS11", 132 "POLARIS12", 133 "VEGAM", 134 "VEGA10", 135 "VEGA12", 136 "VEGA20", 137 "RAVEN", 138 "ARCTURUS", 139 "RENOIR", 140 "ALDEBARAN", 141 "NAVI10", 142 "CYAN_SKILLFISH", 143 "NAVI14", 144 "NAVI12", 145 "SIENNA_CICHLID", 146 "NAVY_FLOUNDER", 147 "VANGOGH", 148 "DIMGREY_CAVEFISH", 149 "BEIGE_GOBY", 150 "YELLOW_CARP", 151 "IP DISCOVERY", 152 "LAST", 153 }; 154 155 #define AMDGPU_IP_BLK_MASK_ALL GENMASK(AMD_IP_BLOCK_TYPE_NUM - 1, 0) 156 /* 157 * Default init level where all blocks are expected to be initialized. This is 158 * the level of initialization expected by default and also after a full reset 159 * of the device. 160 */ 161 struct amdgpu_init_level amdgpu_init_default = { 162 .level = AMDGPU_INIT_LEVEL_DEFAULT, 163 .hwini_ip_block_mask = AMDGPU_IP_BLK_MASK_ALL, 164 }; 165 166 struct amdgpu_init_level amdgpu_init_recovery = { 167 .level = AMDGPU_INIT_LEVEL_RESET_RECOVERY, 168 .hwini_ip_block_mask = AMDGPU_IP_BLK_MASK_ALL, 169 }; 170 171 /* 172 * Minimal blocks needed to be initialized before a XGMI hive can be reset. This 173 * is used for cases like reset on initialization where the entire hive needs to 174 * be reset before first use. 175 */ 176 struct amdgpu_init_level amdgpu_init_minimal_xgmi = { 177 .level = AMDGPU_INIT_LEVEL_MINIMAL_XGMI, 178 .hwini_ip_block_mask = 179 BIT(AMD_IP_BLOCK_TYPE_GMC) | BIT(AMD_IP_BLOCK_TYPE_SMC) | 180 BIT(AMD_IP_BLOCK_TYPE_COMMON) | BIT(AMD_IP_BLOCK_TYPE_IH) | 181 BIT(AMD_IP_BLOCK_TYPE_PSP) 182 }; 183 184 static int amdgpu_device_ip_resume_phase1(struct amdgpu_device *adev); 185 static int amdgpu_device_ip_resume_phase2(struct amdgpu_device *adev); 186 static int amdgpu_device_ip_resume_phase3(struct amdgpu_device *adev); 187 188 static void amdgpu_device_load_switch_state(struct amdgpu_device *adev); 189 190 static inline bool amdgpu_ip_member_of_hwini(struct amdgpu_device *adev, 191 enum amd_ip_block_type block) 192 { 193 return (adev->init_lvl->hwini_ip_block_mask & (1U << block)) != 0; 194 } 195 196 void amdgpu_set_init_level(struct amdgpu_device *adev, 197 enum amdgpu_init_lvl_id lvl) 198 { 199 switch (lvl) { 200 case AMDGPU_INIT_LEVEL_MINIMAL_XGMI: 201 adev->init_lvl = &amdgpu_init_minimal_xgmi; 202 break; 203 case AMDGPU_INIT_LEVEL_RESET_RECOVERY: 204 adev->init_lvl = &amdgpu_init_recovery; 205 break; 206 case AMDGPU_INIT_LEVEL_DEFAULT: 207 fallthrough; 208 default: 209 adev->init_lvl = &amdgpu_init_default; 210 break; 211 } 212 } 213 214 static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev); 215 static int amdgpu_device_pm_notifier(struct notifier_block *nb, unsigned long mode, 216 void *data); 217 218 /** 219 * DOC: pcie_replay_count 220 * 221 * The amdgpu driver provides a sysfs API for reporting the total number 222 * of PCIe replays (NAKs). 223 * The file pcie_replay_count is used for this and returns the total 224 * number of replays as a sum of the NAKs generated and NAKs received. 225 */ 226 227 static ssize_t amdgpu_device_get_pcie_replay_count(struct device *dev, 228 struct device_attribute *attr, char *buf) 229 { 230 struct drm_device *ddev = dev_get_drvdata(dev); 231 struct amdgpu_device *adev = drm_to_adev(ddev); 232 uint64_t cnt = amdgpu_asic_get_pcie_replay_count(adev); 233 234 return sysfs_emit(buf, "%llu\n", cnt); 235 } 236 237 static DEVICE_ATTR(pcie_replay_count, 0444, 238 amdgpu_device_get_pcie_replay_count, NULL); 239 240 static int amdgpu_device_attr_sysfs_init(struct amdgpu_device *adev) 241 { 242 int ret = 0; 243 244 if (amdgpu_nbio_is_replay_cnt_supported(adev)) 245 ret = sysfs_create_file(&adev->dev->kobj, 246 &dev_attr_pcie_replay_count.attr); 247 248 return ret; 249 } 250 251 static void amdgpu_device_attr_sysfs_fini(struct amdgpu_device *adev) 252 { 253 if (amdgpu_nbio_is_replay_cnt_supported(adev)) 254 sysfs_remove_file(&adev->dev->kobj, 255 &dev_attr_pcie_replay_count.attr); 256 } 257 258 static ssize_t amdgpu_sysfs_reg_state_get(struct file *f, struct kobject *kobj, 259 const struct bin_attribute *attr, char *buf, 260 loff_t ppos, size_t count) 261 { 262 struct device *dev = kobj_to_dev(kobj); 263 struct drm_device *ddev = dev_get_drvdata(dev); 264 struct amdgpu_device *adev = drm_to_adev(ddev); 265 ssize_t bytes_read; 266 267 switch (ppos) { 268 case AMDGPU_SYS_REG_STATE_XGMI: 269 bytes_read = amdgpu_asic_get_reg_state( 270 adev, AMDGPU_REG_STATE_TYPE_XGMI, buf, count); 271 break; 272 case AMDGPU_SYS_REG_STATE_WAFL: 273 bytes_read = amdgpu_asic_get_reg_state( 274 adev, AMDGPU_REG_STATE_TYPE_WAFL, buf, count); 275 break; 276 case AMDGPU_SYS_REG_STATE_PCIE: 277 bytes_read = amdgpu_asic_get_reg_state( 278 adev, AMDGPU_REG_STATE_TYPE_PCIE, buf, count); 279 break; 280 case AMDGPU_SYS_REG_STATE_USR: 281 bytes_read = amdgpu_asic_get_reg_state( 282 adev, AMDGPU_REG_STATE_TYPE_USR, buf, count); 283 break; 284 case AMDGPU_SYS_REG_STATE_USR_1: 285 bytes_read = amdgpu_asic_get_reg_state( 286 adev, AMDGPU_REG_STATE_TYPE_USR_1, buf, count); 287 break; 288 default: 289 return -EINVAL; 290 } 291 292 return bytes_read; 293 } 294 295 static const BIN_ATTR(reg_state, 0444, amdgpu_sysfs_reg_state_get, NULL, 296 AMDGPU_SYS_REG_STATE_END); 297 298 int amdgpu_reg_state_sysfs_init(struct amdgpu_device *adev) 299 { 300 int ret; 301 302 if (!amdgpu_asic_get_reg_state_supported(adev)) 303 return 0; 304 305 ret = sysfs_create_bin_file(&adev->dev->kobj, &bin_attr_reg_state); 306 307 return ret; 308 } 309 310 void amdgpu_reg_state_sysfs_fini(struct amdgpu_device *adev) 311 { 312 if (!amdgpu_asic_get_reg_state_supported(adev)) 313 return; 314 sysfs_remove_bin_file(&adev->dev->kobj, &bin_attr_reg_state); 315 } 316 317 /** 318 * DOC: board_info 319 * 320 * The amdgpu driver provides a sysfs API for giving board related information. 321 * It provides the form factor information in the format 322 * 323 * type : form factor 324 * 325 * Possible form factor values 326 * 327 * - "cem" - PCIE CEM card 328 * - "oam" - Open Compute Accelerator Module 329 * - "unknown" - Not known 330 * 331 */ 332 333 static ssize_t amdgpu_device_get_board_info(struct device *dev, 334 struct device_attribute *attr, 335 char *buf) 336 { 337 struct drm_device *ddev = dev_get_drvdata(dev); 338 struct amdgpu_device *adev = drm_to_adev(ddev); 339 enum amdgpu_pkg_type pkg_type = AMDGPU_PKG_TYPE_CEM; 340 const char *pkg; 341 342 if (adev->smuio.funcs && adev->smuio.funcs->get_pkg_type) 343 pkg_type = adev->smuio.funcs->get_pkg_type(adev); 344 345 switch (pkg_type) { 346 case AMDGPU_PKG_TYPE_CEM: 347 pkg = "cem"; 348 break; 349 case AMDGPU_PKG_TYPE_OAM: 350 pkg = "oam"; 351 break; 352 default: 353 pkg = "unknown"; 354 break; 355 } 356 357 return sysfs_emit(buf, "%s : %s\n", "type", pkg); 358 } 359 360 static DEVICE_ATTR(board_info, 0444, amdgpu_device_get_board_info, NULL); 361 362 static struct attribute *amdgpu_board_attrs[] = { 363 &dev_attr_board_info.attr, 364 NULL, 365 }; 366 367 static umode_t amdgpu_board_attrs_is_visible(struct kobject *kobj, 368 struct attribute *attr, int n) 369 { 370 struct device *dev = kobj_to_dev(kobj); 371 struct drm_device *ddev = dev_get_drvdata(dev); 372 struct amdgpu_device *adev = drm_to_adev(ddev); 373 374 if (adev->flags & AMD_IS_APU) 375 return 0; 376 377 return attr->mode; 378 } 379 380 static const struct attribute_group amdgpu_board_attrs_group = { 381 .attrs = amdgpu_board_attrs, 382 .is_visible = amdgpu_board_attrs_is_visible 383 }; 384 385 /** 386 * DOC: uma/carveout_options 387 * 388 * This is a read-only file that lists all available UMA allocation 389 * options and their corresponding indices. Example output:: 390 * 391 * $ cat uma/carveout_options 392 * 0: Minimum (512 MB) 393 * 1: (1 GB) 394 * 2: (2 GB) 395 * 3: (4 GB) 396 * 4: (6 GB) 397 * 5: (8 GB) 398 * 6: (12 GB) 399 * 7: Medium (16 GB) 400 * 8: (24 GB) 401 * 9: High (32 GB) 402 */ 403 static ssize_t carveout_options_show(struct device *dev, 404 struct device_attribute *attr, 405 char *buf) 406 { 407 struct drm_device *ddev = dev_get_drvdata(dev); 408 struct amdgpu_device *adev = drm_to_adev(ddev); 409 struct amdgpu_uma_carveout_info *uma_info = &adev->uma_info; 410 uint32_t memory_carved; 411 ssize_t size = 0; 412 413 if (!uma_info || !uma_info->num_entries) 414 return -ENODEV; 415 416 for (int i = 0; i < uma_info->num_entries; i++) { 417 memory_carved = uma_info->entries[i].memory_carved_mb; 418 if (memory_carved >= SZ_1G/SZ_1M) { 419 size += sysfs_emit_at(buf, size, "%d: %s (%u GB)\n", 420 i, 421 uma_info->entries[i].name, 422 memory_carved >> 10); 423 } else { 424 size += sysfs_emit_at(buf, size, "%d: %s (%u MB)\n", 425 i, 426 uma_info->entries[i].name, 427 memory_carved); 428 } 429 } 430 431 return size; 432 } 433 static DEVICE_ATTR_RO(carveout_options); 434 435 /** 436 * DOC: uma/carveout 437 * 438 * This file is both readable and writable. When read, it shows the 439 * index of the current setting. Writing a valid index to this file 440 * allows users to change the UMA carveout size to the selected option 441 * on the next boot. 442 * 443 * The available options and their corresponding indices can be read 444 * from the uma/carveout_options file. 445 */ 446 static ssize_t carveout_show(struct device *dev, 447 struct device_attribute *attr, 448 char *buf) 449 { 450 struct drm_device *ddev = dev_get_drvdata(dev); 451 struct amdgpu_device *adev = drm_to_adev(ddev); 452 453 return sysfs_emit(buf, "%u\n", adev->uma_info.uma_option_index); 454 } 455 456 static ssize_t carveout_store(struct device *dev, 457 struct device_attribute *attr, 458 const char *buf, size_t count) 459 { 460 struct drm_device *ddev = dev_get_drvdata(dev); 461 struct amdgpu_device *adev = drm_to_adev(ddev); 462 struct amdgpu_uma_carveout_info *uma_info = &adev->uma_info; 463 struct amdgpu_uma_carveout_option *opt; 464 unsigned long val; 465 uint8_t flags; 466 int r; 467 468 r = kstrtoul(buf, 10, &val); 469 if (r) 470 return r; 471 472 if (val >= uma_info->num_entries) 473 return -EINVAL; 474 475 val = array_index_nospec(val, uma_info->num_entries); 476 opt = &uma_info->entries[val]; 477 478 if (!(opt->flags & AMDGPU_UMA_FLAG_AUTO) && 479 !(opt->flags & AMDGPU_UMA_FLAG_CUSTOM)) { 480 drm_err_once(ddev, "Option %lu not supported due to lack of Custom/Auto flag", val); 481 return -EINVAL; 482 } 483 484 flags = opt->flags; 485 flags &= ~((flags & AMDGPU_UMA_FLAG_AUTO) >> 1); 486 487 guard(mutex)(&uma_info->update_lock); 488 489 r = amdgpu_acpi_set_uma_allocation_size(adev, val, flags); 490 if (r) 491 return r; 492 493 uma_info->uma_option_index = val; 494 495 return count; 496 } 497 static DEVICE_ATTR_RW(carveout); 498 499 static struct attribute *amdgpu_uma_attrs[] = { 500 &dev_attr_carveout.attr, 501 &dev_attr_carveout_options.attr, 502 NULL 503 }; 504 505 const struct attribute_group amdgpu_uma_attr_group = { 506 .name = "uma", 507 .attrs = amdgpu_uma_attrs 508 }; 509 510 static void amdgpu_uma_sysfs_init(struct amdgpu_device *adev) 511 { 512 int rc; 513 514 if (!(adev->flags & AMD_IS_APU)) 515 return; 516 517 if (!amdgpu_acpi_is_set_uma_allocation_size_supported()) 518 return; 519 520 rc = amdgpu_atomfirmware_get_uma_carveout_info(adev, &adev->uma_info); 521 if (rc) { 522 drm_dbg(adev_to_drm(adev), 523 "Failed to parse UMA carveout info from VBIOS: %d\n", rc); 524 goto out_info; 525 } 526 527 mutex_init(&adev->uma_info.update_lock); 528 529 rc = devm_device_add_group(adev->dev, &amdgpu_uma_attr_group); 530 if (rc) { 531 drm_dbg(adev_to_drm(adev), "Failed to add UMA carveout sysfs interfaces %d\n", rc); 532 goto out_attr; 533 } 534 535 return; 536 537 out_attr: 538 mutex_destroy(&adev->uma_info.update_lock); 539 out_info: 540 return; 541 } 542 543 static void amdgpu_uma_sysfs_fini(struct amdgpu_device *adev) 544 { 545 struct amdgpu_uma_carveout_info *uma_info = &adev->uma_info; 546 547 if (!amdgpu_acpi_is_set_uma_allocation_size_supported()) 548 return; 549 550 mutex_destroy(&uma_info->update_lock); 551 uma_info->num_entries = 0; 552 } 553 554 static void amdgpu_device_get_pcie_info(struct amdgpu_device *adev); 555 556 /** 557 * amdgpu_device_supports_px - Is the device a dGPU with ATPX power control 558 * 559 * @adev: amdgpu device pointer 560 * 561 * Returns true if the device is a dGPU with ATPX power control, 562 * otherwise return false. 563 */ 564 bool amdgpu_device_supports_px(struct amdgpu_device *adev) 565 { 566 if ((adev->flags & AMD_IS_PX) && !amdgpu_is_atpx_hybrid()) 567 return true; 568 return false; 569 } 570 571 /** 572 * amdgpu_device_supports_boco - Is the device a dGPU with ACPI power resources 573 * 574 * @adev: amdgpu device pointer 575 * 576 * Returns true if the device is a dGPU with ACPI power control, 577 * otherwise return false. 578 */ 579 bool amdgpu_device_supports_boco(struct amdgpu_device *adev) 580 { 581 if (!IS_ENABLED(CONFIG_HOTPLUG_PCI_PCIE)) 582 return false; 583 584 if (adev->has_pr3 || 585 ((adev->flags & AMD_IS_PX) && amdgpu_is_atpx_hybrid())) 586 return true; 587 return false; 588 } 589 590 /** 591 * amdgpu_device_supports_baco - Does the device support BACO 592 * 593 * @adev: amdgpu device pointer 594 * 595 * Return: 596 * 1 if the device supports BACO; 597 * 3 if the device supports MACO (only works if BACO is supported) 598 * otherwise return 0. 599 */ 600 int amdgpu_device_supports_baco(struct amdgpu_device *adev) 601 { 602 return amdgpu_asic_supports_baco(adev); 603 } 604 605 void amdgpu_device_detect_runtime_pm_mode(struct amdgpu_device *adev) 606 { 607 int bamaco_support; 608 609 adev->pm.rpm_mode = AMDGPU_RUNPM_NONE; 610 bamaco_support = amdgpu_device_supports_baco(adev); 611 612 switch (amdgpu_runtime_pm) { 613 case 2: 614 if (bamaco_support & MACO_SUPPORT) { 615 adev->pm.rpm_mode = AMDGPU_RUNPM_BAMACO; 616 dev_info(adev->dev, "Forcing BAMACO for runtime pm\n"); 617 } else if (bamaco_support == BACO_SUPPORT) { 618 adev->pm.rpm_mode = AMDGPU_RUNPM_BACO; 619 dev_info(adev->dev, "Requested mode BAMACO not available,fallback to use BACO\n"); 620 } 621 break; 622 case 1: 623 if (bamaco_support & BACO_SUPPORT) { 624 adev->pm.rpm_mode = AMDGPU_RUNPM_BACO; 625 dev_info(adev->dev, "Forcing BACO for runtime pm\n"); 626 } 627 break; 628 case -1: 629 case -2: 630 if (amdgpu_device_supports_px(adev)) { 631 /* enable PX as runtime mode */ 632 adev->pm.rpm_mode = AMDGPU_RUNPM_PX; 633 dev_info(adev->dev, "Using ATPX for runtime pm\n"); 634 } else if (amdgpu_device_supports_boco(adev)) { 635 /* enable boco as runtime mode */ 636 adev->pm.rpm_mode = AMDGPU_RUNPM_BOCO; 637 dev_info(adev->dev, "Using BOCO for runtime pm\n"); 638 } else { 639 if (!bamaco_support) 640 goto no_runtime_pm; 641 642 switch (adev->asic_type) { 643 case CHIP_VEGA20: 644 case CHIP_ARCTURUS: 645 /* BACO are not supported on vega20 and arctrus */ 646 break; 647 case CHIP_VEGA10: 648 /* enable BACO as runpm mode if noretry=0 */ 649 if (!adev->gmc.noretry && !amdgpu_passthrough(adev)) 650 adev->pm.rpm_mode = AMDGPU_RUNPM_BACO; 651 break; 652 default: 653 /* enable BACO as runpm mode on CI+ */ 654 if (!amdgpu_passthrough(adev)) 655 adev->pm.rpm_mode = AMDGPU_RUNPM_BACO; 656 break; 657 } 658 659 if (adev->pm.rpm_mode == AMDGPU_RUNPM_BACO) { 660 if (bamaco_support & MACO_SUPPORT) { 661 adev->pm.rpm_mode = AMDGPU_RUNPM_BAMACO; 662 dev_info(adev->dev, "Using BAMACO for runtime pm\n"); 663 } else { 664 dev_info(adev->dev, "Using BACO for runtime pm\n"); 665 } 666 } 667 } 668 break; 669 case 0: 670 dev_info(adev->dev, "runtime pm is manually disabled\n"); 671 break; 672 default: 673 break; 674 } 675 676 no_runtime_pm: 677 if (adev->pm.rpm_mode == AMDGPU_RUNPM_NONE) 678 dev_info(adev->dev, "Runtime PM not available\n"); 679 } 680 /** 681 * amdgpu_device_supports_smart_shift - Is the device dGPU with 682 * smart shift support 683 * 684 * @adev: amdgpu device pointer 685 * 686 * Returns true if the device is a dGPU with Smart Shift support, 687 * otherwise returns false. 688 */ 689 bool amdgpu_device_supports_smart_shift(struct amdgpu_device *adev) 690 { 691 return (amdgpu_device_supports_boco(adev) && 692 amdgpu_acpi_is_power_shift_control_supported()); 693 } 694 695 /* 696 * VRAM access helper functions 697 */ 698 699 /** 700 * amdgpu_device_mm_access - access vram by MM_INDEX/MM_DATA 701 * 702 * @adev: amdgpu_device pointer 703 * @pos: offset of the buffer in vram 704 * @buf: virtual address of the buffer in system memory 705 * @size: read/write size, sizeof(@buf) must > @size 706 * @write: true - write to vram, otherwise - read from vram 707 */ 708 void amdgpu_device_mm_access(struct amdgpu_device *adev, loff_t pos, 709 void *buf, size_t size, bool write) 710 { 711 unsigned long flags; 712 uint32_t hi = ~0, tmp = 0; 713 uint32_t *data = buf; 714 uint64_t last; 715 int idx; 716 717 if (!drm_dev_enter(adev_to_drm(adev), &idx)) 718 return; 719 720 BUG_ON(!IS_ALIGNED(pos, 4) || !IS_ALIGNED(size, 4)); 721 722 spin_lock_irqsave(&adev->mmio_idx_lock, flags); 723 for (last = pos + size; pos < last; pos += 4) { 724 tmp = pos >> 31; 725 726 WREG32_NO_KIQ(mmMM_INDEX, ((uint32_t)pos) | 0x80000000); 727 if (tmp != hi) { 728 WREG32_NO_KIQ(mmMM_INDEX_HI, tmp); 729 hi = tmp; 730 } 731 if (write) 732 WREG32_NO_KIQ(mmMM_DATA, *data++); 733 else 734 *data++ = RREG32_NO_KIQ(mmMM_DATA); 735 } 736 737 spin_unlock_irqrestore(&adev->mmio_idx_lock, flags); 738 drm_dev_exit(idx); 739 } 740 741 /** 742 * amdgpu_device_aper_access - access vram by vram aperture 743 * 744 * @adev: amdgpu_device pointer 745 * @pos: offset of the buffer in vram 746 * @buf: virtual address of the buffer in system memory 747 * @size: read/write size, sizeof(@buf) must > @size 748 * @write: true - write to vram, otherwise - read from vram 749 * 750 * The return value means how many bytes have been transferred. 751 */ 752 size_t amdgpu_device_aper_access(struct amdgpu_device *adev, loff_t pos, 753 void *buf, size_t size, bool write) 754 { 755 #ifdef CONFIG_64BIT 756 void __iomem *addr; 757 size_t count = 0; 758 uint64_t last; 759 760 if (!adev->mman.aper_base_kaddr) 761 return 0; 762 763 last = min(pos + size, adev->gmc.visible_vram_size); 764 if (last > pos) { 765 addr = adev->mman.aper_base_kaddr + pos; 766 count = last - pos; 767 768 if (write) { 769 memcpy_toio(addr, buf, count); 770 /* Make sure HDP write cache flush happens without any reordering 771 * after the system memory contents are sent over PCIe device 772 */ 773 mb(); 774 amdgpu_device_flush_hdp(adev, NULL); 775 } else { 776 amdgpu_device_invalidate_hdp(adev, NULL); 777 /* Make sure HDP read cache is invalidated before issuing a read 778 * to the PCIe device 779 */ 780 mb(); 781 memcpy_fromio(buf, addr, count); 782 } 783 784 } 785 786 return count; 787 #else 788 return 0; 789 #endif 790 } 791 792 /** 793 * amdgpu_device_vram_access - read/write a buffer in vram 794 * 795 * @adev: amdgpu_device pointer 796 * @pos: offset of the buffer in vram 797 * @buf: virtual address of the buffer in system memory 798 * @size: read/write size, sizeof(@buf) must > @size 799 * @write: true - write to vram, otherwise - read from vram 800 */ 801 void amdgpu_device_vram_access(struct amdgpu_device *adev, loff_t pos, 802 void *buf, size_t size, bool write) 803 { 804 size_t count; 805 806 /* try to using vram apreature to access vram first */ 807 count = amdgpu_device_aper_access(adev, pos, buf, size, write); 808 size -= count; 809 if (size) { 810 /* using MM to access rest vram */ 811 pos += count; 812 buf += count; 813 amdgpu_device_mm_access(adev, pos, buf, size, write); 814 } 815 } 816 817 /* 818 * register access helper functions. 819 */ 820 821 /* Check if hw access should be skipped because of hotplug or device error */ 822 bool amdgpu_device_skip_hw_access(struct amdgpu_device *adev) 823 { 824 if (adev->no_hw_access) 825 return true; 826 827 #ifdef CONFIG_LOCKDEP 828 /* 829 * This is a bit complicated to understand, so worth a comment. What we assert 830 * here is that the GPU reset is not running on another thread in parallel. 831 * 832 * For this we trylock the read side of the reset semaphore, if that succeeds 833 * we know that the reset is not running in parallel. 834 * 835 * If the trylock fails we assert that we are either already holding the read 836 * side of the lock or are the reset thread itself and hold the write side of 837 * the lock. 838 */ 839 if (in_task()) { 840 if (down_read_trylock(&adev->reset_domain->sem)) 841 up_read(&adev->reset_domain->sem); 842 else 843 lockdep_assert_held(&adev->reset_domain->sem); 844 } 845 #endif 846 return false; 847 } 848 849 /** 850 * amdgpu_device_get_rev_id - query device rev_id 851 * 852 * @adev: amdgpu_device pointer 853 * 854 * Return device rev_id 855 */ 856 u32 amdgpu_device_get_rev_id(struct amdgpu_device *adev) 857 { 858 return adev->nbio.funcs->get_rev_id(adev); 859 } 860 861 static uint32_t amdgpu_device_get_vbios_flags(struct amdgpu_device *adev) 862 { 863 if (hweight32(adev->aid_mask) && (adev->flags & AMD_IS_APU)) 864 return AMDGPU_VBIOS_SKIP; 865 866 if (hweight32(adev->aid_mask) && amdgpu_passthrough(adev)) 867 return AMDGPU_VBIOS_OPTIONAL; 868 869 return 0; 870 } 871 872 /** 873 * amdgpu_device_asic_init - Wrapper for atom asic_init 874 * 875 * @adev: amdgpu_device pointer 876 * 877 * Does any asic specific work and then calls atom asic init. 878 */ 879 static int amdgpu_device_asic_init(struct amdgpu_device *adev) 880 { 881 uint32_t flags; 882 bool optional; 883 int ret; 884 885 amdgpu_asic_pre_asic_init(adev); 886 flags = amdgpu_device_get_vbios_flags(adev); 887 optional = !!(flags & (AMDGPU_VBIOS_OPTIONAL | AMDGPU_VBIOS_SKIP)); 888 889 if (amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 4, 3) || 890 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 4, 4) || 891 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 5, 0) || 892 amdgpu_ip_version(adev, GC_HWIP, 0) >= IP_VERSION(11, 0, 0)) { 893 amdgpu_psp_wait_for_bootloader(adev); 894 if (optional && !adev->bios) 895 return 0; 896 897 ret = amdgpu_atomfirmware_asic_init(adev, true); 898 return ret; 899 } else { 900 if (optional && !adev->bios) 901 return 0; 902 903 return amdgpu_atom_asic_init(adev->mode_info.atom_context); 904 } 905 906 return 0; 907 } 908 909 /** 910 * amdgpu_device_mem_scratch_init - allocate the VRAM scratch page 911 * 912 * @adev: amdgpu_device pointer 913 * 914 * Allocates a scratch page of VRAM for use by various things in the 915 * driver. 916 */ 917 static int amdgpu_device_mem_scratch_init(struct amdgpu_device *adev) 918 { 919 return amdgpu_bo_create_kernel(adev, AMDGPU_GPU_PAGE_SIZE, PAGE_SIZE, 920 AMDGPU_GEM_DOMAIN_VRAM | 921 AMDGPU_GEM_DOMAIN_GTT, 922 &adev->mem_scratch.robj, 923 &adev->mem_scratch.gpu_addr, 924 (void **)&adev->mem_scratch.ptr); 925 } 926 927 /** 928 * amdgpu_device_mem_scratch_fini - Free the VRAM scratch page 929 * 930 * @adev: amdgpu_device pointer 931 * 932 * Frees the VRAM scratch page. 933 */ 934 static void amdgpu_device_mem_scratch_fini(struct amdgpu_device *adev) 935 { 936 amdgpu_bo_free_kernel(&adev->mem_scratch.robj, NULL, NULL); 937 } 938 939 /** 940 * amdgpu_device_program_register_sequence - program an array of registers. 941 * 942 * @adev: amdgpu_device pointer 943 * @registers: pointer to the register array 944 * @array_size: size of the register array 945 * 946 * Programs an array or registers with and or masks. 947 * This is a helper for setting golden registers. 948 */ 949 void amdgpu_device_program_register_sequence(struct amdgpu_device *adev, 950 const u32 *registers, 951 const u32 array_size) 952 { 953 u32 tmp, reg, and_mask, or_mask; 954 int i; 955 956 if (array_size % 3) 957 return; 958 959 for (i = 0; i < array_size; i += 3) { 960 reg = registers[i + 0]; 961 and_mask = registers[i + 1]; 962 or_mask = registers[i + 2]; 963 964 if (and_mask == 0xffffffff) { 965 tmp = or_mask; 966 } else { 967 tmp = RREG32(reg); 968 tmp &= ~and_mask; 969 if (adev->family >= AMDGPU_FAMILY_AI) 970 tmp |= (or_mask & and_mask); 971 else 972 tmp |= or_mask; 973 } 974 WREG32(reg, tmp); 975 } 976 } 977 978 /** 979 * amdgpu_device_pci_config_reset - reset the GPU 980 * 981 * @adev: amdgpu_device pointer 982 * 983 * Resets the GPU using the pci config reset sequence. 984 * Only applicable to asics prior to vega10. 985 */ 986 void amdgpu_device_pci_config_reset(struct amdgpu_device *adev) 987 { 988 pci_write_config_dword(adev->pdev, 0x7c, AMDGPU_ASIC_RESET_DATA); 989 } 990 991 /** 992 * amdgpu_device_pci_reset - reset the GPU using generic PCI means 993 * 994 * @adev: amdgpu_device pointer 995 * 996 * Resets the GPU using generic pci reset interfaces (FLR, SBR, etc.). 997 */ 998 int amdgpu_device_pci_reset(struct amdgpu_device *adev) 999 { 1000 return pci_reset_function(adev->pdev); 1001 } 1002 1003 /* 1004 * amdgpu_device_wb_*() 1005 * Writeback is the method by which the GPU updates special pages in memory 1006 * with the status of certain GPU events (fences, ring pointers,etc.). 1007 */ 1008 1009 /** 1010 * amdgpu_device_wb_fini - Disable Writeback and free memory 1011 * 1012 * @adev: amdgpu_device pointer 1013 * 1014 * Disables Writeback and frees the Writeback memory (all asics). 1015 * Used at driver shutdown. 1016 */ 1017 static void amdgpu_device_wb_fini(struct amdgpu_device *adev) 1018 { 1019 if (adev->wb.wb_obj) { 1020 amdgpu_bo_free_kernel(&adev->wb.wb_obj, 1021 &adev->wb.gpu_addr, 1022 (void **)&adev->wb.wb); 1023 adev->wb.wb_obj = NULL; 1024 } 1025 } 1026 1027 /** 1028 * amdgpu_device_wb_init - Init Writeback driver info and allocate memory 1029 * 1030 * @adev: amdgpu_device pointer 1031 * 1032 * Initializes writeback and allocates writeback memory (all asics). 1033 * Used at driver startup. 1034 * Returns 0 on success or an -error on failure. 1035 */ 1036 static int amdgpu_device_wb_init(struct amdgpu_device *adev) 1037 { 1038 int r; 1039 1040 if (adev->wb.wb_obj == NULL) { 1041 /* AMDGPU_MAX_WB * sizeof(uint32_t) * 8 = AMDGPU_MAX_WB 256bit slots */ 1042 r = amdgpu_bo_create_kernel(adev, AMDGPU_MAX_WB * sizeof(uint32_t) * 8, 1043 PAGE_SIZE, AMDGPU_GEM_DOMAIN_GTT, 1044 &adev->wb.wb_obj, &adev->wb.gpu_addr, 1045 (void **)&adev->wb.wb); 1046 if (r) { 1047 dev_warn(adev->dev, "(%d) create WB bo failed\n", r); 1048 return r; 1049 } 1050 1051 adev->wb.num_wb = AMDGPU_MAX_WB; 1052 memset(&adev->wb.used, 0, sizeof(adev->wb.used)); 1053 1054 /* clear wb memory */ 1055 memset((char *)adev->wb.wb, 0, AMDGPU_MAX_WB * sizeof(uint32_t) * 8); 1056 } 1057 1058 return 0; 1059 } 1060 1061 /** 1062 * amdgpu_device_wb_get - Allocate a wb entry 1063 * 1064 * @adev: amdgpu_device pointer 1065 * @wb: wb index 1066 * 1067 * Allocate a wb slot for use by the driver (all asics). 1068 * Returns 0 on success or -EINVAL on failure. 1069 */ 1070 int amdgpu_device_wb_get(struct amdgpu_device *adev, u32 *wb) 1071 { 1072 unsigned long flags, offset; 1073 1074 spin_lock_irqsave(&adev->wb.lock, flags); 1075 offset = find_first_zero_bit(adev->wb.used, adev->wb.num_wb); 1076 if (offset < adev->wb.num_wb) { 1077 __set_bit(offset, adev->wb.used); 1078 spin_unlock_irqrestore(&adev->wb.lock, flags); 1079 *wb = offset << 3; /* convert to dw offset */ 1080 return 0; 1081 } else { 1082 spin_unlock_irqrestore(&adev->wb.lock, flags); 1083 return -EINVAL; 1084 } 1085 } 1086 1087 /** 1088 * amdgpu_device_wb_free - Free a wb entry 1089 * 1090 * @adev: amdgpu_device pointer 1091 * @wb: wb index 1092 * 1093 * Free a wb slot allocated for use by the driver (all asics) 1094 */ 1095 void amdgpu_device_wb_free(struct amdgpu_device *adev, u32 wb) 1096 { 1097 unsigned long flags; 1098 1099 wb >>= 3; 1100 spin_lock_irqsave(&adev->wb.lock, flags); 1101 if (wb < adev->wb.num_wb) 1102 __clear_bit(wb, adev->wb.used); 1103 spin_unlock_irqrestore(&adev->wb.lock, flags); 1104 } 1105 1106 /** 1107 * amdgpu_device_resize_fb_bar - try to resize FB BAR 1108 * 1109 * @adev: amdgpu_device pointer 1110 * 1111 * Try to resize FB BAR to make all VRAM CPU accessible. We try very hard not 1112 * to fail, but if any of the BARs is not accessible after the size we abort 1113 * driver loading by returning -ENODEV. 1114 */ 1115 int amdgpu_device_resize_fb_bar(struct amdgpu_device *adev) 1116 { 1117 int rbar_size = pci_rebar_bytes_to_size(adev->gmc.real_vram_size); 1118 struct pci_bus *root; 1119 struct resource *res; 1120 int max_size, r; 1121 unsigned int i; 1122 u16 cmd; 1123 1124 if (!IS_ENABLED(CONFIG_PHYS_ADDR_T_64BIT)) 1125 return 0; 1126 1127 /* Bypass for VF */ 1128 if (amdgpu_sriov_vf(adev)) 1129 return 0; 1130 1131 if (!amdgpu_rebar) 1132 return 0; 1133 1134 /* resizing on Dell G5 SE platforms causes problems with runtime pm */ 1135 if ((amdgpu_runtime_pm != 0) && 1136 adev->pdev->vendor == PCI_VENDOR_ID_ATI && 1137 adev->pdev->device == 0x731f && 1138 adev->pdev->subsystem_vendor == PCI_VENDOR_ID_DELL) 1139 return 0; 1140 1141 /* PCI_EXT_CAP_ID_VNDR extended capability is located at 0x100 */ 1142 if (!pci_find_ext_capability(adev->pdev, PCI_EXT_CAP_ID_VNDR)) 1143 dev_warn( 1144 adev->dev, 1145 "System can't access extended configuration space, please check!!\n"); 1146 1147 /* skip if the bios has already enabled large BAR */ 1148 if (adev->gmc.real_vram_size && 1149 (pci_resource_len(adev->pdev, 0) >= adev->gmc.real_vram_size)) 1150 return 0; 1151 1152 /* Check if the root BUS has 64bit memory resources */ 1153 root = adev->pdev->bus; 1154 while (root->parent) 1155 root = root->parent; 1156 1157 pci_bus_for_each_resource(root, res, i) { 1158 if (res && res->flags & (IORESOURCE_MEM | IORESOURCE_MEM_64) && 1159 res->start > 0x100000000ull) 1160 break; 1161 } 1162 1163 /* Trying to resize is pointless without a root hub window above 4GB */ 1164 if (!res) 1165 return 0; 1166 1167 /* Limit the BAR size to what is available */ 1168 max_size = pci_rebar_get_max_size(adev->pdev, 0); 1169 if (max_size < 0) 1170 return 0; 1171 rbar_size = min(max_size, rbar_size); 1172 1173 /* Disable memory decoding while we change the BAR addresses and size */ 1174 pci_read_config_word(adev->pdev, PCI_COMMAND, &cmd); 1175 pci_write_config_word(adev->pdev, PCI_COMMAND, 1176 cmd & ~PCI_COMMAND_MEMORY); 1177 1178 /* Tear down doorbell as resizing will release BARs */ 1179 amdgpu_doorbell_fini(adev); 1180 1181 r = pci_resize_resource(adev->pdev, 0, rbar_size, 1182 (adev->asic_type >= CHIP_BONAIRE) ? 1 << 5 1183 : 1 << 2); 1184 if (r == -ENOSPC) 1185 dev_info(adev->dev, 1186 "Not enough PCI address space for a large BAR."); 1187 else if (r && r != -ENOTSUPP) 1188 dev_err(adev->dev, "Problem resizing BAR0 (%d).", r); 1189 1190 /* When the doorbell or fb BAR isn't available we have no chance of 1191 * using the device. 1192 */ 1193 r = amdgpu_doorbell_init(adev); 1194 if (r || (pci_resource_flags(adev->pdev, 0) & IORESOURCE_UNSET)) 1195 return -ENODEV; 1196 1197 pci_write_config_word(adev->pdev, PCI_COMMAND, cmd); 1198 1199 return 0; 1200 } 1201 1202 /* 1203 * GPU helpers function. 1204 */ 1205 /** 1206 * amdgpu_device_need_post - check if the hw need post or not 1207 * 1208 * @adev: amdgpu_device pointer 1209 * 1210 * Check if the asic has been initialized (all asics) at driver startup 1211 * or post is needed if hw reset is performed. 1212 * Returns true if need or false if not. 1213 */ 1214 bool amdgpu_device_need_post(struct amdgpu_device *adev) 1215 { 1216 uint32_t reg, flags; 1217 1218 if (amdgpu_sriov_vf(adev)) 1219 return false; 1220 1221 flags = amdgpu_device_get_vbios_flags(adev); 1222 if (flags & AMDGPU_VBIOS_SKIP) 1223 return false; 1224 if ((flags & AMDGPU_VBIOS_OPTIONAL) && !adev->bios) 1225 return false; 1226 1227 if (amdgpu_passthrough(adev)) { 1228 /* for FIJI: In whole GPU pass-through virtualization case, after VM reboot 1229 * some old smc fw still need driver do vPost otherwise gpu hang, while 1230 * those smc fw version above 22.15 doesn't have this flaw, so we force 1231 * vpost executed for smc version below 22.15 1232 */ 1233 if (adev->asic_type == CHIP_FIJI) { 1234 int err; 1235 uint32_t fw_ver; 1236 1237 err = request_firmware(&adev->pm.fw, "amdgpu/fiji_smc.bin", adev->dev); 1238 /* force vPost if error occurred */ 1239 if (err) 1240 return true; 1241 1242 fw_ver = *((uint32_t *)adev->pm.fw->data + 69); 1243 release_firmware(adev->pm.fw); 1244 if (fw_ver < 0x00160e00) 1245 return true; 1246 } 1247 } 1248 1249 /* Don't post if we need to reset whole hive on init */ 1250 if (adev->init_lvl->level == AMDGPU_INIT_LEVEL_MINIMAL_XGMI) 1251 return false; 1252 1253 if (adev->has_hw_reset) { 1254 adev->has_hw_reset = false; 1255 return true; 1256 } 1257 1258 /* bios scratch used on CIK+ */ 1259 if (adev->asic_type >= CHIP_BONAIRE) 1260 return amdgpu_atombios_scratch_need_asic_init(adev); 1261 1262 /* check MEM_SIZE for older asics */ 1263 reg = amdgpu_asic_get_config_memsize(adev); 1264 1265 if ((reg != 0) && (reg != 0xffffffff)) 1266 return false; 1267 1268 return true; 1269 } 1270 1271 /* 1272 * Check whether seamless boot is supported. 1273 * 1274 * So far we only support seamless boot on DCE 3.0 or later. 1275 * If users report that it works on older ASICS as well, we may 1276 * loosen this. 1277 */ 1278 bool amdgpu_device_seamless_boot_supported(struct amdgpu_device *adev) 1279 { 1280 switch (amdgpu_seamless) { 1281 case -1: 1282 break; 1283 case 1: 1284 return true; 1285 case 0: 1286 return false; 1287 default: 1288 dev_err(adev->dev, "Invalid value for amdgpu.seamless: %d\n", 1289 amdgpu_seamless); 1290 return false; 1291 } 1292 1293 if (!(adev->flags & AMD_IS_APU)) 1294 return false; 1295 1296 if (adev->mman.keep_stolen_vga_memory) 1297 return false; 1298 1299 return amdgpu_ip_version(adev, DCE_HWIP, 0) >= IP_VERSION(3, 0, 0); 1300 } 1301 1302 /* 1303 * Intel hosts such as Rocket Lake, Alder Lake, Raptor Lake and Sapphire Rapids 1304 * don't support dynamic speed switching. Until we have confirmation from Intel 1305 * that a specific host supports it, it's safer that we keep it disabled for all. 1306 * 1307 * https://edc.intel.com/content/www/us/en/design/products/platforms/details/raptor-lake-s/13th-generation-core-processors-datasheet-volume-1-of-2/005/pci-express-support/ 1308 * https://gitlab.freedesktop.org/drm/amd/-/issues/2663 1309 */ 1310 static bool amdgpu_device_pcie_dynamic_switching_supported(struct amdgpu_device *adev) 1311 { 1312 #if IS_ENABLED(CONFIG_X86) 1313 struct cpuinfo_x86 *c = &cpu_data(0); 1314 1315 /* eGPU change speeds based on USB4 fabric conditions */ 1316 if (dev_is_removable(adev->dev)) 1317 return true; 1318 1319 if (c->x86_vendor == X86_VENDOR_INTEL) 1320 return false; 1321 #endif 1322 return true; 1323 } 1324 1325 static bool amdgpu_device_aspm_support_quirk(struct amdgpu_device *adev) 1326 { 1327 /* Enabling ASPM causes randoms hangs on Tahiti and Oland on Zen4. 1328 * It's unclear if this is a platform-specific or GPU-specific issue. 1329 * Disable ASPM on SI for the time being. 1330 */ 1331 if (adev->family == AMDGPU_FAMILY_SI) 1332 return true; 1333 1334 #if IS_ENABLED(CONFIG_X86) 1335 struct cpuinfo_x86 *c = &cpu_data(0); 1336 1337 if (!(amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(12, 0, 0) || 1338 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(12, 0, 1))) 1339 return false; 1340 1341 if (c->x86 == 6 && 1342 adev->pm.pcie_gen_mask & CAIL_PCIE_LINK_SPEED_SUPPORT_GEN5) { 1343 switch (c->x86_model) { 1344 case VFM_MODEL(INTEL_ALDERLAKE): 1345 case VFM_MODEL(INTEL_ALDERLAKE_L): 1346 case VFM_MODEL(INTEL_RAPTORLAKE): 1347 case VFM_MODEL(INTEL_RAPTORLAKE_P): 1348 case VFM_MODEL(INTEL_RAPTORLAKE_S): 1349 return true; 1350 default: 1351 return false; 1352 } 1353 } else { 1354 return false; 1355 } 1356 #else 1357 return false; 1358 #endif 1359 } 1360 1361 /** 1362 * amdgpu_device_should_use_aspm - check if the device should program ASPM 1363 * 1364 * @adev: amdgpu_device pointer 1365 * 1366 * Confirm whether the module parameter and pcie bridge agree that ASPM should 1367 * be set for this device. 1368 * 1369 * Returns true if it should be used or false if not. 1370 */ 1371 bool amdgpu_device_should_use_aspm(struct amdgpu_device *adev) 1372 { 1373 switch (amdgpu_aspm) { 1374 case -1: 1375 break; 1376 case 0: 1377 return false; 1378 case 1: 1379 return true; 1380 default: 1381 return false; 1382 } 1383 if (adev->flags & AMD_IS_APU) 1384 return false; 1385 if (amdgpu_device_aspm_support_quirk(adev)) 1386 return false; 1387 return pcie_aspm_enabled(adev->pdev); 1388 } 1389 1390 /* if we get transitioned to only one device, take VGA back */ 1391 /** 1392 * amdgpu_device_vga_set_decode - enable/disable vga decode 1393 * 1394 * @pdev: PCI device pointer 1395 * @state: enable/disable vga decode 1396 * 1397 * Enable/disable vga decode (all asics). 1398 * Returns VGA resource flags. 1399 */ 1400 static unsigned int amdgpu_device_vga_set_decode(struct pci_dev *pdev, 1401 bool state) 1402 { 1403 struct amdgpu_device *adev = drm_to_adev(pci_get_drvdata(pdev)); 1404 1405 amdgpu_asic_set_vga_state(adev, state); 1406 if (state) 1407 return VGA_RSRC_LEGACY_IO | VGA_RSRC_LEGACY_MEM | 1408 VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM; 1409 else 1410 return VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM; 1411 } 1412 1413 /** 1414 * amdgpu_device_check_block_size - validate the vm block size 1415 * 1416 * @adev: amdgpu_device pointer 1417 * 1418 * Validates the vm block size specified via module parameter. 1419 * The vm block size defines number of bits in page table versus page directory, 1420 * a page is 4KB so we have 12 bits offset, minimum 9 bits in the 1421 * page table and the remaining bits are in the page directory. 1422 */ 1423 static void amdgpu_device_check_block_size(struct amdgpu_device *adev) 1424 { 1425 /* defines number of bits in page table versus page directory, 1426 * a page is 4KB so we have 12 bits offset, minimum 9 bits in the 1427 * page table and the remaining bits are in the page directory 1428 */ 1429 if (amdgpu_vm_block_size == -1) 1430 return; 1431 1432 if (amdgpu_vm_block_size < 9) { 1433 dev_warn(adev->dev, "VM page table size (%d) too small\n", 1434 amdgpu_vm_block_size); 1435 amdgpu_vm_block_size = -1; 1436 } 1437 } 1438 1439 /** 1440 * amdgpu_device_check_vm_size - validate the vm size 1441 * 1442 * @adev: amdgpu_device pointer 1443 * 1444 * Validates the vm size in GB specified via module parameter. 1445 * The VM size is the size of the GPU virtual memory space in GB. 1446 */ 1447 static void amdgpu_device_check_vm_size(struct amdgpu_device *adev) 1448 { 1449 /* no need to check the default value */ 1450 if (amdgpu_vm_size == -1) 1451 return; 1452 1453 if (amdgpu_vm_size < 1) { 1454 dev_warn(adev->dev, "VM size (%d) too small, min is 1GB\n", 1455 amdgpu_vm_size); 1456 amdgpu_vm_size = -1; 1457 } 1458 } 1459 1460 static void amdgpu_device_check_smu_prv_buffer_size(struct amdgpu_device *adev) 1461 { 1462 struct sysinfo si; 1463 bool is_os_64 = (sizeof(void *) == 8); 1464 uint64_t total_memory; 1465 uint64_t dram_size_seven_GB = 0x1B8000000; 1466 uint64_t dram_size_three_GB = 0xB8000000; 1467 1468 if (amdgpu_smu_memory_pool_size == 0) 1469 return; 1470 1471 if (!is_os_64) { 1472 dev_warn(adev->dev, "Not 64-bit OS, feature not supported\n"); 1473 goto def_value; 1474 } 1475 si_meminfo(&si); 1476 total_memory = (uint64_t)si.totalram * si.mem_unit; 1477 1478 if ((amdgpu_smu_memory_pool_size == 1) || 1479 (amdgpu_smu_memory_pool_size == 2)) { 1480 if (total_memory < dram_size_three_GB) 1481 goto def_value1; 1482 } else if ((amdgpu_smu_memory_pool_size == 4) || 1483 (amdgpu_smu_memory_pool_size == 8)) { 1484 if (total_memory < dram_size_seven_GB) 1485 goto def_value1; 1486 } else { 1487 dev_warn(adev->dev, "Smu memory pool size not supported\n"); 1488 goto def_value; 1489 } 1490 adev->pm.smu_prv_buffer_size = amdgpu_smu_memory_pool_size << 28; 1491 1492 return; 1493 1494 def_value1: 1495 dev_warn(adev->dev, "No enough system memory\n"); 1496 def_value: 1497 adev->pm.smu_prv_buffer_size = 0; 1498 } 1499 1500 static int amdgpu_device_init_apu_flags(struct amdgpu_device *adev) 1501 { 1502 if (!(adev->flags & AMD_IS_APU) || 1503 adev->asic_type < CHIP_RAVEN) 1504 return 0; 1505 1506 switch (adev->asic_type) { 1507 case CHIP_RAVEN: 1508 if (adev->pdev->device == 0x15dd) 1509 adev->apu_flags |= AMD_APU_IS_RAVEN; 1510 if (adev->pdev->device == 0x15d8) 1511 adev->apu_flags |= AMD_APU_IS_PICASSO; 1512 break; 1513 case CHIP_RENOIR: 1514 if ((adev->pdev->device == 0x1636) || 1515 (adev->pdev->device == 0x164c)) 1516 adev->apu_flags |= AMD_APU_IS_RENOIR; 1517 else 1518 adev->apu_flags |= AMD_APU_IS_GREEN_SARDINE; 1519 break; 1520 case CHIP_VANGOGH: 1521 adev->apu_flags |= AMD_APU_IS_VANGOGH; 1522 break; 1523 case CHIP_YELLOW_CARP: 1524 break; 1525 case CHIP_CYAN_SKILLFISH: 1526 if ((adev->pdev->device == 0x13FE) || 1527 (adev->pdev->device == 0x143F)) 1528 adev->apu_flags |= AMD_APU_IS_CYAN_SKILLFISH2; 1529 break; 1530 default: 1531 break; 1532 } 1533 1534 return 0; 1535 } 1536 1537 /** 1538 * amdgpu_device_check_arguments - validate module params 1539 * 1540 * @adev: amdgpu_device pointer 1541 * 1542 * Validates certain module parameters and updates 1543 * the associated values used by the driver (all asics). 1544 */ 1545 static int amdgpu_device_check_arguments(struct amdgpu_device *adev) 1546 { 1547 int i; 1548 1549 if (amdgpu_sched_jobs < 4) { 1550 dev_warn(adev->dev, "sched jobs (%d) must be at least 4\n", 1551 amdgpu_sched_jobs); 1552 amdgpu_sched_jobs = 4; 1553 } else if (!is_power_of_2(amdgpu_sched_jobs)) { 1554 dev_warn(adev->dev, "sched jobs (%d) must be a power of 2\n", 1555 amdgpu_sched_jobs); 1556 amdgpu_sched_jobs = roundup_pow_of_two(amdgpu_sched_jobs); 1557 } 1558 1559 if (amdgpu_gart_size != -1 && amdgpu_gart_size < 32) { 1560 /* gart size must be greater or equal to 32M */ 1561 dev_warn(adev->dev, "gart size (%d) too small\n", 1562 amdgpu_gart_size); 1563 amdgpu_gart_size = -1; 1564 } 1565 1566 if (amdgpu_gtt_size != -1 && amdgpu_gtt_size < 32) { 1567 /* gtt size must be greater or equal to 32M */ 1568 dev_warn(adev->dev, "gtt size (%d) too small\n", 1569 amdgpu_gtt_size); 1570 amdgpu_gtt_size = -1; 1571 } 1572 1573 /* valid range is between 4 and 9 inclusive */ 1574 if (amdgpu_vm_fragment_size != -1 && 1575 (amdgpu_vm_fragment_size > 9 || amdgpu_vm_fragment_size < 4)) { 1576 dev_warn(adev->dev, "valid range is between 4 and 9\n"); 1577 amdgpu_vm_fragment_size = -1; 1578 } 1579 1580 if (amdgpu_sched_hw_submission < 2) { 1581 dev_warn(adev->dev, "sched hw submission jobs (%d) must be at least 2\n", 1582 amdgpu_sched_hw_submission); 1583 amdgpu_sched_hw_submission = 2; 1584 } else if (!is_power_of_2(amdgpu_sched_hw_submission)) { 1585 dev_warn(adev->dev, "sched hw submission jobs (%d) must be a power of 2\n", 1586 amdgpu_sched_hw_submission); 1587 amdgpu_sched_hw_submission = roundup_pow_of_two(amdgpu_sched_hw_submission); 1588 } 1589 1590 if (amdgpu_reset_method < -1 || amdgpu_reset_method > 4) { 1591 dev_warn(adev->dev, "invalid option for reset method, reverting to default\n"); 1592 amdgpu_reset_method = -1; 1593 } 1594 1595 amdgpu_device_check_smu_prv_buffer_size(adev); 1596 1597 amdgpu_device_check_vm_size(adev); 1598 1599 amdgpu_device_check_block_size(adev); 1600 1601 adev->firmware.load_type = amdgpu_ucode_get_load_type(adev, amdgpu_fw_load_type); 1602 1603 for (i = 0; i < MAX_XCP; i++) { 1604 switch (amdgpu_enforce_isolation) { 1605 case -1: 1606 case 0: 1607 default: 1608 /* disable */ 1609 adev->enforce_isolation[i] = AMDGPU_ENFORCE_ISOLATION_DISABLE; 1610 break; 1611 case 1: 1612 /* enable */ 1613 adev->enforce_isolation[i] = 1614 AMDGPU_ENFORCE_ISOLATION_ENABLE; 1615 break; 1616 case 2: 1617 /* enable legacy mode */ 1618 adev->enforce_isolation[i] = 1619 AMDGPU_ENFORCE_ISOLATION_ENABLE_LEGACY; 1620 break; 1621 case 3: 1622 /* enable only process isolation without submitting cleaner shader */ 1623 adev->enforce_isolation[i] = 1624 AMDGPU_ENFORCE_ISOLATION_NO_CLEANER_SHADER; 1625 break; 1626 } 1627 } 1628 1629 return 0; 1630 } 1631 1632 /** 1633 * amdgpu_switcheroo_set_state - set switcheroo state 1634 * 1635 * @pdev: pci dev pointer 1636 * @state: vga_switcheroo state 1637 * 1638 * Callback for the switcheroo driver. Suspends or resumes 1639 * the asics before or after it is powered up using ACPI methods. 1640 */ 1641 static void amdgpu_switcheroo_set_state(struct pci_dev *pdev, 1642 enum vga_switcheroo_state state) 1643 { 1644 struct drm_device *dev = pci_get_drvdata(pdev); 1645 int r; 1646 1647 if (amdgpu_device_supports_px(drm_to_adev(dev)) && 1648 state == VGA_SWITCHEROO_OFF) 1649 return; 1650 1651 if (state == VGA_SWITCHEROO_ON) { 1652 pr_info("switched on\n"); 1653 /* don't suspend or resume card normally */ 1654 dev->switch_power_state = DRM_SWITCH_POWER_CHANGING; 1655 1656 pci_set_power_state(pdev, PCI_D0); 1657 amdgpu_device_load_pci_state(pdev); 1658 r = pci_enable_device(pdev); 1659 if (r) 1660 dev_warn(&pdev->dev, "pci_enable_device failed (%d)\n", 1661 r); 1662 amdgpu_device_resume(dev, true); 1663 1664 dev->switch_power_state = DRM_SWITCH_POWER_ON; 1665 } else { 1666 dev_info(&pdev->dev, "switched off\n"); 1667 dev->switch_power_state = DRM_SWITCH_POWER_CHANGING; 1668 amdgpu_device_prepare(dev); 1669 amdgpu_device_suspend(dev, true); 1670 amdgpu_device_cache_pci_state(pdev); 1671 /* Shut down the device */ 1672 pci_disable_device(pdev); 1673 pci_set_power_state(pdev, PCI_D3cold); 1674 dev->switch_power_state = DRM_SWITCH_POWER_OFF; 1675 } 1676 } 1677 1678 /** 1679 * amdgpu_switcheroo_can_switch - see if switcheroo state can change 1680 * 1681 * @pdev: pci dev pointer 1682 * 1683 * Callback for the switcheroo driver. Check of the switcheroo 1684 * state can be changed. 1685 * Returns true if the state can be changed, false if not. 1686 */ 1687 static bool amdgpu_switcheroo_can_switch(struct pci_dev *pdev) 1688 { 1689 struct drm_device *dev = pci_get_drvdata(pdev); 1690 1691 /* 1692 * FIXME: open_count is protected by drm_global_mutex but that would lead to 1693 * locking inversion with the driver load path. And the access here is 1694 * completely racy anyway. So don't bother with locking for now. 1695 */ 1696 return atomic_read(&dev->open_count) == 0; 1697 } 1698 1699 static const struct vga_switcheroo_client_ops amdgpu_switcheroo_ops = { 1700 .set_gpu_state = amdgpu_switcheroo_set_state, 1701 .reprobe = NULL, 1702 .can_switch = amdgpu_switcheroo_can_switch, 1703 }; 1704 1705 /** 1706 * amdgpu_device_enable_virtual_display - enable virtual display feature 1707 * 1708 * @adev: amdgpu_device pointer 1709 * 1710 * Enabled the virtual display feature if the user has enabled it via 1711 * the module parameter virtual_display. This feature provides a virtual 1712 * display hardware on headless boards or in virtualized environments. 1713 * This function parses and validates the configuration string specified by 1714 * the user and configures the virtual display configuration (number of 1715 * virtual connectors, crtcs, etc.) specified. 1716 */ 1717 static void amdgpu_device_enable_virtual_display(struct amdgpu_device *adev) 1718 { 1719 adev->enable_virtual_display = false; 1720 1721 if (amdgpu_virtual_display) { 1722 const char *pci_address_name = pci_name(adev->pdev); 1723 char *pciaddstr, *pciaddstr_tmp, *pciaddname_tmp, *pciaddname; 1724 1725 pciaddstr = kstrdup(amdgpu_virtual_display, GFP_KERNEL); 1726 pciaddstr_tmp = pciaddstr; 1727 while ((pciaddname_tmp = strsep(&pciaddstr_tmp, ";"))) { 1728 pciaddname = strsep(&pciaddname_tmp, ","); 1729 if (!strcmp("all", pciaddname) 1730 || !strcmp(pci_address_name, pciaddname)) { 1731 long num_crtc; 1732 int res = -1; 1733 1734 adev->enable_virtual_display = true; 1735 1736 if (pciaddname_tmp) 1737 res = kstrtol(pciaddname_tmp, 10, 1738 &num_crtc); 1739 1740 if (!res) { 1741 if (num_crtc < 1) 1742 num_crtc = 1; 1743 if (num_crtc > 6) 1744 num_crtc = 6; 1745 adev->mode_info.num_crtc = num_crtc; 1746 } else { 1747 adev->mode_info.num_crtc = 1; 1748 } 1749 break; 1750 } 1751 } 1752 1753 dev_info( 1754 adev->dev, 1755 "virtual display string:%s, %s:virtual_display:%d, num_crtc:%d\n", 1756 amdgpu_virtual_display, pci_address_name, 1757 adev->enable_virtual_display, adev->mode_info.num_crtc); 1758 1759 kfree(pciaddstr); 1760 } 1761 } 1762 1763 void amdgpu_device_set_sriov_virtual_display(struct amdgpu_device *adev) 1764 { 1765 if (amdgpu_sriov_vf(adev) && !adev->enable_virtual_display) { 1766 adev->mode_info.num_crtc = 1; 1767 adev->enable_virtual_display = true; 1768 dev_info(adev->dev, "virtual_display:%d, num_crtc:%d\n", 1769 adev->enable_virtual_display, 1770 adev->mode_info.num_crtc); 1771 } 1772 } 1773 1774 /** 1775 * amdgpu_device_parse_gpu_info_fw - parse gpu info firmware 1776 * 1777 * @adev: amdgpu_device pointer 1778 * 1779 * Parses the asic configuration parameters specified in the gpu info 1780 * firmware and makes them available to the driver for use in configuring 1781 * the asic. 1782 * Returns 0 on success, -EINVAL on failure. 1783 */ 1784 static int amdgpu_device_parse_gpu_info_fw(struct amdgpu_device *adev) 1785 { 1786 const char *chip_name; 1787 int err; 1788 const struct gpu_info_firmware_header_v1_0 *hdr; 1789 1790 adev->firmware.gpu_info_fw = NULL; 1791 1792 switch (adev->asic_type) { 1793 default: 1794 return 0; 1795 case CHIP_VEGA10: 1796 chip_name = "vega10"; 1797 break; 1798 case CHIP_VEGA12: 1799 chip_name = "vega12"; 1800 break; 1801 case CHIP_RAVEN: 1802 if (adev->apu_flags & AMD_APU_IS_RAVEN2) 1803 chip_name = "raven2"; 1804 else if (adev->apu_flags & AMD_APU_IS_PICASSO) 1805 chip_name = "picasso"; 1806 else 1807 chip_name = "raven"; 1808 break; 1809 case CHIP_ARCTURUS: 1810 chip_name = "arcturus"; 1811 break; 1812 case CHIP_NAVI12: 1813 if (adev->discovery.bin) 1814 return 0; 1815 chip_name = "navi12"; 1816 break; 1817 case CHIP_CYAN_SKILLFISH: 1818 if (adev->discovery.bin) 1819 return 0; 1820 chip_name = "cyan_skillfish"; 1821 break; 1822 } 1823 1824 err = amdgpu_ucode_request(adev, &adev->firmware.gpu_info_fw, 1825 AMDGPU_UCODE_OPTIONAL, 1826 "amdgpu/%s_gpu_info.bin", chip_name); 1827 if (err) { 1828 dev_err(adev->dev, 1829 "Failed to get gpu_info firmware \"%s_gpu_info.bin\"\n", 1830 chip_name); 1831 goto out; 1832 } 1833 1834 hdr = (const struct gpu_info_firmware_header_v1_0 *)adev->firmware.gpu_info_fw->data; 1835 amdgpu_ucode_print_gpu_info_hdr(&hdr->header); 1836 1837 switch (hdr->version_major) { 1838 case 1: 1839 { 1840 const struct gpu_info_firmware_v1_0 *gpu_info_fw = 1841 (const struct gpu_info_firmware_v1_0 *)(adev->firmware.gpu_info_fw->data + 1842 le32_to_cpu(hdr->header.ucode_array_offset_bytes)); 1843 1844 /* 1845 * Should be dropped when DAL no longer needs it. 1846 */ 1847 if (adev->asic_type == CHIP_NAVI12) 1848 goto parse_soc_bounding_box; 1849 1850 adev->gfx.config.max_shader_engines = le32_to_cpu(gpu_info_fw->gc_num_se); 1851 adev->gfx.config.max_cu_per_sh = le32_to_cpu(gpu_info_fw->gc_num_cu_per_sh); 1852 adev->gfx.config.max_sh_per_se = le32_to_cpu(gpu_info_fw->gc_num_sh_per_se); 1853 adev->gfx.config.max_backends_per_se = le32_to_cpu(gpu_info_fw->gc_num_rb_per_se); 1854 adev->gfx.config.max_texture_channel_caches = 1855 le32_to_cpu(gpu_info_fw->gc_num_tccs); 1856 adev->gfx.config.max_gprs = le32_to_cpu(gpu_info_fw->gc_num_gprs); 1857 adev->gfx.config.max_gs_threads = le32_to_cpu(gpu_info_fw->gc_num_max_gs_thds); 1858 adev->gfx.config.gs_vgt_table_depth = le32_to_cpu(gpu_info_fw->gc_gs_table_depth); 1859 adev->gfx.config.gs_prim_buffer_depth = le32_to_cpu(gpu_info_fw->gc_gsprim_buff_depth); 1860 adev->gfx.config.double_offchip_lds_buf = 1861 le32_to_cpu(gpu_info_fw->gc_double_offchip_lds_buffer); 1862 adev->gfx.cu_info.wave_front_size = le32_to_cpu(gpu_info_fw->gc_wave_size); 1863 adev->gfx.cu_info.max_waves_per_simd = 1864 le32_to_cpu(gpu_info_fw->gc_max_waves_per_simd); 1865 adev->gfx.cu_info.max_scratch_slots_per_cu = 1866 le32_to_cpu(gpu_info_fw->gc_max_scratch_slots_per_cu); 1867 adev->gfx.cu_info.lds_size = le32_to_cpu(gpu_info_fw->gc_lds_size); 1868 if (hdr->version_minor >= 1) { 1869 const struct gpu_info_firmware_v1_1 *gpu_info_fw = 1870 (const struct gpu_info_firmware_v1_1 *)(adev->firmware.gpu_info_fw->data + 1871 le32_to_cpu(hdr->header.ucode_array_offset_bytes)); 1872 adev->gfx.config.num_sc_per_sh = 1873 le32_to_cpu(gpu_info_fw->num_sc_per_sh); 1874 adev->gfx.config.num_packer_per_sc = 1875 le32_to_cpu(gpu_info_fw->num_packer_per_sc); 1876 } 1877 1878 parse_soc_bounding_box: 1879 /* 1880 * soc bounding box info is not integrated in disocovery table, 1881 * we always need to parse it from gpu info firmware if needed. 1882 */ 1883 if (hdr->version_minor == 2) { 1884 const struct gpu_info_firmware_v1_2 *gpu_info_fw = 1885 (const struct gpu_info_firmware_v1_2 *)(adev->firmware.gpu_info_fw->data + 1886 le32_to_cpu(hdr->header.ucode_array_offset_bytes)); 1887 adev->dm.soc_bounding_box = &gpu_info_fw->soc_bounding_box; 1888 } 1889 break; 1890 } 1891 default: 1892 dev_err(adev->dev, 1893 "Unsupported gpu_info table %d\n", hdr->header.ucode_version); 1894 err = -EINVAL; 1895 goto out; 1896 } 1897 out: 1898 return err; 1899 } 1900 1901 static void amdgpu_uid_init(struct amdgpu_device *adev) 1902 { 1903 /* Initialize the UID for the device */ 1904 adev->uid_info = kzalloc_obj(struct amdgpu_uid); 1905 if (!adev->uid_info) { 1906 dev_warn(adev->dev, "Failed to allocate memory for UID\n"); 1907 return; 1908 } 1909 adev->uid_info->adev = adev; 1910 } 1911 1912 static void amdgpu_uid_fini(struct amdgpu_device *adev) 1913 { 1914 /* Free the UID memory */ 1915 kfree(adev->uid_info); 1916 adev->uid_info = NULL; 1917 } 1918 1919 /** 1920 * amdgpu_device_ip_early_init - run early init for hardware IPs 1921 * 1922 * @adev: amdgpu_device pointer 1923 * 1924 * Early initialization pass for hardware IPs. The hardware IPs that make 1925 * up each asic are discovered each IP's early_init callback is run. This 1926 * is the first stage in initializing the asic. 1927 * Returns 0 on success, negative error code on failure. 1928 */ 1929 static int amdgpu_device_ip_early_init(struct amdgpu_device *adev) 1930 { 1931 struct amdgpu_ip_block *ip_block; 1932 struct pci_dev *parent; 1933 bool total, skip_bios; 1934 uint32_t bios_flags; 1935 int i, r; 1936 1937 amdgpu_device_enable_virtual_display(adev); 1938 1939 if (amdgpu_sriov_vf(adev)) { 1940 r = amdgpu_virt_request_full_gpu(adev, true); 1941 if (r) 1942 return r; 1943 1944 r = amdgpu_virt_init_critical_region(adev); 1945 if (r) 1946 return r; 1947 } 1948 1949 switch (adev->asic_type) { 1950 #ifdef CONFIG_DRM_AMDGPU_SI 1951 case CHIP_VERDE: 1952 case CHIP_TAHITI: 1953 case CHIP_PITCAIRN: 1954 case CHIP_OLAND: 1955 case CHIP_HAINAN: 1956 adev->family = AMDGPU_FAMILY_SI; 1957 r = si_set_ip_blocks(adev); 1958 if (r) 1959 return r; 1960 break; 1961 #endif 1962 #ifdef CONFIG_DRM_AMDGPU_CIK 1963 case CHIP_BONAIRE: 1964 case CHIP_HAWAII: 1965 case CHIP_KAVERI: 1966 case CHIP_KABINI: 1967 case CHIP_MULLINS: 1968 if (adev->flags & AMD_IS_APU) 1969 adev->family = AMDGPU_FAMILY_KV; 1970 else 1971 adev->family = AMDGPU_FAMILY_CI; 1972 1973 r = cik_set_ip_blocks(adev); 1974 if (r) 1975 return r; 1976 break; 1977 #endif 1978 case CHIP_TOPAZ: 1979 case CHIP_TONGA: 1980 case CHIP_FIJI: 1981 case CHIP_POLARIS10: 1982 case CHIP_POLARIS11: 1983 case CHIP_POLARIS12: 1984 case CHIP_VEGAM: 1985 case CHIP_CARRIZO: 1986 case CHIP_STONEY: 1987 if (adev->flags & AMD_IS_APU) 1988 adev->family = AMDGPU_FAMILY_CZ; 1989 else 1990 adev->family = AMDGPU_FAMILY_VI; 1991 1992 r = vi_set_ip_blocks(adev); 1993 if (r) 1994 return r; 1995 break; 1996 default: 1997 r = amdgpu_discovery_set_ip_blocks(adev); 1998 if (r) { 1999 adev->num_ip_blocks = 0; 2000 return r; 2001 } 2002 break; 2003 } 2004 2005 /* Check for IP version 9.4.3 with A0 hardware */ 2006 if (amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 4, 3) && 2007 !amdgpu_device_get_rev_id(adev)) { 2008 dev_err(adev->dev, "Unsupported A0 hardware\n"); 2009 return -ENODEV; /* device unsupported - no device error */ 2010 } 2011 2012 if (amdgpu_has_atpx() && 2013 (amdgpu_is_atpx_hybrid() || 2014 amdgpu_has_atpx_dgpu_power_cntl()) && 2015 ((adev->flags & AMD_IS_APU) == 0) && 2016 !dev_is_removable(&adev->pdev->dev)) 2017 adev->flags |= AMD_IS_PX; 2018 2019 if (!(adev->flags & AMD_IS_APU)) { 2020 parent = pcie_find_root_port(adev->pdev); 2021 adev->has_pr3 = parent ? pci_pr3_present(parent) : false; 2022 } 2023 2024 adev->pm.pp_feature = amdgpu_pp_feature_mask; 2025 if (amdgpu_sriov_vf(adev) || sched_policy == KFD_SCHED_POLICY_NO_HWS) 2026 adev->pm.pp_feature &= ~PP_GFXOFF_MASK; 2027 if (amdgpu_sriov_vf(adev) && adev->asic_type == CHIP_SIENNA_CICHLID) 2028 adev->pm.pp_feature &= ~PP_OVERDRIVE_MASK; 2029 if (!amdgpu_device_pcie_dynamic_switching_supported(adev)) 2030 adev->pm.pp_feature &= ~PP_PCIE_DPM_MASK; 2031 2032 adev->virt.is_xgmi_node_migrate_enabled = false; 2033 if (amdgpu_sriov_vf(adev)) { 2034 adev->virt.is_xgmi_node_migrate_enabled = 2035 amdgpu_ip_version((adev), GC_HWIP, 0) == IP_VERSION(9, 4, 4); 2036 } 2037 2038 total = true; 2039 for (i = 0; i < adev->num_ip_blocks; i++) { 2040 ip_block = &adev->ip_blocks[i]; 2041 2042 if ((amdgpu_ip_block_mask & (1 << i)) == 0) { 2043 dev_warn(adev->dev, "disabled ip block: %d <%s>\n", i, 2044 adev->ip_blocks[i].version->funcs->name); 2045 adev->ip_blocks[i].status.valid = false; 2046 } else if (ip_block->version->funcs->early_init) { 2047 r = ip_block->version->funcs->early_init(ip_block); 2048 if (r == -ENOENT) { 2049 adev->ip_blocks[i].status.valid = false; 2050 } else if (r) { 2051 dev_err(adev->dev, 2052 "early_init of IP block <%s> failed %d\n", 2053 adev->ip_blocks[i].version->funcs->name, 2054 r); 2055 total = false; 2056 } else { 2057 adev->ip_blocks[i].status.valid = true; 2058 } 2059 } else { 2060 adev->ip_blocks[i].status.valid = true; 2061 } 2062 /* get the vbios after the asic_funcs are set up */ 2063 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_COMMON) { 2064 r = amdgpu_device_parse_gpu_info_fw(adev); 2065 if (r) 2066 return r; 2067 2068 bios_flags = amdgpu_device_get_vbios_flags(adev); 2069 skip_bios = !!(bios_flags & AMDGPU_VBIOS_SKIP); 2070 /* Read BIOS */ 2071 if (!skip_bios) { 2072 bool optional = 2073 !!(bios_flags & AMDGPU_VBIOS_OPTIONAL); 2074 if (!amdgpu_get_bios(adev) && !optional) 2075 return -EINVAL; 2076 2077 if (optional && !adev->bios) 2078 dev_info( 2079 adev->dev, 2080 "VBIOS image optional, proceeding without VBIOS image"); 2081 2082 if (adev->bios) { 2083 r = amdgpu_atombios_init(adev); 2084 if (r) { 2085 dev_err(adev->dev, 2086 "amdgpu_atombios_init failed\n"); 2087 amdgpu_vf_error_put( 2088 adev, 2089 AMDGIM_ERROR_VF_ATOMBIOS_INIT_FAIL, 2090 0, 0); 2091 return r; 2092 } 2093 } 2094 } 2095 2096 /*get pf2vf msg info at it's earliest time*/ 2097 if (amdgpu_sriov_vf(adev)) 2098 amdgpu_virt_init_data_exchange(adev); 2099 2100 } 2101 } 2102 if (!total) 2103 return -ENODEV; 2104 2105 if (adev->gmc.xgmi.supported) 2106 amdgpu_xgmi_early_init(adev); 2107 2108 if (amdgpu_is_multi_aid(adev)) 2109 amdgpu_uid_init(adev); 2110 ip_block = amdgpu_device_ip_get_ip_block(adev, AMD_IP_BLOCK_TYPE_GFX); 2111 if (ip_block->status.valid != false) 2112 amdgpu_amdkfd_device_probe(adev); 2113 2114 adev->cg_flags &= amdgpu_cg_mask; 2115 adev->pg_flags &= amdgpu_pg_mask; 2116 2117 return 0; 2118 } 2119 2120 static int amdgpu_device_ip_hw_init_phase1(struct amdgpu_device *adev) 2121 { 2122 int i, r; 2123 2124 for (i = 0; i < adev->num_ip_blocks; i++) { 2125 if (!adev->ip_blocks[i].status.sw) 2126 continue; 2127 if (adev->ip_blocks[i].status.hw) 2128 continue; 2129 if (!amdgpu_ip_member_of_hwini( 2130 adev, adev->ip_blocks[i].version->type)) 2131 continue; 2132 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_COMMON || 2133 (amdgpu_sriov_vf(adev) && (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP)) || 2134 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_IH) { 2135 r = adev->ip_blocks[i].version->funcs->hw_init(&adev->ip_blocks[i]); 2136 if (r) { 2137 dev_err(adev->dev, 2138 "hw_init of IP block <%s> failed %d\n", 2139 adev->ip_blocks[i].version->funcs->name, 2140 r); 2141 return r; 2142 } 2143 adev->ip_blocks[i].status.hw = true; 2144 } 2145 } 2146 2147 return 0; 2148 } 2149 2150 static int amdgpu_device_ip_hw_init_phase2(struct amdgpu_device *adev) 2151 { 2152 int i, r; 2153 2154 for (i = 0; i < adev->num_ip_blocks; i++) { 2155 if (!adev->ip_blocks[i].status.sw) 2156 continue; 2157 if (adev->ip_blocks[i].status.hw) 2158 continue; 2159 if (!amdgpu_ip_member_of_hwini( 2160 adev, adev->ip_blocks[i].version->type)) 2161 continue; 2162 r = adev->ip_blocks[i].version->funcs->hw_init(&adev->ip_blocks[i]); 2163 if (r) { 2164 dev_err(adev->dev, 2165 "hw_init of IP block <%s> failed %d\n", 2166 adev->ip_blocks[i].version->funcs->name, r); 2167 return r; 2168 } 2169 adev->ip_blocks[i].status.hw = true; 2170 } 2171 2172 return 0; 2173 } 2174 2175 static int amdgpu_device_fw_loading(struct amdgpu_device *adev) 2176 { 2177 int r = 0; 2178 int i; 2179 uint32_t smu_version; 2180 2181 if (adev->asic_type >= CHIP_VEGA10) { 2182 for (i = 0; i < adev->num_ip_blocks; i++) { 2183 if (adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_PSP) 2184 continue; 2185 2186 if (!amdgpu_ip_member_of_hwini(adev, 2187 AMD_IP_BLOCK_TYPE_PSP)) 2188 break; 2189 2190 if (!adev->ip_blocks[i].status.sw) 2191 continue; 2192 2193 /* no need to do the fw loading again if already done*/ 2194 if (adev->ip_blocks[i].status.hw == true) 2195 break; 2196 2197 if (amdgpu_in_reset(adev) || adev->in_suspend) { 2198 r = amdgpu_ip_block_resume(&adev->ip_blocks[i]); 2199 if (r) 2200 return r; 2201 } else { 2202 r = adev->ip_blocks[i].version->funcs->hw_init(&adev->ip_blocks[i]); 2203 if (r) { 2204 dev_err(adev->dev, 2205 "hw_init of IP block <%s> failed %d\n", 2206 adev->ip_blocks[i] 2207 .version->funcs->name, 2208 r); 2209 return r; 2210 } 2211 adev->ip_blocks[i].status.hw = true; 2212 } 2213 break; 2214 } 2215 } 2216 2217 if (!amdgpu_sriov_vf(adev) || adev->asic_type == CHIP_TONGA) 2218 r = amdgpu_pm_load_smu_firmware(adev, &smu_version); 2219 2220 return r; 2221 } 2222 2223 static int amdgpu_device_init_schedulers(struct amdgpu_device *adev) 2224 { 2225 struct drm_sched_init_args args = { 2226 .ops = &amdgpu_sched_ops, 2227 .num_rqs = DRM_SCHED_PRIORITY_COUNT, 2228 .timeout_wq = adev->reset_domain->wq, 2229 .dev = adev->dev, 2230 }; 2231 long timeout; 2232 int r, i; 2233 2234 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 2235 struct amdgpu_ring *ring = adev->rings[i]; 2236 2237 /* No need to setup the GPU scheduler for rings that don't need it */ 2238 if (!ring || ring->no_scheduler) 2239 continue; 2240 2241 switch (ring->funcs->type) { 2242 case AMDGPU_RING_TYPE_GFX: 2243 timeout = adev->gfx_timeout; 2244 break; 2245 case AMDGPU_RING_TYPE_COMPUTE: 2246 timeout = adev->compute_timeout; 2247 break; 2248 case AMDGPU_RING_TYPE_SDMA: 2249 timeout = adev->sdma_timeout; 2250 break; 2251 default: 2252 timeout = adev->video_timeout; 2253 break; 2254 } 2255 2256 args.timeout = timeout; 2257 args.credit_limit = ring->num_hw_submission; 2258 args.score = ring->sched_score; 2259 args.name = ring->name; 2260 2261 r = drm_sched_init(&ring->sched, &args); 2262 if (r) { 2263 dev_err(adev->dev, 2264 "Failed to create scheduler on ring %s.\n", 2265 ring->name); 2266 return r; 2267 } 2268 r = amdgpu_uvd_entity_init(adev, ring); 2269 if (r) { 2270 dev_err(adev->dev, 2271 "Failed to create UVD scheduling entity on ring %s.\n", 2272 ring->name); 2273 return r; 2274 } 2275 r = amdgpu_vce_entity_init(adev, ring); 2276 if (r) { 2277 dev_err(adev->dev, 2278 "Failed to create VCE scheduling entity on ring %s.\n", 2279 ring->name); 2280 return r; 2281 } 2282 } 2283 2284 if (adev->xcp_mgr) 2285 amdgpu_xcp_update_partition_sched_list(adev); 2286 2287 return 0; 2288 } 2289 2290 2291 /** 2292 * amdgpu_device_ip_init - run init for hardware IPs 2293 * 2294 * @adev: amdgpu_device pointer 2295 * 2296 * Main initialization pass for hardware IPs. The list of all the hardware 2297 * IPs that make up the asic is walked and the sw_init and hw_init callbacks 2298 * are run. sw_init initializes the software state associated with each IP 2299 * and hw_init initializes the hardware associated with each IP. 2300 * Returns 0 on success, negative error code on failure. 2301 */ 2302 static int amdgpu_device_ip_init(struct amdgpu_device *adev) 2303 { 2304 bool init_badpage; 2305 int i, r; 2306 2307 r = amdgpu_ras_init(adev); 2308 if (r) 2309 return r; 2310 2311 for (i = 0; i < adev->num_ip_blocks; i++) { 2312 if (!adev->ip_blocks[i].status.valid) 2313 continue; 2314 if (adev->ip_blocks[i].version->funcs->sw_init) { 2315 r = adev->ip_blocks[i].version->funcs->sw_init(&adev->ip_blocks[i]); 2316 if (r) { 2317 dev_err(adev->dev, 2318 "sw_init of IP block <%s> failed %d\n", 2319 adev->ip_blocks[i].version->funcs->name, 2320 r); 2321 goto init_failed; 2322 } 2323 } 2324 adev->ip_blocks[i].status.sw = true; 2325 2326 if (!amdgpu_ip_member_of_hwini( 2327 adev, adev->ip_blocks[i].version->type)) 2328 continue; 2329 2330 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_COMMON) { 2331 /* need to do common hw init early so everything is set up for gmc */ 2332 r = adev->ip_blocks[i].version->funcs->hw_init(&adev->ip_blocks[i]); 2333 if (r) { 2334 dev_err(adev->dev, "hw_init %d failed %d\n", i, 2335 r); 2336 goto init_failed; 2337 } 2338 adev->ip_blocks[i].status.hw = true; 2339 } else if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GMC) { 2340 /* need to do gmc hw init early so we can allocate gpu mem */ 2341 /* Try to reserve bad pages early */ 2342 if (amdgpu_sriov_vf(adev)) 2343 amdgpu_virt_exchange_data(adev); 2344 2345 r = amdgpu_device_mem_scratch_init(adev); 2346 if (r) { 2347 dev_err(adev->dev, 2348 "amdgpu_mem_scratch_init failed %d\n", 2349 r); 2350 goto init_failed; 2351 } 2352 r = adev->ip_blocks[i].version->funcs->hw_init(&adev->ip_blocks[i]); 2353 if (r) { 2354 dev_err(adev->dev, "hw_init %d failed %d\n", i, 2355 r); 2356 goto init_failed; 2357 } 2358 r = amdgpu_device_wb_init(adev); 2359 if (r) { 2360 dev_err(adev->dev, 2361 "amdgpu_device_wb_init failed %d\n", r); 2362 goto init_failed; 2363 } 2364 adev->ip_blocks[i].status.hw = true; 2365 2366 /* right after GMC hw init, we create CSA */ 2367 if (adev->gfx.mcbp) { 2368 r = amdgpu_allocate_static_csa(adev, &adev->virt.csa_obj, 2369 AMDGPU_GEM_DOMAIN_VRAM | 2370 AMDGPU_GEM_DOMAIN_GTT, 2371 AMDGPU_CSA_SIZE); 2372 if (r) { 2373 dev_err(adev->dev, 2374 "allocate CSA failed %d\n", r); 2375 goto init_failed; 2376 } 2377 } 2378 2379 r = amdgpu_seq64_init(adev); 2380 if (r) { 2381 dev_err(adev->dev, "allocate seq64 failed %d\n", 2382 r); 2383 goto init_failed; 2384 } 2385 } 2386 } 2387 2388 if (amdgpu_sriov_vf(adev)) 2389 amdgpu_virt_init_data_exchange(adev); 2390 2391 r = amdgpu_ib_pool_init(adev); 2392 if (r) { 2393 dev_err(adev->dev, "IB initialization failed (%d).\n", r); 2394 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_IB_INIT_FAIL, 0, r); 2395 goto init_failed; 2396 } 2397 2398 r = amdgpu_ucode_create_bo(adev); /* create ucode bo when sw_init complete*/ 2399 if (r) 2400 goto init_failed; 2401 2402 r = amdgpu_device_ip_hw_init_phase1(adev); 2403 if (r) 2404 goto init_failed; 2405 2406 r = amdgpu_device_fw_loading(adev); 2407 if (r) 2408 goto init_failed; 2409 2410 r = amdgpu_device_ip_hw_init_phase2(adev); 2411 if (r) 2412 goto init_failed; 2413 2414 /* 2415 * retired pages will be loaded from eeprom and reserved here, 2416 * it should be called after amdgpu_device_ip_hw_init_phase2 since 2417 * for some ASICs the RAS EEPROM code relies on SMU fully functioning 2418 * for I2C communication which only true at this point. 2419 * 2420 * amdgpu_ras_recovery_init may fail, but the upper only cares the 2421 * failure from bad gpu situation and stop amdgpu init process 2422 * accordingly. For other failed cases, it will still release all 2423 * the resource and print error message, rather than returning one 2424 * negative value to upper level. 2425 * 2426 * Note: theoretically, this should be called before all vram allocations 2427 * to protect retired page from abusing 2428 */ 2429 init_badpage = (adev->init_lvl->level != AMDGPU_INIT_LEVEL_MINIMAL_XGMI); 2430 r = amdgpu_ras_recovery_init(adev, init_badpage); 2431 if (r) 2432 goto init_failed; 2433 2434 /** 2435 * In case of XGMI grab extra reference for reset domain for this device 2436 */ 2437 if (adev->gmc.xgmi.num_physical_nodes > 1) { 2438 if (amdgpu_xgmi_add_device(adev) == 0) { 2439 if (!amdgpu_sriov_vf(adev)) { 2440 struct amdgpu_hive_info *hive = amdgpu_get_xgmi_hive(adev); 2441 2442 if (WARN_ON(!hive)) { 2443 r = -ENOENT; 2444 goto init_failed; 2445 } 2446 2447 if (!hive->reset_domain || 2448 !amdgpu_reset_get_reset_domain(hive->reset_domain)) { 2449 r = -ENOENT; 2450 amdgpu_put_xgmi_hive(hive); 2451 goto init_failed; 2452 } 2453 2454 /* Drop the early temporary reset domain we created for device */ 2455 amdgpu_reset_put_reset_domain(adev->reset_domain); 2456 adev->reset_domain = hive->reset_domain; 2457 amdgpu_put_xgmi_hive(hive); 2458 } 2459 } 2460 } 2461 2462 r = amdgpu_device_init_schedulers(adev); 2463 if (r) 2464 goto init_failed; 2465 2466 amdgpu_ttm_set_buffer_funcs_status(adev, true); 2467 2468 /* Don't init kfd if whole hive need to be reset during init */ 2469 if (adev->init_lvl->level != AMDGPU_INIT_LEVEL_MINIMAL_XGMI) { 2470 amdgpu_amdkfd_device_init(adev); 2471 } 2472 2473 amdgpu_fru_get_product_info(adev); 2474 2475 r = amdgpu_cper_init(adev); 2476 2477 init_failed: 2478 2479 return r; 2480 } 2481 2482 /** 2483 * amdgpu_device_fill_reset_magic - writes reset magic to gart pointer 2484 * 2485 * @adev: amdgpu_device pointer 2486 * 2487 * Writes a reset magic value to the gart pointer in VRAM. The driver calls 2488 * this function before a GPU reset. If the value is retained after a 2489 * GPU reset, VRAM has not been lost. Some GPU resets may destroy VRAM contents. 2490 */ 2491 static void amdgpu_device_fill_reset_magic(struct amdgpu_device *adev) 2492 { 2493 memcpy(adev->reset_magic, adev->gart.ptr, AMDGPU_RESET_MAGIC_NUM); 2494 } 2495 2496 /** 2497 * amdgpu_device_check_vram_lost - check if vram is valid 2498 * 2499 * @adev: amdgpu_device pointer 2500 * 2501 * Checks the reset magic value written to the gart pointer in VRAM. 2502 * The driver calls this after a GPU reset to see if the contents of 2503 * VRAM is lost or now. 2504 * returns true if vram is lost, false if not. 2505 */ 2506 static bool amdgpu_device_check_vram_lost(struct amdgpu_device *adev) 2507 { 2508 if (memcmp(adev->gart.ptr, adev->reset_magic, 2509 AMDGPU_RESET_MAGIC_NUM)) 2510 return true; 2511 2512 if (!amdgpu_in_reset(adev)) 2513 return false; 2514 2515 /* 2516 * For all ASICs with baco/mode1 reset, the VRAM is 2517 * always assumed to be lost. 2518 */ 2519 switch (amdgpu_asic_reset_method(adev)) { 2520 case AMD_RESET_METHOD_LEGACY: 2521 case AMD_RESET_METHOD_LINK: 2522 case AMD_RESET_METHOD_BACO: 2523 case AMD_RESET_METHOD_MODE1: 2524 return true; 2525 default: 2526 return false; 2527 } 2528 } 2529 2530 /** 2531 * amdgpu_device_set_cg_state - set clockgating for amdgpu device 2532 * 2533 * @adev: amdgpu_device pointer 2534 * @state: clockgating state (gate or ungate) 2535 * 2536 * The list of all the hardware IPs that make up the asic is walked and the 2537 * set_clockgating_state callbacks are run. 2538 * Late initialization pass enabling clockgating for hardware IPs. 2539 * Fini or suspend, pass disabling clockgating for hardware IPs. 2540 * Returns 0 on success, negative error code on failure. 2541 */ 2542 2543 int amdgpu_device_set_cg_state(struct amdgpu_device *adev, 2544 enum amd_clockgating_state state) 2545 { 2546 int i, j, r; 2547 2548 if (amdgpu_emu_mode == 1) 2549 return 0; 2550 2551 for (j = 0; j < adev->num_ip_blocks; j++) { 2552 i = state == AMD_CG_STATE_GATE ? j : adev->num_ip_blocks - j - 1; 2553 if (!adev->ip_blocks[i].status.late_initialized) 2554 continue; 2555 if (!adev->ip_blocks[i].version) 2556 continue; 2557 /* skip CG for GFX, SDMA on S0ix */ 2558 if (adev->in_s0ix && 2559 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GFX || 2560 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_SDMA)) 2561 continue; 2562 /* skip CG for VCE/UVD, it's handled specially */ 2563 if (adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_UVD && 2564 adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_VCE && 2565 adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_VCN && 2566 adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_JPEG && 2567 adev->ip_blocks[i].version->funcs->set_clockgating_state) { 2568 /* enable clockgating to save power */ 2569 r = adev->ip_blocks[i].version->funcs->set_clockgating_state(&adev->ip_blocks[i], 2570 state); 2571 if (r) { 2572 dev_err(adev->dev, 2573 "set_clockgating_state(gate) of IP block <%s> failed %d\n", 2574 adev->ip_blocks[i].version->funcs->name, 2575 r); 2576 return r; 2577 } 2578 } 2579 } 2580 2581 return 0; 2582 } 2583 2584 int amdgpu_device_set_pg_state(struct amdgpu_device *adev, 2585 enum amd_powergating_state state) 2586 { 2587 int i, j, r; 2588 2589 if (amdgpu_emu_mode == 1) 2590 return 0; 2591 2592 for (j = 0; j < adev->num_ip_blocks; j++) { 2593 i = state == AMD_PG_STATE_GATE ? j : adev->num_ip_blocks - j - 1; 2594 if (!adev->ip_blocks[i].status.late_initialized) 2595 continue; 2596 if (!adev->ip_blocks[i].version) 2597 continue; 2598 /* skip PG for GFX, SDMA on S0ix */ 2599 if (adev->in_s0ix && 2600 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GFX || 2601 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_SDMA)) 2602 continue; 2603 /* skip CG for VCE/UVD, it's handled specially */ 2604 if (adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_UVD && 2605 adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_VCE && 2606 adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_VCN && 2607 adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_JPEG && 2608 adev->ip_blocks[i].version->funcs->set_powergating_state) { 2609 /* enable powergating to save power */ 2610 r = adev->ip_blocks[i].version->funcs->set_powergating_state(&adev->ip_blocks[i], 2611 state); 2612 if (r) { 2613 dev_err(adev->dev, 2614 "set_powergating_state(gate) of IP block <%s> failed %d\n", 2615 adev->ip_blocks[i].version->funcs->name, 2616 r); 2617 return r; 2618 } 2619 } 2620 } 2621 return 0; 2622 } 2623 2624 static int amdgpu_device_enable_mgpu_fan_boost(void) 2625 { 2626 struct amdgpu_gpu_instance *gpu_ins; 2627 struct amdgpu_device *adev; 2628 int i, ret = 0; 2629 2630 mutex_lock(&mgpu_info.mutex); 2631 2632 /* 2633 * MGPU fan boost feature should be enabled 2634 * only when there are two or more dGPUs in 2635 * the system 2636 */ 2637 if (mgpu_info.num_dgpu < 2) 2638 goto out; 2639 2640 for (i = 0; i < mgpu_info.num_dgpu; i++) { 2641 gpu_ins = &(mgpu_info.gpu_ins[i]); 2642 adev = gpu_ins->adev; 2643 if (!(adev->flags & AMD_IS_APU || amdgpu_sriov_multi_vf_mode(adev)) && 2644 !gpu_ins->mgpu_fan_enabled) { 2645 ret = amdgpu_dpm_enable_mgpu_fan_boost(adev); 2646 if (ret) 2647 break; 2648 2649 gpu_ins->mgpu_fan_enabled = 1; 2650 } 2651 } 2652 2653 out: 2654 mutex_unlock(&mgpu_info.mutex); 2655 2656 return ret; 2657 } 2658 2659 /** 2660 * amdgpu_device_ip_late_init - run late init for hardware IPs 2661 * 2662 * @adev: amdgpu_device pointer 2663 * 2664 * Late initialization pass for hardware IPs. The list of all the hardware 2665 * IPs that make up the asic is walked and the late_init callbacks are run. 2666 * late_init covers any special initialization that an IP requires 2667 * after all of the have been initialized or something that needs to happen 2668 * late in the init process. 2669 * Returns 0 on success, negative error code on failure. 2670 */ 2671 static int amdgpu_device_ip_late_init(struct amdgpu_device *adev) 2672 { 2673 struct amdgpu_gpu_instance *gpu_instance; 2674 int i = 0, r; 2675 2676 for (i = 0; i < adev->num_ip_blocks; i++) { 2677 if (!adev->ip_blocks[i].status.hw) 2678 continue; 2679 if (adev->ip_blocks[i].version->funcs->late_init) { 2680 r = adev->ip_blocks[i].version->funcs->late_init(&adev->ip_blocks[i]); 2681 if (r) { 2682 dev_err(adev->dev, 2683 "late_init of IP block <%s> failed %d\n", 2684 adev->ip_blocks[i].version->funcs->name, 2685 r); 2686 return r; 2687 } 2688 } 2689 adev->ip_blocks[i].status.late_initialized = true; 2690 } 2691 2692 r = amdgpu_ras_late_init(adev); 2693 if (r) { 2694 dev_err(adev->dev, "amdgpu_ras_late_init failed %d", r); 2695 return r; 2696 } 2697 2698 if (!amdgpu_reset_in_recovery(adev)) 2699 amdgpu_ras_set_error_query_ready(adev, true); 2700 2701 amdgpu_device_set_cg_state(adev, AMD_CG_STATE_GATE); 2702 amdgpu_device_set_pg_state(adev, AMD_PG_STATE_GATE); 2703 2704 amdgpu_device_fill_reset_magic(adev); 2705 2706 r = amdgpu_device_enable_mgpu_fan_boost(); 2707 if (r) 2708 dev_err(adev->dev, "enable mgpu fan boost failed (%d).\n", r); 2709 2710 /* For passthrough configuration on arcturus and aldebaran, enable special handling SBR */ 2711 if (amdgpu_passthrough(adev) && 2712 ((adev->asic_type == CHIP_ARCTURUS && adev->gmc.xgmi.num_physical_nodes > 1) || 2713 adev->asic_type == CHIP_ALDEBARAN)) 2714 amdgpu_dpm_handle_passthrough_sbr(adev, true); 2715 2716 if (adev->gmc.xgmi.num_physical_nodes > 1) { 2717 mutex_lock(&mgpu_info.mutex); 2718 2719 /* 2720 * Reset device p-state to low as this was booted with high. 2721 * 2722 * This should be performed only after all devices from the same 2723 * hive get initialized. 2724 * 2725 * However, it's unknown how many device in the hive in advance. 2726 * As this is counted one by one during devices initializations. 2727 * 2728 * So, we wait for all XGMI interlinked devices initialized. 2729 * This may bring some delays as those devices may come from 2730 * different hives. But that should be OK. 2731 */ 2732 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { 2733 for (i = 0; i < mgpu_info.num_gpu; i++) { 2734 gpu_instance = &(mgpu_info.gpu_ins[i]); 2735 if (gpu_instance->adev->flags & AMD_IS_APU) 2736 continue; 2737 2738 r = amdgpu_xgmi_set_pstate(gpu_instance->adev, 2739 AMDGPU_XGMI_PSTATE_MIN); 2740 if (r) { 2741 dev_err(adev->dev, 2742 "pstate setting failed (%d).\n", 2743 r); 2744 break; 2745 } 2746 } 2747 } 2748 2749 mutex_unlock(&mgpu_info.mutex); 2750 } 2751 2752 return 0; 2753 } 2754 2755 static void amdgpu_ip_block_hw_fini(struct amdgpu_ip_block *ip_block) 2756 { 2757 struct amdgpu_device *adev = ip_block->adev; 2758 int r; 2759 2760 if (!ip_block->version->funcs->hw_fini) { 2761 dev_err(adev->dev, "hw_fini of IP block <%s> not defined\n", 2762 ip_block->version->funcs->name); 2763 } else { 2764 r = ip_block->version->funcs->hw_fini(ip_block); 2765 /* XXX handle errors */ 2766 if (r) { 2767 dev_dbg(adev->dev, 2768 "hw_fini of IP block <%s> failed %d\n", 2769 ip_block->version->funcs->name, r); 2770 } 2771 } 2772 2773 ip_block->status.hw = false; 2774 } 2775 2776 /** 2777 * amdgpu_device_smu_fini_early - smu hw_fini wrapper 2778 * 2779 * @adev: amdgpu_device pointer 2780 * 2781 * For ASICs need to disable SMC first 2782 */ 2783 static void amdgpu_device_smu_fini_early(struct amdgpu_device *adev) 2784 { 2785 int i; 2786 2787 if (amdgpu_ip_version(adev, GC_HWIP, 0) > IP_VERSION(9, 0, 0)) 2788 return; 2789 2790 for (i = 0; i < adev->num_ip_blocks; i++) { 2791 if (!adev->ip_blocks[i].status.hw) 2792 continue; 2793 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_SMC) { 2794 amdgpu_ip_block_hw_fini(&adev->ip_blocks[i]); 2795 break; 2796 } 2797 } 2798 } 2799 2800 static int amdgpu_device_ip_fini_early(struct amdgpu_device *adev) 2801 { 2802 int i, r; 2803 2804 for (i = 0; i < adev->num_ip_blocks; i++) { 2805 if (!adev->ip_blocks[i].version) 2806 continue; 2807 if (!adev->ip_blocks[i].version->funcs->early_fini) 2808 continue; 2809 2810 r = adev->ip_blocks[i].version->funcs->early_fini(&adev->ip_blocks[i]); 2811 if (r) { 2812 dev_dbg(adev->dev, 2813 "early_fini of IP block <%s> failed %d\n", 2814 adev->ip_blocks[i].version->funcs->name, r); 2815 } 2816 } 2817 2818 amdgpu_amdkfd_suspend(adev, true); 2819 amdgpu_amdkfd_teardown_processes(adev); 2820 amdgpu_userq_suspend(adev); 2821 2822 /* Workaround for ASICs need to disable SMC first */ 2823 amdgpu_device_smu_fini_early(adev); 2824 2825 for (i = adev->num_ip_blocks - 1; i >= 0; i--) { 2826 if (!adev->ip_blocks[i].status.hw) 2827 continue; 2828 2829 amdgpu_ip_block_hw_fini(&adev->ip_blocks[i]); 2830 } 2831 2832 if (amdgpu_sriov_vf(adev)) { 2833 if (amdgpu_virt_release_full_gpu(adev, false)) 2834 dev_err(adev->dev, 2835 "failed to release exclusive mode on fini\n"); 2836 } 2837 2838 /* 2839 * Driver reload on the APU can fail due to firmware validation because 2840 * the PSP is always running, as it is shared across the whole SoC. 2841 * This same issue does not occur on dGPU because it has a mechanism 2842 * that checks whether the PSP is running. A solution for those issues 2843 * in the APU is to trigger a GPU reset, but this should be done during 2844 * the unload phase to avoid adding boot latency and screen flicker. 2845 */ 2846 if ((adev->flags & AMD_IS_APU) && !adev->gmc.is_app_apu) { 2847 r = amdgpu_asic_reset(adev); 2848 if (r) 2849 dev_err(adev->dev, "asic reset on %s failed\n", __func__); 2850 } 2851 2852 return 0; 2853 } 2854 2855 /** 2856 * amdgpu_device_ip_fini - run fini for hardware IPs 2857 * 2858 * @adev: amdgpu_device pointer 2859 * 2860 * Main teardown pass for hardware IPs. The list of all the hardware 2861 * IPs that make up the asic is walked and the hw_fini and sw_fini callbacks 2862 * are run. hw_fini tears down the hardware associated with each IP 2863 * and sw_fini tears down any software state associated with each IP. 2864 * Returns 0 on success, negative error code on failure. 2865 */ 2866 static int amdgpu_device_ip_fini(struct amdgpu_device *adev) 2867 { 2868 int i, r; 2869 2870 amdgpu_cper_fini(adev); 2871 2872 if (amdgpu_sriov_vf(adev) && adev->virt.ras_init_done) 2873 amdgpu_virt_release_ras_err_handler_data(adev); 2874 2875 if (adev->gmc.xgmi.num_physical_nodes > 1) 2876 amdgpu_xgmi_remove_device(adev); 2877 2878 amdgpu_amdkfd_device_fini_sw(adev); 2879 2880 for (i = adev->num_ip_blocks - 1; i >= 0; i--) { 2881 if (!adev->ip_blocks[i].status.sw) 2882 continue; 2883 2884 if (!adev->ip_blocks[i].version) 2885 continue; 2886 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GMC) { 2887 amdgpu_ucode_free_bo(adev); 2888 amdgpu_free_static_csa(&adev->virt.csa_obj); 2889 amdgpu_device_wb_fini(adev); 2890 amdgpu_device_mem_scratch_fini(adev); 2891 amdgpu_ib_pool_fini(adev); 2892 amdgpu_seq64_fini(adev); 2893 amdgpu_doorbell_fini(adev); 2894 } 2895 if (adev->ip_blocks[i].version->funcs->sw_fini) { 2896 r = adev->ip_blocks[i].version->funcs->sw_fini(&adev->ip_blocks[i]); 2897 /* XXX handle errors */ 2898 if (r) { 2899 dev_dbg(adev->dev, 2900 "sw_fini of IP block <%s> failed %d\n", 2901 adev->ip_blocks[i].version->funcs->name, 2902 r); 2903 } 2904 } 2905 adev->ip_blocks[i].status.sw = false; 2906 adev->ip_blocks[i].status.valid = false; 2907 } 2908 2909 for (i = adev->num_ip_blocks - 1; i >= 0; i--) { 2910 if (!adev->ip_blocks[i].status.late_initialized) 2911 continue; 2912 if (!adev->ip_blocks[i].version) 2913 continue; 2914 if (adev->ip_blocks[i].version->funcs->late_fini) 2915 adev->ip_blocks[i].version->funcs->late_fini(&adev->ip_blocks[i]); 2916 adev->ip_blocks[i].status.late_initialized = false; 2917 } 2918 2919 amdgpu_ras_fini(adev); 2920 amdgpu_uid_fini(adev); 2921 2922 return 0; 2923 } 2924 2925 /** 2926 * amdgpu_device_delayed_init_work_handler - work handler for IB tests 2927 * 2928 * @work: work_struct. 2929 */ 2930 static void amdgpu_device_delayed_init_work_handler(struct work_struct *work) 2931 { 2932 struct amdgpu_device *adev = 2933 container_of(work, struct amdgpu_device, delayed_init_work.work); 2934 int r; 2935 2936 r = amdgpu_ib_ring_tests(adev); 2937 if (r) 2938 dev_err(adev->dev, "ib ring test failed (%d).\n", r); 2939 } 2940 2941 static void amdgpu_device_delay_enable_gfx_off(struct work_struct *work) 2942 { 2943 struct amdgpu_device *adev = 2944 container_of(work, struct amdgpu_device, gfx.gfx_off_delay_work.work); 2945 2946 WARN_ON_ONCE(adev->gfx.gfx_off_state); 2947 WARN_ON_ONCE(adev->gfx.gfx_off_req_count); 2948 2949 if (!amdgpu_dpm_set_powergating_by_smu(adev, AMD_IP_BLOCK_TYPE_GFX, true, 0)) 2950 adev->gfx.gfx_off_state = true; 2951 } 2952 2953 /** 2954 * amdgpu_device_ip_suspend_phase1 - run suspend for hardware IPs (phase 1) 2955 * 2956 * @adev: amdgpu_device pointer 2957 * 2958 * Main suspend function for hardware IPs. The list of all the hardware 2959 * IPs that make up the asic is walked, clockgating is disabled and the 2960 * suspend callbacks are run. suspend puts the hardware and software state 2961 * in each IP into a state suitable for suspend. 2962 * Returns 0 on success, negative error code on failure. 2963 */ 2964 static int amdgpu_device_ip_suspend_phase1(struct amdgpu_device *adev) 2965 { 2966 int i, r, rec; 2967 2968 amdgpu_device_set_pg_state(adev, AMD_PG_STATE_UNGATE); 2969 amdgpu_device_set_cg_state(adev, AMD_CG_STATE_UNGATE); 2970 2971 /* 2972 * Per PMFW team's suggestion, driver needs to handle gfxoff 2973 * and df cstate features disablement for gpu reset(e.g. Mode1Reset) 2974 * scenario. Add the missing df cstate disablement here. 2975 */ 2976 if (amdgpu_dpm_set_df_cstate(adev, DF_CSTATE_DISALLOW)) 2977 dev_warn(adev->dev, "Failed to disallow df cstate"); 2978 2979 for (i = adev->num_ip_blocks - 1; i >= 0; i--) { 2980 if (!adev->ip_blocks[i].status.valid) 2981 continue; 2982 2983 /* displays are handled separately */ 2984 if (adev->ip_blocks[i].version->type != AMD_IP_BLOCK_TYPE_DCE) 2985 continue; 2986 2987 r = amdgpu_ip_block_suspend(&adev->ip_blocks[i]); 2988 if (r) 2989 goto unwind; 2990 } 2991 2992 return 0; 2993 unwind: 2994 rec = amdgpu_device_ip_resume_phase3(adev); 2995 if (rec) 2996 dev_err(adev->dev, 2997 "amdgpu_device_ip_resume_phase3 failed during unwind: %d\n", 2998 rec); 2999 3000 amdgpu_dpm_set_df_cstate(adev, DF_CSTATE_ALLOW); 3001 3002 amdgpu_device_set_pg_state(adev, AMD_PG_STATE_GATE); 3003 amdgpu_device_set_cg_state(adev, AMD_CG_STATE_GATE); 3004 3005 return r; 3006 } 3007 3008 /** 3009 * amdgpu_device_ip_suspend_phase2 - run suspend for hardware IPs (phase 2) 3010 * 3011 * @adev: amdgpu_device pointer 3012 * 3013 * Main suspend function for hardware IPs. The list of all the hardware 3014 * IPs that make up the asic is walked, clockgating is disabled and the 3015 * suspend callbacks are run. suspend puts the hardware and software state 3016 * in each IP into a state suitable for suspend. 3017 * Returns 0 on success, negative error code on failure. 3018 */ 3019 static int amdgpu_device_ip_suspend_phase2(struct amdgpu_device *adev) 3020 { 3021 int i, r, rec; 3022 3023 if (adev->in_s0ix) 3024 amdgpu_dpm_gfx_state_change(adev, sGpuChangeState_D3Entry); 3025 3026 for (i = adev->num_ip_blocks - 1; i >= 0; i--) { 3027 if (!adev->ip_blocks[i].status.valid) 3028 continue; 3029 /* displays are handled in phase1 */ 3030 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_DCE) 3031 continue; 3032 /* PSP lost connection when err_event_athub occurs */ 3033 if (amdgpu_ras_intr_triggered() && 3034 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP) { 3035 adev->ip_blocks[i].status.hw = false; 3036 continue; 3037 } 3038 3039 /* skip unnecessary suspend if we do not initialize them yet */ 3040 if (!amdgpu_ip_member_of_hwini( 3041 adev, adev->ip_blocks[i].version->type)) 3042 continue; 3043 3044 /* Since we skip suspend for S0i3, we need to cancel the delayed 3045 * idle work here as the suspend callback never gets called. 3046 */ 3047 if (adev->in_s0ix && 3048 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GFX && 3049 amdgpu_ip_version(adev, GC_HWIP, 0) >= IP_VERSION(10, 0, 0)) 3050 cancel_delayed_work_sync(&adev->gfx.idle_work); 3051 /* skip suspend of gfx/mes and psp for S0ix 3052 * gfx is in gfxoff state, so on resume it will exit gfxoff just 3053 * like at runtime. PSP is also part of the always on hardware 3054 * so no need to suspend it. 3055 */ 3056 if (adev->in_s0ix && 3057 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP || 3058 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GFX || 3059 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_MES)) 3060 continue; 3061 3062 /* SDMA 5.x+ is part of GFX power domain so it's covered by GFXOFF */ 3063 if (adev->in_s0ix && 3064 (amdgpu_ip_version(adev, SDMA0_HWIP, 0) >= 3065 IP_VERSION(5, 0, 0)) && 3066 (adev->ip_blocks[i].version->type == 3067 AMD_IP_BLOCK_TYPE_SDMA)) 3068 continue; 3069 3070 /* Once swPSP provides the IMU, RLC FW binaries to TOS during cold-boot. 3071 * These are in TMR, hence are expected to be reused by PSP-TOS to reload 3072 * from this location and RLC Autoload automatically also gets loaded 3073 * from here based on PMFW -> PSP message during re-init sequence. 3074 * Therefore, the psp suspend & resume should be skipped to avoid destroy 3075 * the TMR and reload FWs again for IMU enabled APU ASICs. 3076 */ 3077 if (amdgpu_in_reset(adev) && 3078 (adev->flags & AMD_IS_APU) && adev->gfx.imu.funcs && 3079 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP) 3080 continue; 3081 3082 r = amdgpu_ip_block_suspend(&adev->ip_blocks[i]); 3083 if (r) 3084 goto unwind; 3085 3086 /* handle putting the SMC in the appropriate state */ 3087 if (!amdgpu_sriov_vf(adev)) { 3088 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_SMC) { 3089 r = amdgpu_dpm_set_mp1_state(adev, adev->mp1_state); 3090 if (r) { 3091 dev_err(adev->dev, 3092 "SMC failed to set mp1 state %d, %d\n", 3093 adev->mp1_state, r); 3094 goto unwind; 3095 } 3096 } 3097 } 3098 } 3099 3100 return 0; 3101 unwind: 3102 /* suspend phase 2 = resume phase 1 + resume phase 2 */ 3103 rec = amdgpu_device_ip_resume_phase1(adev); 3104 if (rec) { 3105 dev_err(adev->dev, 3106 "amdgpu_device_ip_resume_phase1 failed during unwind: %d\n", 3107 rec); 3108 return r; 3109 } 3110 3111 rec = amdgpu_device_fw_loading(adev); 3112 if (rec) { 3113 dev_err(adev->dev, 3114 "amdgpu_device_fw_loading failed during unwind: %d\n", 3115 rec); 3116 return r; 3117 } 3118 3119 rec = amdgpu_device_ip_resume_phase2(adev); 3120 if (rec) { 3121 dev_err(adev->dev, 3122 "amdgpu_device_ip_resume_phase2 failed during unwind: %d\n", 3123 rec); 3124 return r; 3125 } 3126 3127 return r; 3128 } 3129 3130 /** 3131 * amdgpu_device_ip_suspend - run suspend for hardware IPs 3132 * 3133 * @adev: amdgpu_device pointer 3134 * 3135 * Main suspend function for hardware IPs. The list of all the hardware 3136 * IPs that make up the asic is walked, clockgating is disabled and the 3137 * suspend callbacks are run. suspend puts the hardware and software state 3138 * in each IP into a state suitable for suspend. 3139 * Returns 0 on success, negative error code on failure. 3140 */ 3141 static int amdgpu_device_ip_suspend(struct amdgpu_device *adev) 3142 { 3143 int r; 3144 3145 if (amdgpu_sriov_vf(adev)) { 3146 amdgpu_virt_fini_data_exchange(adev); 3147 amdgpu_virt_request_full_gpu(adev, false); 3148 } 3149 3150 amdgpu_ttm_set_buffer_funcs_status(adev, false); 3151 3152 r = amdgpu_device_ip_suspend_phase1(adev); 3153 if (r) 3154 return r; 3155 r = amdgpu_device_ip_suspend_phase2(adev); 3156 3157 if (amdgpu_sriov_vf(adev)) 3158 amdgpu_virt_release_full_gpu(adev, false); 3159 3160 return r; 3161 } 3162 3163 static int amdgpu_device_ip_reinit_early_sriov(struct amdgpu_device *adev) 3164 { 3165 int i, r; 3166 3167 static enum amd_ip_block_type ip_order[] = { 3168 AMD_IP_BLOCK_TYPE_COMMON, 3169 AMD_IP_BLOCK_TYPE_GMC, 3170 AMD_IP_BLOCK_TYPE_PSP, 3171 AMD_IP_BLOCK_TYPE_IH, 3172 }; 3173 3174 for (i = 0; i < adev->num_ip_blocks; i++) { 3175 int j; 3176 struct amdgpu_ip_block *block; 3177 3178 block = &adev->ip_blocks[i]; 3179 block->status.hw = false; 3180 3181 for (j = 0; j < ARRAY_SIZE(ip_order); j++) { 3182 3183 if (block->version->type != ip_order[j] || 3184 !block->status.valid) 3185 continue; 3186 3187 r = block->version->funcs->hw_init(&adev->ip_blocks[i]); 3188 if (r) { 3189 dev_err(adev->dev, "RE-INIT-early: %s failed\n", 3190 block->version->funcs->name); 3191 return r; 3192 } 3193 block->status.hw = true; 3194 } 3195 } 3196 3197 return 0; 3198 } 3199 3200 static int amdgpu_device_ip_reinit_late_sriov(struct amdgpu_device *adev) 3201 { 3202 struct amdgpu_ip_block *block; 3203 int i, r = 0; 3204 3205 static enum amd_ip_block_type ip_order[] = { 3206 AMD_IP_BLOCK_TYPE_SMC, 3207 AMD_IP_BLOCK_TYPE_DCE, 3208 AMD_IP_BLOCK_TYPE_GFX, 3209 AMD_IP_BLOCK_TYPE_SDMA, 3210 AMD_IP_BLOCK_TYPE_MES, 3211 AMD_IP_BLOCK_TYPE_UVD, 3212 AMD_IP_BLOCK_TYPE_VCE, 3213 AMD_IP_BLOCK_TYPE_VCN, 3214 AMD_IP_BLOCK_TYPE_JPEG 3215 }; 3216 3217 for (i = 0; i < ARRAY_SIZE(ip_order); i++) { 3218 block = amdgpu_device_ip_get_ip_block(adev, ip_order[i]); 3219 3220 if (!block) 3221 continue; 3222 3223 if (block->status.valid && !block->status.hw) { 3224 if (block->version->type == AMD_IP_BLOCK_TYPE_SMC) { 3225 r = amdgpu_ip_block_resume(block); 3226 } else { 3227 r = block->version->funcs->hw_init(block); 3228 } 3229 3230 if (r) { 3231 dev_err(adev->dev, "RE-INIT-late: %s failed\n", 3232 block->version->funcs->name); 3233 break; 3234 } 3235 block->status.hw = true; 3236 } 3237 } 3238 3239 return r; 3240 } 3241 3242 /** 3243 * amdgpu_device_ip_resume_phase1 - run resume for hardware IPs 3244 * 3245 * @adev: amdgpu_device pointer 3246 * 3247 * First resume function for hardware IPs. The list of all the hardware 3248 * IPs that make up the asic is walked and the resume callbacks are run for 3249 * COMMON, GMC, and IH. resume puts the hardware into a functional state 3250 * after a suspend and updates the software state as necessary. This 3251 * function is also used for restoring the GPU after a GPU reset. 3252 * Returns 0 on success, negative error code on failure. 3253 */ 3254 static int amdgpu_device_ip_resume_phase1(struct amdgpu_device *adev) 3255 { 3256 int i, r; 3257 3258 for (i = 0; i < adev->num_ip_blocks; i++) { 3259 if (!adev->ip_blocks[i].status.valid || adev->ip_blocks[i].status.hw) 3260 continue; 3261 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_COMMON || 3262 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GMC || 3263 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_IH || 3264 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP && amdgpu_sriov_vf(adev))) { 3265 3266 r = amdgpu_ip_block_resume(&adev->ip_blocks[i]); 3267 if (r) 3268 return r; 3269 } 3270 } 3271 3272 return 0; 3273 } 3274 3275 /** 3276 * amdgpu_device_ip_resume_phase2 - run resume for hardware IPs 3277 * 3278 * @adev: amdgpu_device pointer 3279 * 3280 * Second resume function for hardware IPs. The list of all the hardware 3281 * IPs that make up the asic is walked and the resume callbacks are run for 3282 * all blocks except COMMON, GMC, and IH. resume puts the hardware into a 3283 * functional state after a suspend and updates the software state as 3284 * necessary. This function is also used for restoring the GPU after a GPU 3285 * reset. 3286 * Returns 0 on success, negative error code on failure. 3287 */ 3288 static int amdgpu_device_ip_resume_phase2(struct amdgpu_device *adev) 3289 { 3290 int i, r; 3291 3292 for (i = 0; i < adev->num_ip_blocks; i++) { 3293 if (!adev->ip_blocks[i].status.valid || adev->ip_blocks[i].status.hw) 3294 continue; 3295 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_COMMON || 3296 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GMC || 3297 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_IH || 3298 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_DCE || 3299 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP) 3300 continue; 3301 r = amdgpu_ip_block_resume(&adev->ip_blocks[i]); 3302 if (r) 3303 return r; 3304 } 3305 3306 return 0; 3307 } 3308 3309 /** 3310 * amdgpu_device_ip_resume_phase3 - run resume for hardware IPs 3311 * 3312 * @adev: amdgpu_device pointer 3313 * 3314 * Third resume function for hardware IPs. The list of all the hardware 3315 * IPs that make up the asic is walked and the resume callbacks are run for 3316 * all DCE. resume puts the hardware into a functional state after a suspend 3317 * and updates the software state as necessary. This function is also used 3318 * for restoring the GPU after a GPU reset. 3319 * 3320 * Returns 0 on success, negative error code on failure. 3321 */ 3322 static int amdgpu_device_ip_resume_phase3(struct amdgpu_device *adev) 3323 { 3324 int i, r; 3325 3326 for (i = 0; i < adev->num_ip_blocks; i++) { 3327 if (!adev->ip_blocks[i].status.valid || adev->ip_blocks[i].status.hw) 3328 continue; 3329 if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_DCE) { 3330 r = amdgpu_ip_block_resume(&adev->ip_blocks[i]); 3331 if (r) 3332 return r; 3333 } 3334 } 3335 3336 return 0; 3337 } 3338 3339 /** 3340 * amdgpu_device_ip_resume - run resume for hardware IPs 3341 * 3342 * @adev: amdgpu_device pointer 3343 * 3344 * Main resume function for hardware IPs. The hardware IPs 3345 * are split into two resume functions because they are 3346 * also used in recovering from a GPU reset and some additional 3347 * steps need to be take between them. In this case (S3/S4) they are 3348 * run sequentially. 3349 * Returns 0 on success, negative error code on failure. 3350 */ 3351 static int amdgpu_device_ip_resume(struct amdgpu_device *adev) 3352 { 3353 int r; 3354 3355 r = amdgpu_device_ip_resume_phase1(adev); 3356 if (r) 3357 return r; 3358 3359 r = amdgpu_device_fw_loading(adev); 3360 if (r) 3361 return r; 3362 3363 r = amdgpu_device_ip_resume_phase2(adev); 3364 3365 amdgpu_ttm_set_buffer_funcs_status(adev, true); 3366 3367 if (r) 3368 return r; 3369 3370 amdgpu_fence_driver_hw_init(adev); 3371 3372 r = amdgpu_device_ip_resume_phase3(adev); 3373 3374 return r; 3375 } 3376 3377 /** 3378 * amdgpu_device_detect_sriov_bios - determine if the board supports SR-IOV 3379 * 3380 * @adev: amdgpu_device pointer 3381 * 3382 * Query the VBIOS data tables to determine if the board supports SR-IOV. 3383 */ 3384 static void amdgpu_device_detect_sriov_bios(struct amdgpu_device *adev) 3385 { 3386 if (amdgpu_sriov_vf(adev)) { 3387 if (adev->is_atom_fw) { 3388 if (amdgpu_atomfirmware_gpu_virtualization_supported(adev)) 3389 adev->virt.caps |= AMDGPU_SRIOV_CAPS_SRIOV_VBIOS; 3390 } else { 3391 if (amdgpu_atombios_has_gpu_virtualization_table(adev)) 3392 adev->virt.caps |= AMDGPU_SRIOV_CAPS_SRIOV_VBIOS; 3393 } 3394 3395 if (!(adev->virt.caps & AMDGPU_SRIOV_CAPS_SRIOV_VBIOS)) 3396 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_NO_VBIOS, 0, 0); 3397 } 3398 } 3399 3400 /** 3401 * amdgpu_device_asic_has_dc_support - determine if DC supports the asic 3402 * 3403 * @pdev : pci device context 3404 * @asic_type: AMD asic type 3405 * 3406 * Check if there is DC (new modesetting infrastructre) support for an asic. 3407 * returns true if DC has support, false if not. 3408 */ 3409 bool amdgpu_device_asic_has_dc_support(struct pci_dev *pdev, 3410 enum amd_asic_type asic_type) 3411 { 3412 switch (asic_type) { 3413 #ifdef CONFIG_DRM_AMDGPU_SI 3414 case CHIP_HAINAN: 3415 #endif 3416 case CHIP_TOPAZ: 3417 /* chips with no display hardware */ 3418 return false; 3419 #if defined(CONFIG_DRM_AMD_DC) 3420 case CHIP_TAHITI: 3421 case CHIP_PITCAIRN: 3422 case CHIP_VERDE: 3423 case CHIP_OLAND: 3424 return amdgpu_dc != 0 && IS_ENABLED(CONFIG_DRM_AMD_DC_SI); 3425 default: 3426 return amdgpu_dc != 0; 3427 #else 3428 default: 3429 if (amdgpu_dc > 0) 3430 dev_info_once( 3431 &pdev->dev, 3432 "Display Core has been requested via kernel parameter but isn't supported by ASIC, ignoring\n"); 3433 return false; 3434 #endif 3435 } 3436 } 3437 3438 /** 3439 * amdgpu_device_has_dc_support - check if dc is supported 3440 * 3441 * @adev: amdgpu_device pointer 3442 * 3443 * Returns true for supported, false for not supported 3444 */ 3445 bool amdgpu_device_has_dc_support(struct amdgpu_device *adev) 3446 { 3447 if (adev->enable_virtual_display || 3448 (adev->harvest_ip_mask & AMD_HARVEST_IP_DMU_MASK)) 3449 return false; 3450 3451 return amdgpu_device_asic_has_dc_support(adev->pdev, adev->asic_type); 3452 } 3453 3454 static void amdgpu_device_xgmi_reset_func(struct work_struct *__work) 3455 { 3456 struct amdgpu_device *adev = 3457 container_of(__work, struct amdgpu_device, xgmi_reset_work); 3458 struct amdgpu_hive_info *hive = amdgpu_get_xgmi_hive(adev); 3459 3460 /* It's a bug to not have a hive within this function */ 3461 if (WARN_ON(!hive)) 3462 return; 3463 3464 /* 3465 * Use task barrier to synchronize all xgmi reset works across the 3466 * hive. task_barrier_enter and task_barrier_exit will block 3467 * until all the threads running the xgmi reset works reach 3468 * those points. task_barrier_full will do both blocks. 3469 */ 3470 if (amdgpu_asic_reset_method(adev) == AMD_RESET_METHOD_BACO) { 3471 3472 task_barrier_enter(&hive->tb); 3473 adev->asic_reset_res = amdgpu_device_baco_enter(adev); 3474 3475 if (adev->asic_reset_res) 3476 goto fail; 3477 3478 task_barrier_exit(&hive->tb); 3479 adev->asic_reset_res = amdgpu_device_baco_exit(adev); 3480 3481 if (adev->asic_reset_res) 3482 goto fail; 3483 3484 amdgpu_ras_reset_error_count(adev, AMDGPU_RAS_BLOCK__MMHUB); 3485 } else { 3486 3487 task_barrier_full(&hive->tb); 3488 adev->asic_reset_res = amdgpu_asic_reset(adev); 3489 } 3490 3491 fail: 3492 if (adev->asic_reset_res) 3493 dev_warn(adev->dev, 3494 "ASIC reset failed with error, %d for drm dev, %s", 3495 adev->asic_reset_res, adev_to_drm(adev)->unique); 3496 amdgpu_put_xgmi_hive(hive); 3497 } 3498 3499 static int amdgpu_device_get_job_timeout_settings(struct amdgpu_device *adev) 3500 { 3501 char *input = amdgpu_lockup_timeout; 3502 char *timeout_setting = NULL; 3503 int index = 0; 3504 long timeout; 3505 int ret = 0; 3506 3507 /* By default timeout for all queues is 2 sec */ 3508 adev->gfx_timeout = adev->compute_timeout = adev->sdma_timeout = 3509 adev->video_timeout = msecs_to_jiffies(2000); 3510 3511 if (!strnlen(input, AMDGPU_MAX_TIMEOUT_PARAM_LENGTH)) 3512 return 0; 3513 3514 while ((timeout_setting = strsep(&input, ",")) && 3515 strnlen(timeout_setting, AMDGPU_MAX_TIMEOUT_PARAM_LENGTH)) { 3516 ret = kstrtol(timeout_setting, 0, &timeout); 3517 if (ret) 3518 return ret; 3519 3520 if (timeout == 0) { 3521 index++; 3522 continue; 3523 } else if (timeout < 0) { 3524 timeout = MAX_SCHEDULE_TIMEOUT; 3525 dev_warn(adev->dev, "lockup timeout disabled"); 3526 add_taint(TAINT_SOFTLOCKUP, LOCKDEP_STILL_OK); 3527 } else { 3528 timeout = msecs_to_jiffies(timeout); 3529 } 3530 3531 switch (index++) { 3532 case 0: 3533 adev->gfx_timeout = timeout; 3534 break; 3535 case 1: 3536 adev->compute_timeout = timeout; 3537 break; 3538 case 2: 3539 adev->sdma_timeout = timeout; 3540 break; 3541 case 3: 3542 adev->video_timeout = timeout; 3543 break; 3544 default: 3545 break; 3546 } 3547 } 3548 3549 /* When only one value specified apply it to all queues. */ 3550 if (index == 1) 3551 adev->gfx_timeout = adev->compute_timeout = adev->sdma_timeout = 3552 adev->video_timeout = timeout; 3553 3554 return ret; 3555 } 3556 3557 /** 3558 * amdgpu_device_check_iommu_direct_map - check if RAM direct mapped to GPU 3559 * 3560 * @adev: amdgpu_device pointer 3561 * 3562 * RAM direct mapped to GPU if IOMMU is not enabled or is pass through mode 3563 */ 3564 static void amdgpu_device_check_iommu_direct_map(struct amdgpu_device *adev) 3565 { 3566 struct iommu_domain *domain; 3567 3568 domain = iommu_get_domain_for_dev(adev->dev); 3569 if (!domain || domain->type == IOMMU_DOMAIN_IDENTITY) 3570 adev->ram_is_direct_mapped = true; 3571 } 3572 3573 #if defined(CONFIG_HSA_AMD_P2P) 3574 /** 3575 * amdgpu_device_check_iommu_remap - Check if DMA remapping is enabled. 3576 * 3577 * @adev: amdgpu_device pointer 3578 * 3579 * return if IOMMU remapping bar address 3580 */ 3581 static bool amdgpu_device_check_iommu_remap(struct amdgpu_device *adev) 3582 { 3583 struct iommu_domain *domain; 3584 3585 domain = iommu_get_domain_for_dev(adev->dev); 3586 if (domain && (domain->type == IOMMU_DOMAIN_DMA || 3587 domain->type == IOMMU_DOMAIN_DMA_FQ)) 3588 return true; 3589 3590 return false; 3591 } 3592 #endif 3593 3594 static void amdgpu_device_set_mcbp(struct amdgpu_device *adev) 3595 { 3596 if (amdgpu_mcbp == 1) 3597 adev->gfx.mcbp = true; 3598 else if (amdgpu_mcbp == 0) 3599 adev->gfx.mcbp = false; 3600 3601 if (amdgpu_sriov_vf(adev)) 3602 adev->gfx.mcbp = true; 3603 3604 if (adev->gfx.mcbp) 3605 dev_info(adev->dev, "MCBP is enabled\n"); 3606 } 3607 3608 static int amdgpu_device_sys_interface_init(struct amdgpu_device *adev) 3609 { 3610 int r; 3611 3612 r = amdgpu_atombios_sysfs_init(adev); 3613 if (r) 3614 drm_err(&adev->ddev, 3615 "registering atombios sysfs failed (%d).\n", r); 3616 3617 r = amdgpu_pm_sysfs_init(adev); 3618 if (r) 3619 dev_err(adev->dev, "registering pm sysfs failed (%d).\n", r); 3620 3621 r = amdgpu_ucode_sysfs_init(adev); 3622 if (r) { 3623 adev->ucode_sysfs_en = false; 3624 dev_err(adev->dev, "Creating firmware sysfs failed (%d).\n", r); 3625 } else 3626 adev->ucode_sysfs_en = true; 3627 3628 r = amdgpu_device_attr_sysfs_init(adev); 3629 if (r) 3630 dev_err(adev->dev, "Could not create amdgpu device attr\n"); 3631 3632 r = devm_device_add_group(adev->dev, &amdgpu_board_attrs_group); 3633 if (r) 3634 dev_err(adev->dev, 3635 "Could not create amdgpu board attributes\n"); 3636 3637 amdgpu_fru_sysfs_init(adev); 3638 amdgpu_reg_state_sysfs_init(adev); 3639 amdgpu_xcp_sysfs_init(adev); 3640 amdgpu_uma_sysfs_init(adev); 3641 3642 return r; 3643 } 3644 3645 static void amdgpu_device_sys_interface_fini(struct amdgpu_device *adev) 3646 { 3647 if (adev->pm.sysfs_initialized) 3648 amdgpu_pm_sysfs_fini(adev); 3649 if (adev->ucode_sysfs_en) 3650 amdgpu_ucode_sysfs_fini(adev); 3651 amdgpu_device_attr_sysfs_fini(adev); 3652 amdgpu_fru_sysfs_fini(adev); 3653 3654 amdgpu_reg_state_sysfs_fini(adev); 3655 amdgpu_xcp_sysfs_fini(adev); 3656 amdgpu_uma_sysfs_fini(adev); 3657 } 3658 3659 /** 3660 * amdgpu_device_init - initialize the driver 3661 * 3662 * @adev: amdgpu_device pointer 3663 * @flags: driver flags 3664 * 3665 * Initializes the driver info and hw (all asics). 3666 * Returns 0 for success or an error on failure. 3667 * Called at driver startup. 3668 */ 3669 int amdgpu_device_init(struct amdgpu_device *adev, 3670 uint32_t flags) 3671 { 3672 struct pci_dev *pdev = adev->pdev; 3673 int r, i; 3674 bool px = false; 3675 u32 max_MBps; 3676 int tmp; 3677 3678 adev->shutdown = false; 3679 adev->flags = flags; 3680 3681 if (amdgpu_force_asic_type >= 0 && amdgpu_force_asic_type < CHIP_LAST) 3682 adev->asic_type = amdgpu_force_asic_type; 3683 else 3684 adev->asic_type = flags & AMD_ASIC_MASK; 3685 3686 adev->usec_timeout = AMDGPU_MAX_USEC_TIMEOUT; 3687 if (amdgpu_emu_mode == 1) 3688 adev->usec_timeout *= 10; 3689 adev->gmc.gart_size = 512 * 1024 * 1024; 3690 adev->accel_working = false; 3691 adev->num_rings = 0; 3692 RCU_INIT_POINTER(adev->gang_submit, dma_fence_get_stub()); 3693 adev->mman.buffer_funcs = NULL; 3694 adev->mman.buffer_funcs_ring = NULL; 3695 adev->vm_manager.vm_pte_funcs = NULL; 3696 adev->vm_manager.vm_pte_num_scheds = 0; 3697 adev->gmc.gmc_funcs = NULL; 3698 adev->harvest_ip_mask = 0x0; 3699 adev->fence_context = dma_fence_context_alloc(AMDGPU_MAX_RINGS); 3700 bitmap_zero(adev->gfx.pipe_reserve_bitmap, AMDGPU_MAX_COMPUTE_QUEUES); 3701 3702 amdgpu_reg_access_init(adev); 3703 3704 dev_info( 3705 adev->dev, 3706 "initializing kernel modesetting (%s 0x%04X:0x%04X 0x%04X:0x%04X 0x%02X).\n", 3707 amdgpu_asic_name[adev->asic_type], pdev->vendor, pdev->device, 3708 pdev->subsystem_vendor, pdev->subsystem_device, pdev->revision); 3709 3710 /* mutex initialization are all done here so we 3711 * can recall function without having locking issues 3712 */ 3713 mutex_init(&adev->firmware.mutex); 3714 mutex_init(&adev->pm.mutex); 3715 mutex_init(&adev->gfx.gpu_clock_mutex); 3716 mutex_init(&adev->srbm_mutex); 3717 mutex_init(&adev->gfx.pipe_reserve_mutex); 3718 mutex_init(&adev->gfx.gfx_off_mutex); 3719 mutex_init(&adev->gfx.partition_mutex); 3720 mutex_init(&adev->grbm_idx_mutex); 3721 mutex_init(&adev->mn_lock); 3722 mutex_init(&adev->virt.vf_errors.lock); 3723 hash_init(adev->mn_hash); 3724 mutex_init(&adev->psp.mutex); 3725 mutex_init(&adev->notifier_lock); 3726 mutex_init(&adev->pm.stable_pstate_ctx_lock); 3727 mutex_init(&adev->benchmark_mutex); 3728 mutex_init(&adev->gfx.reset_sem_mutex); 3729 /* Initialize the mutex for cleaner shader isolation between GFX and compute processes */ 3730 mutex_init(&adev->enforce_isolation_mutex); 3731 for (i = 0; i < MAX_XCP; ++i) { 3732 adev->isolation[i].spearhead = dma_fence_get_stub(); 3733 amdgpu_sync_create(&adev->isolation[i].active); 3734 amdgpu_sync_create(&adev->isolation[i].prev); 3735 } 3736 mutex_init(&adev->gfx.userq_sch_mutex); 3737 mutex_init(&adev->gfx.workload_profile_mutex); 3738 mutex_init(&adev->vcn.workload_profile_mutex); 3739 3740 amdgpu_device_init_apu_flags(adev); 3741 3742 r = amdgpu_device_check_arguments(adev); 3743 if (r) 3744 return r; 3745 3746 spin_lock_init(&adev->mmio_idx_lock); 3747 spin_lock_init(&adev->mm_stats.lock); 3748 spin_lock_init(&adev->virt.rlcg_reg_lock); 3749 spin_lock_init(&adev->wb.lock); 3750 3751 xa_init_flags(&adev->userq_xa, XA_FLAGS_LOCK_IRQ); 3752 3753 INIT_LIST_HEAD(&adev->reset_list); 3754 3755 INIT_LIST_HEAD(&adev->ras_list); 3756 3757 INIT_LIST_HEAD(&adev->pm.od_kobj_list); 3758 3759 xa_init(&adev->userq_doorbell_xa); 3760 3761 INIT_DELAYED_WORK(&adev->delayed_init_work, 3762 amdgpu_device_delayed_init_work_handler); 3763 INIT_DELAYED_WORK(&adev->gfx.gfx_off_delay_work, 3764 amdgpu_device_delay_enable_gfx_off); 3765 /* 3766 * Initialize the enforce_isolation work structures for each XCP 3767 * partition. This work handler is responsible for enforcing shader 3768 * isolation on AMD GPUs. It counts the number of emitted fences for 3769 * each GFX and compute ring. If there are any fences, it schedules 3770 * the `enforce_isolation_work` to be run after a delay. If there are 3771 * no fences, it signals the Kernel Fusion Driver (KFD) to resume the 3772 * runqueue. 3773 */ 3774 for (i = 0; i < MAX_XCP; i++) { 3775 INIT_DELAYED_WORK(&adev->gfx.enforce_isolation[i].work, 3776 amdgpu_gfx_enforce_isolation_handler); 3777 adev->gfx.enforce_isolation[i].adev = adev; 3778 adev->gfx.enforce_isolation[i].xcp_id = i; 3779 } 3780 3781 INIT_WORK(&adev->xgmi_reset_work, amdgpu_device_xgmi_reset_func); 3782 INIT_WORK(&adev->userq_reset_work, amdgpu_userq_reset_work); 3783 3784 adev->gfx.gfx_off_req_count = 1; 3785 adev->gfx.gfx_off_residency = 0; 3786 adev->gfx.gfx_off_entrycount = 0; 3787 adev->pm.ac_power = power_supply_is_system_supplied() > 0; 3788 3789 atomic_set(&adev->throttling_logging_enabled, 1); 3790 /* 3791 * If throttling continues, logging will be performed every minute 3792 * to avoid log flooding. "-1" is subtracted since the thermal 3793 * throttling interrupt comes every second. Thus, the total logging 3794 * interval is 59 seconds(retelimited printk interval) + 1(waiting 3795 * for throttling interrupt) = 60 seconds. 3796 */ 3797 ratelimit_state_init(&adev->throttling_logging_rs, (60 - 1) * HZ, 1); 3798 3799 ratelimit_set_flags(&adev->throttling_logging_rs, RATELIMIT_MSG_ON_RELEASE); 3800 3801 /* Registers mapping */ 3802 /* TODO: block userspace mapping of io register */ 3803 if (adev->asic_type >= CHIP_BONAIRE) { 3804 adev->rmmio_base = pci_resource_start(adev->pdev, 5); 3805 adev->rmmio_size = pci_resource_len(adev->pdev, 5); 3806 } else { 3807 adev->rmmio_base = pci_resource_start(adev->pdev, 2); 3808 adev->rmmio_size = pci_resource_len(adev->pdev, 2); 3809 } 3810 3811 for (i = 0; i < AMD_IP_BLOCK_TYPE_NUM; i++) 3812 atomic_set(&adev->pm.pwr_state[i], POWER_STATE_UNKNOWN); 3813 3814 adev->rmmio = ioremap(adev->rmmio_base, adev->rmmio_size); 3815 if (!adev->rmmio) 3816 return -ENOMEM; 3817 3818 dev_info(adev->dev, "register mmio base: 0x%08X\n", 3819 (uint32_t)adev->rmmio_base); 3820 dev_info(adev->dev, "register mmio size: %u\n", 3821 (unsigned int)adev->rmmio_size); 3822 3823 /* 3824 * Reset domain needs to be present early, before XGMI hive discovered 3825 * (if any) and initialized to use reset sem and in_gpu reset flag 3826 * early on during init and before calling to RREG32. 3827 */ 3828 adev->reset_domain = amdgpu_reset_create_reset_domain(SINGLE_DEVICE, "amdgpu-reset-dev"); 3829 if (!adev->reset_domain) 3830 return -ENOMEM; 3831 3832 /* detect hw virtualization here */ 3833 amdgpu_virt_init(adev); 3834 3835 amdgpu_device_get_pcie_info(adev); 3836 3837 r = amdgpu_device_get_job_timeout_settings(adev); 3838 if (r) { 3839 dev_err(adev->dev, "invalid lockup_timeout parameter syntax\n"); 3840 return r; 3841 } 3842 3843 amdgpu_device_set_mcbp(adev); 3844 3845 /* 3846 * By default, use default mode where all blocks are expected to be 3847 * initialized. At present a 'swinit' of blocks is required to be 3848 * completed before the need for a different level is detected. 3849 */ 3850 amdgpu_set_init_level(adev, AMDGPU_INIT_LEVEL_DEFAULT); 3851 /* early init functions */ 3852 r = amdgpu_device_ip_early_init(adev); 3853 if (r) 3854 return r; 3855 3856 /* 3857 * No need to remove conflicting FBs for non-display class devices. 3858 * This prevents the sysfb from being freed accidently. 3859 */ 3860 if ((pdev->class >> 8) == PCI_CLASS_DISPLAY_VGA || 3861 (pdev->class >> 8) == PCI_CLASS_DISPLAY_OTHER) { 3862 /* Get rid of things like offb */ 3863 r = aperture_remove_conflicting_pci_devices(adev->pdev, amdgpu_kms_driver.name); 3864 if (r) 3865 return r; 3866 } 3867 3868 /* Enable TMZ based on IP_VERSION */ 3869 amdgpu_gmc_tmz_set(adev); 3870 3871 if (amdgpu_sriov_vf(adev) && 3872 amdgpu_ip_version(adev, GC_HWIP, 0) >= IP_VERSION(10, 3, 0)) 3873 /* VF MMIO access (except mailbox range) from CPU 3874 * will be blocked during sriov runtime 3875 */ 3876 adev->virt.caps |= AMDGPU_VF_MMIO_ACCESS_PROTECT; 3877 3878 amdgpu_gmc_noretry_set(adev); 3879 /* Need to get xgmi info early to decide the reset behavior*/ 3880 if (adev->gmc.xgmi.supported) { 3881 r = adev->gfxhub.funcs->get_xgmi_info(adev); 3882 if (r) 3883 return r; 3884 } 3885 3886 /* enable PCIE atomic ops */ 3887 if (amdgpu_sriov_vf(adev)) { 3888 if (adev->virt.fw_reserve.p_pf2vf) 3889 adev->have_atomics_support = ((struct amd_sriov_msg_pf2vf_info *) 3890 adev->virt.fw_reserve.p_pf2vf)->pcie_atomic_ops_support_flags == 3891 (PCI_EXP_DEVCAP2_ATOMIC_COMP32 | PCI_EXP_DEVCAP2_ATOMIC_COMP64); 3892 /* APUs w/ gfx9 onwards doesn't reply on PCIe atomics, rather it is a 3893 * internal path natively support atomics, set have_atomics_support to true. 3894 */ 3895 } else if ((adev->flags & AMD_IS_APU && 3896 amdgpu_ip_version(adev, GC_HWIP, 0) > IP_VERSION(9, 0, 0)) || 3897 (adev->gmc.xgmi.connected_to_cpu && 3898 amdgpu_ip_version(adev, GC_HWIP, 0) >= IP_VERSION(12, 1, 0))) { 3899 adev->have_atomics_support = true; 3900 } else { 3901 adev->have_atomics_support = 3902 !pci_enable_atomic_ops_to_root(adev->pdev, 3903 PCI_EXP_DEVCAP2_ATOMIC_COMP32 | 3904 PCI_EXP_DEVCAP2_ATOMIC_COMP64); 3905 } 3906 3907 if (!adev->have_atomics_support) 3908 dev_info(adev->dev, "PCIE atomic ops is not supported\n"); 3909 3910 /* doorbell bar mapping and doorbell index init*/ 3911 amdgpu_doorbell_init(adev); 3912 3913 if (amdgpu_emu_mode == 1) { 3914 /* post the asic on emulation mode */ 3915 emu_soc_asic_init(adev); 3916 goto fence_driver_init; 3917 } 3918 3919 amdgpu_reset_init(adev); 3920 3921 /* detect if we are with an SRIOV vbios */ 3922 if (adev->bios) 3923 amdgpu_device_detect_sriov_bios(adev); 3924 3925 /* check if we need to reset the asic 3926 * E.g., driver was not cleanly unloaded previously, etc. 3927 */ 3928 if (!amdgpu_sriov_vf(adev) && amdgpu_asic_need_reset_on_init(adev)) { 3929 if (adev->gmc.xgmi.num_physical_nodes) { 3930 dev_info(adev->dev, "Pending hive reset.\n"); 3931 amdgpu_set_init_level(adev, 3932 AMDGPU_INIT_LEVEL_MINIMAL_XGMI); 3933 } else { 3934 tmp = amdgpu_reset_method; 3935 /* It should do a default reset when loading or reloading the driver, 3936 * regardless of the module parameter reset_method. 3937 */ 3938 amdgpu_reset_method = AMD_RESET_METHOD_NONE; 3939 r = amdgpu_asic_reset(adev); 3940 amdgpu_reset_method = tmp; 3941 } 3942 3943 if (r) { 3944 dev_err(adev->dev, "asic reset on init failed\n"); 3945 goto failed; 3946 } 3947 } 3948 3949 /* Post card if necessary */ 3950 if (amdgpu_device_need_post(adev)) { 3951 if (!adev->bios) { 3952 dev_err(adev->dev, "no vBIOS found\n"); 3953 r = -EINVAL; 3954 goto failed; 3955 } 3956 dev_info(adev->dev, "GPU posting now...\n"); 3957 r = amdgpu_device_asic_init(adev); 3958 if (r) { 3959 dev_err(adev->dev, "gpu post error!\n"); 3960 goto failed; 3961 } 3962 } 3963 3964 if (adev->bios) { 3965 if (adev->is_atom_fw) { 3966 /* Initialize clocks */ 3967 r = amdgpu_atomfirmware_get_clock_info(adev); 3968 if (r) { 3969 dev_err(adev->dev, "amdgpu_atomfirmware_get_clock_info failed\n"); 3970 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_ATOMBIOS_GET_CLOCK_FAIL, 0, 0); 3971 goto failed; 3972 } 3973 } else { 3974 /* Initialize clocks */ 3975 r = amdgpu_atombios_get_clock_info(adev); 3976 if (r) { 3977 dev_err(adev->dev, "amdgpu_atombios_get_clock_info failed\n"); 3978 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_ATOMBIOS_GET_CLOCK_FAIL, 0, 0); 3979 goto failed; 3980 } 3981 /* init i2c buses */ 3982 amdgpu_i2c_init(adev); 3983 } 3984 } 3985 3986 fence_driver_init: 3987 /* Fence driver */ 3988 r = amdgpu_fence_driver_sw_init(adev); 3989 if (r) { 3990 dev_err(adev->dev, "amdgpu_fence_driver_sw_init failed\n"); 3991 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_FENCE_INIT_FAIL, 0, 0); 3992 goto failed; 3993 } 3994 3995 /* init the mode config */ 3996 drm_mode_config_init(adev_to_drm(adev)); 3997 3998 r = amdgpu_device_ip_init(adev); 3999 if (r) { 4000 dev_err(adev->dev, "amdgpu_device_ip_init failed\n"); 4001 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_AMDGPU_INIT_FAIL, 0, 0); 4002 goto release_ras_con; 4003 } 4004 4005 amdgpu_fence_driver_hw_init(adev); 4006 4007 dev_info(adev->dev, 4008 "SE %d, SH per SE %d, CU per SH %d, active_cu_number %d\n", 4009 adev->gfx.config.max_shader_engines, 4010 adev->gfx.config.max_sh_per_se, 4011 adev->gfx.config.max_cu_per_sh, 4012 adev->gfx.cu_info.number); 4013 4014 adev->accel_working = true; 4015 4016 amdgpu_vm_check_compute_bug(adev); 4017 4018 /* Initialize the buffer migration limit. */ 4019 if (amdgpu_moverate >= 0) 4020 max_MBps = amdgpu_moverate; 4021 else 4022 max_MBps = 8; /* Allow 8 MB/s. */ 4023 /* Get a log2 for easy divisions. */ 4024 adev->mm_stats.log2_max_MBps = ilog2(max(1u, max_MBps)); 4025 4026 /* 4027 * Register gpu instance before amdgpu_device_enable_mgpu_fan_boost. 4028 * Otherwise the mgpu fan boost feature will be skipped due to the 4029 * gpu instance is counted less. 4030 */ 4031 amdgpu_register_gpu_instance(adev); 4032 4033 /* enable clockgating, etc. after ib tests, etc. since some blocks require 4034 * explicit gating rather than handling it automatically. 4035 */ 4036 if (adev->init_lvl->level != AMDGPU_INIT_LEVEL_MINIMAL_XGMI) { 4037 r = amdgpu_device_ip_late_init(adev); 4038 if (r) { 4039 dev_err(adev->dev, "amdgpu_device_ip_late_init failed\n"); 4040 amdgpu_vf_error_put(adev, AMDGIM_ERROR_VF_AMDGPU_LATE_INIT_FAIL, 0, r); 4041 goto release_ras_con; 4042 } 4043 /* must succeed. */ 4044 amdgpu_ras_resume(adev); 4045 queue_delayed_work(system_wq, &adev->delayed_init_work, 4046 msecs_to_jiffies(AMDGPU_RESUME_MS)); 4047 } 4048 4049 if (amdgpu_sriov_vf(adev)) { 4050 amdgpu_virt_release_full_gpu(adev, true); 4051 flush_delayed_work(&adev->delayed_init_work); 4052 } 4053 4054 /* Don't init kfd if whole hive need to be reset during init */ 4055 if (adev->init_lvl->level != AMDGPU_INIT_LEVEL_MINIMAL_XGMI) { 4056 kgd2kfd_init_zone_device(adev); 4057 kfd_update_svm_support_properties(adev); 4058 } 4059 4060 if (adev->init_lvl->level == AMDGPU_INIT_LEVEL_MINIMAL_XGMI) 4061 amdgpu_xgmi_reset_on_init(adev); 4062 4063 /* 4064 * Place those sysfs registering after `late_init`. As some of those 4065 * operations performed in `late_init` might affect the sysfs 4066 * interfaces creating. 4067 */ 4068 r = amdgpu_device_sys_interface_init(adev); 4069 4070 if (IS_ENABLED(CONFIG_PERF_EVENTS)) 4071 r = amdgpu_pmu_init(adev); 4072 if (r) 4073 dev_err(adev->dev, "amdgpu_pmu_init failed\n"); 4074 4075 /* Have stored pci confspace at hand for restore in sudden PCI error */ 4076 if (amdgpu_device_cache_pci_state(adev->pdev)) 4077 pci_restore_state(pdev); 4078 4079 /* if we have > 1 VGA cards, then disable the amdgpu VGA resources */ 4080 /* this will fail for cards that aren't VGA class devices, just 4081 * ignore it 4082 */ 4083 if ((adev->pdev->class >> 8) == PCI_CLASS_DISPLAY_VGA) 4084 vga_client_register(adev->pdev, amdgpu_device_vga_set_decode); 4085 4086 px = amdgpu_device_supports_px(adev); 4087 4088 if (px || (!dev_is_removable(&adev->pdev->dev) && 4089 apple_gmux_detect(NULL, NULL))) 4090 vga_switcheroo_register_client(adev->pdev, 4091 &amdgpu_switcheroo_ops, px); 4092 4093 if (px) 4094 vga_switcheroo_init_domain_pm_ops(adev->dev, &adev->vga_pm_domain); 4095 4096 amdgpu_device_check_iommu_direct_map(adev); 4097 4098 adev->pm_nb.notifier_call = amdgpu_device_pm_notifier; 4099 r = register_pm_notifier(&adev->pm_nb); 4100 if (r) 4101 goto failed; 4102 4103 return 0; 4104 4105 release_ras_con: 4106 if (amdgpu_sriov_vf(adev)) 4107 amdgpu_virt_release_full_gpu(adev, true); 4108 4109 /* failed in exclusive mode due to timeout */ 4110 if (amdgpu_sriov_vf(adev) && 4111 !amdgpu_sriov_runtime(adev) && 4112 amdgpu_virt_mmio_blocked(adev) && 4113 !amdgpu_virt_wait_reset(adev)) { 4114 dev_err(adev->dev, "VF exclusive mode timeout\n"); 4115 /* Don't send request since VF is inactive. */ 4116 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; 4117 adev->virt.ops = NULL; 4118 r = -EAGAIN; 4119 } 4120 amdgpu_release_ras_context(adev); 4121 4122 failed: 4123 amdgpu_vf_error_trans_all(adev); 4124 4125 return r; 4126 } 4127 4128 static void amdgpu_device_unmap_mmio(struct amdgpu_device *adev) 4129 { 4130 4131 /* Clear all CPU mappings pointing to this device */ 4132 unmap_mapping_range(adev->ddev.anon_inode->i_mapping, 0, 0, 1); 4133 4134 /* Unmap all mapped bars - Doorbell, registers and VRAM */ 4135 amdgpu_doorbell_fini(adev); 4136 4137 iounmap(adev->rmmio); 4138 adev->rmmio = NULL; 4139 if (adev->mman.aper_base_kaddr) 4140 iounmap(adev->mman.aper_base_kaddr); 4141 adev->mman.aper_base_kaddr = NULL; 4142 4143 /* Memory manager related */ 4144 if (!adev->gmc.xgmi.connected_to_cpu && !adev->gmc.is_app_apu) { 4145 arch_phys_wc_del(adev->gmc.vram_mtrr); 4146 arch_io_free_memtype_wc(adev->gmc.aper_base, adev->gmc.aper_size); 4147 } 4148 } 4149 4150 /** 4151 * amdgpu_device_fini_hw - tear down the driver 4152 * 4153 * @adev: amdgpu_device pointer 4154 * 4155 * Tear down the driver info (all asics). 4156 * Called at driver shutdown. 4157 */ 4158 void amdgpu_device_fini_hw(struct amdgpu_device *adev) 4159 { 4160 dev_info(adev->dev, "finishing device.\n"); 4161 flush_delayed_work(&adev->delayed_init_work); 4162 4163 if (adev->mman.initialized) 4164 drain_workqueue(adev->mman.bdev.wq); 4165 adev->shutdown = true; 4166 4167 unregister_pm_notifier(&adev->pm_nb); 4168 4169 /* make sure IB test finished before entering exclusive mode 4170 * to avoid preemption on IB test 4171 */ 4172 if (amdgpu_sriov_vf(adev)) { 4173 amdgpu_virt_request_full_gpu(adev, false); 4174 amdgpu_virt_fini_data_exchange(adev); 4175 } 4176 4177 amdgpu_device_set_pg_state(adev, AMD_PG_STATE_UNGATE); 4178 amdgpu_device_set_cg_state(adev, AMD_CG_STATE_UNGATE); 4179 4180 /* disable all interrupts */ 4181 amdgpu_irq_disable_all(adev); 4182 if (adev->mode_info.mode_config_initialized) { 4183 if (!drm_drv_uses_atomic_modeset(adev_to_drm(adev))) 4184 drm_helper_force_disable_all(adev_to_drm(adev)); 4185 else 4186 drm_atomic_helper_shutdown(adev_to_drm(adev)); 4187 } 4188 amdgpu_fence_driver_hw_fini(adev); 4189 4190 amdgpu_device_sys_interface_fini(adev); 4191 4192 /* disable ras feature must before hw fini */ 4193 amdgpu_ras_pre_fini(adev); 4194 4195 amdgpu_ttm_set_buffer_funcs_status(adev, false); 4196 4197 /* 4198 * device went through surprise hotplug; we need to destroy topology 4199 * before ip_fini_early to prevent kfd locking refcount issues by calling 4200 * amdgpu_amdkfd_suspend() 4201 */ 4202 if (pci_dev_is_disconnected(adev->pdev)) 4203 amdgpu_amdkfd_device_fini_sw(adev); 4204 4205 amdgpu_device_ip_fini_early(adev); 4206 4207 amdgpu_irq_fini_hw(adev); 4208 4209 if (adev->mman.initialized) 4210 ttm_device_clear_dma_mappings(&adev->mman.bdev); 4211 4212 amdgpu_gart_dummy_page_fini(adev); 4213 4214 if (pci_dev_is_disconnected(adev->pdev)) 4215 amdgpu_device_unmap_mmio(adev); 4216 4217 } 4218 4219 void amdgpu_device_fini_sw(struct amdgpu_device *adev) 4220 { 4221 int i, idx; 4222 bool px; 4223 4224 amdgpu_device_ip_fini(adev); 4225 amdgpu_fence_driver_sw_fini(adev); 4226 amdgpu_ucode_release(&adev->firmware.gpu_info_fw); 4227 adev->accel_working = false; 4228 dma_fence_put(rcu_dereference_protected(adev->gang_submit, true)); 4229 for (i = 0; i < MAX_XCP; ++i) { 4230 dma_fence_put(adev->isolation[i].spearhead); 4231 amdgpu_sync_free(&adev->isolation[i].active); 4232 amdgpu_sync_free(&adev->isolation[i].prev); 4233 } 4234 4235 amdgpu_reset_fini(adev); 4236 4237 /* free i2c buses */ 4238 amdgpu_i2c_fini(adev); 4239 4240 if (adev->bios) { 4241 if (amdgpu_emu_mode != 1) 4242 amdgpu_atombios_fini(adev); 4243 amdgpu_bios_release(adev); 4244 } 4245 4246 kfree(adev->fru_info); 4247 adev->fru_info = NULL; 4248 4249 kfree(adev->xcp_mgr); 4250 adev->xcp_mgr = NULL; 4251 4252 px = amdgpu_device_supports_px(adev); 4253 4254 if (px || (!dev_is_removable(&adev->pdev->dev) && 4255 apple_gmux_detect(NULL, NULL))) 4256 vga_switcheroo_unregister_client(adev->pdev); 4257 4258 if (px) 4259 vga_switcheroo_fini_domain_pm_ops(adev->dev); 4260 4261 if ((adev->pdev->class >> 8) == PCI_CLASS_DISPLAY_VGA) 4262 vga_client_unregister(adev->pdev); 4263 4264 if (drm_dev_enter(adev_to_drm(adev), &idx)) { 4265 4266 iounmap(adev->rmmio); 4267 adev->rmmio = NULL; 4268 drm_dev_exit(idx); 4269 } 4270 4271 if (IS_ENABLED(CONFIG_PERF_EVENTS)) 4272 amdgpu_pmu_fini(adev); 4273 if (adev->discovery.bin) 4274 amdgpu_discovery_fini(adev); 4275 4276 amdgpu_reset_put_reset_domain(adev->reset_domain); 4277 adev->reset_domain = NULL; 4278 4279 kfree(adev->pci_state); 4280 kfree(adev->pcie_reset_ctx.swds_pcistate); 4281 kfree(adev->pcie_reset_ctx.swus_pcistate); 4282 } 4283 4284 /** 4285 * amdgpu_device_evict_resources - evict device resources 4286 * @adev: amdgpu device object 4287 * 4288 * Evicts all ttm device resources(vram BOs, gart table) from the lru list 4289 * of the vram memory type. Mainly used for evicting device resources 4290 * at suspend time. 4291 * 4292 */ 4293 static int amdgpu_device_evict_resources(struct amdgpu_device *adev) 4294 { 4295 int ret; 4296 4297 /* No need to evict vram on APUs unless going to S4 */ 4298 if (!adev->in_s4 && (adev->flags & AMD_IS_APU)) 4299 return 0; 4300 4301 /* No need to evict when going to S5 through S4 callbacks */ 4302 if (system_state == SYSTEM_POWER_OFF) 4303 return 0; 4304 4305 ret = amdgpu_ttm_evict_resources(adev, TTM_PL_VRAM); 4306 if (ret) { 4307 dev_warn(adev->dev, "evicting device resources failed\n"); 4308 return ret; 4309 } 4310 4311 if (adev->in_s4) { 4312 ret = ttm_device_prepare_hibernation(&adev->mman.bdev); 4313 if (ret) 4314 dev_err(adev->dev, "prepare hibernation failed, %d\n", ret); 4315 } 4316 return ret; 4317 } 4318 4319 /* 4320 * Suspend & resume. 4321 */ 4322 /** 4323 * amdgpu_device_pm_notifier - Notification block for Suspend/Hibernate events 4324 * @nb: notifier block 4325 * @mode: suspend mode 4326 * @data: data 4327 * 4328 * This function is called when the system is about to suspend or hibernate. 4329 * It is used to set the appropriate flags so that eviction can be optimized 4330 * in the pm prepare callback. 4331 */ 4332 static int amdgpu_device_pm_notifier(struct notifier_block *nb, unsigned long mode, 4333 void *data) 4334 { 4335 struct amdgpu_device *adev = container_of(nb, struct amdgpu_device, pm_nb); 4336 4337 switch (mode) { 4338 case PM_HIBERNATION_PREPARE: 4339 adev->in_s4 = true; 4340 break; 4341 case PM_POST_HIBERNATION: 4342 adev->in_s4 = false; 4343 break; 4344 } 4345 4346 return NOTIFY_DONE; 4347 } 4348 4349 /** 4350 * amdgpu_device_prepare - prepare for device suspend 4351 * 4352 * @dev: drm dev pointer 4353 * 4354 * Prepare to put the hw in the suspend state (all asics). 4355 * Returns 0 for success or an error on failure. 4356 * Called at driver suspend. 4357 */ 4358 int amdgpu_device_prepare(struct drm_device *dev) 4359 { 4360 struct amdgpu_device *adev = drm_to_adev(dev); 4361 int i, r; 4362 4363 if (dev->switch_power_state == DRM_SWITCH_POWER_OFF) 4364 return 0; 4365 4366 /* Evict the majority of BOs before starting suspend sequence */ 4367 r = amdgpu_device_evict_resources(adev); 4368 if (r) 4369 return r; 4370 4371 flush_delayed_work(&adev->gfx.gfx_off_delay_work); 4372 4373 for (i = 0; i < adev->num_ip_blocks; i++) { 4374 if (!adev->ip_blocks[i].status.valid) 4375 continue; 4376 if (!adev->ip_blocks[i].version->funcs->prepare_suspend) 4377 continue; 4378 r = adev->ip_blocks[i].version->funcs->prepare_suspend(&adev->ip_blocks[i]); 4379 if (r) 4380 return r; 4381 } 4382 4383 return 0; 4384 } 4385 4386 /** 4387 * amdgpu_device_complete - complete power state transition 4388 * 4389 * @dev: drm dev pointer 4390 * 4391 * Undo the changes from amdgpu_device_prepare. This will be 4392 * called on all resume transitions, including those that failed. 4393 */ 4394 void amdgpu_device_complete(struct drm_device *dev) 4395 { 4396 struct amdgpu_device *adev = drm_to_adev(dev); 4397 int i; 4398 4399 for (i = 0; i < adev->num_ip_blocks; i++) { 4400 if (!adev->ip_blocks[i].status.valid) 4401 continue; 4402 if (!adev->ip_blocks[i].version->funcs->complete) 4403 continue; 4404 adev->ip_blocks[i].version->funcs->complete(&adev->ip_blocks[i]); 4405 } 4406 } 4407 4408 /** 4409 * amdgpu_device_suspend - initiate device suspend 4410 * 4411 * @dev: drm dev pointer 4412 * @notify_clients: notify in-kernel DRM clients 4413 * 4414 * Puts the hw in the suspend state (all asics). 4415 * Returns 0 for success or an error on failure. 4416 * Called at driver suspend. 4417 */ 4418 int amdgpu_device_suspend(struct drm_device *dev, bool notify_clients) 4419 { 4420 struct amdgpu_device *adev = drm_to_adev(dev); 4421 int r, rec; 4422 4423 if (dev->switch_power_state == DRM_SWITCH_POWER_OFF) 4424 return 0; 4425 4426 adev->in_suspend = true; 4427 4428 if (amdgpu_sriov_vf(adev)) { 4429 if (!adev->in_runpm) 4430 amdgpu_amdkfd_suspend_process(adev); 4431 amdgpu_virt_fini_data_exchange(adev); 4432 r = amdgpu_virt_request_full_gpu(adev, false); 4433 if (r) 4434 return r; 4435 } 4436 4437 r = amdgpu_acpi_smart_shift_update(adev, AMDGPU_SS_DEV_D3); 4438 if (r) 4439 goto unwind_sriov; 4440 4441 if (notify_clients) 4442 drm_client_dev_suspend(adev_to_drm(adev)); 4443 4444 cancel_delayed_work_sync(&adev->delayed_init_work); 4445 4446 amdgpu_ras_suspend(adev); 4447 4448 r = amdgpu_device_ip_suspend_phase1(adev); 4449 if (r) 4450 goto unwind_smartshift; 4451 4452 amdgpu_amdkfd_suspend(adev, !amdgpu_sriov_vf(adev) && !adev->in_runpm); 4453 r = amdgpu_userq_suspend(adev); 4454 if (r) 4455 goto unwind_ip_phase1; 4456 4457 r = amdgpu_device_evict_resources(adev); 4458 if (r) 4459 goto unwind_userq; 4460 4461 amdgpu_ttm_set_buffer_funcs_status(adev, false); 4462 4463 amdgpu_fence_driver_hw_fini(adev); 4464 4465 r = amdgpu_device_ip_suspend_phase2(adev); 4466 if (r) 4467 goto unwind_evict; 4468 4469 if (amdgpu_sriov_vf(adev)) 4470 amdgpu_virt_release_full_gpu(adev, false); 4471 4472 return 0; 4473 4474 unwind_evict: 4475 amdgpu_ttm_set_buffer_funcs_status(adev, true); 4476 amdgpu_fence_driver_hw_init(adev); 4477 4478 unwind_userq: 4479 rec = amdgpu_userq_resume(adev); 4480 if (rec) { 4481 dev_warn(adev->dev, "failed to re-initialize user queues: %d\n", rec); 4482 return r; 4483 } 4484 rec = amdgpu_amdkfd_resume(adev, !amdgpu_sriov_vf(adev) && !adev->in_runpm); 4485 if (rec) { 4486 dev_warn(adev->dev, "failed to re-initialize kfd: %d\n", rec); 4487 return r; 4488 } 4489 4490 unwind_ip_phase1: 4491 /* suspend phase 1 = resume phase 3 */ 4492 rec = amdgpu_device_ip_resume_phase3(adev); 4493 if (rec) { 4494 dev_warn(adev->dev, "failed to re-initialize IPs phase1: %d\n", rec); 4495 return r; 4496 } 4497 4498 unwind_smartshift: 4499 rec = amdgpu_acpi_smart_shift_update(adev, AMDGPU_SS_DEV_D0); 4500 if (rec) { 4501 dev_warn(adev->dev, "failed to re-update smart shift: %d\n", rec); 4502 return r; 4503 } 4504 4505 if (notify_clients) 4506 drm_client_dev_resume(adev_to_drm(adev)); 4507 4508 amdgpu_ras_resume(adev); 4509 4510 unwind_sriov: 4511 if (amdgpu_sriov_vf(adev)) { 4512 rec = amdgpu_virt_request_full_gpu(adev, true); 4513 if (rec) { 4514 dev_warn(adev->dev, "failed to reinitialize sriov: %d\n", rec); 4515 return r; 4516 } 4517 } 4518 4519 adev->in_suspend = adev->in_s0ix = adev->in_s3 = false; 4520 4521 return r; 4522 } 4523 4524 static inline int amdgpu_virt_resume(struct amdgpu_device *adev) 4525 { 4526 int r; 4527 unsigned int prev_physical_node_id = adev->gmc.xgmi.physical_node_id; 4528 4529 /* During VM resume, QEMU programming of VF MSIX table (register GFXMSIX_VECT0_ADDR_LO) 4530 * may not work. The access could be blocked by nBIF protection as VF isn't in 4531 * exclusive access mode. Exclusive access is enabled now, disable/enable MSIX 4532 * so that QEMU reprograms MSIX table. 4533 */ 4534 amdgpu_restore_msix(adev); 4535 4536 r = adev->gfxhub.funcs->get_xgmi_info(adev); 4537 if (r) 4538 return r; 4539 4540 dev_info(adev->dev, "xgmi node, old id %d, new id %d\n", 4541 prev_physical_node_id, adev->gmc.xgmi.physical_node_id); 4542 4543 adev->vm_manager.vram_base_offset = adev->gfxhub.funcs->get_mc_fb_offset(adev); 4544 adev->vm_manager.vram_base_offset += 4545 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; 4546 4547 return 0; 4548 } 4549 4550 /** 4551 * amdgpu_device_resume - initiate device resume 4552 * 4553 * @dev: drm dev pointer 4554 * @notify_clients: notify in-kernel DRM clients 4555 * 4556 * Bring the hw back to operating state (all asics). 4557 * Returns 0 for success or an error on failure. 4558 * Called at driver resume. 4559 */ 4560 int amdgpu_device_resume(struct drm_device *dev, bool notify_clients) 4561 { 4562 struct amdgpu_device *adev = drm_to_adev(dev); 4563 int r = 0; 4564 4565 if (amdgpu_sriov_vf(adev)) { 4566 r = amdgpu_virt_request_full_gpu(adev, true); 4567 if (r) 4568 return r; 4569 } 4570 4571 if (amdgpu_virt_xgmi_migrate_enabled(adev)) { 4572 r = amdgpu_virt_resume(adev); 4573 if (r) 4574 goto exit; 4575 } 4576 4577 if (dev->switch_power_state == DRM_SWITCH_POWER_OFF) 4578 return 0; 4579 4580 if (adev->in_s0ix) 4581 amdgpu_dpm_gfx_state_change(adev, sGpuChangeState_D0Entry); 4582 4583 /* post card */ 4584 if (amdgpu_device_need_post(adev)) { 4585 r = amdgpu_device_asic_init(adev); 4586 if (r) 4587 dev_err(adev->dev, "amdgpu asic init failed\n"); 4588 } 4589 4590 r = amdgpu_device_ip_resume(adev); 4591 4592 if (r) { 4593 dev_err(adev->dev, "amdgpu_device_ip_resume failed (%d).\n", r); 4594 goto exit; 4595 } 4596 4597 r = amdgpu_amdkfd_resume(adev, !amdgpu_sriov_vf(adev) && !adev->in_runpm); 4598 if (r) 4599 goto exit; 4600 4601 r = amdgpu_userq_resume(adev); 4602 if (r) 4603 goto exit; 4604 4605 r = amdgpu_device_ip_late_init(adev); 4606 if (r) 4607 goto exit; 4608 4609 queue_delayed_work(system_wq, &adev->delayed_init_work, 4610 msecs_to_jiffies(AMDGPU_RESUME_MS)); 4611 exit: 4612 if (amdgpu_sriov_vf(adev)) { 4613 amdgpu_virt_init_data_exchange(adev); 4614 amdgpu_virt_release_full_gpu(adev, true); 4615 4616 if (!r && !adev->in_runpm) 4617 r = amdgpu_amdkfd_resume_process(adev); 4618 } 4619 4620 if (r) 4621 return r; 4622 4623 /* Make sure IB tests flushed */ 4624 flush_delayed_work(&adev->delayed_init_work); 4625 4626 if (notify_clients) 4627 drm_client_dev_resume(adev_to_drm(adev)); 4628 4629 amdgpu_ras_resume(adev); 4630 4631 if (adev->mode_info.num_crtc) { 4632 /* 4633 * Most of the connector probing functions try to acquire runtime pm 4634 * refs to ensure that the GPU is powered on when connector polling is 4635 * performed. Since we're calling this from a runtime PM callback, 4636 * trying to acquire rpm refs will cause us to deadlock. 4637 * 4638 * Since we're guaranteed to be holding the rpm lock, it's safe to 4639 * temporarily disable the rpm helpers so this doesn't deadlock us. 4640 */ 4641 #ifdef CONFIG_PM 4642 dev->dev->power.disable_depth++; 4643 #endif 4644 if (!adev->dc_enabled) 4645 drm_helper_hpd_irq_event(dev); 4646 else 4647 drm_kms_helper_hotplug_event(dev); 4648 #ifdef CONFIG_PM 4649 dev->dev->power.disable_depth--; 4650 #endif 4651 } 4652 4653 amdgpu_vram_mgr_clear_reset_blocks(adev); 4654 adev->in_suspend = false; 4655 4656 if (amdgpu_acpi_smart_shift_update(adev, AMDGPU_SS_DEV_D0)) 4657 dev_warn(adev->dev, "smart shift update failed\n"); 4658 4659 return 0; 4660 } 4661 4662 /** 4663 * amdgpu_device_ip_check_soft_reset - did soft reset succeed 4664 * 4665 * @adev: amdgpu_device pointer 4666 * 4667 * The list of all the hardware IPs that make up the asic is walked and 4668 * the check_soft_reset callbacks are run. check_soft_reset determines 4669 * if the asic is still hung or not. 4670 * Returns true if any of the IPs are still in a hung state, false if not. 4671 */ 4672 static bool amdgpu_device_ip_check_soft_reset(struct amdgpu_device *adev) 4673 { 4674 int i; 4675 bool asic_hang = false; 4676 4677 if (amdgpu_sriov_vf(adev)) 4678 return true; 4679 4680 if (amdgpu_asic_need_full_reset(adev)) 4681 return true; 4682 4683 for (i = 0; i < adev->num_ip_blocks; i++) { 4684 if (!adev->ip_blocks[i].status.valid) 4685 continue; 4686 if (adev->ip_blocks[i].version->funcs->check_soft_reset) 4687 adev->ip_blocks[i].status.hang = 4688 adev->ip_blocks[i].version->funcs->check_soft_reset( 4689 &adev->ip_blocks[i]); 4690 if (adev->ip_blocks[i].status.hang) { 4691 dev_info(adev->dev, "IP block:%s is hung!\n", adev->ip_blocks[i].version->funcs->name); 4692 asic_hang = true; 4693 } 4694 } 4695 return asic_hang; 4696 } 4697 4698 /** 4699 * amdgpu_device_ip_pre_soft_reset - prepare for soft reset 4700 * 4701 * @adev: amdgpu_device pointer 4702 * 4703 * The list of all the hardware IPs that make up the asic is walked and the 4704 * pre_soft_reset callbacks are run if the block is hung. pre_soft_reset 4705 * handles any IP specific hardware or software state changes that are 4706 * necessary for a soft reset to succeed. 4707 * Returns 0 on success, negative error code on failure. 4708 */ 4709 static int amdgpu_device_ip_pre_soft_reset(struct amdgpu_device *adev) 4710 { 4711 int i, r = 0; 4712 4713 for (i = 0; i < adev->num_ip_blocks; i++) { 4714 if (!adev->ip_blocks[i].status.valid) 4715 continue; 4716 if (adev->ip_blocks[i].status.hang && 4717 adev->ip_blocks[i].version->funcs->pre_soft_reset) { 4718 r = adev->ip_blocks[i].version->funcs->pre_soft_reset(&adev->ip_blocks[i]); 4719 if (r) 4720 return r; 4721 } 4722 } 4723 4724 return 0; 4725 } 4726 4727 /** 4728 * amdgpu_device_ip_need_full_reset - check if a full asic reset is needed 4729 * 4730 * @adev: amdgpu_device pointer 4731 * 4732 * Some hardware IPs cannot be soft reset. If they are hung, a full gpu 4733 * reset is necessary to recover. 4734 * Returns true if a full asic reset is required, false if not. 4735 */ 4736 static bool amdgpu_device_ip_need_full_reset(struct amdgpu_device *adev) 4737 { 4738 int i; 4739 4740 if (amdgpu_asic_need_full_reset(adev)) 4741 return true; 4742 4743 for (i = 0; i < adev->num_ip_blocks; i++) { 4744 if (!adev->ip_blocks[i].status.valid) 4745 continue; 4746 if ((adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_GMC) || 4747 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_SMC) || 4748 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_ACP) || 4749 (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_DCE) || 4750 adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP) { 4751 if (adev->ip_blocks[i].status.hang) { 4752 dev_info(adev->dev, "Some block need full reset!\n"); 4753 return true; 4754 } 4755 } 4756 } 4757 return false; 4758 } 4759 4760 /** 4761 * amdgpu_device_ip_soft_reset - do a soft reset 4762 * 4763 * @adev: amdgpu_device pointer 4764 * 4765 * The list of all the hardware IPs that make up the asic is walked and the 4766 * soft_reset callbacks are run if the block is hung. soft_reset handles any 4767 * IP specific hardware or software state changes that are necessary to soft 4768 * reset the IP. 4769 * Returns 0 on success, negative error code on failure. 4770 */ 4771 static int amdgpu_device_ip_soft_reset(struct amdgpu_device *adev) 4772 { 4773 int i, r = 0; 4774 4775 for (i = 0; i < adev->num_ip_blocks; i++) { 4776 if (!adev->ip_blocks[i].status.valid) 4777 continue; 4778 if (adev->ip_blocks[i].status.hang && 4779 adev->ip_blocks[i].version->funcs->soft_reset) { 4780 r = adev->ip_blocks[i].version->funcs->soft_reset(&adev->ip_blocks[i]); 4781 if (r) 4782 return r; 4783 } 4784 } 4785 4786 return 0; 4787 } 4788 4789 /** 4790 * amdgpu_device_ip_post_soft_reset - clean up from soft reset 4791 * 4792 * @adev: amdgpu_device pointer 4793 * 4794 * The list of all the hardware IPs that make up the asic is walked and the 4795 * post_soft_reset callbacks are run if the asic was hung. post_soft_reset 4796 * handles any IP specific hardware or software state changes that are 4797 * necessary after the IP has been soft reset. 4798 * Returns 0 on success, negative error code on failure. 4799 */ 4800 static int amdgpu_device_ip_post_soft_reset(struct amdgpu_device *adev) 4801 { 4802 int i, r = 0; 4803 4804 for (i = 0; i < adev->num_ip_blocks; i++) { 4805 if (!adev->ip_blocks[i].status.valid) 4806 continue; 4807 if (adev->ip_blocks[i].status.hang && 4808 adev->ip_blocks[i].version->funcs->post_soft_reset) 4809 r = adev->ip_blocks[i].version->funcs->post_soft_reset(&adev->ip_blocks[i]); 4810 if (r) 4811 return r; 4812 } 4813 4814 return 0; 4815 } 4816 4817 /** 4818 * amdgpu_device_reset_sriov - reset ASIC for SR-IOV vf 4819 * 4820 * @adev: amdgpu_device pointer 4821 * @reset_context: amdgpu reset context pointer 4822 * 4823 * do VF FLR and reinitialize Asic 4824 * return 0 means succeeded otherwise failed 4825 */ 4826 static int amdgpu_device_reset_sriov(struct amdgpu_device *adev, 4827 struct amdgpu_reset_context *reset_context) 4828 { 4829 int r; 4830 struct amdgpu_hive_info *hive = NULL; 4831 4832 if (test_bit(AMDGPU_HOST_FLR, &reset_context->flags)) { 4833 if (!amdgpu_ras_get_fed_status(adev)) 4834 amdgpu_virt_ready_to_reset(adev); 4835 amdgpu_virt_wait_reset(adev); 4836 clear_bit(AMDGPU_HOST_FLR, &reset_context->flags); 4837 r = amdgpu_virt_request_full_gpu(adev, true); 4838 } else { 4839 r = amdgpu_virt_reset_gpu(adev); 4840 } 4841 if (r) 4842 return r; 4843 4844 amdgpu_ras_clear_err_state(adev); 4845 amdgpu_irq_gpu_reset_resume_helper(adev); 4846 4847 /* some sw clean up VF needs to do before recover */ 4848 amdgpu_virt_post_reset(adev); 4849 4850 /* Resume IP prior to SMC */ 4851 r = amdgpu_device_ip_reinit_early_sriov(adev); 4852 if (r) 4853 return r; 4854 4855 amdgpu_virt_init_data_exchange(adev); 4856 4857 r = amdgpu_device_fw_loading(adev); 4858 if (r) 4859 return r; 4860 4861 /* now we are okay to resume SMC/CP/SDMA */ 4862 r = amdgpu_device_ip_reinit_late_sriov(adev); 4863 if (r) 4864 return r; 4865 4866 hive = amdgpu_get_xgmi_hive(adev); 4867 /* Update PSP FW topology after reset */ 4868 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) 4869 r = amdgpu_xgmi_update_topology(hive, adev); 4870 if (hive) 4871 amdgpu_put_xgmi_hive(hive); 4872 if (r) 4873 return r; 4874 4875 r = amdgpu_ib_ring_tests(adev); 4876 if (r) 4877 return r; 4878 4879 if (adev->virt.gim_feature & AMDGIM_FEATURE_GIM_FLR_VRAMLOST) 4880 amdgpu_inc_vram_lost(adev); 4881 4882 /* need to be called during full access so we can't do it later like 4883 * bare-metal does. 4884 */ 4885 amdgpu_amdkfd_post_reset(adev); 4886 amdgpu_virt_release_full_gpu(adev, true); 4887 4888 /* Aldebaran and gfx_11_0_3 support ras in SRIOV, so need resume ras during reset */ 4889 if (amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 4, 2) || 4890 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 4, 3) || 4891 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 4, 4) || 4892 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(9, 5, 0) || 4893 amdgpu_ip_version(adev, GC_HWIP, 0) == IP_VERSION(11, 0, 3)) 4894 amdgpu_ras_resume(adev); 4895 4896 amdgpu_virt_ras_telemetry_post_reset(adev); 4897 4898 return 0; 4899 } 4900 4901 /** 4902 * amdgpu_device_has_job_running - check if there is any unfinished job 4903 * 4904 * @adev: amdgpu_device pointer 4905 * 4906 * check if there is any job running on the device when guest driver receives 4907 * FLR notification from host driver. If there are still jobs running, then 4908 * the guest driver will not respond the FLR reset. Instead, let the job hit 4909 * the timeout and guest driver then issue the reset request. 4910 */ 4911 bool amdgpu_device_has_job_running(struct amdgpu_device *adev) 4912 { 4913 int i; 4914 4915 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 4916 struct amdgpu_ring *ring = adev->rings[i]; 4917 4918 if (!amdgpu_ring_sched_ready(ring)) 4919 continue; 4920 4921 if (amdgpu_fence_count_emitted(ring)) 4922 return true; 4923 } 4924 return false; 4925 } 4926 4927 /** 4928 * amdgpu_device_should_recover_gpu - check if we should try GPU recovery 4929 * 4930 * @adev: amdgpu_device pointer 4931 * 4932 * Check amdgpu_gpu_recovery and SRIOV status to see if we should try to recover 4933 * a hung GPU. 4934 */ 4935 bool amdgpu_device_should_recover_gpu(struct amdgpu_device *adev) 4936 { 4937 4938 if (amdgpu_gpu_recovery == 0) 4939 goto disabled; 4940 4941 /* Skip soft reset check in fatal error mode */ 4942 if (!amdgpu_ras_is_poison_mode_supported(adev)) 4943 return true; 4944 4945 if (amdgpu_sriov_vf(adev)) 4946 return true; 4947 4948 if (amdgpu_gpu_recovery == -1) { 4949 switch (adev->asic_type) { 4950 #ifdef CONFIG_DRM_AMDGPU_SI 4951 case CHIP_VERDE: 4952 case CHIP_TAHITI: 4953 case CHIP_PITCAIRN: 4954 case CHIP_OLAND: 4955 case CHIP_HAINAN: 4956 #endif 4957 #ifdef CONFIG_DRM_AMDGPU_CIK 4958 case CHIP_KAVERI: 4959 case CHIP_KABINI: 4960 case CHIP_MULLINS: 4961 #endif 4962 case CHIP_CARRIZO: 4963 case CHIP_STONEY: 4964 case CHIP_CYAN_SKILLFISH: 4965 goto disabled; 4966 default: 4967 break; 4968 } 4969 } 4970 4971 return true; 4972 4973 disabled: 4974 dev_info(adev->dev, "GPU recovery disabled.\n"); 4975 return false; 4976 } 4977 4978 int amdgpu_device_mode1_reset(struct amdgpu_device *adev) 4979 { 4980 u32 i; 4981 int ret = 0; 4982 4983 if (adev->bios) 4984 amdgpu_atombios_scratch_regs_engine_hung(adev, true); 4985 4986 dev_info(adev->dev, "GPU mode1 reset\n"); 4987 4988 /* Cache the state before bus master disable. The saved config space 4989 * values are used in other cases like restore after mode-2 reset. 4990 */ 4991 amdgpu_device_cache_pci_state(adev->pdev); 4992 4993 /* disable BM */ 4994 pci_clear_master(adev->pdev); 4995 4996 if (amdgpu_dpm_is_mode1_reset_supported(adev)) { 4997 dev_info(adev->dev, "GPU smu mode1 reset\n"); 4998 ret = amdgpu_dpm_mode1_reset(adev); 4999 } else { 5000 dev_info(adev->dev, "GPU psp mode1 reset\n"); 5001 ret = psp_gpu_reset(adev); 5002 } 5003 5004 if (ret) 5005 goto mode1_reset_failed; 5006 5007 /* enable mmio access after mode 1 reset completed */ 5008 adev->no_hw_access = false; 5009 5010 /* ensure no_hw_access is updated before we access hw */ 5011 smp_mb(); 5012 5013 amdgpu_device_load_pci_state(adev->pdev); 5014 ret = amdgpu_psp_wait_for_bootloader(adev); 5015 if (ret) 5016 goto mode1_reset_failed; 5017 5018 /* wait for asic to come out of reset */ 5019 for (i = 0; i < adev->usec_timeout; i++) { 5020 u32 memsize = adev->nbio.funcs->get_memsize(adev); 5021 5022 if (memsize != 0xffffffff) 5023 break; 5024 udelay(1); 5025 } 5026 5027 if (i >= adev->usec_timeout) { 5028 ret = -ETIMEDOUT; 5029 goto mode1_reset_failed; 5030 } 5031 5032 if (adev->bios) 5033 amdgpu_atombios_scratch_regs_engine_hung(adev, false); 5034 5035 return 0; 5036 5037 mode1_reset_failed: 5038 dev_err(adev->dev, "GPU mode1 reset failed\n"); 5039 return ret; 5040 } 5041 5042 int amdgpu_device_link_reset(struct amdgpu_device *adev) 5043 { 5044 int ret = 0; 5045 5046 dev_info(adev->dev, "GPU link reset\n"); 5047 5048 if (!amdgpu_reset_in_dpc(adev)) 5049 ret = amdgpu_dpm_link_reset(adev); 5050 5051 if (ret) 5052 goto link_reset_failed; 5053 5054 ret = amdgpu_psp_wait_for_bootloader(adev); 5055 if (ret) 5056 goto link_reset_failed; 5057 5058 return 0; 5059 5060 link_reset_failed: 5061 dev_err(adev->dev, "GPU link reset failed\n"); 5062 return ret; 5063 } 5064 5065 int amdgpu_device_pre_asic_reset(struct amdgpu_device *adev, 5066 struct amdgpu_reset_context *reset_context) 5067 { 5068 int i, r = 0; 5069 struct amdgpu_job *job = NULL; 5070 struct amdgpu_device *tmp_adev = reset_context->reset_req_dev; 5071 bool need_full_reset = 5072 test_bit(AMDGPU_NEED_FULL_RESET, &reset_context->flags); 5073 5074 if (reset_context->reset_req_dev == adev) 5075 job = reset_context->job; 5076 5077 if (amdgpu_sriov_vf(adev)) 5078 amdgpu_virt_pre_reset(adev); 5079 5080 amdgpu_fence_driver_isr_toggle(adev, true); 5081 5082 /* block all schedulers and reset given job's ring */ 5083 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 5084 struct amdgpu_ring *ring = adev->rings[i]; 5085 5086 if (!amdgpu_ring_sched_ready(ring)) 5087 continue; 5088 5089 /* after all hw jobs are reset, hw fence is meaningless, so force_completion */ 5090 amdgpu_fence_driver_force_completion(ring); 5091 } 5092 5093 amdgpu_fence_driver_isr_toggle(adev, false); 5094 5095 if (job && job->vm) 5096 drm_sched_increase_karma(&job->base); 5097 5098 r = amdgpu_reset_prepare_hwcontext(adev, reset_context); 5099 /* If reset handler not implemented, continue; otherwise return */ 5100 if (r == -EOPNOTSUPP) 5101 r = 0; 5102 else 5103 return r; 5104 5105 /* Don't suspend on bare metal if we are not going to HW reset the ASIC */ 5106 if (!amdgpu_sriov_vf(adev)) { 5107 5108 if (!need_full_reset) 5109 need_full_reset = amdgpu_device_ip_need_full_reset(adev); 5110 5111 if (!need_full_reset && amdgpu_gpu_recovery && 5112 amdgpu_device_ip_check_soft_reset(adev)) { 5113 amdgpu_device_ip_pre_soft_reset(adev); 5114 r = amdgpu_device_ip_soft_reset(adev); 5115 amdgpu_device_ip_post_soft_reset(adev); 5116 if (r || amdgpu_device_ip_check_soft_reset(adev)) { 5117 dev_info(adev->dev, "soft reset failed, will fallback to full reset!\n"); 5118 need_full_reset = true; 5119 } 5120 } 5121 5122 if (!test_bit(AMDGPU_SKIP_COREDUMP, &reset_context->flags)) { 5123 dev_info(tmp_adev->dev, "Dumping IP State\n"); 5124 /* Trigger ip dump before we reset the asic */ 5125 for (i = 0; i < tmp_adev->num_ip_blocks; i++) 5126 if (tmp_adev->ip_blocks[i].version->funcs->dump_ip_state) 5127 tmp_adev->ip_blocks[i].version->funcs 5128 ->dump_ip_state((void *)&tmp_adev->ip_blocks[i]); 5129 dev_info(tmp_adev->dev, "Dumping IP State Completed\n"); 5130 } 5131 5132 if (need_full_reset) 5133 r = amdgpu_device_ip_suspend(adev); 5134 if (need_full_reset) 5135 set_bit(AMDGPU_NEED_FULL_RESET, &reset_context->flags); 5136 else 5137 clear_bit(AMDGPU_NEED_FULL_RESET, 5138 &reset_context->flags); 5139 } 5140 5141 return r; 5142 } 5143 5144 int amdgpu_device_reinit_after_reset(struct amdgpu_reset_context *reset_context) 5145 { 5146 struct list_head *device_list_handle; 5147 bool full_reset, vram_lost = false; 5148 struct amdgpu_device *tmp_adev; 5149 int r, init_level; 5150 5151 device_list_handle = reset_context->reset_device_list; 5152 5153 if (!device_list_handle) 5154 return -EINVAL; 5155 5156 full_reset = test_bit(AMDGPU_NEED_FULL_RESET, &reset_context->flags); 5157 5158 /** 5159 * If it's reset on init, it's default init level, otherwise keep level 5160 * as recovery level. 5161 */ 5162 if (reset_context->method == AMD_RESET_METHOD_ON_INIT) 5163 init_level = AMDGPU_INIT_LEVEL_DEFAULT; 5164 else 5165 init_level = AMDGPU_INIT_LEVEL_RESET_RECOVERY; 5166 5167 r = 0; 5168 list_for_each_entry(tmp_adev, device_list_handle, reset_list) { 5169 amdgpu_set_init_level(tmp_adev, init_level); 5170 if (full_reset) { 5171 /* post card */ 5172 amdgpu_reset_set_dpc_status(tmp_adev, false); 5173 amdgpu_ras_clear_err_state(tmp_adev); 5174 r = amdgpu_device_asic_init(tmp_adev); 5175 if (r) { 5176 dev_warn(tmp_adev->dev, "asic atom init failed!"); 5177 } else { 5178 dev_info(tmp_adev->dev, "GPU reset succeeded, trying to resume\n"); 5179 5180 r = amdgpu_device_ip_resume_phase1(tmp_adev); 5181 if (r) 5182 goto out; 5183 5184 vram_lost = amdgpu_device_check_vram_lost(tmp_adev); 5185 5186 if (!test_bit(AMDGPU_SKIP_COREDUMP, &reset_context->flags)) 5187 amdgpu_coredump(tmp_adev, false, vram_lost, reset_context->job); 5188 5189 if (vram_lost) { 5190 dev_info( 5191 tmp_adev->dev, 5192 "VRAM is lost due to GPU reset!\n"); 5193 amdgpu_inc_vram_lost(tmp_adev); 5194 } 5195 5196 r = amdgpu_device_fw_loading(tmp_adev); 5197 if (r) 5198 return r; 5199 5200 r = amdgpu_xcp_restore_partition_mode( 5201 tmp_adev->xcp_mgr); 5202 if (r) 5203 goto out; 5204 5205 r = amdgpu_device_ip_resume_phase2(tmp_adev); 5206 if (r) 5207 goto out; 5208 5209 amdgpu_ttm_set_buffer_funcs_status(tmp_adev, true); 5210 5211 r = amdgpu_device_ip_resume_phase3(tmp_adev); 5212 if (r) 5213 goto out; 5214 5215 if (vram_lost) 5216 amdgpu_device_fill_reset_magic(tmp_adev); 5217 5218 /* 5219 * Add this ASIC as tracked as reset was already 5220 * complete successfully. 5221 */ 5222 amdgpu_register_gpu_instance(tmp_adev); 5223 5224 if (!reset_context->hive && 5225 tmp_adev->gmc.xgmi.num_physical_nodes > 1) 5226 amdgpu_xgmi_add_device(tmp_adev); 5227 5228 r = amdgpu_device_ip_late_init(tmp_adev); 5229 if (r) 5230 goto out; 5231 5232 r = amdgpu_userq_post_reset(tmp_adev, vram_lost); 5233 if (r) 5234 goto out; 5235 5236 drm_client_dev_resume(adev_to_drm(tmp_adev)); 5237 5238 /* 5239 * The GPU enters bad state once faulty pages 5240 * by ECC has reached the threshold, and ras 5241 * recovery is scheduled next. So add one check 5242 * here to break recovery if it indeed exceeds 5243 * bad page threshold, and remind user to 5244 * retire this GPU or setting one bigger 5245 * bad_page_threshold value to fix this once 5246 * probing driver again. 5247 */ 5248 if (!amdgpu_ras_is_rma(tmp_adev)) { 5249 /* must succeed. */ 5250 amdgpu_ras_resume(tmp_adev); 5251 } else { 5252 r = -EINVAL; 5253 goto out; 5254 } 5255 5256 /* Update PSP FW topology after reset */ 5257 if (reset_context->hive && 5258 tmp_adev->gmc.xgmi.num_physical_nodes > 1) 5259 r = amdgpu_xgmi_update_topology( 5260 reset_context->hive, tmp_adev); 5261 } 5262 } 5263 5264 out: 5265 if (!r) { 5266 /* IP init is complete now, set level as default */ 5267 amdgpu_set_init_level(tmp_adev, 5268 AMDGPU_INIT_LEVEL_DEFAULT); 5269 amdgpu_irq_gpu_reset_resume_helper(tmp_adev); 5270 r = amdgpu_ib_ring_tests(tmp_adev); 5271 if (r) { 5272 dev_err(tmp_adev->dev, "ib ring test failed (%d).\n", r); 5273 r = -EAGAIN; 5274 goto end; 5275 } 5276 } 5277 5278 if (r) 5279 tmp_adev->asic_reset_res = r; 5280 } 5281 5282 end: 5283 return r; 5284 } 5285 5286 int amdgpu_do_asic_reset(struct list_head *device_list_handle, 5287 struct amdgpu_reset_context *reset_context) 5288 { 5289 struct amdgpu_device *tmp_adev = NULL; 5290 bool need_full_reset, skip_hw_reset; 5291 int r = 0; 5292 5293 /* Try reset handler method first */ 5294 tmp_adev = list_first_entry(device_list_handle, struct amdgpu_device, 5295 reset_list); 5296 5297 reset_context->reset_device_list = device_list_handle; 5298 r = amdgpu_reset_perform_reset(tmp_adev, reset_context); 5299 /* If reset handler not implemented, continue; otherwise return */ 5300 if (r == -EOPNOTSUPP) 5301 r = 0; 5302 else 5303 return r; 5304 5305 /* Reset handler not implemented, use the default method */ 5306 need_full_reset = 5307 test_bit(AMDGPU_NEED_FULL_RESET, &reset_context->flags); 5308 skip_hw_reset = test_bit(AMDGPU_SKIP_HW_RESET, &reset_context->flags); 5309 5310 /* 5311 * ASIC reset has to be done on all XGMI hive nodes ASAP 5312 * to allow proper links negotiation in FW (within 1 sec) 5313 */ 5314 if (!skip_hw_reset && need_full_reset) { 5315 list_for_each_entry(tmp_adev, device_list_handle, reset_list) { 5316 /* For XGMI run all resets in parallel to speed up the process */ 5317 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { 5318 if (!queue_work(system_unbound_wq, 5319 &tmp_adev->xgmi_reset_work)) 5320 r = -EALREADY; 5321 } else 5322 r = amdgpu_asic_reset(tmp_adev); 5323 5324 if (r) { 5325 dev_err(tmp_adev->dev, 5326 "ASIC reset failed with error, %d for drm dev, %s", 5327 r, adev_to_drm(tmp_adev)->unique); 5328 goto out; 5329 } 5330 } 5331 5332 /* For XGMI wait for all resets to complete before proceed */ 5333 if (!r) { 5334 list_for_each_entry(tmp_adev, device_list_handle, 5335 reset_list) { 5336 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { 5337 flush_work(&tmp_adev->xgmi_reset_work); 5338 r = tmp_adev->asic_reset_res; 5339 if (r) 5340 break; 5341 } 5342 } 5343 } 5344 } 5345 5346 if (!r && amdgpu_ras_intr_triggered()) { 5347 list_for_each_entry(tmp_adev, device_list_handle, reset_list) { 5348 amdgpu_ras_reset_error_count(tmp_adev, 5349 AMDGPU_RAS_BLOCK__MMHUB); 5350 } 5351 5352 amdgpu_ras_intr_cleared(); 5353 } 5354 5355 r = amdgpu_device_reinit_after_reset(reset_context); 5356 if (r == -EAGAIN) 5357 set_bit(AMDGPU_NEED_FULL_RESET, &reset_context->flags); 5358 else 5359 clear_bit(AMDGPU_NEED_FULL_RESET, &reset_context->flags); 5360 5361 out: 5362 return r; 5363 } 5364 5365 static void amdgpu_device_set_mp1_state(struct amdgpu_device *adev) 5366 { 5367 5368 switch (amdgpu_asic_reset_method(adev)) { 5369 case AMD_RESET_METHOD_MODE1: 5370 case AMD_RESET_METHOD_LINK: 5371 adev->mp1_state = PP_MP1_STATE_SHUTDOWN; 5372 break; 5373 case AMD_RESET_METHOD_MODE2: 5374 adev->mp1_state = PP_MP1_STATE_RESET; 5375 break; 5376 default: 5377 adev->mp1_state = PP_MP1_STATE_NONE; 5378 break; 5379 } 5380 } 5381 5382 static void amdgpu_device_unset_mp1_state(struct amdgpu_device *adev) 5383 { 5384 amdgpu_vf_error_trans_all(adev); 5385 adev->mp1_state = PP_MP1_STATE_NONE; 5386 } 5387 5388 static void amdgpu_device_resume_display_audio(struct amdgpu_device *adev) 5389 { 5390 struct pci_dev *p = NULL; 5391 5392 p = pci_get_domain_bus_and_slot(pci_domain_nr(adev->pdev->bus), 5393 adev->pdev->bus->number, 1); 5394 if (p) { 5395 pm_runtime_enable(&(p->dev)); 5396 pm_runtime_resume(&(p->dev)); 5397 } 5398 5399 pci_dev_put(p); 5400 } 5401 5402 static int amdgpu_device_suspend_display_audio(struct amdgpu_device *adev) 5403 { 5404 enum amd_reset_method reset_method; 5405 struct pci_dev *p = NULL; 5406 u64 expires; 5407 5408 /* 5409 * For now, only BACO and mode1 reset are confirmed 5410 * to suffer the audio issue without proper suspended. 5411 */ 5412 reset_method = amdgpu_asic_reset_method(adev); 5413 if ((reset_method != AMD_RESET_METHOD_BACO) && 5414 (reset_method != AMD_RESET_METHOD_MODE1)) 5415 return -EINVAL; 5416 5417 p = pci_get_domain_bus_and_slot(pci_domain_nr(adev->pdev->bus), 5418 adev->pdev->bus->number, 1); 5419 if (!p) 5420 return -ENODEV; 5421 5422 expires = pm_runtime_autosuspend_expiration(&(p->dev)); 5423 if (!expires) 5424 /* 5425 * If we cannot get the audio device autosuspend delay, 5426 * a fixed 4S interval will be used. Considering 3S is 5427 * the audio controller default autosuspend delay setting. 5428 * 4S used here is guaranteed to cover that. 5429 */ 5430 expires = ktime_get_mono_fast_ns() + NSEC_PER_SEC * 4ULL; 5431 5432 while (!pm_runtime_status_suspended(&(p->dev))) { 5433 if (!pm_runtime_suspend(&(p->dev))) 5434 break; 5435 5436 if (expires < ktime_get_mono_fast_ns()) { 5437 dev_warn(adev->dev, "failed to suspend display audio\n"); 5438 pci_dev_put(p); 5439 /* TODO: abort the succeeding gpu reset? */ 5440 return -ETIMEDOUT; 5441 } 5442 } 5443 5444 pm_runtime_disable(&(p->dev)); 5445 5446 pci_dev_put(p); 5447 return 0; 5448 } 5449 5450 static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) 5451 { 5452 struct amdgpu_ras *con = amdgpu_ras_get_context(adev); 5453 5454 #if defined(CONFIG_DEBUG_FS) 5455 if (!amdgpu_sriov_vf(adev)) 5456 cancel_work(&adev->reset_work); 5457 #endif 5458 cancel_work(&adev->userq_reset_work); 5459 5460 if (adev->kfd.dev) 5461 cancel_work(&adev->kfd.reset_work); 5462 5463 if (amdgpu_sriov_vf(adev)) 5464 cancel_work(&adev->virt.flr_work); 5465 5466 if (con && adev->ras_enabled) 5467 cancel_work(&con->recovery_work); 5468 5469 } 5470 5471 static int amdgpu_device_health_check(struct list_head *device_list_handle) 5472 { 5473 struct amdgpu_device *tmp_adev; 5474 int ret = 0; 5475 5476 list_for_each_entry(tmp_adev, device_list_handle, reset_list) { 5477 ret |= amdgpu_device_bus_status_check(tmp_adev); 5478 } 5479 5480 return ret; 5481 } 5482 5483 static void amdgpu_device_recovery_prepare(struct amdgpu_device *adev, 5484 struct list_head *device_list, 5485 struct amdgpu_hive_info *hive) 5486 { 5487 struct amdgpu_device *tmp_adev = NULL; 5488 5489 /* 5490 * Build list of devices to reset. 5491 * In case we are in XGMI hive mode, resort the device list 5492 * to put adev in the 1st position. 5493 */ 5494 if (!amdgpu_sriov_vf(adev) && (adev->gmc.xgmi.num_physical_nodes > 1) && hive) { 5495 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { 5496 list_add_tail(&tmp_adev->reset_list, device_list); 5497 if (adev->shutdown) 5498 tmp_adev->shutdown = true; 5499 if (amdgpu_reset_in_dpc(adev)) 5500 tmp_adev->pcie_reset_ctx.in_link_reset = true; 5501 } 5502 if (!list_is_first(&adev->reset_list, device_list)) 5503 list_rotate_to_front(&adev->reset_list, device_list); 5504 } else { 5505 list_add_tail(&adev->reset_list, device_list); 5506 } 5507 } 5508 5509 static void amdgpu_device_recovery_get_reset_lock(struct amdgpu_device *adev, 5510 struct list_head *device_list) 5511 { 5512 struct amdgpu_device *tmp_adev = NULL; 5513 5514 if (list_empty(device_list)) 5515 return; 5516 tmp_adev = 5517 list_first_entry(device_list, struct amdgpu_device, reset_list); 5518 amdgpu_device_lock_reset_domain(tmp_adev->reset_domain); 5519 } 5520 5521 static void amdgpu_device_recovery_put_reset_lock(struct amdgpu_device *adev, 5522 struct list_head *device_list) 5523 { 5524 struct amdgpu_device *tmp_adev = NULL; 5525 5526 if (list_empty(device_list)) 5527 return; 5528 tmp_adev = 5529 list_first_entry(device_list, struct amdgpu_device, reset_list); 5530 amdgpu_device_unlock_reset_domain(tmp_adev->reset_domain); 5531 } 5532 5533 static void amdgpu_device_halt_activities(struct amdgpu_device *adev, 5534 struct amdgpu_job *job, 5535 struct amdgpu_reset_context *reset_context, 5536 struct list_head *device_list, 5537 struct amdgpu_hive_info *hive, 5538 bool need_emergency_restart) 5539 { 5540 struct amdgpu_device *tmp_adev = NULL; 5541 int i; 5542 5543 /* block all schedulers and reset given job's ring */ 5544 list_for_each_entry(tmp_adev, device_list, reset_list) { 5545 amdgpu_device_set_mp1_state(tmp_adev); 5546 5547 /* 5548 * Try to put the audio codec into suspend state 5549 * before gpu reset started. 5550 * 5551 * Due to the power domain of the graphics device 5552 * is shared with AZ power domain. Without this, 5553 * we may change the audio hardware from behind 5554 * the audio driver's back. That will trigger 5555 * some audio codec errors. 5556 */ 5557 if (!amdgpu_device_suspend_display_audio(tmp_adev)) 5558 tmp_adev->pcie_reset_ctx.audio_suspended = true; 5559 5560 amdgpu_ras_set_error_query_ready(tmp_adev, false); 5561 5562 cancel_delayed_work_sync(&tmp_adev->delayed_init_work); 5563 5564 amdgpu_amdkfd_pre_reset(tmp_adev, reset_context); 5565 5566 /* 5567 * Mark these ASICs to be reset as untracked first 5568 * And add them back after reset completed 5569 */ 5570 amdgpu_unregister_gpu_instance(tmp_adev); 5571 5572 drm_client_dev_suspend(adev_to_drm(tmp_adev)); 5573 5574 /* disable ras on ALL IPs */ 5575 if (!need_emergency_restart && !amdgpu_reset_in_dpc(adev) && 5576 amdgpu_device_ip_need_full_reset(tmp_adev)) 5577 amdgpu_ras_suspend(tmp_adev); 5578 5579 amdgpu_userq_pre_reset(tmp_adev); 5580 5581 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 5582 struct amdgpu_ring *ring = tmp_adev->rings[i]; 5583 5584 if (!amdgpu_ring_sched_ready(ring)) 5585 continue; 5586 5587 drm_sched_wqueue_stop(&ring->sched); 5588 5589 if (need_emergency_restart) 5590 amdgpu_job_stop_all_jobs_on_sched(&ring->sched); 5591 } 5592 atomic_inc(&tmp_adev->gpu_reset_counter); 5593 } 5594 } 5595 5596 static int amdgpu_device_asic_reset(struct amdgpu_device *adev, 5597 struct list_head *device_list, 5598 struct amdgpu_reset_context *reset_context) 5599 { 5600 struct amdgpu_device *tmp_adev = NULL; 5601 int retry_limit = AMDGPU_MAX_RETRY_LIMIT; 5602 int r = 0; 5603 5604 retry: /* Rest of adevs pre asic reset from XGMI hive. */ 5605 list_for_each_entry(tmp_adev, device_list, reset_list) { 5606 r = amdgpu_device_pre_asic_reset(tmp_adev, reset_context); 5607 /*TODO Should we stop ?*/ 5608 if (r) { 5609 dev_err(tmp_adev->dev, "GPU pre asic reset failed with err, %d for drm dev, %s ", 5610 r, adev_to_drm(tmp_adev)->unique); 5611 tmp_adev->asic_reset_res = r; 5612 } 5613 } 5614 5615 /* Actual ASIC resets if needed.*/ 5616 /* Host driver will handle XGMI hive reset for SRIOV */ 5617 if (amdgpu_sriov_vf(adev)) { 5618 5619 /* Bail out of reset early */ 5620 if (amdgpu_ras_is_rma(adev)) 5621 return -ENODEV; 5622 5623 if (amdgpu_ras_get_fed_status(adev) || amdgpu_virt_rcvd_ras_interrupt(adev)) { 5624 dev_dbg(adev->dev, "Detected RAS error, wait for FLR completion\n"); 5625 amdgpu_ras_set_fed(adev, true); 5626 set_bit(AMDGPU_HOST_FLR, &reset_context->flags); 5627 } 5628 5629 r = amdgpu_device_reset_sriov(adev, reset_context); 5630 if (AMDGPU_RETRY_SRIOV_RESET(r) && (retry_limit--) > 0) { 5631 amdgpu_virt_release_full_gpu(adev, true); 5632 goto retry; 5633 } 5634 if (r) 5635 adev->asic_reset_res = r; 5636 } else { 5637 r = amdgpu_do_asic_reset(device_list, reset_context); 5638 if (r && r == -EAGAIN) 5639 goto retry; 5640 } 5641 5642 list_for_each_entry(tmp_adev, device_list, reset_list) { 5643 /* 5644 * Drop any pending non scheduler resets queued before reset is done. 5645 * Any reset scheduled after this point would be valid. Scheduler resets 5646 * were already dropped during drm_sched_stop and no new ones can come 5647 * in before drm_sched_start. 5648 */ 5649 amdgpu_device_stop_pending_resets(tmp_adev); 5650 } 5651 5652 return r; 5653 } 5654 5655 static int amdgpu_device_sched_resume(struct list_head *device_list, 5656 struct amdgpu_reset_context *reset_context, 5657 bool job_signaled) 5658 { 5659 struct amdgpu_device *tmp_adev = NULL; 5660 int i, r = 0; 5661 5662 /* Post ASIC reset for all devs .*/ 5663 list_for_each_entry(tmp_adev, device_list, reset_list) { 5664 5665 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 5666 struct amdgpu_ring *ring = tmp_adev->rings[i]; 5667 5668 if (!amdgpu_ring_sched_ready(ring)) 5669 continue; 5670 5671 drm_sched_wqueue_start(&ring->sched); 5672 } 5673 5674 if (!drm_drv_uses_atomic_modeset(adev_to_drm(tmp_adev)) && !job_signaled) 5675 drm_helper_resume_force_mode(adev_to_drm(tmp_adev)); 5676 5677 if (tmp_adev->asic_reset_res) { 5678 /* bad news, how to tell it to userspace ? 5679 * for ras error, we should report GPU bad status instead of 5680 * reset failure 5681 */ 5682 if (reset_context->src != AMDGPU_RESET_SRC_RAS || 5683 !amdgpu_ras_eeprom_check_err_threshold(tmp_adev)) 5684 dev_info( 5685 tmp_adev->dev, 5686 "GPU reset(%d) failed with error %d\n", 5687 atomic_read( 5688 &tmp_adev->gpu_reset_counter), 5689 tmp_adev->asic_reset_res); 5690 amdgpu_vf_error_put(tmp_adev, 5691 AMDGIM_ERROR_VF_GPU_RESET_FAIL, 0, 5692 tmp_adev->asic_reset_res); 5693 if (!r) 5694 r = tmp_adev->asic_reset_res; 5695 tmp_adev->asic_reset_res = 0; 5696 } else { 5697 dev_info(tmp_adev->dev, "GPU reset(%d) succeeded!\n", 5698 atomic_read(&tmp_adev->gpu_reset_counter)); 5699 if (amdgpu_acpi_smart_shift_update(tmp_adev, 5700 AMDGPU_SS_DEV_D0)) 5701 dev_warn(tmp_adev->dev, 5702 "smart shift update failed\n"); 5703 } 5704 } 5705 5706 return r; 5707 } 5708 5709 static void amdgpu_device_gpu_resume(struct amdgpu_device *adev, 5710 struct list_head *device_list, 5711 bool need_emergency_restart) 5712 { 5713 struct amdgpu_device *tmp_adev = NULL; 5714 5715 list_for_each_entry(tmp_adev, device_list, reset_list) { 5716 /* unlock kfd: SRIOV would do it separately */ 5717 if (!need_emergency_restart && !amdgpu_sriov_vf(tmp_adev)) 5718 amdgpu_amdkfd_post_reset(tmp_adev); 5719 5720 /* kfd_post_reset will do nothing if kfd device is not initialized, 5721 * need to bring up kfd here if it's not be initialized before 5722 */ 5723 if (!adev->kfd.init_complete) 5724 amdgpu_amdkfd_device_init(adev); 5725 5726 if (tmp_adev->pcie_reset_ctx.audio_suspended) 5727 amdgpu_device_resume_display_audio(tmp_adev); 5728 5729 amdgpu_device_unset_mp1_state(tmp_adev); 5730 5731 amdgpu_ras_set_error_query_ready(tmp_adev, true); 5732 5733 } 5734 } 5735 5736 5737 /** 5738 * amdgpu_device_gpu_recover - reset the asic and recover scheduler 5739 * 5740 * @adev: amdgpu_device pointer 5741 * @job: which job trigger hang 5742 * @reset_context: amdgpu reset context pointer 5743 * 5744 * Attempt to reset the GPU if it has hung (all asics). 5745 * Attempt to do soft-reset or full-reset and reinitialize Asic 5746 * Returns 0 for success or an error on failure. 5747 */ 5748 5749 int amdgpu_device_gpu_recover(struct amdgpu_device *adev, 5750 struct amdgpu_job *job, 5751 struct amdgpu_reset_context *reset_context) 5752 { 5753 struct list_head device_list; 5754 bool job_signaled = false; 5755 struct amdgpu_hive_info *hive = NULL; 5756 int r = 0; 5757 bool need_emergency_restart = false; 5758 /* save the pasid here as the job may be freed before the end of the reset */ 5759 int pasid = job ? job->pasid : -EINVAL; 5760 5761 /* 5762 * If it reaches here because of hang/timeout and a RAS error is 5763 * detected at the same time, let RAS recovery take care of it. 5764 */ 5765 if (amdgpu_ras_is_err_state(adev, AMDGPU_RAS_BLOCK__ANY) && 5766 !amdgpu_sriov_vf(adev) && 5767 reset_context->src != AMDGPU_RESET_SRC_RAS) { 5768 dev_dbg(adev->dev, 5769 "Gpu recovery from source: %d yielding to RAS error recovery handling", 5770 reset_context->src); 5771 return 0; 5772 } 5773 5774 /* 5775 * Special case: RAS triggered and full reset isn't supported 5776 */ 5777 need_emergency_restart = amdgpu_ras_need_emergency_restart(adev); 5778 5779 /* 5780 * Flush RAM to disk so that after reboot 5781 * the user can read log and see why the system rebooted. 5782 */ 5783 if (need_emergency_restart && amdgpu_ras_get_context(adev) && 5784 amdgpu_ras_get_context(adev)->reboot) { 5785 dev_warn(adev->dev, "Emergency reboot."); 5786 5787 ksys_sync_helper(); 5788 emergency_restart(); 5789 } 5790 5791 dev_info(adev->dev, "GPU %s begin!. Source: %d\n", 5792 need_emergency_restart ? "jobs stop" : "reset", 5793 reset_context->src); 5794 5795 if (!amdgpu_sriov_vf(adev)) 5796 hive = amdgpu_get_xgmi_hive(adev); 5797 if (hive) 5798 mutex_lock(&hive->hive_lock); 5799 5800 reset_context->job = job; 5801 reset_context->hive = hive; 5802 INIT_LIST_HEAD(&device_list); 5803 5804 amdgpu_device_recovery_prepare(adev, &device_list, hive); 5805 5806 if (!amdgpu_sriov_vf(adev)) { 5807 r = amdgpu_device_health_check(&device_list); 5808 if (r) 5809 goto end_reset; 5810 } 5811 5812 /* Cannot be called after locking reset domain */ 5813 amdgpu_ras_pre_reset(adev, &device_list); 5814 5815 /* We need to lock reset domain only once both for XGMI and single device */ 5816 amdgpu_device_recovery_get_reset_lock(adev, &device_list); 5817 5818 amdgpu_device_halt_activities(adev, job, reset_context, &device_list, 5819 hive, need_emergency_restart); 5820 if (need_emergency_restart) 5821 goto skip_sched_resume; 5822 /* 5823 * Must check guilty signal here since after this point all old 5824 * HW fences are force signaled. 5825 * 5826 * job->base holds a reference to parent fence 5827 */ 5828 if (job && (dma_fence_get_status(&job->hw_fence->base) > 0)) { 5829 job_signaled = true; 5830 dev_info(adev->dev, "Guilty job already signaled, skipping HW reset"); 5831 goto skip_hw_reset; 5832 } 5833 5834 r = amdgpu_device_asic_reset(adev, &device_list, reset_context); 5835 if (r) 5836 goto reset_unlock; 5837 skip_hw_reset: 5838 r = amdgpu_device_sched_resume(&device_list, reset_context, job_signaled); 5839 if (r) 5840 goto reset_unlock; 5841 skip_sched_resume: 5842 amdgpu_device_gpu_resume(adev, &device_list, need_emergency_restart); 5843 reset_unlock: 5844 amdgpu_device_recovery_put_reset_lock(adev, &device_list); 5845 amdgpu_ras_post_reset(adev, &device_list); 5846 end_reset: 5847 if (hive) { 5848 mutex_unlock(&hive->hive_lock); 5849 amdgpu_put_xgmi_hive(hive); 5850 } 5851 5852 if (r) 5853 dev_info(adev->dev, "GPU reset end with ret = %d\n", r); 5854 5855 atomic_set(&adev->reset_domain->reset_res, r); 5856 5857 if (!r) { 5858 struct amdgpu_task_info *ti = NULL; 5859 5860 /* 5861 * The job may already be freed at this point via the sched tdr workqueue so 5862 * use the cached pasid. 5863 */ 5864 if (pasid >= 0) 5865 ti = amdgpu_vm_get_task_info_pasid(adev, pasid); 5866 5867 drm_dev_wedged_event(adev_to_drm(adev), DRM_WEDGE_RECOVERY_NONE, 5868 ti ? &ti->task : NULL); 5869 5870 amdgpu_vm_put_task_info(ti); 5871 } 5872 5873 return r; 5874 } 5875 5876 /** 5877 * amdgpu_device_partner_bandwidth - find the bandwidth of appropriate partner 5878 * 5879 * @adev: amdgpu_device pointer 5880 * @speed: pointer to the speed of the link 5881 * @width: pointer to the width of the link 5882 * 5883 * Evaluate the hierarchy to find the speed and bandwidth capabilities of the 5884 * first physical partner to an AMD dGPU. 5885 * This will exclude any virtual switches and links. 5886 */ 5887 static void amdgpu_device_partner_bandwidth(struct amdgpu_device *adev, 5888 enum pci_bus_speed *speed, 5889 enum pcie_link_width *width) 5890 { 5891 struct pci_dev *parent = adev->pdev; 5892 5893 if (!speed || !width) 5894 return; 5895 5896 *speed = PCI_SPEED_UNKNOWN; 5897 *width = PCIE_LNK_WIDTH_UNKNOWN; 5898 5899 if (amdgpu_device_pcie_dynamic_switching_supported(adev)) { 5900 while ((parent = pci_upstream_bridge(parent))) { 5901 /* skip upstream/downstream switches internal to dGPU*/ 5902 if (parent->vendor == PCI_VENDOR_ID_ATI) 5903 continue; 5904 *speed = pcie_get_speed_cap(parent); 5905 *width = pcie_get_width_cap(parent); 5906 break; 5907 } 5908 } else { 5909 /* use the current speeds rather than max if switching is not supported */ 5910 pcie_bandwidth_available(adev->pdev, NULL, speed, width); 5911 } 5912 } 5913 5914 /** 5915 * amdgpu_device_gpu_bandwidth - find the bandwidth of the GPU 5916 * 5917 * @adev: amdgpu_device pointer 5918 * @speed: pointer to the speed of the link 5919 * @width: pointer to the width of the link 5920 * 5921 * Evaluate the hierarchy to find the speed and bandwidth capabilities of the 5922 * AMD dGPU which may be a virtual upstream bridge. 5923 */ 5924 static void amdgpu_device_gpu_bandwidth(struct amdgpu_device *adev, 5925 enum pci_bus_speed *speed, 5926 enum pcie_link_width *width) 5927 { 5928 struct pci_dev *parent = adev->pdev; 5929 5930 if (!speed || !width) 5931 return; 5932 5933 parent = pci_upstream_bridge(parent); 5934 if (parent && parent->vendor == PCI_VENDOR_ID_ATI) { 5935 /* use the upstream/downstream switches internal to dGPU */ 5936 *speed = pcie_get_speed_cap(parent); 5937 *width = pcie_get_width_cap(parent); 5938 while ((parent = pci_upstream_bridge(parent))) { 5939 if (parent->vendor == PCI_VENDOR_ID_ATI) { 5940 /* use the upstream/downstream switches internal to dGPU */ 5941 *speed = pcie_get_speed_cap(parent); 5942 *width = pcie_get_width_cap(parent); 5943 } 5944 } 5945 } else { 5946 /* use the device itself */ 5947 *speed = pcie_get_speed_cap(adev->pdev); 5948 *width = pcie_get_width_cap(adev->pdev); 5949 } 5950 } 5951 5952 /** 5953 * amdgpu_device_get_pcie_info - fence pcie info about the PCIE slot 5954 * 5955 * @adev: amdgpu_device pointer 5956 * 5957 * Fetches and stores in the driver the PCIE capabilities (gen speed 5958 * and lanes) of the slot the device is in. Handles APUs and 5959 * virtualized environments where PCIE config space may not be available. 5960 */ 5961 static void amdgpu_device_get_pcie_info(struct amdgpu_device *adev) 5962 { 5963 enum pci_bus_speed speed_cap, platform_speed_cap; 5964 enum pcie_link_width platform_link_width, link_width; 5965 5966 if (amdgpu_pcie_gen_cap) 5967 adev->pm.pcie_gen_mask = amdgpu_pcie_gen_cap; 5968 5969 if (amdgpu_pcie_lane_cap) 5970 adev->pm.pcie_mlw_mask = amdgpu_pcie_lane_cap; 5971 5972 /* covers APUs as well */ 5973 if (pci_is_root_bus(adev->pdev->bus) && !amdgpu_passthrough(adev)) { 5974 if (adev->pm.pcie_gen_mask == 0) 5975 adev->pm.pcie_gen_mask = AMDGPU_DEFAULT_PCIE_GEN_MASK; 5976 if (adev->pm.pcie_mlw_mask == 0) 5977 adev->pm.pcie_mlw_mask = AMDGPU_DEFAULT_PCIE_MLW_MASK; 5978 return; 5979 } 5980 5981 if (adev->pm.pcie_gen_mask && adev->pm.pcie_mlw_mask) 5982 return; 5983 5984 amdgpu_device_partner_bandwidth(adev, &platform_speed_cap, 5985 &platform_link_width); 5986 amdgpu_device_gpu_bandwidth(adev, &speed_cap, &link_width); 5987 5988 if (adev->pm.pcie_gen_mask == 0) { 5989 /* asic caps */ 5990 if (speed_cap == PCI_SPEED_UNKNOWN) { 5991 adev->pm.pcie_gen_mask |= (CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN1 | 5992 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN2 | 5993 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN3); 5994 } else { 5995 if (speed_cap == PCIE_SPEED_32_0GT) 5996 adev->pm.pcie_gen_mask |= (CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN1 | 5997 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN2 | 5998 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN3 | 5999 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN4 | 6000 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN5); 6001 else if (speed_cap == PCIE_SPEED_16_0GT) 6002 adev->pm.pcie_gen_mask |= (CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6003 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN2 | 6004 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN3 | 6005 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN4); 6006 else if (speed_cap == PCIE_SPEED_8_0GT) 6007 adev->pm.pcie_gen_mask |= (CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6008 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN2 | 6009 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN3); 6010 else if (speed_cap == PCIE_SPEED_5_0GT) 6011 adev->pm.pcie_gen_mask |= (CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6012 CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN2); 6013 else 6014 adev->pm.pcie_gen_mask |= CAIL_ASIC_PCIE_LINK_SPEED_SUPPORT_GEN1; 6015 } 6016 /* platform caps */ 6017 if (platform_speed_cap == PCI_SPEED_UNKNOWN) { 6018 adev->pm.pcie_gen_mask |= (CAIL_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6019 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2); 6020 } else { 6021 if (platform_speed_cap == PCIE_SPEED_32_0GT) 6022 adev->pm.pcie_gen_mask |= (CAIL_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6023 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2 | 6024 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN3 | 6025 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN4 | 6026 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN5); 6027 else if (platform_speed_cap == PCIE_SPEED_16_0GT) 6028 adev->pm.pcie_gen_mask |= (CAIL_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6029 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2 | 6030 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN3 | 6031 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN4); 6032 else if (platform_speed_cap == PCIE_SPEED_8_0GT) 6033 adev->pm.pcie_gen_mask |= (CAIL_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6034 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2 | 6035 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN3); 6036 else if (platform_speed_cap == PCIE_SPEED_5_0GT) 6037 adev->pm.pcie_gen_mask |= (CAIL_PCIE_LINK_SPEED_SUPPORT_GEN1 | 6038 CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2); 6039 else 6040 adev->pm.pcie_gen_mask |= CAIL_PCIE_LINK_SPEED_SUPPORT_GEN1; 6041 6042 } 6043 } 6044 if (adev->pm.pcie_mlw_mask == 0) { 6045 /* asic caps */ 6046 if (link_width == PCIE_LNK_WIDTH_UNKNOWN) { 6047 adev->pm.pcie_mlw_mask |= AMDGPU_DEFAULT_ASIC_PCIE_MLW_MASK; 6048 } else { 6049 switch (link_width) { 6050 case PCIE_LNK_X32: 6051 adev->pm.pcie_mlw_mask |= (CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X32 | 6052 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X16 | 6053 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X12 | 6054 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X8 | 6055 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X4 | 6056 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X2 | 6057 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1); 6058 break; 6059 case PCIE_LNK_X16: 6060 adev->pm.pcie_mlw_mask |= (CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X16 | 6061 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X12 | 6062 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X8 | 6063 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X4 | 6064 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X2 | 6065 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1); 6066 break; 6067 case PCIE_LNK_X12: 6068 adev->pm.pcie_mlw_mask |= (CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X12 | 6069 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X8 | 6070 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X4 | 6071 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X2 | 6072 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1); 6073 break; 6074 case PCIE_LNK_X8: 6075 adev->pm.pcie_mlw_mask |= (CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X8 | 6076 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X4 | 6077 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X2 | 6078 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1); 6079 break; 6080 case PCIE_LNK_X4: 6081 adev->pm.pcie_mlw_mask |= (CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X4 | 6082 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X2 | 6083 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1); 6084 break; 6085 case PCIE_LNK_X2: 6086 adev->pm.pcie_mlw_mask |= (CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X2 | 6087 CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1); 6088 break; 6089 case PCIE_LNK_X1: 6090 adev->pm.pcie_mlw_mask |= CAIL_ASIC_PCIE_LINK_WIDTH_SUPPORT_X1; 6091 break; 6092 default: 6093 break; 6094 } 6095 } 6096 /* platform caps */ 6097 if (platform_link_width == PCIE_LNK_WIDTH_UNKNOWN) { 6098 adev->pm.pcie_mlw_mask |= AMDGPU_DEFAULT_PCIE_MLW_MASK; 6099 } else { 6100 switch (platform_link_width) { 6101 case PCIE_LNK_X32: 6102 adev->pm.pcie_mlw_mask |= (CAIL_PCIE_LINK_WIDTH_SUPPORT_X32 | 6103 CAIL_PCIE_LINK_WIDTH_SUPPORT_X16 | 6104 CAIL_PCIE_LINK_WIDTH_SUPPORT_X12 | 6105 CAIL_PCIE_LINK_WIDTH_SUPPORT_X8 | 6106 CAIL_PCIE_LINK_WIDTH_SUPPORT_X4 | 6107 CAIL_PCIE_LINK_WIDTH_SUPPORT_X2 | 6108 CAIL_PCIE_LINK_WIDTH_SUPPORT_X1); 6109 break; 6110 case PCIE_LNK_X16: 6111 adev->pm.pcie_mlw_mask |= (CAIL_PCIE_LINK_WIDTH_SUPPORT_X16 | 6112 CAIL_PCIE_LINK_WIDTH_SUPPORT_X12 | 6113 CAIL_PCIE_LINK_WIDTH_SUPPORT_X8 | 6114 CAIL_PCIE_LINK_WIDTH_SUPPORT_X4 | 6115 CAIL_PCIE_LINK_WIDTH_SUPPORT_X2 | 6116 CAIL_PCIE_LINK_WIDTH_SUPPORT_X1); 6117 break; 6118 case PCIE_LNK_X12: 6119 adev->pm.pcie_mlw_mask |= (CAIL_PCIE_LINK_WIDTH_SUPPORT_X12 | 6120 CAIL_PCIE_LINK_WIDTH_SUPPORT_X8 | 6121 CAIL_PCIE_LINK_WIDTH_SUPPORT_X4 | 6122 CAIL_PCIE_LINK_WIDTH_SUPPORT_X2 | 6123 CAIL_PCIE_LINK_WIDTH_SUPPORT_X1); 6124 break; 6125 case PCIE_LNK_X8: 6126 adev->pm.pcie_mlw_mask |= (CAIL_PCIE_LINK_WIDTH_SUPPORT_X8 | 6127 CAIL_PCIE_LINK_WIDTH_SUPPORT_X4 | 6128 CAIL_PCIE_LINK_WIDTH_SUPPORT_X2 | 6129 CAIL_PCIE_LINK_WIDTH_SUPPORT_X1); 6130 break; 6131 case PCIE_LNK_X4: 6132 adev->pm.pcie_mlw_mask |= (CAIL_PCIE_LINK_WIDTH_SUPPORT_X4 | 6133 CAIL_PCIE_LINK_WIDTH_SUPPORT_X2 | 6134 CAIL_PCIE_LINK_WIDTH_SUPPORT_X1); 6135 break; 6136 case PCIE_LNK_X2: 6137 adev->pm.pcie_mlw_mask |= (CAIL_PCIE_LINK_WIDTH_SUPPORT_X2 | 6138 CAIL_PCIE_LINK_WIDTH_SUPPORT_X1); 6139 break; 6140 case PCIE_LNK_X1: 6141 adev->pm.pcie_mlw_mask |= CAIL_PCIE_LINK_WIDTH_SUPPORT_X1; 6142 break; 6143 default: 6144 break; 6145 } 6146 } 6147 } 6148 } 6149 6150 /** 6151 * amdgpu_device_is_peer_accessible - Check peer access through PCIe BAR 6152 * 6153 * @adev: amdgpu_device pointer 6154 * @peer_adev: amdgpu_device pointer for peer device trying to access @adev 6155 * 6156 * Return true if @peer_adev can access (DMA) @adev through the PCIe 6157 * BAR, i.e. @adev is "large BAR" and the BAR matches the DMA mask of 6158 * @peer_adev. 6159 */ 6160 bool amdgpu_device_is_peer_accessible(struct amdgpu_device *adev, 6161 struct amdgpu_device *peer_adev) 6162 { 6163 #ifdef CONFIG_HSA_AMD_P2P 6164 bool p2p_access = 6165 !adev->gmc.xgmi.connected_to_cpu && 6166 !(pci_p2pdma_distance(adev->pdev, peer_adev->dev, false) < 0); 6167 if (!p2p_access) 6168 dev_info(adev->dev, "PCIe P2P access from peer device %s is not supported by the chipset\n", 6169 pci_name(peer_adev->pdev)); 6170 6171 bool is_large_bar = adev->gmc.visible_vram_size && 6172 adev->gmc.real_vram_size == adev->gmc.visible_vram_size; 6173 bool p2p_addressable = amdgpu_device_check_iommu_remap(peer_adev); 6174 6175 if (!p2p_addressable) { 6176 uint64_t address_mask = peer_adev->dev->dma_mask ? 6177 ~*peer_adev->dev->dma_mask : ~((1ULL << 32) - 1); 6178 resource_size_t aper_limit = 6179 adev->gmc.aper_base + adev->gmc.aper_size - 1; 6180 6181 p2p_addressable = !(adev->gmc.aper_base & address_mask || 6182 aper_limit & address_mask); 6183 } 6184 return pcie_p2p && is_large_bar && p2p_access && p2p_addressable; 6185 #else 6186 return false; 6187 #endif 6188 } 6189 6190 int amdgpu_device_baco_enter(struct amdgpu_device *adev) 6191 { 6192 struct amdgpu_ras *ras = amdgpu_ras_get_context(adev); 6193 6194 if (!amdgpu_device_supports_baco(adev)) 6195 return -ENOTSUPP; 6196 6197 if (ras && adev->ras_enabled && 6198 adev->nbio.funcs->enable_doorbell_interrupt) 6199 adev->nbio.funcs->enable_doorbell_interrupt(adev, false); 6200 6201 return amdgpu_dpm_baco_enter(adev); 6202 } 6203 6204 int amdgpu_device_baco_exit(struct amdgpu_device *adev) 6205 { 6206 struct amdgpu_ras *ras = amdgpu_ras_get_context(adev); 6207 int ret = 0; 6208 6209 if (!amdgpu_device_supports_baco(adev)) 6210 return -ENOTSUPP; 6211 6212 ret = amdgpu_dpm_baco_exit(adev); 6213 if (ret) 6214 return ret; 6215 6216 if (ras && adev->ras_enabled && 6217 adev->nbio.funcs->enable_doorbell_interrupt) 6218 adev->nbio.funcs->enable_doorbell_interrupt(adev, true); 6219 6220 if (amdgpu_passthrough(adev) && adev->nbio.funcs && 6221 adev->nbio.funcs->clear_doorbell_interrupt) 6222 adev->nbio.funcs->clear_doorbell_interrupt(adev); 6223 6224 return 0; 6225 } 6226 6227 /** 6228 * amdgpu_pci_error_detected - Called when a PCI error is detected. 6229 * @pdev: PCI device struct 6230 * @state: PCI channel state 6231 * 6232 * Description: Called when a PCI error is detected. 6233 * 6234 * Return: PCI_ERS_RESULT_NEED_RESET or PCI_ERS_RESULT_DISCONNECT. 6235 */ 6236 pci_ers_result_t amdgpu_pci_error_detected(struct pci_dev *pdev, pci_channel_state_t state) 6237 { 6238 struct drm_device *dev = pci_get_drvdata(pdev); 6239 struct amdgpu_device *adev = drm_to_adev(dev); 6240 struct amdgpu_hive_info *hive __free(xgmi_put_hive) = 6241 amdgpu_get_xgmi_hive(adev); 6242 struct amdgpu_reset_context reset_context; 6243 struct list_head device_list; 6244 6245 dev_info(adev->dev, "PCI error: detected callback!!\n"); 6246 6247 adev->pci_channel_state = state; 6248 6249 switch (state) { 6250 case pci_channel_io_normal: 6251 dev_info(adev->dev, "pci_channel_io_normal: state(%d)!!\n", state); 6252 return PCI_ERS_RESULT_CAN_RECOVER; 6253 case pci_channel_io_frozen: 6254 /* Fatal error, prepare for slot reset */ 6255 dev_info(adev->dev, "pci_channel_io_frozen: state(%d)!!\n", state); 6256 if (hive) { 6257 /* Hive devices should be able to support FW based 6258 * link reset on other devices, if not return. 6259 */ 6260 if (!amdgpu_dpm_is_link_reset_supported(adev)) { 6261 dev_warn(adev->dev, 6262 "No support for XGMI hive yet...\n"); 6263 return PCI_ERS_RESULT_DISCONNECT; 6264 } 6265 /* Set dpc status only if device is part of hive 6266 * Non-hive devices should be able to recover after 6267 * link reset. 6268 */ 6269 amdgpu_reset_set_dpc_status(adev, true); 6270 6271 mutex_lock(&hive->hive_lock); 6272 } 6273 memset(&reset_context, 0, sizeof(reset_context)); 6274 INIT_LIST_HEAD(&device_list); 6275 6276 amdgpu_device_recovery_prepare(adev, &device_list, hive); 6277 amdgpu_device_recovery_get_reset_lock(adev, &device_list); 6278 amdgpu_device_halt_activities(adev, NULL, &reset_context, &device_list, 6279 hive, false); 6280 if (hive) 6281 mutex_unlock(&hive->hive_lock); 6282 return PCI_ERS_RESULT_NEED_RESET; 6283 case pci_channel_io_perm_failure: 6284 /* Permanent error, prepare for device removal */ 6285 dev_info(adev->dev, "pci_channel_io_perm_failure: state(%d)!!\n", state); 6286 return PCI_ERS_RESULT_DISCONNECT; 6287 } 6288 6289 return PCI_ERS_RESULT_NEED_RESET; 6290 } 6291 6292 /** 6293 * amdgpu_pci_mmio_enabled - Enable MMIO and dump debug registers 6294 * @pdev: pointer to PCI device 6295 */ 6296 pci_ers_result_t amdgpu_pci_mmio_enabled(struct pci_dev *pdev) 6297 { 6298 struct drm_device *dev = pci_get_drvdata(pdev); 6299 struct amdgpu_device *adev = drm_to_adev(dev); 6300 6301 dev_info(adev->dev, "PCI error: mmio enabled callback!!\n"); 6302 6303 /* TODO - dump whatever for debugging purposes */ 6304 6305 /* This called only if amdgpu_pci_error_detected returns 6306 * PCI_ERS_RESULT_CAN_RECOVER. Read/write to the device still 6307 * works, no need to reset slot. 6308 */ 6309 6310 return PCI_ERS_RESULT_RECOVERED; 6311 } 6312 6313 /** 6314 * amdgpu_pci_slot_reset - Called when PCI slot has been reset. 6315 * @pdev: PCI device struct 6316 * 6317 * Description: This routine is called by the pci error recovery 6318 * code after the PCI slot has been reset, just before we 6319 * should resume normal operations. 6320 */ 6321 pci_ers_result_t amdgpu_pci_slot_reset(struct pci_dev *pdev) 6322 { 6323 struct drm_device *dev = pci_get_drvdata(pdev); 6324 struct amdgpu_device *adev = drm_to_adev(dev); 6325 struct amdgpu_reset_context reset_context; 6326 struct amdgpu_device *tmp_adev; 6327 struct amdgpu_hive_info *hive; 6328 struct list_head device_list; 6329 struct pci_dev *link_dev; 6330 int r = 0, i, timeout; 6331 u32 memsize; 6332 u16 status; 6333 6334 dev_info(adev->dev, "PCI error: slot reset callback!!\n"); 6335 6336 memset(&reset_context, 0, sizeof(reset_context)); 6337 INIT_LIST_HEAD(&device_list); 6338 hive = amdgpu_get_xgmi_hive(adev); 6339 if (hive) { 6340 mutex_lock(&hive->hive_lock); 6341 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) 6342 list_add_tail(&tmp_adev->reset_list, &device_list); 6343 } else { 6344 list_add_tail(&adev->reset_list, &device_list); 6345 } 6346 6347 if (adev->pcie_reset_ctx.swus) 6348 link_dev = adev->pcie_reset_ctx.swus; 6349 else 6350 link_dev = adev->pdev; 6351 /* wait for asic to come out of reset, timeout = 10s */ 6352 timeout = 10000; 6353 do { 6354 usleep_range(10000, 10500); 6355 r = pci_read_config_word(link_dev, PCI_VENDOR_ID, &status); 6356 timeout -= 10; 6357 } while (timeout > 0 && (status != PCI_VENDOR_ID_ATI) && 6358 (status != PCI_VENDOR_ID_AMD)); 6359 6360 if ((status != PCI_VENDOR_ID_ATI) && (status != PCI_VENDOR_ID_AMD)) { 6361 r = -ETIME; 6362 goto out; 6363 } 6364 6365 amdgpu_device_load_switch_state(adev); 6366 /* Restore PCI confspace */ 6367 amdgpu_device_load_pci_state(pdev); 6368 6369 /* confirm ASIC came out of reset */ 6370 for (i = 0; i < adev->usec_timeout; i++) { 6371 memsize = amdgpu_asic_get_config_memsize(adev); 6372 6373 if (memsize != 0xffffffff) 6374 break; 6375 udelay(1); 6376 } 6377 if (memsize == 0xffffffff) { 6378 r = -ETIME; 6379 goto out; 6380 } 6381 6382 reset_context.method = AMD_RESET_METHOD_NONE; 6383 reset_context.reset_req_dev = adev; 6384 set_bit(AMDGPU_NEED_FULL_RESET, &reset_context.flags); 6385 set_bit(AMDGPU_SKIP_COREDUMP, &reset_context.flags); 6386 6387 if (hive) { 6388 reset_context.hive = hive; 6389 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) 6390 tmp_adev->pcie_reset_ctx.in_link_reset = true; 6391 } else { 6392 set_bit(AMDGPU_SKIP_HW_RESET, &reset_context.flags); 6393 } 6394 6395 r = amdgpu_device_asic_reset(adev, &device_list, &reset_context); 6396 out: 6397 if (!r) { 6398 if (amdgpu_device_cache_pci_state(adev->pdev)) 6399 pci_restore_state(adev->pdev); 6400 dev_info(adev->dev, "PCIe error recovery succeeded\n"); 6401 } else { 6402 dev_err(adev->dev, "PCIe error recovery failed, err:%d\n", r); 6403 if (hive) { 6404 list_for_each_entry(tmp_adev, &device_list, reset_list) 6405 amdgpu_device_unset_mp1_state(tmp_adev); 6406 } 6407 amdgpu_device_recovery_put_reset_lock(adev, &device_list); 6408 } 6409 6410 if (hive) { 6411 mutex_unlock(&hive->hive_lock); 6412 amdgpu_put_xgmi_hive(hive); 6413 } 6414 6415 return r ? PCI_ERS_RESULT_DISCONNECT : PCI_ERS_RESULT_RECOVERED; 6416 } 6417 6418 /** 6419 * amdgpu_pci_resume() - resume normal ops after PCI reset 6420 * @pdev: pointer to PCI device 6421 * 6422 * Called when the error recovery driver tells us that its 6423 * OK to resume normal operation. 6424 */ 6425 void amdgpu_pci_resume(struct pci_dev *pdev) 6426 { 6427 struct drm_device *dev = pci_get_drvdata(pdev); 6428 struct amdgpu_device *adev = drm_to_adev(dev); 6429 struct list_head device_list; 6430 struct amdgpu_hive_info *hive = NULL; 6431 struct amdgpu_device *tmp_adev = NULL; 6432 6433 dev_info(adev->dev, "PCI error: resume callback!!\n"); 6434 6435 /* Only continue execution for the case of pci_channel_io_frozen */ 6436 if (adev->pci_channel_state != pci_channel_io_frozen) 6437 return; 6438 6439 INIT_LIST_HEAD(&device_list); 6440 6441 hive = amdgpu_get_xgmi_hive(adev); 6442 if (hive) { 6443 mutex_lock(&hive->hive_lock); 6444 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { 6445 tmp_adev->pcie_reset_ctx.in_link_reset = false; 6446 list_add_tail(&tmp_adev->reset_list, &device_list); 6447 } 6448 } else 6449 list_add_tail(&adev->reset_list, &device_list); 6450 6451 amdgpu_device_sched_resume(&device_list, NULL, NULL); 6452 amdgpu_device_gpu_resume(adev, &device_list, false); 6453 amdgpu_device_recovery_put_reset_lock(adev, &device_list); 6454 6455 if (hive) { 6456 mutex_unlock(&hive->hive_lock); 6457 amdgpu_put_xgmi_hive(hive); 6458 } 6459 } 6460 6461 static void amdgpu_device_cache_switch_state(struct amdgpu_device *adev) 6462 { 6463 struct pci_dev *swus, *swds; 6464 int r; 6465 6466 swds = pci_upstream_bridge(adev->pdev); 6467 if (!swds || swds->vendor != PCI_VENDOR_ID_ATI || 6468 pci_pcie_type(swds) != PCI_EXP_TYPE_DOWNSTREAM) 6469 return; 6470 swus = pci_upstream_bridge(swds); 6471 if (!swus || 6472 (swus->vendor != PCI_VENDOR_ID_ATI && 6473 swus->vendor != PCI_VENDOR_ID_AMD) || 6474 pci_pcie_type(swus) != PCI_EXP_TYPE_UPSTREAM) 6475 return; 6476 6477 /* If already saved, return */ 6478 if (adev->pcie_reset_ctx.swus) 6479 return; 6480 /* Upstream bridge is ATI, assume it's SWUS/DS architecture */ 6481 r = pci_save_state(swds); 6482 if (r) 6483 return; 6484 adev->pcie_reset_ctx.swds_pcistate = pci_store_saved_state(swds); 6485 6486 r = pci_save_state(swus); 6487 if (r) 6488 return; 6489 adev->pcie_reset_ctx.swus_pcistate = pci_store_saved_state(swus); 6490 6491 adev->pcie_reset_ctx.swus = swus; 6492 } 6493 6494 static void amdgpu_device_load_switch_state(struct amdgpu_device *adev) 6495 { 6496 struct pci_dev *pdev; 6497 int r; 6498 6499 if (!adev->pcie_reset_ctx.swds_pcistate || 6500 !adev->pcie_reset_ctx.swus_pcistate) 6501 return; 6502 6503 pdev = adev->pcie_reset_ctx.swus; 6504 r = pci_load_saved_state(pdev, adev->pcie_reset_ctx.swus_pcistate); 6505 if (!r) { 6506 pci_restore_state(pdev); 6507 } else { 6508 dev_warn(adev->dev, "Failed to load SWUS state, err:%d\n", r); 6509 return; 6510 } 6511 6512 pdev = pci_upstream_bridge(adev->pdev); 6513 r = pci_load_saved_state(pdev, adev->pcie_reset_ctx.swds_pcistate); 6514 if (!r) 6515 pci_restore_state(pdev); 6516 else 6517 dev_warn(adev->dev, "Failed to load SWDS state, err:%d\n", r); 6518 } 6519 6520 bool amdgpu_device_cache_pci_state(struct pci_dev *pdev) 6521 { 6522 struct drm_device *dev = pci_get_drvdata(pdev); 6523 struct amdgpu_device *adev = drm_to_adev(dev); 6524 int r; 6525 6526 if (amdgpu_sriov_vf(adev)) 6527 return false; 6528 6529 r = pci_save_state(pdev); 6530 if (!r) { 6531 kfree(adev->pci_state); 6532 6533 adev->pci_state = pci_store_saved_state(pdev); 6534 6535 if (!adev->pci_state) { 6536 dev_err(adev->dev, "Failed to store PCI saved state"); 6537 return false; 6538 } 6539 } else { 6540 dev_warn(adev->dev, "Failed to save PCI state, err:%d\n", r); 6541 return false; 6542 } 6543 6544 amdgpu_device_cache_switch_state(adev); 6545 6546 return true; 6547 } 6548 6549 bool amdgpu_device_load_pci_state(struct pci_dev *pdev) 6550 { 6551 struct drm_device *dev = pci_get_drvdata(pdev); 6552 struct amdgpu_device *adev = drm_to_adev(dev); 6553 int r; 6554 6555 if (!adev->pci_state) 6556 return false; 6557 6558 r = pci_load_saved_state(pdev, adev->pci_state); 6559 6560 if (!r) { 6561 pci_restore_state(pdev); 6562 } else { 6563 dev_warn(adev->dev, "Failed to load PCI state, err:%d\n", r); 6564 return false; 6565 } 6566 6567 return true; 6568 } 6569 6570 void amdgpu_device_flush_hdp(struct amdgpu_device *adev, 6571 struct amdgpu_ring *ring) 6572 { 6573 #ifdef CONFIG_X86_64 6574 if ((adev->flags & AMD_IS_APU) && !amdgpu_passthrough(adev)) 6575 return; 6576 #endif 6577 if (adev->gmc.xgmi.connected_to_cpu) 6578 return; 6579 6580 if (ring && ring->funcs->emit_hdp_flush) { 6581 amdgpu_ring_emit_hdp_flush(ring); 6582 return; 6583 } 6584 6585 if (!ring && amdgpu_sriov_runtime(adev)) { 6586 if (!amdgpu_kiq_hdp_flush(adev)) 6587 return; 6588 } 6589 6590 amdgpu_hdp_flush(adev, ring); 6591 } 6592 6593 void amdgpu_device_invalidate_hdp(struct amdgpu_device *adev, 6594 struct amdgpu_ring *ring) 6595 { 6596 #ifdef CONFIG_X86_64 6597 if ((adev->flags & AMD_IS_APU) && !amdgpu_passthrough(adev)) 6598 return; 6599 #endif 6600 if (adev->gmc.xgmi.connected_to_cpu) 6601 return; 6602 6603 amdgpu_hdp_invalidate(adev, ring); 6604 } 6605 6606 int amdgpu_in_reset(struct amdgpu_device *adev) 6607 { 6608 return atomic_read(&adev->reset_domain->in_gpu_reset); 6609 } 6610 6611 /** 6612 * amdgpu_device_halt() - bring hardware to some kind of halt state 6613 * 6614 * @adev: amdgpu_device pointer 6615 * 6616 * Bring hardware to some kind of halt state so that no one can touch it 6617 * any more. It will help to maintain error context when error occurred. 6618 * Compare to a simple hang, the system will keep stable at least for SSH 6619 * access. Then it should be trivial to inspect the hardware state and 6620 * see what's going on. Implemented as following: 6621 * 6622 * 1. drm_dev_unplug() makes device inaccessible to user space(IOCTLs, etc), 6623 * clears all CPU mappings to device, disallows remappings through page faults 6624 * 2. amdgpu_irq_disable_all() disables all interrupts 6625 * 3. amdgpu_fence_driver_hw_fini() signals all HW fences 6626 * 4. set adev->no_hw_access to avoid potential crashes after setp 5 6627 * 5. amdgpu_device_unmap_mmio() clears all MMIO mappings 6628 * 6. pci_disable_device() and pci_wait_for_pending_transaction() 6629 * flush any in flight DMA operations 6630 */ 6631 void amdgpu_device_halt(struct amdgpu_device *adev) 6632 { 6633 struct pci_dev *pdev = adev->pdev; 6634 struct drm_device *ddev = adev_to_drm(adev); 6635 6636 amdgpu_xcp_dev_unplug(adev); 6637 drm_dev_unplug(ddev); 6638 6639 amdgpu_device_set_pg_state(adev, AMD_PG_STATE_UNGATE); 6640 amdgpu_device_set_cg_state(adev, AMD_CG_STATE_UNGATE); 6641 6642 amdgpu_irq_disable_all(adev); 6643 6644 amdgpu_fence_driver_hw_fini(adev); 6645 6646 adev->no_hw_access = true; 6647 6648 amdgpu_device_unmap_mmio(adev); 6649 6650 pci_disable_device(pdev); 6651 pci_wait_for_pending_transaction(pdev); 6652 } 6653 6654 /** 6655 * amdgpu_device_get_gang - return a reference to the current gang 6656 * @adev: amdgpu_device pointer 6657 * 6658 * Returns: A new reference to the current gang leader. 6659 */ 6660 struct dma_fence *amdgpu_device_get_gang(struct amdgpu_device *adev) 6661 { 6662 struct dma_fence *fence; 6663 6664 rcu_read_lock(); 6665 fence = dma_fence_get_rcu_safe(&adev->gang_submit); 6666 rcu_read_unlock(); 6667 return fence; 6668 } 6669 6670 /** 6671 * amdgpu_device_switch_gang - switch to a new gang 6672 * @adev: amdgpu_device pointer 6673 * @gang: the gang to switch to 6674 * 6675 * Try to switch to a new gang. 6676 * Returns: NULL if we switched to the new gang or a reference to the current 6677 * gang leader. 6678 */ 6679 struct dma_fence *amdgpu_device_switch_gang(struct amdgpu_device *adev, 6680 struct dma_fence *gang) 6681 { 6682 struct dma_fence *old = NULL; 6683 6684 dma_fence_get(gang); 6685 do { 6686 dma_fence_put(old); 6687 old = amdgpu_device_get_gang(adev); 6688 if (old == gang) 6689 break; 6690 6691 if (!dma_fence_is_signaled(old)) { 6692 dma_fence_put(gang); 6693 return old; 6694 } 6695 6696 } while (cmpxchg((struct dma_fence __force **)&adev->gang_submit, 6697 old, gang) != old); 6698 6699 /* 6700 * Drop it once for the exchanged reference in adev and once for the 6701 * thread local reference acquired in amdgpu_device_get_gang(). 6702 */ 6703 dma_fence_put(old); 6704 dma_fence_put(old); 6705 return NULL; 6706 } 6707 6708 /** 6709 * amdgpu_device_enforce_isolation - enforce HW isolation 6710 * @adev: the amdgpu device pointer 6711 * @ring: the HW ring the job is supposed to run on 6712 * @job: the job which is about to be pushed to the HW ring 6713 * 6714 * Makes sure that only one client at a time can use the GFX block. 6715 * Returns: The dependency to wait on before the job can be pushed to the HW. 6716 * The function is called multiple times until NULL is returned. 6717 */ 6718 struct dma_fence *amdgpu_device_enforce_isolation(struct amdgpu_device *adev, 6719 struct amdgpu_ring *ring, 6720 struct amdgpu_job *job) 6721 { 6722 struct amdgpu_isolation *isolation = &adev->isolation[ring->xcp_id]; 6723 struct drm_sched_fence *f = job->base.s_fence; 6724 struct dma_fence *dep; 6725 void *owner; 6726 int r; 6727 6728 /* 6729 * For now enforce isolation only for the GFX block since we only need 6730 * the cleaner shader on those rings. 6731 */ 6732 if (ring->funcs->type != AMDGPU_RING_TYPE_GFX && 6733 ring->funcs->type != AMDGPU_RING_TYPE_COMPUTE) 6734 return NULL; 6735 6736 /* 6737 * All submissions where enforce isolation is false are handled as if 6738 * they come from a single client. Use ~0l as the owner to distinct it 6739 * from kernel submissions where the owner is NULL. 6740 */ 6741 owner = job->enforce_isolation ? f->owner : (void *)~0l; 6742 6743 mutex_lock(&adev->enforce_isolation_mutex); 6744 6745 /* 6746 * The "spearhead" submission is the first one which changes the 6747 * ownership to its client. We always need to wait for it to be 6748 * pushed to the HW before proceeding with anything. 6749 */ 6750 if (&f->scheduled != isolation->spearhead && 6751 !dma_fence_is_signaled(isolation->spearhead)) { 6752 dep = isolation->spearhead; 6753 goto out_grab_ref; 6754 } 6755 6756 if (isolation->owner != owner) { 6757 6758 /* 6759 * Wait for any gang to be assembled before switching to a 6760 * different owner or otherwise we could deadlock the 6761 * submissions. 6762 */ 6763 if (!job->gang_submit) { 6764 dep = amdgpu_device_get_gang(adev); 6765 if (!dma_fence_is_signaled(dep)) 6766 goto out_return_dep; 6767 dma_fence_put(dep); 6768 } 6769 6770 dma_fence_put(isolation->spearhead); 6771 isolation->spearhead = dma_fence_get(&f->scheduled); 6772 amdgpu_sync_move(&isolation->active, &isolation->prev); 6773 trace_amdgpu_isolation(isolation->owner, owner); 6774 isolation->owner = owner; 6775 } 6776 6777 /* 6778 * Specifying the ring here helps to pipeline submissions even when 6779 * isolation is enabled. If that is not desired for testing NULL can be 6780 * used instead of the ring to enforce a CPU round trip while switching 6781 * between clients. 6782 */ 6783 dep = amdgpu_sync_peek_fence(&isolation->prev, ring); 6784 r = amdgpu_sync_fence(&isolation->active, &f->finished, GFP_NOWAIT); 6785 if (r) 6786 dev_warn(adev->dev, "OOM tracking isolation\n"); 6787 6788 out_grab_ref: 6789 dma_fence_get(dep); 6790 out_return_dep: 6791 mutex_unlock(&adev->enforce_isolation_mutex); 6792 return dep; 6793 } 6794 6795 bool amdgpu_device_has_display_hardware(struct amdgpu_device *adev) 6796 { 6797 switch (adev->asic_type) { 6798 #ifdef CONFIG_DRM_AMDGPU_SI 6799 case CHIP_HAINAN: 6800 #endif 6801 case CHIP_TOPAZ: 6802 /* chips with no display hardware */ 6803 return false; 6804 #ifdef CONFIG_DRM_AMDGPU_SI 6805 case CHIP_TAHITI: 6806 case CHIP_PITCAIRN: 6807 case CHIP_VERDE: 6808 case CHIP_OLAND: 6809 #endif 6810 #ifdef CONFIG_DRM_AMDGPU_CIK 6811 case CHIP_BONAIRE: 6812 case CHIP_HAWAII: 6813 case CHIP_KAVERI: 6814 case CHIP_KABINI: 6815 case CHIP_MULLINS: 6816 #endif 6817 case CHIP_TONGA: 6818 case CHIP_FIJI: 6819 case CHIP_POLARIS10: 6820 case CHIP_POLARIS11: 6821 case CHIP_POLARIS12: 6822 case CHIP_VEGAM: 6823 case CHIP_CARRIZO: 6824 case CHIP_STONEY: 6825 /* chips with display hardware */ 6826 return true; 6827 default: 6828 /* IP discovery */ 6829 if (!amdgpu_ip_version(adev, DCE_HWIP, 0) || 6830 (adev->harvest_ip_mask & AMD_HARVEST_IP_DMU_MASK)) 6831 return false; 6832 return true; 6833 } 6834 } 6835 6836 ssize_t amdgpu_get_soft_full_reset_mask(struct amdgpu_ring *ring) 6837 { 6838 ssize_t size = 0; 6839 6840 if (!ring || !ring->adev) 6841 return size; 6842 6843 if (amdgpu_device_should_recover_gpu(ring->adev)) 6844 size |= AMDGPU_RESET_TYPE_FULL; 6845 6846 if (unlikely(!ring->adev->debug_disable_soft_recovery) && 6847 !amdgpu_sriov_vf(ring->adev) && ring->funcs->soft_recovery) 6848 size |= AMDGPU_RESET_TYPE_SOFT_RESET; 6849 6850 return size; 6851 } 6852 6853 ssize_t amdgpu_show_reset_mask(char *buf, uint32_t supported_reset) 6854 { 6855 ssize_t size = 0; 6856 6857 if (supported_reset == 0) { 6858 size += sysfs_emit_at(buf, size, "unsupported"); 6859 size += sysfs_emit_at(buf, size, "\n"); 6860 return size; 6861 6862 } 6863 6864 if (supported_reset & AMDGPU_RESET_TYPE_SOFT_RESET) 6865 size += sysfs_emit_at(buf, size, "soft "); 6866 6867 if (supported_reset & AMDGPU_RESET_TYPE_PER_QUEUE) 6868 size += sysfs_emit_at(buf, size, "queue "); 6869 6870 if (supported_reset & AMDGPU_RESET_TYPE_PER_PIPE) 6871 size += sysfs_emit_at(buf, size, "pipe "); 6872 6873 if (supported_reset & AMDGPU_RESET_TYPE_FULL) 6874 size += sysfs_emit_at(buf, size, "full "); 6875 6876 size += sysfs_emit_at(buf, size, "\n"); 6877 return size; 6878 } 6879 6880 void amdgpu_device_set_uid(struct amdgpu_uid *uid_info, 6881 enum amdgpu_uid_type type, uint8_t inst, 6882 uint64_t uid) 6883 { 6884 if (!uid_info) 6885 return; 6886 6887 if (type >= AMDGPU_UID_TYPE_MAX) { 6888 dev_err_once(uid_info->adev->dev, "Invalid UID type %d\n", 6889 type); 6890 return; 6891 } 6892 6893 if (inst >= AMDGPU_UID_INST_MAX) { 6894 dev_err_once(uid_info->adev->dev, "Invalid UID instance %d\n", 6895 inst); 6896 return; 6897 } 6898 6899 if (uid_info->uid[type][inst] != 0) { 6900 dev_warn_once( 6901 uid_info->adev->dev, 6902 "Overwriting existing UID %llu for type %d instance %d\n", 6903 uid_info->uid[type][inst], type, inst); 6904 } 6905 6906 uid_info->uid[type][inst] = uid; 6907 } 6908 6909 u64 amdgpu_device_get_uid(struct amdgpu_uid *uid_info, 6910 enum amdgpu_uid_type type, uint8_t inst) 6911 { 6912 if (!uid_info) 6913 return 0; 6914 6915 if (type >= AMDGPU_UID_TYPE_MAX) { 6916 dev_err_once(uid_info->adev->dev, "Invalid UID type %d\n", 6917 type); 6918 return 0; 6919 } 6920 6921 if (inst >= AMDGPU_UID_INST_MAX) { 6922 dev_err_once(uid_info->adev->dev, "Invalid UID instance %d\n", 6923 inst); 6924 return 0; 6925 } 6926 6927 return uid_info->uid[type][inst]; 6928 } 6929