1 /* 2 * Copyright 2023 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 */ 23 24 #include <linux/list.h> 25 #include "amdgpu.h" 26 #include "amdgpu_aca.h" 27 #include "amdgpu_ras.h" 28 29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype} 30 31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data); 32 33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = { 34 ACA_BANK_HWID(SMU, 0x01, 0x01), 35 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00), 36 ACA_BANK_HWID(UMC, 0x96, 0x00), 37 }; 38 39 static void aca_banks_init(struct aca_banks *banks) 40 { 41 if (!banks) 42 return; 43 44 memset(banks, 0, sizeof(*banks)); 45 INIT_LIST_HEAD(&banks->list); 46 } 47 48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank) 49 { 50 struct aca_bank_node *node; 51 52 if (!bank) 53 return -EINVAL; 54 55 node = kvzalloc(sizeof(*node), GFP_KERNEL); 56 if (!node) 57 return -ENOMEM; 58 59 memcpy(&node->bank, bank, sizeof(*bank)); 60 61 INIT_LIST_HEAD(&node->node); 62 list_add_tail(&node->node, &banks->list); 63 64 banks->nr_banks++; 65 66 return 0; 67 } 68 69 static void aca_banks_release(struct aca_banks *banks) 70 { 71 struct aca_bank_node *node, *tmp; 72 73 if (list_empty(&banks->list)) 74 return; 75 76 list_for_each_entry_safe(node, tmp, &banks->list, node) { 77 list_del(&node->node); 78 kvfree(node); 79 } 80 } 81 82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count) 83 { 84 struct amdgpu_aca *aca = &adev->aca; 85 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 86 87 if (!count) 88 return -EINVAL; 89 90 if (!smu_funcs || !smu_funcs->get_valid_aca_count) 91 return -EOPNOTSUPP; 92 93 return smu_funcs->get_valid_aca_count(adev, type, count); 94 } 95 96 static struct aca_regs_dump { 97 const char *name; 98 int reg_idx; 99 } aca_regs[] = { 100 {"CONTROL", ACA_REG_IDX_CTL}, 101 {"STATUS", ACA_REG_IDX_STATUS}, 102 {"ADDR", ACA_REG_IDX_ADDR}, 103 {"MISC", ACA_REG_IDX_MISC0}, 104 {"CONFIG", ACA_REG_IDX_CONFIG}, 105 {"IPID", ACA_REG_IDX_IPID}, 106 {"SYND", ACA_REG_IDX_SYND}, 107 {"DESTAT", ACA_REG_IDX_DESTAT}, 108 {"DEADDR", ACA_REG_IDX_DEADDR}, 109 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK}, 110 }; 111 112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank, 113 struct ras_query_context *qctx) 114 { 115 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID; 116 int i; 117 118 if (adev->debug_disable_ce_logs && 119 bank->smu_err_type == ACA_SMU_TYPE_CE && 120 !ACA_BANK_ERR_IS_DEFFERED(bank)) 121 return; 122 123 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n"); 124 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */ 125 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 126 RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n", 127 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]); 128 129 if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS])) 130 RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n"); 131 } 132 133 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type) 134 { 135 136 struct aca_hwip *hwip; 137 int hwid, mcatype; 138 u64 ipid; 139 140 if (!bank || type == ACA_HWIP_TYPE_UNKNOW) 141 return false; 142 143 hwip = &aca_hwid_mcatypes[type]; 144 if (!hwip->hwid) 145 return false; 146 147 ipid = bank->regs[ACA_REG_IDX_IPID]; 148 hwid = ACA_REG__IPID__HARDWAREID(ipid); 149 mcatype = ACA_REG__IPID__MCATYPE(ipid); 150 151 return hwip->hwid == hwid && hwip->mcatype == mcatype; 152 } 153 154 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type, 155 int start, int count, 156 struct aca_banks *banks, struct ras_query_context *qctx) 157 { 158 struct amdgpu_aca *aca = &adev->aca; 159 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 160 struct aca_bank bank; 161 int i, max_count, ret; 162 163 if (!count) 164 return 0; 165 166 if (!smu_funcs || !smu_funcs->get_valid_aca_bank) 167 return -EOPNOTSUPP; 168 169 switch (type) { 170 case ACA_SMU_TYPE_UE: 171 max_count = smu_funcs->max_ue_bank_count; 172 break; 173 case ACA_SMU_TYPE_CE: 174 max_count = smu_funcs->max_ce_bank_count; 175 break; 176 default: 177 return -EINVAL; 178 } 179 180 if (start + count > max_count) 181 return -EINVAL; 182 183 count = min_t(int, count, max_count); 184 for (i = 0; i < count; i++) { 185 memset(&bank, 0, sizeof(bank)); 186 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank); 187 if (ret) 188 return ret; 189 190 bank.smu_err_type = type; 191 192 /* 193 * Poison being consumed when injecting a UE while running background workloads, 194 * which are unexpected. 195 */ 196 if (type == ACA_SMU_TYPE_UE && 197 ACA_REG__STATUS__POISON(bank.regs[ACA_REG_IDX_STATUS]) && 198 !aca_bank_hwip_is_matched(&bank, ACA_HWIP_TYPE_UMC)) 199 continue; 200 201 aca_smu_bank_dump(adev, i, count, &bank, qctx); 202 203 ret = aca_banks_add_bank(banks, &bank); 204 if (ret) 205 return ret; 206 } 207 208 return 0; 209 } 210 211 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 212 { 213 const struct aca_bank_ops *bank_ops = handle->bank_ops; 214 215 /* Parse all deferred errors with UMC aca handle */ 216 if (ACA_BANK_ERR_IS_DEFFERED(bank)) 217 return handle->hwip == ACA_HWIP_TYPE_UMC; 218 219 if (!aca_bank_hwip_is_matched(bank, handle->hwip)) 220 return false; 221 222 if (!bank_ops->aca_bank_is_valid) 223 return true; 224 225 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data); 226 } 227 228 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 229 { 230 struct aca_bank_error *bank_error; 231 232 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL); 233 if (!bank_error) 234 return NULL; 235 236 INIT_LIST_HEAD(&bank_error->node); 237 memcpy(&bank_error->info, info, sizeof(*info)); 238 239 mutex_lock(&aerr->lock); 240 list_add_tail(&bank_error->node, &aerr->list); 241 mutex_unlock(&aerr->lock); 242 243 return bank_error; 244 } 245 246 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 247 { 248 struct aca_bank_error *bank_error = NULL; 249 struct aca_bank_info *tmp_info; 250 bool found = false; 251 252 mutex_lock(&aerr->lock); 253 list_for_each_entry(bank_error, &aerr->list, node) { 254 tmp_info = &bank_error->info; 255 if (tmp_info->socket_id == info->socket_id && 256 tmp_info->die_id == info->die_id) { 257 found = true; 258 goto out_unlock; 259 } 260 } 261 262 out_unlock: 263 mutex_unlock(&aerr->lock); 264 265 return found ? bank_error : NULL; 266 } 267 268 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error) 269 { 270 if (!aerr || !bank_error) 271 return; 272 273 list_del(&bank_error->node); 274 aerr->nr_errors--; 275 276 kvfree(bank_error); 277 } 278 279 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 280 { 281 struct aca_bank_error *bank_error; 282 283 if (!aerr || !info) 284 return NULL; 285 286 bank_error = find_bank_error(aerr, info); 287 if (bank_error) 288 return bank_error; 289 290 return new_bank_error(aerr, info); 291 } 292 293 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info, 294 enum aca_error_type type, u64 count) 295 { 296 struct aca_error_cache *error_cache = &handle->error_cache; 297 struct aca_bank_error *bank_error; 298 struct aca_error *aerr; 299 300 if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT) 301 return -EINVAL; 302 303 if (!count) 304 return 0; 305 306 aerr = &error_cache->errors[type]; 307 bank_error = get_bank_error(aerr, info); 308 if (!bank_error) 309 return -ENOMEM; 310 311 bank_error->count += count; 312 313 return 0; 314 } 315 316 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 317 { 318 const struct aca_bank_ops *bank_ops = handle->bank_ops; 319 320 if (!bank) 321 return -EINVAL; 322 323 if (!bank_ops->aca_bank_parser) 324 return -EOPNOTSUPP; 325 326 return bank_ops->aca_bank_parser(handle, bank, type, 327 handle->data); 328 } 329 330 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank, 331 enum aca_smu_type type, void *data) 332 { 333 int ret; 334 335 ret = aca_bank_parser(handle, bank, type); 336 if (ret) 337 return ret; 338 339 return 0; 340 } 341 342 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank, 343 enum aca_smu_type type, bank_handler_t handler, void *data) 344 { 345 struct aca_handle *handle; 346 int ret; 347 348 if (list_empty(&mgr->list)) 349 return 0; 350 351 list_for_each_entry(handle, &mgr->list, node) { 352 if (!aca_bank_is_valid(handle, bank, type)) 353 continue; 354 355 ret = handler(handle, bank, type, data); 356 if (ret) 357 return ret; 358 } 359 360 return 0; 361 } 362 363 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks, 364 enum aca_smu_type type, bank_handler_t handler, void *data) 365 { 366 struct aca_bank_node *node; 367 struct aca_bank *bank; 368 int ret; 369 370 if (!mgr || !banks) 371 return -EINVAL; 372 373 /* pre check to avoid unnecessary operations */ 374 if (list_empty(&mgr->list) || list_empty(&banks->list)) 375 return 0; 376 377 list_for_each_entry(node, &banks->list, node) { 378 bank = &node->bank; 379 380 ret = aca_dispatch_bank(mgr, bank, type, handler, data); 381 if (ret) 382 return ret; 383 } 384 385 return 0; 386 } 387 388 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type) 389 { 390 struct amdgpu_aca *aca = &adev->aca; 391 bool ret = true; 392 393 /* 394 * Because the UE Valid MCA count will only be cleared after reset, 395 * in order to avoid repeated counting of the error count, 396 * the aca bank is only updated once during the gpu recovery stage. 397 */ 398 if (type == ACA_SMU_TYPE_UE) { 399 if (amdgpu_ras_intr_triggered()) 400 ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0; 401 else 402 atomic_set(&aca->ue_update_flag, 0); 403 } 404 405 return ret; 406 } 407 408 static void aca_banks_generate_cper(struct amdgpu_device *adev, 409 enum aca_smu_type type, 410 struct aca_banks *banks, 411 int count) 412 { 413 struct aca_bank_node *node; 414 struct aca_bank *bank; 415 int r; 416 417 if (!adev->cper.enabled) 418 return; 419 420 if (!banks || !count) { 421 dev_warn(adev->dev, "fail to generate cper records\n"); 422 return; 423 } 424 425 /* UEs must be encoded into separate CPER entries */ 426 if (type == ACA_SMU_TYPE_UE) { 427 struct aca_banks de_banks; 428 429 aca_banks_init(&de_banks); 430 list_for_each_entry(node, &banks->list, node) { 431 bank = &node->bank; 432 if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) { 433 r = aca_banks_add_bank(&de_banks, bank); 434 if (r) 435 dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r); 436 } else { 437 if (amdgpu_cper_generate_ue_record(adev, bank)) 438 dev_warn(adev->dev, "fail to generate ue cper records\n"); 439 } 440 } 441 442 if (!list_empty(&de_banks.list)) { 443 if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks)) 444 dev_warn(adev->dev, "fail to generate de cper records\n"); 445 } 446 447 aca_banks_release(&de_banks); 448 } else { 449 /* 450 * SMU_TYPE_CE banks are combined into 1 CPER entries, 451 * they could be CEs or DEs or both 452 */ 453 if (amdgpu_cper_generate_ce_records(adev, banks, count)) 454 dev_warn(adev->dev, "fail to generate ce cper records\n"); 455 } 456 } 457 458 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type, 459 bank_handler_t handler, struct ras_query_context *qctx, void *data) 460 { 461 struct amdgpu_aca *aca = &adev->aca; 462 struct aca_banks banks; 463 u32 count = 0; 464 int ret; 465 466 if (list_empty(&aca->mgr.list)) 467 return 0; 468 469 if (!aca_bank_should_update(adev, type)) 470 return 0; 471 472 ret = aca_smu_get_valid_aca_count(adev, type, &count); 473 if (ret) 474 return ret; 475 476 if (!count) 477 return 0; 478 479 aca_banks_init(&banks); 480 481 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx); 482 if (ret) 483 goto err_release_banks; 484 485 if (list_empty(&banks.list)) { 486 ret = 0; 487 goto err_release_banks; 488 } 489 490 ret = aca_dispatch_banks(&aca->mgr, &banks, type, 491 handler, data); 492 if (ret) 493 goto err_release_banks; 494 495 aca_banks_generate_cper(adev, type, &banks, count); 496 497 err_release_banks: 498 aca_banks_release(&banks); 499 500 return ret; 501 } 502 503 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data) 504 { 505 struct aca_bank_info *info; 506 struct amdgpu_smuio_mcm_config_info mcm_info; 507 u64 count; 508 509 if (type >= ACA_ERROR_TYPE_COUNT) 510 return -EINVAL; 511 512 count = bank_error->count; 513 if (!count) 514 return 0; 515 516 info = &bank_error->info; 517 mcm_info.die_id = info->die_id; 518 mcm_info.socket_id = info->socket_id; 519 520 switch (type) { 521 case ACA_ERROR_TYPE_UE: 522 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count); 523 break; 524 case ACA_ERROR_TYPE_CE: 525 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count); 526 break; 527 case ACA_ERROR_TYPE_DEFERRED: 528 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count); 529 break; 530 default: 531 break; 532 } 533 534 return 0; 535 } 536 537 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data) 538 { 539 struct aca_error_cache *error_cache = &handle->error_cache; 540 struct aca_error *aerr = &error_cache->errors[type]; 541 struct aca_bank_error *bank_error, *tmp; 542 543 mutex_lock(&aerr->lock); 544 545 if (list_empty(&aerr->list)) 546 goto out_unlock; 547 548 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) { 549 aca_log_aca_error_data(bank_error, type, err_data); 550 aca_bank_error_remove(aerr, bank_error); 551 } 552 553 out_unlock: 554 mutex_unlock(&aerr->lock); 555 556 return 0; 557 } 558 559 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type, 560 struct ras_err_data *err_data, struct ras_query_context *qctx) 561 { 562 enum aca_smu_type smu_type; 563 int ret; 564 565 switch (type) { 566 case ACA_ERROR_TYPE_UE: 567 smu_type = ACA_SMU_TYPE_UE; 568 break; 569 case ACA_ERROR_TYPE_CE: 570 case ACA_ERROR_TYPE_DEFERRED: 571 smu_type = ACA_SMU_TYPE_CE; 572 break; 573 default: 574 return -EINVAL; 575 } 576 577 /* update aca bank to aca source error_cache first */ 578 ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL); 579 if (ret) 580 return ret; 581 582 /* DEs may contain in CEs or UEs */ 583 if (type != ACA_ERROR_TYPE_DEFERRED) 584 aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data); 585 586 return aca_log_aca_error(handle, type, err_data); 587 } 588 589 static bool aca_handle_is_valid(struct aca_handle *handle) 590 { 591 if (!handle->mask || !list_empty(&handle->node)) 592 return false; 593 594 return true; 595 } 596 597 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, 598 enum aca_error_type type, struct ras_err_data *err_data, 599 struct ras_query_context *qctx) 600 { 601 if (!handle || !err_data) 602 return -EINVAL; 603 604 if (aca_handle_is_valid(handle)) 605 return -EOPNOTSUPP; 606 607 if ((type < 0) || (!(BIT(type) & handle->mask))) 608 return 0; 609 610 return __aca_get_error_data(adev, handle, type, err_data, qctx); 611 } 612 613 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type) 614 { 615 mutex_init(&aerr->lock); 616 INIT_LIST_HEAD(&aerr->list); 617 aerr->type = type; 618 aerr->nr_errors = 0; 619 } 620 621 static void aca_init_error_cache(struct aca_handle *handle) 622 { 623 struct aca_error_cache *error_cache = &handle->error_cache; 624 int type; 625 626 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 627 aca_error_init(&error_cache->errors[type], type); 628 } 629 630 static void aca_error_fini(struct aca_error *aerr) 631 { 632 struct aca_bank_error *bank_error, *tmp; 633 634 mutex_lock(&aerr->lock); 635 if (list_empty(&aerr->list)) 636 goto out_unlock; 637 638 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) 639 aca_bank_error_remove(aerr, bank_error); 640 641 out_unlock: 642 mutex_destroy(&aerr->lock); 643 } 644 645 static void aca_fini_error_cache(struct aca_handle *handle) 646 { 647 struct aca_error_cache *error_cache = &handle->error_cache; 648 int type; 649 650 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 651 aca_error_fini(&error_cache->errors[type]); 652 } 653 654 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle, 655 const char *name, const struct aca_info *ras_info, void *data) 656 { 657 memset(handle, 0, sizeof(*handle)); 658 659 handle->adev = adev; 660 handle->mgr = mgr; 661 handle->name = name; 662 handle->hwip = ras_info->hwip; 663 handle->mask = ras_info->mask; 664 handle->bank_ops = ras_info->bank_ops; 665 handle->data = data; 666 aca_init_error_cache(handle); 667 668 INIT_LIST_HEAD(&handle->node); 669 list_add_tail(&handle->node, &mgr->list); 670 mgr->nr_handles++; 671 672 return 0; 673 } 674 675 static ssize_t aca_sysfs_read(struct device *dev, 676 struct device_attribute *attr, char *buf) 677 { 678 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr); 679 680 /* NOTE: the aca cache will be auto cleared once read, 681 * So the driver should unify the query entry point, forward request to ras query interface directly */ 682 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data); 683 } 684 685 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle) 686 { 687 struct device_attribute *aca_attr = &handle->aca_attr; 688 689 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name); 690 aca_attr->show = aca_sysfs_read; 691 aca_attr->attr.name = handle->attr_name; 692 aca_attr->attr.mode = S_IRUGO; 693 sysfs_attr_init(&aca_attr->attr); 694 695 return sysfs_add_file_to_group(&adev->dev->kobj, 696 &aca_attr->attr, 697 "ras"); 698 } 699 700 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle, 701 const char *name, const struct aca_info *ras_info, void *data) 702 { 703 struct amdgpu_aca *aca = &adev->aca; 704 int ret; 705 706 if (!amdgpu_aca_is_enabled(adev)) 707 return 0; 708 709 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data); 710 if (ret) 711 return ret; 712 713 return add_aca_sysfs(adev, handle); 714 } 715 716 static void remove_aca_handle(struct aca_handle *handle) 717 { 718 struct aca_handle_manager *mgr = handle->mgr; 719 720 aca_fini_error_cache(handle); 721 list_del(&handle->node); 722 mgr->nr_handles--; 723 } 724 725 static void remove_aca_sysfs(struct aca_handle *handle) 726 { 727 struct amdgpu_device *adev = handle->adev; 728 struct device_attribute *aca_attr = &handle->aca_attr; 729 730 if (adev->dev->kobj.sd) 731 sysfs_remove_file_from_group(&adev->dev->kobj, 732 &aca_attr->attr, 733 "ras"); 734 } 735 736 void amdgpu_aca_remove_handle(struct aca_handle *handle) 737 { 738 if (!handle || list_empty(&handle->node)) 739 return; 740 741 remove_aca_sysfs(handle); 742 remove_aca_handle(handle); 743 } 744 745 static int aca_manager_init(struct aca_handle_manager *mgr) 746 { 747 INIT_LIST_HEAD(&mgr->list); 748 mgr->nr_handles = 0; 749 750 return 0; 751 } 752 753 static void aca_manager_fini(struct aca_handle_manager *mgr) 754 { 755 struct aca_handle *handle, *tmp; 756 757 if (list_empty(&mgr->list)) 758 return; 759 760 list_for_each_entry_safe(handle, tmp, &mgr->list, node) 761 amdgpu_aca_remove_handle(handle); 762 } 763 764 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev) 765 { 766 return (adev->aca.is_enabled || 767 adev->debug_enable_ras_aca); 768 } 769 770 int amdgpu_aca_init(struct amdgpu_device *adev) 771 { 772 struct amdgpu_aca *aca = &adev->aca; 773 int ret; 774 775 atomic_set(&aca->ue_update_flag, 0); 776 777 ret = aca_manager_init(&aca->mgr); 778 if (ret) 779 return ret; 780 781 return 0; 782 } 783 784 void amdgpu_aca_fini(struct amdgpu_device *adev) 785 { 786 struct amdgpu_aca *aca = &adev->aca; 787 788 aca_manager_fini(&aca->mgr); 789 790 atomic_set(&aca->ue_update_flag, 0); 791 } 792 793 int amdgpu_aca_reset(struct amdgpu_device *adev) 794 { 795 struct amdgpu_aca *aca = &adev->aca; 796 797 atomic_set(&aca->ue_update_flag, 0); 798 799 return 0; 800 } 801 802 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs) 803 { 804 struct amdgpu_aca *aca = &adev->aca; 805 806 WARN_ON(aca->smu_funcs); 807 aca->smu_funcs = smu_funcs; 808 } 809 810 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info) 811 { 812 u64 ipid; 813 u32 instidhi, instidlo; 814 815 if (!bank || !info) 816 return -EINVAL; 817 818 ipid = bank->regs[ACA_REG_IDX_IPID]; 819 info->hwid = ACA_REG__IPID__HARDWAREID(ipid); 820 info->mcatype = ACA_REG__IPID__MCATYPE(ipid); 821 /* 822 * Unfied DieID Format: SAASS. A:AID, S:Socket. 823 * Unfied DieID[4:4] = InstanceId[0:0] 824 * Unfied DieID[0:3] = InstanceIdHi[0:3] 825 */ 826 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid); 827 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid); 828 info->die_id = ((instidhi >> 2) & 0x03); 829 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03); 830 831 return 0; 832 } 833 834 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank) 835 { 836 struct amdgpu_aca *aca = &adev->aca; 837 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 838 839 if (!smu_funcs || !smu_funcs->parse_error_code) 840 return -EOPNOTSUPP; 841 842 return smu_funcs->parse_error_code(adev, bank); 843 } 844 845 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size) 846 { 847 int i, error_code; 848 849 if (!bank || !err_codes) 850 return -EINVAL; 851 852 error_code = aca_bank_get_error_code(adev, bank); 853 if (error_code < 0) 854 return error_code; 855 856 for (i = 0; i < size; i++) { 857 if (err_codes[i] == error_code) 858 return 0; 859 } 860 861 return -EINVAL; 862 } 863 864 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en) 865 { 866 struct amdgpu_aca *aca = &adev->aca; 867 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 868 869 if (!smu_funcs || !smu_funcs->set_debug_mode) 870 return -EOPNOTSUPP; 871 872 return smu_funcs->set_debug_mode(adev, en); 873 } 874 875 #if defined(CONFIG_DEBUG_FS) 876 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val) 877 { 878 struct amdgpu_device *adev = (struct amdgpu_device *)data; 879 int ret; 880 881 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false); 882 if (ret) 883 return ret; 884 885 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off"); 886 887 return 0; 888 } 889 890 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx) 891 { 892 struct aca_bank_info info; 893 int i, ret; 894 895 ret = aca_bank_info_decode(bank, &info); 896 if (ret) 897 return; 898 899 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE"); 900 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n", 901 idx, info.socket_id, info.die_id, info.hwid, info.mcatype); 902 903 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 904 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]); 905 } 906 907 struct aca_dump_context { 908 struct seq_file *m; 909 int idx; 910 }; 911 912 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank, 913 enum aca_smu_type type, void *data) 914 { 915 struct aca_dump_context *ctx = (struct aca_dump_context *)data; 916 917 aca_dump_entry(ctx->m, bank, type, ctx->idx++); 918 919 return handler_aca_log_bank_error(handle, bank, type, NULL); 920 } 921 922 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type) 923 { 924 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 925 struct aca_dump_context context = { 926 .m = m, 927 .idx = 0, 928 }; 929 930 return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context); 931 } 932 933 static int aca_dump_ce_show(struct seq_file *m, void *unused) 934 { 935 return aca_dump_show(m, ACA_SMU_TYPE_CE); 936 } 937 938 static int aca_dump_ce_open(struct inode *inode, struct file *file) 939 { 940 return single_open(file, aca_dump_ce_show, inode->i_private); 941 } 942 943 static const struct file_operations aca_ce_dump_debug_fops = { 944 .owner = THIS_MODULE, 945 .open = aca_dump_ce_open, 946 .read = seq_read, 947 .llseek = seq_lseek, 948 .release = single_release, 949 }; 950 951 static int aca_dump_ue_show(struct seq_file *m, void *unused) 952 { 953 return aca_dump_show(m, ACA_SMU_TYPE_UE); 954 } 955 956 static int aca_dump_ue_open(struct inode *inode, struct file *file) 957 { 958 return single_open(file, aca_dump_ue_show, inode->i_private); 959 } 960 961 static const struct file_operations aca_ue_dump_debug_fops = { 962 .owner = THIS_MODULE, 963 .open = aca_dump_ue_open, 964 .read = seq_read, 965 .llseek = seq_lseek, 966 .release = single_release, 967 }; 968 969 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n"); 970 #endif 971 972 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root) 973 { 974 #if defined(CONFIG_DEBUG_FS) 975 if (!root) 976 return; 977 978 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops); 979 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops); 980 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops); 981 #endif 982 } 983