1 /* 2 * Copyright 2023 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 */ 23 24 #include <linux/list.h> 25 #include "amdgpu.h" 26 #include "amdgpu_aca.h" 27 #include "amdgpu_ras.h" 28 29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype} 30 31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data); 32 33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = { 34 ACA_BANK_HWID(SMU, 0x01, 0x01), 35 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00), 36 ACA_BANK_HWID(UMC, 0x96, 0x00), 37 }; 38 39 static void aca_banks_init(struct aca_banks *banks) 40 { 41 if (!banks) 42 return; 43 44 memset(banks, 0, sizeof(*banks)); 45 INIT_LIST_HEAD(&banks->list); 46 } 47 48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank) 49 { 50 struct aca_bank_node *node; 51 52 if (!bank) 53 return -EINVAL; 54 55 node = kvzalloc(sizeof(*node), GFP_KERNEL); 56 if (!node) 57 return -ENOMEM; 58 59 memcpy(&node->bank, bank, sizeof(*bank)); 60 61 INIT_LIST_HEAD(&node->node); 62 list_add_tail(&node->node, &banks->list); 63 64 banks->nr_banks++; 65 66 return 0; 67 } 68 69 static void aca_banks_release(struct aca_banks *banks) 70 { 71 struct aca_bank_node *node, *tmp; 72 73 if (list_empty(&banks->list)) 74 return; 75 76 list_for_each_entry_safe(node, tmp, &banks->list, node) { 77 list_del(&node->node); 78 kvfree(node); 79 } 80 } 81 82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count) 83 { 84 struct amdgpu_aca *aca = &adev->aca; 85 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 86 87 if (!count) 88 return -EINVAL; 89 90 if (!smu_funcs || !smu_funcs->get_valid_aca_count) 91 return -EOPNOTSUPP; 92 93 return smu_funcs->get_valid_aca_count(adev, type, count); 94 } 95 96 static struct aca_regs_dump { 97 const char *name; 98 int reg_idx; 99 } aca_regs[] = { 100 {"CONTROL", ACA_REG_IDX_CTL}, 101 {"STATUS", ACA_REG_IDX_STATUS}, 102 {"ADDR", ACA_REG_IDX_ADDR}, 103 {"MISC", ACA_REG_IDX_MISC0}, 104 {"CONFIG", ACA_REG_IDX_CONFIG}, 105 {"IPID", ACA_REG_IDX_IPID}, 106 {"SYND", ACA_REG_IDX_SYND}, 107 {"DESTAT", ACA_REG_IDX_DESTAT}, 108 {"DEADDR", ACA_REG_IDX_DEADDR}, 109 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK}, 110 }; 111 112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank, 113 struct ras_query_context *qctx) 114 { 115 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID; 116 int i; 117 118 if (adev->debug_disable_ce_logs && 119 bank->smu_err_type == ACA_SMU_TYPE_CE && 120 !ACA_BANK_ERR_IS_DEFFERED(bank)) 121 return; 122 123 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n"); 124 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */ 125 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 126 RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n", 127 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]); 128 129 if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS])) 130 RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n"); 131 } 132 133 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type, 134 int start, int count, 135 struct aca_banks *banks, struct ras_query_context *qctx) 136 { 137 struct amdgpu_aca *aca = &adev->aca; 138 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 139 struct aca_bank bank; 140 int i, max_count, ret; 141 142 if (!count) 143 return 0; 144 145 if (!smu_funcs || !smu_funcs->get_valid_aca_bank) 146 return -EOPNOTSUPP; 147 148 switch (type) { 149 case ACA_SMU_TYPE_UE: 150 max_count = smu_funcs->max_ue_bank_count; 151 break; 152 case ACA_SMU_TYPE_CE: 153 max_count = smu_funcs->max_ce_bank_count; 154 break; 155 default: 156 return -EINVAL; 157 } 158 159 if (start + count > max_count) 160 return -EINVAL; 161 162 count = min_t(int, count, max_count); 163 for (i = 0; i < count; i++) { 164 memset(&bank, 0, sizeof(bank)); 165 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank); 166 if (ret) 167 return ret; 168 169 bank.smu_err_type = type; 170 171 aca_smu_bank_dump(adev, i, count, &bank, qctx); 172 173 ret = aca_banks_add_bank(banks, &bank); 174 if (ret) 175 return ret; 176 } 177 178 return 0; 179 } 180 181 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type) 182 { 183 184 struct aca_hwip *hwip; 185 int hwid, mcatype; 186 u64 ipid; 187 188 if (!bank || type == ACA_HWIP_TYPE_UNKNOW) 189 return false; 190 191 hwip = &aca_hwid_mcatypes[type]; 192 if (!hwip->hwid) 193 return false; 194 195 ipid = bank->regs[ACA_REG_IDX_IPID]; 196 hwid = ACA_REG__IPID__HARDWAREID(ipid); 197 mcatype = ACA_REG__IPID__MCATYPE(ipid); 198 199 return hwip->hwid == hwid && hwip->mcatype == mcatype; 200 } 201 202 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 203 { 204 const struct aca_bank_ops *bank_ops = handle->bank_ops; 205 206 /* Parse all deferred errors with UMC aca handle */ 207 if (ACA_BANK_ERR_IS_DEFFERED(bank)) 208 return handle->hwip == ACA_HWIP_TYPE_UMC; 209 210 if (!aca_bank_hwip_is_matched(bank, handle->hwip)) 211 return false; 212 213 if (!bank_ops->aca_bank_is_valid) 214 return true; 215 216 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data); 217 } 218 219 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 220 { 221 struct aca_bank_error *bank_error; 222 223 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL); 224 if (!bank_error) 225 return NULL; 226 227 INIT_LIST_HEAD(&bank_error->node); 228 memcpy(&bank_error->info, info, sizeof(*info)); 229 230 mutex_lock(&aerr->lock); 231 list_add_tail(&bank_error->node, &aerr->list); 232 mutex_unlock(&aerr->lock); 233 234 return bank_error; 235 } 236 237 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 238 { 239 struct aca_bank_error *bank_error = NULL; 240 struct aca_bank_info *tmp_info; 241 bool found = false; 242 243 mutex_lock(&aerr->lock); 244 list_for_each_entry(bank_error, &aerr->list, node) { 245 tmp_info = &bank_error->info; 246 if (tmp_info->socket_id == info->socket_id && 247 tmp_info->die_id == info->die_id) { 248 found = true; 249 goto out_unlock; 250 } 251 } 252 253 out_unlock: 254 mutex_unlock(&aerr->lock); 255 256 return found ? bank_error : NULL; 257 } 258 259 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error) 260 { 261 if (!aerr || !bank_error) 262 return; 263 264 list_del(&bank_error->node); 265 aerr->nr_errors--; 266 267 kvfree(bank_error); 268 } 269 270 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 271 { 272 struct aca_bank_error *bank_error; 273 274 if (!aerr || !info) 275 return NULL; 276 277 bank_error = find_bank_error(aerr, info); 278 if (bank_error) 279 return bank_error; 280 281 return new_bank_error(aerr, info); 282 } 283 284 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info, 285 enum aca_error_type type, u64 count) 286 { 287 struct aca_error_cache *error_cache = &handle->error_cache; 288 struct aca_bank_error *bank_error; 289 struct aca_error *aerr; 290 291 if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT) 292 return -EINVAL; 293 294 if (!count) 295 return 0; 296 297 aerr = &error_cache->errors[type]; 298 bank_error = get_bank_error(aerr, info); 299 if (!bank_error) 300 return -ENOMEM; 301 302 bank_error->count += count; 303 304 return 0; 305 } 306 307 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 308 { 309 const struct aca_bank_ops *bank_ops = handle->bank_ops; 310 311 if (!bank) 312 return -EINVAL; 313 314 if (!bank_ops->aca_bank_parser) 315 return -EOPNOTSUPP; 316 317 return bank_ops->aca_bank_parser(handle, bank, type, 318 handle->data); 319 } 320 321 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank, 322 enum aca_smu_type type, void *data) 323 { 324 int ret; 325 326 ret = aca_bank_parser(handle, bank, type); 327 if (ret) 328 return ret; 329 330 return 0; 331 } 332 333 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank, 334 enum aca_smu_type type, bank_handler_t handler, void *data) 335 { 336 struct aca_handle *handle; 337 int ret; 338 339 if (list_empty(&mgr->list)) 340 return 0; 341 342 list_for_each_entry(handle, &mgr->list, node) { 343 if (!aca_bank_is_valid(handle, bank, type)) 344 continue; 345 346 ret = handler(handle, bank, type, data); 347 if (ret) 348 return ret; 349 } 350 351 return 0; 352 } 353 354 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks, 355 enum aca_smu_type type, bank_handler_t handler, void *data) 356 { 357 struct aca_bank_node *node; 358 struct aca_bank *bank; 359 int ret; 360 361 if (!mgr || !banks) 362 return -EINVAL; 363 364 /* pre check to avoid unnecessary operations */ 365 if (list_empty(&mgr->list) || list_empty(&banks->list)) 366 return 0; 367 368 list_for_each_entry(node, &banks->list, node) { 369 bank = &node->bank; 370 371 ret = aca_dispatch_bank(mgr, bank, type, handler, data); 372 if (ret) 373 return ret; 374 } 375 376 return 0; 377 } 378 379 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type) 380 { 381 struct amdgpu_aca *aca = &adev->aca; 382 bool ret = true; 383 384 /* 385 * Because the UE Valid MCA count will only be cleared after reset, 386 * in order to avoid repeated counting of the error count, 387 * the aca bank is only updated once during the gpu recovery stage. 388 */ 389 if (type == ACA_SMU_TYPE_UE) { 390 if (amdgpu_ras_intr_triggered()) 391 ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0; 392 else 393 atomic_set(&aca->ue_update_flag, 0); 394 } 395 396 return ret; 397 } 398 399 static void aca_banks_generate_cper(struct amdgpu_device *adev, 400 enum aca_smu_type type, 401 struct aca_banks *banks, 402 int count) 403 { 404 struct aca_bank_node *node; 405 struct aca_bank *bank; 406 int r; 407 408 if (!adev->cper.enabled) 409 return; 410 411 if (!banks || !count) { 412 dev_warn(adev->dev, "fail to generate cper records\n"); 413 return; 414 } 415 416 /* UEs must be encoded into separate CPER entries */ 417 if (type == ACA_SMU_TYPE_UE) { 418 struct aca_banks de_banks; 419 420 aca_banks_init(&de_banks); 421 list_for_each_entry(node, &banks->list, node) { 422 bank = &node->bank; 423 if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) { 424 r = aca_banks_add_bank(&de_banks, bank); 425 if (r) 426 dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r); 427 } else { 428 if (amdgpu_cper_generate_ue_record(adev, bank)) 429 dev_warn(adev->dev, "fail to generate ue cper records\n"); 430 } 431 } 432 433 if (!list_empty(&de_banks.list)) { 434 if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks)) 435 dev_warn(adev->dev, "fail to generate de cper records\n"); 436 } 437 438 aca_banks_release(&de_banks); 439 } else { 440 /* 441 * SMU_TYPE_CE banks are combined into 1 CPER entries, 442 * they could be CEs or DEs or both 443 */ 444 if (amdgpu_cper_generate_ce_records(adev, banks, count)) 445 dev_warn(adev->dev, "fail to generate ce cper records\n"); 446 } 447 } 448 449 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type, 450 bank_handler_t handler, struct ras_query_context *qctx, void *data) 451 { 452 struct amdgpu_aca *aca = &adev->aca; 453 struct aca_banks banks; 454 u32 count = 0; 455 int ret; 456 457 if (list_empty(&aca->mgr.list)) 458 return 0; 459 460 if (!aca_bank_should_update(adev, type)) 461 return 0; 462 463 ret = aca_smu_get_valid_aca_count(adev, type, &count); 464 if (ret) 465 return ret; 466 467 if (!count) 468 return 0; 469 470 aca_banks_init(&banks); 471 472 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx); 473 if (ret) 474 goto err_release_banks; 475 476 if (list_empty(&banks.list)) { 477 ret = 0; 478 goto err_release_banks; 479 } 480 481 ret = aca_dispatch_banks(&aca->mgr, &banks, type, 482 handler, data); 483 if (ret) 484 goto err_release_banks; 485 486 aca_banks_generate_cper(adev, type, &banks, count); 487 488 err_release_banks: 489 aca_banks_release(&banks); 490 491 return ret; 492 } 493 494 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data) 495 { 496 struct aca_bank_info *info; 497 struct amdgpu_smuio_mcm_config_info mcm_info; 498 u64 count; 499 500 if (type >= ACA_ERROR_TYPE_COUNT) 501 return -EINVAL; 502 503 count = bank_error->count; 504 if (!count) 505 return 0; 506 507 info = &bank_error->info; 508 mcm_info.die_id = info->die_id; 509 mcm_info.socket_id = info->socket_id; 510 511 switch (type) { 512 case ACA_ERROR_TYPE_UE: 513 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count); 514 break; 515 case ACA_ERROR_TYPE_CE: 516 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count); 517 break; 518 case ACA_ERROR_TYPE_DEFERRED: 519 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count); 520 break; 521 default: 522 break; 523 } 524 525 return 0; 526 } 527 528 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data) 529 { 530 struct aca_error_cache *error_cache = &handle->error_cache; 531 struct aca_error *aerr = &error_cache->errors[type]; 532 struct aca_bank_error *bank_error, *tmp; 533 534 mutex_lock(&aerr->lock); 535 536 if (list_empty(&aerr->list)) 537 goto out_unlock; 538 539 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) { 540 aca_log_aca_error_data(bank_error, type, err_data); 541 aca_bank_error_remove(aerr, bank_error); 542 } 543 544 out_unlock: 545 mutex_unlock(&aerr->lock); 546 547 return 0; 548 } 549 550 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type, 551 struct ras_err_data *err_data, struct ras_query_context *qctx) 552 { 553 enum aca_smu_type smu_type; 554 int ret; 555 556 switch (type) { 557 case ACA_ERROR_TYPE_UE: 558 smu_type = ACA_SMU_TYPE_UE; 559 break; 560 case ACA_ERROR_TYPE_CE: 561 case ACA_ERROR_TYPE_DEFERRED: 562 smu_type = ACA_SMU_TYPE_CE; 563 break; 564 default: 565 return -EINVAL; 566 } 567 568 /* update aca bank to aca source error_cache first */ 569 ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL); 570 if (ret) 571 return ret; 572 573 /* DEs may contain in CEs or UEs */ 574 if (type != ACA_ERROR_TYPE_DEFERRED) 575 aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data); 576 577 return aca_log_aca_error(handle, type, err_data); 578 } 579 580 static bool aca_handle_is_valid(struct aca_handle *handle) 581 { 582 if (!handle->mask || !list_empty(&handle->node)) 583 return false; 584 585 return true; 586 } 587 588 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, 589 enum aca_error_type type, struct ras_err_data *err_data, 590 struct ras_query_context *qctx) 591 { 592 if (!handle || !err_data) 593 return -EINVAL; 594 595 if (aca_handle_is_valid(handle)) 596 return -EOPNOTSUPP; 597 598 if ((type < 0) || (!(BIT(type) & handle->mask))) 599 return 0; 600 601 return __aca_get_error_data(adev, handle, type, err_data, qctx); 602 } 603 604 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type) 605 { 606 mutex_init(&aerr->lock); 607 INIT_LIST_HEAD(&aerr->list); 608 aerr->type = type; 609 aerr->nr_errors = 0; 610 } 611 612 static void aca_init_error_cache(struct aca_handle *handle) 613 { 614 struct aca_error_cache *error_cache = &handle->error_cache; 615 int type; 616 617 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 618 aca_error_init(&error_cache->errors[type], type); 619 } 620 621 static void aca_error_fini(struct aca_error *aerr) 622 { 623 struct aca_bank_error *bank_error, *tmp; 624 625 mutex_lock(&aerr->lock); 626 if (list_empty(&aerr->list)) 627 goto out_unlock; 628 629 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) 630 aca_bank_error_remove(aerr, bank_error); 631 632 out_unlock: 633 mutex_destroy(&aerr->lock); 634 } 635 636 static void aca_fini_error_cache(struct aca_handle *handle) 637 { 638 struct aca_error_cache *error_cache = &handle->error_cache; 639 int type; 640 641 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 642 aca_error_fini(&error_cache->errors[type]); 643 } 644 645 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle, 646 const char *name, const struct aca_info *ras_info, void *data) 647 { 648 memset(handle, 0, sizeof(*handle)); 649 650 handle->adev = adev; 651 handle->mgr = mgr; 652 handle->name = name; 653 handle->hwip = ras_info->hwip; 654 handle->mask = ras_info->mask; 655 handle->bank_ops = ras_info->bank_ops; 656 handle->data = data; 657 aca_init_error_cache(handle); 658 659 INIT_LIST_HEAD(&handle->node); 660 list_add_tail(&handle->node, &mgr->list); 661 mgr->nr_handles++; 662 663 return 0; 664 } 665 666 static ssize_t aca_sysfs_read(struct device *dev, 667 struct device_attribute *attr, char *buf) 668 { 669 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr); 670 671 /* NOTE: the aca cache will be auto cleared once read, 672 * So the driver should unify the query entry point, forward request to ras query interface directly */ 673 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data); 674 } 675 676 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle) 677 { 678 struct device_attribute *aca_attr = &handle->aca_attr; 679 680 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name); 681 aca_attr->show = aca_sysfs_read; 682 aca_attr->attr.name = handle->attr_name; 683 aca_attr->attr.mode = S_IRUGO; 684 sysfs_attr_init(&aca_attr->attr); 685 686 return sysfs_add_file_to_group(&adev->dev->kobj, 687 &aca_attr->attr, 688 "ras"); 689 } 690 691 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle, 692 const char *name, const struct aca_info *ras_info, void *data) 693 { 694 struct amdgpu_aca *aca = &adev->aca; 695 int ret; 696 697 if (!amdgpu_aca_is_enabled(adev)) 698 return 0; 699 700 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data); 701 if (ret) 702 return ret; 703 704 return add_aca_sysfs(adev, handle); 705 } 706 707 static void remove_aca_handle(struct aca_handle *handle) 708 { 709 struct aca_handle_manager *mgr = handle->mgr; 710 711 aca_fini_error_cache(handle); 712 list_del(&handle->node); 713 mgr->nr_handles--; 714 } 715 716 static void remove_aca_sysfs(struct aca_handle *handle) 717 { 718 struct amdgpu_device *adev = handle->adev; 719 struct device_attribute *aca_attr = &handle->aca_attr; 720 721 if (adev->dev->kobj.sd) 722 sysfs_remove_file_from_group(&adev->dev->kobj, 723 &aca_attr->attr, 724 "ras"); 725 } 726 727 void amdgpu_aca_remove_handle(struct aca_handle *handle) 728 { 729 if (!handle || list_empty(&handle->node)) 730 return; 731 732 remove_aca_sysfs(handle); 733 remove_aca_handle(handle); 734 } 735 736 static int aca_manager_init(struct aca_handle_manager *mgr) 737 { 738 INIT_LIST_HEAD(&mgr->list); 739 mgr->nr_handles = 0; 740 741 return 0; 742 } 743 744 static void aca_manager_fini(struct aca_handle_manager *mgr) 745 { 746 struct aca_handle *handle, *tmp; 747 748 if (list_empty(&mgr->list)) 749 return; 750 751 list_for_each_entry_safe(handle, tmp, &mgr->list, node) 752 amdgpu_aca_remove_handle(handle); 753 } 754 755 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev) 756 { 757 return (adev->aca.is_enabled || 758 adev->debug_enable_ras_aca); 759 } 760 761 int amdgpu_aca_init(struct amdgpu_device *adev) 762 { 763 struct amdgpu_aca *aca = &adev->aca; 764 int ret; 765 766 atomic_set(&aca->ue_update_flag, 0); 767 768 ret = aca_manager_init(&aca->mgr); 769 if (ret) 770 return ret; 771 772 return 0; 773 } 774 775 void amdgpu_aca_fini(struct amdgpu_device *adev) 776 { 777 struct amdgpu_aca *aca = &adev->aca; 778 779 aca_manager_fini(&aca->mgr); 780 781 atomic_set(&aca->ue_update_flag, 0); 782 } 783 784 int amdgpu_aca_reset(struct amdgpu_device *adev) 785 { 786 struct amdgpu_aca *aca = &adev->aca; 787 788 atomic_set(&aca->ue_update_flag, 0); 789 790 return 0; 791 } 792 793 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs) 794 { 795 struct amdgpu_aca *aca = &adev->aca; 796 797 WARN_ON(aca->smu_funcs); 798 aca->smu_funcs = smu_funcs; 799 } 800 801 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info) 802 { 803 u64 ipid; 804 u32 instidhi, instidlo; 805 806 if (!bank || !info) 807 return -EINVAL; 808 809 ipid = bank->regs[ACA_REG_IDX_IPID]; 810 info->hwid = ACA_REG__IPID__HARDWAREID(ipid); 811 info->mcatype = ACA_REG__IPID__MCATYPE(ipid); 812 /* 813 * Unfied DieID Format: SAASS. A:AID, S:Socket. 814 * Unfied DieID[4:4] = InstanceId[0:0] 815 * Unfied DieID[0:3] = InstanceIdHi[0:3] 816 */ 817 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid); 818 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid); 819 info->die_id = ((instidhi >> 2) & 0x03); 820 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03); 821 822 return 0; 823 } 824 825 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank) 826 { 827 struct amdgpu_aca *aca = &adev->aca; 828 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 829 830 if (!smu_funcs || !smu_funcs->parse_error_code) 831 return -EOPNOTSUPP; 832 833 return smu_funcs->parse_error_code(adev, bank); 834 } 835 836 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size) 837 { 838 int i, error_code; 839 840 if (!bank || !err_codes) 841 return -EINVAL; 842 843 error_code = aca_bank_get_error_code(adev, bank); 844 if (error_code < 0) 845 return error_code; 846 847 for (i = 0; i < size; i++) { 848 if (err_codes[i] == error_code) 849 return 0; 850 } 851 852 return -EINVAL; 853 } 854 855 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en) 856 { 857 struct amdgpu_aca *aca = &adev->aca; 858 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 859 860 if (!smu_funcs || !smu_funcs->set_debug_mode) 861 return -EOPNOTSUPP; 862 863 return smu_funcs->set_debug_mode(adev, en); 864 } 865 866 #if defined(CONFIG_DEBUG_FS) 867 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val) 868 { 869 struct amdgpu_device *adev = (struct amdgpu_device *)data; 870 int ret; 871 872 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false); 873 if (ret) 874 return ret; 875 876 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off"); 877 878 return 0; 879 } 880 881 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx) 882 { 883 struct aca_bank_info info; 884 int i, ret; 885 886 ret = aca_bank_info_decode(bank, &info); 887 if (ret) 888 return; 889 890 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE"); 891 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n", 892 idx, info.socket_id, info.die_id, info.hwid, info.mcatype); 893 894 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 895 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]); 896 } 897 898 struct aca_dump_context { 899 struct seq_file *m; 900 int idx; 901 }; 902 903 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank, 904 enum aca_smu_type type, void *data) 905 { 906 struct aca_dump_context *ctx = (struct aca_dump_context *)data; 907 908 aca_dump_entry(ctx->m, bank, type, ctx->idx++); 909 910 return handler_aca_log_bank_error(handle, bank, type, NULL); 911 } 912 913 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type) 914 { 915 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 916 struct aca_dump_context context = { 917 .m = m, 918 .idx = 0, 919 }; 920 921 return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context); 922 } 923 924 static int aca_dump_ce_show(struct seq_file *m, void *unused) 925 { 926 return aca_dump_show(m, ACA_SMU_TYPE_CE); 927 } 928 929 static int aca_dump_ce_open(struct inode *inode, struct file *file) 930 { 931 return single_open(file, aca_dump_ce_show, inode->i_private); 932 } 933 934 static const struct file_operations aca_ce_dump_debug_fops = { 935 .owner = THIS_MODULE, 936 .open = aca_dump_ce_open, 937 .read = seq_read, 938 .llseek = seq_lseek, 939 .release = single_release, 940 }; 941 942 static int aca_dump_ue_show(struct seq_file *m, void *unused) 943 { 944 return aca_dump_show(m, ACA_SMU_TYPE_UE); 945 } 946 947 static int aca_dump_ue_open(struct inode *inode, struct file *file) 948 { 949 return single_open(file, aca_dump_ue_show, inode->i_private); 950 } 951 952 static const struct file_operations aca_ue_dump_debug_fops = { 953 .owner = THIS_MODULE, 954 .open = aca_dump_ue_open, 955 .read = seq_read, 956 .llseek = seq_lseek, 957 .release = single_release, 958 }; 959 960 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n"); 961 #endif 962 963 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root) 964 { 965 #if defined(CONFIG_DEBUG_FS) 966 if (!root) 967 return; 968 969 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops); 970 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops); 971 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops); 972 #endif 973 } 974