1 /* 2 * Copyright 2023 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 */ 23 24 #include <linux/list.h> 25 #include "amdgpu.h" 26 #include "amdgpu_aca.h" 27 #include "amdgpu_ras.h" 28 29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype} 30 31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data); 32 33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = { 34 ACA_BANK_HWID(SMU, 0x01, 0x01), 35 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00), 36 ACA_BANK_HWID(UMC, 0x96, 0x00), 37 }; 38 39 static void aca_banks_init(struct aca_banks *banks) 40 { 41 if (!banks) 42 return; 43 44 memset(banks, 0, sizeof(*banks)); 45 INIT_LIST_HEAD(&banks->list); 46 } 47 48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank) 49 { 50 struct aca_bank_node *node; 51 52 if (!bank) 53 return -EINVAL; 54 55 node = kvzalloc(sizeof(*node), GFP_KERNEL); 56 if (!node) 57 return -ENOMEM; 58 59 memcpy(&node->bank, bank, sizeof(*bank)); 60 61 INIT_LIST_HEAD(&node->node); 62 list_add_tail(&node->node, &banks->list); 63 64 banks->nr_banks++; 65 66 return 0; 67 } 68 69 static void aca_banks_release(struct aca_banks *banks) 70 { 71 struct aca_bank_node *node, *tmp; 72 73 if (list_empty(&banks->list)) 74 return; 75 76 list_for_each_entry_safe(node, tmp, &banks->list, node) { 77 list_del(&node->node); 78 kvfree(node); 79 } 80 } 81 82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count) 83 { 84 struct amdgpu_aca *aca = &adev->aca; 85 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 86 87 if (!count) 88 return -EINVAL; 89 90 if (!smu_funcs || !smu_funcs->get_valid_aca_count) 91 return -EOPNOTSUPP; 92 93 return smu_funcs->get_valid_aca_count(adev, type, count); 94 } 95 96 static struct aca_regs_dump { 97 const char *name; 98 int reg_idx; 99 } aca_regs[] = { 100 {"CONTROL", ACA_REG_IDX_CTL}, 101 {"STATUS", ACA_REG_IDX_STATUS}, 102 {"ADDR", ACA_REG_IDX_ADDR}, 103 {"MISC", ACA_REG_IDX_MISC0}, 104 {"CONFIG", ACA_REG_IDX_CONFIG}, 105 {"IPID", ACA_REG_IDX_IPID}, 106 {"SYND", ACA_REG_IDX_SYND}, 107 {"DESTAT", ACA_REG_IDX_DESTAT}, 108 {"DEADDR", ACA_REG_IDX_DEADDR}, 109 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK}, 110 }; 111 112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank, 113 struct ras_query_context *qctx) 114 { 115 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID; 116 int i; 117 118 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n"); 119 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */ 120 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 121 RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n", 122 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]); 123 124 if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS])) 125 RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n"); 126 } 127 128 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type, 129 int start, int count, 130 struct aca_banks *banks, struct ras_query_context *qctx) 131 { 132 struct amdgpu_aca *aca = &adev->aca; 133 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 134 struct aca_bank bank; 135 int i, max_count, ret; 136 137 if (!count) 138 return 0; 139 140 if (!smu_funcs || !smu_funcs->get_valid_aca_bank) 141 return -EOPNOTSUPP; 142 143 switch (type) { 144 case ACA_SMU_TYPE_UE: 145 max_count = smu_funcs->max_ue_bank_count; 146 break; 147 case ACA_SMU_TYPE_CE: 148 max_count = smu_funcs->max_ce_bank_count; 149 break; 150 default: 151 return -EINVAL; 152 } 153 154 if (start + count > max_count) 155 return -EINVAL; 156 157 count = min_t(int, count, max_count); 158 for (i = 0; i < count; i++) { 159 memset(&bank, 0, sizeof(bank)); 160 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank); 161 if (ret) 162 return ret; 163 164 bank.smu_err_type = type; 165 166 aca_smu_bank_dump(adev, i, count, &bank, qctx); 167 168 ret = aca_banks_add_bank(banks, &bank); 169 if (ret) 170 return ret; 171 } 172 173 return 0; 174 } 175 176 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type) 177 { 178 179 struct aca_hwip *hwip; 180 int hwid, mcatype; 181 u64 ipid; 182 183 if (!bank || type == ACA_HWIP_TYPE_UNKNOW) 184 return false; 185 186 hwip = &aca_hwid_mcatypes[type]; 187 if (!hwip->hwid) 188 return false; 189 190 ipid = bank->regs[ACA_REG_IDX_IPID]; 191 hwid = ACA_REG__IPID__HARDWAREID(ipid); 192 mcatype = ACA_REG__IPID__MCATYPE(ipid); 193 194 return hwip->hwid == hwid && hwip->mcatype == mcatype; 195 } 196 197 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 198 { 199 const struct aca_bank_ops *bank_ops = handle->bank_ops; 200 201 /* Parse all deferred errors with UMC aca handle */ 202 if (ACA_BANK_ERR_IS_DEFFERED(bank)) 203 return handle->hwip == ACA_HWIP_TYPE_UMC; 204 205 if (!aca_bank_hwip_is_matched(bank, handle->hwip)) 206 return false; 207 208 if (!bank_ops->aca_bank_is_valid) 209 return true; 210 211 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data); 212 } 213 214 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 215 { 216 struct aca_bank_error *bank_error; 217 218 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL); 219 if (!bank_error) 220 return NULL; 221 222 INIT_LIST_HEAD(&bank_error->node); 223 memcpy(&bank_error->info, info, sizeof(*info)); 224 225 mutex_lock(&aerr->lock); 226 list_add_tail(&bank_error->node, &aerr->list); 227 mutex_unlock(&aerr->lock); 228 229 return bank_error; 230 } 231 232 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 233 { 234 struct aca_bank_error *bank_error = NULL; 235 struct aca_bank_info *tmp_info; 236 bool found = false; 237 238 mutex_lock(&aerr->lock); 239 list_for_each_entry(bank_error, &aerr->list, node) { 240 tmp_info = &bank_error->info; 241 if (tmp_info->socket_id == info->socket_id && 242 tmp_info->die_id == info->die_id) { 243 found = true; 244 goto out_unlock; 245 } 246 } 247 248 out_unlock: 249 mutex_unlock(&aerr->lock); 250 251 return found ? bank_error : NULL; 252 } 253 254 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error) 255 { 256 if (!aerr || !bank_error) 257 return; 258 259 list_del(&bank_error->node); 260 aerr->nr_errors--; 261 262 kvfree(bank_error); 263 } 264 265 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 266 { 267 struct aca_bank_error *bank_error; 268 269 if (!aerr || !info) 270 return NULL; 271 272 bank_error = find_bank_error(aerr, info); 273 if (bank_error) 274 return bank_error; 275 276 return new_bank_error(aerr, info); 277 } 278 279 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info, 280 enum aca_error_type type, u64 count) 281 { 282 struct aca_error_cache *error_cache = &handle->error_cache; 283 struct aca_bank_error *bank_error; 284 struct aca_error *aerr; 285 286 if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT) 287 return -EINVAL; 288 289 if (!count) 290 return 0; 291 292 aerr = &error_cache->errors[type]; 293 bank_error = get_bank_error(aerr, info); 294 if (!bank_error) 295 return -ENOMEM; 296 297 bank_error->count += count; 298 299 return 0; 300 } 301 302 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 303 { 304 const struct aca_bank_ops *bank_ops = handle->bank_ops; 305 306 if (!bank) 307 return -EINVAL; 308 309 if (!bank_ops->aca_bank_parser) 310 return -EOPNOTSUPP; 311 312 return bank_ops->aca_bank_parser(handle, bank, type, 313 handle->data); 314 } 315 316 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank, 317 enum aca_smu_type type, void *data) 318 { 319 int ret; 320 321 ret = aca_bank_parser(handle, bank, type); 322 if (ret) 323 return ret; 324 325 return 0; 326 } 327 328 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank, 329 enum aca_smu_type type, bank_handler_t handler, void *data) 330 { 331 struct aca_handle *handle; 332 int ret; 333 334 if (list_empty(&mgr->list)) 335 return 0; 336 337 list_for_each_entry(handle, &mgr->list, node) { 338 if (!aca_bank_is_valid(handle, bank, type)) 339 continue; 340 341 ret = handler(handle, bank, type, data); 342 if (ret) 343 return ret; 344 } 345 346 return 0; 347 } 348 349 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks, 350 enum aca_smu_type type, bank_handler_t handler, void *data) 351 { 352 struct aca_bank_node *node; 353 struct aca_bank *bank; 354 int ret; 355 356 if (!mgr || !banks) 357 return -EINVAL; 358 359 /* pre check to avoid unnecessary operations */ 360 if (list_empty(&mgr->list) || list_empty(&banks->list)) 361 return 0; 362 363 list_for_each_entry(node, &banks->list, node) { 364 bank = &node->bank; 365 366 ret = aca_dispatch_bank(mgr, bank, type, handler, data); 367 if (ret) 368 return ret; 369 } 370 371 return 0; 372 } 373 374 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type) 375 { 376 struct amdgpu_aca *aca = &adev->aca; 377 bool ret = true; 378 379 /* 380 * Because the UE Valid MCA count will only be cleared after reset, 381 * in order to avoid repeated counting of the error count, 382 * the aca bank is only updated once during the gpu recovery stage. 383 */ 384 if (type == ACA_SMU_TYPE_UE) { 385 if (amdgpu_ras_intr_triggered()) 386 ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0; 387 else 388 atomic_set(&aca->ue_update_flag, 0); 389 } 390 391 return ret; 392 } 393 394 static void aca_banks_generate_cper(struct amdgpu_device *adev, 395 enum aca_smu_type type, 396 struct aca_banks *banks, 397 int count) 398 { 399 struct aca_bank_node *node; 400 struct aca_bank *bank; 401 int r; 402 403 if (!adev->cper.enabled) 404 return; 405 406 if (!banks || !count) { 407 dev_warn(adev->dev, "fail to generate cper records\n"); 408 return; 409 } 410 411 /* UEs must be encoded into separate CPER entries */ 412 if (type == ACA_SMU_TYPE_UE) { 413 struct aca_banks de_banks; 414 415 aca_banks_init(&de_banks); 416 list_for_each_entry(node, &banks->list, node) { 417 bank = &node->bank; 418 if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) { 419 r = aca_banks_add_bank(&de_banks, bank); 420 if (r) 421 dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r); 422 } else { 423 if (amdgpu_cper_generate_ue_record(adev, bank)) 424 dev_warn(adev->dev, "fail to generate ue cper records\n"); 425 } 426 } 427 428 if (!list_empty(&de_banks.list)) { 429 if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks)) 430 dev_warn(adev->dev, "fail to generate de cper records\n"); 431 } 432 433 aca_banks_release(&de_banks); 434 } else { 435 /* 436 * SMU_TYPE_CE banks are combined into 1 CPER entries, 437 * they could be CEs or DEs or both 438 */ 439 if (amdgpu_cper_generate_ce_records(adev, banks, count)) 440 dev_warn(adev->dev, "fail to generate ce cper records\n"); 441 } 442 } 443 444 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type, 445 bank_handler_t handler, struct ras_query_context *qctx, void *data) 446 { 447 struct amdgpu_aca *aca = &adev->aca; 448 struct aca_banks banks; 449 u32 count = 0; 450 int ret; 451 452 if (list_empty(&aca->mgr.list)) 453 return 0; 454 455 if (!aca_bank_should_update(adev, type)) 456 return 0; 457 458 ret = aca_smu_get_valid_aca_count(adev, type, &count); 459 if (ret) 460 return ret; 461 462 if (!count) 463 return 0; 464 465 aca_banks_init(&banks); 466 467 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx); 468 if (ret) 469 goto err_release_banks; 470 471 if (list_empty(&banks.list)) { 472 ret = 0; 473 goto err_release_banks; 474 } 475 476 ret = aca_dispatch_banks(&aca->mgr, &banks, type, 477 handler, data); 478 if (ret) 479 goto err_release_banks; 480 481 aca_banks_generate_cper(adev, type, &banks, count); 482 483 err_release_banks: 484 aca_banks_release(&banks); 485 486 return ret; 487 } 488 489 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data) 490 { 491 struct aca_bank_info *info; 492 struct amdgpu_smuio_mcm_config_info mcm_info; 493 u64 count; 494 495 if (type >= ACA_ERROR_TYPE_COUNT) 496 return -EINVAL; 497 498 count = bank_error->count; 499 if (!count) 500 return 0; 501 502 info = &bank_error->info; 503 mcm_info.die_id = info->die_id; 504 mcm_info.socket_id = info->socket_id; 505 506 switch (type) { 507 case ACA_ERROR_TYPE_UE: 508 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count); 509 break; 510 case ACA_ERROR_TYPE_CE: 511 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count); 512 break; 513 case ACA_ERROR_TYPE_DEFERRED: 514 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count); 515 break; 516 default: 517 break; 518 } 519 520 return 0; 521 } 522 523 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data) 524 { 525 struct aca_error_cache *error_cache = &handle->error_cache; 526 struct aca_error *aerr = &error_cache->errors[type]; 527 struct aca_bank_error *bank_error, *tmp; 528 529 mutex_lock(&aerr->lock); 530 531 if (list_empty(&aerr->list)) 532 goto out_unlock; 533 534 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) { 535 aca_log_aca_error_data(bank_error, type, err_data); 536 aca_bank_error_remove(aerr, bank_error); 537 } 538 539 out_unlock: 540 mutex_unlock(&aerr->lock); 541 542 return 0; 543 } 544 545 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type, 546 struct ras_err_data *err_data, struct ras_query_context *qctx) 547 { 548 enum aca_smu_type smu_type; 549 int ret; 550 551 switch (type) { 552 case ACA_ERROR_TYPE_UE: 553 smu_type = ACA_SMU_TYPE_UE; 554 break; 555 case ACA_ERROR_TYPE_CE: 556 case ACA_ERROR_TYPE_DEFERRED: 557 smu_type = ACA_SMU_TYPE_CE; 558 break; 559 default: 560 return -EINVAL; 561 } 562 563 /* update aca bank to aca source error_cache first */ 564 ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL); 565 if (ret) 566 return ret; 567 568 /* DEs may contain in CEs or UEs */ 569 if (type != ACA_ERROR_TYPE_DEFERRED) 570 aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data); 571 572 return aca_log_aca_error(handle, type, err_data); 573 } 574 575 static bool aca_handle_is_valid(struct aca_handle *handle) 576 { 577 if (!handle->mask || !list_empty(&handle->node)) 578 return false; 579 580 return true; 581 } 582 583 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, 584 enum aca_error_type type, struct ras_err_data *err_data, 585 struct ras_query_context *qctx) 586 { 587 if (!handle || !err_data) 588 return -EINVAL; 589 590 if (aca_handle_is_valid(handle)) 591 return -EOPNOTSUPP; 592 593 if ((type < 0) || (!(BIT(type) & handle->mask))) 594 return 0; 595 596 return __aca_get_error_data(adev, handle, type, err_data, qctx); 597 } 598 599 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type) 600 { 601 mutex_init(&aerr->lock); 602 INIT_LIST_HEAD(&aerr->list); 603 aerr->type = type; 604 aerr->nr_errors = 0; 605 } 606 607 static void aca_init_error_cache(struct aca_handle *handle) 608 { 609 struct aca_error_cache *error_cache = &handle->error_cache; 610 int type; 611 612 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 613 aca_error_init(&error_cache->errors[type], type); 614 } 615 616 static void aca_error_fini(struct aca_error *aerr) 617 { 618 struct aca_bank_error *bank_error, *tmp; 619 620 mutex_lock(&aerr->lock); 621 if (list_empty(&aerr->list)) 622 goto out_unlock; 623 624 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) 625 aca_bank_error_remove(aerr, bank_error); 626 627 out_unlock: 628 mutex_destroy(&aerr->lock); 629 } 630 631 static void aca_fini_error_cache(struct aca_handle *handle) 632 { 633 struct aca_error_cache *error_cache = &handle->error_cache; 634 int type; 635 636 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 637 aca_error_fini(&error_cache->errors[type]); 638 } 639 640 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle, 641 const char *name, const struct aca_info *ras_info, void *data) 642 { 643 memset(handle, 0, sizeof(*handle)); 644 645 handle->adev = adev; 646 handle->mgr = mgr; 647 handle->name = name; 648 handle->hwip = ras_info->hwip; 649 handle->mask = ras_info->mask; 650 handle->bank_ops = ras_info->bank_ops; 651 handle->data = data; 652 aca_init_error_cache(handle); 653 654 INIT_LIST_HEAD(&handle->node); 655 list_add_tail(&handle->node, &mgr->list); 656 mgr->nr_handles++; 657 658 return 0; 659 } 660 661 static ssize_t aca_sysfs_read(struct device *dev, 662 struct device_attribute *attr, char *buf) 663 { 664 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr); 665 666 /* NOTE: the aca cache will be auto cleared once read, 667 * So the driver should unify the query entry point, forward request to ras query interface directly */ 668 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data); 669 } 670 671 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle) 672 { 673 struct device_attribute *aca_attr = &handle->aca_attr; 674 675 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name); 676 aca_attr->show = aca_sysfs_read; 677 aca_attr->attr.name = handle->attr_name; 678 aca_attr->attr.mode = S_IRUGO; 679 sysfs_attr_init(&aca_attr->attr); 680 681 return sysfs_add_file_to_group(&adev->dev->kobj, 682 &aca_attr->attr, 683 "ras"); 684 } 685 686 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle, 687 const char *name, const struct aca_info *ras_info, void *data) 688 { 689 struct amdgpu_aca *aca = &adev->aca; 690 int ret; 691 692 if (!amdgpu_aca_is_enabled(adev)) 693 return 0; 694 695 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data); 696 if (ret) 697 return ret; 698 699 return add_aca_sysfs(adev, handle); 700 } 701 702 static void remove_aca_handle(struct aca_handle *handle) 703 { 704 struct aca_handle_manager *mgr = handle->mgr; 705 706 aca_fini_error_cache(handle); 707 list_del(&handle->node); 708 mgr->nr_handles--; 709 } 710 711 static void remove_aca_sysfs(struct aca_handle *handle) 712 { 713 struct amdgpu_device *adev = handle->adev; 714 struct device_attribute *aca_attr = &handle->aca_attr; 715 716 if (adev->dev->kobj.sd) 717 sysfs_remove_file_from_group(&adev->dev->kobj, 718 &aca_attr->attr, 719 "ras"); 720 } 721 722 void amdgpu_aca_remove_handle(struct aca_handle *handle) 723 { 724 if (!handle || list_empty(&handle->node)) 725 return; 726 727 remove_aca_sysfs(handle); 728 remove_aca_handle(handle); 729 } 730 731 static int aca_manager_init(struct aca_handle_manager *mgr) 732 { 733 INIT_LIST_HEAD(&mgr->list); 734 mgr->nr_handles = 0; 735 736 return 0; 737 } 738 739 static void aca_manager_fini(struct aca_handle_manager *mgr) 740 { 741 struct aca_handle *handle, *tmp; 742 743 if (list_empty(&mgr->list)) 744 return; 745 746 list_for_each_entry_safe(handle, tmp, &mgr->list, node) 747 amdgpu_aca_remove_handle(handle); 748 } 749 750 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev) 751 { 752 return (adev->aca.is_enabled || 753 adev->debug_enable_ras_aca); 754 } 755 756 int amdgpu_aca_init(struct amdgpu_device *adev) 757 { 758 struct amdgpu_aca *aca = &adev->aca; 759 int ret; 760 761 atomic_set(&aca->ue_update_flag, 0); 762 763 ret = aca_manager_init(&aca->mgr); 764 if (ret) 765 return ret; 766 767 return 0; 768 } 769 770 void amdgpu_aca_fini(struct amdgpu_device *adev) 771 { 772 struct amdgpu_aca *aca = &adev->aca; 773 774 aca_manager_fini(&aca->mgr); 775 776 atomic_set(&aca->ue_update_flag, 0); 777 } 778 779 int amdgpu_aca_reset(struct amdgpu_device *adev) 780 { 781 struct amdgpu_aca *aca = &adev->aca; 782 783 atomic_set(&aca->ue_update_flag, 0); 784 785 return 0; 786 } 787 788 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs) 789 { 790 struct amdgpu_aca *aca = &adev->aca; 791 792 WARN_ON(aca->smu_funcs); 793 aca->smu_funcs = smu_funcs; 794 } 795 796 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info) 797 { 798 u64 ipid; 799 u32 instidhi, instidlo; 800 801 if (!bank || !info) 802 return -EINVAL; 803 804 ipid = bank->regs[ACA_REG_IDX_IPID]; 805 info->hwid = ACA_REG__IPID__HARDWAREID(ipid); 806 info->mcatype = ACA_REG__IPID__MCATYPE(ipid); 807 /* 808 * Unfied DieID Format: SAASS. A:AID, S:Socket. 809 * Unfied DieID[4:4] = InstanceId[0:0] 810 * Unfied DieID[0:3] = InstanceIdHi[0:3] 811 */ 812 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid); 813 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid); 814 info->die_id = ((instidhi >> 2) & 0x03); 815 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03); 816 817 return 0; 818 } 819 820 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank) 821 { 822 struct amdgpu_aca *aca = &adev->aca; 823 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 824 825 if (!smu_funcs || !smu_funcs->parse_error_code) 826 return -EOPNOTSUPP; 827 828 return smu_funcs->parse_error_code(adev, bank); 829 } 830 831 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size) 832 { 833 int i, error_code; 834 835 if (!bank || !err_codes) 836 return -EINVAL; 837 838 error_code = aca_bank_get_error_code(adev, bank); 839 if (error_code < 0) 840 return error_code; 841 842 for (i = 0; i < size; i++) { 843 if (err_codes[i] == error_code) 844 return 0; 845 } 846 847 return -EINVAL; 848 } 849 850 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en) 851 { 852 struct amdgpu_aca *aca = &adev->aca; 853 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 854 855 if (!smu_funcs || !smu_funcs->set_debug_mode) 856 return -EOPNOTSUPP; 857 858 return smu_funcs->set_debug_mode(adev, en); 859 } 860 861 #if defined(CONFIG_DEBUG_FS) 862 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val) 863 { 864 struct amdgpu_device *adev = (struct amdgpu_device *)data; 865 int ret; 866 867 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false); 868 if (ret) 869 return ret; 870 871 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off"); 872 873 return 0; 874 } 875 876 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx) 877 { 878 struct aca_bank_info info; 879 int i, ret; 880 881 ret = aca_bank_info_decode(bank, &info); 882 if (ret) 883 return; 884 885 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE"); 886 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n", 887 idx, info.socket_id, info.die_id, info.hwid, info.mcatype); 888 889 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 890 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]); 891 } 892 893 struct aca_dump_context { 894 struct seq_file *m; 895 int idx; 896 }; 897 898 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank, 899 enum aca_smu_type type, void *data) 900 { 901 struct aca_dump_context *ctx = (struct aca_dump_context *)data; 902 903 aca_dump_entry(ctx->m, bank, type, ctx->idx++); 904 905 return handler_aca_log_bank_error(handle, bank, type, NULL); 906 } 907 908 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type) 909 { 910 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 911 struct aca_dump_context context = { 912 .m = m, 913 .idx = 0, 914 }; 915 916 return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context); 917 } 918 919 static int aca_dump_ce_show(struct seq_file *m, void *unused) 920 { 921 return aca_dump_show(m, ACA_SMU_TYPE_CE); 922 } 923 924 static int aca_dump_ce_open(struct inode *inode, struct file *file) 925 { 926 return single_open(file, aca_dump_ce_show, inode->i_private); 927 } 928 929 static const struct file_operations aca_ce_dump_debug_fops = { 930 .owner = THIS_MODULE, 931 .open = aca_dump_ce_open, 932 .read = seq_read, 933 .llseek = seq_lseek, 934 .release = single_release, 935 }; 936 937 static int aca_dump_ue_show(struct seq_file *m, void *unused) 938 { 939 return aca_dump_show(m, ACA_SMU_TYPE_UE); 940 } 941 942 static int aca_dump_ue_open(struct inode *inode, struct file *file) 943 { 944 return single_open(file, aca_dump_ue_show, inode->i_private); 945 } 946 947 static const struct file_operations aca_ue_dump_debug_fops = { 948 .owner = THIS_MODULE, 949 .open = aca_dump_ue_open, 950 .read = seq_read, 951 .llseek = seq_lseek, 952 .release = single_release, 953 }; 954 955 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n"); 956 #endif 957 958 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root) 959 { 960 #if defined(CONFIG_DEBUG_FS) 961 if (!root) 962 return; 963 964 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops); 965 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops); 966 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops); 967 #endif 968 } 969