1 /* 2 * Copyright 2023 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 */ 23 24 #include <linux/list.h> 25 #include "amdgpu.h" 26 #include "amdgpu_aca.h" 27 #include "amdgpu_ras.h" 28 29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype} 30 31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data); 32 33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = { 34 ACA_BANK_HWID(SMU, 0x01, 0x01), 35 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00), 36 ACA_BANK_HWID(UMC, 0x96, 0x00), 37 }; 38 39 static void aca_banks_init(struct aca_banks *banks) 40 { 41 if (!banks) 42 return; 43 44 memset(banks, 0, sizeof(*banks)); 45 INIT_LIST_HEAD(&banks->list); 46 } 47 48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank) 49 { 50 struct aca_bank_node *node; 51 52 if (!bank) 53 return -EINVAL; 54 55 node = kvzalloc(sizeof(*node), GFP_KERNEL); 56 if (!node) 57 return -ENOMEM; 58 59 memcpy(&node->bank, bank, sizeof(*bank)); 60 61 INIT_LIST_HEAD(&node->node); 62 list_add_tail(&node->node, &banks->list); 63 64 banks->nr_banks++; 65 66 return 0; 67 } 68 69 static void aca_banks_release(struct aca_banks *banks) 70 { 71 struct aca_bank_node *node, *tmp; 72 73 if (list_empty(&banks->list)) 74 return; 75 76 list_for_each_entry_safe(node, tmp, &banks->list, node) { 77 list_del(&node->node); 78 kvfree(node); 79 } 80 } 81 82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count) 83 { 84 struct amdgpu_aca *aca = &adev->aca; 85 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 86 87 if (!count) 88 return -EINVAL; 89 90 if (!smu_funcs || !smu_funcs->get_valid_aca_count) 91 return -EOPNOTSUPP; 92 93 return smu_funcs->get_valid_aca_count(adev, type, count); 94 } 95 96 static struct aca_regs_dump { 97 const char *name; 98 int reg_idx; 99 } aca_regs[] = { 100 {"CONTROL", ACA_REG_IDX_CTL}, 101 {"STATUS", ACA_REG_IDX_STATUS}, 102 {"ADDR", ACA_REG_IDX_ADDR}, 103 {"MISC", ACA_REG_IDX_MISC0}, 104 {"CONFIG", ACA_REG_IDX_CONFIG}, 105 {"IPID", ACA_REG_IDX_IPID}, 106 {"SYND", ACA_REG_IDX_SYND}, 107 {"DESTAT", ACA_REG_IDX_DESTAT}, 108 {"DEADDR", ACA_REG_IDX_DEADDR}, 109 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK}, 110 }; 111 112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank, 113 struct ras_query_context *qctx) 114 { 115 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID; 116 int i; 117 118 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n"); 119 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */ 120 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 121 RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n", 122 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]); 123 } 124 125 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type, 126 int start, int count, 127 struct aca_banks *banks, struct ras_query_context *qctx) 128 { 129 struct amdgpu_aca *aca = &adev->aca; 130 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 131 struct aca_bank bank; 132 int i, max_count, ret; 133 134 if (!count) 135 return 0; 136 137 if (!smu_funcs || !smu_funcs->get_valid_aca_bank) 138 return -EOPNOTSUPP; 139 140 switch (type) { 141 case ACA_SMU_TYPE_UE: 142 max_count = smu_funcs->max_ue_bank_count; 143 break; 144 case ACA_SMU_TYPE_CE: 145 max_count = smu_funcs->max_ce_bank_count; 146 break; 147 default: 148 return -EINVAL; 149 } 150 151 if (start + count > max_count) 152 return -EINVAL; 153 154 count = min_t(int, count, max_count); 155 for (i = 0; i < count; i++) { 156 memset(&bank, 0, sizeof(bank)); 157 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank); 158 if (ret) 159 return ret; 160 161 bank.smu_err_type = type; 162 163 aca_smu_bank_dump(adev, i, count, &bank, qctx); 164 165 ret = aca_banks_add_bank(banks, &bank); 166 if (ret) 167 return ret; 168 } 169 170 return 0; 171 } 172 173 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type) 174 { 175 176 struct aca_hwip *hwip; 177 int hwid, mcatype; 178 u64 ipid; 179 180 if (!bank || type == ACA_HWIP_TYPE_UNKNOW) 181 return false; 182 183 hwip = &aca_hwid_mcatypes[type]; 184 if (!hwip->hwid) 185 return false; 186 187 ipid = bank->regs[ACA_REG_IDX_IPID]; 188 hwid = ACA_REG__IPID__HARDWAREID(ipid); 189 mcatype = ACA_REG__IPID__MCATYPE(ipid); 190 191 return hwip->hwid == hwid && hwip->mcatype == mcatype; 192 } 193 194 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 195 { 196 const struct aca_bank_ops *bank_ops = handle->bank_ops; 197 198 /* Parse all deferred errors with UMC aca handle */ 199 if (ACA_BANK_ERR_IS_DEFFERED(bank)) 200 return handle->hwip == ACA_HWIP_TYPE_UMC; 201 202 if (!aca_bank_hwip_is_matched(bank, handle->hwip)) 203 return false; 204 205 if (!bank_ops->aca_bank_is_valid) 206 return true; 207 208 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data); 209 } 210 211 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 212 { 213 struct aca_bank_error *bank_error; 214 215 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL); 216 if (!bank_error) 217 return NULL; 218 219 INIT_LIST_HEAD(&bank_error->node); 220 memcpy(&bank_error->info, info, sizeof(*info)); 221 222 mutex_lock(&aerr->lock); 223 list_add_tail(&bank_error->node, &aerr->list); 224 mutex_unlock(&aerr->lock); 225 226 return bank_error; 227 } 228 229 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 230 { 231 struct aca_bank_error *bank_error = NULL; 232 struct aca_bank_info *tmp_info; 233 bool found = false; 234 235 mutex_lock(&aerr->lock); 236 list_for_each_entry(bank_error, &aerr->list, node) { 237 tmp_info = &bank_error->info; 238 if (tmp_info->socket_id == info->socket_id && 239 tmp_info->die_id == info->die_id) { 240 found = true; 241 goto out_unlock; 242 } 243 } 244 245 out_unlock: 246 mutex_unlock(&aerr->lock); 247 248 return found ? bank_error : NULL; 249 } 250 251 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error) 252 { 253 if (!aerr || !bank_error) 254 return; 255 256 list_del(&bank_error->node); 257 aerr->nr_errors--; 258 259 kvfree(bank_error); 260 } 261 262 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info) 263 { 264 struct aca_bank_error *bank_error; 265 266 if (!aerr || !info) 267 return NULL; 268 269 bank_error = find_bank_error(aerr, info); 270 if (bank_error) 271 return bank_error; 272 273 return new_bank_error(aerr, info); 274 } 275 276 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info, 277 enum aca_error_type type, u64 count) 278 { 279 struct aca_error_cache *error_cache = &handle->error_cache; 280 struct aca_bank_error *bank_error; 281 struct aca_error *aerr; 282 283 if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT) 284 return -EINVAL; 285 286 if (!count) 287 return 0; 288 289 aerr = &error_cache->errors[type]; 290 bank_error = get_bank_error(aerr, info); 291 if (!bank_error) 292 return -ENOMEM; 293 294 bank_error->count += count; 295 296 return 0; 297 } 298 299 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type) 300 { 301 const struct aca_bank_ops *bank_ops = handle->bank_ops; 302 303 if (!bank) 304 return -EINVAL; 305 306 if (!bank_ops->aca_bank_parser) 307 return -EOPNOTSUPP; 308 309 return bank_ops->aca_bank_parser(handle, bank, type, 310 handle->data); 311 } 312 313 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank, 314 enum aca_smu_type type, void *data) 315 { 316 int ret; 317 318 ret = aca_bank_parser(handle, bank, type); 319 if (ret) 320 return ret; 321 322 return 0; 323 } 324 325 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank, 326 enum aca_smu_type type, bank_handler_t handler, void *data) 327 { 328 struct aca_handle *handle; 329 int ret; 330 331 if (list_empty(&mgr->list)) 332 return 0; 333 334 list_for_each_entry(handle, &mgr->list, node) { 335 if (!aca_bank_is_valid(handle, bank, type)) 336 continue; 337 338 ret = handler(handle, bank, type, data); 339 if (ret) 340 return ret; 341 } 342 343 return 0; 344 } 345 346 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks, 347 enum aca_smu_type type, bank_handler_t handler, void *data) 348 { 349 struct aca_bank_node *node; 350 struct aca_bank *bank; 351 int ret; 352 353 if (!mgr || !banks) 354 return -EINVAL; 355 356 /* pre check to avoid unnecessary operations */ 357 if (list_empty(&mgr->list) || list_empty(&banks->list)) 358 return 0; 359 360 list_for_each_entry(node, &banks->list, node) { 361 bank = &node->bank; 362 363 ret = aca_dispatch_bank(mgr, bank, type, handler, data); 364 if (ret) 365 return ret; 366 } 367 368 return 0; 369 } 370 371 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type) 372 { 373 struct amdgpu_aca *aca = &adev->aca; 374 bool ret = true; 375 376 /* 377 * Because the UE Valid MCA count will only be cleared after reset, 378 * in order to avoid repeated counting of the error count, 379 * the aca bank is only updated once during the gpu recovery stage. 380 */ 381 if (type == ACA_SMU_TYPE_UE) { 382 if (amdgpu_ras_intr_triggered()) 383 ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0; 384 else 385 atomic_set(&aca->ue_update_flag, 0); 386 } 387 388 return ret; 389 } 390 391 static void aca_banks_generate_cper(struct amdgpu_device *adev, 392 enum aca_smu_type type, 393 struct aca_banks *banks, 394 int count) 395 { 396 struct aca_bank_node *node; 397 struct aca_bank *bank; 398 int r; 399 400 if (!adev->cper.enabled) 401 return; 402 403 if (!banks || !count) { 404 dev_warn(adev->dev, "fail to generate cper records\n"); 405 return; 406 } 407 408 /* UEs must be encoded into separate CPER entries */ 409 if (type == ACA_SMU_TYPE_UE) { 410 struct aca_banks de_banks; 411 412 aca_banks_init(&de_banks); 413 list_for_each_entry(node, &banks->list, node) { 414 bank = &node->bank; 415 if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) { 416 r = aca_banks_add_bank(&de_banks, bank); 417 if (r) 418 dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r); 419 } else { 420 if (amdgpu_cper_generate_ue_record(adev, bank)) 421 dev_warn(adev->dev, "fail to generate ue cper records\n"); 422 } 423 } 424 425 if (!list_empty(&de_banks.list)) { 426 if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks)) 427 dev_warn(adev->dev, "fail to generate de cper records\n"); 428 } 429 430 aca_banks_release(&de_banks); 431 } else { 432 /* 433 * SMU_TYPE_CE banks are combined into 1 CPER entries, 434 * they could be CEs or DEs or both 435 */ 436 if (amdgpu_cper_generate_ce_records(adev, banks, count)) 437 dev_warn(adev->dev, "fail to generate ce cper records\n"); 438 } 439 } 440 441 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type, 442 bank_handler_t handler, struct ras_query_context *qctx, void *data) 443 { 444 struct amdgpu_aca *aca = &adev->aca; 445 struct aca_banks banks; 446 u32 count = 0; 447 int ret; 448 449 if (list_empty(&aca->mgr.list)) 450 return 0; 451 452 if (!aca_bank_should_update(adev, type)) 453 return 0; 454 455 ret = aca_smu_get_valid_aca_count(adev, type, &count); 456 if (ret) 457 return ret; 458 459 if (!count) 460 return 0; 461 462 aca_banks_init(&banks); 463 464 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx); 465 if (ret) 466 goto err_release_banks; 467 468 if (list_empty(&banks.list)) { 469 ret = 0; 470 goto err_release_banks; 471 } 472 473 ret = aca_dispatch_banks(&aca->mgr, &banks, type, 474 handler, data); 475 if (ret) 476 goto err_release_banks; 477 478 aca_banks_generate_cper(adev, type, &banks, count); 479 480 err_release_banks: 481 aca_banks_release(&banks); 482 483 return ret; 484 } 485 486 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data) 487 { 488 struct aca_bank_info *info; 489 struct amdgpu_smuio_mcm_config_info mcm_info; 490 u64 count; 491 492 if (type >= ACA_ERROR_TYPE_COUNT) 493 return -EINVAL; 494 495 count = bank_error->count; 496 if (!count) 497 return 0; 498 499 info = &bank_error->info; 500 mcm_info.die_id = info->die_id; 501 mcm_info.socket_id = info->socket_id; 502 503 switch (type) { 504 case ACA_ERROR_TYPE_UE: 505 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count); 506 break; 507 case ACA_ERROR_TYPE_CE: 508 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count); 509 break; 510 case ACA_ERROR_TYPE_DEFERRED: 511 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count); 512 break; 513 default: 514 break; 515 } 516 517 return 0; 518 } 519 520 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data) 521 { 522 struct aca_error_cache *error_cache = &handle->error_cache; 523 struct aca_error *aerr = &error_cache->errors[type]; 524 struct aca_bank_error *bank_error, *tmp; 525 526 mutex_lock(&aerr->lock); 527 528 if (list_empty(&aerr->list)) 529 goto out_unlock; 530 531 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) { 532 aca_log_aca_error_data(bank_error, type, err_data); 533 aca_bank_error_remove(aerr, bank_error); 534 } 535 536 out_unlock: 537 mutex_unlock(&aerr->lock); 538 539 return 0; 540 } 541 542 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type, 543 struct ras_err_data *err_data, struct ras_query_context *qctx) 544 { 545 enum aca_smu_type smu_type; 546 int ret; 547 548 switch (type) { 549 case ACA_ERROR_TYPE_UE: 550 smu_type = ACA_SMU_TYPE_UE; 551 break; 552 case ACA_ERROR_TYPE_CE: 553 case ACA_ERROR_TYPE_DEFERRED: 554 smu_type = ACA_SMU_TYPE_CE; 555 break; 556 default: 557 return -EINVAL; 558 } 559 560 /* update aca bank to aca source error_cache first */ 561 ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL); 562 if (ret) 563 return ret; 564 565 /* DEs may contain in CEs or UEs */ 566 if (type != ACA_ERROR_TYPE_DEFERRED) 567 aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data); 568 569 return aca_log_aca_error(handle, type, err_data); 570 } 571 572 static bool aca_handle_is_valid(struct aca_handle *handle) 573 { 574 if (!handle->mask || !list_empty(&handle->node)) 575 return false; 576 577 return true; 578 } 579 580 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, 581 enum aca_error_type type, struct ras_err_data *err_data, 582 struct ras_query_context *qctx) 583 { 584 if (!handle || !err_data) 585 return -EINVAL; 586 587 if (aca_handle_is_valid(handle)) 588 return -EOPNOTSUPP; 589 590 if ((type < 0) || (!(BIT(type) & handle->mask))) 591 return 0; 592 593 return __aca_get_error_data(adev, handle, type, err_data, qctx); 594 } 595 596 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type) 597 { 598 mutex_init(&aerr->lock); 599 INIT_LIST_HEAD(&aerr->list); 600 aerr->type = type; 601 aerr->nr_errors = 0; 602 } 603 604 static void aca_init_error_cache(struct aca_handle *handle) 605 { 606 struct aca_error_cache *error_cache = &handle->error_cache; 607 int type; 608 609 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 610 aca_error_init(&error_cache->errors[type], type); 611 } 612 613 static void aca_error_fini(struct aca_error *aerr) 614 { 615 struct aca_bank_error *bank_error, *tmp; 616 617 mutex_lock(&aerr->lock); 618 if (list_empty(&aerr->list)) 619 goto out_unlock; 620 621 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) 622 aca_bank_error_remove(aerr, bank_error); 623 624 out_unlock: 625 mutex_destroy(&aerr->lock); 626 } 627 628 static void aca_fini_error_cache(struct aca_handle *handle) 629 { 630 struct aca_error_cache *error_cache = &handle->error_cache; 631 int type; 632 633 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++) 634 aca_error_fini(&error_cache->errors[type]); 635 } 636 637 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle, 638 const char *name, const struct aca_info *ras_info, void *data) 639 { 640 memset(handle, 0, sizeof(*handle)); 641 642 handle->adev = adev; 643 handle->mgr = mgr; 644 handle->name = name; 645 handle->hwip = ras_info->hwip; 646 handle->mask = ras_info->mask; 647 handle->bank_ops = ras_info->bank_ops; 648 handle->data = data; 649 aca_init_error_cache(handle); 650 651 INIT_LIST_HEAD(&handle->node); 652 list_add_tail(&handle->node, &mgr->list); 653 mgr->nr_handles++; 654 655 return 0; 656 } 657 658 static ssize_t aca_sysfs_read(struct device *dev, 659 struct device_attribute *attr, char *buf) 660 { 661 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr); 662 663 /* NOTE: the aca cache will be auto cleared once read, 664 * So the driver should unify the query entry point, forward request to ras query interface directly */ 665 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data); 666 } 667 668 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle) 669 { 670 struct device_attribute *aca_attr = &handle->aca_attr; 671 672 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name); 673 aca_attr->show = aca_sysfs_read; 674 aca_attr->attr.name = handle->attr_name; 675 aca_attr->attr.mode = S_IRUGO; 676 sysfs_attr_init(&aca_attr->attr); 677 678 return sysfs_add_file_to_group(&adev->dev->kobj, 679 &aca_attr->attr, 680 "ras"); 681 } 682 683 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle, 684 const char *name, const struct aca_info *ras_info, void *data) 685 { 686 struct amdgpu_aca *aca = &adev->aca; 687 int ret; 688 689 if (!amdgpu_aca_is_enabled(adev)) 690 return 0; 691 692 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data); 693 if (ret) 694 return ret; 695 696 return add_aca_sysfs(adev, handle); 697 } 698 699 static void remove_aca_handle(struct aca_handle *handle) 700 { 701 struct aca_handle_manager *mgr = handle->mgr; 702 703 aca_fini_error_cache(handle); 704 list_del(&handle->node); 705 mgr->nr_handles--; 706 } 707 708 static void remove_aca_sysfs(struct aca_handle *handle) 709 { 710 struct amdgpu_device *adev = handle->adev; 711 struct device_attribute *aca_attr = &handle->aca_attr; 712 713 if (adev->dev->kobj.sd) 714 sysfs_remove_file_from_group(&adev->dev->kobj, 715 &aca_attr->attr, 716 "ras"); 717 } 718 719 void amdgpu_aca_remove_handle(struct aca_handle *handle) 720 { 721 if (!handle || list_empty(&handle->node)) 722 return; 723 724 remove_aca_sysfs(handle); 725 remove_aca_handle(handle); 726 } 727 728 static int aca_manager_init(struct aca_handle_manager *mgr) 729 { 730 INIT_LIST_HEAD(&mgr->list); 731 mgr->nr_handles = 0; 732 733 return 0; 734 } 735 736 static void aca_manager_fini(struct aca_handle_manager *mgr) 737 { 738 struct aca_handle *handle, *tmp; 739 740 if (list_empty(&mgr->list)) 741 return; 742 743 list_for_each_entry_safe(handle, tmp, &mgr->list, node) 744 amdgpu_aca_remove_handle(handle); 745 } 746 747 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev) 748 { 749 return (adev->aca.is_enabled || 750 adev->debug_enable_ras_aca); 751 } 752 753 int amdgpu_aca_init(struct amdgpu_device *adev) 754 { 755 struct amdgpu_aca *aca = &adev->aca; 756 int ret; 757 758 atomic_set(&aca->ue_update_flag, 0); 759 760 ret = aca_manager_init(&aca->mgr); 761 if (ret) 762 return ret; 763 764 return 0; 765 } 766 767 void amdgpu_aca_fini(struct amdgpu_device *adev) 768 { 769 struct amdgpu_aca *aca = &adev->aca; 770 771 aca_manager_fini(&aca->mgr); 772 773 atomic_set(&aca->ue_update_flag, 0); 774 } 775 776 int amdgpu_aca_reset(struct amdgpu_device *adev) 777 { 778 struct amdgpu_aca *aca = &adev->aca; 779 780 atomic_set(&aca->ue_update_flag, 0); 781 782 return 0; 783 } 784 785 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs) 786 { 787 struct amdgpu_aca *aca = &adev->aca; 788 789 WARN_ON(aca->smu_funcs); 790 aca->smu_funcs = smu_funcs; 791 } 792 793 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info) 794 { 795 u64 ipid; 796 u32 instidhi, instidlo; 797 798 if (!bank || !info) 799 return -EINVAL; 800 801 ipid = bank->regs[ACA_REG_IDX_IPID]; 802 info->hwid = ACA_REG__IPID__HARDWAREID(ipid); 803 info->mcatype = ACA_REG__IPID__MCATYPE(ipid); 804 /* 805 * Unfied DieID Format: SAASS. A:AID, S:Socket. 806 * Unfied DieID[4:4] = InstanceId[0:0] 807 * Unfied DieID[0:3] = InstanceIdHi[0:3] 808 */ 809 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid); 810 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid); 811 info->die_id = ((instidhi >> 2) & 0x03); 812 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03); 813 814 return 0; 815 } 816 817 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank) 818 { 819 struct amdgpu_aca *aca = &adev->aca; 820 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 821 822 if (!smu_funcs || !smu_funcs->parse_error_code) 823 return -EOPNOTSUPP; 824 825 return smu_funcs->parse_error_code(adev, bank); 826 } 827 828 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size) 829 { 830 int i, error_code; 831 832 if (!bank || !err_codes) 833 return -EINVAL; 834 835 error_code = aca_bank_get_error_code(adev, bank); 836 if (error_code < 0) 837 return error_code; 838 839 for (i = 0; i < size; i++) { 840 if (err_codes[i] == error_code) 841 return 0; 842 } 843 844 return -EINVAL; 845 } 846 847 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en) 848 { 849 struct amdgpu_aca *aca = &adev->aca; 850 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs; 851 852 if (!smu_funcs || !smu_funcs->set_debug_mode) 853 return -EOPNOTSUPP; 854 855 return smu_funcs->set_debug_mode(adev, en); 856 } 857 858 #if defined(CONFIG_DEBUG_FS) 859 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val) 860 { 861 struct amdgpu_device *adev = (struct amdgpu_device *)data; 862 int ret; 863 864 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false); 865 if (ret) 866 return ret; 867 868 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off"); 869 870 return 0; 871 } 872 873 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx) 874 { 875 struct aca_bank_info info; 876 int i, ret; 877 878 ret = aca_bank_info_decode(bank, &info); 879 if (ret) 880 return; 881 882 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE"); 883 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n", 884 idx, info.socket_id, info.die_id, info.hwid, info.mcatype); 885 886 for (i = 0; i < ARRAY_SIZE(aca_regs); i++) 887 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]); 888 } 889 890 struct aca_dump_context { 891 struct seq_file *m; 892 int idx; 893 }; 894 895 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank, 896 enum aca_smu_type type, void *data) 897 { 898 struct aca_dump_context *ctx = (struct aca_dump_context *)data; 899 900 aca_dump_entry(ctx->m, bank, type, ctx->idx++); 901 902 return handler_aca_log_bank_error(handle, bank, type, NULL); 903 } 904 905 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type) 906 { 907 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 908 struct aca_dump_context context = { 909 .m = m, 910 .idx = 0, 911 }; 912 913 return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context); 914 } 915 916 static int aca_dump_ce_show(struct seq_file *m, void *unused) 917 { 918 return aca_dump_show(m, ACA_SMU_TYPE_CE); 919 } 920 921 static int aca_dump_ce_open(struct inode *inode, struct file *file) 922 { 923 return single_open(file, aca_dump_ce_show, inode->i_private); 924 } 925 926 static const struct file_operations aca_ce_dump_debug_fops = { 927 .owner = THIS_MODULE, 928 .open = aca_dump_ce_open, 929 .read = seq_read, 930 .llseek = seq_lseek, 931 .release = single_release, 932 }; 933 934 static int aca_dump_ue_show(struct seq_file *m, void *unused) 935 { 936 return aca_dump_show(m, ACA_SMU_TYPE_UE); 937 } 938 939 static int aca_dump_ue_open(struct inode *inode, struct file *file) 940 { 941 return single_open(file, aca_dump_ue_show, inode->i_private); 942 } 943 944 static const struct file_operations aca_ue_dump_debug_fops = { 945 .owner = THIS_MODULE, 946 .open = aca_dump_ue_open, 947 .read = seq_read, 948 .llseek = seq_lseek, 949 .release = single_release, 950 }; 951 952 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n"); 953 #endif 954 955 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root) 956 { 957 #if defined(CONFIG_DEBUG_FS) 958 if (!root) 959 return; 960 961 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops); 962 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops); 963 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops); 964 #endif 965 } 966