xref: /linux/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c (revision 25489a4f556414445d342951615178368ee45cde)
1 /*
2  * Copyright 2023 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  */
23 
24 #include <linux/list.h>
25 #include "amdgpu.h"
26 #include "amdgpu_aca.h"
27 #include "amdgpu_ras.h"
28 
29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30 
31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
32 
33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
34 	ACA_BANK_HWID(SMU,	0x01,	0x01),
35 	ACA_BANK_HWID(PCS_XGMI, 0x50,	0x00),
36 	ACA_BANK_HWID(UMC,	0x96,	0x00),
37 };
38 
39 static void aca_banks_init(struct aca_banks *banks)
40 {
41 	if (!banks)
42 		return;
43 
44 	memset(banks, 0, sizeof(*banks));
45 	INIT_LIST_HEAD(&banks->list);
46 }
47 
48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
49 {
50 	struct aca_bank_node *node;
51 
52 	if (!bank)
53 		return -EINVAL;
54 
55 	node = kvzalloc(sizeof(*node), GFP_KERNEL);
56 	if (!node)
57 		return -ENOMEM;
58 
59 	memcpy(&node->bank, bank, sizeof(*bank));
60 
61 	INIT_LIST_HEAD(&node->node);
62 	list_add_tail(&node->node, &banks->list);
63 
64 	banks->nr_banks++;
65 
66 	return 0;
67 }
68 
69 static void aca_banks_release(struct aca_banks *banks)
70 {
71 	struct aca_bank_node *node, *tmp;
72 
73 	if (list_empty(&banks->list))
74 		return;
75 
76 	list_for_each_entry_safe(node, tmp, &banks->list, node) {
77 		list_del(&node->node);
78 		kvfree(node);
79 	}
80 }
81 
82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
83 {
84 	struct amdgpu_aca *aca = &adev->aca;
85 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
86 
87 	if (!count)
88 		return -EINVAL;
89 
90 	if (!smu_funcs || !smu_funcs->get_valid_aca_count)
91 		return -EOPNOTSUPP;
92 
93 	return smu_funcs->get_valid_aca_count(adev, type, count);
94 }
95 
96 static struct aca_regs_dump {
97 	const char *name;
98 	int reg_idx;
99 } aca_regs[] = {
100 	{"CONTROL",		ACA_REG_IDX_CTL},
101 	{"STATUS",		ACA_REG_IDX_STATUS},
102 	{"ADDR",		ACA_REG_IDX_ADDR},
103 	{"MISC",		ACA_REG_IDX_MISC0},
104 	{"CONFIG",		ACA_REG_IDX_CONFIG},
105 	{"IPID",		ACA_REG_IDX_IPID},
106 	{"SYND",		ACA_REG_IDX_SYND},
107 	{"DESTAT",		ACA_REG_IDX_DESTAT},
108 	{"DEADDR",		ACA_REG_IDX_DEADDR},
109 	{"CONTROL_MASK",	ACA_REG_IDX_CTL_MASK},
110 };
111 
112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
113 			      struct ras_query_context *qctx)
114 {
115 	u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
116 	int i;
117 
118 	RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
119 	/* plus 1 for output format, e.g: ACA[08/08]: xxxx */
120 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
121 		RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
122 			      idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
123 
124 	if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS]))
125 		RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n");
126 }
127 
128 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
129 				       int start, int count,
130 				       struct aca_banks *banks, struct ras_query_context *qctx)
131 {
132 	struct amdgpu_aca *aca = &adev->aca;
133 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
134 	struct aca_bank bank;
135 	int i, max_count, ret;
136 
137 	if (!count)
138 		return 0;
139 
140 	if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
141 		return -EOPNOTSUPP;
142 
143 	switch (type) {
144 	case ACA_SMU_TYPE_UE:
145 		max_count = smu_funcs->max_ue_bank_count;
146 		break;
147 	case ACA_SMU_TYPE_CE:
148 		max_count = smu_funcs->max_ce_bank_count;
149 		break;
150 	default:
151 		return -EINVAL;
152 	}
153 
154 	if (start + count > max_count)
155 		return -EINVAL;
156 
157 	count = min_t(int, count, max_count);
158 	for (i = 0; i < count; i++) {
159 		memset(&bank, 0, sizeof(bank));
160 		ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
161 		if (ret)
162 			return ret;
163 
164 		bank.smu_err_type = type;
165 
166 		aca_smu_bank_dump(adev, i, count, &bank, qctx);
167 
168 		ret = aca_banks_add_bank(banks, &bank);
169 		if (ret)
170 			return ret;
171 	}
172 
173 	return 0;
174 }
175 
176 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
177 {
178 
179 	struct aca_hwip *hwip;
180 	int hwid, mcatype;
181 	u64 ipid;
182 
183 	if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
184 		return false;
185 
186 	hwip = &aca_hwid_mcatypes[type];
187 	if (!hwip->hwid)
188 		return false;
189 
190 	ipid = bank->regs[ACA_REG_IDX_IPID];
191 	hwid = ACA_REG__IPID__HARDWAREID(ipid);
192 	mcatype = ACA_REG__IPID__MCATYPE(ipid);
193 
194 	return hwip->hwid == hwid && hwip->mcatype == mcatype;
195 }
196 
197 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
198 {
199 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
200 
201 	/* Parse all deferred errors with UMC aca handle */
202 	if (ACA_BANK_ERR_IS_DEFFERED(bank))
203 		return handle->hwip == ACA_HWIP_TYPE_UMC;
204 
205 	if (!aca_bank_hwip_is_matched(bank, handle->hwip))
206 		return false;
207 
208 	if (!bank_ops->aca_bank_is_valid)
209 		return true;
210 
211 	return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
212 }
213 
214 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
215 {
216 	struct aca_bank_error *bank_error;
217 
218 	bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
219 	if (!bank_error)
220 		return NULL;
221 
222 	INIT_LIST_HEAD(&bank_error->node);
223 	memcpy(&bank_error->info, info, sizeof(*info));
224 
225 	mutex_lock(&aerr->lock);
226 	list_add_tail(&bank_error->node, &aerr->list);
227 	mutex_unlock(&aerr->lock);
228 
229 	return bank_error;
230 }
231 
232 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
233 {
234 	struct aca_bank_error *bank_error = NULL;
235 	struct aca_bank_info *tmp_info;
236 	bool found = false;
237 
238 	mutex_lock(&aerr->lock);
239 	list_for_each_entry(bank_error, &aerr->list, node) {
240 		tmp_info = &bank_error->info;
241 		if (tmp_info->socket_id == info->socket_id &&
242 		    tmp_info->die_id == info->die_id) {
243 			found = true;
244 			goto out_unlock;
245 		}
246 	}
247 
248 out_unlock:
249 	mutex_unlock(&aerr->lock);
250 
251 	return found ? bank_error : NULL;
252 }
253 
254 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
255 {
256 	if (!aerr || !bank_error)
257 		return;
258 
259 	list_del(&bank_error->node);
260 	aerr->nr_errors--;
261 
262 	kvfree(bank_error);
263 }
264 
265 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
266 {
267 	struct aca_bank_error *bank_error;
268 
269 	if (!aerr || !info)
270 		return NULL;
271 
272 	bank_error = find_bank_error(aerr, info);
273 	if (bank_error)
274 		return bank_error;
275 
276 	return new_bank_error(aerr, info);
277 }
278 
279 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
280 				   enum aca_error_type type, u64 count)
281 {
282 	struct aca_error_cache *error_cache = &handle->error_cache;
283 	struct aca_bank_error *bank_error;
284 	struct aca_error *aerr;
285 
286 	if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
287 		return -EINVAL;
288 
289 	if (!count)
290 		return 0;
291 
292 	aerr = &error_cache->errors[type];
293 	bank_error = get_bank_error(aerr, info);
294 	if (!bank_error)
295 		return -ENOMEM;
296 
297 	bank_error->count += count;
298 
299 	return 0;
300 }
301 
302 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
303 {
304 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
305 
306 	if (!bank)
307 		return -EINVAL;
308 
309 	if (!bank_ops->aca_bank_parser)
310 		return -EOPNOTSUPP;
311 
312 	return bank_ops->aca_bank_parser(handle, bank, type,
313 					 handle->data);
314 }
315 
316 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
317 				      enum aca_smu_type type, void *data)
318 {
319 	int ret;
320 
321 	ret = aca_bank_parser(handle, bank, type);
322 	if (ret)
323 		return ret;
324 
325 	return 0;
326 }
327 
328 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
329 			     enum aca_smu_type type, bank_handler_t handler, void *data)
330 {
331 	struct aca_handle *handle;
332 	int ret;
333 
334 	if (list_empty(&mgr->list))
335 		return 0;
336 
337 	list_for_each_entry(handle, &mgr->list, node) {
338 		if (!aca_bank_is_valid(handle, bank, type))
339 			continue;
340 
341 		ret = handler(handle, bank, type, data);
342 		if (ret)
343 			return ret;
344 	}
345 
346 	return 0;
347 }
348 
349 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
350 			      enum aca_smu_type type, bank_handler_t handler, void *data)
351 {
352 	struct aca_bank_node *node;
353 	struct aca_bank *bank;
354 	int ret;
355 
356 	if (!mgr || !banks)
357 		return -EINVAL;
358 
359 	/* pre check to avoid unnecessary operations */
360 	if (list_empty(&mgr->list) || list_empty(&banks->list))
361 		return 0;
362 
363 	list_for_each_entry(node, &banks->list, node) {
364 		bank = &node->bank;
365 
366 		ret = aca_dispatch_bank(mgr, bank, type, handler, data);
367 		if (ret)
368 			return ret;
369 	}
370 
371 	return 0;
372 }
373 
374 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
375 {
376 	struct amdgpu_aca *aca = &adev->aca;
377 	bool ret = true;
378 
379 	/*
380 	 * Because the UE Valid MCA count will only be cleared after reset,
381 	 * in order to avoid repeated counting of the error count,
382 	 * the aca bank is only updated once during the gpu recovery stage.
383 	 */
384 	if (type == ACA_SMU_TYPE_UE) {
385 		if (amdgpu_ras_intr_triggered())
386 			ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
387 		else
388 			atomic_set(&aca->ue_update_flag, 0);
389 	}
390 
391 	return ret;
392 }
393 
394 static void aca_banks_generate_cper(struct amdgpu_device *adev,
395 				    enum aca_smu_type type,
396 				    struct aca_banks *banks,
397 				    int count)
398 {
399 	struct aca_bank_node *node;
400 	struct aca_bank *bank;
401 	int r;
402 
403 	if (!adev->cper.enabled)
404 		return;
405 
406 	if (!banks || !count) {
407 		dev_warn(adev->dev, "fail to generate cper records\n");
408 		return;
409 	}
410 
411 	/* UEs must be encoded into separate CPER entries */
412 	if (type == ACA_SMU_TYPE_UE) {
413 		struct aca_banks de_banks;
414 
415 		aca_banks_init(&de_banks);
416 		list_for_each_entry(node, &banks->list, node) {
417 			bank = &node->bank;
418 			if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {
419 				r = aca_banks_add_bank(&de_banks, bank);
420 				if (r)
421 					dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);
422 			} else {
423 				if (amdgpu_cper_generate_ue_record(adev, bank))
424 					dev_warn(adev->dev, "fail to generate ue cper records\n");
425 			}
426 		}
427 
428 		if (!list_empty(&de_banks.list)) {
429 			if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))
430 				dev_warn(adev->dev, "fail to generate de cper records\n");
431 		}
432 
433 		aca_banks_release(&de_banks);
434 	} else {
435 		/*
436 		 * SMU_TYPE_CE banks are combined into 1 CPER entries,
437 		 * they could be CEs or DEs or both
438 		 */
439 		if (amdgpu_cper_generate_ce_records(adev, banks, count))
440 			dev_warn(adev->dev, "fail to generate ce cper records\n");
441 	}
442 }
443 
444 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
445 			    bank_handler_t handler, struct ras_query_context *qctx, void *data)
446 {
447 	struct amdgpu_aca *aca = &adev->aca;
448 	struct aca_banks banks;
449 	u32 count = 0;
450 	int ret;
451 
452 	if (list_empty(&aca->mgr.list))
453 		return 0;
454 
455 	if (!aca_bank_should_update(adev, type))
456 		return 0;
457 
458 	ret = aca_smu_get_valid_aca_count(adev, type, &count);
459 	if (ret)
460 		return ret;
461 
462 	if (!count)
463 		return 0;
464 
465 	aca_banks_init(&banks);
466 
467 	ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
468 	if (ret)
469 		goto err_release_banks;
470 
471 	if (list_empty(&banks.list)) {
472 		ret = 0;
473 		goto err_release_banks;
474 	}
475 
476 	ret = aca_dispatch_banks(&aca->mgr, &banks, type,
477 				 handler, data);
478 	if (ret)
479 		goto err_release_banks;
480 
481 	aca_banks_generate_cper(adev, type, &banks, count);
482 
483 err_release_banks:
484 	aca_banks_release(&banks);
485 
486 	return ret;
487 }
488 
489 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
490 {
491 	struct aca_bank_info *info;
492 	struct amdgpu_smuio_mcm_config_info mcm_info;
493 	u64 count;
494 
495 	if (type >= ACA_ERROR_TYPE_COUNT)
496 		return -EINVAL;
497 
498 	count = bank_error->count;
499 	if (!count)
500 		return 0;
501 
502 	info = &bank_error->info;
503 	mcm_info.die_id = info->die_id;
504 	mcm_info.socket_id = info->socket_id;
505 
506 	switch (type) {
507 	case ACA_ERROR_TYPE_UE:
508 		amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);
509 		break;
510 	case ACA_ERROR_TYPE_CE:
511 		amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);
512 		break;
513 	case ACA_ERROR_TYPE_DEFERRED:
514 		amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);
515 		break;
516 	default:
517 		break;
518 	}
519 
520 	return 0;
521 }
522 
523 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
524 {
525 	struct aca_error_cache *error_cache = &handle->error_cache;
526 	struct aca_error *aerr = &error_cache->errors[type];
527 	struct aca_bank_error *bank_error, *tmp;
528 
529 	mutex_lock(&aerr->lock);
530 
531 	if (list_empty(&aerr->list))
532 		goto out_unlock;
533 
534 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
535 		aca_log_aca_error_data(bank_error, type, err_data);
536 		aca_bank_error_remove(aerr, bank_error);
537 	}
538 
539 out_unlock:
540 	mutex_unlock(&aerr->lock);
541 
542 	return 0;
543 }
544 
545 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
546 				struct ras_err_data *err_data, struct ras_query_context *qctx)
547 {
548 	enum aca_smu_type smu_type;
549 	int ret;
550 
551 	switch (type) {
552 	case ACA_ERROR_TYPE_UE:
553 		smu_type = ACA_SMU_TYPE_UE;
554 		break;
555 	case ACA_ERROR_TYPE_CE:
556 	case ACA_ERROR_TYPE_DEFERRED:
557 		smu_type = ACA_SMU_TYPE_CE;
558 		break;
559 	default:
560 		return -EINVAL;
561 	}
562 
563 	/* update aca bank to aca source error_cache first */
564 	ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
565 	if (ret)
566 		return ret;
567 
568 	/* DEs may contain in CEs or UEs */
569 	if (type != ACA_ERROR_TYPE_DEFERRED)
570 		aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);
571 
572 	return aca_log_aca_error(handle, type, err_data);
573 }
574 
575 static bool aca_handle_is_valid(struct aca_handle *handle)
576 {
577 	if (!handle->mask || !list_empty(&handle->node))
578 		return false;
579 
580 	return true;
581 }
582 
583 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
584 			      enum aca_error_type type, struct ras_err_data *err_data,
585 			      struct ras_query_context *qctx)
586 {
587 	if (!handle || !err_data)
588 		return -EINVAL;
589 
590 	if (aca_handle_is_valid(handle))
591 		return -EOPNOTSUPP;
592 
593 	if ((type < 0) || (!(BIT(type) & handle->mask)))
594 		return  0;
595 
596 	return __aca_get_error_data(adev, handle, type, err_data, qctx);
597 }
598 
599 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
600 {
601 	mutex_init(&aerr->lock);
602 	INIT_LIST_HEAD(&aerr->list);
603 	aerr->type = type;
604 	aerr->nr_errors = 0;
605 }
606 
607 static void aca_init_error_cache(struct aca_handle *handle)
608 {
609 	struct aca_error_cache *error_cache = &handle->error_cache;
610 	int type;
611 
612 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
613 		aca_error_init(&error_cache->errors[type], type);
614 }
615 
616 static void aca_error_fini(struct aca_error *aerr)
617 {
618 	struct aca_bank_error *bank_error, *tmp;
619 
620 	mutex_lock(&aerr->lock);
621 	if (list_empty(&aerr->list))
622 		goto out_unlock;
623 
624 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
625 		aca_bank_error_remove(aerr, bank_error);
626 
627 out_unlock:
628 	mutex_destroy(&aerr->lock);
629 }
630 
631 static void aca_fini_error_cache(struct aca_handle *handle)
632 {
633 	struct aca_error_cache *error_cache = &handle->error_cache;
634 	int type;
635 
636 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
637 		aca_error_fini(&error_cache->errors[type]);
638 }
639 
640 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
641 			  const char *name, const struct aca_info *ras_info, void *data)
642 {
643 	memset(handle, 0, sizeof(*handle));
644 
645 	handle->adev = adev;
646 	handle->mgr = mgr;
647 	handle->name = name;
648 	handle->hwip = ras_info->hwip;
649 	handle->mask = ras_info->mask;
650 	handle->bank_ops = ras_info->bank_ops;
651 	handle->data = data;
652 	aca_init_error_cache(handle);
653 
654 	INIT_LIST_HEAD(&handle->node);
655 	list_add_tail(&handle->node, &mgr->list);
656 	mgr->nr_handles++;
657 
658 	return 0;
659 }
660 
661 static ssize_t aca_sysfs_read(struct device *dev,
662 			      struct device_attribute *attr, char *buf)
663 {
664 	struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
665 
666 	/* NOTE: the aca cache will be auto cleared once read,
667 	 * So the driver should unify the query entry point, forward request to ras query interface directly */
668 	return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
669 }
670 
671 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
672 {
673 	struct device_attribute *aca_attr = &handle->aca_attr;
674 
675 	snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
676 	aca_attr->show = aca_sysfs_read;
677 	aca_attr->attr.name = handle->attr_name;
678 	aca_attr->attr.mode = S_IRUGO;
679 	sysfs_attr_init(&aca_attr->attr);
680 
681 	return sysfs_add_file_to_group(&adev->dev->kobj,
682 				       &aca_attr->attr,
683 				       "ras");
684 }
685 
686 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
687 			  const char *name, const struct aca_info *ras_info, void *data)
688 {
689 	struct amdgpu_aca *aca = &adev->aca;
690 	int ret;
691 
692 	if (!amdgpu_aca_is_enabled(adev))
693 		return 0;
694 
695 	ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
696 	if (ret)
697 		return ret;
698 
699 	return add_aca_sysfs(adev, handle);
700 }
701 
702 static void remove_aca_handle(struct aca_handle *handle)
703 {
704 	struct aca_handle_manager *mgr = handle->mgr;
705 
706 	aca_fini_error_cache(handle);
707 	list_del(&handle->node);
708 	mgr->nr_handles--;
709 }
710 
711 static void remove_aca_sysfs(struct aca_handle *handle)
712 {
713 	struct amdgpu_device *adev = handle->adev;
714 	struct device_attribute *aca_attr = &handle->aca_attr;
715 
716 	if (adev->dev->kobj.sd)
717 		sysfs_remove_file_from_group(&adev->dev->kobj,
718 					     &aca_attr->attr,
719 					     "ras");
720 }
721 
722 void amdgpu_aca_remove_handle(struct aca_handle *handle)
723 {
724 	if (!handle || list_empty(&handle->node))
725 		return;
726 
727 	remove_aca_sysfs(handle);
728 	remove_aca_handle(handle);
729 }
730 
731 static int aca_manager_init(struct aca_handle_manager *mgr)
732 {
733 	INIT_LIST_HEAD(&mgr->list);
734 	mgr->nr_handles = 0;
735 
736 	return 0;
737 }
738 
739 static void aca_manager_fini(struct aca_handle_manager *mgr)
740 {
741 	struct aca_handle *handle, *tmp;
742 
743 	if (list_empty(&mgr->list))
744 		return;
745 
746 	list_for_each_entry_safe(handle, tmp, &mgr->list, node)
747 		amdgpu_aca_remove_handle(handle);
748 }
749 
750 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
751 {
752 	return (adev->aca.is_enabled ||
753 		adev->debug_enable_ras_aca);
754 }
755 
756 int amdgpu_aca_init(struct amdgpu_device *adev)
757 {
758 	struct amdgpu_aca *aca = &adev->aca;
759 	int ret;
760 
761 	atomic_set(&aca->ue_update_flag, 0);
762 
763 	ret = aca_manager_init(&aca->mgr);
764 	if (ret)
765 		return ret;
766 
767 	return 0;
768 }
769 
770 void amdgpu_aca_fini(struct amdgpu_device *adev)
771 {
772 	struct amdgpu_aca *aca = &adev->aca;
773 
774 	aca_manager_fini(&aca->mgr);
775 
776 	atomic_set(&aca->ue_update_flag, 0);
777 }
778 
779 int amdgpu_aca_reset(struct amdgpu_device *adev)
780 {
781 	struct amdgpu_aca *aca = &adev->aca;
782 
783 	atomic_set(&aca->ue_update_flag, 0);
784 
785 	return 0;
786 }
787 
788 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
789 {
790 	struct amdgpu_aca *aca = &adev->aca;
791 
792 	WARN_ON(aca->smu_funcs);
793 	aca->smu_funcs = smu_funcs;
794 }
795 
796 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
797 {
798 	u64 ipid;
799 	u32 instidhi, instidlo;
800 
801 	if (!bank || !info)
802 		return -EINVAL;
803 
804 	ipid = bank->regs[ACA_REG_IDX_IPID];
805 	info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
806 	info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
807 	/*
808 	 * Unfied DieID Format: SAASS. A:AID, S:Socket.
809 	 * Unfied DieID[4:4] = InstanceId[0:0]
810 	 * Unfied DieID[0:3] = InstanceIdHi[0:3]
811 	 */
812 	instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
813 	instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
814 	info->die_id = ((instidhi >> 2) & 0x03);
815 	info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
816 
817 	return 0;
818 }
819 
820 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
821 {
822 	struct amdgpu_aca *aca = &adev->aca;
823 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
824 
825 	if (!smu_funcs || !smu_funcs->parse_error_code)
826 		return -EOPNOTSUPP;
827 
828 	return smu_funcs->parse_error_code(adev, bank);
829 }
830 
831 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
832 {
833 	int i, error_code;
834 
835 	if (!bank || !err_codes)
836 		return -EINVAL;
837 
838 	error_code = aca_bank_get_error_code(adev, bank);
839 	if (error_code < 0)
840 		return error_code;
841 
842 	for (i = 0; i < size; i++) {
843 		if (err_codes[i] == error_code)
844 			return 0;
845 	}
846 
847 	return -EINVAL;
848 }
849 
850 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
851 {
852 	struct amdgpu_aca *aca = &adev->aca;
853 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
854 
855 	if (!smu_funcs || !smu_funcs->set_debug_mode)
856 		return -EOPNOTSUPP;
857 
858 	return smu_funcs->set_debug_mode(adev, en);
859 }
860 
861 #if defined(CONFIG_DEBUG_FS)
862 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
863 {
864 	struct amdgpu_device *adev = (struct amdgpu_device *)data;
865 	int ret;
866 
867 	ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
868 	if (ret)
869 		return ret;
870 
871 	dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
872 
873 	return 0;
874 }
875 
876 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
877 {
878 	struct aca_bank_info info;
879 	int i, ret;
880 
881 	ret = aca_bank_info_decode(bank, &info);
882 	if (ret)
883 		return;
884 
885 	seq_printf(m, "aca entry[%d].type: %s\n", idx, type ==  ACA_SMU_TYPE_UE ? "UE" : "CE");
886 	seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
887 		   idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
888 
889 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
890 		seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
891 }
892 
893 struct aca_dump_context {
894 	struct seq_file *m;
895 	int idx;
896 };
897 
898 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
899 				 enum aca_smu_type type, void *data)
900 {
901 	struct aca_dump_context *ctx = (struct aca_dump_context *)data;
902 
903 	aca_dump_entry(ctx->m, bank, type, ctx->idx++);
904 
905 	return handler_aca_log_bank_error(handle, bank, type, NULL);
906 }
907 
908 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
909 {
910 	struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
911 	struct aca_dump_context context = {
912 		.m = m,
913 		.idx = 0,
914 	};
915 
916 	return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
917 }
918 
919 static int aca_dump_ce_show(struct seq_file *m, void *unused)
920 {
921 	return aca_dump_show(m, ACA_SMU_TYPE_CE);
922 }
923 
924 static int aca_dump_ce_open(struct inode *inode, struct file *file)
925 {
926 	return single_open(file, aca_dump_ce_show, inode->i_private);
927 }
928 
929 static const struct file_operations aca_ce_dump_debug_fops = {
930 	.owner = THIS_MODULE,
931 	.open = aca_dump_ce_open,
932 	.read = seq_read,
933 	.llseek = seq_lseek,
934 	.release = single_release,
935 };
936 
937 static int aca_dump_ue_show(struct seq_file *m, void *unused)
938 {
939 	return aca_dump_show(m, ACA_SMU_TYPE_UE);
940 }
941 
942 static int aca_dump_ue_open(struct inode *inode, struct file *file)
943 {
944 	return single_open(file, aca_dump_ue_show, inode->i_private);
945 }
946 
947 static const struct file_operations aca_ue_dump_debug_fops = {
948 	.owner = THIS_MODULE,
949 	.open = aca_dump_ue_open,
950 	.read = seq_read,
951 	.llseek = seq_lseek,
952 	.release = single_release,
953 };
954 
955 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
956 #endif
957 
958 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
959 {
960 #if defined(CONFIG_DEBUG_FS)
961 	if (!root)
962 		return;
963 
964 	debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
965 	debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
966 	debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
967 #endif
968 }
969