xref: /linux/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c (revision 07fdad3a93756b872da7b53647715c48d0f4a2d0)
1 /*
2  * Copyright 2023 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  */
23 
24 #include <linux/list.h>
25 #include "amdgpu.h"
26 #include "amdgpu_aca.h"
27 #include "amdgpu_ras.h"
28 
29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30 
31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
32 
33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
34 	ACA_BANK_HWID(SMU,	0x01,	0x01),
35 	ACA_BANK_HWID(PCS_XGMI, 0x50,	0x00),
36 	ACA_BANK_HWID(UMC,	0x96,	0x00),
37 };
38 
39 static void aca_banks_init(struct aca_banks *banks)
40 {
41 	if (!banks)
42 		return;
43 
44 	memset(banks, 0, sizeof(*banks));
45 	INIT_LIST_HEAD(&banks->list);
46 }
47 
48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
49 {
50 	struct aca_bank_node *node;
51 
52 	if (!bank)
53 		return -EINVAL;
54 
55 	node = kvzalloc(sizeof(*node), GFP_KERNEL);
56 	if (!node)
57 		return -ENOMEM;
58 
59 	memcpy(&node->bank, bank, sizeof(*bank));
60 
61 	INIT_LIST_HEAD(&node->node);
62 	list_add_tail(&node->node, &banks->list);
63 
64 	banks->nr_banks++;
65 
66 	return 0;
67 }
68 
69 static void aca_banks_release(struct aca_banks *banks)
70 {
71 	struct aca_bank_node *node, *tmp;
72 
73 	if (list_empty(&banks->list))
74 		return;
75 
76 	list_for_each_entry_safe(node, tmp, &banks->list, node) {
77 		list_del(&node->node);
78 		kvfree(node);
79 		banks->nr_banks--;
80 	}
81 }
82 
83 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
84 {
85 	struct amdgpu_aca *aca = &adev->aca;
86 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
87 
88 	if (!count)
89 		return -EINVAL;
90 
91 	if (!smu_funcs || !smu_funcs->get_valid_aca_count)
92 		return -EOPNOTSUPP;
93 
94 	return smu_funcs->get_valid_aca_count(adev, type, count);
95 }
96 
97 static struct aca_regs_dump {
98 	const char *name;
99 	int reg_idx;
100 } aca_regs[] = {
101 	{"CONTROL",		ACA_REG_IDX_CTL},
102 	{"STATUS",		ACA_REG_IDX_STATUS},
103 	{"ADDR",		ACA_REG_IDX_ADDR},
104 	{"MISC",		ACA_REG_IDX_MISC0},
105 	{"CONFIG",		ACA_REG_IDX_CONFIG},
106 	{"IPID",		ACA_REG_IDX_IPID},
107 	{"SYND",		ACA_REG_IDX_SYND},
108 	{"DESTAT",		ACA_REG_IDX_DESTAT},
109 	{"DEADDR",		ACA_REG_IDX_DEADDR},
110 	{"CONTROL_MASK",	ACA_REG_IDX_CTL_MASK},
111 };
112 
113 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
114 			      struct ras_query_context *qctx)
115 {
116 	u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
117 	int i;
118 
119 	if (adev->debug_disable_ce_logs &&
120 	    bank->smu_err_type == ACA_SMU_TYPE_CE &&
121 	    !ACA_BANK_ERR_IS_DEFFERED(bank))
122 		return;
123 
124 	RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
125 	/* plus 1 for output format, e.g: ACA[08/08]: xxxx */
126 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
127 		RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
128 			      idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
129 
130 	if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS]))
131 		RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n");
132 }
133 
134 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
135 {
136 
137 	struct aca_hwip *hwip;
138 	int hwid, mcatype;
139 	u64 ipid;
140 
141 	if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
142 		return false;
143 
144 	hwip = &aca_hwid_mcatypes[type];
145 	if (!hwip->hwid)
146 		return false;
147 
148 	ipid = bank->regs[ACA_REG_IDX_IPID];
149 	hwid = ACA_REG__IPID__HARDWAREID(ipid);
150 	mcatype = ACA_REG__IPID__MCATYPE(ipid);
151 
152 	return hwip->hwid == hwid && hwip->mcatype == mcatype;
153 }
154 
155 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
156 				       int start, int count,
157 				       struct aca_banks *banks, struct ras_query_context *qctx)
158 {
159 	struct amdgpu_aca *aca = &adev->aca;
160 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
161 	struct aca_bank bank;
162 	int i, max_count, ret;
163 
164 	if (!count)
165 		return 0;
166 
167 	if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
168 		return -EOPNOTSUPP;
169 
170 	switch (type) {
171 	case ACA_SMU_TYPE_UE:
172 		max_count = smu_funcs->max_ue_bank_count;
173 		break;
174 	case ACA_SMU_TYPE_CE:
175 		max_count = smu_funcs->max_ce_bank_count;
176 		break;
177 	default:
178 		return -EINVAL;
179 	}
180 
181 	if (start + count > max_count)
182 		return -EINVAL;
183 
184 	count = min_t(int, count, max_count);
185 	for (i = 0; i < count; i++) {
186 		memset(&bank, 0, sizeof(bank));
187 		ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
188 		if (ret)
189 			return ret;
190 
191 		bank.smu_err_type = type;
192 
193 		/*
194 		 * Poison being consumed when injecting a UE while running background workloads,
195 		 * which are unexpected.
196 		 */
197 		if (type == ACA_SMU_TYPE_UE &&
198 		    ACA_REG__STATUS__POISON(bank.regs[ACA_REG_IDX_STATUS]) &&
199 		    !aca_bank_hwip_is_matched(&bank, ACA_HWIP_TYPE_UMC))
200 			continue;
201 
202 		aca_smu_bank_dump(adev, i, count, &bank, qctx);
203 
204 		ret = aca_banks_add_bank(banks, &bank);
205 		if (ret)
206 			return ret;
207 	}
208 
209 	return 0;
210 }
211 
212 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
213 {
214 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
215 
216 	/* Parse all deferred errors with UMC aca handle */
217 	if (ACA_BANK_ERR_IS_DEFFERED(bank))
218 		return handle->hwip == ACA_HWIP_TYPE_UMC;
219 
220 	if (!aca_bank_hwip_is_matched(bank, handle->hwip))
221 		return false;
222 
223 	if (!bank_ops->aca_bank_is_valid)
224 		return true;
225 
226 	return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
227 }
228 
229 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
230 {
231 	struct aca_bank_error *bank_error;
232 
233 	bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
234 	if (!bank_error)
235 		return NULL;
236 
237 	INIT_LIST_HEAD(&bank_error->node);
238 	memcpy(&bank_error->info, info, sizeof(*info));
239 
240 	mutex_lock(&aerr->lock);
241 	list_add_tail(&bank_error->node, &aerr->list);
242 	aerr->nr_errors++;
243 	mutex_unlock(&aerr->lock);
244 
245 	return bank_error;
246 }
247 
248 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
249 {
250 	struct aca_bank_error *bank_error = NULL;
251 	struct aca_bank_info *tmp_info;
252 	bool found = false;
253 
254 	mutex_lock(&aerr->lock);
255 	list_for_each_entry(bank_error, &aerr->list, node) {
256 		tmp_info = &bank_error->info;
257 		if (tmp_info->socket_id == info->socket_id &&
258 		    tmp_info->die_id == info->die_id) {
259 			found = true;
260 			goto out_unlock;
261 		}
262 	}
263 
264 out_unlock:
265 	mutex_unlock(&aerr->lock);
266 
267 	return found ? bank_error : NULL;
268 }
269 
270 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
271 {
272 	if (!aerr || !bank_error)
273 		return;
274 
275 	list_del(&bank_error->node);
276 	aerr->nr_errors--;
277 
278 	kvfree(bank_error);
279 }
280 
281 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
282 {
283 	struct aca_bank_error *bank_error;
284 
285 	if (!aerr || !info)
286 		return NULL;
287 
288 	bank_error = find_bank_error(aerr, info);
289 	if (bank_error)
290 		return bank_error;
291 
292 	return new_bank_error(aerr, info);
293 }
294 
295 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
296 				   enum aca_error_type type, u64 count)
297 {
298 	struct aca_error_cache *error_cache = &handle->error_cache;
299 	struct aca_bank_error *bank_error;
300 	struct aca_error *aerr;
301 
302 	if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
303 		return -EINVAL;
304 
305 	if (!count)
306 		return 0;
307 
308 	aerr = &error_cache->errors[type];
309 	bank_error = get_bank_error(aerr, info);
310 	if (!bank_error)
311 		return -ENOMEM;
312 
313 	bank_error->count += count;
314 
315 	return 0;
316 }
317 
318 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
319 {
320 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
321 
322 	if (!bank)
323 		return -EINVAL;
324 
325 	if (!bank_ops->aca_bank_parser)
326 		return -EOPNOTSUPP;
327 
328 	return bank_ops->aca_bank_parser(handle, bank, type,
329 					 handle->data);
330 }
331 
332 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
333 				      enum aca_smu_type type, void *data)
334 {
335 	int ret;
336 
337 	ret = aca_bank_parser(handle, bank, type);
338 	if (ret)
339 		return ret;
340 
341 	return 0;
342 }
343 
344 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
345 			     enum aca_smu_type type, bank_handler_t handler, void *data)
346 {
347 	struct aca_handle *handle;
348 	int ret;
349 
350 	if (list_empty(&mgr->list))
351 		return 0;
352 
353 	list_for_each_entry(handle, &mgr->list, node) {
354 		if (!aca_bank_is_valid(handle, bank, type))
355 			continue;
356 
357 		ret = handler(handle, bank, type, data);
358 		if (ret)
359 			return ret;
360 	}
361 
362 	return 0;
363 }
364 
365 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
366 			      enum aca_smu_type type, bank_handler_t handler, void *data)
367 {
368 	struct aca_bank_node *node;
369 	struct aca_bank *bank;
370 	int ret;
371 
372 	if (!mgr || !banks)
373 		return -EINVAL;
374 
375 	/* pre check to avoid unnecessary operations */
376 	if (list_empty(&mgr->list) || list_empty(&banks->list))
377 		return 0;
378 
379 	list_for_each_entry(node, &banks->list, node) {
380 		bank = &node->bank;
381 
382 		ret = aca_dispatch_bank(mgr, bank, type, handler, data);
383 		if (ret)
384 			return ret;
385 	}
386 
387 	return 0;
388 }
389 
390 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
391 {
392 	struct amdgpu_aca *aca = &adev->aca;
393 	bool ret = true;
394 
395 	/*
396 	 * Because the UE Valid MCA count will only be cleared after reset,
397 	 * in order to avoid repeated counting of the error count,
398 	 * the aca bank is only updated once during the gpu recovery stage.
399 	 */
400 	if (type == ACA_SMU_TYPE_UE) {
401 		if (amdgpu_ras_intr_triggered())
402 			ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
403 		else
404 			atomic_set(&aca->ue_update_flag, 0);
405 	}
406 
407 	return ret;
408 }
409 
410 static void aca_banks_generate_cper(struct amdgpu_device *adev,
411 				    enum aca_smu_type type,
412 				    struct aca_banks *banks,
413 				    int count)
414 {
415 	struct aca_bank_node *node;
416 	struct aca_bank *bank;
417 	int r;
418 
419 	if (!adev->cper.enabled)
420 		return;
421 
422 	if (!banks || !count) {
423 		dev_warn(adev->dev, "fail to generate cper records\n");
424 		return;
425 	}
426 
427 	/* UEs must be encoded into separate CPER entries */
428 	if (type == ACA_SMU_TYPE_UE) {
429 		struct aca_banks de_banks;
430 
431 		aca_banks_init(&de_banks);
432 		list_for_each_entry(node, &banks->list, node) {
433 			bank = &node->bank;
434 			if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {
435 				r = aca_banks_add_bank(&de_banks, bank);
436 				if (r)
437 					dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);
438 			} else {
439 				if (amdgpu_cper_generate_ue_record(adev, bank))
440 					dev_warn(adev->dev, "fail to generate ue cper records\n");
441 			}
442 		}
443 
444 		if (!list_empty(&de_banks.list)) {
445 			if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))
446 				dev_warn(adev->dev, "fail to generate de cper records\n");
447 		}
448 
449 		aca_banks_release(&de_banks);
450 	} else {
451 		/*
452 		 * SMU_TYPE_CE banks are combined into 1 CPER entries,
453 		 * they could be CEs or DEs or both
454 		 */
455 		if (amdgpu_cper_generate_ce_records(adev, banks, count))
456 			dev_warn(adev->dev, "fail to generate ce cper records\n");
457 	}
458 }
459 
460 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
461 			    bank_handler_t handler, struct ras_query_context *qctx, void *data)
462 {
463 	struct amdgpu_aca *aca = &adev->aca;
464 	struct aca_banks banks;
465 	u32 count = 0;
466 	int ret;
467 
468 	if (list_empty(&aca->mgr.list))
469 		return 0;
470 
471 	if (!aca_bank_should_update(adev, type))
472 		return 0;
473 
474 	ret = aca_smu_get_valid_aca_count(adev, type, &count);
475 	if (ret)
476 		return ret;
477 
478 	if (!count)
479 		return 0;
480 
481 	aca_banks_init(&banks);
482 
483 	ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
484 	if (ret)
485 		goto err_release_banks;
486 
487 	if (list_empty(&banks.list)) {
488 		ret = 0;
489 		goto err_release_banks;
490 	}
491 
492 	ret = aca_dispatch_banks(&aca->mgr, &banks, type,
493 				 handler, data);
494 	if (ret)
495 		goto err_release_banks;
496 
497 	aca_banks_generate_cper(adev, type, &banks, count);
498 
499 err_release_banks:
500 	aca_banks_release(&banks);
501 
502 	return ret;
503 }
504 
505 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
506 {
507 	struct aca_bank_info *info;
508 	struct amdgpu_smuio_mcm_config_info mcm_info;
509 	u64 count;
510 
511 	if (type >= ACA_ERROR_TYPE_COUNT)
512 		return -EINVAL;
513 
514 	count = bank_error->count;
515 	if (!count)
516 		return 0;
517 
518 	info = &bank_error->info;
519 	mcm_info.die_id = info->die_id;
520 	mcm_info.socket_id = info->socket_id;
521 
522 	switch (type) {
523 	case ACA_ERROR_TYPE_UE:
524 		amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);
525 		break;
526 	case ACA_ERROR_TYPE_CE:
527 		amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);
528 		break;
529 	case ACA_ERROR_TYPE_DEFERRED:
530 		amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);
531 		break;
532 	default:
533 		break;
534 	}
535 
536 	return 0;
537 }
538 
539 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
540 {
541 	struct aca_error_cache *error_cache = &handle->error_cache;
542 	struct aca_error *aerr = &error_cache->errors[type];
543 	struct aca_bank_error *bank_error, *tmp;
544 
545 	mutex_lock(&aerr->lock);
546 
547 	if (list_empty(&aerr->list))
548 		goto out_unlock;
549 
550 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
551 		aca_log_aca_error_data(bank_error, type, err_data);
552 		aca_bank_error_remove(aerr, bank_error);
553 	}
554 
555 out_unlock:
556 	mutex_unlock(&aerr->lock);
557 
558 	return 0;
559 }
560 
561 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
562 				struct ras_err_data *err_data, struct ras_query_context *qctx)
563 {
564 	enum aca_smu_type smu_type;
565 	int ret;
566 
567 	switch (type) {
568 	case ACA_ERROR_TYPE_UE:
569 		smu_type = ACA_SMU_TYPE_UE;
570 		break;
571 	case ACA_ERROR_TYPE_CE:
572 	case ACA_ERROR_TYPE_DEFERRED:
573 		smu_type = ACA_SMU_TYPE_CE;
574 		break;
575 	default:
576 		return -EINVAL;
577 	}
578 
579 	/* update aca bank to aca source error_cache first */
580 	ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
581 	if (ret)
582 		return ret;
583 
584 	/* DEs may contain in CEs or UEs */
585 	if (type != ACA_ERROR_TYPE_DEFERRED)
586 		aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);
587 
588 	return aca_log_aca_error(handle, type, err_data);
589 }
590 
591 static bool aca_handle_is_valid(struct aca_handle *handle)
592 {
593 	if (!handle->mask || !list_empty(&handle->node))
594 		return false;
595 
596 	return true;
597 }
598 
599 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
600 			      enum aca_error_type type, struct ras_err_data *err_data,
601 			      struct ras_query_context *qctx)
602 {
603 	if (!handle || !err_data)
604 		return -EINVAL;
605 
606 	if (aca_handle_is_valid(handle))
607 		return -EOPNOTSUPP;
608 
609 	if ((type < 0) || (!(BIT(type) & handle->mask)))
610 		return  0;
611 
612 	return __aca_get_error_data(adev, handle, type, err_data, qctx);
613 }
614 
615 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
616 {
617 	mutex_init(&aerr->lock);
618 	INIT_LIST_HEAD(&aerr->list);
619 	aerr->type = type;
620 	aerr->nr_errors = 0;
621 }
622 
623 static void aca_init_error_cache(struct aca_handle *handle)
624 {
625 	struct aca_error_cache *error_cache = &handle->error_cache;
626 	int type;
627 
628 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
629 		aca_error_init(&error_cache->errors[type], type);
630 }
631 
632 static void aca_error_fini(struct aca_error *aerr)
633 {
634 	struct aca_bank_error *bank_error, *tmp;
635 
636 	mutex_lock(&aerr->lock);
637 	if (list_empty(&aerr->list))
638 		goto out_unlock;
639 
640 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
641 		aca_bank_error_remove(aerr, bank_error);
642 
643 out_unlock:
644 	mutex_destroy(&aerr->lock);
645 }
646 
647 static void aca_fini_error_cache(struct aca_handle *handle)
648 {
649 	struct aca_error_cache *error_cache = &handle->error_cache;
650 	int type;
651 
652 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
653 		aca_error_fini(&error_cache->errors[type]);
654 }
655 
656 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
657 			  const char *name, const struct aca_info *ras_info, void *data)
658 {
659 	memset(handle, 0, sizeof(*handle));
660 
661 	handle->adev = adev;
662 	handle->mgr = mgr;
663 	handle->name = name;
664 	handle->hwip = ras_info->hwip;
665 	handle->mask = ras_info->mask;
666 	handle->bank_ops = ras_info->bank_ops;
667 	handle->data = data;
668 	aca_init_error_cache(handle);
669 
670 	INIT_LIST_HEAD(&handle->node);
671 	list_add_tail(&handle->node, &mgr->list);
672 	mgr->nr_handles++;
673 
674 	return 0;
675 }
676 
677 static ssize_t aca_sysfs_read(struct device *dev,
678 			      struct device_attribute *attr, char *buf)
679 {
680 	struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
681 
682 	/* NOTE: the aca cache will be auto cleared once read,
683 	 * So the driver should unify the query entry point, forward request to ras query interface directly */
684 	return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
685 }
686 
687 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
688 {
689 	struct device_attribute *aca_attr = &handle->aca_attr;
690 
691 	snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
692 	aca_attr->show = aca_sysfs_read;
693 	aca_attr->attr.name = handle->attr_name;
694 	aca_attr->attr.mode = S_IRUGO;
695 	sysfs_attr_init(&aca_attr->attr);
696 
697 	return sysfs_add_file_to_group(&adev->dev->kobj,
698 				       &aca_attr->attr,
699 				       "ras");
700 }
701 
702 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
703 			  const char *name, const struct aca_info *ras_info, void *data)
704 {
705 	struct amdgpu_aca *aca = &adev->aca;
706 	int ret;
707 
708 	if (!amdgpu_aca_is_enabled(adev))
709 		return 0;
710 
711 	ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
712 	if (ret)
713 		return ret;
714 
715 	return add_aca_sysfs(adev, handle);
716 }
717 
718 static void remove_aca_handle(struct aca_handle *handle)
719 {
720 	struct aca_handle_manager *mgr = handle->mgr;
721 
722 	aca_fini_error_cache(handle);
723 	list_del(&handle->node);
724 	mgr->nr_handles--;
725 }
726 
727 static void remove_aca_sysfs(struct aca_handle *handle)
728 {
729 	struct amdgpu_device *adev = handle->adev;
730 	struct device_attribute *aca_attr = &handle->aca_attr;
731 
732 	if (adev->dev->kobj.sd)
733 		sysfs_remove_file_from_group(&adev->dev->kobj,
734 					     &aca_attr->attr,
735 					     "ras");
736 }
737 
738 void amdgpu_aca_remove_handle(struct aca_handle *handle)
739 {
740 	if (!handle || list_empty(&handle->node))
741 		return;
742 
743 	remove_aca_sysfs(handle);
744 	remove_aca_handle(handle);
745 }
746 
747 static int aca_manager_init(struct aca_handle_manager *mgr)
748 {
749 	INIT_LIST_HEAD(&mgr->list);
750 	mgr->nr_handles = 0;
751 
752 	return 0;
753 }
754 
755 static void aca_manager_fini(struct aca_handle_manager *mgr)
756 {
757 	struct aca_handle *handle, *tmp;
758 
759 	if (list_empty(&mgr->list))
760 		return;
761 
762 	list_for_each_entry_safe(handle, tmp, &mgr->list, node)
763 		amdgpu_aca_remove_handle(handle);
764 }
765 
766 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
767 {
768 	return (adev->aca.is_enabled ||
769 		adev->debug_enable_ras_aca);
770 }
771 
772 int amdgpu_aca_init(struct amdgpu_device *adev)
773 {
774 	struct amdgpu_aca *aca = &adev->aca;
775 	int ret;
776 
777 	atomic_set(&aca->ue_update_flag, 0);
778 
779 	ret = aca_manager_init(&aca->mgr);
780 	if (ret)
781 		return ret;
782 
783 	return 0;
784 }
785 
786 void amdgpu_aca_fini(struct amdgpu_device *adev)
787 {
788 	struct amdgpu_aca *aca = &adev->aca;
789 
790 	aca_manager_fini(&aca->mgr);
791 
792 	atomic_set(&aca->ue_update_flag, 0);
793 }
794 
795 int amdgpu_aca_reset(struct amdgpu_device *adev)
796 {
797 	struct amdgpu_aca *aca = &adev->aca;
798 
799 	atomic_set(&aca->ue_update_flag, 0);
800 
801 	return 0;
802 }
803 
804 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
805 {
806 	struct amdgpu_aca *aca = &adev->aca;
807 
808 	WARN_ON(aca->smu_funcs);
809 	aca->smu_funcs = smu_funcs;
810 }
811 
812 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
813 {
814 	u64 ipid;
815 	u32 instidhi, instidlo;
816 
817 	if (!bank || !info)
818 		return -EINVAL;
819 
820 	ipid = bank->regs[ACA_REG_IDX_IPID];
821 	info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
822 	info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
823 	/*
824 	 * Unfied DieID Format: SAASS. A:AID, S:Socket.
825 	 * Unfied DieID[4:4] = InstanceId[0:0]
826 	 * Unfied DieID[0:3] = InstanceIdHi[0:3]
827 	 */
828 	instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
829 	instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
830 	info->die_id = ((instidhi >> 2) & 0x03);
831 	info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
832 
833 	return 0;
834 }
835 
836 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
837 {
838 	struct amdgpu_aca *aca = &adev->aca;
839 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
840 
841 	if (!smu_funcs || !smu_funcs->parse_error_code)
842 		return -EOPNOTSUPP;
843 
844 	return smu_funcs->parse_error_code(adev, bank);
845 }
846 
847 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
848 {
849 	int i, error_code;
850 
851 	if (!bank || !err_codes)
852 		return -EINVAL;
853 
854 	error_code = aca_bank_get_error_code(adev, bank);
855 	if (error_code < 0)
856 		return error_code;
857 
858 	for (i = 0; i < size; i++) {
859 		if (err_codes[i] == error_code)
860 			return 0;
861 	}
862 
863 	return -EINVAL;
864 }
865 
866 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
867 {
868 	struct amdgpu_aca *aca = &adev->aca;
869 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
870 
871 	if (!smu_funcs || !smu_funcs->set_debug_mode)
872 		return -EOPNOTSUPP;
873 
874 	return smu_funcs->set_debug_mode(adev, en);
875 }
876 
877 #if defined(CONFIG_DEBUG_FS)
878 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
879 {
880 	struct amdgpu_device *adev = (struct amdgpu_device *)data;
881 	int ret;
882 
883 	ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
884 	if (ret)
885 		return ret;
886 
887 	dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
888 
889 	return 0;
890 }
891 
892 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
893 {
894 	struct aca_bank_info info;
895 	int i, ret;
896 
897 	ret = aca_bank_info_decode(bank, &info);
898 	if (ret)
899 		return;
900 
901 	seq_printf(m, "aca entry[%d].type: %s\n", idx, type ==  ACA_SMU_TYPE_UE ? "UE" : "CE");
902 	seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
903 		   idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
904 
905 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
906 		seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
907 }
908 
909 struct aca_dump_context {
910 	struct seq_file *m;
911 	int idx;
912 };
913 
914 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
915 				 enum aca_smu_type type, void *data)
916 {
917 	struct aca_dump_context *ctx = (struct aca_dump_context *)data;
918 
919 	aca_dump_entry(ctx->m, bank, type, ctx->idx++);
920 
921 	return handler_aca_log_bank_error(handle, bank, type, NULL);
922 }
923 
924 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
925 {
926 	struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
927 	struct aca_dump_context context = {
928 		.m = m,
929 		.idx = 0,
930 	};
931 
932 	return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
933 }
934 
935 static int aca_dump_ce_show(struct seq_file *m, void *unused)
936 {
937 	return aca_dump_show(m, ACA_SMU_TYPE_CE);
938 }
939 
940 static int aca_dump_ce_open(struct inode *inode, struct file *file)
941 {
942 	return single_open(file, aca_dump_ce_show, inode->i_private);
943 }
944 
945 static const struct file_operations aca_ce_dump_debug_fops = {
946 	.owner = THIS_MODULE,
947 	.open = aca_dump_ce_open,
948 	.read = seq_read,
949 	.llseek = seq_lseek,
950 	.release = single_release,
951 };
952 
953 static int aca_dump_ue_show(struct seq_file *m, void *unused)
954 {
955 	return aca_dump_show(m, ACA_SMU_TYPE_UE);
956 }
957 
958 static int aca_dump_ue_open(struct inode *inode, struct file *file)
959 {
960 	return single_open(file, aca_dump_ue_show, inode->i_private);
961 }
962 
963 static const struct file_operations aca_ue_dump_debug_fops = {
964 	.owner = THIS_MODULE,
965 	.open = aca_dump_ue_open,
966 	.read = seq_read,
967 	.llseek = seq_lseek,
968 	.release = single_release,
969 };
970 
971 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
972 #endif
973 
974 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
975 {
976 #if defined(CONFIG_DEBUG_FS)
977 	if (!root)
978 		return;
979 
980 	debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
981 	debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
982 	debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
983 #endif
984 }
985