xref: /linux/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c (revision 1b392348de8ffc340545c62079645cb4695f5602)
1 /*
2  * Copyright 2023 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  */
23 
24 #include <linux/list.h>
25 #include "amdgpu.h"
26 #include "amdgpu_aca.h"
27 #include "amdgpu_ras.h"
28 
29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30 
31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
32 
33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
34 	ACA_BANK_HWID(SMU,	0x01,	0x01),
35 	ACA_BANK_HWID(PCS_XGMI, 0x50,	0x00),
36 	ACA_BANK_HWID(UMC,	0x96,	0x00),
37 };
38 
39 static void aca_banks_init(struct aca_banks *banks)
40 {
41 	if (!banks)
42 		return;
43 
44 	memset(banks, 0, sizeof(*banks));
45 	INIT_LIST_HEAD(&banks->list);
46 }
47 
48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
49 {
50 	struct aca_bank_node *node;
51 
52 	if (!bank)
53 		return -EINVAL;
54 
55 	node = kvzalloc(sizeof(*node), GFP_KERNEL);
56 	if (!node)
57 		return -ENOMEM;
58 
59 	memcpy(&node->bank, bank, sizeof(*bank));
60 
61 	INIT_LIST_HEAD(&node->node);
62 	list_add_tail(&node->node, &banks->list);
63 
64 	banks->nr_banks++;
65 
66 	return 0;
67 }
68 
69 static void aca_banks_release(struct aca_banks *banks)
70 {
71 	struct aca_bank_node *node, *tmp;
72 
73 	if (list_empty(&banks->list))
74 		return;
75 
76 	list_for_each_entry_safe(node, tmp, &banks->list, node) {
77 		list_del(&node->node);
78 		kvfree(node);
79 	}
80 }
81 
82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
83 {
84 	struct amdgpu_aca *aca = &adev->aca;
85 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
86 
87 	if (!count)
88 		return -EINVAL;
89 
90 	if (!smu_funcs || !smu_funcs->get_valid_aca_count)
91 		return -EOPNOTSUPP;
92 
93 	return smu_funcs->get_valid_aca_count(adev, type, count);
94 }
95 
96 static struct aca_regs_dump {
97 	const char *name;
98 	int reg_idx;
99 } aca_regs[] = {
100 	{"CONTROL",		ACA_REG_IDX_CTL},
101 	{"STATUS",		ACA_REG_IDX_STATUS},
102 	{"ADDR",		ACA_REG_IDX_ADDR},
103 	{"MISC",		ACA_REG_IDX_MISC0},
104 	{"CONFIG",		ACA_REG_IDX_CONFIG},
105 	{"IPID",		ACA_REG_IDX_IPID},
106 	{"SYND",		ACA_REG_IDX_SYND},
107 	{"DESTAT",		ACA_REG_IDX_DESTAT},
108 	{"DEADDR",		ACA_REG_IDX_DEADDR},
109 	{"CONTROL_MASK",	ACA_REG_IDX_CTL_MASK},
110 };
111 
112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
113 			      struct ras_query_context *qctx)
114 {
115 	u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
116 	int i;
117 
118 	if (adev->debug_disable_ce_logs &&
119 	    bank->smu_err_type == ACA_SMU_TYPE_CE &&
120 	    !ACA_BANK_ERR_IS_DEFFERED(bank))
121 		return;
122 
123 	RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
124 	/* plus 1 for output format, e.g: ACA[08/08]: xxxx */
125 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
126 		RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
127 			      idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
128 
129 	if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS]))
130 		RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n");
131 }
132 
133 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
134 {
135 
136 	struct aca_hwip *hwip;
137 	int hwid, mcatype;
138 	u64 ipid;
139 
140 	if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
141 		return false;
142 
143 	hwip = &aca_hwid_mcatypes[type];
144 	if (!hwip->hwid)
145 		return false;
146 
147 	ipid = bank->regs[ACA_REG_IDX_IPID];
148 	hwid = ACA_REG__IPID__HARDWAREID(ipid);
149 	mcatype = ACA_REG__IPID__MCATYPE(ipid);
150 
151 	return hwip->hwid == hwid && hwip->mcatype == mcatype;
152 }
153 
154 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
155 				       int start, int count,
156 				       struct aca_banks *banks, struct ras_query_context *qctx)
157 {
158 	struct amdgpu_aca *aca = &adev->aca;
159 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
160 	struct aca_bank bank;
161 	int i, max_count, ret;
162 
163 	if (!count)
164 		return 0;
165 
166 	if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
167 		return -EOPNOTSUPP;
168 
169 	switch (type) {
170 	case ACA_SMU_TYPE_UE:
171 		max_count = smu_funcs->max_ue_bank_count;
172 		break;
173 	case ACA_SMU_TYPE_CE:
174 		max_count = smu_funcs->max_ce_bank_count;
175 		break;
176 	default:
177 		return -EINVAL;
178 	}
179 
180 	if (start + count > max_count)
181 		return -EINVAL;
182 
183 	count = min_t(int, count, max_count);
184 	for (i = 0; i < count; i++) {
185 		memset(&bank, 0, sizeof(bank));
186 		ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
187 		if (ret)
188 			return ret;
189 
190 		bank.smu_err_type = type;
191 
192 		/*
193 		 * Poison being consumed when injecting a UE while running background workloads,
194 		 * which are unexpected.
195 		 */
196 		if (type == ACA_SMU_TYPE_UE &&
197 		    ACA_REG__STATUS__POISON(bank.regs[ACA_REG_IDX_STATUS]) &&
198 		    !aca_bank_hwip_is_matched(&bank, ACA_HWIP_TYPE_UMC))
199 			continue;
200 
201 		aca_smu_bank_dump(adev, i, count, &bank, qctx);
202 
203 		ret = aca_banks_add_bank(banks, &bank);
204 		if (ret)
205 			return ret;
206 	}
207 
208 	return 0;
209 }
210 
211 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
212 {
213 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
214 
215 	/* Parse all deferred errors with UMC aca handle */
216 	if (ACA_BANK_ERR_IS_DEFFERED(bank))
217 		return handle->hwip == ACA_HWIP_TYPE_UMC;
218 
219 	if (!aca_bank_hwip_is_matched(bank, handle->hwip))
220 		return false;
221 
222 	if (!bank_ops->aca_bank_is_valid)
223 		return true;
224 
225 	return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
226 }
227 
228 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
229 {
230 	struct aca_bank_error *bank_error;
231 
232 	bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
233 	if (!bank_error)
234 		return NULL;
235 
236 	INIT_LIST_HEAD(&bank_error->node);
237 	memcpy(&bank_error->info, info, sizeof(*info));
238 
239 	mutex_lock(&aerr->lock);
240 	list_add_tail(&bank_error->node, &aerr->list);
241 	mutex_unlock(&aerr->lock);
242 
243 	return bank_error;
244 }
245 
246 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
247 {
248 	struct aca_bank_error *bank_error = NULL;
249 	struct aca_bank_info *tmp_info;
250 	bool found = false;
251 
252 	mutex_lock(&aerr->lock);
253 	list_for_each_entry(bank_error, &aerr->list, node) {
254 		tmp_info = &bank_error->info;
255 		if (tmp_info->socket_id == info->socket_id &&
256 		    tmp_info->die_id == info->die_id) {
257 			found = true;
258 			goto out_unlock;
259 		}
260 	}
261 
262 out_unlock:
263 	mutex_unlock(&aerr->lock);
264 
265 	return found ? bank_error : NULL;
266 }
267 
268 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
269 {
270 	if (!aerr || !bank_error)
271 		return;
272 
273 	list_del(&bank_error->node);
274 	aerr->nr_errors--;
275 
276 	kvfree(bank_error);
277 }
278 
279 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
280 {
281 	struct aca_bank_error *bank_error;
282 
283 	if (!aerr || !info)
284 		return NULL;
285 
286 	bank_error = find_bank_error(aerr, info);
287 	if (bank_error)
288 		return bank_error;
289 
290 	return new_bank_error(aerr, info);
291 }
292 
293 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
294 				   enum aca_error_type type, u64 count)
295 {
296 	struct aca_error_cache *error_cache = &handle->error_cache;
297 	struct aca_bank_error *bank_error;
298 	struct aca_error *aerr;
299 
300 	if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
301 		return -EINVAL;
302 
303 	if (!count)
304 		return 0;
305 
306 	aerr = &error_cache->errors[type];
307 	bank_error = get_bank_error(aerr, info);
308 	if (!bank_error)
309 		return -ENOMEM;
310 
311 	bank_error->count += count;
312 
313 	return 0;
314 }
315 
316 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
317 {
318 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
319 
320 	if (!bank)
321 		return -EINVAL;
322 
323 	if (!bank_ops->aca_bank_parser)
324 		return -EOPNOTSUPP;
325 
326 	return bank_ops->aca_bank_parser(handle, bank, type,
327 					 handle->data);
328 }
329 
330 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
331 				      enum aca_smu_type type, void *data)
332 {
333 	int ret;
334 
335 	ret = aca_bank_parser(handle, bank, type);
336 	if (ret)
337 		return ret;
338 
339 	return 0;
340 }
341 
342 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
343 			     enum aca_smu_type type, bank_handler_t handler, void *data)
344 {
345 	struct aca_handle *handle;
346 	int ret;
347 
348 	if (list_empty(&mgr->list))
349 		return 0;
350 
351 	list_for_each_entry(handle, &mgr->list, node) {
352 		if (!aca_bank_is_valid(handle, bank, type))
353 			continue;
354 
355 		ret = handler(handle, bank, type, data);
356 		if (ret)
357 			return ret;
358 	}
359 
360 	return 0;
361 }
362 
363 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
364 			      enum aca_smu_type type, bank_handler_t handler, void *data)
365 {
366 	struct aca_bank_node *node;
367 	struct aca_bank *bank;
368 	int ret;
369 
370 	if (!mgr || !banks)
371 		return -EINVAL;
372 
373 	/* pre check to avoid unnecessary operations */
374 	if (list_empty(&mgr->list) || list_empty(&banks->list))
375 		return 0;
376 
377 	list_for_each_entry(node, &banks->list, node) {
378 		bank = &node->bank;
379 
380 		ret = aca_dispatch_bank(mgr, bank, type, handler, data);
381 		if (ret)
382 			return ret;
383 	}
384 
385 	return 0;
386 }
387 
388 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
389 {
390 	struct amdgpu_aca *aca = &adev->aca;
391 	bool ret = true;
392 
393 	/*
394 	 * Because the UE Valid MCA count will only be cleared after reset,
395 	 * in order to avoid repeated counting of the error count,
396 	 * the aca bank is only updated once during the gpu recovery stage.
397 	 */
398 	if (type == ACA_SMU_TYPE_UE) {
399 		if (amdgpu_ras_intr_triggered())
400 			ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
401 		else
402 			atomic_set(&aca->ue_update_flag, 0);
403 	}
404 
405 	return ret;
406 }
407 
408 static void aca_banks_generate_cper(struct amdgpu_device *adev,
409 				    enum aca_smu_type type,
410 				    struct aca_banks *banks,
411 				    int count)
412 {
413 	struct aca_bank_node *node;
414 	struct aca_bank *bank;
415 	int r;
416 
417 	if (!adev->cper.enabled)
418 		return;
419 
420 	if (!banks || !count) {
421 		dev_warn(adev->dev, "fail to generate cper records\n");
422 		return;
423 	}
424 
425 	/* UEs must be encoded into separate CPER entries */
426 	if (type == ACA_SMU_TYPE_UE) {
427 		struct aca_banks de_banks;
428 
429 		aca_banks_init(&de_banks);
430 		list_for_each_entry(node, &banks->list, node) {
431 			bank = &node->bank;
432 			if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {
433 				r = aca_banks_add_bank(&de_banks, bank);
434 				if (r)
435 					dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);
436 			} else {
437 				if (amdgpu_cper_generate_ue_record(adev, bank))
438 					dev_warn(adev->dev, "fail to generate ue cper records\n");
439 			}
440 		}
441 
442 		if (!list_empty(&de_banks.list)) {
443 			if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))
444 				dev_warn(adev->dev, "fail to generate de cper records\n");
445 		}
446 
447 		aca_banks_release(&de_banks);
448 	} else {
449 		/*
450 		 * SMU_TYPE_CE banks are combined into 1 CPER entries,
451 		 * they could be CEs or DEs or both
452 		 */
453 		if (amdgpu_cper_generate_ce_records(adev, banks, count))
454 			dev_warn(adev->dev, "fail to generate ce cper records\n");
455 	}
456 }
457 
458 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
459 			    bank_handler_t handler, struct ras_query_context *qctx, void *data)
460 {
461 	struct amdgpu_aca *aca = &adev->aca;
462 	struct aca_banks banks;
463 	u32 count = 0;
464 	int ret;
465 
466 	if (list_empty(&aca->mgr.list))
467 		return 0;
468 
469 	if (!aca_bank_should_update(adev, type))
470 		return 0;
471 
472 	ret = aca_smu_get_valid_aca_count(adev, type, &count);
473 	if (ret)
474 		return ret;
475 
476 	if (!count)
477 		return 0;
478 
479 	aca_banks_init(&banks);
480 
481 	ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
482 	if (ret)
483 		goto err_release_banks;
484 
485 	if (list_empty(&banks.list)) {
486 		ret = 0;
487 		goto err_release_banks;
488 	}
489 
490 	ret = aca_dispatch_banks(&aca->mgr, &banks, type,
491 				 handler, data);
492 	if (ret)
493 		goto err_release_banks;
494 
495 	aca_banks_generate_cper(adev, type, &banks, count);
496 
497 err_release_banks:
498 	aca_banks_release(&banks);
499 
500 	return ret;
501 }
502 
503 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
504 {
505 	struct aca_bank_info *info;
506 	struct amdgpu_smuio_mcm_config_info mcm_info;
507 	u64 count;
508 
509 	if (type >= ACA_ERROR_TYPE_COUNT)
510 		return -EINVAL;
511 
512 	count = bank_error->count;
513 	if (!count)
514 		return 0;
515 
516 	info = &bank_error->info;
517 	mcm_info.die_id = info->die_id;
518 	mcm_info.socket_id = info->socket_id;
519 
520 	switch (type) {
521 	case ACA_ERROR_TYPE_UE:
522 		amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);
523 		break;
524 	case ACA_ERROR_TYPE_CE:
525 		amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);
526 		break;
527 	case ACA_ERROR_TYPE_DEFERRED:
528 		amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);
529 		break;
530 	default:
531 		break;
532 	}
533 
534 	return 0;
535 }
536 
537 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
538 {
539 	struct aca_error_cache *error_cache = &handle->error_cache;
540 	struct aca_error *aerr = &error_cache->errors[type];
541 	struct aca_bank_error *bank_error, *tmp;
542 
543 	mutex_lock(&aerr->lock);
544 
545 	if (list_empty(&aerr->list))
546 		goto out_unlock;
547 
548 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
549 		aca_log_aca_error_data(bank_error, type, err_data);
550 		aca_bank_error_remove(aerr, bank_error);
551 	}
552 
553 out_unlock:
554 	mutex_unlock(&aerr->lock);
555 
556 	return 0;
557 }
558 
559 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
560 				struct ras_err_data *err_data, struct ras_query_context *qctx)
561 {
562 	enum aca_smu_type smu_type;
563 	int ret;
564 
565 	switch (type) {
566 	case ACA_ERROR_TYPE_UE:
567 		smu_type = ACA_SMU_TYPE_UE;
568 		break;
569 	case ACA_ERROR_TYPE_CE:
570 	case ACA_ERROR_TYPE_DEFERRED:
571 		smu_type = ACA_SMU_TYPE_CE;
572 		break;
573 	default:
574 		return -EINVAL;
575 	}
576 
577 	/* update aca bank to aca source error_cache first */
578 	ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
579 	if (ret)
580 		return ret;
581 
582 	/* DEs may contain in CEs or UEs */
583 	if (type != ACA_ERROR_TYPE_DEFERRED)
584 		aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);
585 
586 	return aca_log_aca_error(handle, type, err_data);
587 }
588 
589 static bool aca_handle_is_valid(struct aca_handle *handle)
590 {
591 	if (!handle->mask || !list_empty(&handle->node))
592 		return false;
593 
594 	return true;
595 }
596 
597 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
598 			      enum aca_error_type type, struct ras_err_data *err_data,
599 			      struct ras_query_context *qctx)
600 {
601 	if (!handle || !err_data)
602 		return -EINVAL;
603 
604 	if (aca_handle_is_valid(handle))
605 		return -EOPNOTSUPP;
606 
607 	if ((type < 0) || (!(BIT(type) & handle->mask)))
608 		return  0;
609 
610 	return __aca_get_error_data(adev, handle, type, err_data, qctx);
611 }
612 
613 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
614 {
615 	mutex_init(&aerr->lock);
616 	INIT_LIST_HEAD(&aerr->list);
617 	aerr->type = type;
618 	aerr->nr_errors = 0;
619 }
620 
621 static void aca_init_error_cache(struct aca_handle *handle)
622 {
623 	struct aca_error_cache *error_cache = &handle->error_cache;
624 	int type;
625 
626 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
627 		aca_error_init(&error_cache->errors[type], type);
628 }
629 
630 static void aca_error_fini(struct aca_error *aerr)
631 {
632 	struct aca_bank_error *bank_error, *tmp;
633 
634 	mutex_lock(&aerr->lock);
635 	if (list_empty(&aerr->list))
636 		goto out_unlock;
637 
638 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
639 		aca_bank_error_remove(aerr, bank_error);
640 
641 out_unlock:
642 	mutex_destroy(&aerr->lock);
643 }
644 
645 static void aca_fini_error_cache(struct aca_handle *handle)
646 {
647 	struct aca_error_cache *error_cache = &handle->error_cache;
648 	int type;
649 
650 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
651 		aca_error_fini(&error_cache->errors[type]);
652 }
653 
654 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
655 			  const char *name, const struct aca_info *ras_info, void *data)
656 {
657 	memset(handle, 0, sizeof(*handle));
658 
659 	handle->adev = adev;
660 	handle->mgr = mgr;
661 	handle->name = name;
662 	handle->hwip = ras_info->hwip;
663 	handle->mask = ras_info->mask;
664 	handle->bank_ops = ras_info->bank_ops;
665 	handle->data = data;
666 	aca_init_error_cache(handle);
667 
668 	INIT_LIST_HEAD(&handle->node);
669 	list_add_tail(&handle->node, &mgr->list);
670 	mgr->nr_handles++;
671 
672 	return 0;
673 }
674 
675 static ssize_t aca_sysfs_read(struct device *dev,
676 			      struct device_attribute *attr, char *buf)
677 {
678 	struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
679 
680 	/* NOTE: the aca cache will be auto cleared once read,
681 	 * So the driver should unify the query entry point, forward request to ras query interface directly */
682 	return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
683 }
684 
685 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
686 {
687 	struct device_attribute *aca_attr = &handle->aca_attr;
688 
689 	snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
690 	aca_attr->show = aca_sysfs_read;
691 	aca_attr->attr.name = handle->attr_name;
692 	aca_attr->attr.mode = S_IRUGO;
693 	sysfs_attr_init(&aca_attr->attr);
694 
695 	return sysfs_add_file_to_group(&adev->dev->kobj,
696 				       &aca_attr->attr,
697 				       "ras");
698 }
699 
700 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
701 			  const char *name, const struct aca_info *ras_info, void *data)
702 {
703 	struct amdgpu_aca *aca = &adev->aca;
704 	int ret;
705 
706 	if (!amdgpu_aca_is_enabled(adev))
707 		return 0;
708 
709 	ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
710 	if (ret)
711 		return ret;
712 
713 	return add_aca_sysfs(adev, handle);
714 }
715 
716 static void remove_aca_handle(struct aca_handle *handle)
717 {
718 	struct aca_handle_manager *mgr = handle->mgr;
719 
720 	aca_fini_error_cache(handle);
721 	list_del(&handle->node);
722 	mgr->nr_handles--;
723 }
724 
725 static void remove_aca_sysfs(struct aca_handle *handle)
726 {
727 	struct amdgpu_device *adev = handle->adev;
728 	struct device_attribute *aca_attr = &handle->aca_attr;
729 
730 	if (adev->dev->kobj.sd)
731 		sysfs_remove_file_from_group(&adev->dev->kobj,
732 					     &aca_attr->attr,
733 					     "ras");
734 }
735 
736 void amdgpu_aca_remove_handle(struct aca_handle *handle)
737 {
738 	if (!handle || list_empty(&handle->node))
739 		return;
740 
741 	remove_aca_sysfs(handle);
742 	remove_aca_handle(handle);
743 }
744 
745 static int aca_manager_init(struct aca_handle_manager *mgr)
746 {
747 	INIT_LIST_HEAD(&mgr->list);
748 	mgr->nr_handles = 0;
749 
750 	return 0;
751 }
752 
753 static void aca_manager_fini(struct aca_handle_manager *mgr)
754 {
755 	struct aca_handle *handle, *tmp;
756 
757 	if (list_empty(&mgr->list))
758 		return;
759 
760 	list_for_each_entry_safe(handle, tmp, &mgr->list, node)
761 		amdgpu_aca_remove_handle(handle);
762 }
763 
764 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
765 {
766 	return (adev->aca.is_enabled ||
767 		adev->debug_enable_ras_aca);
768 }
769 
770 int amdgpu_aca_init(struct amdgpu_device *adev)
771 {
772 	struct amdgpu_aca *aca = &adev->aca;
773 	int ret;
774 
775 	atomic_set(&aca->ue_update_flag, 0);
776 
777 	ret = aca_manager_init(&aca->mgr);
778 	if (ret)
779 		return ret;
780 
781 	return 0;
782 }
783 
784 void amdgpu_aca_fini(struct amdgpu_device *adev)
785 {
786 	struct amdgpu_aca *aca = &adev->aca;
787 
788 	aca_manager_fini(&aca->mgr);
789 
790 	atomic_set(&aca->ue_update_flag, 0);
791 }
792 
793 int amdgpu_aca_reset(struct amdgpu_device *adev)
794 {
795 	struct amdgpu_aca *aca = &adev->aca;
796 
797 	atomic_set(&aca->ue_update_flag, 0);
798 
799 	return 0;
800 }
801 
802 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
803 {
804 	struct amdgpu_aca *aca = &adev->aca;
805 
806 	WARN_ON(aca->smu_funcs);
807 	aca->smu_funcs = smu_funcs;
808 }
809 
810 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
811 {
812 	u64 ipid;
813 	u32 instidhi, instidlo;
814 
815 	if (!bank || !info)
816 		return -EINVAL;
817 
818 	ipid = bank->regs[ACA_REG_IDX_IPID];
819 	info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
820 	info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
821 	/*
822 	 * Unfied DieID Format: SAASS. A:AID, S:Socket.
823 	 * Unfied DieID[4:4] = InstanceId[0:0]
824 	 * Unfied DieID[0:3] = InstanceIdHi[0:3]
825 	 */
826 	instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
827 	instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
828 	info->die_id = ((instidhi >> 2) & 0x03);
829 	info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
830 
831 	return 0;
832 }
833 
834 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
835 {
836 	struct amdgpu_aca *aca = &adev->aca;
837 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
838 
839 	if (!smu_funcs || !smu_funcs->parse_error_code)
840 		return -EOPNOTSUPP;
841 
842 	return smu_funcs->parse_error_code(adev, bank);
843 }
844 
845 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
846 {
847 	int i, error_code;
848 
849 	if (!bank || !err_codes)
850 		return -EINVAL;
851 
852 	error_code = aca_bank_get_error_code(adev, bank);
853 	if (error_code < 0)
854 		return error_code;
855 
856 	for (i = 0; i < size; i++) {
857 		if (err_codes[i] == error_code)
858 			return 0;
859 	}
860 
861 	return -EINVAL;
862 }
863 
864 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
865 {
866 	struct amdgpu_aca *aca = &adev->aca;
867 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
868 
869 	if (!smu_funcs || !smu_funcs->set_debug_mode)
870 		return -EOPNOTSUPP;
871 
872 	return smu_funcs->set_debug_mode(adev, en);
873 }
874 
875 #if defined(CONFIG_DEBUG_FS)
876 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
877 {
878 	struct amdgpu_device *adev = (struct amdgpu_device *)data;
879 	int ret;
880 
881 	ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
882 	if (ret)
883 		return ret;
884 
885 	dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
886 
887 	return 0;
888 }
889 
890 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
891 {
892 	struct aca_bank_info info;
893 	int i, ret;
894 
895 	ret = aca_bank_info_decode(bank, &info);
896 	if (ret)
897 		return;
898 
899 	seq_printf(m, "aca entry[%d].type: %s\n", idx, type ==  ACA_SMU_TYPE_UE ? "UE" : "CE");
900 	seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
901 		   idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
902 
903 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
904 		seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
905 }
906 
907 struct aca_dump_context {
908 	struct seq_file *m;
909 	int idx;
910 };
911 
912 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
913 				 enum aca_smu_type type, void *data)
914 {
915 	struct aca_dump_context *ctx = (struct aca_dump_context *)data;
916 
917 	aca_dump_entry(ctx->m, bank, type, ctx->idx++);
918 
919 	return handler_aca_log_bank_error(handle, bank, type, NULL);
920 }
921 
922 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
923 {
924 	struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
925 	struct aca_dump_context context = {
926 		.m = m,
927 		.idx = 0,
928 	};
929 
930 	return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
931 }
932 
933 static int aca_dump_ce_show(struct seq_file *m, void *unused)
934 {
935 	return aca_dump_show(m, ACA_SMU_TYPE_CE);
936 }
937 
938 static int aca_dump_ce_open(struct inode *inode, struct file *file)
939 {
940 	return single_open(file, aca_dump_ce_show, inode->i_private);
941 }
942 
943 static const struct file_operations aca_ce_dump_debug_fops = {
944 	.owner = THIS_MODULE,
945 	.open = aca_dump_ce_open,
946 	.read = seq_read,
947 	.llseek = seq_lseek,
948 	.release = single_release,
949 };
950 
951 static int aca_dump_ue_show(struct seq_file *m, void *unused)
952 {
953 	return aca_dump_show(m, ACA_SMU_TYPE_UE);
954 }
955 
956 static int aca_dump_ue_open(struct inode *inode, struct file *file)
957 {
958 	return single_open(file, aca_dump_ue_show, inode->i_private);
959 }
960 
961 static const struct file_operations aca_ue_dump_debug_fops = {
962 	.owner = THIS_MODULE,
963 	.open = aca_dump_ue_open,
964 	.read = seq_read,
965 	.llseek = seq_lseek,
966 	.release = single_release,
967 };
968 
969 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
970 #endif
971 
972 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
973 {
974 #if defined(CONFIG_DEBUG_FS)
975 	if (!root)
976 		return;
977 
978 	debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
979 	debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
980 	debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
981 #endif
982 }
983