xref: /linux/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c (revision de848da12f752170c2ebe114804a985314fd5a6a)
1 /*
2  * Copyright 2023 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  */
23 
24 #include <linux/list.h>
25 #include "amdgpu.h"
26 #include "amdgpu_aca.h"
27 #include "amdgpu_ras.h"
28 
29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30 
31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
32 
33 struct aca_banks {
34 	int nr_banks;
35 	struct list_head list;
36 };
37 
38 struct aca_hwip {
39 	int hwid;
40 	int mcatype;
41 };
42 
43 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
44 	ACA_BANK_HWID(SMU,	0x01,	0x01),
45 	ACA_BANK_HWID(PCS_XGMI, 0x50,	0x00),
46 	ACA_BANK_HWID(UMC,	0x96,	0x00),
47 };
48 
49 static void aca_banks_init(struct aca_banks *banks)
50 {
51 	if (!banks)
52 		return;
53 
54 	memset(banks, 0, sizeof(*banks));
55 	INIT_LIST_HEAD(&banks->list);
56 }
57 
58 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
59 {
60 	struct aca_bank_node *node;
61 
62 	if (!bank)
63 		return -EINVAL;
64 
65 	node = kvzalloc(sizeof(*node), GFP_KERNEL);
66 	if (!node)
67 		return -ENOMEM;
68 
69 	memcpy(&node->bank, bank, sizeof(*bank));
70 
71 	INIT_LIST_HEAD(&node->node);
72 	list_add_tail(&node->node, &banks->list);
73 
74 	banks->nr_banks++;
75 
76 	return 0;
77 }
78 
79 static void aca_banks_release(struct aca_banks *banks)
80 {
81 	struct aca_bank_node *node, *tmp;
82 
83 	if (list_empty(&banks->list))
84 		return;
85 
86 	list_for_each_entry_safe(node, tmp, &banks->list, node) {
87 		list_del(&node->node);
88 		kvfree(node);
89 	}
90 }
91 
92 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
93 {
94 	struct amdgpu_aca *aca = &adev->aca;
95 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
96 
97 	if (!count)
98 		return -EINVAL;
99 
100 	if (!smu_funcs || !smu_funcs->get_valid_aca_count)
101 		return -EOPNOTSUPP;
102 
103 	return smu_funcs->get_valid_aca_count(adev, type, count);
104 }
105 
106 static struct aca_regs_dump {
107 	const char *name;
108 	int reg_idx;
109 } aca_regs[] = {
110 	{"CONTROL",		ACA_REG_IDX_CTL},
111 	{"STATUS",		ACA_REG_IDX_STATUS},
112 	{"ADDR",		ACA_REG_IDX_ADDR},
113 	{"MISC",		ACA_REG_IDX_MISC0},
114 	{"CONFIG",		ACA_REG_IDX_CONFG},
115 	{"IPID",		ACA_REG_IDX_IPID},
116 	{"SYND",		ACA_REG_IDX_SYND},
117 	{"DESTAT",		ACA_REG_IDX_DESTAT},
118 	{"DEADDR",		ACA_REG_IDX_DEADDR},
119 	{"CONTROL_MASK",	ACA_REG_IDX_CTL_MASK},
120 };
121 
122 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
123 			      struct ras_query_context *qctx)
124 {
125 	u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
126 	int i;
127 
128 	RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
129 	/* plus 1 for output format, e.g: ACA[08/08]: xxxx */
130 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
131 		RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
132 			      idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
133 }
134 
135 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
136 				       int start, int count,
137 				       struct aca_banks *banks, struct ras_query_context *qctx)
138 {
139 	struct amdgpu_aca *aca = &adev->aca;
140 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
141 	struct aca_bank bank;
142 	int i, max_count, ret;
143 
144 	if (!count)
145 		return 0;
146 
147 	if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
148 		return -EOPNOTSUPP;
149 
150 	switch (type) {
151 	case ACA_SMU_TYPE_UE:
152 		max_count = smu_funcs->max_ue_bank_count;
153 		break;
154 	case ACA_SMU_TYPE_CE:
155 		max_count = smu_funcs->max_ce_bank_count;
156 		break;
157 	default:
158 		return -EINVAL;
159 	}
160 
161 	if (start + count >= max_count)
162 		return -EINVAL;
163 
164 	count = min_t(int, count, max_count);
165 	for (i = 0; i < count; i++) {
166 		memset(&bank, 0, sizeof(bank));
167 		ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
168 		if (ret)
169 			return ret;
170 
171 		bank.type = type;
172 
173 		aca_smu_bank_dump(adev, i, count, &bank, qctx);
174 
175 		ret = aca_banks_add_bank(banks, &bank);
176 		if (ret)
177 			return ret;
178 	}
179 
180 	return 0;
181 }
182 
183 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
184 {
185 
186 	struct aca_hwip *hwip;
187 	int hwid, mcatype;
188 	u64 ipid;
189 
190 	if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
191 		return false;
192 
193 	hwip = &aca_hwid_mcatypes[type];
194 	if (!hwip->hwid)
195 		return false;
196 
197 	ipid = bank->regs[ACA_REG_IDX_IPID];
198 	hwid = ACA_REG__IPID__HARDWAREID(ipid);
199 	mcatype = ACA_REG__IPID__MCATYPE(ipid);
200 
201 	return hwip->hwid == hwid && hwip->mcatype == mcatype;
202 }
203 
204 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
205 {
206 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
207 
208 	if (!aca_bank_hwip_is_matched(bank, handle->hwip))
209 		return false;
210 
211 	if (!bank_ops->aca_bank_is_valid)
212 		return true;
213 
214 	return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
215 }
216 
217 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
218 {
219 	struct aca_bank_error *bank_error;
220 
221 	bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
222 	if (!bank_error)
223 		return NULL;
224 
225 	INIT_LIST_HEAD(&bank_error->node);
226 	memcpy(&bank_error->info, info, sizeof(*info));
227 
228 	mutex_lock(&aerr->lock);
229 	list_add_tail(&bank_error->node, &aerr->list);
230 	mutex_unlock(&aerr->lock);
231 
232 	return bank_error;
233 }
234 
235 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
236 {
237 	struct aca_bank_error *bank_error = NULL;
238 	struct aca_bank_info *tmp_info;
239 	bool found = false;
240 
241 	mutex_lock(&aerr->lock);
242 	list_for_each_entry(bank_error, &aerr->list, node) {
243 		tmp_info = &bank_error->info;
244 		if (tmp_info->socket_id == info->socket_id &&
245 		    tmp_info->die_id == info->die_id) {
246 			found = true;
247 			goto out_unlock;
248 		}
249 	}
250 
251 out_unlock:
252 	mutex_unlock(&aerr->lock);
253 
254 	return found ? bank_error : NULL;
255 }
256 
257 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
258 {
259 	if (!aerr || !bank_error)
260 		return;
261 
262 	list_del(&bank_error->node);
263 	aerr->nr_errors--;
264 
265 	kvfree(bank_error);
266 }
267 
268 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
269 {
270 	struct aca_bank_error *bank_error;
271 
272 	if (!aerr || !info)
273 		return NULL;
274 
275 	bank_error = find_bank_error(aerr, info);
276 	if (bank_error)
277 		return bank_error;
278 
279 	return new_bank_error(aerr, info);
280 }
281 
282 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
283 				   enum aca_error_type type, u64 count)
284 {
285 	struct aca_error_cache *error_cache = &handle->error_cache;
286 	struct aca_bank_error *bank_error;
287 	struct aca_error *aerr;
288 
289 	if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
290 		return -EINVAL;
291 
292 	if (!count)
293 		return 0;
294 
295 	aerr = &error_cache->errors[type];
296 	bank_error = get_bank_error(aerr, info);
297 	if (!bank_error)
298 		return -ENOMEM;
299 
300 	bank_error->count += count;
301 
302 	return 0;
303 }
304 
305 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
306 {
307 	const struct aca_bank_ops *bank_ops = handle->bank_ops;
308 
309 	if (!bank)
310 		return -EINVAL;
311 
312 	if (!bank_ops->aca_bank_parser)
313 		return -EOPNOTSUPP;
314 
315 	return bank_ops->aca_bank_parser(handle, bank, type,
316 					 handle->data);
317 }
318 
319 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
320 				      enum aca_smu_type type, void *data)
321 {
322 	int ret;
323 
324 	ret = aca_bank_parser(handle, bank, type);
325 	if (ret)
326 		return ret;
327 
328 	return 0;
329 }
330 
331 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
332 			     enum aca_smu_type type, bank_handler_t handler, void *data)
333 {
334 	struct aca_handle *handle;
335 	int ret;
336 
337 	if (list_empty(&mgr->list))
338 		return 0;
339 
340 	list_for_each_entry(handle, &mgr->list, node) {
341 		if (!aca_bank_is_valid(handle, bank, type))
342 			continue;
343 
344 		ret = handler(handle, bank, type, data);
345 		if (ret)
346 			return ret;
347 	}
348 
349 	return 0;
350 }
351 
352 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
353 			      enum aca_smu_type type, bank_handler_t handler, void *data)
354 {
355 	struct aca_bank_node *node;
356 	struct aca_bank *bank;
357 	int ret;
358 
359 	if (!mgr || !banks)
360 		return -EINVAL;
361 
362 	/* pre check to avoid unnecessary operations */
363 	if (list_empty(&mgr->list) || list_empty(&banks->list))
364 		return 0;
365 
366 	list_for_each_entry(node, &banks->list, node) {
367 		bank = &node->bank;
368 
369 		ret = aca_dispatch_bank(mgr, bank, type, handler, data);
370 		if (ret)
371 			return ret;
372 	}
373 
374 	return 0;
375 }
376 
377 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
378 {
379 	struct amdgpu_aca *aca = &adev->aca;
380 	bool ret = true;
381 
382 	/*
383 	 * Because the UE Valid MCA count will only be cleared after reset,
384 	 * in order to avoid repeated counting of the error count,
385 	 * the aca bank is only updated once during the gpu recovery stage.
386 	 */
387 	if (type == ACA_SMU_TYPE_UE) {
388 		if (amdgpu_ras_intr_triggered())
389 			ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
390 		else
391 			atomic_set(&aca->ue_update_flag, 0);
392 	}
393 
394 	return ret;
395 }
396 
397 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
398 			    bank_handler_t handler, struct ras_query_context *qctx, void *data)
399 {
400 	struct amdgpu_aca *aca = &adev->aca;
401 	struct aca_banks banks;
402 	u32 count = 0;
403 	int ret;
404 
405 	if (list_empty(&aca->mgr.list))
406 		return 0;
407 
408 	if (!aca_bank_should_update(adev, type))
409 		return 0;
410 
411 	ret = aca_smu_get_valid_aca_count(adev, type, &count);
412 	if (ret)
413 		return ret;
414 
415 	if (!count)
416 		return 0;
417 
418 	aca_banks_init(&banks);
419 
420 	ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
421 	if (ret)
422 		goto err_release_banks;
423 
424 	if (list_empty(&banks.list)) {
425 		ret = 0;
426 		goto err_release_banks;
427 	}
428 
429 	ret = aca_dispatch_banks(&aca->mgr, &banks, type,
430 				 handler, data);
431 	if (ret)
432 		goto err_release_banks;
433 
434 err_release_banks:
435 	aca_banks_release(&banks);
436 
437 	return ret;
438 }
439 
440 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
441 {
442 	struct aca_bank_info *info;
443 	struct amdgpu_smuio_mcm_config_info mcm_info;
444 	u64 count;
445 
446 	if (type >= ACA_ERROR_TYPE_COUNT)
447 		return -EINVAL;
448 
449 	count = bank_error->count;
450 	if (!count)
451 		return 0;
452 
453 	info = &bank_error->info;
454 	mcm_info.die_id = info->die_id;
455 	mcm_info.socket_id = info->socket_id;
456 
457 	switch (type) {
458 	case ACA_ERROR_TYPE_UE:
459 		amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);
460 		break;
461 	case ACA_ERROR_TYPE_CE:
462 		amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);
463 		break;
464 	case ACA_ERROR_TYPE_DEFERRED:
465 		amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);
466 		break;
467 	default:
468 		break;
469 	}
470 
471 	return 0;
472 }
473 
474 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
475 {
476 	struct aca_error_cache *error_cache = &handle->error_cache;
477 	struct aca_error *aerr = &error_cache->errors[type];
478 	struct aca_bank_error *bank_error, *tmp;
479 
480 	mutex_lock(&aerr->lock);
481 
482 	if (list_empty(&aerr->list))
483 		goto out_unlock;
484 
485 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
486 		aca_log_aca_error_data(bank_error, type, err_data);
487 		aca_bank_error_remove(aerr, bank_error);
488 	}
489 
490 out_unlock:
491 	mutex_unlock(&aerr->lock);
492 
493 	return 0;
494 }
495 
496 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
497 				struct ras_err_data *err_data, struct ras_query_context *qctx)
498 {
499 	enum aca_smu_type smu_type;
500 	int ret;
501 
502 	switch (type) {
503 	case ACA_ERROR_TYPE_UE:
504 		smu_type = ACA_SMU_TYPE_UE;
505 		break;
506 	case ACA_ERROR_TYPE_CE:
507 	case ACA_ERROR_TYPE_DEFERRED:
508 		smu_type = ACA_SMU_TYPE_CE;
509 		break;
510 	default:
511 		return -EINVAL;
512 	}
513 
514 	/* udpate aca bank to aca source error_cache first */
515 	ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
516 	if (ret)
517 		return ret;
518 
519 	return aca_log_aca_error(handle, type, err_data);
520 }
521 
522 static bool aca_handle_is_valid(struct aca_handle *handle)
523 {
524 	if (!handle->mask || !list_empty(&handle->node))
525 		return false;
526 
527 	return true;
528 }
529 
530 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
531 			      enum aca_error_type type, struct ras_err_data *err_data,
532 			      struct ras_query_context *qctx)
533 {
534 	if (!handle || !err_data)
535 		return -EINVAL;
536 
537 	if (aca_handle_is_valid(handle))
538 		return -EOPNOTSUPP;
539 
540 	if ((type < 0) || (!(BIT(type) & handle->mask)))
541 		return  0;
542 
543 	return __aca_get_error_data(adev, handle, type, err_data, qctx);
544 }
545 
546 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
547 {
548 	mutex_init(&aerr->lock);
549 	INIT_LIST_HEAD(&aerr->list);
550 	aerr->type = type;
551 	aerr->nr_errors = 0;
552 }
553 
554 static void aca_init_error_cache(struct aca_handle *handle)
555 {
556 	struct aca_error_cache *error_cache = &handle->error_cache;
557 	int type;
558 
559 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
560 		aca_error_init(&error_cache->errors[type], type);
561 }
562 
563 static void aca_error_fini(struct aca_error *aerr)
564 {
565 	struct aca_bank_error *bank_error, *tmp;
566 
567 	mutex_lock(&aerr->lock);
568 	if (list_empty(&aerr->list))
569 		goto out_unlock;
570 
571 	list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
572 		aca_bank_error_remove(aerr, bank_error);
573 
574 out_unlock:
575 	mutex_destroy(&aerr->lock);
576 }
577 
578 static void aca_fini_error_cache(struct aca_handle *handle)
579 {
580 	struct aca_error_cache *error_cache = &handle->error_cache;
581 	int type;
582 
583 	for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
584 		aca_error_fini(&error_cache->errors[type]);
585 }
586 
587 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
588 			  const char *name, const struct aca_info *ras_info, void *data)
589 {
590 	memset(handle, 0, sizeof(*handle));
591 
592 	handle->adev = adev;
593 	handle->mgr = mgr;
594 	handle->name = name;
595 	handle->hwip = ras_info->hwip;
596 	handle->mask = ras_info->mask;
597 	handle->bank_ops = ras_info->bank_ops;
598 	handle->data = data;
599 	aca_init_error_cache(handle);
600 
601 	INIT_LIST_HEAD(&handle->node);
602 	list_add_tail(&handle->node, &mgr->list);
603 	mgr->nr_handles++;
604 
605 	return 0;
606 }
607 
608 static ssize_t aca_sysfs_read(struct device *dev,
609 			      struct device_attribute *attr, char *buf)
610 {
611 	struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
612 
613 	/* NOTE: the aca cache will be auto cleared once read,
614 	 * So the driver should unify the query entry point, forward request to ras query interface directly */
615 	return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
616 }
617 
618 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
619 {
620 	struct device_attribute *aca_attr = &handle->aca_attr;
621 
622 	snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
623 	aca_attr->show = aca_sysfs_read;
624 	aca_attr->attr.name = handle->attr_name;
625 	aca_attr->attr.mode = S_IRUGO;
626 	sysfs_attr_init(&aca_attr->attr);
627 
628 	return sysfs_add_file_to_group(&adev->dev->kobj,
629 				       &aca_attr->attr,
630 				       "ras");
631 }
632 
633 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
634 			  const char *name, const struct aca_info *ras_info, void *data)
635 {
636 	struct amdgpu_aca *aca = &adev->aca;
637 	int ret;
638 
639 	if (!amdgpu_aca_is_enabled(adev))
640 		return 0;
641 
642 	ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
643 	if (ret)
644 		return ret;
645 
646 	return add_aca_sysfs(adev, handle);
647 }
648 
649 static void remove_aca_handle(struct aca_handle *handle)
650 {
651 	struct aca_handle_manager *mgr = handle->mgr;
652 
653 	aca_fini_error_cache(handle);
654 	list_del(&handle->node);
655 	mgr->nr_handles--;
656 }
657 
658 static void remove_aca_sysfs(struct aca_handle *handle)
659 {
660 	struct amdgpu_device *adev = handle->adev;
661 	struct device_attribute *aca_attr = &handle->aca_attr;
662 
663 	if (adev->dev->kobj.sd)
664 		sysfs_remove_file_from_group(&adev->dev->kobj,
665 					     &aca_attr->attr,
666 					     "ras");
667 }
668 
669 void amdgpu_aca_remove_handle(struct aca_handle *handle)
670 {
671 	if (!handle || list_empty(&handle->node))
672 		return;
673 
674 	remove_aca_sysfs(handle);
675 	remove_aca_handle(handle);
676 }
677 
678 static int aca_manager_init(struct aca_handle_manager *mgr)
679 {
680 	INIT_LIST_HEAD(&mgr->list);
681 	mgr->nr_handles = 0;
682 
683 	return 0;
684 }
685 
686 static void aca_manager_fini(struct aca_handle_manager *mgr)
687 {
688 	struct aca_handle *handle, *tmp;
689 
690 	if (list_empty(&mgr->list))
691 		return;
692 
693 	list_for_each_entry_safe(handle, tmp, &mgr->list, node)
694 		amdgpu_aca_remove_handle(handle);
695 }
696 
697 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
698 {
699 	return (adev->aca.is_enabled ||
700 		adev->debug_enable_ras_aca);
701 }
702 
703 int amdgpu_aca_init(struct amdgpu_device *adev)
704 {
705 	struct amdgpu_aca *aca = &adev->aca;
706 	int ret;
707 
708 	atomic_set(&aca->ue_update_flag, 0);
709 
710 	ret = aca_manager_init(&aca->mgr);
711 	if (ret)
712 		return ret;
713 
714 	return 0;
715 }
716 
717 void amdgpu_aca_fini(struct amdgpu_device *adev)
718 {
719 	struct amdgpu_aca *aca = &adev->aca;
720 
721 	aca_manager_fini(&aca->mgr);
722 
723 	atomic_set(&aca->ue_update_flag, 0);
724 }
725 
726 int amdgpu_aca_reset(struct amdgpu_device *adev)
727 {
728 	struct amdgpu_aca *aca = &adev->aca;
729 
730 	atomic_set(&aca->ue_update_flag, 0);
731 
732 	return 0;
733 }
734 
735 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
736 {
737 	struct amdgpu_aca *aca = &adev->aca;
738 
739 	WARN_ON(aca->smu_funcs);
740 	aca->smu_funcs = smu_funcs;
741 }
742 
743 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
744 {
745 	u64 ipid;
746 	u32 instidhi, instidlo;
747 
748 	if (!bank || !info)
749 		return -EINVAL;
750 
751 	ipid = bank->regs[ACA_REG_IDX_IPID];
752 	info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
753 	info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
754 	/*
755 	 * Unfied DieID Format: SAASS. A:AID, S:Socket.
756 	 * Unfied DieID[4:4] = InstanceId[0:0]
757 	 * Unfied DieID[0:3] = InstanceIdHi[0:3]
758 	 */
759 	instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
760 	instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
761 	info->die_id = ((instidhi >> 2) & 0x03);
762 	info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
763 
764 	return 0;
765 }
766 
767 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
768 {
769 	struct amdgpu_aca *aca = &adev->aca;
770 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
771 
772 	if (!smu_funcs || !smu_funcs->parse_error_code)
773 		return -EOPNOTSUPP;
774 
775 	return smu_funcs->parse_error_code(adev, bank);
776 }
777 
778 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
779 {
780 	int i, error_code;
781 
782 	if (!bank || !err_codes)
783 		return -EINVAL;
784 
785 	error_code = aca_bank_get_error_code(adev, bank);
786 	if (error_code < 0)
787 		return error_code;
788 
789 	for (i = 0; i < size; i++) {
790 		if (err_codes[i] == error_code)
791 			return 0;
792 	}
793 
794 	return -EINVAL;
795 }
796 
797 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
798 {
799 	struct amdgpu_aca *aca = &adev->aca;
800 	const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
801 
802 	if (!smu_funcs || !smu_funcs->set_debug_mode)
803 		return -EOPNOTSUPP;
804 
805 	return smu_funcs->set_debug_mode(adev, en);
806 }
807 
808 #if defined(CONFIG_DEBUG_FS)
809 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
810 {
811 	struct amdgpu_device *adev = (struct amdgpu_device *)data;
812 	int ret;
813 
814 	ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
815 	if (ret)
816 		return ret;
817 
818 	dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
819 
820 	return 0;
821 }
822 
823 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
824 {
825 	struct aca_bank_info info;
826 	int i, ret;
827 
828 	ret = aca_bank_info_decode(bank, &info);
829 	if (ret)
830 		return;
831 
832 	seq_printf(m, "aca entry[%d].type: %s\n", idx, type ==  ACA_SMU_TYPE_UE ? "UE" : "CE");
833 	seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
834 		   idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
835 
836 	for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
837 		seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
838 }
839 
840 struct aca_dump_context {
841 	struct seq_file *m;
842 	int idx;
843 };
844 
845 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
846 				 enum aca_smu_type type, void *data)
847 {
848 	struct aca_dump_context *ctx = (struct aca_dump_context *)data;
849 
850 	aca_dump_entry(ctx->m, bank, type, ctx->idx++);
851 
852 	return handler_aca_log_bank_error(handle, bank, type, NULL);
853 }
854 
855 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
856 {
857 	struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
858 	struct aca_dump_context context = {
859 		.m = m,
860 		.idx = 0,
861 	};
862 
863 	return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
864 }
865 
866 static int aca_dump_ce_show(struct seq_file *m, void *unused)
867 {
868 	return aca_dump_show(m, ACA_SMU_TYPE_CE);
869 }
870 
871 static int aca_dump_ce_open(struct inode *inode, struct file *file)
872 {
873 	return single_open(file, aca_dump_ce_show, inode->i_private);
874 }
875 
876 static const struct file_operations aca_ce_dump_debug_fops = {
877 	.owner = THIS_MODULE,
878 	.open = aca_dump_ce_open,
879 	.read = seq_read,
880 	.llseek = seq_lseek,
881 	.release = single_release,
882 };
883 
884 static int aca_dump_ue_show(struct seq_file *m, void *unused)
885 {
886 	return aca_dump_show(m, ACA_SMU_TYPE_UE);
887 }
888 
889 static int aca_dump_ue_open(struct inode *inode, struct file *file)
890 {
891 	return single_open(file, aca_dump_ue_show, inode->i_private);
892 }
893 
894 static const struct file_operations aca_ue_dump_debug_fops = {
895 	.owner = THIS_MODULE,
896 	.open = aca_dump_ue_open,
897 	.read = seq_read,
898 	.llseek = seq_lseek,
899 	.release = single_release,
900 };
901 
902 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
903 #endif
904 
905 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
906 {
907 #if defined(CONFIG_DEBUG_FS)
908 	if (!root)
909 		return;
910 
911 	debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
912 	debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
913 	debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
914 #endif
915 }
916