Loading...
Note: File does not exist in v3.1.
1/*
2 * Copyright 2023 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 */
23
24#include <linux/list.h>
25#include "amdgpu.h"
26#include "amdgpu_aca.h"
27#include "amdgpu_ras.h"
28
29#define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30
31typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_error_type type, void *data);
32
33struct aca_banks {
34 int nr_banks;
35 struct list_head list;
36};
37
38struct aca_hwip {
39 int hwid;
40 int mcatype;
41};
42
43static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
44 ACA_BANK_HWID(SMU, 0x01, 0x01),
45 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00),
46 ACA_BANK_HWID(UMC, 0x96, 0x00),
47};
48
49static void aca_banks_init(struct aca_banks *banks)
50{
51 if (!banks)
52 return;
53
54 memset(banks, 0, sizeof(*banks));
55 INIT_LIST_HEAD(&banks->list);
56}
57
58static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
59{
60 struct aca_bank_node *node;
61
62 if (!bank)
63 return -EINVAL;
64
65 node = kvzalloc(sizeof(*node), GFP_KERNEL);
66 if (!node)
67 return -ENOMEM;
68
69 memcpy(&node->bank, bank, sizeof(*bank));
70
71 INIT_LIST_HEAD(&node->node);
72 list_add_tail(&node->node, &banks->list);
73
74 banks->nr_banks++;
75
76 return 0;
77}
78
79static void aca_banks_release(struct aca_banks *banks)
80{
81 struct aca_bank_node *node, *tmp;
82
83 list_for_each_entry_safe(node, tmp, &banks->list, node) {
84 list_del(&node->node);
85 kvfree(node);
86 }
87}
88
89static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_error_type type, u32 *count)
90{
91 struct amdgpu_aca *aca = &adev->aca;
92 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
93
94 if (!count)
95 return -EINVAL;
96
97 if (!smu_funcs || !smu_funcs->get_valid_aca_count)
98 return -EOPNOTSUPP;
99
100 return smu_funcs->get_valid_aca_count(adev, type, count);
101}
102
103static struct aca_regs_dump {
104 const char *name;
105 int reg_idx;
106} aca_regs[] = {
107 {"CONTROL", ACA_REG_IDX_CTL},
108 {"STATUS", ACA_REG_IDX_STATUS},
109 {"ADDR", ACA_REG_IDX_ADDR},
110 {"MISC", ACA_REG_IDX_MISC0},
111 {"CONFIG", ACA_REG_IDX_CONFG},
112 {"IPID", ACA_REG_IDX_IPID},
113 {"SYND", ACA_REG_IDX_SYND},
114 {"DESTAT", ACA_REG_IDX_DESTAT},
115 {"DEADDR", ACA_REG_IDX_DEADDR},
116 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK},
117};
118
119static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank)
120{
121 int i;
122
123 dev_info(adev->dev, HW_ERR "Accelerator Check Architecture events logged\n");
124 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */
125 for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
126 dev_info(adev->dev, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
127 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
128}
129
130static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_error_type type,
131 int start, int count,
132 struct aca_banks *banks)
133{
134 struct amdgpu_aca *aca = &adev->aca;
135 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
136 struct aca_bank bank;
137 int i, max_count, ret;
138
139 if (!count)
140 return 0;
141
142 if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
143 return -EOPNOTSUPP;
144
145 switch (type) {
146 case ACA_ERROR_TYPE_UE:
147 max_count = smu_funcs->max_ue_bank_count;
148 break;
149 case ACA_ERROR_TYPE_CE:
150 max_count = smu_funcs->max_ce_bank_count;
151 break;
152 case ACA_ERROR_TYPE_DEFERRED:
153 default:
154 return -EINVAL;
155 }
156
157 if (start + count >= max_count)
158 return -EINVAL;
159
160 count = min_t(int, count, max_count);
161 for (i = 0; i < count; i++) {
162 memset(&bank, 0, sizeof(bank));
163 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
164 if (ret)
165 return ret;
166
167 aca_smu_bank_dump(adev, i, count, &bank);
168
169 ret = aca_banks_add_bank(banks, &bank);
170 if (ret)
171 return ret;
172 }
173
174 return 0;
175}
176
177static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
178{
179
180 struct aca_hwip *hwip;
181 int hwid, mcatype;
182 u64 ipid;
183
184 if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
185 return false;
186
187 hwip = &aca_hwid_mcatypes[type];
188 if (!hwip->hwid)
189 return false;
190
191 ipid = bank->regs[ACA_REG_IDX_IPID];
192 hwid = ACA_REG__IPID__HARDWAREID(ipid);
193 mcatype = ACA_REG__IPID__MCATYPE(ipid);
194
195 return hwip->hwid == hwid && hwip->mcatype == mcatype;
196}
197
198static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_error_type type)
199{
200 const struct aca_bank_ops *bank_ops = handle->bank_ops;
201
202 if (!aca_bank_hwip_is_matched(bank, handle->hwip))
203 return false;
204
205 if (!bank_ops->aca_bank_is_valid)
206 return true;
207
208 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
209}
210
211static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
212{
213 struct aca_bank_error *bank_error;
214
215 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
216 if (!bank_error)
217 return NULL;
218
219 INIT_LIST_HEAD(&bank_error->node);
220 memcpy(&bank_error->info, info, sizeof(*info));
221
222 mutex_lock(&aerr->lock);
223 list_add_tail(&bank_error->node, &aerr->list);
224 mutex_unlock(&aerr->lock);
225
226 return bank_error;
227}
228
229static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
230{
231 struct aca_bank_error *bank_error = NULL;
232 struct aca_bank_info *tmp_info;
233 bool found = false;
234
235 mutex_lock(&aerr->lock);
236 list_for_each_entry(bank_error, &aerr->list, node) {
237 tmp_info = &bank_error->info;
238 if (tmp_info->socket_id == info->socket_id &&
239 tmp_info->die_id == info->die_id) {
240 found = true;
241 goto out_unlock;
242 }
243 }
244
245out_unlock:
246 mutex_unlock(&aerr->lock);
247
248 return found ? bank_error : NULL;
249}
250
251static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
252{
253 if (!aerr || !bank_error)
254 return;
255
256 list_del(&bank_error->node);
257 aerr->nr_errors--;
258
259 kvfree(bank_error);
260}
261
262static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
263{
264 struct aca_bank_error *bank_error;
265
266 if (!aerr || !info)
267 return NULL;
268
269 bank_error = find_bank_error(aerr, info);
270 if (bank_error)
271 return bank_error;
272
273 return new_bank_error(aerr, info);
274}
275
276static int aca_log_errors(struct aca_handle *handle, enum aca_error_type type,
277 struct aca_bank_report *report)
278{
279 struct aca_error_cache *error_cache = &handle->error_cache;
280 struct aca_bank_error *bank_error;
281 struct aca_error *aerr;
282
283 if (!handle || !report)
284 return -EINVAL;
285
286 if (!report->count[type])
287 return 0;
288
289 aerr = &error_cache->errors[type];
290 bank_error = get_bank_error(aerr, &report->info);
291 if (!bank_error)
292 return -ENOMEM;
293
294 bank_error->count[type] += report->count[type];
295
296 return 0;
297}
298
299static int aca_generate_bank_report(struct aca_handle *handle, struct aca_bank *bank,
300 enum aca_error_type type, struct aca_bank_report *report)
301{
302 const struct aca_bank_ops *bank_ops = handle->bank_ops;
303
304 if (!bank || !report)
305 return -EINVAL;
306
307 if (!bank_ops->aca_bank_generate_report)
308 return -EOPNOTSUPP;
309
310 memset(report, 0, sizeof(*report));
311 return bank_ops->aca_bank_generate_report(handle, bank, type,
312 report, handle->data);
313}
314
315static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
316 enum aca_error_type type, void *data)
317{
318 struct aca_bank_report report;
319 int ret;
320
321 ret = aca_generate_bank_report(handle, bank, type, &report);
322 if (ret)
323 return ret;
324
325 if (!report.count[type])
326 return 0;
327
328 ret = aca_log_errors(handle, type, &report);
329 if (ret)
330 return ret;
331
332 return 0;
333}
334
335static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
336 enum aca_error_type type, bank_handler_t handler, void *data)
337{
338 struct aca_handle *handle;
339 int ret;
340
341 if (list_empty(&mgr->list))
342 return 0;
343
344 list_for_each_entry(handle, &mgr->list, node) {
345 if (!aca_bank_is_valid(handle, bank, type))
346 continue;
347
348 ret = handler(handle, bank, type, data);
349 if (ret)
350 return ret;
351 }
352
353 return 0;
354}
355
356static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
357 enum aca_error_type type, bank_handler_t handler, void *data)
358{
359 struct aca_bank_node *node;
360 struct aca_bank *bank;
361 int ret;
362
363 if (!mgr || !banks)
364 return -EINVAL;
365
366 /* pre check to avoid unnecessary operations */
367 if (list_empty(&mgr->list) || list_empty(&banks->list))
368 return 0;
369
370 list_for_each_entry(node, &banks->list, node) {
371 bank = &node->bank;
372
373 ret = aca_dispatch_bank(mgr, bank, type, handler, data);
374 if (ret)
375 return ret;
376 }
377
378 return 0;
379}
380
381static int aca_banks_update(struct amdgpu_device *adev, enum aca_error_type type,
382 bank_handler_t handler, void *data)
383{
384 struct amdgpu_aca *aca = &adev->aca;
385 struct aca_banks banks;
386 u32 count = 0;
387 int ret;
388
389 if (list_empty(&aca->mgr.list))
390 return 0;
391
392 /* NOTE: pmfw is only support UE and CE */
393 if (type == ACA_ERROR_TYPE_DEFERRED)
394 type = ACA_ERROR_TYPE_CE;
395
396 ret = aca_smu_get_valid_aca_count(adev, type, &count);
397 if (ret)
398 return ret;
399
400 if (!count)
401 return 0;
402
403 aca_banks_init(&banks);
404
405 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks);
406 if (ret)
407 goto err_release_banks;
408
409 if (list_empty(&banks.list)) {
410 ret = 0;
411 goto err_release_banks;
412 }
413
414 ret = aca_dispatch_banks(&aca->mgr, &banks, type,
415 handler, data);
416 if (ret)
417 goto err_release_banks;
418
419err_release_banks:
420 aca_banks_release(&banks);
421
422 return ret;
423}
424
425static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
426{
427 struct aca_bank_info *info;
428 struct amdgpu_smuio_mcm_config_info mcm_info;
429 u64 count;
430
431 if (type >= ACA_ERROR_TYPE_COUNT)
432 return -EINVAL;
433
434 count = bank_error->count[type];
435 if (!count)
436 return 0;
437
438 info = &bank_error->info;
439 mcm_info.die_id = info->die_id;
440 mcm_info.socket_id = info->socket_id;
441
442 switch (type) {
443 case ACA_ERROR_TYPE_UE:
444 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, NULL, count);
445 break;
446 case ACA_ERROR_TYPE_CE:
447 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, NULL, count);
448 break;
449 case ACA_ERROR_TYPE_DEFERRED:
450 default:
451 break;
452 }
453
454 return 0;
455}
456
457static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
458{
459 struct aca_error_cache *error_cache = &handle->error_cache;
460 struct aca_error *aerr = &error_cache->errors[type];
461 struct aca_bank_error *bank_error, *tmp;
462
463 mutex_lock(&aerr->lock);
464
465 if (list_empty(&aerr->list))
466 goto out_unlock;
467
468 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
469 aca_log_aca_error_data(bank_error, type, err_data);
470 aca_bank_error_remove(aerr, bank_error);
471 }
472
473out_unlock:
474 mutex_unlock(&aerr->lock);
475
476 return 0;
477}
478
479static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
480 struct ras_err_data *err_data)
481{
482 int ret;
483
484 /* udpate aca bank to aca source error_cache first */
485 ret = aca_banks_update(adev, type, handler_aca_log_bank_error, NULL);
486 if (ret)
487 return ret;
488
489 return aca_log_aca_error(handle, type, err_data);
490}
491
492static bool aca_handle_is_valid(struct aca_handle *handle)
493{
494 if (!handle->mask || !list_empty(&handle->node))
495 return false;
496
497 return true;
498}
499
500int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
501 enum aca_error_type type, void *data)
502{
503 struct ras_err_data *err_data = (struct ras_err_data *)data;
504
505 if (!handle || !err_data)
506 return -EINVAL;
507
508 if (aca_handle_is_valid(handle))
509 return -EOPNOTSUPP;
510
511 if (!(BIT(type) & handle->mask))
512 return 0;
513
514 return __aca_get_error_data(adev, handle, type, err_data);
515}
516
517static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
518{
519 mutex_init(&aerr->lock);
520 INIT_LIST_HEAD(&aerr->list);
521 aerr->type = type;
522 aerr->nr_errors = 0;
523}
524
525static void aca_init_error_cache(struct aca_handle *handle)
526{
527 struct aca_error_cache *error_cache = &handle->error_cache;
528 int type;
529
530 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
531 aca_error_init(&error_cache->errors[type], type);
532}
533
534static void aca_error_fini(struct aca_error *aerr)
535{
536 struct aca_bank_error *bank_error, *tmp;
537
538 mutex_lock(&aerr->lock);
539 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
540 aca_bank_error_remove(aerr, bank_error);
541
542 mutex_destroy(&aerr->lock);
543}
544
545static void aca_fini_error_cache(struct aca_handle *handle)
546{
547 struct aca_error_cache *error_cache = &handle->error_cache;
548 int type;
549
550 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
551 aca_error_fini(&error_cache->errors[type]);
552}
553
554static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
555 const char *name, const struct aca_info *ras_info, void *data)
556{
557 memset(handle, 0, sizeof(*handle));
558
559 handle->adev = adev;
560 handle->mgr = mgr;
561 handle->name = name;
562 handle->hwip = ras_info->hwip;
563 handle->mask = ras_info->mask;
564 handle->bank_ops = ras_info->bank_ops;
565 handle->data = data;
566 aca_init_error_cache(handle);
567
568 INIT_LIST_HEAD(&handle->node);
569 list_add_tail(&handle->node, &mgr->list);
570 mgr->nr_handles++;
571
572 return 0;
573}
574
575static ssize_t aca_sysfs_read(struct device *dev,
576 struct device_attribute *attr, char *buf)
577{
578 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
579
580 /* NOTE: the aca cache will be auto cleared once read,
581 * So the driver should unify the query entry point, forward request to ras query interface directly */
582 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
583}
584
585static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
586{
587 struct device_attribute *aca_attr = &handle->aca_attr;
588
589 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
590 aca_attr->show = aca_sysfs_read;
591 aca_attr->attr.name = handle->attr_name;
592 aca_attr->attr.mode = S_IRUGO;
593 sysfs_attr_init(&aca_attr->attr);
594
595 return sysfs_add_file_to_group(&adev->dev->kobj,
596 &aca_attr->attr,
597 "ras");
598}
599
600int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
601 const char *name, const struct aca_info *ras_info, void *data)
602{
603 struct amdgpu_aca *aca = &adev->aca;
604 int ret;
605
606 if (!amdgpu_aca_is_enabled(adev))
607 return 0;
608
609 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
610 if (ret)
611 return ret;
612
613 return add_aca_sysfs(adev, handle);
614}
615
616static void remove_aca_handle(struct aca_handle *handle)
617{
618 struct aca_handle_manager *mgr = handle->mgr;
619
620 aca_fini_error_cache(handle);
621 list_del(&handle->node);
622 mgr->nr_handles--;
623}
624
625static void remove_aca_sysfs(struct aca_handle *handle)
626{
627 struct amdgpu_device *adev = handle->adev;
628 struct device_attribute *aca_attr = &handle->aca_attr;
629
630 if (adev->dev->kobj.sd)
631 sysfs_remove_file_from_group(&adev->dev->kobj,
632 &aca_attr->attr,
633 "ras");
634}
635
636void amdgpu_aca_remove_handle(struct aca_handle *handle)
637{
638 if (!handle || list_empty(&handle->node))
639 return;
640
641 remove_aca_sysfs(handle);
642 remove_aca_handle(handle);
643}
644
645static int aca_manager_init(struct aca_handle_manager *mgr)
646{
647 INIT_LIST_HEAD(&mgr->list);
648 mgr->nr_handles = 0;
649
650 return 0;
651}
652
653static void aca_manager_fini(struct aca_handle_manager *mgr)
654{
655 struct aca_handle *handle, *tmp;
656
657 list_for_each_entry_safe(handle, tmp, &mgr->list, node)
658 amdgpu_aca_remove_handle(handle);
659}
660
661bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
662{
663 return adev->aca.is_enabled;
664}
665
666int amdgpu_aca_init(struct amdgpu_device *adev)
667{
668 struct amdgpu_aca *aca = &adev->aca;
669 int ret;
670
671 ret = aca_manager_init(&aca->mgr);
672 if (ret)
673 return ret;
674
675 return 0;
676}
677
678void amdgpu_aca_fini(struct amdgpu_device *adev)
679{
680 struct amdgpu_aca *aca = &adev->aca;
681
682 aca_manager_fini(&aca->mgr);
683}
684
685int amdgpu_aca_reset(struct amdgpu_device *adev)
686{
687 amdgpu_aca_fini(adev);
688
689 return amdgpu_aca_init(adev);
690}
691
692void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
693{
694 struct amdgpu_aca *aca = &adev->aca;
695
696 WARN_ON(aca->smu_funcs);
697 aca->smu_funcs = smu_funcs;
698}
699
700int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
701{
702 u64 ipid;
703 u32 instidhi, instidlo;
704
705 if (!bank || !info)
706 return -EINVAL;
707
708 ipid = bank->regs[ACA_REG_IDX_IPID];
709 info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
710 info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
711 /*
712 * Unfied DieID Format: SAASS. A:AID, S:Socket.
713 * Unfied DieID[4:4] = InstanceId[0:0]
714 * Unfied DieID[0:3] = InstanceIdHi[0:3]
715 */
716 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
717 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
718 info->die_id = ((instidhi >> 2) & 0x03);
719 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
720
721 return 0;
722}
723
724static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
725{
726 int error_code;
727
728 switch (amdgpu_ip_version(adev, MP1_HWIP, 0)) {
729 case IP_VERSION(13, 0, 6):
730 if (!(adev->flags & AMD_IS_APU) && adev->pm.fw_version >= 0x00555600) {
731 error_code = ACA_REG__SYND__ERRORINFORMATION(bank->regs[ACA_REG_IDX_SYND]);
732 return error_code & 0xff;
733 }
734 break;
735 default:
736 break;
737 }
738
739 /* NOTE: the true error code is encoded in status.errorcode[0:7] */
740 error_code = ACA_REG__STATUS__ERRORCODE(bank->regs[ACA_REG_IDX_STATUS]);
741
742 return error_code & 0xff;
743}
744
745int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
746{
747 int i, error_code;
748
749 if (!bank || !err_codes)
750 return -EINVAL;
751
752 error_code = aca_bank_get_error_code(adev, bank);
753 for (i = 0; i < size; i++) {
754 if (err_codes[i] == error_code)
755 return 0;
756 }
757
758 return -EINVAL;
759}
760
761int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
762{
763 struct amdgpu_aca *aca = &adev->aca;
764 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
765
766 if (!smu_funcs || !smu_funcs->set_debug_mode)
767 return -EOPNOTSUPP;
768
769 return smu_funcs->set_debug_mode(adev, en);
770}
771
772#if defined(CONFIG_DEBUG_FS)
773static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
774{
775 struct amdgpu_device *adev = (struct amdgpu_device *)data;
776 int ret;
777
778 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
779 if (ret)
780 return ret;
781
782 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
783
784 return 0;
785}
786
787static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_error_type type, int idx)
788{
789 struct aca_bank_info info;
790 int i, ret;
791
792 ret = aca_bank_info_decode(bank, &info);
793 if (ret)
794 return;
795
796 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_ERROR_TYPE_UE ? "UE" : "CE");
797 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
798 idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
799
800 for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
801 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
802}
803
804struct aca_dump_context {
805 struct seq_file *m;
806 int idx;
807};
808
809static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
810 enum aca_error_type type, void *data)
811{
812 struct aca_dump_context *ctx = (struct aca_dump_context *)data;
813
814 aca_dump_entry(ctx->m, bank, type, ctx->idx++);
815
816 return handler_aca_log_bank_error(handle, bank, type, NULL);
817}
818
819static int aca_dump_show(struct seq_file *m, enum aca_error_type type)
820{
821 struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
822 struct aca_dump_context context = {
823 .m = m,
824 .idx = 0,
825 };
826
827 return aca_banks_update(adev, type, handler_aca_bank_dump, (void *)&context);
828}
829
830static int aca_dump_ce_show(struct seq_file *m, void *unused)
831{
832 return aca_dump_show(m, ACA_ERROR_TYPE_CE);
833}
834
835static int aca_dump_ce_open(struct inode *inode, struct file *file)
836{
837 return single_open(file, aca_dump_ce_show, inode->i_private);
838}
839
840static const struct file_operations aca_ce_dump_debug_fops = {
841 .owner = THIS_MODULE,
842 .open = aca_dump_ce_open,
843 .read = seq_read,
844 .llseek = seq_lseek,
845 .release = single_release,
846};
847
848static int aca_dump_ue_show(struct seq_file *m, void *unused)
849{
850 return aca_dump_show(m, ACA_ERROR_TYPE_UE);
851}
852
853static int aca_dump_ue_open(struct inode *inode, struct file *file)
854{
855 return single_open(file, aca_dump_ue_show, inode->i_private);
856}
857
858static const struct file_operations aca_ue_dump_debug_fops = {
859 .owner = THIS_MODULE,
860 .open = aca_dump_ue_open,
861 .read = seq_read,
862 .llseek = seq_lseek,
863 .release = single_release,
864};
865
866DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
867#endif
868
869void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
870{
871#if defined(CONFIG_DEBUG_FS)
872 if (!root || adev->ip_versions[MP1_HWIP][0] != IP_VERSION(13, 0, 6))
873 return;
874
875 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
876 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
877 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
878#endif
879}