Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c
29285 views
1
/*
2
* Copyright 2023 Advanced Micro Devices, Inc.
3
*
4
* Permission is hereby granted, free of charge, to any person obtaining a
5
* copy of this software and associated documentation files (the "Software"),
6
* to deal in the Software without restriction, including without limitation
7
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
8
* and/or sell copies of the Software, and to permit persons to whom the
9
* Software is furnished to do so, subject to the following conditions:
10
*
11
* The above copyright notice and this permission notice shall be included in
12
* all copies or substantial portions of the Software.
13
*
14
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20
* OTHER DEALINGS IN THE SOFTWARE.
21
*
22
*/
23
24
#include <linux/list.h>
25
#include "amdgpu.h"
26
#include "amdgpu_aca.h"
27
#include "amdgpu_ras.h"
28
29
#define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30
31
typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
32
33
static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
34
ACA_BANK_HWID(SMU, 0x01, 0x01),
35
ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00),
36
ACA_BANK_HWID(UMC, 0x96, 0x00),
37
};
38
39
static void aca_banks_init(struct aca_banks *banks)
40
{
41
if (!banks)
42
return;
43
44
memset(banks, 0, sizeof(*banks));
45
INIT_LIST_HEAD(&banks->list);
46
}
47
48
static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
49
{
50
struct aca_bank_node *node;
51
52
if (!bank)
53
return -EINVAL;
54
55
node = kvzalloc(sizeof(*node), GFP_KERNEL);
56
if (!node)
57
return -ENOMEM;
58
59
memcpy(&node->bank, bank, sizeof(*bank));
60
61
INIT_LIST_HEAD(&node->node);
62
list_add_tail(&node->node, &banks->list);
63
64
banks->nr_banks++;
65
66
return 0;
67
}
68
69
static void aca_banks_release(struct aca_banks *banks)
70
{
71
struct aca_bank_node *node, *tmp;
72
73
if (list_empty(&banks->list))
74
return;
75
76
list_for_each_entry_safe(node, tmp, &banks->list, node) {
77
list_del(&node->node);
78
kvfree(node);
79
banks->nr_banks--;
80
}
81
}
82
83
static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
84
{
85
struct amdgpu_aca *aca = &adev->aca;
86
const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
87
88
if (!count)
89
return -EINVAL;
90
91
if (!smu_funcs || !smu_funcs->get_valid_aca_count)
92
return -EOPNOTSUPP;
93
94
return smu_funcs->get_valid_aca_count(adev, type, count);
95
}
96
97
static struct aca_regs_dump {
98
const char *name;
99
int reg_idx;
100
} aca_regs[] = {
101
{"CONTROL", ACA_REG_IDX_CTL},
102
{"STATUS", ACA_REG_IDX_STATUS},
103
{"ADDR", ACA_REG_IDX_ADDR},
104
{"MISC", ACA_REG_IDX_MISC0},
105
{"CONFIG", ACA_REG_IDX_CONFIG},
106
{"IPID", ACA_REG_IDX_IPID},
107
{"SYND", ACA_REG_IDX_SYND},
108
{"DESTAT", ACA_REG_IDX_DESTAT},
109
{"DEADDR", ACA_REG_IDX_DEADDR},
110
{"CONTROL_MASK", ACA_REG_IDX_CTL_MASK},
111
};
112
113
static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
114
struct ras_query_context *qctx)
115
{
116
u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
117
int i;
118
119
if (adev->debug_disable_ce_logs &&
120
bank->smu_err_type == ACA_SMU_TYPE_CE &&
121
!ACA_BANK_ERR_IS_DEFFERED(bank))
122
return;
123
124
RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
125
/* plus 1 for output format, e.g: ACA[08/08]: xxxx */
126
for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
127
RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
128
idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
129
130
if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS]))
131
RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n");
132
}
133
134
static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
135
{
136
137
struct aca_hwip *hwip;
138
int hwid, mcatype;
139
u64 ipid;
140
141
if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
142
return false;
143
144
hwip = &aca_hwid_mcatypes[type];
145
if (!hwip->hwid)
146
return false;
147
148
ipid = bank->regs[ACA_REG_IDX_IPID];
149
hwid = ACA_REG__IPID__HARDWAREID(ipid);
150
mcatype = ACA_REG__IPID__MCATYPE(ipid);
151
152
return hwip->hwid == hwid && hwip->mcatype == mcatype;
153
}
154
155
static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
156
int start, int count,
157
struct aca_banks *banks, struct ras_query_context *qctx)
158
{
159
struct amdgpu_aca *aca = &adev->aca;
160
const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
161
struct aca_bank bank;
162
int i, max_count, ret;
163
164
if (!count)
165
return 0;
166
167
if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
168
return -EOPNOTSUPP;
169
170
switch (type) {
171
case ACA_SMU_TYPE_UE:
172
max_count = smu_funcs->max_ue_bank_count;
173
break;
174
case ACA_SMU_TYPE_CE:
175
max_count = smu_funcs->max_ce_bank_count;
176
break;
177
default:
178
return -EINVAL;
179
}
180
181
if (start + count > max_count)
182
return -EINVAL;
183
184
count = min_t(int, count, max_count);
185
for (i = 0; i < count; i++) {
186
memset(&bank, 0, sizeof(bank));
187
ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
188
if (ret)
189
return ret;
190
191
bank.smu_err_type = type;
192
193
/*
194
* Poison being consumed when injecting a UE while running background workloads,
195
* which are unexpected.
196
*/
197
if (type == ACA_SMU_TYPE_UE &&
198
ACA_REG__STATUS__POISON(bank.regs[ACA_REG_IDX_STATUS]) &&
199
!aca_bank_hwip_is_matched(&bank, ACA_HWIP_TYPE_UMC))
200
continue;
201
202
aca_smu_bank_dump(adev, i, count, &bank, qctx);
203
204
ret = aca_banks_add_bank(banks, &bank);
205
if (ret)
206
return ret;
207
}
208
209
return 0;
210
}
211
212
static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
213
{
214
const struct aca_bank_ops *bank_ops = handle->bank_ops;
215
216
/* Parse all deferred errors with UMC aca handle */
217
if (ACA_BANK_ERR_IS_DEFFERED(bank))
218
return handle->hwip == ACA_HWIP_TYPE_UMC;
219
220
if (!aca_bank_hwip_is_matched(bank, handle->hwip))
221
return false;
222
223
if (!bank_ops->aca_bank_is_valid)
224
return true;
225
226
return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
227
}
228
229
static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
230
{
231
struct aca_bank_error *bank_error;
232
233
bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
234
if (!bank_error)
235
return NULL;
236
237
INIT_LIST_HEAD(&bank_error->node);
238
memcpy(&bank_error->info, info, sizeof(*info));
239
240
mutex_lock(&aerr->lock);
241
list_add_tail(&bank_error->node, &aerr->list);
242
aerr->nr_errors++;
243
mutex_unlock(&aerr->lock);
244
245
return bank_error;
246
}
247
248
static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
249
{
250
struct aca_bank_error *bank_error = NULL;
251
struct aca_bank_info *tmp_info;
252
bool found = false;
253
254
mutex_lock(&aerr->lock);
255
list_for_each_entry(bank_error, &aerr->list, node) {
256
tmp_info = &bank_error->info;
257
if (tmp_info->socket_id == info->socket_id &&
258
tmp_info->die_id == info->die_id) {
259
found = true;
260
goto out_unlock;
261
}
262
}
263
264
out_unlock:
265
mutex_unlock(&aerr->lock);
266
267
return found ? bank_error : NULL;
268
}
269
270
static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
271
{
272
if (!aerr || !bank_error)
273
return;
274
275
list_del(&bank_error->node);
276
aerr->nr_errors--;
277
278
kvfree(bank_error);
279
}
280
281
static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
282
{
283
struct aca_bank_error *bank_error;
284
285
if (!aerr || !info)
286
return NULL;
287
288
bank_error = find_bank_error(aerr, info);
289
if (bank_error)
290
return bank_error;
291
292
return new_bank_error(aerr, info);
293
}
294
295
int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
296
enum aca_error_type type, u64 count)
297
{
298
struct aca_error_cache *error_cache = &handle->error_cache;
299
struct aca_bank_error *bank_error;
300
struct aca_error *aerr;
301
302
if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
303
return -EINVAL;
304
305
if (!count)
306
return 0;
307
308
aerr = &error_cache->errors[type];
309
bank_error = get_bank_error(aerr, info);
310
if (!bank_error)
311
return -ENOMEM;
312
313
bank_error->count += count;
314
315
return 0;
316
}
317
318
static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
319
{
320
const struct aca_bank_ops *bank_ops = handle->bank_ops;
321
322
if (!bank)
323
return -EINVAL;
324
325
if (!bank_ops->aca_bank_parser)
326
return -EOPNOTSUPP;
327
328
return bank_ops->aca_bank_parser(handle, bank, type,
329
handle->data);
330
}
331
332
static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
333
enum aca_smu_type type, void *data)
334
{
335
int ret;
336
337
ret = aca_bank_parser(handle, bank, type);
338
if (ret)
339
return ret;
340
341
return 0;
342
}
343
344
static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
345
enum aca_smu_type type, bank_handler_t handler, void *data)
346
{
347
struct aca_handle *handle;
348
int ret;
349
350
if (list_empty(&mgr->list))
351
return 0;
352
353
list_for_each_entry(handle, &mgr->list, node) {
354
if (!aca_bank_is_valid(handle, bank, type))
355
continue;
356
357
ret = handler(handle, bank, type, data);
358
if (ret)
359
return ret;
360
}
361
362
return 0;
363
}
364
365
static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
366
enum aca_smu_type type, bank_handler_t handler, void *data)
367
{
368
struct aca_bank_node *node;
369
struct aca_bank *bank;
370
int ret;
371
372
if (!mgr || !banks)
373
return -EINVAL;
374
375
/* pre check to avoid unnecessary operations */
376
if (list_empty(&mgr->list) || list_empty(&banks->list))
377
return 0;
378
379
list_for_each_entry(node, &banks->list, node) {
380
bank = &node->bank;
381
382
ret = aca_dispatch_bank(mgr, bank, type, handler, data);
383
if (ret)
384
return ret;
385
}
386
387
return 0;
388
}
389
390
static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
391
{
392
struct amdgpu_aca *aca = &adev->aca;
393
bool ret = true;
394
395
/*
396
* Because the UE Valid MCA count will only be cleared after reset,
397
* in order to avoid repeated counting of the error count,
398
* the aca bank is only updated once during the gpu recovery stage.
399
*/
400
if (type == ACA_SMU_TYPE_UE) {
401
if (amdgpu_ras_intr_triggered())
402
ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
403
else
404
atomic_set(&aca->ue_update_flag, 0);
405
}
406
407
return ret;
408
}
409
410
static void aca_banks_generate_cper(struct amdgpu_device *adev,
411
enum aca_smu_type type,
412
struct aca_banks *banks,
413
int count)
414
{
415
struct aca_bank_node *node;
416
struct aca_bank *bank;
417
int r;
418
419
if (!adev->cper.enabled)
420
return;
421
422
if (!banks || !count) {
423
dev_warn(adev->dev, "fail to generate cper records\n");
424
return;
425
}
426
427
/* UEs must be encoded into separate CPER entries */
428
if (type == ACA_SMU_TYPE_UE) {
429
struct aca_banks de_banks;
430
431
aca_banks_init(&de_banks);
432
list_for_each_entry(node, &banks->list, node) {
433
bank = &node->bank;
434
if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {
435
r = aca_banks_add_bank(&de_banks, bank);
436
if (r)
437
dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);
438
} else {
439
if (amdgpu_cper_generate_ue_record(adev, bank))
440
dev_warn(adev->dev, "fail to generate ue cper records\n");
441
}
442
}
443
444
if (!list_empty(&de_banks.list)) {
445
if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))
446
dev_warn(adev->dev, "fail to generate de cper records\n");
447
}
448
449
aca_banks_release(&de_banks);
450
} else {
451
/*
452
* SMU_TYPE_CE banks are combined into 1 CPER entries,
453
* they could be CEs or DEs or both
454
*/
455
if (amdgpu_cper_generate_ce_records(adev, banks, count))
456
dev_warn(adev->dev, "fail to generate ce cper records\n");
457
}
458
}
459
460
static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
461
bank_handler_t handler, struct ras_query_context *qctx, void *data)
462
{
463
struct amdgpu_aca *aca = &adev->aca;
464
struct aca_banks banks;
465
u32 count = 0;
466
int ret;
467
468
if (list_empty(&aca->mgr.list))
469
return 0;
470
471
if (!aca_bank_should_update(adev, type))
472
return 0;
473
474
ret = aca_smu_get_valid_aca_count(adev, type, &count);
475
if (ret)
476
return ret;
477
478
if (!count)
479
return 0;
480
481
aca_banks_init(&banks);
482
483
ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
484
if (ret)
485
goto err_release_banks;
486
487
if (list_empty(&banks.list)) {
488
ret = 0;
489
goto err_release_banks;
490
}
491
492
ret = aca_dispatch_banks(&aca->mgr, &banks, type,
493
handler, data);
494
if (ret)
495
goto err_release_banks;
496
497
aca_banks_generate_cper(adev, type, &banks, count);
498
499
err_release_banks:
500
aca_banks_release(&banks);
501
502
return ret;
503
}
504
505
static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
506
{
507
struct aca_bank_info *info;
508
struct amdgpu_smuio_mcm_config_info mcm_info;
509
u64 count;
510
511
if (type >= ACA_ERROR_TYPE_COUNT)
512
return -EINVAL;
513
514
count = bank_error->count;
515
if (!count)
516
return 0;
517
518
info = &bank_error->info;
519
mcm_info.die_id = info->die_id;
520
mcm_info.socket_id = info->socket_id;
521
522
switch (type) {
523
case ACA_ERROR_TYPE_UE:
524
amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);
525
break;
526
case ACA_ERROR_TYPE_CE:
527
amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);
528
break;
529
case ACA_ERROR_TYPE_DEFERRED:
530
amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);
531
break;
532
default:
533
break;
534
}
535
536
return 0;
537
}
538
539
static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
540
{
541
struct aca_error_cache *error_cache = &handle->error_cache;
542
struct aca_error *aerr = &error_cache->errors[type];
543
struct aca_bank_error *bank_error, *tmp;
544
545
mutex_lock(&aerr->lock);
546
547
if (list_empty(&aerr->list))
548
goto out_unlock;
549
550
list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
551
aca_log_aca_error_data(bank_error, type, err_data);
552
aca_bank_error_remove(aerr, bank_error);
553
}
554
555
out_unlock:
556
mutex_unlock(&aerr->lock);
557
558
return 0;
559
}
560
561
static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
562
struct ras_err_data *err_data, struct ras_query_context *qctx)
563
{
564
enum aca_smu_type smu_type;
565
int ret;
566
567
switch (type) {
568
case ACA_ERROR_TYPE_UE:
569
smu_type = ACA_SMU_TYPE_UE;
570
break;
571
case ACA_ERROR_TYPE_CE:
572
case ACA_ERROR_TYPE_DEFERRED:
573
smu_type = ACA_SMU_TYPE_CE;
574
break;
575
default:
576
return -EINVAL;
577
}
578
579
/* update aca bank to aca source error_cache first */
580
ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
581
if (ret)
582
return ret;
583
584
/* DEs may contain in CEs or UEs */
585
if (type != ACA_ERROR_TYPE_DEFERRED)
586
aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);
587
588
return aca_log_aca_error(handle, type, err_data);
589
}
590
591
static bool aca_handle_is_valid(struct aca_handle *handle)
592
{
593
if (!handle->mask || !list_empty(&handle->node))
594
return false;
595
596
return true;
597
}
598
599
int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
600
enum aca_error_type type, struct ras_err_data *err_data,
601
struct ras_query_context *qctx)
602
{
603
if (!handle || !err_data)
604
return -EINVAL;
605
606
if (aca_handle_is_valid(handle))
607
return -EOPNOTSUPP;
608
609
if ((type < 0) || (!(BIT(type) & handle->mask)))
610
return 0;
611
612
return __aca_get_error_data(adev, handle, type, err_data, qctx);
613
}
614
615
static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
616
{
617
mutex_init(&aerr->lock);
618
INIT_LIST_HEAD(&aerr->list);
619
aerr->type = type;
620
aerr->nr_errors = 0;
621
}
622
623
static void aca_init_error_cache(struct aca_handle *handle)
624
{
625
struct aca_error_cache *error_cache = &handle->error_cache;
626
int type;
627
628
for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
629
aca_error_init(&error_cache->errors[type], type);
630
}
631
632
static void aca_error_fini(struct aca_error *aerr)
633
{
634
struct aca_bank_error *bank_error, *tmp;
635
636
mutex_lock(&aerr->lock);
637
if (list_empty(&aerr->list))
638
goto out_unlock;
639
640
list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
641
aca_bank_error_remove(aerr, bank_error);
642
643
out_unlock:
644
mutex_destroy(&aerr->lock);
645
}
646
647
static void aca_fini_error_cache(struct aca_handle *handle)
648
{
649
struct aca_error_cache *error_cache = &handle->error_cache;
650
int type;
651
652
for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
653
aca_error_fini(&error_cache->errors[type]);
654
}
655
656
static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
657
const char *name, const struct aca_info *ras_info, void *data)
658
{
659
memset(handle, 0, sizeof(*handle));
660
661
handle->adev = adev;
662
handle->mgr = mgr;
663
handle->name = name;
664
handle->hwip = ras_info->hwip;
665
handle->mask = ras_info->mask;
666
handle->bank_ops = ras_info->bank_ops;
667
handle->data = data;
668
aca_init_error_cache(handle);
669
670
INIT_LIST_HEAD(&handle->node);
671
list_add_tail(&handle->node, &mgr->list);
672
mgr->nr_handles++;
673
674
return 0;
675
}
676
677
static ssize_t aca_sysfs_read(struct device *dev,
678
struct device_attribute *attr, char *buf)
679
{
680
struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
681
682
/* NOTE: the aca cache will be auto cleared once read,
683
* So the driver should unify the query entry point, forward request to ras query interface directly */
684
return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
685
}
686
687
static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
688
{
689
struct device_attribute *aca_attr = &handle->aca_attr;
690
691
snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
692
aca_attr->show = aca_sysfs_read;
693
aca_attr->attr.name = handle->attr_name;
694
aca_attr->attr.mode = S_IRUGO;
695
sysfs_attr_init(&aca_attr->attr);
696
697
return sysfs_add_file_to_group(&adev->dev->kobj,
698
&aca_attr->attr,
699
"ras");
700
}
701
702
int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
703
const char *name, const struct aca_info *ras_info, void *data)
704
{
705
struct amdgpu_aca *aca = &adev->aca;
706
int ret;
707
708
if (!amdgpu_aca_is_enabled(adev))
709
return 0;
710
711
ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
712
if (ret)
713
return ret;
714
715
return add_aca_sysfs(adev, handle);
716
}
717
718
static void remove_aca_handle(struct aca_handle *handle)
719
{
720
struct aca_handle_manager *mgr = handle->mgr;
721
722
aca_fini_error_cache(handle);
723
list_del(&handle->node);
724
mgr->nr_handles--;
725
}
726
727
static void remove_aca_sysfs(struct aca_handle *handle)
728
{
729
struct amdgpu_device *adev = handle->adev;
730
struct device_attribute *aca_attr = &handle->aca_attr;
731
732
if (adev->dev->kobj.sd)
733
sysfs_remove_file_from_group(&adev->dev->kobj,
734
&aca_attr->attr,
735
"ras");
736
}
737
738
void amdgpu_aca_remove_handle(struct aca_handle *handle)
739
{
740
if (!handle || list_empty(&handle->node))
741
return;
742
743
remove_aca_sysfs(handle);
744
remove_aca_handle(handle);
745
}
746
747
static int aca_manager_init(struct aca_handle_manager *mgr)
748
{
749
INIT_LIST_HEAD(&mgr->list);
750
mgr->nr_handles = 0;
751
752
return 0;
753
}
754
755
static void aca_manager_fini(struct aca_handle_manager *mgr)
756
{
757
struct aca_handle *handle, *tmp;
758
759
if (list_empty(&mgr->list))
760
return;
761
762
list_for_each_entry_safe(handle, tmp, &mgr->list, node)
763
amdgpu_aca_remove_handle(handle);
764
}
765
766
bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
767
{
768
return (adev->aca.is_enabled ||
769
adev->debug_enable_ras_aca);
770
}
771
772
int amdgpu_aca_init(struct amdgpu_device *adev)
773
{
774
struct amdgpu_aca *aca = &adev->aca;
775
int ret;
776
777
atomic_set(&aca->ue_update_flag, 0);
778
779
ret = aca_manager_init(&aca->mgr);
780
if (ret)
781
return ret;
782
783
return 0;
784
}
785
786
void amdgpu_aca_fini(struct amdgpu_device *adev)
787
{
788
struct amdgpu_aca *aca = &adev->aca;
789
790
aca_manager_fini(&aca->mgr);
791
792
atomic_set(&aca->ue_update_flag, 0);
793
}
794
795
int amdgpu_aca_reset(struct amdgpu_device *adev)
796
{
797
struct amdgpu_aca *aca = &adev->aca;
798
799
atomic_set(&aca->ue_update_flag, 0);
800
801
return 0;
802
}
803
804
void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
805
{
806
struct amdgpu_aca *aca = &adev->aca;
807
808
WARN_ON(aca->smu_funcs);
809
aca->smu_funcs = smu_funcs;
810
}
811
812
int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
813
{
814
u64 ipid;
815
u32 instidhi, instidlo;
816
817
if (!bank || !info)
818
return -EINVAL;
819
820
ipid = bank->regs[ACA_REG_IDX_IPID];
821
info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
822
info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
823
/*
824
* Unfied DieID Format: SAASS. A:AID, S:Socket.
825
* Unfied DieID[4:4] = InstanceId[0:0]
826
* Unfied DieID[0:3] = InstanceIdHi[0:3]
827
*/
828
instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
829
instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
830
info->die_id = ((instidhi >> 2) & 0x03);
831
info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
832
833
return 0;
834
}
835
836
static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
837
{
838
struct amdgpu_aca *aca = &adev->aca;
839
const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
840
841
if (!smu_funcs || !smu_funcs->parse_error_code)
842
return -EOPNOTSUPP;
843
844
return smu_funcs->parse_error_code(adev, bank);
845
}
846
847
int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
848
{
849
int i, error_code;
850
851
if (!bank || !err_codes)
852
return -EINVAL;
853
854
error_code = aca_bank_get_error_code(adev, bank);
855
if (error_code < 0)
856
return error_code;
857
858
for (i = 0; i < size; i++) {
859
if (err_codes[i] == error_code)
860
return 0;
861
}
862
863
return -EINVAL;
864
}
865
866
int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
867
{
868
struct amdgpu_aca *aca = &adev->aca;
869
const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
870
871
if (!smu_funcs || !smu_funcs->set_debug_mode)
872
return -EOPNOTSUPP;
873
874
return smu_funcs->set_debug_mode(adev, en);
875
}
876
877
#if defined(CONFIG_DEBUG_FS)
878
static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
879
{
880
struct amdgpu_device *adev = (struct amdgpu_device *)data;
881
int ret;
882
883
ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
884
if (ret)
885
return ret;
886
887
dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
888
889
return 0;
890
}
891
892
static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
893
{
894
struct aca_bank_info info;
895
int i, ret;
896
897
ret = aca_bank_info_decode(bank, &info);
898
if (ret)
899
return;
900
901
seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE");
902
seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
903
idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
904
905
for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
906
seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
907
}
908
909
struct aca_dump_context {
910
struct seq_file *m;
911
int idx;
912
};
913
914
static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
915
enum aca_smu_type type, void *data)
916
{
917
struct aca_dump_context *ctx = (struct aca_dump_context *)data;
918
919
aca_dump_entry(ctx->m, bank, type, ctx->idx++);
920
921
return handler_aca_log_bank_error(handle, bank, type, NULL);
922
}
923
924
static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
925
{
926
struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
927
struct aca_dump_context context = {
928
.m = m,
929
.idx = 0,
930
};
931
932
return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
933
}
934
935
static int aca_dump_ce_show(struct seq_file *m, void *unused)
936
{
937
return aca_dump_show(m, ACA_SMU_TYPE_CE);
938
}
939
940
static int aca_dump_ce_open(struct inode *inode, struct file *file)
941
{
942
return single_open(file, aca_dump_ce_show, inode->i_private);
943
}
944
945
static const struct file_operations aca_ce_dump_debug_fops = {
946
.owner = THIS_MODULE,
947
.open = aca_dump_ce_open,
948
.read = seq_read,
949
.llseek = seq_lseek,
950
.release = single_release,
951
};
952
953
static int aca_dump_ue_show(struct seq_file *m, void *unused)
954
{
955
return aca_dump_show(m, ACA_SMU_TYPE_UE);
956
}
957
958
static int aca_dump_ue_open(struct inode *inode, struct file *file)
959
{
960
return single_open(file, aca_dump_ue_show, inode->i_private);
961
}
962
963
static const struct file_operations aca_ue_dump_debug_fops = {
964
.owner = THIS_MODULE,
965
.open = aca_dump_ue_open,
966
.read = seq_read,
967
.llseek = seq_lseek,
968
.release = single_release,
969
};
970
971
DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
972
#endif
973
974
void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
975
{
976
#if defined(CONFIG_DEBUG_FS)
977
if (!root)
978
return;
979
980
debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
981
debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
982
debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
983
#endif
984
}
985
986