1 /*
2 * Copyright 2023 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 */
23
24 #include <linux/list.h>
25 #include "amdgpu.h"
26 #include "amdgpu_aca.h"
27 #include "amdgpu_ras.h"
28
29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
30
31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
32
33 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
34 ACA_BANK_HWID(SMU, 0x01, 0x01),
35 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00),
36 ACA_BANK_HWID(UMC, 0x96, 0x00),
37 };
38
aca_banks_init(struct aca_banks * banks)39 static void aca_banks_init(struct aca_banks *banks)
40 {
41 if (!banks)
42 return;
43
44 memset(banks, 0, sizeof(*banks));
45 INIT_LIST_HEAD(&banks->list);
46 }
47
aca_banks_add_bank(struct aca_banks * banks,struct aca_bank * bank)48 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
49 {
50 struct aca_bank_node *node;
51
52 if (!bank)
53 return -EINVAL;
54
55 node = kvzalloc(sizeof(*node), GFP_KERNEL);
56 if (!node)
57 return -ENOMEM;
58
59 memcpy(&node->bank, bank, sizeof(*bank));
60
61 INIT_LIST_HEAD(&node->node);
62 list_add_tail(&node->node, &banks->list);
63
64 banks->nr_banks++;
65
66 return 0;
67 }
68
aca_banks_release(struct aca_banks * banks)69 static void aca_banks_release(struct aca_banks *banks)
70 {
71 struct aca_bank_node *node, *tmp;
72
73 if (list_empty(&banks->list))
74 return;
75
76 list_for_each_entry_safe(node, tmp, &banks->list, node) {
77 list_del(&node->node);
78 kvfree(node);
79 }
80 }
81
aca_smu_get_valid_aca_count(struct amdgpu_device * adev,enum aca_smu_type type,u32 * count)82 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
83 {
84 struct amdgpu_aca *aca = &adev->aca;
85 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
86
87 if (!count)
88 return -EINVAL;
89
90 if (!smu_funcs || !smu_funcs->get_valid_aca_count)
91 return -EOPNOTSUPP;
92
93 return smu_funcs->get_valid_aca_count(adev, type, count);
94 }
95
96 static struct aca_regs_dump {
97 const char *name;
98 int reg_idx;
99 } aca_regs[] = {
100 {"CONTROL", ACA_REG_IDX_CTL},
101 {"STATUS", ACA_REG_IDX_STATUS},
102 {"ADDR", ACA_REG_IDX_ADDR},
103 {"MISC", ACA_REG_IDX_MISC0},
104 {"CONFIG", ACA_REG_IDX_CONFIG},
105 {"IPID", ACA_REG_IDX_IPID},
106 {"SYND", ACA_REG_IDX_SYND},
107 {"DESTAT", ACA_REG_IDX_DESTAT},
108 {"DEADDR", ACA_REG_IDX_DEADDR},
109 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK},
110 };
111
aca_smu_bank_dump(struct amdgpu_device * adev,int idx,int total,struct aca_bank * bank,struct ras_query_context * qctx)112 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
113 struct ras_query_context *qctx)
114 {
115 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
116 int i;
117
118 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
119 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */
120 for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
121 RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
122 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
123 }
124
aca_smu_get_valid_aca_banks(struct amdgpu_device * adev,enum aca_smu_type type,int start,int count,struct aca_banks * banks,struct ras_query_context * qctx)125 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
126 int start, int count,
127 struct aca_banks *banks, struct ras_query_context *qctx)
128 {
129 struct amdgpu_aca *aca = &adev->aca;
130 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
131 struct aca_bank bank;
132 int i, max_count, ret;
133
134 if (!count)
135 return 0;
136
137 if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
138 return -EOPNOTSUPP;
139
140 switch (type) {
141 case ACA_SMU_TYPE_UE:
142 max_count = smu_funcs->max_ue_bank_count;
143 break;
144 case ACA_SMU_TYPE_CE:
145 max_count = smu_funcs->max_ce_bank_count;
146 break;
147 default:
148 return -EINVAL;
149 }
150
151 if (start + count > max_count)
152 return -EINVAL;
153
154 count = min_t(int, count, max_count);
155 for (i = 0; i < count; i++) {
156 memset(&bank, 0, sizeof(bank));
157 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
158 if (ret)
159 return ret;
160
161 bank.smu_err_type = type;
162
163 aca_smu_bank_dump(adev, i, count, &bank, qctx);
164
165 ret = aca_banks_add_bank(banks, &bank);
166 if (ret)
167 return ret;
168 }
169
170 return 0;
171 }
172
aca_bank_hwip_is_matched(struct aca_bank * bank,enum aca_hwip_type type)173 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
174 {
175
176 struct aca_hwip *hwip;
177 int hwid, mcatype;
178 u64 ipid;
179
180 if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
181 return false;
182
183 hwip = &aca_hwid_mcatypes[type];
184 if (!hwip->hwid)
185 return false;
186
187 ipid = bank->regs[ACA_REG_IDX_IPID];
188 hwid = ACA_REG__IPID__HARDWAREID(ipid);
189 mcatype = ACA_REG__IPID__MCATYPE(ipid);
190
191 return hwip->hwid == hwid && hwip->mcatype == mcatype;
192 }
193
aca_bank_is_valid(struct aca_handle * handle,struct aca_bank * bank,enum aca_smu_type type)194 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
195 {
196 const struct aca_bank_ops *bank_ops = handle->bank_ops;
197
198 /* Parse all deferred errors with UMC aca handle */
199 if (ACA_BANK_ERR_IS_DEFFERED(bank))
200 return handle->hwip == ACA_HWIP_TYPE_UMC;
201
202 if (!aca_bank_hwip_is_matched(bank, handle->hwip))
203 return false;
204
205 if (!bank_ops->aca_bank_is_valid)
206 return true;
207
208 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
209 }
210
new_bank_error(struct aca_error * aerr,struct aca_bank_info * info)211 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
212 {
213 struct aca_bank_error *bank_error;
214
215 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
216 if (!bank_error)
217 return NULL;
218
219 INIT_LIST_HEAD(&bank_error->node);
220 memcpy(&bank_error->info, info, sizeof(*info));
221
222 mutex_lock(&aerr->lock);
223 list_add_tail(&bank_error->node, &aerr->list);
224 mutex_unlock(&aerr->lock);
225
226 return bank_error;
227 }
228
find_bank_error(struct aca_error * aerr,struct aca_bank_info * info)229 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
230 {
231 struct aca_bank_error *bank_error = NULL;
232 struct aca_bank_info *tmp_info;
233 bool found = false;
234
235 mutex_lock(&aerr->lock);
236 list_for_each_entry(bank_error, &aerr->list, node) {
237 tmp_info = &bank_error->info;
238 if (tmp_info->socket_id == info->socket_id &&
239 tmp_info->die_id == info->die_id) {
240 found = true;
241 goto out_unlock;
242 }
243 }
244
245 out_unlock:
246 mutex_unlock(&aerr->lock);
247
248 return found ? bank_error : NULL;
249 }
250
aca_bank_error_remove(struct aca_error * aerr,struct aca_bank_error * bank_error)251 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
252 {
253 if (!aerr || !bank_error)
254 return;
255
256 list_del(&bank_error->node);
257 aerr->nr_errors--;
258
259 kvfree(bank_error);
260 }
261
get_bank_error(struct aca_error * aerr,struct aca_bank_info * info)262 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
263 {
264 struct aca_bank_error *bank_error;
265
266 if (!aerr || !info)
267 return NULL;
268
269 bank_error = find_bank_error(aerr, info);
270 if (bank_error)
271 return bank_error;
272
273 return new_bank_error(aerr, info);
274 }
275
aca_error_cache_log_bank_error(struct aca_handle * handle,struct aca_bank_info * info,enum aca_error_type type,u64 count)276 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
277 enum aca_error_type type, u64 count)
278 {
279 struct aca_error_cache *error_cache = &handle->error_cache;
280 struct aca_bank_error *bank_error;
281 struct aca_error *aerr;
282
283 if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
284 return -EINVAL;
285
286 if (!count)
287 return 0;
288
289 aerr = &error_cache->errors[type];
290 bank_error = get_bank_error(aerr, info);
291 if (!bank_error)
292 return -ENOMEM;
293
294 bank_error->count += count;
295
296 return 0;
297 }
298
aca_bank_parser(struct aca_handle * handle,struct aca_bank * bank,enum aca_smu_type type)299 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
300 {
301 const struct aca_bank_ops *bank_ops = handle->bank_ops;
302
303 if (!bank)
304 return -EINVAL;
305
306 if (!bank_ops->aca_bank_parser)
307 return -EOPNOTSUPP;
308
309 return bank_ops->aca_bank_parser(handle, bank, type,
310 handle->data);
311 }
312
handler_aca_log_bank_error(struct aca_handle * handle,struct aca_bank * bank,enum aca_smu_type type,void * data)313 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
314 enum aca_smu_type type, void *data)
315 {
316 int ret;
317
318 ret = aca_bank_parser(handle, bank, type);
319 if (ret)
320 return ret;
321
322 return 0;
323 }
324
aca_dispatch_bank(struct aca_handle_manager * mgr,struct aca_bank * bank,enum aca_smu_type type,bank_handler_t handler,void * data)325 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
326 enum aca_smu_type type, bank_handler_t handler, void *data)
327 {
328 struct aca_handle *handle;
329 int ret;
330
331 if (list_empty(&mgr->list))
332 return 0;
333
334 list_for_each_entry(handle, &mgr->list, node) {
335 if (!aca_bank_is_valid(handle, bank, type))
336 continue;
337
338 ret = handler(handle, bank, type, data);
339 if (ret)
340 return ret;
341 }
342
343 return 0;
344 }
345
aca_dispatch_banks(struct aca_handle_manager * mgr,struct aca_banks * banks,enum aca_smu_type type,bank_handler_t handler,void * data)346 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
347 enum aca_smu_type type, bank_handler_t handler, void *data)
348 {
349 struct aca_bank_node *node;
350 struct aca_bank *bank;
351 int ret;
352
353 if (!mgr || !banks)
354 return -EINVAL;
355
356 /* pre check to avoid unnecessary operations */
357 if (list_empty(&mgr->list) || list_empty(&banks->list))
358 return 0;
359
360 list_for_each_entry(node, &banks->list, node) {
361 bank = &node->bank;
362
363 ret = aca_dispatch_bank(mgr, bank, type, handler, data);
364 if (ret)
365 return ret;
366 }
367
368 return 0;
369 }
370
aca_bank_should_update(struct amdgpu_device * adev,enum aca_smu_type type)371 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
372 {
373 struct amdgpu_aca *aca = &adev->aca;
374 bool ret = true;
375
376 /*
377 * Because the UE Valid MCA count will only be cleared after reset,
378 * in order to avoid repeated counting of the error count,
379 * the aca bank is only updated once during the gpu recovery stage.
380 */
381 if (type == ACA_SMU_TYPE_UE) {
382 if (amdgpu_ras_intr_triggered())
383 ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
384 else
385 atomic_set(&aca->ue_update_flag, 0);
386 }
387
388 return ret;
389 }
390
aca_banks_generate_cper(struct amdgpu_device * adev,enum aca_smu_type type,struct aca_banks * banks,int count)391 static void aca_banks_generate_cper(struct amdgpu_device *adev,
392 enum aca_smu_type type,
393 struct aca_banks *banks,
394 int count)
395 {
396 struct aca_bank_node *node;
397 struct aca_bank *bank;
398 int r;
399
400 if (!adev->cper.enabled)
401 return;
402
403 if (!banks || !count) {
404 dev_warn(adev->dev, "fail to generate cper records\n");
405 return;
406 }
407
408 /* UEs must be encoded into separate CPER entries */
409 if (type == ACA_SMU_TYPE_UE) {
410 struct aca_banks de_banks;
411
412 aca_banks_init(&de_banks);
413 list_for_each_entry(node, &banks->list, node) {
414 bank = &node->bank;
415 if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {
416 r = aca_banks_add_bank(&de_banks, bank);
417 if (r)
418 dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);
419 } else {
420 if (amdgpu_cper_generate_ue_record(adev, bank))
421 dev_warn(adev->dev, "fail to generate ue cper records\n");
422 }
423 }
424
425 if (!list_empty(&de_banks.list)) {
426 if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))
427 dev_warn(adev->dev, "fail to generate de cper records\n");
428 }
429
430 aca_banks_release(&de_banks);
431 } else {
432 /*
433 * SMU_TYPE_CE banks are combined into 1 CPER entries,
434 * they could be CEs or DEs or both
435 */
436 if (amdgpu_cper_generate_ce_records(adev, banks, count))
437 dev_warn(adev->dev, "fail to generate ce cper records\n");
438 }
439 }
440
aca_banks_update(struct amdgpu_device * adev,enum aca_smu_type type,bank_handler_t handler,struct ras_query_context * qctx,void * data)441 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
442 bank_handler_t handler, struct ras_query_context *qctx, void *data)
443 {
444 struct amdgpu_aca *aca = &adev->aca;
445 struct aca_banks banks;
446 u32 count = 0;
447 int ret;
448
449 if (list_empty(&aca->mgr.list))
450 return 0;
451
452 if (!aca_bank_should_update(adev, type))
453 return 0;
454
455 ret = aca_smu_get_valid_aca_count(adev, type, &count);
456 if (ret)
457 return ret;
458
459 if (!count)
460 return 0;
461
462 aca_banks_init(&banks);
463
464 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
465 if (ret)
466 goto err_release_banks;
467
468 if (list_empty(&banks.list)) {
469 ret = 0;
470 goto err_release_banks;
471 }
472
473 ret = aca_dispatch_banks(&aca->mgr, &banks, type,
474 handler, data);
475 if (ret)
476 goto err_release_banks;
477
478 aca_banks_generate_cper(adev, type, &banks, count);
479
480 err_release_banks:
481 aca_banks_release(&banks);
482
483 return ret;
484 }
485
aca_log_aca_error_data(struct aca_bank_error * bank_error,enum aca_error_type type,struct ras_err_data * err_data)486 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
487 {
488 struct aca_bank_info *info;
489 struct amdgpu_smuio_mcm_config_info mcm_info;
490 u64 count;
491
492 if (type >= ACA_ERROR_TYPE_COUNT)
493 return -EINVAL;
494
495 count = bank_error->count;
496 if (!count)
497 return 0;
498
499 info = &bank_error->info;
500 mcm_info.die_id = info->die_id;
501 mcm_info.socket_id = info->socket_id;
502
503 switch (type) {
504 case ACA_ERROR_TYPE_UE:
505 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);
506 break;
507 case ACA_ERROR_TYPE_CE:
508 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);
509 break;
510 case ACA_ERROR_TYPE_DEFERRED:
511 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);
512 break;
513 default:
514 break;
515 }
516
517 return 0;
518 }
519
aca_log_aca_error(struct aca_handle * handle,enum aca_error_type type,struct ras_err_data * err_data)520 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
521 {
522 struct aca_error_cache *error_cache = &handle->error_cache;
523 struct aca_error *aerr = &error_cache->errors[type];
524 struct aca_bank_error *bank_error, *tmp;
525
526 mutex_lock(&aerr->lock);
527
528 if (list_empty(&aerr->list))
529 goto out_unlock;
530
531 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
532 aca_log_aca_error_data(bank_error, type, err_data);
533 aca_bank_error_remove(aerr, bank_error);
534 }
535
536 out_unlock:
537 mutex_unlock(&aerr->lock);
538
539 return 0;
540 }
541
__aca_get_error_data(struct amdgpu_device * adev,struct aca_handle * handle,enum aca_error_type type,struct ras_err_data * err_data,struct ras_query_context * qctx)542 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
543 struct ras_err_data *err_data, struct ras_query_context *qctx)
544 {
545 enum aca_smu_type smu_type;
546 int ret;
547
548 switch (type) {
549 case ACA_ERROR_TYPE_UE:
550 smu_type = ACA_SMU_TYPE_UE;
551 break;
552 case ACA_ERROR_TYPE_CE:
553 case ACA_ERROR_TYPE_DEFERRED:
554 smu_type = ACA_SMU_TYPE_CE;
555 break;
556 default:
557 return -EINVAL;
558 }
559
560 /* update aca bank to aca source error_cache first */
561 ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
562 if (ret)
563 return ret;
564
565 /* DEs may contain in CEs or UEs */
566 if (type != ACA_ERROR_TYPE_DEFERRED)
567 aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);
568
569 return aca_log_aca_error(handle, type, err_data);
570 }
571
aca_handle_is_valid(struct aca_handle * handle)572 static bool aca_handle_is_valid(struct aca_handle *handle)
573 {
574 if (!handle->mask || !list_empty(&handle->node))
575 return false;
576
577 return true;
578 }
579
amdgpu_aca_get_error_data(struct amdgpu_device * adev,struct aca_handle * handle,enum aca_error_type type,struct ras_err_data * err_data,struct ras_query_context * qctx)580 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
581 enum aca_error_type type, struct ras_err_data *err_data,
582 struct ras_query_context *qctx)
583 {
584 if (!handle || !err_data)
585 return -EINVAL;
586
587 if (aca_handle_is_valid(handle))
588 return -EOPNOTSUPP;
589
590 if ((type < 0) || (!(BIT(type) & handle->mask)))
591 return 0;
592
593 return __aca_get_error_data(adev, handle, type, err_data, qctx);
594 }
595
aca_error_init(struct aca_error * aerr,enum aca_error_type type)596 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
597 {
598 mutex_init(&aerr->lock);
599 INIT_LIST_HEAD(&aerr->list);
600 aerr->type = type;
601 aerr->nr_errors = 0;
602 }
603
aca_init_error_cache(struct aca_handle * handle)604 static void aca_init_error_cache(struct aca_handle *handle)
605 {
606 struct aca_error_cache *error_cache = &handle->error_cache;
607 int type;
608
609 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
610 aca_error_init(&error_cache->errors[type], type);
611 }
612
aca_error_fini(struct aca_error * aerr)613 static void aca_error_fini(struct aca_error *aerr)
614 {
615 struct aca_bank_error *bank_error, *tmp;
616
617 mutex_lock(&aerr->lock);
618 if (list_empty(&aerr->list))
619 goto out_unlock;
620
621 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
622 aca_bank_error_remove(aerr, bank_error);
623
624 out_unlock:
625 mutex_destroy(&aerr->lock);
626 }
627
aca_fini_error_cache(struct aca_handle * handle)628 static void aca_fini_error_cache(struct aca_handle *handle)
629 {
630 struct aca_error_cache *error_cache = &handle->error_cache;
631 int type;
632
633 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
634 aca_error_fini(&error_cache->errors[type]);
635 }
636
add_aca_handle(struct amdgpu_device * adev,struct aca_handle_manager * mgr,struct aca_handle * handle,const char * name,const struct aca_info * ras_info,void * data)637 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
638 const char *name, const struct aca_info *ras_info, void *data)
639 {
640 memset(handle, 0, sizeof(*handle));
641
642 handle->adev = adev;
643 handle->mgr = mgr;
644 handle->name = name;
645 handle->hwip = ras_info->hwip;
646 handle->mask = ras_info->mask;
647 handle->bank_ops = ras_info->bank_ops;
648 handle->data = data;
649 aca_init_error_cache(handle);
650
651 INIT_LIST_HEAD(&handle->node);
652 list_add_tail(&handle->node, &mgr->list);
653 mgr->nr_handles++;
654
655 return 0;
656 }
657
aca_sysfs_read(struct device * dev,struct device_attribute * attr,char * buf)658 static ssize_t aca_sysfs_read(struct device *dev,
659 struct device_attribute *attr, char *buf)
660 {
661 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
662
663 /* NOTE: the aca cache will be auto cleared once read,
664 * So the driver should unify the query entry point, forward request to ras query interface directly */
665 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
666 }
667
add_aca_sysfs(struct amdgpu_device * adev,struct aca_handle * handle)668 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
669 {
670 struct device_attribute *aca_attr = &handle->aca_attr;
671
672 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
673 aca_attr->show = aca_sysfs_read;
674 aca_attr->attr.name = handle->attr_name;
675 aca_attr->attr.mode = S_IRUGO;
676 sysfs_attr_init(&aca_attr->attr);
677
678 return sysfs_add_file_to_group(&adev->dev->kobj,
679 &aca_attr->attr,
680 "ras");
681 }
682
amdgpu_aca_add_handle(struct amdgpu_device * adev,struct aca_handle * handle,const char * name,const struct aca_info * ras_info,void * data)683 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
684 const char *name, const struct aca_info *ras_info, void *data)
685 {
686 struct amdgpu_aca *aca = &adev->aca;
687 int ret;
688
689 if (!amdgpu_aca_is_enabled(adev))
690 return 0;
691
692 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
693 if (ret)
694 return ret;
695
696 return add_aca_sysfs(adev, handle);
697 }
698
remove_aca_handle(struct aca_handle * handle)699 static void remove_aca_handle(struct aca_handle *handle)
700 {
701 struct aca_handle_manager *mgr = handle->mgr;
702
703 aca_fini_error_cache(handle);
704 list_del(&handle->node);
705 mgr->nr_handles--;
706 }
707
remove_aca_sysfs(struct aca_handle * handle)708 static void remove_aca_sysfs(struct aca_handle *handle)
709 {
710 struct amdgpu_device *adev = handle->adev;
711 struct device_attribute *aca_attr = &handle->aca_attr;
712
713 if (adev->dev->kobj.sd)
714 sysfs_remove_file_from_group(&adev->dev->kobj,
715 &aca_attr->attr,
716 "ras");
717 }
718
amdgpu_aca_remove_handle(struct aca_handle * handle)719 void amdgpu_aca_remove_handle(struct aca_handle *handle)
720 {
721 if (!handle || list_empty(&handle->node))
722 return;
723
724 remove_aca_sysfs(handle);
725 remove_aca_handle(handle);
726 }
727
aca_manager_init(struct aca_handle_manager * mgr)728 static int aca_manager_init(struct aca_handle_manager *mgr)
729 {
730 INIT_LIST_HEAD(&mgr->list);
731 mgr->nr_handles = 0;
732
733 return 0;
734 }
735
aca_manager_fini(struct aca_handle_manager * mgr)736 static void aca_manager_fini(struct aca_handle_manager *mgr)
737 {
738 struct aca_handle *handle, *tmp;
739
740 if (list_empty(&mgr->list))
741 return;
742
743 list_for_each_entry_safe(handle, tmp, &mgr->list, node)
744 amdgpu_aca_remove_handle(handle);
745 }
746
amdgpu_aca_is_enabled(struct amdgpu_device * adev)747 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
748 {
749 return (adev->aca.is_enabled ||
750 adev->debug_enable_ras_aca);
751 }
752
amdgpu_aca_init(struct amdgpu_device * adev)753 int amdgpu_aca_init(struct amdgpu_device *adev)
754 {
755 struct amdgpu_aca *aca = &adev->aca;
756 int ret;
757
758 atomic_set(&aca->ue_update_flag, 0);
759
760 ret = aca_manager_init(&aca->mgr);
761 if (ret)
762 return ret;
763
764 return 0;
765 }
766
amdgpu_aca_fini(struct amdgpu_device * adev)767 void amdgpu_aca_fini(struct amdgpu_device *adev)
768 {
769 struct amdgpu_aca *aca = &adev->aca;
770
771 aca_manager_fini(&aca->mgr);
772
773 atomic_set(&aca->ue_update_flag, 0);
774 }
775
amdgpu_aca_reset(struct amdgpu_device * adev)776 int amdgpu_aca_reset(struct amdgpu_device *adev)
777 {
778 struct amdgpu_aca *aca = &adev->aca;
779
780 atomic_set(&aca->ue_update_flag, 0);
781
782 return 0;
783 }
784
amdgpu_aca_set_smu_funcs(struct amdgpu_device * adev,const struct aca_smu_funcs * smu_funcs)785 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
786 {
787 struct amdgpu_aca *aca = &adev->aca;
788
789 WARN_ON(aca->smu_funcs);
790 aca->smu_funcs = smu_funcs;
791 }
792
aca_bank_info_decode(struct aca_bank * bank,struct aca_bank_info * info)793 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
794 {
795 u64 ipid;
796 u32 instidhi, instidlo;
797
798 if (!bank || !info)
799 return -EINVAL;
800
801 ipid = bank->regs[ACA_REG_IDX_IPID];
802 info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
803 info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
804 /*
805 * Unfied DieID Format: SAASS. A:AID, S:Socket.
806 * Unfied DieID[4:4] = InstanceId[0:0]
807 * Unfied DieID[0:3] = InstanceIdHi[0:3]
808 */
809 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
810 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
811 info->die_id = ((instidhi >> 2) & 0x03);
812 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
813
814 return 0;
815 }
816
aca_bank_get_error_code(struct amdgpu_device * adev,struct aca_bank * bank)817 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
818 {
819 struct amdgpu_aca *aca = &adev->aca;
820 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
821
822 if (!smu_funcs || !smu_funcs->parse_error_code)
823 return -EOPNOTSUPP;
824
825 return smu_funcs->parse_error_code(adev, bank);
826 }
827
aca_bank_check_error_codes(struct amdgpu_device * adev,struct aca_bank * bank,int * err_codes,int size)828 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
829 {
830 int i, error_code;
831
832 if (!bank || !err_codes)
833 return -EINVAL;
834
835 error_code = aca_bank_get_error_code(adev, bank);
836 if (error_code < 0)
837 return error_code;
838
839 for (i = 0; i < size; i++) {
840 if (err_codes[i] == error_code)
841 return 0;
842 }
843
844 return -EINVAL;
845 }
846
amdgpu_aca_smu_set_debug_mode(struct amdgpu_device * adev,bool en)847 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
848 {
849 struct amdgpu_aca *aca = &adev->aca;
850 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
851
852 if (!smu_funcs || !smu_funcs->set_debug_mode)
853 return -EOPNOTSUPP;
854
855 return smu_funcs->set_debug_mode(adev, en);
856 }
857
858 #if defined(CONFIG_DEBUG_FS)
amdgpu_aca_smu_debug_mode_set(void * data,u64 val)859 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
860 {
861 struct amdgpu_device *adev = (struct amdgpu_device *)data;
862 int ret;
863
864 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
865 if (ret)
866 return ret;
867
868 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
869
870 return 0;
871 }
872
aca_dump_entry(struct seq_file * m,struct aca_bank * bank,enum aca_smu_type type,int idx)873 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
874 {
875 struct aca_bank_info info;
876 int i, ret;
877
878 ret = aca_bank_info_decode(bank, &info);
879 if (ret)
880 return;
881
882 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE");
883 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
884 idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
885
886 for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
887 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
888 }
889
890 struct aca_dump_context {
891 struct seq_file *m;
892 int idx;
893 };
894
handler_aca_bank_dump(struct aca_handle * handle,struct aca_bank * bank,enum aca_smu_type type,void * data)895 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
896 enum aca_smu_type type, void *data)
897 {
898 struct aca_dump_context *ctx = (struct aca_dump_context *)data;
899
900 aca_dump_entry(ctx->m, bank, type, ctx->idx++);
901
902 return handler_aca_log_bank_error(handle, bank, type, NULL);
903 }
904
aca_dump_show(struct seq_file * m,enum aca_smu_type type)905 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
906 {
907 struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
908 struct aca_dump_context context = {
909 .m = m,
910 .idx = 0,
911 };
912
913 return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
914 }
915
aca_dump_ce_show(struct seq_file * m,void * unused)916 static int aca_dump_ce_show(struct seq_file *m, void *unused)
917 {
918 return aca_dump_show(m, ACA_SMU_TYPE_CE);
919 }
920
aca_dump_ce_open(struct inode * inode,struct file * file)921 static int aca_dump_ce_open(struct inode *inode, struct file *file)
922 {
923 return single_open(file, aca_dump_ce_show, inode->i_private);
924 }
925
926 static const struct file_operations aca_ce_dump_debug_fops = {
927 .owner = THIS_MODULE,
928 .open = aca_dump_ce_open,
929 .read = seq_read,
930 .llseek = seq_lseek,
931 .release = single_release,
932 };
933
aca_dump_ue_show(struct seq_file * m,void * unused)934 static int aca_dump_ue_show(struct seq_file *m, void *unused)
935 {
936 return aca_dump_show(m, ACA_SMU_TYPE_UE);
937 }
938
aca_dump_ue_open(struct inode * inode,struct file * file)939 static int aca_dump_ue_open(struct inode *inode, struct file *file)
940 {
941 return single_open(file, aca_dump_ue_show, inode->i_private);
942 }
943
944 static const struct file_operations aca_ue_dump_debug_fops = {
945 .owner = THIS_MODULE,
946 .open = aca_dump_ue_open,
947 .read = seq_read,
948 .llseek = seq_lseek,
949 .release = single_release,
950 };
951
952 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
953 #endif
954
amdgpu_aca_smu_debugfs_init(struct amdgpu_device * adev,struct dentry * root)955 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
956 {
957 #if defined(CONFIG_DEBUG_FS)
958 if (!root)
959 return;
960
961 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
962 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
963 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);
964 #endif
965 }
966