Path: blob/master/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c
29285 views
/*1* Copyright 2023 Advanced Micro Devices, Inc.2*3* Permission is hereby granted, free of charge, to any person obtaining a4* copy of this software and associated documentation files (the "Software"),5* to deal in the Software without restriction, including without limitation6* the rights to use, copy, modify, merge, publish, distribute, sublicense,7* and/or sell copies of the Software, and to permit persons to whom the8* Software is furnished to do so, subject to the following conditions:9*10* The above copyright notice and this permission notice shall be included in11* all copies or substantial portions of the Software.12*13* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR14* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,15* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL16* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR17* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,18* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR19* OTHER DEALINGS IN THE SOFTWARE.20*21*/2223#include <linux/list.h>24#include "amdgpu.h"25#include "amdgpu_aca.h"26#include "amdgpu_ras.h"2728#define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}2930typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);3132static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {33ACA_BANK_HWID(SMU, 0x01, 0x01),34ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00),35ACA_BANK_HWID(UMC, 0x96, 0x00),36};3738static void aca_banks_init(struct aca_banks *banks)39{40if (!banks)41return;4243memset(banks, 0, sizeof(*banks));44INIT_LIST_HEAD(&banks->list);45}4647static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)48{49struct aca_bank_node *node;5051if (!bank)52return -EINVAL;5354node = kvzalloc(sizeof(*node), GFP_KERNEL);55if (!node)56return -ENOMEM;5758memcpy(&node->bank, bank, sizeof(*bank));5960INIT_LIST_HEAD(&node->node);61list_add_tail(&node->node, &banks->list);6263banks->nr_banks++;6465return 0;66}6768static void aca_banks_release(struct aca_banks *banks)69{70struct aca_bank_node *node, *tmp;7172if (list_empty(&banks->list))73return;7475list_for_each_entry_safe(node, tmp, &banks->list, node) {76list_del(&node->node);77kvfree(node);78banks->nr_banks--;79}80}8182static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)83{84struct amdgpu_aca *aca = &adev->aca;85const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;8687if (!count)88return -EINVAL;8990if (!smu_funcs || !smu_funcs->get_valid_aca_count)91return -EOPNOTSUPP;9293return smu_funcs->get_valid_aca_count(adev, type, count);94}9596static struct aca_regs_dump {97const char *name;98int reg_idx;99} aca_regs[] = {100{"CONTROL", ACA_REG_IDX_CTL},101{"STATUS", ACA_REG_IDX_STATUS},102{"ADDR", ACA_REG_IDX_ADDR},103{"MISC", ACA_REG_IDX_MISC0},104{"CONFIG", ACA_REG_IDX_CONFIG},105{"IPID", ACA_REG_IDX_IPID},106{"SYND", ACA_REG_IDX_SYND},107{"DESTAT", ACA_REG_IDX_DESTAT},108{"DEADDR", ACA_REG_IDX_DEADDR},109{"CONTROL_MASK", ACA_REG_IDX_CTL_MASK},110};111112static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,113struct ras_query_context *qctx)114{115u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;116int i;117118if (adev->debug_disable_ce_logs &&119bank->smu_err_type == ACA_SMU_TYPE_CE &&120!ACA_BANK_ERR_IS_DEFFERED(bank))121return;122123RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");124/* plus 1 for output format, e.g: ACA[08/08]: xxxx */125for (i = 0; i < ARRAY_SIZE(aca_regs); i++)126RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",127idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);128129if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS]))130RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n");131}132133static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)134{135136struct aca_hwip *hwip;137int hwid, mcatype;138u64 ipid;139140if (!bank || type == ACA_HWIP_TYPE_UNKNOW)141return false;142143hwip = &aca_hwid_mcatypes[type];144if (!hwip->hwid)145return false;146147ipid = bank->regs[ACA_REG_IDX_IPID];148hwid = ACA_REG__IPID__HARDWAREID(ipid);149mcatype = ACA_REG__IPID__MCATYPE(ipid);150151return hwip->hwid == hwid && hwip->mcatype == mcatype;152}153154static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,155int start, int count,156struct aca_banks *banks, struct ras_query_context *qctx)157{158struct amdgpu_aca *aca = &adev->aca;159const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;160struct aca_bank bank;161int i, max_count, ret;162163if (!count)164return 0;165166if (!smu_funcs || !smu_funcs->get_valid_aca_bank)167return -EOPNOTSUPP;168169switch (type) {170case ACA_SMU_TYPE_UE:171max_count = smu_funcs->max_ue_bank_count;172break;173case ACA_SMU_TYPE_CE:174max_count = smu_funcs->max_ce_bank_count;175break;176default:177return -EINVAL;178}179180if (start + count > max_count)181return -EINVAL;182183count = min_t(int, count, max_count);184for (i = 0; i < count; i++) {185memset(&bank, 0, sizeof(bank));186ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);187if (ret)188return ret;189190bank.smu_err_type = type;191192/*193* Poison being consumed when injecting a UE while running background workloads,194* which are unexpected.195*/196if (type == ACA_SMU_TYPE_UE &&197ACA_REG__STATUS__POISON(bank.regs[ACA_REG_IDX_STATUS]) &&198!aca_bank_hwip_is_matched(&bank, ACA_HWIP_TYPE_UMC))199continue;200201aca_smu_bank_dump(adev, i, count, &bank, qctx);202203ret = aca_banks_add_bank(banks, &bank);204if (ret)205return ret;206}207208return 0;209}210211static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)212{213const struct aca_bank_ops *bank_ops = handle->bank_ops;214215/* Parse all deferred errors with UMC aca handle */216if (ACA_BANK_ERR_IS_DEFFERED(bank))217return handle->hwip == ACA_HWIP_TYPE_UMC;218219if (!aca_bank_hwip_is_matched(bank, handle->hwip))220return false;221222if (!bank_ops->aca_bank_is_valid)223return true;224225return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);226}227228static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)229{230struct aca_bank_error *bank_error;231232bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);233if (!bank_error)234return NULL;235236INIT_LIST_HEAD(&bank_error->node);237memcpy(&bank_error->info, info, sizeof(*info));238239mutex_lock(&aerr->lock);240list_add_tail(&bank_error->node, &aerr->list);241aerr->nr_errors++;242mutex_unlock(&aerr->lock);243244return bank_error;245}246247static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)248{249struct aca_bank_error *bank_error = NULL;250struct aca_bank_info *tmp_info;251bool found = false;252253mutex_lock(&aerr->lock);254list_for_each_entry(bank_error, &aerr->list, node) {255tmp_info = &bank_error->info;256if (tmp_info->socket_id == info->socket_id &&257tmp_info->die_id == info->die_id) {258found = true;259goto out_unlock;260}261}262263out_unlock:264mutex_unlock(&aerr->lock);265266return found ? bank_error : NULL;267}268269static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)270{271if (!aerr || !bank_error)272return;273274list_del(&bank_error->node);275aerr->nr_errors--;276277kvfree(bank_error);278}279280static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)281{282struct aca_bank_error *bank_error;283284if (!aerr || !info)285return NULL;286287bank_error = find_bank_error(aerr, info);288if (bank_error)289return bank_error;290291return new_bank_error(aerr, info);292}293294int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,295enum aca_error_type type, u64 count)296{297struct aca_error_cache *error_cache = &handle->error_cache;298struct aca_bank_error *bank_error;299struct aca_error *aerr;300301if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)302return -EINVAL;303304if (!count)305return 0;306307aerr = &error_cache->errors[type];308bank_error = get_bank_error(aerr, info);309if (!bank_error)310return -ENOMEM;311312bank_error->count += count;313314return 0;315}316317static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)318{319const struct aca_bank_ops *bank_ops = handle->bank_ops;320321if (!bank)322return -EINVAL;323324if (!bank_ops->aca_bank_parser)325return -EOPNOTSUPP;326327return bank_ops->aca_bank_parser(handle, bank, type,328handle->data);329}330331static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,332enum aca_smu_type type, void *data)333{334int ret;335336ret = aca_bank_parser(handle, bank, type);337if (ret)338return ret;339340return 0;341}342343static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,344enum aca_smu_type type, bank_handler_t handler, void *data)345{346struct aca_handle *handle;347int ret;348349if (list_empty(&mgr->list))350return 0;351352list_for_each_entry(handle, &mgr->list, node) {353if (!aca_bank_is_valid(handle, bank, type))354continue;355356ret = handler(handle, bank, type, data);357if (ret)358return ret;359}360361return 0;362}363364static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,365enum aca_smu_type type, bank_handler_t handler, void *data)366{367struct aca_bank_node *node;368struct aca_bank *bank;369int ret;370371if (!mgr || !banks)372return -EINVAL;373374/* pre check to avoid unnecessary operations */375if (list_empty(&mgr->list) || list_empty(&banks->list))376return 0;377378list_for_each_entry(node, &banks->list, node) {379bank = &node->bank;380381ret = aca_dispatch_bank(mgr, bank, type, handler, data);382if (ret)383return ret;384}385386return 0;387}388389static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)390{391struct amdgpu_aca *aca = &adev->aca;392bool ret = true;393394/*395* Because the UE Valid MCA count will only be cleared after reset,396* in order to avoid repeated counting of the error count,397* the aca bank is only updated once during the gpu recovery stage.398*/399if (type == ACA_SMU_TYPE_UE) {400if (amdgpu_ras_intr_triggered())401ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;402else403atomic_set(&aca->ue_update_flag, 0);404}405406return ret;407}408409static void aca_banks_generate_cper(struct amdgpu_device *adev,410enum aca_smu_type type,411struct aca_banks *banks,412int count)413{414struct aca_bank_node *node;415struct aca_bank *bank;416int r;417418if (!adev->cper.enabled)419return;420421if (!banks || !count) {422dev_warn(adev->dev, "fail to generate cper records\n");423return;424}425426/* UEs must be encoded into separate CPER entries */427if (type == ACA_SMU_TYPE_UE) {428struct aca_banks de_banks;429430aca_banks_init(&de_banks);431list_for_each_entry(node, &banks->list, node) {432bank = &node->bank;433if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {434r = aca_banks_add_bank(&de_banks, bank);435if (r)436dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);437} else {438if (amdgpu_cper_generate_ue_record(adev, bank))439dev_warn(adev->dev, "fail to generate ue cper records\n");440}441}442443if (!list_empty(&de_banks.list)) {444if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))445dev_warn(adev->dev, "fail to generate de cper records\n");446}447448aca_banks_release(&de_banks);449} else {450/*451* SMU_TYPE_CE banks are combined into 1 CPER entries,452* they could be CEs or DEs or both453*/454if (amdgpu_cper_generate_ce_records(adev, banks, count))455dev_warn(adev->dev, "fail to generate ce cper records\n");456}457}458459static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,460bank_handler_t handler, struct ras_query_context *qctx, void *data)461{462struct amdgpu_aca *aca = &adev->aca;463struct aca_banks banks;464u32 count = 0;465int ret;466467if (list_empty(&aca->mgr.list))468return 0;469470if (!aca_bank_should_update(adev, type))471return 0;472473ret = aca_smu_get_valid_aca_count(adev, type, &count);474if (ret)475return ret;476477if (!count)478return 0;479480aca_banks_init(&banks);481482ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);483if (ret)484goto err_release_banks;485486if (list_empty(&banks.list)) {487ret = 0;488goto err_release_banks;489}490491ret = aca_dispatch_banks(&aca->mgr, &banks, type,492handler, data);493if (ret)494goto err_release_banks;495496aca_banks_generate_cper(adev, type, &banks, count);497498err_release_banks:499aca_banks_release(&banks);500501return ret;502}503504static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)505{506struct aca_bank_info *info;507struct amdgpu_smuio_mcm_config_info mcm_info;508u64 count;509510if (type >= ACA_ERROR_TYPE_COUNT)511return -EINVAL;512513count = bank_error->count;514if (!count)515return 0;516517info = &bank_error->info;518mcm_info.die_id = info->die_id;519mcm_info.socket_id = info->socket_id;520521switch (type) {522case ACA_ERROR_TYPE_UE:523amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);524break;525case ACA_ERROR_TYPE_CE:526amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);527break;528case ACA_ERROR_TYPE_DEFERRED:529amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);530break;531default:532break;533}534535return 0;536}537538static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)539{540struct aca_error_cache *error_cache = &handle->error_cache;541struct aca_error *aerr = &error_cache->errors[type];542struct aca_bank_error *bank_error, *tmp;543544mutex_lock(&aerr->lock);545546if (list_empty(&aerr->list))547goto out_unlock;548549list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {550aca_log_aca_error_data(bank_error, type, err_data);551aca_bank_error_remove(aerr, bank_error);552}553554out_unlock:555mutex_unlock(&aerr->lock);556557return 0;558}559560static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,561struct ras_err_data *err_data, struct ras_query_context *qctx)562{563enum aca_smu_type smu_type;564int ret;565566switch (type) {567case ACA_ERROR_TYPE_UE:568smu_type = ACA_SMU_TYPE_UE;569break;570case ACA_ERROR_TYPE_CE:571case ACA_ERROR_TYPE_DEFERRED:572smu_type = ACA_SMU_TYPE_CE;573break;574default:575return -EINVAL;576}577578/* update aca bank to aca source error_cache first */579ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);580if (ret)581return ret;582583/* DEs may contain in CEs or UEs */584if (type != ACA_ERROR_TYPE_DEFERRED)585aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);586587return aca_log_aca_error(handle, type, err_data);588}589590static bool aca_handle_is_valid(struct aca_handle *handle)591{592if (!handle->mask || !list_empty(&handle->node))593return false;594595return true;596}597598int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,599enum aca_error_type type, struct ras_err_data *err_data,600struct ras_query_context *qctx)601{602if (!handle || !err_data)603return -EINVAL;604605if (aca_handle_is_valid(handle))606return -EOPNOTSUPP;607608if ((type < 0) || (!(BIT(type) & handle->mask)))609return 0;610611return __aca_get_error_data(adev, handle, type, err_data, qctx);612}613614static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)615{616mutex_init(&aerr->lock);617INIT_LIST_HEAD(&aerr->list);618aerr->type = type;619aerr->nr_errors = 0;620}621622static void aca_init_error_cache(struct aca_handle *handle)623{624struct aca_error_cache *error_cache = &handle->error_cache;625int type;626627for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)628aca_error_init(&error_cache->errors[type], type);629}630631static void aca_error_fini(struct aca_error *aerr)632{633struct aca_bank_error *bank_error, *tmp;634635mutex_lock(&aerr->lock);636if (list_empty(&aerr->list))637goto out_unlock;638639list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)640aca_bank_error_remove(aerr, bank_error);641642out_unlock:643mutex_destroy(&aerr->lock);644}645646static void aca_fini_error_cache(struct aca_handle *handle)647{648struct aca_error_cache *error_cache = &handle->error_cache;649int type;650651for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)652aca_error_fini(&error_cache->errors[type]);653}654655static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,656const char *name, const struct aca_info *ras_info, void *data)657{658memset(handle, 0, sizeof(*handle));659660handle->adev = adev;661handle->mgr = mgr;662handle->name = name;663handle->hwip = ras_info->hwip;664handle->mask = ras_info->mask;665handle->bank_ops = ras_info->bank_ops;666handle->data = data;667aca_init_error_cache(handle);668669INIT_LIST_HEAD(&handle->node);670list_add_tail(&handle->node, &mgr->list);671mgr->nr_handles++;672673return 0;674}675676static ssize_t aca_sysfs_read(struct device *dev,677struct device_attribute *attr, char *buf)678{679struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);680681/* NOTE: the aca cache will be auto cleared once read,682* So the driver should unify the query entry point, forward request to ras query interface directly */683return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);684}685686static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)687{688struct device_attribute *aca_attr = &handle->aca_attr;689690snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);691aca_attr->show = aca_sysfs_read;692aca_attr->attr.name = handle->attr_name;693aca_attr->attr.mode = S_IRUGO;694sysfs_attr_init(&aca_attr->attr);695696return sysfs_add_file_to_group(&adev->dev->kobj,697&aca_attr->attr,698"ras");699}700701int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,702const char *name, const struct aca_info *ras_info, void *data)703{704struct amdgpu_aca *aca = &adev->aca;705int ret;706707if (!amdgpu_aca_is_enabled(adev))708return 0;709710ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);711if (ret)712return ret;713714return add_aca_sysfs(adev, handle);715}716717static void remove_aca_handle(struct aca_handle *handle)718{719struct aca_handle_manager *mgr = handle->mgr;720721aca_fini_error_cache(handle);722list_del(&handle->node);723mgr->nr_handles--;724}725726static void remove_aca_sysfs(struct aca_handle *handle)727{728struct amdgpu_device *adev = handle->adev;729struct device_attribute *aca_attr = &handle->aca_attr;730731if (adev->dev->kobj.sd)732sysfs_remove_file_from_group(&adev->dev->kobj,733&aca_attr->attr,734"ras");735}736737void amdgpu_aca_remove_handle(struct aca_handle *handle)738{739if (!handle || list_empty(&handle->node))740return;741742remove_aca_sysfs(handle);743remove_aca_handle(handle);744}745746static int aca_manager_init(struct aca_handle_manager *mgr)747{748INIT_LIST_HEAD(&mgr->list);749mgr->nr_handles = 0;750751return 0;752}753754static void aca_manager_fini(struct aca_handle_manager *mgr)755{756struct aca_handle *handle, *tmp;757758if (list_empty(&mgr->list))759return;760761list_for_each_entry_safe(handle, tmp, &mgr->list, node)762amdgpu_aca_remove_handle(handle);763}764765bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)766{767return (adev->aca.is_enabled ||768adev->debug_enable_ras_aca);769}770771int amdgpu_aca_init(struct amdgpu_device *adev)772{773struct amdgpu_aca *aca = &adev->aca;774int ret;775776atomic_set(&aca->ue_update_flag, 0);777778ret = aca_manager_init(&aca->mgr);779if (ret)780return ret;781782return 0;783}784785void amdgpu_aca_fini(struct amdgpu_device *adev)786{787struct amdgpu_aca *aca = &adev->aca;788789aca_manager_fini(&aca->mgr);790791atomic_set(&aca->ue_update_flag, 0);792}793794int amdgpu_aca_reset(struct amdgpu_device *adev)795{796struct amdgpu_aca *aca = &adev->aca;797798atomic_set(&aca->ue_update_flag, 0);799800return 0;801}802803void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)804{805struct amdgpu_aca *aca = &adev->aca;806807WARN_ON(aca->smu_funcs);808aca->smu_funcs = smu_funcs;809}810811int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)812{813u64 ipid;814u32 instidhi, instidlo;815816if (!bank || !info)817return -EINVAL;818819ipid = bank->regs[ACA_REG_IDX_IPID];820info->hwid = ACA_REG__IPID__HARDWAREID(ipid);821info->mcatype = ACA_REG__IPID__MCATYPE(ipid);822/*823* Unfied DieID Format: SAASS. A:AID, S:Socket.824* Unfied DieID[4:4] = InstanceId[0:0]825* Unfied DieID[0:3] = InstanceIdHi[0:3]826*/827instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);828instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);829info->die_id = ((instidhi >> 2) & 0x03);830info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);831832return 0;833}834835static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)836{837struct amdgpu_aca *aca = &adev->aca;838const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;839840if (!smu_funcs || !smu_funcs->parse_error_code)841return -EOPNOTSUPP;842843return smu_funcs->parse_error_code(adev, bank);844}845846int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)847{848int i, error_code;849850if (!bank || !err_codes)851return -EINVAL;852853error_code = aca_bank_get_error_code(adev, bank);854if (error_code < 0)855return error_code;856857for (i = 0; i < size; i++) {858if (err_codes[i] == error_code)859return 0;860}861862return -EINVAL;863}864865int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)866{867struct amdgpu_aca *aca = &adev->aca;868const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;869870if (!smu_funcs || !smu_funcs->set_debug_mode)871return -EOPNOTSUPP;872873return smu_funcs->set_debug_mode(adev, en);874}875876#if defined(CONFIG_DEBUG_FS)877static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)878{879struct amdgpu_device *adev = (struct amdgpu_device *)data;880int ret;881882ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);883if (ret)884return ret;885886dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");887888return 0;889}890891static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)892{893struct aca_bank_info info;894int i, ret;895896ret = aca_bank_info_decode(bank, &info);897if (ret)898return;899900seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE");901seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",902idx, info.socket_id, info.die_id, info.hwid, info.mcatype);903904for (i = 0; i < ARRAY_SIZE(aca_regs); i++)905seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);906}907908struct aca_dump_context {909struct seq_file *m;910int idx;911};912913static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,914enum aca_smu_type type, void *data)915{916struct aca_dump_context *ctx = (struct aca_dump_context *)data;917918aca_dump_entry(ctx->m, bank, type, ctx->idx++);919920return handler_aca_log_bank_error(handle, bank, type, NULL);921}922923static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)924{925struct amdgpu_device *adev = (struct amdgpu_device *)m->private;926struct aca_dump_context context = {927.m = m,928.idx = 0,929};930931return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);932}933934static int aca_dump_ce_show(struct seq_file *m, void *unused)935{936return aca_dump_show(m, ACA_SMU_TYPE_CE);937}938939static int aca_dump_ce_open(struct inode *inode, struct file *file)940{941return single_open(file, aca_dump_ce_show, inode->i_private);942}943944static const struct file_operations aca_ce_dump_debug_fops = {945.owner = THIS_MODULE,946.open = aca_dump_ce_open,947.read = seq_read,948.llseek = seq_lseek,949.release = single_release,950};951952static int aca_dump_ue_show(struct seq_file *m, void *unused)953{954return aca_dump_show(m, ACA_SMU_TYPE_UE);955}956957static int aca_dump_ue_open(struct inode *inode, struct file *file)958{959return single_open(file, aca_dump_ue_show, inode->i_private);960}961962static const struct file_operations aca_ue_dump_debug_fops = {963.owner = THIS_MODULE,964.open = aca_dump_ue_open,965.read = seq_read,966.llseek = seq_lseek,967.release = single_release,968};969970DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");971#endif972973void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)974{975#if defined(CONFIG_DEBUG_FS)976if (!root)977return;978979debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);980debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);981debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);982#endif983}984985986