Path: blob/master/drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c
26517 views
/*1* Copyright 2023 Advanced Micro Devices, Inc.2*3* Permission is hereby granted, free of charge, to any person obtaining a4* copy of this software and associated documentation files (the "Software"),5* to deal in the Software without restriction, including without limitation6* the rights to use, copy, modify, merge, publish, distribute, sublicense,7* and/or sell copies of the Software, and to permit persons to whom the8* Software is furnished to do so, subject to the following conditions:9*10* The above copyright notice and this permission notice shall be included in11* all copies or substantial portions of the Software.12*13* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR14* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,15* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL16* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR17* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,18* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR19* OTHER DEALINGS IN THE SOFTWARE.20*21*/2223#include <linux/list.h>24#include "amdgpu.h"25#include "amdgpu_aca.h"26#include "amdgpu_ras.h"2728#define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}2930typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);3132static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {33ACA_BANK_HWID(SMU, 0x01, 0x01),34ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00),35ACA_BANK_HWID(UMC, 0x96, 0x00),36};3738static void aca_banks_init(struct aca_banks *banks)39{40if (!banks)41return;4243memset(banks, 0, sizeof(*banks));44INIT_LIST_HEAD(&banks->list);45}4647static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)48{49struct aca_bank_node *node;5051if (!bank)52return -EINVAL;5354node = kvzalloc(sizeof(*node), GFP_KERNEL);55if (!node)56return -ENOMEM;5758memcpy(&node->bank, bank, sizeof(*bank));5960INIT_LIST_HEAD(&node->node);61list_add_tail(&node->node, &banks->list);6263banks->nr_banks++;6465return 0;66}6768static void aca_banks_release(struct aca_banks *banks)69{70struct aca_bank_node *node, *tmp;7172if (list_empty(&banks->list))73return;7475list_for_each_entry_safe(node, tmp, &banks->list, node) {76list_del(&node->node);77kvfree(node);78}79}8081static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)82{83struct amdgpu_aca *aca = &adev->aca;84const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;8586if (!count)87return -EINVAL;8889if (!smu_funcs || !smu_funcs->get_valid_aca_count)90return -EOPNOTSUPP;9192return smu_funcs->get_valid_aca_count(adev, type, count);93}9495static struct aca_regs_dump {96const char *name;97int reg_idx;98} aca_regs[] = {99{"CONTROL", ACA_REG_IDX_CTL},100{"STATUS", ACA_REG_IDX_STATUS},101{"ADDR", ACA_REG_IDX_ADDR},102{"MISC", ACA_REG_IDX_MISC0},103{"CONFIG", ACA_REG_IDX_CONFIG},104{"IPID", ACA_REG_IDX_IPID},105{"SYND", ACA_REG_IDX_SYND},106{"DESTAT", ACA_REG_IDX_DESTAT},107{"DEADDR", ACA_REG_IDX_DEADDR},108{"CONTROL_MASK", ACA_REG_IDX_CTL_MASK},109};110111static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,112struct ras_query_context *qctx)113{114u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;115int i;116117if (adev->debug_disable_ce_logs &&118bank->smu_err_type == ACA_SMU_TYPE_CE &&119!ACA_BANK_ERR_IS_DEFFERED(bank))120return;121122RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");123/* plus 1 for output format, e.g: ACA[08/08]: xxxx */124for (i = 0; i < ARRAY_SIZE(aca_regs); i++)125RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",126idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);127128if (ACA_REG__STATUS__SCRUB(bank->regs[ACA_REG_IDX_STATUS]))129RAS_EVENT_LOG(adev, event_id, HW_ERR "hardware error logged by the scrubber\n");130}131132static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,133int start, int count,134struct aca_banks *banks, struct ras_query_context *qctx)135{136struct amdgpu_aca *aca = &adev->aca;137const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;138struct aca_bank bank;139int i, max_count, ret;140141if (!count)142return 0;143144if (!smu_funcs || !smu_funcs->get_valid_aca_bank)145return -EOPNOTSUPP;146147switch (type) {148case ACA_SMU_TYPE_UE:149max_count = smu_funcs->max_ue_bank_count;150break;151case ACA_SMU_TYPE_CE:152max_count = smu_funcs->max_ce_bank_count;153break;154default:155return -EINVAL;156}157158if (start + count > max_count)159return -EINVAL;160161count = min_t(int, count, max_count);162for (i = 0; i < count; i++) {163memset(&bank, 0, sizeof(bank));164ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);165if (ret)166return ret;167168bank.smu_err_type = type;169170aca_smu_bank_dump(adev, i, count, &bank, qctx);171172ret = aca_banks_add_bank(banks, &bank);173if (ret)174return ret;175}176177return 0;178}179180static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)181{182183struct aca_hwip *hwip;184int hwid, mcatype;185u64 ipid;186187if (!bank || type == ACA_HWIP_TYPE_UNKNOW)188return false;189190hwip = &aca_hwid_mcatypes[type];191if (!hwip->hwid)192return false;193194ipid = bank->regs[ACA_REG_IDX_IPID];195hwid = ACA_REG__IPID__HARDWAREID(ipid);196mcatype = ACA_REG__IPID__MCATYPE(ipid);197198return hwip->hwid == hwid && hwip->mcatype == mcatype;199}200201static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)202{203const struct aca_bank_ops *bank_ops = handle->bank_ops;204205/* Parse all deferred errors with UMC aca handle */206if (ACA_BANK_ERR_IS_DEFFERED(bank))207return handle->hwip == ACA_HWIP_TYPE_UMC;208209if (!aca_bank_hwip_is_matched(bank, handle->hwip))210return false;211212if (!bank_ops->aca_bank_is_valid)213return true;214215return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);216}217218static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)219{220struct aca_bank_error *bank_error;221222bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);223if (!bank_error)224return NULL;225226INIT_LIST_HEAD(&bank_error->node);227memcpy(&bank_error->info, info, sizeof(*info));228229mutex_lock(&aerr->lock);230list_add_tail(&bank_error->node, &aerr->list);231mutex_unlock(&aerr->lock);232233return bank_error;234}235236static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)237{238struct aca_bank_error *bank_error = NULL;239struct aca_bank_info *tmp_info;240bool found = false;241242mutex_lock(&aerr->lock);243list_for_each_entry(bank_error, &aerr->list, node) {244tmp_info = &bank_error->info;245if (tmp_info->socket_id == info->socket_id &&246tmp_info->die_id == info->die_id) {247found = true;248goto out_unlock;249}250}251252out_unlock:253mutex_unlock(&aerr->lock);254255return found ? bank_error : NULL;256}257258static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)259{260if (!aerr || !bank_error)261return;262263list_del(&bank_error->node);264aerr->nr_errors--;265266kvfree(bank_error);267}268269static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)270{271struct aca_bank_error *bank_error;272273if (!aerr || !info)274return NULL;275276bank_error = find_bank_error(aerr, info);277if (bank_error)278return bank_error;279280return new_bank_error(aerr, info);281}282283int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,284enum aca_error_type type, u64 count)285{286struct aca_error_cache *error_cache = &handle->error_cache;287struct aca_bank_error *bank_error;288struct aca_error *aerr;289290if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)291return -EINVAL;292293if (!count)294return 0;295296aerr = &error_cache->errors[type];297bank_error = get_bank_error(aerr, info);298if (!bank_error)299return -ENOMEM;300301bank_error->count += count;302303return 0;304}305306static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)307{308const struct aca_bank_ops *bank_ops = handle->bank_ops;309310if (!bank)311return -EINVAL;312313if (!bank_ops->aca_bank_parser)314return -EOPNOTSUPP;315316return bank_ops->aca_bank_parser(handle, bank, type,317handle->data);318}319320static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,321enum aca_smu_type type, void *data)322{323int ret;324325ret = aca_bank_parser(handle, bank, type);326if (ret)327return ret;328329return 0;330}331332static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,333enum aca_smu_type type, bank_handler_t handler, void *data)334{335struct aca_handle *handle;336int ret;337338if (list_empty(&mgr->list))339return 0;340341list_for_each_entry(handle, &mgr->list, node) {342if (!aca_bank_is_valid(handle, bank, type))343continue;344345ret = handler(handle, bank, type, data);346if (ret)347return ret;348}349350return 0;351}352353static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,354enum aca_smu_type type, bank_handler_t handler, void *data)355{356struct aca_bank_node *node;357struct aca_bank *bank;358int ret;359360if (!mgr || !banks)361return -EINVAL;362363/* pre check to avoid unnecessary operations */364if (list_empty(&mgr->list) || list_empty(&banks->list))365return 0;366367list_for_each_entry(node, &banks->list, node) {368bank = &node->bank;369370ret = aca_dispatch_bank(mgr, bank, type, handler, data);371if (ret)372return ret;373}374375return 0;376}377378static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)379{380struct amdgpu_aca *aca = &adev->aca;381bool ret = true;382383/*384* Because the UE Valid MCA count will only be cleared after reset,385* in order to avoid repeated counting of the error count,386* the aca bank is only updated once during the gpu recovery stage.387*/388if (type == ACA_SMU_TYPE_UE) {389if (amdgpu_ras_intr_triggered())390ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;391else392atomic_set(&aca->ue_update_flag, 0);393}394395return ret;396}397398static void aca_banks_generate_cper(struct amdgpu_device *adev,399enum aca_smu_type type,400struct aca_banks *banks,401int count)402{403struct aca_bank_node *node;404struct aca_bank *bank;405int r;406407if (!adev->cper.enabled)408return;409410if (!banks || !count) {411dev_warn(adev->dev, "fail to generate cper records\n");412return;413}414415/* UEs must be encoded into separate CPER entries */416if (type == ACA_SMU_TYPE_UE) {417struct aca_banks de_banks;418419aca_banks_init(&de_banks);420list_for_each_entry(node, &banks->list, node) {421bank = &node->bank;422if (bank->aca_err_type == ACA_ERROR_TYPE_DEFERRED) {423r = aca_banks_add_bank(&de_banks, bank);424if (r)425dev_warn(adev->dev, "fail to add de banks, ret = %d\n", r);426} else {427if (amdgpu_cper_generate_ue_record(adev, bank))428dev_warn(adev->dev, "fail to generate ue cper records\n");429}430}431432if (!list_empty(&de_banks.list)) {433if (amdgpu_cper_generate_ce_records(adev, &de_banks, de_banks.nr_banks))434dev_warn(adev->dev, "fail to generate de cper records\n");435}436437aca_banks_release(&de_banks);438} else {439/*440* SMU_TYPE_CE banks are combined into 1 CPER entries,441* they could be CEs or DEs or both442*/443if (amdgpu_cper_generate_ce_records(adev, banks, count))444dev_warn(adev->dev, "fail to generate ce cper records\n");445}446}447448static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,449bank_handler_t handler, struct ras_query_context *qctx, void *data)450{451struct amdgpu_aca *aca = &adev->aca;452struct aca_banks banks;453u32 count = 0;454int ret;455456if (list_empty(&aca->mgr.list))457return 0;458459if (!aca_bank_should_update(adev, type))460return 0;461462ret = aca_smu_get_valid_aca_count(adev, type, &count);463if (ret)464return ret;465466if (!count)467return 0;468469aca_banks_init(&banks);470471ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);472if (ret)473goto err_release_banks;474475if (list_empty(&banks.list)) {476ret = 0;477goto err_release_banks;478}479480ret = aca_dispatch_banks(&aca->mgr, &banks, type,481handler, data);482if (ret)483goto err_release_banks;484485aca_banks_generate_cper(adev, type, &banks, count);486487err_release_banks:488aca_banks_release(&banks);489490return ret;491}492493static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)494{495struct aca_bank_info *info;496struct amdgpu_smuio_mcm_config_info mcm_info;497u64 count;498499if (type >= ACA_ERROR_TYPE_COUNT)500return -EINVAL;501502count = bank_error->count;503if (!count)504return 0;505506info = &bank_error->info;507mcm_info.die_id = info->die_id;508mcm_info.socket_id = info->socket_id;509510switch (type) {511case ACA_ERROR_TYPE_UE:512amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count);513break;514case ACA_ERROR_TYPE_CE:515amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count);516break;517case ACA_ERROR_TYPE_DEFERRED:518amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count);519break;520default:521break;522}523524return 0;525}526527static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)528{529struct aca_error_cache *error_cache = &handle->error_cache;530struct aca_error *aerr = &error_cache->errors[type];531struct aca_bank_error *bank_error, *tmp;532533mutex_lock(&aerr->lock);534535if (list_empty(&aerr->list))536goto out_unlock;537538list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {539aca_log_aca_error_data(bank_error, type, err_data);540aca_bank_error_remove(aerr, bank_error);541}542543out_unlock:544mutex_unlock(&aerr->lock);545546return 0;547}548549static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,550struct ras_err_data *err_data, struct ras_query_context *qctx)551{552enum aca_smu_type smu_type;553int ret;554555switch (type) {556case ACA_ERROR_TYPE_UE:557smu_type = ACA_SMU_TYPE_UE;558break;559case ACA_ERROR_TYPE_CE:560case ACA_ERROR_TYPE_DEFERRED:561smu_type = ACA_SMU_TYPE_CE;562break;563default:564return -EINVAL;565}566567/* update aca bank to aca source error_cache first */568ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);569if (ret)570return ret;571572/* DEs may contain in CEs or UEs */573if (type != ACA_ERROR_TYPE_DEFERRED)574aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data);575576return aca_log_aca_error(handle, type, err_data);577}578579static bool aca_handle_is_valid(struct aca_handle *handle)580{581if (!handle->mask || !list_empty(&handle->node))582return false;583584return true;585}586587int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,588enum aca_error_type type, struct ras_err_data *err_data,589struct ras_query_context *qctx)590{591if (!handle || !err_data)592return -EINVAL;593594if (aca_handle_is_valid(handle))595return -EOPNOTSUPP;596597if ((type < 0) || (!(BIT(type) & handle->mask)))598return 0;599600return __aca_get_error_data(adev, handle, type, err_data, qctx);601}602603static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)604{605mutex_init(&aerr->lock);606INIT_LIST_HEAD(&aerr->list);607aerr->type = type;608aerr->nr_errors = 0;609}610611static void aca_init_error_cache(struct aca_handle *handle)612{613struct aca_error_cache *error_cache = &handle->error_cache;614int type;615616for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)617aca_error_init(&error_cache->errors[type], type);618}619620static void aca_error_fini(struct aca_error *aerr)621{622struct aca_bank_error *bank_error, *tmp;623624mutex_lock(&aerr->lock);625if (list_empty(&aerr->list))626goto out_unlock;627628list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)629aca_bank_error_remove(aerr, bank_error);630631out_unlock:632mutex_destroy(&aerr->lock);633}634635static void aca_fini_error_cache(struct aca_handle *handle)636{637struct aca_error_cache *error_cache = &handle->error_cache;638int type;639640for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)641aca_error_fini(&error_cache->errors[type]);642}643644static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,645const char *name, const struct aca_info *ras_info, void *data)646{647memset(handle, 0, sizeof(*handle));648649handle->adev = adev;650handle->mgr = mgr;651handle->name = name;652handle->hwip = ras_info->hwip;653handle->mask = ras_info->mask;654handle->bank_ops = ras_info->bank_ops;655handle->data = data;656aca_init_error_cache(handle);657658INIT_LIST_HEAD(&handle->node);659list_add_tail(&handle->node, &mgr->list);660mgr->nr_handles++;661662return 0;663}664665static ssize_t aca_sysfs_read(struct device *dev,666struct device_attribute *attr, char *buf)667{668struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);669670/* NOTE: the aca cache will be auto cleared once read,671* So the driver should unify the query entry point, forward request to ras query interface directly */672return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);673}674675static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)676{677struct device_attribute *aca_attr = &handle->aca_attr;678679snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);680aca_attr->show = aca_sysfs_read;681aca_attr->attr.name = handle->attr_name;682aca_attr->attr.mode = S_IRUGO;683sysfs_attr_init(&aca_attr->attr);684685return sysfs_add_file_to_group(&adev->dev->kobj,686&aca_attr->attr,687"ras");688}689690int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,691const char *name, const struct aca_info *ras_info, void *data)692{693struct amdgpu_aca *aca = &adev->aca;694int ret;695696if (!amdgpu_aca_is_enabled(adev))697return 0;698699ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);700if (ret)701return ret;702703return add_aca_sysfs(adev, handle);704}705706static void remove_aca_handle(struct aca_handle *handle)707{708struct aca_handle_manager *mgr = handle->mgr;709710aca_fini_error_cache(handle);711list_del(&handle->node);712mgr->nr_handles--;713}714715static void remove_aca_sysfs(struct aca_handle *handle)716{717struct amdgpu_device *adev = handle->adev;718struct device_attribute *aca_attr = &handle->aca_attr;719720if (adev->dev->kobj.sd)721sysfs_remove_file_from_group(&adev->dev->kobj,722&aca_attr->attr,723"ras");724}725726void amdgpu_aca_remove_handle(struct aca_handle *handle)727{728if (!handle || list_empty(&handle->node))729return;730731remove_aca_sysfs(handle);732remove_aca_handle(handle);733}734735static int aca_manager_init(struct aca_handle_manager *mgr)736{737INIT_LIST_HEAD(&mgr->list);738mgr->nr_handles = 0;739740return 0;741}742743static void aca_manager_fini(struct aca_handle_manager *mgr)744{745struct aca_handle *handle, *tmp;746747if (list_empty(&mgr->list))748return;749750list_for_each_entry_safe(handle, tmp, &mgr->list, node)751amdgpu_aca_remove_handle(handle);752}753754bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)755{756return (adev->aca.is_enabled ||757adev->debug_enable_ras_aca);758}759760int amdgpu_aca_init(struct amdgpu_device *adev)761{762struct amdgpu_aca *aca = &adev->aca;763int ret;764765atomic_set(&aca->ue_update_flag, 0);766767ret = aca_manager_init(&aca->mgr);768if (ret)769return ret;770771return 0;772}773774void amdgpu_aca_fini(struct amdgpu_device *adev)775{776struct amdgpu_aca *aca = &adev->aca;777778aca_manager_fini(&aca->mgr);779780atomic_set(&aca->ue_update_flag, 0);781}782783int amdgpu_aca_reset(struct amdgpu_device *adev)784{785struct amdgpu_aca *aca = &adev->aca;786787atomic_set(&aca->ue_update_flag, 0);788789return 0;790}791792void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)793{794struct amdgpu_aca *aca = &adev->aca;795796WARN_ON(aca->smu_funcs);797aca->smu_funcs = smu_funcs;798}799800int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)801{802u64 ipid;803u32 instidhi, instidlo;804805if (!bank || !info)806return -EINVAL;807808ipid = bank->regs[ACA_REG_IDX_IPID];809info->hwid = ACA_REG__IPID__HARDWAREID(ipid);810info->mcatype = ACA_REG__IPID__MCATYPE(ipid);811/*812* Unfied DieID Format: SAASS. A:AID, S:Socket.813* Unfied DieID[4:4] = InstanceId[0:0]814* Unfied DieID[0:3] = InstanceIdHi[0:3]815*/816instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);817instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);818info->die_id = ((instidhi >> 2) & 0x03);819info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);820821return 0;822}823824static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)825{826struct amdgpu_aca *aca = &adev->aca;827const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;828829if (!smu_funcs || !smu_funcs->parse_error_code)830return -EOPNOTSUPP;831832return smu_funcs->parse_error_code(adev, bank);833}834835int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)836{837int i, error_code;838839if (!bank || !err_codes)840return -EINVAL;841842error_code = aca_bank_get_error_code(adev, bank);843if (error_code < 0)844return error_code;845846for (i = 0; i < size; i++) {847if (err_codes[i] == error_code)848return 0;849}850851return -EINVAL;852}853854int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)855{856struct amdgpu_aca *aca = &adev->aca;857const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;858859if (!smu_funcs || !smu_funcs->set_debug_mode)860return -EOPNOTSUPP;861862return smu_funcs->set_debug_mode(adev, en);863}864865#if defined(CONFIG_DEBUG_FS)866static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)867{868struct amdgpu_device *adev = (struct amdgpu_device *)data;869int ret;870871ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);872if (ret)873return ret;874875dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");876877return 0;878}879880static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)881{882struct aca_bank_info info;883int i, ret;884885ret = aca_bank_info_decode(bank, &info);886if (ret)887return;888889seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE");890seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",891idx, info.socket_id, info.die_id, info.hwid, info.mcatype);892893for (i = 0; i < ARRAY_SIZE(aca_regs); i++)894seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);895}896897struct aca_dump_context {898struct seq_file *m;899int idx;900};901902static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,903enum aca_smu_type type, void *data)904{905struct aca_dump_context *ctx = (struct aca_dump_context *)data;906907aca_dump_entry(ctx->m, bank, type, ctx->idx++);908909return handler_aca_log_bank_error(handle, bank, type, NULL);910}911912static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)913{914struct amdgpu_device *adev = (struct amdgpu_device *)m->private;915struct aca_dump_context context = {916.m = m,917.idx = 0,918};919920return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);921}922923static int aca_dump_ce_show(struct seq_file *m, void *unused)924{925return aca_dump_show(m, ACA_SMU_TYPE_CE);926}927928static int aca_dump_ce_open(struct inode *inode, struct file *file)929{930return single_open(file, aca_dump_ce_show, inode->i_private);931}932933static const struct file_operations aca_ce_dump_debug_fops = {934.owner = THIS_MODULE,935.open = aca_dump_ce_open,936.read = seq_read,937.llseek = seq_lseek,938.release = single_release,939};940941static int aca_dump_ue_show(struct seq_file *m, void *unused)942{943return aca_dump_show(m, ACA_SMU_TYPE_UE);944}945946static int aca_dump_ue_open(struct inode *inode, struct file *file)947{948return single_open(file, aca_dump_ue_show, inode->i_private);949}950951static const struct file_operations aca_ue_dump_debug_fops = {952.owner = THIS_MODULE,953.open = aca_dump_ue_open,954.read = seq_read,955.llseek = seq_lseek,956.release = single_release,957};958959DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");960#endif961962void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)963{964#if defined(CONFIG_DEBUG_FS)965if (!root)966return;967968debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);969debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);970debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);971#endif972}973974975