2 * Copyright 2023 Advanced Micro Devices, Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
24 #include <linux/list.h>
26 #include "amdgpu_aca.h"
27 #include "amdgpu_ras.h"
29 #define ACA_BANK_HWID(type, hwid, mcatype) [ACA_HWIP_TYPE_##type] = {hwid, mcatype}
31 typedef int bank_handler_t(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type, void *data);
35 struct list_head list;
43 static struct aca_hwip aca_hwid_mcatypes[ACA_HWIP_TYPE_COUNT] = {
44 ACA_BANK_HWID(SMU, 0x01, 0x01),
45 ACA_BANK_HWID(PCS_XGMI, 0x50, 0x00),
46 ACA_BANK_HWID(UMC, 0x96, 0x00),
49 static void aca_banks_init(struct aca_banks *banks)
54 memset(banks, 0, sizeof(*banks));
55 INIT_LIST_HEAD(&banks->list);
58 static int aca_banks_add_bank(struct aca_banks *banks, struct aca_bank *bank)
60 struct aca_bank_node *node;
65 node = kvzalloc(sizeof(*node), GFP_KERNEL);
69 memcpy(&node->bank, bank, sizeof(*bank));
71 INIT_LIST_HEAD(&node->node);
72 list_add_tail(&node->node, &banks->list);
79 static void aca_banks_release(struct aca_banks *banks)
81 struct aca_bank_node *node, *tmp;
83 list_for_each_entry_safe(node, tmp, &banks->list, node) {
84 list_del(&node->node);
89 static int aca_smu_get_valid_aca_count(struct amdgpu_device *adev, enum aca_smu_type type, u32 *count)
91 struct amdgpu_aca *aca = &adev->aca;
92 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
97 if (!smu_funcs || !smu_funcs->get_valid_aca_count)
100 return smu_funcs->get_valid_aca_count(adev, type, count);
103 static struct aca_regs_dump {
107 {"CONTROL", ACA_REG_IDX_CTL},
108 {"STATUS", ACA_REG_IDX_STATUS},
109 {"ADDR", ACA_REG_IDX_ADDR},
110 {"MISC", ACA_REG_IDX_MISC0},
111 {"CONFIG", ACA_REG_IDX_CONFG},
112 {"IPID", ACA_REG_IDX_IPID},
113 {"SYND", ACA_REG_IDX_SYND},
114 {"DESTAT", ACA_REG_IDX_DESTAT},
115 {"DEADDR", ACA_REG_IDX_DEADDR},
116 {"CONTROL_MASK", ACA_REG_IDX_CTL_MASK},
119 static void aca_smu_bank_dump(struct amdgpu_device *adev, int idx, int total, struct aca_bank *bank,
120 struct ras_query_context *qctx)
122 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID;
125 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n");
126 /* plus 1 for output format, e.g: ACA[08/08]: xxxx */
127 for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
128 RAS_EVENT_LOG(adev, event_id, HW_ERR "ACA[%02d/%02d].%s=0x%016llx\n",
129 idx + 1, total, aca_regs[i].name, bank->regs[aca_regs[i].reg_idx]);
132 static int aca_smu_get_valid_aca_banks(struct amdgpu_device *adev, enum aca_smu_type type,
133 int start, int count,
134 struct aca_banks *banks, struct ras_query_context *qctx)
136 struct amdgpu_aca *aca = &adev->aca;
137 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
138 struct aca_bank bank;
139 int i, max_count, ret;
144 if (!smu_funcs || !smu_funcs->get_valid_aca_bank)
148 case ACA_SMU_TYPE_UE:
149 max_count = smu_funcs->max_ue_bank_count;
151 case ACA_SMU_TYPE_CE:
152 max_count = smu_funcs->max_ce_bank_count;
158 if (start + count >= max_count)
161 count = min_t(int, count, max_count);
162 for (i = 0; i < count; i++) {
163 memset(&bank, 0, sizeof(bank));
164 ret = smu_funcs->get_valid_aca_bank(adev, type, start + i, &bank);
170 aca_smu_bank_dump(adev, i, count, &bank, qctx);
172 ret = aca_banks_add_bank(banks, &bank);
180 static bool aca_bank_hwip_is_matched(struct aca_bank *bank, enum aca_hwip_type type)
183 struct aca_hwip *hwip;
187 if (!bank || type == ACA_HWIP_TYPE_UNKNOW)
190 hwip = &aca_hwid_mcatypes[type];
194 ipid = bank->regs[ACA_REG_IDX_IPID];
195 hwid = ACA_REG__IPID__HARDWAREID(ipid);
196 mcatype = ACA_REG__IPID__MCATYPE(ipid);
198 return hwip->hwid == hwid && hwip->mcatype == mcatype;
201 static bool aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
203 const struct aca_bank_ops *bank_ops = handle->bank_ops;
205 if (!aca_bank_hwip_is_matched(bank, handle->hwip))
208 if (!bank_ops->aca_bank_is_valid)
211 return bank_ops->aca_bank_is_valid(handle, bank, type, handle->data);
214 static struct aca_bank_error *new_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
216 struct aca_bank_error *bank_error;
218 bank_error = kvzalloc(sizeof(*bank_error), GFP_KERNEL);
222 INIT_LIST_HEAD(&bank_error->node);
223 memcpy(&bank_error->info, info, sizeof(*info));
225 mutex_lock(&aerr->lock);
226 list_add_tail(&bank_error->node, &aerr->list);
227 mutex_unlock(&aerr->lock);
232 static struct aca_bank_error *find_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
234 struct aca_bank_error *bank_error = NULL;
235 struct aca_bank_info *tmp_info;
238 mutex_lock(&aerr->lock);
239 list_for_each_entry(bank_error, &aerr->list, node) {
240 tmp_info = &bank_error->info;
241 if (tmp_info->socket_id == info->socket_id &&
242 tmp_info->die_id == info->die_id) {
249 mutex_unlock(&aerr->lock);
251 return found ? bank_error : NULL;
254 static void aca_bank_error_remove(struct aca_error *aerr, struct aca_bank_error *bank_error)
256 if (!aerr || !bank_error)
259 list_del(&bank_error->node);
265 static struct aca_bank_error *get_bank_error(struct aca_error *aerr, struct aca_bank_info *info)
267 struct aca_bank_error *bank_error;
272 bank_error = find_bank_error(aerr, info);
276 return new_bank_error(aerr, info);
279 int aca_error_cache_log_bank_error(struct aca_handle *handle, struct aca_bank_info *info,
280 enum aca_error_type type, u64 count)
282 struct aca_error_cache *error_cache = &handle->error_cache;
283 struct aca_bank_error *bank_error;
284 struct aca_error *aerr;
286 if (!handle || !info || type >= ACA_ERROR_TYPE_COUNT)
292 aerr = &error_cache->errors[type];
293 bank_error = get_bank_error(aerr, info);
297 bank_error->count += count;
302 static int aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank, enum aca_smu_type type)
304 const struct aca_bank_ops *bank_ops = handle->bank_ops;
309 if (!bank_ops->aca_bank_parser)
312 return bank_ops->aca_bank_parser(handle, bank, type,
316 static int handler_aca_log_bank_error(struct aca_handle *handle, struct aca_bank *bank,
317 enum aca_smu_type type, void *data)
321 ret = aca_bank_parser(handle, bank, type);
328 static int aca_dispatch_bank(struct aca_handle_manager *mgr, struct aca_bank *bank,
329 enum aca_smu_type type, bank_handler_t handler, void *data)
331 struct aca_handle *handle;
334 if (list_empty(&mgr->list))
337 list_for_each_entry(handle, &mgr->list, node) {
338 if (!aca_bank_is_valid(handle, bank, type))
341 ret = handler(handle, bank, type, data);
349 static int aca_dispatch_banks(struct aca_handle_manager *mgr, struct aca_banks *banks,
350 enum aca_smu_type type, bank_handler_t handler, void *data)
352 struct aca_bank_node *node;
353 struct aca_bank *bank;
359 /* pre check to avoid unnecessary operations */
360 if (list_empty(&mgr->list) || list_empty(&banks->list))
363 list_for_each_entry(node, &banks->list, node) {
366 ret = aca_dispatch_bank(mgr, bank, type, handler, data);
374 static bool aca_bank_should_update(struct amdgpu_device *adev, enum aca_smu_type type)
376 struct amdgpu_aca *aca = &adev->aca;
380 * Because the UE Valid MCA count will only be cleared after reset,
381 * in order to avoid repeated counting of the error count,
382 * the aca bank is only updated once during the gpu recovery stage.
384 if (type == ACA_SMU_TYPE_UE) {
385 if (amdgpu_ras_intr_triggered())
386 ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
388 atomic_set(&aca->ue_update_flag, 0);
394 static int aca_banks_update(struct amdgpu_device *adev, enum aca_smu_type type,
395 bank_handler_t handler, struct ras_query_context *qctx, void *data)
397 struct amdgpu_aca *aca = &adev->aca;
398 struct aca_banks banks;
402 if (list_empty(&aca->mgr.list))
405 if (!aca_bank_should_update(adev, type))
408 ret = aca_smu_get_valid_aca_count(adev, type, &count);
415 aca_banks_init(&banks);
417 ret = aca_smu_get_valid_aca_banks(adev, type, 0, count, &banks, qctx);
419 goto err_release_banks;
421 if (list_empty(&banks.list)) {
423 goto err_release_banks;
426 ret = aca_dispatch_banks(&aca->mgr, &banks, type,
429 goto err_release_banks;
432 aca_banks_release(&banks);
437 static int aca_log_aca_error_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data)
439 struct aca_bank_info *info;
440 struct amdgpu_smuio_mcm_config_info mcm_info;
443 if (type >= ACA_ERROR_TYPE_COUNT)
446 count = bank_error->count;
450 info = &bank_error->info;
451 mcm_info.die_id = info->die_id;
452 mcm_info.socket_id = info->socket_id;
455 case ACA_ERROR_TYPE_UE:
456 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, NULL, count);
458 case ACA_ERROR_TYPE_CE:
459 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, NULL, count);
461 case ACA_ERROR_TYPE_DEFERRED:
462 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, NULL, count);
471 static int aca_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data)
473 struct aca_error_cache *error_cache = &handle->error_cache;
474 struct aca_error *aerr = &error_cache->errors[type];
475 struct aca_bank_error *bank_error, *tmp;
477 mutex_lock(&aerr->lock);
479 if (list_empty(&aerr->list))
482 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node) {
483 aca_log_aca_error_data(bank_error, type, err_data);
484 aca_bank_error_remove(aerr, bank_error);
488 mutex_unlock(&aerr->lock);
493 static int __aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle, enum aca_error_type type,
494 struct ras_err_data *err_data, struct ras_query_context *qctx)
496 enum aca_smu_type smu_type;
500 case ACA_ERROR_TYPE_UE:
501 smu_type = ACA_SMU_TYPE_UE;
503 case ACA_ERROR_TYPE_CE:
504 case ACA_ERROR_TYPE_DEFERRED:
505 smu_type = ACA_SMU_TYPE_CE;
511 /* udpate aca bank to aca source error_cache first */
512 ret = aca_banks_update(adev, smu_type, handler_aca_log_bank_error, qctx, NULL);
516 return aca_log_aca_error(handle, type, err_data);
519 static bool aca_handle_is_valid(struct aca_handle *handle)
521 if (!handle->mask || !list_empty(&handle->node))
527 int amdgpu_aca_get_error_data(struct amdgpu_device *adev, struct aca_handle *handle,
528 enum aca_error_type type, struct ras_err_data *err_data,
529 struct ras_query_context *qctx)
531 if (!handle || !err_data)
534 if (aca_handle_is_valid(handle))
537 if ((type < 0) || (!(BIT(type) & handle->mask)))
540 return __aca_get_error_data(adev, handle, type, err_data, qctx);
543 static void aca_error_init(struct aca_error *aerr, enum aca_error_type type)
545 mutex_init(&aerr->lock);
546 INIT_LIST_HEAD(&aerr->list);
551 static void aca_init_error_cache(struct aca_handle *handle)
553 struct aca_error_cache *error_cache = &handle->error_cache;
556 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
557 aca_error_init(&error_cache->errors[type], type);
560 static void aca_error_fini(struct aca_error *aerr)
562 struct aca_bank_error *bank_error, *tmp;
564 mutex_lock(&aerr->lock);
565 list_for_each_entry_safe(bank_error, tmp, &aerr->list, node)
566 aca_bank_error_remove(aerr, bank_error);
568 mutex_destroy(&aerr->lock);
571 static void aca_fini_error_cache(struct aca_handle *handle)
573 struct aca_error_cache *error_cache = &handle->error_cache;
576 for (type = ACA_ERROR_TYPE_UE; type < ACA_ERROR_TYPE_COUNT; type++)
577 aca_error_fini(&error_cache->errors[type]);
580 static int add_aca_handle(struct amdgpu_device *adev, struct aca_handle_manager *mgr, struct aca_handle *handle,
581 const char *name, const struct aca_info *ras_info, void *data)
583 memset(handle, 0, sizeof(*handle));
588 handle->hwip = ras_info->hwip;
589 handle->mask = ras_info->mask;
590 handle->bank_ops = ras_info->bank_ops;
592 aca_init_error_cache(handle);
594 INIT_LIST_HEAD(&handle->node);
595 list_add_tail(&handle->node, &mgr->list);
601 static ssize_t aca_sysfs_read(struct device *dev,
602 struct device_attribute *attr, char *buf)
604 struct aca_handle *handle = container_of(attr, struct aca_handle, aca_attr);
606 /* NOTE: the aca cache will be auto cleared once read,
607 * So the driver should unify the query entry point, forward request to ras query interface directly */
608 return amdgpu_ras_aca_sysfs_read(dev, attr, handle, buf, handle->data);
611 static int add_aca_sysfs(struct amdgpu_device *adev, struct aca_handle *handle)
613 struct device_attribute *aca_attr = &handle->aca_attr;
615 snprintf(handle->attr_name, sizeof(handle->attr_name) - 1, "aca_%s", handle->name);
616 aca_attr->show = aca_sysfs_read;
617 aca_attr->attr.name = handle->attr_name;
618 aca_attr->attr.mode = S_IRUGO;
619 sysfs_attr_init(&aca_attr->attr);
621 return sysfs_add_file_to_group(&adev->dev->kobj,
626 int amdgpu_aca_add_handle(struct amdgpu_device *adev, struct aca_handle *handle,
627 const char *name, const struct aca_info *ras_info, void *data)
629 struct amdgpu_aca *aca = &adev->aca;
632 if (!amdgpu_aca_is_enabled(adev))
635 ret = add_aca_handle(adev, &aca->mgr, handle, name, ras_info, data);
639 return add_aca_sysfs(adev, handle);
642 static void remove_aca_handle(struct aca_handle *handle)
644 struct aca_handle_manager *mgr = handle->mgr;
646 aca_fini_error_cache(handle);
647 list_del(&handle->node);
651 static void remove_aca_sysfs(struct aca_handle *handle)
653 struct amdgpu_device *adev = handle->adev;
654 struct device_attribute *aca_attr = &handle->aca_attr;
656 if (adev->dev->kobj.sd)
657 sysfs_remove_file_from_group(&adev->dev->kobj,
662 void amdgpu_aca_remove_handle(struct aca_handle *handle)
664 if (!handle || list_empty(&handle->node))
667 remove_aca_sysfs(handle);
668 remove_aca_handle(handle);
671 static int aca_manager_init(struct aca_handle_manager *mgr)
673 INIT_LIST_HEAD(&mgr->list);
679 static void aca_manager_fini(struct aca_handle_manager *mgr)
681 struct aca_handle *handle, *tmp;
683 list_for_each_entry_safe(handle, tmp, &mgr->list, node)
684 amdgpu_aca_remove_handle(handle);
687 bool amdgpu_aca_is_enabled(struct amdgpu_device *adev)
689 return (adev->aca.is_enabled ||
690 adev->debug_enable_ras_aca);
693 int amdgpu_aca_init(struct amdgpu_device *adev)
695 struct amdgpu_aca *aca = &adev->aca;
698 atomic_set(&aca->ue_update_flag, 0);
700 ret = aca_manager_init(&aca->mgr);
707 void amdgpu_aca_fini(struct amdgpu_device *adev)
709 struct amdgpu_aca *aca = &adev->aca;
711 aca_manager_fini(&aca->mgr);
713 atomic_set(&aca->ue_update_flag, 0);
716 int amdgpu_aca_reset(struct amdgpu_device *adev)
718 struct amdgpu_aca *aca = &adev->aca;
720 atomic_set(&aca->ue_update_flag, 0);
725 void amdgpu_aca_set_smu_funcs(struct amdgpu_device *adev, const struct aca_smu_funcs *smu_funcs)
727 struct amdgpu_aca *aca = &adev->aca;
729 WARN_ON(aca->smu_funcs);
730 aca->smu_funcs = smu_funcs;
733 int aca_bank_info_decode(struct aca_bank *bank, struct aca_bank_info *info)
736 u32 instidhi, instidlo;
741 ipid = bank->regs[ACA_REG_IDX_IPID];
742 info->hwid = ACA_REG__IPID__HARDWAREID(ipid);
743 info->mcatype = ACA_REG__IPID__MCATYPE(ipid);
745 * Unfied DieID Format: SAASS. A:AID, S:Socket.
746 * Unfied DieID[4:4] = InstanceId[0:0]
747 * Unfied DieID[0:3] = InstanceIdHi[0:3]
749 instidhi = ACA_REG__IPID__INSTANCEIDHI(ipid);
750 instidlo = ACA_REG__IPID__INSTANCEIDLO(ipid);
751 info->die_id = ((instidhi >> 2) & 0x03);
752 info->socket_id = ((instidlo & 0x1) << 2) | (instidhi & 0x03);
757 static int aca_bank_get_error_code(struct amdgpu_device *adev, struct aca_bank *bank)
759 struct amdgpu_aca *aca = &adev->aca;
760 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
762 if (!smu_funcs || !smu_funcs->parse_error_code)
765 return smu_funcs->parse_error_code(adev, bank);
768 int aca_bank_check_error_codes(struct amdgpu_device *adev, struct aca_bank *bank, int *err_codes, int size)
772 if (!bank || !err_codes)
775 error_code = aca_bank_get_error_code(adev, bank);
779 for (i = 0; i < size; i++) {
780 if (err_codes[i] == error_code)
787 int amdgpu_aca_smu_set_debug_mode(struct amdgpu_device *adev, bool en)
789 struct amdgpu_aca *aca = &adev->aca;
790 const struct aca_smu_funcs *smu_funcs = aca->smu_funcs;
792 if (!smu_funcs || !smu_funcs->set_debug_mode)
795 return smu_funcs->set_debug_mode(adev, en);
798 #if defined(CONFIG_DEBUG_FS)
799 static int amdgpu_aca_smu_debug_mode_set(void *data, u64 val)
801 struct amdgpu_device *adev = (struct amdgpu_device *)data;
804 ret = amdgpu_ras_set_aca_debug_mode(adev, val ? true : false);
808 dev_info(adev->dev, "amdgpu set smu aca debug mode %s success\n", val ? "on" : "off");
813 static void aca_dump_entry(struct seq_file *m, struct aca_bank *bank, enum aca_smu_type type, int idx)
815 struct aca_bank_info info;
818 ret = aca_bank_info_decode(bank, &info);
822 seq_printf(m, "aca entry[%d].type: %s\n", idx, type == ACA_SMU_TYPE_UE ? "UE" : "CE");
823 seq_printf(m, "aca entry[%d].info: socketid:%d aid:%d hwid:0x%03x mcatype:0x%04x\n",
824 idx, info.socket_id, info.die_id, info.hwid, info.mcatype);
826 for (i = 0; i < ARRAY_SIZE(aca_regs); i++)
827 seq_printf(m, "aca entry[%d].regs[%d]: 0x%016llx\n", idx, aca_regs[i].reg_idx, bank->regs[aca_regs[i].reg_idx]);
830 struct aca_dump_context {
835 static int handler_aca_bank_dump(struct aca_handle *handle, struct aca_bank *bank,
836 enum aca_smu_type type, void *data)
838 struct aca_dump_context *ctx = (struct aca_dump_context *)data;
840 aca_dump_entry(ctx->m, bank, type, ctx->idx++);
842 return handler_aca_log_bank_error(handle, bank, type, NULL);
845 static int aca_dump_show(struct seq_file *m, enum aca_smu_type type)
847 struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
848 struct aca_dump_context context = {
853 return aca_banks_update(adev, type, handler_aca_bank_dump, NULL, (void *)&context);
856 static int aca_dump_ce_show(struct seq_file *m, void *unused)
858 return aca_dump_show(m, ACA_SMU_TYPE_CE);
861 static int aca_dump_ce_open(struct inode *inode, struct file *file)
863 return single_open(file, aca_dump_ce_show, inode->i_private);
866 static const struct file_operations aca_ce_dump_debug_fops = {
867 .owner = THIS_MODULE,
868 .open = aca_dump_ce_open,
871 .release = single_release,
874 static int aca_dump_ue_show(struct seq_file *m, void *unused)
876 return aca_dump_show(m, ACA_SMU_TYPE_UE);
879 static int aca_dump_ue_open(struct inode *inode, struct file *file)
881 return single_open(file, aca_dump_ue_show, inode->i_private);
884 static const struct file_operations aca_ue_dump_debug_fops = {
885 .owner = THIS_MODULE,
886 .open = aca_dump_ue_open,
889 .release = single_release,
892 DEFINE_DEBUGFS_ATTRIBUTE(aca_debug_mode_fops, NULL, amdgpu_aca_smu_debug_mode_set, "%llu\n");
895 void amdgpu_aca_smu_debugfs_init(struct amdgpu_device *adev, struct dentry *root)
897 #if defined(CONFIG_DEBUG_FS)
901 debugfs_create_file("aca_debug_mode", 0200, root, adev, &aca_debug_mode_fops);
902 debugfs_create_file("aca_ue_dump", 0400, root, adev, &aca_ue_dump_debug_fops);
903 debugfs_create_file("aca_ce_dump", 0400, root, adev, &aca_ce_dump_debug_fops);