2 * Copyright 2019 Advanced Micro Devices, Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
24 #include <linux/sort.h>
27 #define MAX_UMC_POISON_POLLING_TIME_SYNC 20 //ms
29 #define MAX_UMC_HASH_STRING_SIZE 256
31 static int amdgpu_umc_convert_error_address(struct amdgpu_device *adev,
32 struct ras_err_data *err_data, uint64_t err_addr,
33 uint32_t ch_inst, uint32_t umc_inst)
35 switch (amdgpu_ip_version(adev, UMC_HWIP, 0)) {
36 case IP_VERSION(6, 7, 0):
37 umc_v6_7_convert_error_address(adev,
38 err_data, err_addr, ch_inst, umc_inst);
42 "UMC address to Physical address translation is not supported\n");
43 return AMDGPU_RAS_FAIL;
46 return AMDGPU_RAS_SUCCESS;
49 int amdgpu_umc_page_retirement_mca(struct amdgpu_device *adev,
50 uint64_t err_addr, uint32_t ch_inst, uint32_t umc_inst)
52 struct ras_err_data err_data;
55 ret = amdgpu_ras_error_data_init(&err_data);
60 kcalloc(adev->umc.max_ras_err_cnt_per_query,
61 sizeof(struct eeprom_table_record), GFP_KERNEL);
62 if (!err_data.err_addr) {
64 "Failed to alloc memory for umc error record in MCA notifier!\n");
65 ret = AMDGPU_RAS_FAIL;
66 goto out_fini_err_data;
69 err_data.err_addr_len = adev->umc.max_ras_err_cnt_per_query;
72 * Translate UMC channel address to Physical address
74 ret = amdgpu_umc_convert_error_address(adev, &err_data, err_addr,
77 goto out_free_err_addr;
79 if (amdgpu_bad_page_threshold != 0) {
80 amdgpu_ras_add_bad_pages(adev, err_data.err_addr,
81 err_data.err_addr_cnt);
82 amdgpu_ras_save_bad_pages(adev, NULL);
86 kfree(err_data.err_addr);
89 amdgpu_ras_error_data_fini(&err_data);
94 void amdgpu_umc_handle_bad_pages(struct amdgpu_device *adev,
95 void *ras_error_status)
97 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status;
98 struct amdgpu_ras *con = amdgpu_ras_get_context(adev);
99 unsigned int error_query_mode;
101 unsigned long err_count;
103 amdgpu_ras_get_error_query_mode(adev, &error_query_mode);
105 mutex_lock(&con->page_retirement_lock);
106 ret = amdgpu_dpm_get_ecc_info(adev, (void *)&(con->umc_ecc));
107 if (ret == -EOPNOTSUPP &&
108 error_query_mode == AMDGPU_RAS_DIRECT_ERROR_QUERY) {
109 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops &&
110 adev->umc.ras->ras_block.hw_ops->query_ras_error_count)
111 adev->umc.ras->ras_block.hw_ops->query_ras_error_count(adev, ras_error_status);
113 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops &&
114 adev->umc.ras->ras_block.hw_ops->query_ras_error_address &&
115 adev->umc.max_ras_err_cnt_per_query) {
117 kcalloc(adev->umc.max_ras_err_cnt_per_query,
118 sizeof(struct eeprom_table_record), GFP_KERNEL);
120 /* still call query_ras_error_address to clear error status
121 * even NOMEM error is encountered
123 if(!err_data->err_addr)
124 dev_warn(adev->dev, "Failed to alloc memory for "
125 "umc error address record!\n");
127 err_data->err_addr_len = adev->umc.max_ras_err_cnt_per_query;
129 /* umc query_ras_error_address is also responsible for clearing
132 adev->umc.ras->ras_block.hw_ops->query_ras_error_address(adev, ras_error_status);
134 } else if (error_query_mode == AMDGPU_RAS_FIRMWARE_ERROR_QUERY ||
135 (!ret && error_query_mode == AMDGPU_RAS_DIRECT_ERROR_QUERY)) {
137 adev->umc.ras->ecc_info_query_ras_error_count)
138 adev->umc.ras->ecc_info_query_ras_error_count(adev, ras_error_status);
141 adev->umc.ras->ecc_info_query_ras_error_address &&
142 adev->umc.max_ras_err_cnt_per_query) {
144 kcalloc(adev->umc.max_ras_err_cnt_per_query,
145 sizeof(struct eeprom_table_record), GFP_KERNEL);
147 /* still call query_ras_error_address to clear error status
148 * even NOMEM error is encountered
150 if(!err_data->err_addr)
151 dev_warn(adev->dev, "Failed to alloc memory for "
152 "umc error address record!\n");
154 err_data->err_addr_len = adev->umc.max_ras_err_cnt_per_query;
156 /* umc query_ras_error_address is also responsible for clearing
159 adev->umc.ras->ecc_info_query_ras_error_address(adev, ras_error_status);
163 /* only uncorrectable error needs gpu reset */
164 if (err_data->ue_count || err_data->de_count) {
165 err_count = err_data->ue_count + err_data->de_count;
166 if ((amdgpu_bad_page_threshold != 0) &&
167 err_data->err_addr_cnt) {
168 amdgpu_ras_add_bad_pages(adev, err_data->err_addr,
169 err_data->err_addr_cnt);
170 amdgpu_ras_save_bad_pages(adev, &err_count);
172 amdgpu_dpm_send_hbm_bad_pages_num(adev, con->eeprom_control.ras_num_recs);
174 if (con->update_channel_flag == true) {
175 amdgpu_dpm_send_hbm_bad_channel_flag(adev, con->eeprom_control.bad_channel_bitmap);
176 con->update_channel_flag = false;
181 kfree(err_data->err_addr);
182 err_data->err_addr = NULL;
184 mutex_unlock(&con->page_retirement_lock);
187 static int amdgpu_umc_do_page_retirement(struct amdgpu_device *adev,
188 void *ras_error_status,
189 struct amdgpu_iv_entry *entry,
192 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status;
193 struct amdgpu_ras *con = amdgpu_ras_get_context(adev);
195 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev);
196 amdgpu_umc_handle_bad_pages(adev, ras_error_status);
198 if ((err_data->ue_count || err_data->de_count) &&
199 (reset || (con && con->is_rma))) {
200 con->gpu_reset_flags |= reset;
201 amdgpu_ras_reset_gpu(adev);
204 return AMDGPU_RAS_SUCCESS;
207 int amdgpu_umc_bad_page_polling_timeout(struct amdgpu_device *adev,
208 uint32_t reset, uint32_t timeout_ms)
210 struct ras_err_data err_data;
211 struct ras_common_if head = {
212 .block = AMDGPU_RAS_BLOCK__UMC,
214 struct ras_manager *obj = amdgpu_ras_find_obj(adev, &head);
215 struct amdgpu_ras *con = amdgpu_ras_get_context(adev);
216 uint32_t timeout = timeout_ms;
218 memset(&err_data, 0, sizeof(err_data));
219 amdgpu_ras_error_data_init(&err_data);
223 amdgpu_umc_handle_bad_pages(adev, &err_data);
225 if (timeout && !err_data.de_count) {
230 } while (timeout && !err_data.de_count);
233 dev_warn(adev->dev, "Can't find bad pages\n");
235 if (err_data.de_count)
236 dev_info(adev->dev, "%ld new deferred hardware errors detected\n", err_data.de_count);
239 obj->err_data.ue_count += err_data.ue_count;
240 obj->err_data.ce_count += err_data.ce_count;
241 obj->err_data.de_count += err_data.de_count;
244 amdgpu_ras_error_data_fini(&err_data);
246 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev);
248 if (reset || (err_data.err_addr_cnt && con && con->is_rma)) {
249 con->gpu_reset_flags |= reset;
250 amdgpu_ras_reset_gpu(adev);
256 int amdgpu_umc_pasid_poison_handler(struct amdgpu_device *adev,
257 enum amdgpu_ras_block block, uint16_t pasid,
258 pasid_notify pasid_fn, void *data, uint32_t reset)
260 int ret = AMDGPU_RAS_SUCCESS;
262 if (adev->gmc.xgmi.connected_to_cpu ||
263 adev->gmc.is_app_apu) {
265 /* MCA poison handler is only responsible for GPU reset,
266 * let MCA notifier do page retirement.
268 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev);
269 amdgpu_ras_reset_gpu(adev);
274 if (!amdgpu_sriov_vf(adev)) {
275 if (amdgpu_ip_version(adev, UMC_HWIP, 0) < IP_VERSION(12, 0, 0)) {
276 struct ras_err_data err_data;
277 struct ras_common_if head = {
278 .block = AMDGPU_RAS_BLOCK__UMC,
280 struct ras_manager *obj = amdgpu_ras_find_obj(adev, &head);
282 ret = amdgpu_ras_error_data_init(&err_data);
286 ret = amdgpu_umc_do_page_retirement(adev, &err_data, NULL, reset);
288 if (ret == AMDGPU_RAS_SUCCESS && obj) {
289 obj->err_data.ue_count += err_data.ue_count;
290 obj->err_data.ce_count += err_data.ce_count;
291 obj->err_data.de_count += err_data.de_count;
294 amdgpu_ras_error_data_fini(&err_data);
296 struct amdgpu_ras *con = amdgpu_ras_get_context(adev);
299 ret = amdgpu_ras_put_poison_req(adev,
300 block, pasid, pasid_fn, data, reset);
302 atomic_inc(&con->page_retirement_req_cnt);
303 wake_up(&con->page_retirement_wq);
307 if (adev->virt.ops && adev->virt.ops->ras_poison_handler)
308 adev->virt.ops->ras_poison_handler(adev, block);
311 "No ras_poison_handler interface in SRIOV!\n");
317 int amdgpu_umc_poison_handler(struct amdgpu_device *adev,
318 enum amdgpu_ras_block block, uint32_t reset)
320 return amdgpu_umc_pasid_poison_handler(adev,
321 block, 0, NULL, NULL, reset);
324 int amdgpu_umc_process_ras_data_cb(struct amdgpu_device *adev,
325 void *ras_error_status,
326 struct amdgpu_iv_entry *entry)
328 return amdgpu_umc_do_page_retirement(adev, ras_error_status, entry,
329 AMDGPU_RAS_GPU_RESET_MODE1_RESET);
332 int amdgpu_umc_ras_sw_init(struct amdgpu_device *adev)
335 struct amdgpu_umc_ras *ras;
342 err = amdgpu_ras_register_ras_block(adev, &ras->ras_block);
344 dev_err(adev->dev, "Failed to register umc ras block!\n");
348 strcpy(adev->umc.ras->ras_block.ras_comm.name, "umc");
349 ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__UMC;
350 ras->ras_block.ras_comm.type = AMDGPU_RAS_ERROR__MULTI_UNCORRECTABLE;
351 adev->umc.ras_if = &ras->ras_block.ras_comm;
353 if (!ras->ras_block.ras_late_init)
354 ras->ras_block.ras_late_init = amdgpu_umc_ras_late_init;
356 if (!ras->ras_block.ras_cb)
357 ras->ras_block.ras_cb = amdgpu_umc_process_ras_data_cb;
362 int amdgpu_umc_ras_late_init(struct amdgpu_device *adev, struct ras_common_if *ras_block)
366 r = amdgpu_ras_block_late_init(adev, ras_block);
370 if (amdgpu_ras_is_supported(adev, ras_block->block)) {
371 r = amdgpu_irq_get(adev, &adev->gmc.ecc_irq, 0);
376 /* ras init of specific umc version */
378 adev->umc.ras->err_cnt_init)
379 adev->umc.ras->err_cnt_init(adev);
384 amdgpu_ras_block_late_fini(adev, ras_block);
388 int amdgpu_umc_process_ecc_irq(struct amdgpu_device *adev,
389 struct amdgpu_irq_src *source,
390 struct amdgpu_iv_entry *entry)
392 struct ras_common_if *ras_if = adev->umc.ras_if;
393 struct ras_dispatch_if ih_data = {
400 ih_data.head = *ras_if;
402 amdgpu_ras_interrupt_dispatch(adev, &ih_data);
406 int amdgpu_umc_fill_error_record(struct ras_err_data *err_data,
408 uint64_t retired_page,
409 uint32_t channel_index,
412 struct eeprom_table_record *err_rec;
415 !err_data->err_addr ||
416 (err_data->err_addr_cnt >= err_data->err_addr_len))
419 err_rec = &err_data->err_addr[err_data->err_addr_cnt];
421 err_rec->address = err_addr;
422 /* page frame address is saved */
423 err_rec->retired_page = retired_page >> AMDGPU_GPU_PAGE_SHIFT;
424 err_rec->ts = (uint64_t)ktime_get_real_seconds();
425 err_rec->err_type = AMDGPU_RAS_EEPROM_ERR_NON_RECOVERABLE;
427 err_rec->mem_channel = channel_index;
428 err_rec->mcumc_id = umc_inst;
430 err_data->err_addr_cnt++;
435 int amdgpu_umc_loop_channels(struct amdgpu_device *adev,
436 umc_func func, void *data)
438 uint32_t node_inst = 0;
439 uint32_t umc_inst = 0;
440 uint32_t ch_inst = 0;
443 if (adev->umc.node_inst_num) {
444 LOOP_UMC_EACH_NODE_INST_AND_CH(node_inst, umc_inst, ch_inst) {
445 ret = func(adev, node_inst, umc_inst, ch_inst, data);
447 dev_err(adev->dev, "Node %d umc %d ch %d func returns %d\n",
448 node_inst, umc_inst, ch_inst, ret);
453 LOOP_UMC_INST_AND_CH(umc_inst, ch_inst) {
454 ret = func(adev, 0, umc_inst, ch_inst, data);
456 dev_err(adev->dev, "Umc %d ch %d func returns %d\n",
457 umc_inst, ch_inst, ret);
466 int amdgpu_umc_update_ecc_status(struct amdgpu_device *adev,
467 uint64_t status, uint64_t ipid, uint64_t addr)
469 if (adev->umc.ras->update_ecc_status)
470 return adev->umc.ras->update_ecc_status(adev,
475 static int amdgpu_umc_uint64_cmp(const void *a, const void *b)
477 uint64_t *addr_a = (uint64_t *)a;
478 uint64_t *addr_b = (uint64_t *)b;
480 if (*addr_a > *addr_b)
482 else if (*addr_a < *addr_b)
488 /* Use string hash to avoid logging the same bad pages repeatedly */
489 int amdgpu_umc_build_pages_hash(struct amdgpu_device *adev,
490 uint64_t *pfns, int len, uint64_t *val)
492 struct amdgpu_ras *con = amdgpu_ras_get_context(adev);
493 char buf[MAX_UMC_HASH_STRING_SIZE] = {0};
494 int offset = 0, i = 0;
500 sort(pfns, len, sizeof(uint64_t), amdgpu_umc_uint64_cmp, NULL);
502 for (i = 0; i < len; i++)
503 offset += snprintf(&buf[offset], sizeof(buf) - offset, "%llx", pfns[i]);
505 hash_val = siphash(buf, offset, &con->umc_ecc_log.ecc_key);
512 int amdgpu_umc_logs_ecc_err(struct amdgpu_device *adev,
513 struct radix_tree_root *ecc_tree, struct ras_ecc_err *ecc_err)
515 struct amdgpu_ras *con = amdgpu_ras_get_context(adev);
516 struct ras_ecc_log_info *ecc_log;
519 ecc_log = &con->umc_ecc_log;
521 mutex_lock(&ecc_log->lock);
522 ret = radix_tree_insert(ecc_tree, ecc_err->hash_index, ecc_err);
524 struct ras_err_pages *err_pages = &ecc_err->err_pages;
528 for (i = 0; i < err_pages->count; i++)
529 amdgpu_ras_reserve_page(adev, err_pages->pfn[i]);
531 radix_tree_tag_set(ecc_tree,
532 ecc_err->hash_index, UMC_ECC_NEW_DETECTED_TAG);
534 mutex_unlock(&ecc_log->lock);