Home
last modified time | relevance | path

Searched refs:err_data (Results 1 – 25 of 43) sorted by relevance

12

/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_umc.c33 struct ras_err_data *err_data, uint64_t err_addr, in amdgpu_umc_convert_error_address() argument
39 err_data, err_addr, ch_inst, umc_inst); in amdgpu_umc_convert_error_address()
53 struct ras_err_data err_data; in amdgpu_umc_page_retirement_mca() local
56 ret = amdgpu_ras_error_data_init(&err_data); in amdgpu_umc_page_retirement_mca()
60 err_data.err_addr = in amdgpu_umc_page_retirement_mca()
63 if (!err_data.err_addr) { in amdgpu_umc_page_retirement_mca()
70 err_data.err_addr_len = adev->umc.max_ras_err_cnt_per_query; in amdgpu_umc_page_retirement_mca()
75 ret = amdgpu_umc_convert_error_address(adev, &err_data, err_addr, in amdgpu_umc_page_retirement_mca()
81 amdgpu_ras_add_bad_pages(adev, err_data.err_addr, in amdgpu_umc_page_retirement_mca()
82 err_data.err_addr_cnt, false); in amdgpu_umc_page_retirement_mca()
[all …]
H A Damdgpu_ras.c178 struct ras_err_data err_data; in amdgpu_reserve_page_direct() local
195 ret = amdgpu_ras_error_data_init(&err_data); in amdgpu_reserve_page_direct()
200 err_data.err_addr = &err_rec; in amdgpu_reserve_page_direct()
201 amdgpu_umc_fill_error_record(&err_data, address, address, 0, 0); in amdgpu_reserve_page_direct()
204 amdgpu_ras_add_bad_pages(adev, err_data.err_addr, in amdgpu_reserve_page_direct()
205 err_data.err_addr_cnt, false); in amdgpu_reserve_page_direct()
209 amdgpu_ras_error_data_fini(&err_data); in amdgpu_reserve_page_direct()
738 amdgpu_ras_error_data_fini(&obj->err_data); in put_obj()
770 if (amdgpu_ras_error_data_init(&obj->err_data)) in amdgpu_ras_create_obj()
1097 static void amdgpu_ras_get_ecc_info(struct amdgpu_device *adev, struct ras_err_data *err_data) in amdgpu_ras_get_ecc_info() argument
[all …]
H A Dumc_v8_7.c93 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in umc_v8_7_ecc_info_query_ras_error_count() local
104 &(err_data->ce_count)); in umc_v8_7_ecc_info_query_ras_error_count()
107 &(err_data->ue_count)); in umc_v8_7_ecc_info_query_ras_error_count()
112 struct ras_err_data *err_data, uint64_t err_addr, in umc_v8_7_convert_error_address() argument
126 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v8_7_convert_error_address()
131 struct ras_err_data *err_data, in umc_v8_7_ecc_info_query_error_address() argument
145 if (!err_data->err_addr) in umc_v8_7_ecc_info_query_error_address()
155 umc_v8_7_convert_error_address(adev, err_data, err_addr, in umc_v8_7_ecc_info_query_error_address()
163 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in umc_v8_7_ecc_info_query_ras_error_address() local
174 err_data, in umc_v8_7_ecc_info_query_ras_error_address()
[all …]
H A Dumc_v8_10.c147 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_10_query_ecc_error_count() local
153 &(err_data->ce_count)); in umc_v8_10_query_ecc_error_count()
156 &(err_data->ue_count)); in umc_v8_10_query_ecc_error_count()
206 struct ras_err_data *err_data, uint64_t err_addr, in umc_v8_10_convert_error_address() argument
239 amdgpu_umc_fill_error_record(err_data, na_err_addr, in umc_v8_10_convert_error_address()
251 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_10_query_error_address() local
262 if (!err_data->err_addr) { in umc_v8_10_query_error_address()
277 umc_v8_10_convert_error_address(adev, err_data, err_addr, in umc_v8_10_query_error_address()
383 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_10_ecc_info_query_ecc_error_count() local
387 &(err_data->ce_count)); in umc_v8_10_ecc_info_query_ecc_error_count()
[all …]
H A Dumc_v6_7.c167 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v6_7_ecc_info_querry_ecc_error_count() local
171 &(err_data->ce_count)); in umc_v6_7_ecc_info_querry_ecc_error_count()
175 &(err_data->ue_count)); in umc_v6_7_ecc_info_querry_ecc_error_count()
188 struct ras_err_data *err_data, uint64_t err_addr, in umc_v6_7_convert_error_address() argument
211 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v6_7_convert_error_address()
217 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v6_7_convert_error_address()
229 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v6_7_ecc_info_query_error_address() local
237 if (!err_data->err_addr) in umc_v6_7_ecc_info_query_error_address()
247 umc_v6_7_convert_error_address(adev, err_data, err_addr, in umc_v6_7_ecc_info_query_error_address()
416 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v6_7_query_ecc_error_count() local
[all …]
H A Dnbio_v7_4.c363 struct ras_err_data err_data; in nbio_v7_4_handle_ras_controller_intr_no_bifring() local
366 if (amdgpu_ras_error_data_init(&err_data)) in nbio_v7_4_handle_ras_controller_intr_no_bifring()
391 nbio_v7_4_query_ras_error_count(adev, &err_data); in nbio_v7_4_handle_ras_controller_intr_no_bifring()
394 obj->err_data.ue_count += err_data.ue_count; in nbio_v7_4_handle_ras_controller_intr_no_bifring()
395 obj->err_data.ce_count += err_data.ce_count; in nbio_v7_4_handle_ras_controller_intr_no_bifring()
397 if (err_data.ce_count) in nbio_v7_4_handle_ras_controller_intr_no_bifring()
400 obj->err_data.ce_count, in nbio_v7_4_handle_ras_controller_intr_no_bifring()
403 if (err_data.ue_count) in nbio_v7_4_handle_ras_controller_intr_no_bifring()
406 obj->err_data.ue_count, in nbio_v7_4_handle_ras_controller_intr_no_bifring()
419 amdgpu_ras_error_data_fini(&err_data); in nbio_v7_4_handle_ras_controller_intr_no_bifring()
[all …]
H A Dumc_v12_0.c141 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v12_0_query_error_count() local
161 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, ue_count); in umc_v12_0_query_error_count()
162 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, ce_count); in umc_v12_0_query_error_count()
163 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, de_count); in umc_v12_0_query_error_count()
236 struct ras_err_data *err_data, in umc_v12_0_convert_error_address() argument
297 if (!err_data && !dump_addr) in umc_v12_0_convert_error_address()
317 if (err_data) in umc_v12_0_convert_error_address()
318 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v12_0_convert_error_address()
330 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v12_0_query_error_address() local
346 if (!err_data->err_addr) { in umc_v12_0_query_error_address()
[all …]
H A Dnbio_v7_9.c522 struct ras_err_data err_data; in nbio_v7_9_handle_ras_controller_intr_no_bifring() local
525 if (amdgpu_ras_error_data_init(&err_data)) in nbio_v7_9_handle_ras_controller_intr_no_bifring()
544 nbio_v7_9_query_ras_error_count(adev, &err_data); in nbio_v7_9_handle_ras_controller_intr_no_bifring()
547 obj->err_data.ue_count += err_data.ue_count; in nbio_v7_9_handle_ras_controller_intr_no_bifring()
548 obj->err_data.ce_count += err_data.ce_count; in nbio_v7_9_handle_ras_controller_intr_no_bifring()
550 if (err_data.ce_count) in nbio_v7_9_handle_ras_controller_intr_no_bifring()
553 obj->err_data.ce_count, in nbio_v7_9_handle_ras_controller_intr_no_bifring()
556 if (err_data.ue_count) in nbio_v7_9_handle_ras_controller_intr_no_bifring()
559 obj->err_data.ue_count, in nbio_v7_9_handle_ras_controller_intr_no_bifring()
567 amdgpu_ras_error_data_fini(&err_data); in nbio_v7_9_handle_ras_controller_intr_no_bifring()
H A Dhdp_v4_0.c59 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in hdp_v4_0_query_ras_error_count() local
61 err_data->ue_count = 0; in hdp_v4_0_query_ras_error_count()
62 err_data->ce_count = 0; in hdp_v4_0_query_ras_error_count()
68 err_data->ue_count += RREG32_SOC15(HDP, 0, mmHDP_EDC_CNT); in hdp_v4_0_query_ras_error_count()
H A Dgfx_v9_4.c689 struct ras_err_data *err_data) in gfx_v9_4_query_utc_edc_status() argument
715 err_data->ce_count += sec_count; in gfx_v9_4_query_utc_edc_status()
723 err_data->ue_count += ded_count; in gfx_v9_4_query_utc_edc_status()
737 err_data->ce_count += sec_count; in gfx_v9_4_query_utc_edc_status()
746 err_data->ue_count += ded_count; in gfx_v9_4_query_utc_edc_status()
759 err_data->ce_count += sec_count; in gfx_v9_4_query_utc_edc_status()
767 err_data->ue_count += ded_count; in gfx_v9_4_query_utc_edc_status()
781 err_data->ce_count += sec_count; in gfx_v9_4_query_utc_edc_status()
790 err_data->ue_count += ded_count; in gfx_v9_4_query_utc_edc_status()
804 err_data->ce_count += sec_count; in gfx_v9_4_query_utc_edc_status()
[all …]
H A Damdgpu_mca.c76 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in amdgpu_mca_query_ras_error_count() local
78 amdgpu_mca_query_correctable_error_count(adev, mc_status_addr, &(err_data->ce_count)); in amdgpu_mca_query_ras_error_count()
79 amdgpu_mca_query_uncorrectable_error_count(adev, mc_status_addr, &(err_data->ue_count)); in amdgpu_mca_query_ras_error_count()
416 struct mca_bank_set *mca_set, struct ras_err_data *err_data) in amdgpu_mca_dispatch_mca_set() argument
447 amdgpu_ras_error_statistic_ue_count(err_data, in amdgpu_mca_dispatch_mca_set()
451 amdgpu_ras_error_statistic_de_count(err_data, in amdgpu_mca_dispatch_mca_set()
454 amdgpu_ras_error_statistic_ce_count(err_data, in amdgpu_mca_dispatch_mca_set()
477 struct ras_err_data *err_data, struct ras_query_context *qctx) in amdgpu_mca_smu_log_ras_error() argument
489 ret = amdgpu_mca_dispatch_mca_set(adev, blk, type, &mca_set, err_data); in amdgpu_mca_smu_log_ras_error()
503 ret = amdgpu_mca_dispatch_mca_set(adev, blk, type, &mca_cache->mca_set, err_data); in amdgpu_mca_smu_log_ras_error()
H A Damdgpu_aca.c505 …or_data(struct aca_bank_error *bank_error, enum aca_error_type type, struct ras_err_data *err_data) in aca_log_aca_error_data() argument
524 amdgpu_ras_error_statistic_ue_count(err_data, &mcm_info, count); in aca_log_aca_error_data()
527 amdgpu_ras_error_statistic_ce_count(err_data, &mcm_info, count); in aca_log_aca_error_data()
530 amdgpu_ras_error_statistic_de_count(err_data, &mcm_info, count); in aca_log_aca_error_data()
539 …a_log_aca_error(struct aca_handle *handle, enum aca_error_type type, struct ras_err_data *err_data) in aca_log_aca_error() argument
551 aca_log_aca_error_data(bank_error, type, err_data); in aca_log_aca_error()
562 struct ras_err_data *err_data, struct ras_query_context *qctx) in __aca_get_error_data() argument
586 aca_log_aca_error(handle, ACA_ERROR_TYPE_DEFERRED, err_data); in __aca_get_error_data()
588 return aca_log_aca_error(handle, type, err_data); in __aca_get_error_data()
600 enum aca_error_type type, struct ras_err_data *err_data, in amdgpu_aca_get_error_data() argument
[all …]
H A Dumc_v8_14.c96 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_14_query_error_count_per_channel() local
102 &(err_data->ce_count)); in umc_v8_14_query_error_count_per_channel()
105 &(err_data->ue_count)); in umc_v8_14_query_error_count_per_channel()
H A Damdgpu_ras.h654 #define for_each_ras_error(err_node, err_data) \ argument
655 list_for_each_entry(err_node, &(err_data)->err_node_list, node)
668 void *err_data,
704 struct ras_err_data err_data; member
996 int amdgpu_ras_error_data_init(struct ras_err_data *err_data);
997 void amdgpu_ras_error_data_fini(struct ras_err_data *err_data);
998 int amdgpu_ras_error_statistic_ce_count(struct ras_err_data *err_data,
1001 int amdgpu_ras_error_statistic_ue_count(struct ras_err_data *err_data,
1004 int amdgpu_ras_error_statistic_de_count(struct ras_err_data *err_data,
H A Damdgpu_umc.h109 struct ras_err_data *err_data,
164 int amdgpu_umc_fill_error_record(struct ras_err_data *err_data,
187 struct ras_err_data *err_data, uint64_t pa_addr);
H A Dsdma_v4_4.c199 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in sdma_v4_4_query_ras_error_count_by_instance() local
225 err_data->ue_count += sec_count; in sdma_v4_4_query_ras_error_count_by_instance()
231 err_data->ce_count = 0; in sdma_v4_4_query_ras_error_count_by_instance()
H A Damdgpu_xgmi.c1381 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in amdgpu_xgmi_legacy_query_ras_error_count() local
1389 err_data->ue_count = 0; in amdgpu_xgmi_legacy_query_ras_error_count()
1390 err_data->ce_count = 0; in amdgpu_xgmi_legacy_query_ras_error_count()
1471 err_data->ue_count += ue_cnt; in amdgpu_xgmi_legacy_query_ras_error_count()
1472 err_data->ce_count += ce_cnt; in amdgpu_xgmi_legacy_query_ras_error_count()
1500 u64 mca_base, struct ras_err_data *err_data) in __xgmi_v6_4_0_query_error_count() argument
1511 amdgpu_ras_error_statistic_ue_count(err_data, mcm_info, 1ULL); in __xgmi_v6_4_0_query_error_count()
1514 amdgpu_ras_error_statistic_ce_count(err_data, mcm_info, 1ULL); in __xgmi_v6_4_0_query_error_count()
1523 …_v6_4_0_query_error_count(struct amdgpu_device *adev, int xgmi_inst, struct ras_err_data *err_data) in xgmi_v6_4_0_query_error_count() argument
1532 __xgmi_v6_4_0_query_error_count(adev, &mcm_info, xgmi_v6_4_0_mca_base_array[i], err_data); in xgmi_v6_4_0_query_error_count()
[all …]
H A Dmmhub_v1_0.c797 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in mmhub_v1_0_query_ras_error_count() local
802 err_data->ue_count = 0; in mmhub_v1_0_query_ras_error_count()
803 err_data->ce_count = 0; in mmhub_v1_0_query_ras_error_count()
814 err_data->ce_count += sec_count; in mmhub_v1_0_query_ras_error_count()
815 err_data->ue_count += ded_count; in mmhub_v1_0_query_ras_error_count()
H A Dumc_v6_7.h75 struct ras_err_data *err_data, uint64_t err_addr,
/linux/drivers/gpu/drm/amd/ras/rascore/
H A Dras_cmd.c52 struct ras_ecc_count err_data; in ras_get_block_ecc_info() local
58 memset(&err_data, 0, sizeof(err_data)); in ras_get_block_ecc_info()
59 ret = ras_aca_get_block_ecc_count(ras_core, input_data->block_id, &err_data); in ras_get_block_ecc_info()
63 output_data->ce_count = err_data.total_ce_count; in ras_get_block_ecc_info()
64 output_data->ue_count = err_data.total_ue_count; in ras_get_block_ecc_info()
65 output_data->de_count = err_data.total_de_count; in ras_get_block_ecc_info()
H A Dras_aca.c533 struct ras_ecc_count *err_data = (struct ras_ecc_count *)data; in ras_aca_get_block_ecc_count() local
543 if (!err_data) in ras_aca_get_block_ecc_count()
578 err_data->new_ce_count = ecc.new_ce_count; in ras_aca_get_block_ecc_count()
579 err_data->total_ce_count = ecc.total_ce_count; in ras_aca_get_block_ecc_count()
580 err_data->new_ue_count = ecc.new_ue_count; in ras_aca_get_block_ecc_count()
581 err_data->total_ue_count = ecc.total_ue_count; in ras_aca_get_block_ecc_count()
582 err_data->new_de_count = ecc.new_de_count; in ras_aca_get_block_ecc_count()
583 err_data->total_de_count = ecc.total_de_count; in ras_aca_get_block_ecc_count()
/linux/drivers/misc/
H A Dsmpro-errmon.c251 unsigned char err_data[MAX_READ_BLOCK_LENGTH]; in smpro_error_data_read() local
271 memset(err_data, 0x00, MAX_READ_BLOCK_LENGTH); in smpro_error_data_read()
272 ret = regmap_noinc_read(errmon->regmap, err_info->data, err_data, err_length); in smpro_error_data_read()
285 return sysfs_emit(buf, "%*phN\n", MAX_READ_BLOCK_LENGTH, err_data); in smpro_error_data_read()
/linux/drivers/dpll/zl3073x/
H A Dfw.c258 goto err_data; in zl3073x_fw_component_load()
268 err_data: in zl3073x_fw_component_load()
/linux/drivers/mtd/nand/raw/
H A Dnuvoton-ma35d1-nand-controller.c290 u32 err_data[6]; in ma35_nfi_correct() local
317 err_data[i] = readl(nand->regs + MA35_NFI_REG_NANDECCED0 + i * 4); in ma35_nfi_correct()
320 temp_data[i * 4 + 0] = err_data[i] & 0xff; in ma35_nfi_correct()
321 temp_data[i * 4 + 1] = (err_data[i] >> 8) & 0xff; in ma35_nfi_correct()
322 temp_data[i * 4 + 2] = (err_data[i] >> 16) & 0xff; in ma35_nfi_correct()
323 temp_data[i * 4 + 3] = (err_data[i] >> 24) & 0xff; in ma35_nfi_correct()
/linux/drivers/clk/hisilicon/
H A Dclk.c78 goto err_data; in hisi_clk_init()
84 err_data: in hisi_clk_init()

12