| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | umc_v8_7.c | 93 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in umc_v8_7_ecc_info_query_ras_error_count() local 104 &(err_data->ce_count)); in umc_v8_7_ecc_info_query_ras_error_count() 107 &(err_data->ue_count)); in umc_v8_7_ecc_info_query_ras_error_count() 112 struct ras_err_data *err_data, uint64_t err_addr, in umc_v8_7_convert_error_address() argument 126 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v8_7_convert_error_address() 131 struct ras_err_data *err_data, in umc_v8_7_ecc_info_query_error_address() argument 145 if (!err_data->err_addr) in umc_v8_7_ecc_info_query_error_address() 155 umc_v8_7_convert_error_address(adev, err_data, err_addr, in umc_v8_7_ecc_info_query_error_address() 163 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in umc_v8_7_ecc_info_query_ras_error_address() local 174 err_data, in umc_v8_7_ecc_info_query_ras_error_address() [all …]
|
| H A D | amdgpu_ras.c | 178 struct ras_err_data err_data; in amdgpu_reserve_page_direct() local 195 ret = amdgpu_ras_error_data_init(&err_data); in amdgpu_reserve_page_direct() 200 err_data.err_addr = &err_rec; in amdgpu_reserve_page_direct() 201 amdgpu_umc_fill_error_record(&err_data, address, address, 0, 0); in amdgpu_reserve_page_direct() 204 amdgpu_ras_add_bad_pages(adev, err_data.err_addr, in amdgpu_reserve_page_direct() 205 err_data.err_addr_cnt, false); in amdgpu_reserve_page_direct() 209 amdgpu_ras_error_data_fini(&err_data); in amdgpu_reserve_page_direct() 733 amdgpu_ras_error_data_fini(&obj->err_data); in put_obj() 765 if (amdgpu_ras_error_data_init(&obj->err_data)) in amdgpu_ras_create_obj() 1092 static void amdgpu_ras_get_ecc_info(struct amdgpu_device *adev, struct ras_err_data *err_data) in amdgpu_ras_get_ecc_info() argument [all …]
|
| H A D | umc_v8_10.c | 147 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_10_query_ecc_error_count() local 153 &(err_data->ce_count)); in umc_v8_10_query_ecc_error_count() 156 &(err_data->ue_count)); in umc_v8_10_query_ecc_error_count() 206 struct ras_err_data *err_data, uint64_t err_addr, in umc_v8_10_convert_error_address() argument 239 amdgpu_umc_fill_error_record(err_data, na_err_addr, in umc_v8_10_convert_error_address() 251 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_10_query_error_address() local 262 if (!err_data->err_addr) { in umc_v8_10_query_error_address() 277 umc_v8_10_convert_error_address(adev, err_data, err_addr, in umc_v8_10_query_error_address() 383 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_10_ecc_info_query_ecc_error_count() local 387 &(err_data->ce_count)); in umc_v8_10_ecc_info_query_ecc_error_count() [all …]
|
| H A D | umc_v6_7.c | 167 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v6_7_ecc_info_querry_ecc_error_count() local 171 &(err_data->ce_count)); in umc_v6_7_ecc_info_querry_ecc_error_count() 175 &(err_data->ue_count)); in umc_v6_7_ecc_info_querry_ecc_error_count() 188 struct ras_err_data *err_data, uint64_t err_addr, in umc_v6_7_convert_error_address() argument 211 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v6_7_convert_error_address() 217 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v6_7_convert_error_address() 229 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v6_7_ecc_info_query_error_address() local 237 if (!err_data->err_addr) in umc_v6_7_ecc_info_query_error_address() 247 umc_v6_7_convert_error_address(adev, err_data, err_addr, in umc_v6_7_ecc_info_query_error_address() 416 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v6_7_query_ecc_error_count() local [all …]
|
| H A D | umc_v6_1.c | 257 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in umc_v6_1_query_ras_error_count() local 279 &(err_data->ce_count)); in umc_v6_1_query_ras_error_count() 282 &(err_data->ue_count)); in umc_v6_1_query_ras_error_count() 296 struct ras_err_data *err_data, in umc_v6_1_query_error_address() argument 324 if (!err_data->err_addr) { in umc_v6_1_query_error_address() 345 amdgpu_umc_fill_error_record(err_data, err_addr, in umc_v6_1_query_error_address() 356 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in umc_v6_1_query_ras_error_address() local 377 err_data, in umc_v6_1_query_ras_error_address()
|
| H A D | hdp_v4_0.c | 59 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in hdp_v4_0_query_ras_error_count() local 61 err_data->ue_count = 0; in hdp_v4_0_query_ras_error_count() 62 err_data->ce_count = 0; in hdp_v4_0_query_ras_error_count() 68 err_data->ue_count += RREG32_SOC15(HDP, 0, mmHDP_EDC_CNT); in hdp_v4_0_query_ras_error_count()
|
| H A D | amdgpu_mca.c | 76 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in amdgpu_mca_query_ras_error_count() local 78 amdgpu_mca_query_correctable_error_count(adev, mc_status_addr, &(err_data->ce_count)); in amdgpu_mca_query_ras_error_count() 79 amdgpu_mca_query_uncorrectable_error_count(adev, mc_status_addr, &(err_data->ue_count)); in amdgpu_mca_query_ras_error_count() 416 struct mca_bank_set *mca_set, struct ras_err_data *err_data) in amdgpu_mca_dispatch_mca_set() argument 447 amdgpu_ras_error_statistic_ue_count(err_data, in amdgpu_mca_dispatch_mca_set() 451 amdgpu_ras_error_statistic_de_count(err_data, in amdgpu_mca_dispatch_mca_set() 454 amdgpu_ras_error_statistic_ce_count(err_data, in amdgpu_mca_dispatch_mca_set() 477 struct ras_err_data *err_data, struct ras_query_context *qctx) in amdgpu_mca_smu_log_ras_error() argument 489 ret = amdgpu_mca_dispatch_mca_set(adev, blk, type, &mca_set, err_data); in amdgpu_mca_smu_log_ras_error() 503 ret = amdgpu_mca_dispatch_mca_set(adev, blk, type, &mca_cache->mca_set, err_data); in amdgpu_mca_smu_log_ras_error()
|
| H A D | umc_v8_14.c | 96 struct ras_err_data *err_data = (struct ras_err_data *)data; in umc_v8_14_query_error_count_per_channel() local 102 &(err_data->ce_count)); in umc_v8_14_query_error_count_per_channel() 105 &(err_data->ue_count)); in umc_v8_14_query_error_count_per_channel()
|
| H A D | sdma_v4_4.c | 199 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in sdma_v4_4_query_ras_error_count_by_instance() local 225 err_data->ue_count += sec_count; in sdma_v4_4_query_ras_error_count_by_instance() 231 err_data->ce_count = 0; in sdma_v4_4_query_ras_error_count_by_instance()
|
| H A D | umc_v6_7.h | 75 struct ras_err_data *err_data, uint64_t err_addr,
|
| H A D | mmhub_v1_0.c | 797 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in mmhub_v1_0_query_ras_error_count() local 802 err_data->ue_count = 0; in mmhub_v1_0_query_ras_error_count() 803 err_data->ce_count = 0; in mmhub_v1_0_query_ras_error_count() 814 err_data->ce_count += sec_count; in mmhub_v1_0_query_ras_error_count() 815 err_data->ue_count += ded_count; in mmhub_v1_0_query_ras_error_count()
|
| H A D | amdgpu_sdma.h | 188 void *err_data,
|
| H A D | amdgpu_mca.h | 167 struct ras_err_data *err_data, struct ras_query_context *qctx);
|
| H A D | amdgpu_virt.c | 1638 struct ras_err_data *err_data) in amdgpu_virt_req_ras_err_count() argument 1654 err_data->ue_count = adev->virt.count_cache.block[sriov_block].ue_count; in amdgpu_virt_req_ras_err_count() 1655 err_data->ce_count = adev->virt.count_cache.block[sriov_block].ce_count; in amdgpu_virt_req_ras_err_count() 1656 err_data->de_count = adev->virt.count_cache.block[sriov_block].de_count; in amdgpu_virt_req_ras_err_count()
|
| H A D | amdgpu_sdma.c | 119 void *err_data, in amdgpu_sdma_process_ras_data_cb() argument
|
| H A D | vcn_v4_0_3.c | 1914 struct ras_err_data *err_data = (struct ras_err_data *)ras_err_status; in vcn_v4_0_3_inst_query_ras_error_count() local 1922 &err_data->ue_count); in vcn_v4_0_3_inst_query_ras_error_count()
|
| /linux/drivers/misc/ |
| H A D | smpro-errmon.c | 251 unsigned char err_data[MAX_READ_BLOCK_LENGTH]; in smpro_error_data_read() local 271 memset(err_data, 0x00, MAX_READ_BLOCK_LENGTH); in smpro_error_data_read() 272 ret = regmap_noinc_read(errmon->regmap, err_info->data, err_data, err_length); in smpro_error_data_read() 285 return sysfs_emit(buf, "%*phN\n", MAX_READ_BLOCK_LENGTH, err_data); in smpro_error_data_read()
|
| /linux/drivers/dpll/zl3073x/ |
| H A D | fw.c | 258 goto err_data; in zl3073x_fw_component_load() 268 err_data: in zl3073x_fw_component_load()
|
| /linux/drivers/mtd/nand/raw/ |
| H A D | nuvoton-ma35d1-nand-controller.c | 290 u32 err_data[6]; in ma35_nfi_correct() local 317 err_data[i] = readl(nand->regs + MA35_NFI_REG_NANDECCED0 + i * 4); in ma35_nfi_correct() 320 temp_data[i * 4 + 0] = err_data[i] & 0xff; in ma35_nfi_correct() 321 temp_data[i * 4 + 1] = (err_data[i] >> 8) & 0xff; in ma35_nfi_correct() 322 temp_data[i * 4 + 2] = (err_data[i] >> 16) & 0xff; in ma35_nfi_correct() 323 temp_data[i * 4 + 3] = (err_data[i] >> 24) & 0xff; in ma35_nfi_correct()
|
| /linux/drivers/clk/hisilicon/ |
| H A D | clk.c | 78 goto err_data; in hisi_clk_init() 84 err_data: in hisi_clk_init()
|
| /linux/drivers/net/wireless/intersil/p54/ |
| H A D | eeprom.c | 529 goto err_data; in p54_parse_rssical() 546 goto err_data; in p54_parse_rssical() 593 err_data: in p54_parse_rssical()
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| H A D | qed_dev.c | 2369 if (data->err_data.recovery_scope == ERR_SCOPE_FUNC && in qed_fw_err_handler() 2370 le16_to_cpu(data->err_data.entity_id) >= MAX_NUM_PFS) { in qed_fw_err_handler() 2371 qed_sriov_vfpf_malicious(p_hwfn, &data->err_data); in qed_fw_err_handler()
|
| /linux/drivers/usb/typec/tcpm/ |
| H A D | tcpm.c | 1674 goto err_data; in tcpm_queue_vdm_unlocked() 1694 err_data: in tcpm_queue_vdm_unlocked()
|