| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_bios.c | 259 bus_cntl = RREG32(R600_BUS_CNTL); in ni_read_disabled_bios() 260 d1vga_control = RREG32(AVIVO_D1VGA_CONTROL); in ni_read_disabled_bios() 261 d2vga_control = RREG32(AVIVO_D2VGA_CONTROL); in ni_read_disabled_bios() 262 vga_render_control = RREG32(AVIVO_VGA_RENDER_CONTROL); in ni_read_disabled_bios() 263 rom_cntl = RREG32(R600_ROM_CNTL); in ni_read_disabled_bios() 305 viph_control = RREG32(RADEON_VIPH_CONTROL); in r700_read_disabled_bios() 306 bus_cntl = RREG32(R600_BUS_CNTL); in r700_read_disabled_bios() 307 d1vga_control = RREG32(AVIVO_D1VGA_CONTROL); in r700_read_disabled_bios() 308 d2vga_control = RREG32(AVIVO_D2VGA_CONTROL); in r700_read_disabled_bios() 309 vga_render_control = RREG32(AVIVO_VGA_RENDER_CONTROL); in r700_read_disabled_bios() [all …]
|
| H A D | radeon_legacy_encoders.c | 67 lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); in radeon_legacy_lvds_update() 95 disp_pwr_man = RREG32(RADEON_DISP_PWR_MAN); in radeon_legacy_lvds_update() 98 lvds_pll_cntl = RREG32(RADEON_LVDS_PLL_CNTL); in radeon_legacy_lvds_update() 103 lvds_pll_cntl = RREG32(RADEON_LVDS_PLL_CNTL); in radeon_legacy_lvds_update() 198 lvds_pll_cntl = RREG32(RADEON_LVDS_PLL_CNTL); in radeon_legacy_lvds_mode_set() 201 lvds_ss_gen_cntl = RREG32(RADEON_LVDS_SS_GEN_CNTL); in radeon_legacy_lvds_mode_set() 208 lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); in radeon_legacy_lvds_mode_set() 219 lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); in radeon_legacy_lvds_mode_set() 289 backlight_level = (RREG32(RADEON_LVDS_GEN_CNTL) >> in radeon_legacy_get_backlight_level() 361 backlight_level = (RREG32(RADEON_LVDS_GEN_CNTL) >> in radeon_legacy_backlight_get_brightness() [all …]
|
| H A D | radeon_i2c.c | 124 temp = RREG32(rec->mask_clk_reg); in pre_xfer() 130 temp = RREG32(rec->a_clk_reg) & ~rec->a_clk_mask; in pre_xfer() 133 temp = RREG32(rec->a_data_reg) & ~rec->a_data_mask; in pre_xfer() 137 temp = RREG32(rec->en_clk_reg) & ~rec->en_clk_mask; in pre_xfer() 140 temp = RREG32(rec->en_data_reg) & ~rec->en_data_mask; in pre_xfer() 144 temp = RREG32(rec->mask_clk_reg) | rec->mask_clk_mask; in pre_xfer() 146 temp = RREG32(rec->mask_clk_reg); in pre_xfer() 148 temp = RREG32(rec->mask_data_reg) | rec->mask_data_mask; in pre_xfer() 150 temp = RREG32(rec->mask_data_reg); in pre_xfer() 163 temp = RREG32(rec->mask_clk_reg) & ~rec->mask_clk_mask; in post_xfer() [all …]
|
| H A D | vce_v2_0.c | 44 tmp = RREG32(VCE_CLOCK_GATING_B); in vce_v2_0_set_sw_cg() 48 tmp = RREG32(VCE_UENC_CLOCK_GATING); in vce_v2_0_set_sw_cg() 52 tmp = RREG32(VCE_UENC_REG_CLOCK_GATING); in vce_v2_0_set_sw_cg() 58 tmp = RREG32(VCE_CLOCK_GATING_B); in vce_v2_0_set_sw_cg() 63 tmp = RREG32(VCE_UENC_CLOCK_GATING); in vce_v2_0_set_sw_cg() 68 tmp = RREG32(VCE_UENC_REG_CLOCK_GATING); in vce_v2_0_set_sw_cg() 78 tmp = RREG32(VCE_CLOCK_GATING_B); in vce_v2_0_set_dyn_cg() 88 orig = tmp = RREG32(VCE_UENC_CLOCK_GATING); in vce_v2_0_set_dyn_cg() 94 orig = tmp = RREG32(VCE_UENC_REG_CLOCK_GATING); in vce_v2_0_set_dyn_cg() 135 tmp = RREG32(VCE_CLOCK_GATING_A); in vce_v2_0_init_cg() [all …]
|
| H A D | vce_v1_0.c | 64 return RREG32(VCE_RB_RPTR); in vce_v1_0_get_rptr() 66 return RREG32(VCE_RB_RPTR2); in vce_v1_0_get_rptr() 81 return RREG32(VCE_RB_WPTR); in vce_v1_0_get_wptr() 83 return RREG32(VCE_RB_WPTR2); in vce_v1_0_get_wptr() 108 tmp = RREG32(VCE_CLOCK_GATING_A); in vce_v1_0_enable_mgcg() 112 tmp = RREG32(VCE_UENC_CLOCK_GATING); in vce_v1_0_enable_mgcg() 117 tmp = RREG32(VCE_UENC_REG_CLOCK_GATING); in vce_v1_0_enable_mgcg() 121 tmp = RREG32(VCE_CLOCK_GATING_A); in vce_v1_0_enable_mgcg() 125 tmp = RREG32(VCE_UENC_CLOCK_GATING); in vce_v1_0_enable_mgcg() 130 tmp = RREG32(VCE_UENC_REG_CLOCK_GATING); in vce_v1_0_enable_mgcg() [all …]
|
| H A D | rs600.c | 64 if (RREG32(AVIVO_D1CRTC_STATUS + crtc_offsets[crtc]) & AVIVO_D1CRTC_V_BLANK) in avivo_is_in_vblank() 74 pos1 = RREG32(AVIVO_D1CRTC_STATUS_POSITION + crtc_offsets[crtc]); in avivo_is_counter_moving() 75 pos2 = RREG32(AVIVO_D1CRTC_STATUS_POSITION + crtc_offsets[crtc]); in avivo_is_counter_moving() 98 if (!(RREG32(AVIVO_D1CRTC_CONTROL + crtc_offsets[crtc]) & AVIVO_CRTC_EN)) in avivo_wait_for_vblank() 123 u32 tmp = RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset); in rs600_page_flip() 144 if (RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset) & AVIVO_D1GRPH_SURFACE_UPDATE_PENDING) in rs600_page_flip() 160 return !!(RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset) & in rs600_page_flip_pending() 238 tmp = RREG32(voltage->gpio.reg); in rs600_pm_misc() 247 tmp = RREG32(voltage->gpio.reg); in rs600_pm_misc() 333 tmp = RREG32(AVIVO_D1CRTC_CONTROL + radeon_crtc->crtc_offset); in rs600_pm_prepare() [all …]
|
| H A D | r600.c | 127 r = RREG32(R600_RCU_DATA); in r600_rcu_rreg() 149 r = RREG32(R600_UVD_CTX_DATA); in r600_uvd_ctx_rreg() 183 *val = RREG32(reg); in r600_get_allowed_info_register() 352 u32 temp = (RREG32(CG_THERMAL_STATUS) & ASIC_T_MASK) >> in rv6xx_get_temp() 797 if (RREG32(GRBM_STATUS) & GUI_ACTIVE) in r600_gui_idle() 811 if (RREG32(DC_HPD1_INT_STATUS) & DC_HPDx_SENSE) in r600_hpd_sense() 815 if (RREG32(DC_HPD2_INT_STATUS) & DC_HPDx_SENSE) in r600_hpd_sense() 819 if (RREG32(DC_HPD3_INT_STATUS) & DC_HPDx_SENSE) in r600_hpd_sense() 823 if (RREG32(DC_HPD4_INT_STATUS) & DC_HPDx_SENSE) in r600_hpd_sense() 828 if (RREG32(DC_HPD5_INT_STATUS) & DC_HPDx_SENSE) in r600_hpd_sense() [all …]
|
| H A D | rv730_dpm.c | 200 RREG32(CG_SPLL_FUNC_CNTL); in rv730_read_clock_registers() 202 RREG32(CG_SPLL_FUNC_CNTL_2); in rv730_read_clock_registers() 204 RREG32(CG_SPLL_FUNC_CNTL_3); in rv730_read_clock_registers() 206 RREG32(CG_SPLL_SPREAD_SPECTRUM); in rv730_read_clock_registers() 208 RREG32(CG_SPLL_SPREAD_SPECTRUM_2); in rv730_read_clock_registers() 211 RREG32(TCI_MCLK_PWRMGT_CNTL); in rv730_read_clock_registers() 213 RREG32(TCI_DLL_CNTL); in rv730_read_clock_registers() 215 RREG32(CG_MPLL_FUNC_CNTL); in rv730_read_clock_registers() 217 RREG32(CG_MPLL_FUNC_CNTL_2); in rv730_read_clock_registers() 219 RREG32(CG_MPLL_FUNC_CNTL_3); in rv730_read_clock_registers() [all …]
|
| H A D | r100.c | 81 if (RREG32(RADEON_CRTC_STATUS) & RADEON_CRTC_VBLANK_CUR) in r100_is_in_vblank() 86 if (RREG32(RADEON_CRTC2_STATUS) & RADEON_CRTC2_VBLANK_CUR) in r100_is_in_vblank() 98 vline1 = (RREG32(RADEON_CRTC_VLINE_CRNT_VLINE) >> 16) & RADEON_CRTC_V_TOTAL; in r100_is_counter_moving() 99 vline2 = (RREG32(RADEON_CRTC_VLINE_CRNT_VLINE) >> 16) & RADEON_CRTC_V_TOTAL; in r100_is_counter_moving() 101 vline1 = (RREG32(RADEON_CRTC2_VLINE_CRNT_VLINE) >> 16) & RADEON_CRTC_V_TOTAL; in r100_is_counter_moving() 102 vline2 = (RREG32(RADEON_CRTC2_VLINE_CRNT_VLINE) >> 16) & RADEON_CRTC_V_TOTAL; in r100_is_counter_moving() 126 if (!(RREG32(RADEON_CRTC_GEN_CNTL) & RADEON_CRTC_EN)) in r100_wait_for_vblank() 129 if (!(RREG32(RADEON_CRTC2_GEN_CNTL) & RADEON_CRTC2_EN)) in r100_wait_for_vblank() 185 if (RREG32(RADEON_CRTC_OFFSET + radeon_crtc->crtc_offset) & RADEON_CRTC_OFFSET__GUI_TRIG_OFFSET) in r100_page_flip() 211 return !!(RREG32(RADEON_CRTC_OFFSET + radeon_crtc->crtc_offset) & in r100_page_flip_pending() [all …]
|
| H A D | rs400.c | 157 tmp = RREG32(RADEON_BUS_CNTL) & ~RS600_BUS_MASTER_DIS; in rs400_gart_enable() 161 tmp = RREG32(RADEON_BUS_CNTL) & ~RADEON_BUS_MASTER_DIS; in rs400_gart_enable() 248 tmp = RREG32(RADEON_MC_STATUS); in rs400_mc_wait_for_idle() 277 RREG32(RADEON_MC_STATUS)); in rs400_gpu_init() 291 base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; in rs400_mc_init() 305 r = RREG32(RS480_NB_MC_DATA); in rs400_mc_rreg() 328 tmp = RREG32(RADEON_HOST_PATH_CNTL); in rs400_debugfs_gart_info_show() 330 tmp = RREG32(RADEON_BUS_CNTL); in rs400_debugfs_gart_info_show() 343 tmp = RREG32(RS690_HDP_FB_LOCATION); in rs400_debugfs_gart_info_show() 346 tmp = RREG32(RADEON_AGP_BASE); in rs400_debugfs_gart_info_show() [all …]
|
| H A D | r420.c | 104 gb_pipe_select = RREG32(R400_GB_PIPE_SELECT); in r420_pipes_init() 140 tmp = RREG32(R300_DST_PIPE_CONFIG); in r420_pipes_init() 144 RREG32(R300_RB2D_DSTCACHE_MODE) | in r420_pipes_init() 153 tmp = RREG32(RV530_GB_PIPE_SELECT2); in r420_pipes_init() 172 r = RREG32(R_0001FC_MC_IND_DATA); in r420_mc_rreg() 286 rdev->config.r300.hdp_cntl = RREG32(RADEON_HOST_PATH_CNTL); in r420_startup() 318 RREG32(R_000E40_RBBM_STATUS), in r420_resume() 319 RREG32(R_0007C0_CP_STAT)); in r420_resume() 409 RREG32(R_000E40_RBBM_STATUS), in r420_init() 410 RREG32(R_0007C0_CP_STAT)); in r420_init() [all …]
|
| H A D | ni.c | 54 r = RREG32(TN_SMC_IND_DATA_0); in tn_smc_rreg() 652 mem_type = (RREG32(MC_SEQ_MISC0) & MC_SEQ_MISC0_GDDR5_MASK) >> MC_SEQ_MISC0_GDDR5_SHIFT; in ni_mc_load_microcode() 653 running = RREG32(MC_SEQ_SUP_CNTL) & RUN_MASK; in ni_mc_load_microcode() 677 if (RREG32(MC_IO_PAD_CNTL_D0) & MEM_FALL_OUT_CMD) in ni_mc_load_microcode() 847 *val = RREG32(reg); in cayman_get_allowed_info_register() 993 RREG32(MC_SHARED_CHMAP); in cayman_gpu_init() 994 mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG); in cayman_gpu_init() 1071 rb_disable_bitmap = (RREG32(CC_RB_BACKEND_DISABLE) & 0x00ff0000) >> 16; in cayman_gpu_init() 1091 simd_disable_bitmap = (RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffff0000) >> 16; in cayman_gpu_init() 1140 cgts_sm_ctrl_reg = RREG32(CGTS_SM_CTRL_REG); in cayman_gpu_init() [all …]
|
| H A D | rv740_dpm.c | 292 RREG32(CG_SPLL_FUNC_CNTL); in rv740_read_clock_registers() 294 RREG32(CG_SPLL_FUNC_CNTL_2); in rv740_read_clock_registers() 296 RREG32(CG_SPLL_FUNC_CNTL_3); in rv740_read_clock_registers() 298 RREG32(CG_SPLL_SPREAD_SPECTRUM); in rv740_read_clock_registers() 300 RREG32(CG_SPLL_SPREAD_SPECTRUM_2); in rv740_read_clock_registers() 303 RREG32(MPLL_AD_FUNC_CNTL); in rv740_read_clock_registers() 305 RREG32(MPLL_AD_FUNC_CNTL_2); in rv740_read_clock_registers() 307 RREG32(MPLL_DQ_FUNC_CNTL); in rv740_read_clock_registers() 309 RREG32(MPLL_DQ_FUNC_CNTL_2); in rv740_read_clock_registers() 311 RREG32(MCLK_PWRMGT_CNTL); in rv740_read_clock_registers() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | gmc_v8_0.c | 181 blackout = RREG32(mmMC_SHARED_BLACKOUT_CNTL); in gmc_v8_0_mc_stop() 199 tmp = RREG32(mmMC_SHARED_BLACKOUT_CNTL); in gmc_v8_0_mc_resume() 247 if (RREG32(mmMC_SEQ_IO_DEBUG_DATA) == 0x05b4dc40) in gmc_v8_0_init_microcode() 309 running = REG_GET_FIELD(RREG32(mmMC_SEQ_SUP_CNTL), MC_SEQ_SUP_CNTL, RUN); in gmc_v8_0_tonga_mc_load_microcode() 332 if (REG_GET_FIELD(RREG32(mmMC_SEQ_TRAIN_WAKEUP_CNTL), in gmc_v8_0_tonga_mc_load_microcode() 338 if (REG_GET_FIELD(RREG32(mmMC_SEQ_TRAIN_WAKEUP_CNTL), in gmc_v8_0_tonga_mc_load_microcode() 378 data = RREG32(mmMC_SEQ_MISC0); in gmc_v8_0_polaris_mc_load_microcode() 402 data = RREG32(mmMC_SEQ_MISC0); in gmc_v8_0_polaris_mc_load_microcode() 417 base = RREG32(mmMC_VM_FB_LOCATION) & 0xFFFF; in gmc_v8_0_vram_gtt_location() 458 tmp = RREG32(mmVGA_HDP_CONTROL); in gmc_v8_0_mc_program() [all …]
|
| H A D | gmc_v7_0.c | 99 blackout = RREG32(mmMC_SHARED_BLACKOUT_CNTL); in gmc_v7_0_mc_stop() 117 tmp = RREG32(mmMC_SHARED_BLACKOUT_CNTL); in gmc_v7_0_mc_resume() 199 running = REG_GET_FIELD(RREG32(mmMC_SEQ_SUP_CNTL), MC_SEQ_SUP_CNTL, RUN); in gmc_v7_0_mc_load_microcode() 222 if (REG_GET_FIELD(RREG32(mmMC_SEQ_TRAIN_WAKEUP_CNTL), in gmc_v7_0_mc_load_microcode() 228 if (REG_GET_FIELD(RREG32(mmMC_SEQ_TRAIN_WAKEUP_CNTL), in gmc_v7_0_mc_load_microcode() 241 u64 base = RREG32(mmMC_VM_FB_LOCATION) & 0xFFFF; in gmc_v7_0_vram_gtt_location() 283 tmp = RREG32(mmVGA_HDP_CONTROL); in gmc_v7_0_mc_program() 288 tmp = RREG32(mmVGA_RENDER_CONTROL); in gmc_v7_0_mc_program() 307 tmp = RREG32(mmHDP_MISC_CNTL); in gmc_v7_0_mc_program() 311 tmp = RREG32(mmHDP_HOST_PATH_CNTL); in gmc_v7_0_mc_program() [all …]
|
| H A D | amdgpu_i2c.c | 50 temp = RREG32(rec->mask_clk_reg); in amdgpu_i2c_pre_xfer() 56 temp = RREG32(rec->a_clk_reg) & ~rec->a_clk_mask; in amdgpu_i2c_pre_xfer() 59 temp = RREG32(rec->a_data_reg) & ~rec->a_data_mask; in amdgpu_i2c_pre_xfer() 63 temp = RREG32(rec->en_clk_reg) & ~rec->en_clk_mask; in amdgpu_i2c_pre_xfer() 66 temp = RREG32(rec->en_data_reg) & ~rec->en_data_mask; in amdgpu_i2c_pre_xfer() 70 temp = RREG32(rec->mask_clk_reg) | rec->mask_clk_mask; in amdgpu_i2c_pre_xfer() 72 temp = RREG32(rec->mask_clk_reg); in amdgpu_i2c_pre_xfer() 74 temp = RREG32(rec->mask_data_reg) | rec->mask_data_mask; in amdgpu_i2c_pre_xfer() 76 temp = RREG32(rec->mask_data_reg); in amdgpu_i2c_pre_xfer() 89 temp = RREG32(rec->mask_clk_reg) & ~rec->mask_clk_mask; in amdgpu_i2c_post_xfer() [all …]
|
| H A D | iceland_ih.c | 62 u32 ih_cntl = RREG32(mmIH_CNTL); in iceland_ih_enable_interrupts() 63 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in iceland_ih_enable_interrupts() 81 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in iceland_ih_disable_interrupts() 82 u32 ih_cntl = RREG32(mmIH_CNTL); in iceland_ih_disable_interrupts() 117 interrupt_cntl = RREG32(mmINTERRUPT_CNTL); in iceland_ih_irq_init() 148 ih_cntl = RREG32(mmIH_CNTL); in iceland_ih_irq_init() 207 wptr = RREG32(mmIH_RB_WPTR); in iceland_ih_get_wptr() 220 tmp = RREG32(mmIH_RB_CNTL); in iceland_ih_get_wptr() 351 u32 tmp = RREG32(mmSRBM_STATUS); in iceland_ih_is_idle() 367 tmp = RREG32(mmSRBM_STATUS); in iceland_ih_wait_for_idle() [all …]
|
| H A D | cz_ih.c | 62 u32 ih_cntl = RREG32(mmIH_CNTL); in cz_ih_enable_interrupts() 63 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in cz_ih_enable_interrupts() 81 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in cz_ih_disable_interrupts() 82 u32 ih_cntl = RREG32(mmIH_CNTL); in cz_ih_disable_interrupts() 117 interrupt_cntl = RREG32(mmINTERRUPT_CNTL); in cz_ih_irq_init() 148 ih_cntl = RREG32(mmIH_CNTL); in cz_ih_irq_init() 207 wptr = RREG32(mmIH_RB_WPTR); in cz_ih_get_wptr() 221 tmp = RREG32(mmIH_RB_CNTL); in cz_ih_get_wptr() 357 u32 tmp = RREG32(mmSRBM_STATUS); in cz_ih_is_idle() 373 tmp = RREG32(mmSRBM_STATUS); in cz_ih_wait_for_idle() [all …]
|
| H A D | cik_ih.c | 62 u32 ih_cntl = RREG32(mmIH_CNTL); in cik_ih_enable_interrupts() 63 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in cik_ih_enable_interrupts() 81 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in cik_ih_disable_interrupts() 82 u32 ih_cntl = RREG32(mmIH_CNTL); in cik_ih_disable_interrupts() 117 interrupt_cntl = RREG32(mmINTERRUPT_CNTL); in cik_ih_irq_init() 210 tmp = RREG32(mmIH_RB_CNTL); in cik_ih_get_wptr() 363 u32 tmp = RREG32(mmSRBM_STATUS); in cik_ih_is_idle() 379 tmp = RREG32(mmSRBM_STATUS) & SRBM_STATUS__IH_BUSY_MASK; in cik_ih_wait_for_idle() 392 u32 tmp = RREG32(mmSRBM_STATUS); in cik_ih_soft_reset() 398 tmp = RREG32(mmSRBM_SOFT_RESET); in cik_ih_soft_reset() [all …]
|
| H A D | si_ih.c | 38 u32 ih_cntl = RREG32(IH_CNTL); in si_ih_enable_interrupts() 39 u32 ih_rb_cntl = RREG32(IH_RB_CNTL); in si_ih_enable_interrupts() 50 u32 ih_rb_cntl = RREG32(IH_RB_CNTL); in si_ih_disable_interrupts() 51 u32 ih_cntl = RREG32(IH_CNTL); in si_ih_disable_interrupts() 72 interrupt_cntl = RREG32(INTERRUPT_CNTL); in si_ih_irq_init() 126 tmp = RREG32(IH_RB_CNTL); in si_ih_get_wptr() 229 u32 tmp = RREG32(mmSRBM_STATUS); in si_ih_is_idle() 255 u32 tmp = RREG32(mmSRBM_STATUS); in si_ih_soft_reset() 261 tmp = RREG32(mmSRBM_SOFT_RESET); in si_ih_soft_reset() 265 tmp = RREG32(mmSRBM_SOFT_RESET); in si_ih_soft_reset() [all …]
|
| H A D | amdgpu_amdkfd_gfx_v7.c | 107 while (!(RREG32(mmATC_VMID_PASID_MAPPING_UPDATE_STATUS) & (1U << vmid))) in kgd_set_pasid_vmid_mapping() 214 (*dump)[i++][1] = RREG32(addr); \ in kgd_hqd_dump() 255 data = RREG32(sdma_rlc_reg_offset + mmSDMA0_RLC0_CONTEXT_STATUS); in kgd_hqd_sdma_load() 329 act = RREG32(mmCP_HQD_ACTIVE); in kgd_hqd_is_occupied() 334 if (low == RREG32(mmCP_HQD_PQ_BASE) && in kgd_hqd_is_occupied() 335 high == RREG32(mmCP_HQD_PQ_BASE_HI)) in kgd_hqd_is_occupied() 351 sdma_rlc_rb_cntl = RREG32(sdma_rlc_reg_offset + mmSDMA0_RLC0_RB_CNTL); in kgd_hqd_sdma_is_occupied() 397 temp = RREG32(mmCP_HQD_IQ_TIMER); in kgd_hqd_destroy() 426 temp = RREG32(mmCP_HQD_DEQUEUE_REQUEST); in kgd_hqd_destroy() 445 temp = RREG32(mmCP_HQD_ACTIVE); in kgd_hqd_destroy() [all …]
|
| H A D | tonga_ih.c | 62 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in tonga_ih_enable_interrupts() 79 u32 ih_rb_cntl = RREG32(mmIH_RB_CNTL); in tonga_ih_disable_interrupts() 113 interrupt_cntl = RREG32(mmINTERRUPT_CNTL); in tonga_ih_irq_init() 145 ih_doorbell_rtpr = RREG32(mmIH_DOORBELL_RPTR); in tonga_ih_irq_init() 209 wptr = RREG32(mmIH_RB_WPTR); in tonga_ih_get_wptr() 224 tmp = RREG32(mmIH_RB_CNTL); in tonga_ih_get_wptr() 369 u32 tmp = RREG32(mmSRBM_STATUS); in tonga_ih_is_idle() 385 tmp = RREG32(mmSRBM_STATUS); in tonga_ih_wait_for_idle() 397 u32 tmp = RREG32(mmSRBM_STATUS); in tonga_ih_check_soft_reset() 442 tmp = RREG32(mmSRBM_SOFT_RESET); in tonga_ih_soft_reset() [all …]
|
| H A D | amdgpu_amdkfd_gfx_v8.c | 102 while (!(RREG32(mmATC_VMID_PASID_MAPPING_UPDATE_STATUS) & (1U << vmid))) in kgd_set_pasid_vmid_mapping() 177 value = RREG32(mmRLC_CP_SCHEDULERS); in kgd_hqd_load() 238 (*dump)[i++][1] = RREG32(addr); \ in kgd_hqd_dump() 278 data = RREG32(sdma_rlc_reg_offset + mmSDMA0_RLC0_CONTEXT_STATUS); in kgd_hqd_sdma_load() 361 act = RREG32(mmCP_HQD_ACTIVE); in kgd_hqd_is_occupied() 366 if (low == RREG32(mmCP_HQD_PQ_BASE) && in kgd_hqd_is_occupied() 367 high == RREG32(mmCP_HQD_PQ_BASE_HI)) in kgd_hqd_is_occupied() 383 sdma_rlc_rb_cntl = RREG32(sdma_rlc_reg_offset + mmSDMA0_RLC0_RB_CNTL); in kgd_hqd_sdma_is_occupied() 432 temp = RREG32(mmCP_HQD_IQ_TIMER); in kgd_hqd_destroy() 461 temp = RREG32(mmCP_HQD_DEQUEUE_REQUEST); in kgd_hqd_destroy() [all …]
|
| H A D | vpe_v6_1.c | 79 f32_cntl = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_F32_CNTL)); in vpe_v6_1_halt() 109 vpe_colla_cntl = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_COLLABORATE_CNTL)); in vpe_v6_1_set_collaborate_mode() 114 vpe_colla_cfg = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_COLLABORATE_CFG)); in vpe_v6_1_set_collaborate_mode() 136 ret = RREG32(vpe_get_reg_offset(vpe, j, regVPEC_CNTL_6_1_1)); in vpe_v6_1_load_microcode() 138 ret = RREG32(vpe_get_reg_offset(vpe, j, regVPEC_CNTL)); in vpe_v6_1_load_microcode() 162 f32_cntl = RREG32(f32_offset); in vpe_v6_1_load_microcode() 218 rb_cntl = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_QUEUE0_RB_CNTL)); in vpe_v6_1_ring_start() 250 doorbell_offset = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_QUEUE0_DOORBELL_OFFSET)); in vpe_v6_1_ring_start() 254 doorbell = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_QUEUE0_DOORBELL)); in vpe_v6_1_ring_start() 264 ib_cntl = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_QUEUE0_IB_CNTL)); in vpe_v6_1_ring_start() [all …]
|
| H A D | dce_v10_0.c | 180 r = RREG32(mmAZALIA_F0_CODEC_ENDPOINT_DATA + block_offset); in dce_v10_0_audio_endpt_rreg() 202 return RREG32(mmCRTC_STATUS_FRAME_COUNT + crtc_offsets[crtc]); in dce_v10_0_vblank_get_counter() 242 tmp = RREG32(mmGRPH_FLIP_CONTROL + amdgpu_crtc->crtc_offset); in dce_v10_0_page_flip() 256 RREG32(mmGRPH_PRIMARY_SURFACE_ADDRESS + amdgpu_crtc->crtc_offset); in dce_v10_0_page_flip() 265 *vbl = RREG32(mmCRTC_V_BLANK_START_END + crtc_offsets[crtc]); in dce_v10_0_crtc_get_scanoutpos() 266 *position = RREG32(mmCRTC_STATUS_POSITION + crtc_offsets[crtc]); in dce_v10_0_crtc_get_scanoutpos() 288 if (RREG32(mmDC_HPD_INT_STATUS + hpd_offsets[hpd]) & in dce_v10_0_hpd_sense() 312 tmp = RREG32(mmDC_HPD_INT_CONTROL + hpd_offsets[hpd]); in dce_v10_0_hpd_set_polarity() 349 tmp = RREG32(mmDC_HPD_INT_CONTROL + hpd_offsets[amdgpu_connector->hpd.hpd]); in dce_v10_0_hpd_init() 355 tmp = RREG32(mmDC_HPD_CONTROL + hpd_offsets[amdgpu_connector->hpd.hpd]); in dce_v10_0_hpd_init() [all …]
|