1 /*
2 * Copyright 2023 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 */
23
24 #define SWSMU_CODE_LAYER_L2
25
26 #include <linux/firmware.h>
27 #include <linux/pci.h>
28 #include <linux/i2c.h>
29 #include "amdgpu.h"
30 #include "amdgpu_smu.h"
31 #include "atomfirmware.h"
32 #include "amdgpu_atomfirmware.h"
33 #include "amdgpu_atombios.h"
34 #include "smu_v14_0.h"
35 #include "smu14_driver_if_v14_0.h"
36 #include "soc15_common.h"
37 #include "atom.h"
38 #include "smu_v14_0_2_ppt.h"
39 #include "smu_v14_0_2_pptable.h"
40 #include "smu_v14_0_2_ppsmc.h"
41 #include "mp/mp_14_0_2_offset.h"
42 #include "mp/mp_14_0_2_sh_mask.h"
43
44 #include "smu_cmn.h"
45 #include "amdgpu_ras.h"
46
47 /*
48 * DO NOT use these for err/warn/info/debug messages.
49 * Use dev_err, dev_warn, dev_info and dev_dbg instead.
50 * They are more MGPU friendly.
51 */
52 #undef pr_err
53 #undef pr_warn
54 #undef pr_info
55 #undef pr_debug
56
57 #define to_amdgpu_device(x) (container_of(x, struct amdgpu_device, pm.smu_i2c))
58
59 #define FEATURE_MASK(feature) (1ULL << feature)
60 #define SMC_DPM_FEATURE ( \
61 FEATURE_MASK(FEATURE_DPM_GFXCLK_BIT) | \
62 FEATURE_MASK(FEATURE_DPM_UCLK_BIT) | \
63 FEATURE_MASK(FEATURE_DPM_LINK_BIT) | \
64 FEATURE_MASK(FEATURE_DPM_SOCCLK_BIT) | \
65 FEATURE_MASK(FEATURE_DPM_FCLK_BIT))
66
67 #define MP0_MP1_DATA_REGION_SIZE_COMBOPPTABLE 0x4000
68 #define DEBUGSMC_MSG_Mode1Reset 2
69 #define LINK_SPEED_MAX 3
70
71 #define PP_OD_FEATURE_GFXCLK_FMIN 0
72 #define PP_OD_FEATURE_GFXCLK_FMAX 1
73 #define PP_OD_FEATURE_UCLK_FMIN 2
74 #define PP_OD_FEATURE_UCLK_FMAX 3
75 #define PP_OD_FEATURE_GFX_VF_CURVE 4
76 #define PP_OD_FEATURE_FAN_CURVE_TEMP 5
77 #define PP_OD_FEATURE_FAN_CURVE_PWM 6
78 #define PP_OD_FEATURE_FAN_ACOUSTIC_LIMIT 7
79 #define PP_OD_FEATURE_FAN_ACOUSTIC_TARGET 8
80 #define PP_OD_FEATURE_FAN_TARGET_TEMPERATURE 9
81 #define PP_OD_FEATURE_FAN_MINIMUM_PWM 10
82 #define PP_OD_FEATURE_FAN_ZERO_RPM_ENABLE 11
83
84 static struct cmn2asic_msg_mapping smu_v14_0_2_message_map[SMU_MSG_MAX_COUNT] = {
85 MSG_MAP(TestMessage, PPSMC_MSG_TestMessage, 1),
86 MSG_MAP(GetSmuVersion, PPSMC_MSG_GetSmuVersion, 1),
87 MSG_MAP(GetDriverIfVersion, PPSMC_MSG_GetDriverIfVersion, 1),
88 MSG_MAP(SetAllowedFeaturesMaskLow, PPSMC_MSG_SetAllowedFeaturesMaskLow, 0),
89 MSG_MAP(SetAllowedFeaturesMaskHigh, PPSMC_MSG_SetAllowedFeaturesMaskHigh, 0),
90 MSG_MAP(EnableAllSmuFeatures, PPSMC_MSG_EnableAllSmuFeatures, 0),
91 MSG_MAP(DisableAllSmuFeatures, PPSMC_MSG_DisableAllSmuFeatures, 0),
92 MSG_MAP(EnableSmuFeaturesLow, PPSMC_MSG_EnableSmuFeaturesLow, 1),
93 MSG_MAP(EnableSmuFeaturesHigh, PPSMC_MSG_EnableSmuFeaturesHigh, 1),
94 MSG_MAP(DisableSmuFeaturesLow, PPSMC_MSG_DisableSmuFeaturesLow, 1),
95 MSG_MAP(DisableSmuFeaturesHigh, PPSMC_MSG_DisableSmuFeaturesHigh, 1),
96 MSG_MAP(GetEnabledSmuFeaturesLow, PPSMC_MSG_GetRunningSmuFeaturesLow, 1),
97 MSG_MAP(GetEnabledSmuFeaturesHigh, PPSMC_MSG_GetRunningSmuFeaturesHigh, 1),
98 MSG_MAP(SetWorkloadMask, PPSMC_MSG_SetWorkloadMask, 1),
99 MSG_MAP(SetPptLimit, PPSMC_MSG_SetPptLimit, 0),
100 MSG_MAP(SetDriverDramAddrHigh, PPSMC_MSG_SetDriverDramAddrHigh, 1),
101 MSG_MAP(SetDriverDramAddrLow, PPSMC_MSG_SetDriverDramAddrLow, 1),
102 MSG_MAP(SetToolsDramAddrHigh, PPSMC_MSG_SetToolsDramAddrHigh, 0),
103 MSG_MAP(SetToolsDramAddrLow, PPSMC_MSG_SetToolsDramAddrLow, 0),
104 MSG_MAP(TransferTableSmu2Dram, PPSMC_MSG_TransferTableSmu2Dram, 1),
105 MSG_MAP(TransferTableDram2Smu, PPSMC_MSG_TransferTableDram2Smu, 0),
106 MSG_MAP(UseDefaultPPTable, PPSMC_MSG_UseDefaultPPTable, 0),
107 MSG_MAP(RunDcBtc, PPSMC_MSG_RunDcBtc, 0),
108 MSG_MAP(EnterBaco, PPSMC_MSG_EnterBaco, 0),
109 MSG_MAP(ExitBaco, PPSMC_MSG_ExitBaco, 0),
110 MSG_MAP(SetSoftMinByFreq, PPSMC_MSG_SetSoftMinByFreq, 1),
111 MSG_MAP(SetSoftMaxByFreq, PPSMC_MSG_SetSoftMaxByFreq, 1),
112 MSG_MAP(SetHardMinByFreq, PPSMC_MSG_SetHardMinByFreq, 1),
113 MSG_MAP(SetHardMaxByFreq, PPSMC_MSG_SetHardMaxByFreq, 0),
114 MSG_MAP(GetMinDpmFreq, PPSMC_MSG_GetMinDpmFreq, 1),
115 MSG_MAP(GetMaxDpmFreq, PPSMC_MSG_GetMaxDpmFreq, 1),
116 MSG_MAP(GetDpmFreqByIndex, PPSMC_MSG_GetDpmFreqByIndex, 1),
117 MSG_MAP(PowerUpVcn, PPSMC_MSG_PowerUpVcn, 0),
118 MSG_MAP(PowerDownVcn, PPSMC_MSG_PowerDownVcn, 0),
119 MSG_MAP(PowerUpJpeg, PPSMC_MSG_PowerUpJpeg, 0),
120 MSG_MAP(PowerDownJpeg, PPSMC_MSG_PowerDownJpeg, 0),
121 MSG_MAP(GetDcModeMaxDpmFreq, PPSMC_MSG_GetDcModeMaxDpmFreq, 1),
122 MSG_MAP(OverridePcieParameters, PPSMC_MSG_OverridePcieParameters, 0),
123 MSG_MAP(DramLogSetDramAddrHigh, PPSMC_MSG_DramLogSetDramAddrHigh, 0),
124 MSG_MAP(DramLogSetDramAddrLow, PPSMC_MSG_DramLogSetDramAddrLow, 0),
125 MSG_MAP(DramLogSetDramSize, PPSMC_MSG_DramLogSetDramSize, 0),
126 MSG_MAP(AllowGfxOff, PPSMC_MSG_AllowGfxOff, 0),
127 MSG_MAP(DisallowGfxOff, PPSMC_MSG_DisallowGfxOff, 0),
128 MSG_MAP(SetMGpuFanBoostLimitRpm, PPSMC_MSG_SetMGpuFanBoostLimitRpm, 0),
129 MSG_MAP(GetPptLimit, PPSMC_MSG_GetPptLimit, 0),
130 MSG_MAP(NotifyPowerSource, PPSMC_MSG_NotifyPowerSource, 0),
131 MSG_MAP(PrepareMp1ForUnload, PPSMC_MSG_PrepareMp1ForUnload, 0),
132 MSG_MAP(DFCstateControl, PPSMC_MSG_SetExternalClientDfCstateAllow, 0),
133 MSG_MAP(ArmD3, PPSMC_MSG_ArmD3, 0),
134 MSG_MAP(SetNumBadMemoryPagesRetired, PPSMC_MSG_SetNumBadMemoryPagesRetired, 0),
135 MSG_MAP(SetBadMemoryPagesRetiredFlagsPerChannel,
136 PPSMC_MSG_SetBadMemoryPagesRetiredFlagsPerChannel, 0),
137 MSG_MAP(AllowIHHostInterrupt, PPSMC_MSG_AllowIHHostInterrupt, 0),
138 MSG_MAP(ReenableAcDcInterrupt, PPSMC_MSG_ReenableAcDcInterrupt, 0),
139 };
140
141 static struct cmn2asic_mapping smu_v14_0_2_clk_map[SMU_CLK_COUNT] = {
142 CLK_MAP(GFXCLK, PPCLK_GFXCLK),
143 CLK_MAP(SCLK, PPCLK_GFXCLK),
144 CLK_MAP(SOCCLK, PPCLK_SOCCLK),
145 CLK_MAP(FCLK, PPCLK_FCLK),
146 CLK_MAP(UCLK, PPCLK_UCLK),
147 CLK_MAP(MCLK, PPCLK_UCLK),
148 CLK_MAP(VCLK, PPCLK_VCLK_0),
149 CLK_MAP(DCLK, PPCLK_DCLK_0),
150 CLK_MAP(DCEFCLK, PPCLK_DCFCLK),
151 };
152
153 static struct cmn2asic_mapping smu_v14_0_2_feature_mask_map[SMU_FEATURE_COUNT] = {
154 FEA_MAP(FW_DATA_READ),
155 FEA_MAP(DPM_GFXCLK),
156 FEA_MAP(DPM_GFX_POWER_OPTIMIZER),
157 FEA_MAP(DPM_UCLK),
158 FEA_MAP(DPM_FCLK),
159 FEA_MAP(DPM_SOCCLK),
160 FEA_MAP(DPM_LINK),
161 FEA_MAP(DPM_DCN),
162 FEA_MAP(VMEMP_SCALING),
163 FEA_MAP(VDDIO_MEM_SCALING),
164 FEA_MAP(DS_GFXCLK),
165 FEA_MAP(DS_SOCCLK),
166 FEA_MAP(DS_FCLK),
167 FEA_MAP(DS_LCLK),
168 FEA_MAP(DS_DCFCLK),
169 FEA_MAP(DS_UCLK),
170 FEA_MAP(GFX_ULV),
171 FEA_MAP(FW_DSTATE),
172 FEA_MAP(GFXOFF),
173 FEA_MAP(BACO),
174 FEA_MAP(MM_DPM),
175 FEA_MAP(SOC_MPCLK_DS),
176 FEA_MAP(BACO_MPCLK_DS),
177 FEA_MAP(THROTTLERS),
178 FEA_MAP(SMARTSHIFT),
179 FEA_MAP(GTHR),
180 FEA_MAP(ACDC),
181 FEA_MAP(VR0HOT),
182 FEA_MAP(FW_CTF),
183 FEA_MAP(FAN_CONTROL),
184 FEA_MAP(GFX_DCS),
185 FEA_MAP(GFX_READ_MARGIN),
186 FEA_MAP(LED_DISPLAY),
187 FEA_MAP(GFXCLK_SPREAD_SPECTRUM),
188 FEA_MAP(OUT_OF_BAND_MONITOR),
189 FEA_MAP(OPTIMIZED_VMIN),
190 FEA_MAP(GFX_IMU),
191 FEA_MAP(BOOT_TIME_CAL),
192 FEA_MAP(GFX_PCC_DFLL),
193 FEA_MAP(SOC_CG),
194 FEA_MAP(DF_CSTATE),
195 FEA_MAP(GFX_EDC),
196 FEA_MAP(BOOT_POWER_OPT),
197 FEA_MAP(CLOCK_POWER_DOWN_BYPASS),
198 FEA_MAP(DS_VCN),
199 FEA_MAP(BACO_CG),
200 FEA_MAP(MEM_TEMP_READ),
201 FEA_MAP(ATHUB_MMHUB_PG),
202 FEA_MAP(SOC_PCC),
203 FEA_MAP(EDC_PWRBRK),
204 FEA_MAP(SOC_EDC_XVMIN),
205 FEA_MAP(GFX_PSM_DIDT),
206 FEA_MAP(APT_ALL_ENABLE),
207 FEA_MAP(APT_SQ_THROTTLE),
208 FEA_MAP(APT_PF_DCS),
209 FEA_MAP(GFX_EDC_XVMIN),
210 FEA_MAP(GFX_DIDT_XVMIN),
211 FEA_MAP(FAN_ABNORMAL),
212 [SMU_FEATURE_DPM_VCLK_BIT] = {1, FEATURE_MM_DPM_BIT},
213 [SMU_FEATURE_DPM_DCLK_BIT] = {1, FEATURE_MM_DPM_BIT},
214 [SMU_FEATURE_PPT_BIT] = {1, FEATURE_THROTTLERS_BIT},
215 };
216
217 static struct cmn2asic_mapping smu_v14_0_2_table_map[SMU_TABLE_COUNT] = {
218 TAB_MAP(PPTABLE),
219 TAB_MAP(WATERMARKS),
220 TAB_MAP(AVFS_PSM_DEBUG),
221 TAB_MAP(PMSTATUSLOG),
222 TAB_MAP(SMU_METRICS),
223 TAB_MAP(DRIVER_SMU_CONFIG),
224 TAB_MAP(ACTIVITY_MONITOR_COEFF),
225 [SMU_TABLE_COMBO_PPTABLE] = {1, TABLE_COMBO_PPTABLE},
226 TAB_MAP(I2C_COMMANDS),
227 TAB_MAP(ECCINFO),
228 TAB_MAP(OVERDRIVE),
229 };
230
231 static struct cmn2asic_mapping smu_v14_0_2_pwr_src_map[SMU_POWER_SOURCE_COUNT] = {
232 PWR_MAP(AC),
233 PWR_MAP(DC),
234 };
235
236 static struct cmn2asic_mapping smu_v14_0_2_workload_map[PP_SMC_POWER_PROFILE_COUNT] = {
237 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT, WORKLOAD_PPLIB_DEFAULT_BIT),
238 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_FULLSCREEN3D, WORKLOAD_PPLIB_FULL_SCREEN_3D_BIT),
239 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_POWERSAVING, WORKLOAD_PPLIB_POWER_SAVING_BIT),
240 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_VIDEO, WORKLOAD_PPLIB_VIDEO_BIT),
241 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_VR, WORKLOAD_PPLIB_VR_BIT),
242 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_COMPUTE, WORKLOAD_PPLIB_COMPUTE_BIT),
243 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_CUSTOM, WORKLOAD_PPLIB_CUSTOM_BIT),
244 WORKLOAD_MAP(PP_SMC_POWER_PROFILE_WINDOW3D, WORKLOAD_PPLIB_WINDOW_3D_BIT),
245 };
246
247 static const uint8_t smu_v14_0_2_throttler_map[] = {
248 [THROTTLER_PPT0_BIT] = (SMU_THROTTLER_PPT0_BIT),
249 [THROTTLER_PPT1_BIT] = (SMU_THROTTLER_PPT1_BIT),
250 [THROTTLER_PPT2_BIT] = (SMU_THROTTLER_PPT2_BIT),
251 [THROTTLER_PPT3_BIT] = (SMU_THROTTLER_PPT3_BIT),
252 [THROTTLER_TDC_GFX_BIT] = (SMU_THROTTLER_TDC_GFX_BIT),
253 [THROTTLER_TDC_SOC_BIT] = (SMU_THROTTLER_TDC_SOC_BIT),
254 [THROTTLER_TEMP_EDGE_BIT] = (SMU_THROTTLER_TEMP_EDGE_BIT),
255 [THROTTLER_TEMP_HOTSPOT_BIT] = (SMU_THROTTLER_TEMP_HOTSPOT_BIT),
256 [THROTTLER_TEMP_MEM_BIT] = (SMU_THROTTLER_TEMP_MEM_BIT),
257 [THROTTLER_TEMP_VR_GFX_BIT] = (SMU_THROTTLER_TEMP_VR_GFX_BIT),
258 [THROTTLER_TEMP_VR_SOC_BIT] = (SMU_THROTTLER_TEMP_VR_SOC_BIT),
259 [THROTTLER_TEMP_VR_MEM0_BIT] = (SMU_THROTTLER_TEMP_VR_MEM0_BIT),
260 [THROTTLER_TEMP_VR_MEM1_BIT] = (SMU_THROTTLER_TEMP_VR_MEM1_BIT),
261 [THROTTLER_TEMP_LIQUID0_BIT] = (SMU_THROTTLER_TEMP_LIQUID0_BIT),
262 [THROTTLER_TEMP_LIQUID1_BIT] = (SMU_THROTTLER_TEMP_LIQUID1_BIT),
263 [THROTTLER_GFX_APCC_PLUS_BIT] = (SMU_THROTTLER_APCC_BIT),
264 [THROTTLER_FIT_BIT] = (SMU_THROTTLER_FIT_BIT),
265 };
266
267 static int
smu_v14_0_2_get_allowed_feature_mask(struct smu_context * smu,uint32_t * feature_mask,uint32_t num)268 smu_v14_0_2_get_allowed_feature_mask(struct smu_context *smu,
269 uint32_t *feature_mask, uint32_t num)
270 {
271 struct amdgpu_device *adev = smu->adev;
272 /*u32 smu_version;*/
273
274 if (num > 2)
275 return -EINVAL;
276
277 memset(feature_mask, 0xff, sizeof(uint32_t) * num);
278
279 if (adev->pm.pp_feature & PP_SCLK_DPM_MASK) {
280 *(uint64_t *)feature_mask |= FEATURE_MASK(FEATURE_DPM_GFXCLK_BIT);
281 *(uint64_t *)feature_mask |= FEATURE_MASK(FEATURE_GFX_IMU_BIT);
282 }
283 #if 0
284 if (!(adev->pg_flags & AMD_PG_SUPPORT_ATHUB) ||
285 !(adev->pg_flags & AMD_PG_SUPPORT_MMHUB))
286 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_ATHUB_MMHUB_PG_BIT);
287
288 if (!(adev->pm.pp_feature & PP_SOCCLK_DPM_MASK))
289 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_DPM_SOCCLK_BIT);
290
291 /* PMFW 78.58 contains a critical fix for gfxoff feature */
292 smu_cmn_get_smc_version(smu, NULL, &smu_version);
293 if ((smu_version < 0x004e3a00) ||
294 !(adev->pm.pp_feature & PP_GFXOFF_MASK))
295 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_GFXOFF_BIT);
296
297 if (!(adev->pm.pp_feature & PP_MCLK_DPM_MASK)) {
298 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_DPM_UCLK_BIT);
299 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_VMEMP_SCALING_BIT);
300 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_VDDIO_MEM_SCALING_BIT);
301 }
302
303 if (!(adev->pm.pp_feature & PP_SCLK_DEEP_SLEEP_MASK))
304 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_DS_GFXCLK_BIT);
305
306 if (!(adev->pm.pp_feature & PP_PCIE_DPM_MASK)) {
307 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_DPM_LINK_BIT);
308 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_DS_LCLK_BIT);
309 }
310
311 if (!(adev->pm.pp_feature & PP_ULV_MASK))
312 *(uint64_t *)feature_mask &= ~FEATURE_MASK(FEATURE_GFX_ULV_BIT);
313 #endif
314
315 return 0;
316 }
317
smu_v14_0_2_check_powerplay_table(struct smu_context * smu)318 static int smu_v14_0_2_check_powerplay_table(struct smu_context *smu)
319 {
320 struct smu_table_context *table_context = &smu->smu_table;
321 struct smu_14_0_2_powerplay_table *powerplay_table =
322 table_context->power_play_table;
323 struct smu_baco_context *smu_baco = &smu->smu_baco;
324 PPTable_t *pptable = smu->smu_table.driver_pptable;
325 const OverDriveLimits_t * const overdrive_upperlimits =
326 &pptable->SkuTable.OverDriveLimitsBasicMax;
327 const OverDriveLimits_t * const overdrive_lowerlimits =
328 &pptable->SkuTable.OverDriveLimitsBasicMin;
329
330 if (powerplay_table->platform_caps & SMU_14_0_2_PP_PLATFORM_CAP_HARDWAREDC)
331 smu->dc_controlled_by_gpio = true;
332
333 if (powerplay_table->platform_caps & SMU_14_0_2_PP_PLATFORM_CAP_BACO) {
334 smu_baco->platform_support = true;
335
336 if (powerplay_table->platform_caps & SMU_14_0_2_PP_PLATFORM_CAP_MACO)
337 smu_baco->maco_support = true;
338 }
339
340 if (!overdrive_lowerlimits->FeatureCtrlMask ||
341 !overdrive_upperlimits->FeatureCtrlMask)
342 smu->od_enabled = false;
343
344 table_context->thermal_controller_type =
345 powerplay_table->thermal_controller_type;
346
347 /*
348 * Instead of having its own buffer space and get overdrive_table copied,
349 * smu->od_settings just points to the actual overdrive_table
350 */
351 smu->od_settings = &powerplay_table->overdrive_table;
352
353 smu->adev->pm.no_fan =
354 !(pptable->PFE_Settings.FeaturesToRun[0] & (1 << FEATURE_FAN_CONTROL_BIT));
355
356 return 0;
357 }
358
smu_v14_0_2_store_powerplay_table(struct smu_context * smu)359 static int smu_v14_0_2_store_powerplay_table(struct smu_context *smu)
360 {
361 struct smu_table_context *table_context = &smu->smu_table;
362 struct smu_14_0_2_powerplay_table *powerplay_table =
363 table_context->power_play_table;
364
365 memcpy(table_context->driver_pptable, &powerplay_table->smc_pptable,
366 sizeof(PPTable_t));
367
368 return 0;
369 }
370
smu_v14_0_2_get_pptable_from_pmfw(struct smu_context * smu,void ** table,uint32_t * size)371 static int smu_v14_0_2_get_pptable_from_pmfw(struct smu_context *smu,
372 void **table,
373 uint32_t *size)
374 {
375 struct smu_table_context *smu_table = &smu->smu_table;
376 void *combo_pptable = smu_table->combo_pptable;
377 int ret = 0;
378
379 ret = smu_cmn_get_combo_pptable(smu);
380 if (ret)
381 return ret;
382
383 *table = combo_pptable;
384 *size = sizeof(struct smu_14_0_2_powerplay_table);
385
386 return 0;
387 }
388
smu_v14_0_2_setup_pptable(struct smu_context * smu)389 static int smu_v14_0_2_setup_pptable(struct smu_context *smu)
390 {
391 struct smu_table_context *smu_table = &smu->smu_table;
392 int ret = 0;
393
394 if (amdgpu_sriov_vf(smu->adev))
395 return 0;
396
397 ret = smu_v14_0_2_get_pptable_from_pmfw(smu,
398 &smu_table->power_play_table,
399 &smu_table->power_play_table_size);
400 if (ret)
401 return ret;
402
403 ret = smu_v14_0_2_store_powerplay_table(smu);
404 if (ret)
405 return ret;
406
407 ret = smu_v14_0_2_check_powerplay_table(smu);
408 if (ret)
409 return ret;
410
411 return ret;
412 }
413
smu_v14_0_2_tables_init(struct smu_context * smu)414 static int smu_v14_0_2_tables_init(struct smu_context *smu)
415 {
416 struct smu_table_context *smu_table = &smu->smu_table;
417 struct smu_table *tables = smu_table->tables;
418
419 SMU_TABLE_INIT(tables, SMU_TABLE_PPTABLE, sizeof(PPTable_t),
420 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
421 SMU_TABLE_INIT(tables, SMU_TABLE_WATERMARKS, sizeof(Watermarks_t),
422 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
423 SMU_TABLE_INIT(tables, SMU_TABLE_SMU_METRICS, sizeof(SmuMetricsExternal_t),
424 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
425 SMU_TABLE_INIT(tables, SMU_TABLE_I2C_COMMANDS, sizeof(SwI2cRequest_t),
426 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
427 SMU_TABLE_INIT(tables, SMU_TABLE_OVERDRIVE, sizeof(OverDriveTable_t),
428 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
429 SMU_TABLE_INIT(tables, SMU_TABLE_PMSTATUSLOG, SMU14_TOOL_SIZE,
430 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
431 SMU_TABLE_INIT(tables, SMU_TABLE_ACTIVITY_MONITOR_COEFF,
432 sizeof(DpmActivityMonitorCoeffIntExternal_t), PAGE_SIZE,
433 AMDGPU_GEM_DOMAIN_VRAM);
434 SMU_TABLE_INIT(tables, SMU_TABLE_COMBO_PPTABLE, MP0_MP1_DATA_REGION_SIZE_COMBOPPTABLE,
435 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
436 SMU_TABLE_INIT(tables, SMU_TABLE_ECCINFO, sizeof(EccInfoTable_t),
437 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
438
439 smu_table->metrics_table = kzalloc(sizeof(SmuMetricsExternal_t), GFP_KERNEL);
440 if (!smu_table->metrics_table)
441 goto err0_out;
442 smu_table->metrics_time = 0;
443
444 smu_table->gpu_metrics_table_size = sizeof(struct gpu_metrics_v1_3);
445 smu_table->gpu_metrics_table = kzalloc(smu_table->gpu_metrics_table_size, GFP_KERNEL);
446 if (!smu_table->gpu_metrics_table)
447 goto err1_out;
448
449 smu_table->watermarks_table = kzalloc(sizeof(Watermarks_t), GFP_KERNEL);
450 if (!smu_table->watermarks_table)
451 goto err2_out;
452
453 smu_table->ecc_table = kzalloc(tables[SMU_TABLE_ECCINFO].size, GFP_KERNEL);
454 if (!smu_table->ecc_table)
455 goto err3_out;
456
457 return 0;
458
459 err3_out:
460 kfree(smu_table->watermarks_table);
461 err2_out:
462 kfree(smu_table->gpu_metrics_table);
463 err1_out:
464 kfree(smu_table->metrics_table);
465 err0_out:
466 return -ENOMEM;
467 }
468
smu_v14_0_2_allocate_dpm_context(struct smu_context * smu)469 static int smu_v14_0_2_allocate_dpm_context(struct smu_context *smu)
470 {
471 struct smu_dpm_context *smu_dpm = &smu->smu_dpm;
472
473 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_14_0_dpm_context),
474 GFP_KERNEL);
475 if (!smu_dpm->dpm_context)
476 return -ENOMEM;
477
478 smu_dpm->dpm_context_size = sizeof(struct smu_14_0_dpm_context);
479
480 return 0;
481 }
482
smu_v14_0_2_init_smc_tables(struct smu_context * smu)483 static int smu_v14_0_2_init_smc_tables(struct smu_context *smu)
484 {
485 int ret = 0;
486
487 ret = smu_v14_0_2_tables_init(smu);
488 if (ret)
489 return ret;
490
491 ret = smu_v14_0_2_allocate_dpm_context(smu);
492 if (ret)
493 return ret;
494
495 return smu_v14_0_init_smc_tables(smu);
496 }
497
smu_v14_0_2_set_default_dpm_table(struct smu_context * smu)498 static int smu_v14_0_2_set_default_dpm_table(struct smu_context *smu)
499 {
500 struct smu_14_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context;
501 struct smu_table_context *table_context = &smu->smu_table;
502 PPTable_t *pptable = table_context->driver_pptable;
503 SkuTable_t *skutable = &pptable->SkuTable;
504 struct smu_14_0_dpm_table *dpm_table;
505 int ret = 0;
506
507 /* socclk dpm table setup */
508 dpm_table = &dpm_context->dpm_tables.soc_table;
509 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_SOCCLK_BIT)) {
510 ret = smu_v14_0_set_single_dpm_table(smu,
511 SMU_SOCCLK,
512 dpm_table);
513 if (ret)
514 return ret;
515 } else {
516 dpm_table->count = 1;
517 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.socclk / 100;
518 dpm_table->dpm_levels[0].enabled = true;
519 dpm_table->min = dpm_table->dpm_levels[0].value;
520 dpm_table->max = dpm_table->dpm_levels[0].value;
521 }
522
523 /* gfxclk dpm table setup */
524 dpm_table = &dpm_context->dpm_tables.gfx_table;
525 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_GFXCLK_BIT)) {
526 ret = smu_v14_0_set_single_dpm_table(smu,
527 SMU_GFXCLK,
528 dpm_table);
529 if (ret)
530 return ret;
531
532 /*
533 * Update the reported maximum shader clock to the value
534 * which can be guarded to be achieved on all cards. This
535 * is aligned with Window setting. And considering that value
536 * might be not the peak frequency the card can achieve, it
537 * is normal some real-time clock frequency can overtake this
538 * labelled maximum clock frequency(for example in pp_dpm_sclk
539 * sysfs output).
540 */
541 if (skutable->DriverReportedClocks.GameClockAc &&
542 (dpm_table->dpm_levels[dpm_table->count - 1].value >
543 skutable->DriverReportedClocks.GameClockAc)) {
544 dpm_table->dpm_levels[dpm_table->count - 1].value =
545 skutable->DriverReportedClocks.GameClockAc;
546 dpm_table->max = skutable->DriverReportedClocks.GameClockAc;
547 }
548 } else {
549 dpm_table->count = 1;
550 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.gfxclk / 100;
551 dpm_table->dpm_levels[0].enabled = true;
552 dpm_table->min = dpm_table->dpm_levels[0].value;
553 dpm_table->max = dpm_table->dpm_levels[0].value;
554 }
555
556 /* uclk dpm table setup */
557 dpm_table = &dpm_context->dpm_tables.uclk_table;
558 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_UCLK_BIT)) {
559 ret = smu_v14_0_set_single_dpm_table(smu,
560 SMU_UCLK,
561 dpm_table);
562 if (ret)
563 return ret;
564 } else {
565 dpm_table->count = 1;
566 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.uclk / 100;
567 dpm_table->dpm_levels[0].enabled = true;
568 dpm_table->min = dpm_table->dpm_levels[0].value;
569 dpm_table->max = dpm_table->dpm_levels[0].value;
570 }
571
572 /* fclk dpm table setup */
573 dpm_table = &dpm_context->dpm_tables.fclk_table;
574 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_FCLK_BIT)) {
575 ret = smu_v14_0_set_single_dpm_table(smu,
576 SMU_FCLK,
577 dpm_table);
578 if (ret)
579 return ret;
580 } else {
581 dpm_table->count = 1;
582 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.fclk / 100;
583 dpm_table->dpm_levels[0].enabled = true;
584 dpm_table->min = dpm_table->dpm_levels[0].value;
585 dpm_table->max = dpm_table->dpm_levels[0].value;
586 }
587
588 /* vclk dpm table setup */
589 dpm_table = &dpm_context->dpm_tables.vclk_table;
590 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_VCLK_BIT)) {
591 ret = smu_v14_0_set_single_dpm_table(smu,
592 SMU_VCLK,
593 dpm_table);
594 if (ret)
595 return ret;
596 } else {
597 dpm_table->count = 1;
598 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.vclk / 100;
599 dpm_table->dpm_levels[0].enabled = true;
600 dpm_table->min = dpm_table->dpm_levels[0].value;
601 dpm_table->max = dpm_table->dpm_levels[0].value;
602 }
603
604 /* dclk dpm table setup */
605 dpm_table = &dpm_context->dpm_tables.dclk_table;
606 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_DCLK_BIT)) {
607 ret = smu_v14_0_set_single_dpm_table(smu,
608 SMU_DCLK,
609 dpm_table);
610 if (ret)
611 return ret;
612 } else {
613 dpm_table->count = 1;
614 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.dclk / 100;
615 dpm_table->dpm_levels[0].enabled = true;
616 dpm_table->min = dpm_table->dpm_levels[0].value;
617 dpm_table->max = dpm_table->dpm_levels[0].value;
618 }
619
620 /* dcefclk dpm table setup */
621 dpm_table = &dpm_context->dpm_tables.dcef_table;
622 if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_DPM_DCN_BIT)) {
623 ret = smu_v14_0_set_single_dpm_table(smu,
624 SMU_DCEFCLK,
625 dpm_table);
626 if (ret)
627 return ret;
628 } else {
629 dpm_table->count = 1;
630 dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.dcefclk / 100;
631 dpm_table->dpm_levels[0].enabled = true;
632 dpm_table->min = dpm_table->dpm_levels[0].value;
633 dpm_table->max = dpm_table->dpm_levels[0].value;
634 }
635
636 return 0;
637 }
638
smu_v14_0_2_is_dpm_running(struct smu_context * smu)639 static bool smu_v14_0_2_is_dpm_running(struct smu_context *smu)
640 {
641 int ret = 0;
642 uint64_t feature_enabled;
643
644 ret = smu_cmn_get_enabled_mask(smu, &feature_enabled);
645 if (ret)
646 return false;
647
648 return !!(feature_enabled & SMC_DPM_FEATURE);
649 }
650
smu_v14_0_2_get_throttler_status(SmuMetrics_t * metrics)651 static uint32_t smu_v14_0_2_get_throttler_status(SmuMetrics_t *metrics)
652 {
653 uint32_t throttler_status = 0;
654 int i;
655
656 for (i = 0; i < THROTTLER_COUNT; i++)
657 throttler_status |=
658 (metrics->ThrottlingPercentage[i] ? 1U << i : 0);
659
660 return throttler_status;
661 }
662
663 #define SMU_14_0_2_BUSY_THRESHOLD 5
smu_v14_0_2_get_smu_metrics_data(struct smu_context * smu,MetricsMember_t member,uint32_t * value)664 static int smu_v14_0_2_get_smu_metrics_data(struct smu_context *smu,
665 MetricsMember_t member,
666 uint32_t *value)
667 {
668 struct smu_table_context *smu_table = &smu->smu_table;
669 SmuMetrics_t *metrics =
670 &(((SmuMetricsExternal_t *)(smu_table->metrics_table))->SmuMetrics);
671 int ret = 0;
672
673 ret = smu_cmn_get_metrics_table(smu,
674 NULL,
675 false);
676 if (ret)
677 return ret;
678
679 switch (member) {
680 case METRICS_CURR_GFXCLK:
681 *value = metrics->CurrClock[PPCLK_GFXCLK];
682 break;
683 case METRICS_CURR_SOCCLK:
684 *value = metrics->CurrClock[PPCLK_SOCCLK];
685 break;
686 case METRICS_CURR_UCLK:
687 *value = metrics->CurrClock[PPCLK_UCLK];
688 break;
689 case METRICS_CURR_VCLK:
690 *value = metrics->CurrClock[PPCLK_VCLK_0];
691 break;
692 case METRICS_CURR_DCLK:
693 *value = metrics->CurrClock[PPCLK_DCLK_0];
694 break;
695 case METRICS_CURR_FCLK:
696 *value = metrics->CurrClock[PPCLK_FCLK];
697 break;
698 case METRICS_CURR_DCEFCLK:
699 *value = metrics->CurrClock[PPCLK_DCFCLK];
700 break;
701 case METRICS_AVERAGE_GFXCLK:
702 if (metrics->AverageGfxActivity <= SMU_14_0_2_BUSY_THRESHOLD)
703 *value = metrics->AverageGfxclkFrequencyPostDs;
704 else
705 *value = metrics->AverageGfxclkFrequencyPreDs;
706 break;
707 case METRICS_AVERAGE_FCLK:
708 if (metrics->AverageUclkActivity <= SMU_14_0_2_BUSY_THRESHOLD)
709 *value = metrics->AverageFclkFrequencyPostDs;
710 else
711 *value = metrics->AverageFclkFrequencyPreDs;
712 break;
713 case METRICS_AVERAGE_UCLK:
714 if (metrics->AverageUclkActivity <= SMU_14_0_2_BUSY_THRESHOLD)
715 *value = metrics->AverageMemclkFrequencyPostDs;
716 else
717 *value = metrics->AverageMemclkFrequencyPreDs;
718 break;
719 case METRICS_AVERAGE_VCLK:
720 *value = metrics->AverageVclk0Frequency;
721 break;
722 case METRICS_AVERAGE_DCLK:
723 *value = metrics->AverageDclk0Frequency;
724 break;
725 case METRICS_AVERAGE_VCLK1:
726 *value = metrics->AverageVclk1Frequency;
727 break;
728 case METRICS_AVERAGE_DCLK1:
729 *value = metrics->AverageDclk1Frequency;
730 break;
731 case METRICS_AVERAGE_GFXACTIVITY:
732 *value = metrics->AverageGfxActivity;
733 break;
734 case METRICS_AVERAGE_MEMACTIVITY:
735 *value = metrics->AverageUclkActivity;
736 break;
737 case METRICS_AVERAGE_VCNACTIVITY:
738 *value = max(metrics->AverageVcn0ActivityPercentage,
739 metrics->Vcn1ActivityPercentage);
740 break;
741 case METRICS_AVERAGE_SOCKETPOWER:
742 *value = metrics->AverageSocketPower << 8;
743 break;
744 case METRICS_TEMPERATURE_EDGE:
745 *value = metrics->AvgTemperature[TEMP_EDGE] *
746 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
747 break;
748 case METRICS_TEMPERATURE_HOTSPOT:
749 *value = metrics->AvgTemperature[TEMP_HOTSPOT] *
750 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
751 break;
752 case METRICS_TEMPERATURE_MEM:
753 *value = metrics->AvgTemperature[TEMP_MEM] *
754 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
755 break;
756 case METRICS_TEMPERATURE_VRGFX:
757 *value = metrics->AvgTemperature[TEMP_VR_GFX] *
758 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
759 break;
760 case METRICS_TEMPERATURE_VRSOC:
761 *value = metrics->AvgTemperature[TEMP_VR_SOC] *
762 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
763 break;
764 case METRICS_THROTTLER_STATUS:
765 *value = smu_v14_0_2_get_throttler_status(metrics);
766 break;
767 case METRICS_CURR_FANSPEED:
768 *value = metrics->AvgFanRpm;
769 break;
770 case METRICS_CURR_FANPWM:
771 *value = metrics->AvgFanPwm;
772 break;
773 case METRICS_VOLTAGE_VDDGFX:
774 *value = metrics->AvgVoltage[SVI_PLANE_VDD_GFX];
775 break;
776 case METRICS_PCIE_RATE:
777 *value = metrics->PcieRate;
778 break;
779 case METRICS_PCIE_WIDTH:
780 *value = metrics->PcieWidth;
781 break;
782 default:
783 *value = UINT_MAX;
784 break;
785 }
786
787 return ret;
788 }
789
smu_v14_0_2_get_dpm_ultimate_freq(struct smu_context * smu,enum smu_clk_type clk_type,uint32_t * min,uint32_t * max)790 static int smu_v14_0_2_get_dpm_ultimate_freq(struct smu_context *smu,
791 enum smu_clk_type clk_type,
792 uint32_t *min,
793 uint32_t *max)
794 {
795 struct smu_14_0_dpm_context *dpm_context =
796 smu->smu_dpm.dpm_context;
797 struct smu_14_0_dpm_table *dpm_table;
798
799 switch (clk_type) {
800 case SMU_MCLK:
801 case SMU_UCLK:
802 /* uclk dpm table */
803 dpm_table = &dpm_context->dpm_tables.uclk_table;
804 break;
805 case SMU_GFXCLK:
806 case SMU_SCLK:
807 /* gfxclk dpm table */
808 dpm_table = &dpm_context->dpm_tables.gfx_table;
809 break;
810 case SMU_SOCCLK:
811 /* socclk dpm table */
812 dpm_table = &dpm_context->dpm_tables.soc_table;
813 break;
814 case SMU_FCLK:
815 /* fclk dpm table */
816 dpm_table = &dpm_context->dpm_tables.fclk_table;
817 break;
818 case SMU_VCLK:
819 case SMU_VCLK1:
820 /* vclk dpm table */
821 dpm_table = &dpm_context->dpm_tables.vclk_table;
822 break;
823 case SMU_DCLK:
824 case SMU_DCLK1:
825 /* dclk dpm table */
826 dpm_table = &dpm_context->dpm_tables.dclk_table;
827 break;
828 default:
829 dev_err(smu->adev->dev, "Unsupported clock type!\n");
830 return -EINVAL;
831 }
832
833 if (min)
834 *min = dpm_table->min;
835 if (max)
836 *max = dpm_table->max;
837
838 return 0;
839 }
840
smu_v14_0_2_read_sensor(struct smu_context * smu,enum amd_pp_sensors sensor,void * data,uint32_t * size)841 static int smu_v14_0_2_read_sensor(struct smu_context *smu,
842 enum amd_pp_sensors sensor,
843 void *data,
844 uint32_t *size)
845 {
846 struct smu_table_context *table_context = &smu->smu_table;
847 PPTable_t *smc_pptable = table_context->driver_pptable;
848 int ret = 0;
849
850 switch (sensor) {
851 case AMDGPU_PP_SENSOR_MAX_FAN_RPM:
852 *(uint16_t *)data = smc_pptable->CustomSkuTable.FanMaximumRpm;
853 *size = 4;
854 break;
855 case AMDGPU_PP_SENSOR_MEM_LOAD:
856 ret = smu_v14_0_2_get_smu_metrics_data(smu,
857 METRICS_AVERAGE_MEMACTIVITY,
858 (uint32_t *)data);
859 *size = 4;
860 break;
861 case AMDGPU_PP_SENSOR_GPU_LOAD:
862 ret = smu_v14_0_2_get_smu_metrics_data(smu,
863 METRICS_AVERAGE_GFXACTIVITY,
864 (uint32_t *)data);
865 *size = 4;
866 break;
867 case AMDGPU_PP_SENSOR_VCN_LOAD:
868 ret = smu_v14_0_2_get_smu_metrics_data(smu,
869 METRICS_AVERAGE_VCNACTIVITY,
870 (uint32_t *)data);
871 *size = 4;
872 break;
873 case AMDGPU_PP_SENSOR_GPU_AVG_POWER:
874 ret = smu_v14_0_2_get_smu_metrics_data(smu,
875 METRICS_AVERAGE_SOCKETPOWER,
876 (uint32_t *)data);
877 *size = 4;
878 break;
879 case AMDGPU_PP_SENSOR_HOTSPOT_TEMP:
880 ret = smu_v14_0_2_get_smu_metrics_data(smu,
881 METRICS_TEMPERATURE_HOTSPOT,
882 (uint32_t *)data);
883 *size = 4;
884 break;
885 case AMDGPU_PP_SENSOR_EDGE_TEMP:
886 ret = smu_v14_0_2_get_smu_metrics_data(smu,
887 METRICS_TEMPERATURE_EDGE,
888 (uint32_t *)data);
889 *size = 4;
890 break;
891 case AMDGPU_PP_SENSOR_MEM_TEMP:
892 ret = smu_v14_0_2_get_smu_metrics_data(smu,
893 METRICS_TEMPERATURE_MEM,
894 (uint32_t *)data);
895 *size = 4;
896 break;
897 case AMDGPU_PP_SENSOR_GFX_MCLK:
898 ret = smu_v14_0_2_get_smu_metrics_data(smu,
899 METRICS_CURR_UCLK,
900 (uint32_t *)data);
901 *(uint32_t *)data *= 100;
902 *size = 4;
903 break;
904 case AMDGPU_PP_SENSOR_GFX_SCLK:
905 ret = smu_v14_0_2_get_smu_metrics_data(smu,
906 METRICS_AVERAGE_GFXCLK,
907 (uint32_t *)data);
908 *(uint32_t *)data *= 100;
909 *size = 4;
910 break;
911 case AMDGPU_PP_SENSOR_VDDGFX:
912 ret = smu_v14_0_2_get_smu_metrics_data(smu,
913 METRICS_VOLTAGE_VDDGFX,
914 (uint32_t *)data);
915 *size = 4;
916 break;
917 default:
918 ret = -EOPNOTSUPP;
919 break;
920 }
921
922 return ret;
923 }
924
smu_v14_0_2_get_current_clk_freq_by_table(struct smu_context * smu,enum smu_clk_type clk_type,uint32_t * value)925 static int smu_v14_0_2_get_current_clk_freq_by_table(struct smu_context *smu,
926 enum smu_clk_type clk_type,
927 uint32_t *value)
928 {
929 MetricsMember_t member_type;
930 int clk_id = 0;
931
932 clk_id = smu_cmn_to_asic_specific_index(smu,
933 CMN2ASIC_MAPPING_CLK,
934 clk_type);
935 if (clk_id < 0)
936 return -EINVAL;
937
938 switch (clk_id) {
939 case PPCLK_GFXCLK:
940 member_type = METRICS_AVERAGE_GFXCLK;
941 break;
942 case PPCLK_UCLK:
943 member_type = METRICS_CURR_UCLK;
944 break;
945 case PPCLK_FCLK:
946 member_type = METRICS_CURR_FCLK;
947 break;
948 case PPCLK_SOCCLK:
949 member_type = METRICS_CURR_SOCCLK;
950 break;
951 case PPCLK_VCLK_0:
952 member_type = METRICS_AVERAGE_VCLK;
953 break;
954 case PPCLK_DCLK_0:
955 member_type = METRICS_AVERAGE_DCLK;
956 break;
957 case PPCLK_DCFCLK:
958 member_type = METRICS_CURR_DCEFCLK;
959 break;
960 default:
961 return -EINVAL;
962 }
963
964 return smu_v14_0_2_get_smu_metrics_data(smu,
965 member_type,
966 value);
967 }
968
smu_v14_0_2_is_od_feature_supported(struct smu_context * smu,int od_feature_bit)969 static bool smu_v14_0_2_is_od_feature_supported(struct smu_context *smu,
970 int od_feature_bit)
971 {
972 PPTable_t *pptable = smu->smu_table.driver_pptable;
973 const OverDriveLimits_t * const overdrive_upperlimits =
974 &pptable->SkuTable.OverDriveLimitsBasicMax;
975
976 return overdrive_upperlimits->FeatureCtrlMask & (1U << od_feature_bit);
977 }
978
smu_v14_0_2_get_od_setting_limits(struct smu_context * smu,int od_feature_bit,int32_t * min,int32_t * max)979 static void smu_v14_0_2_get_od_setting_limits(struct smu_context *smu,
980 int od_feature_bit,
981 int32_t *min,
982 int32_t *max)
983 {
984 PPTable_t *pptable = smu->smu_table.driver_pptable;
985 const OverDriveLimits_t * const overdrive_upperlimits =
986 &pptable->SkuTable.OverDriveLimitsBasicMax;
987 const OverDriveLimits_t * const overdrive_lowerlimits =
988 &pptable->SkuTable.OverDriveLimitsBasicMin;
989 int32_t od_min_setting, od_max_setting;
990
991 switch (od_feature_bit) {
992 case PP_OD_FEATURE_GFXCLK_FMIN:
993 case PP_OD_FEATURE_GFXCLK_FMAX:
994 od_min_setting = overdrive_lowerlimits->GfxclkFoffset;
995 od_max_setting = overdrive_upperlimits->GfxclkFoffset;
996 break;
997 case PP_OD_FEATURE_UCLK_FMIN:
998 od_min_setting = overdrive_lowerlimits->UclkFmin;
999 od_max_setting = overdrive_upperlimits->UclkFmin;
1000 break;
1001 case PP_OD_FEATURE_UCLK_FMAX:
1002 od_min_setting = overdrive_lowerlimits->UclkFmax;
1003 od_max_setting = overdrive_upperlimits->UclkFmax;
1004 break;
1005 case PP_OD_FEATURE_GFX_VF_CURVE:
1006 od_min_setting = overdrive_lowerlimits->VoltageOffsetPerZoneBoundary[0];
1007 od_max_setting = overdrive_upperlimits->VoltageOffsetPerZoneBoundary[0];
1008 break;
1009 case PP_OD_FEATURE_FAN_CURVE_TEMP:
1010 od_min_setting = overdrive_lowerlimits->FanLinearTempPoints[0];
1011 od_max_setting = overdrive_upperlimits->FanLinearTempPoints[0];
1012 break;
1013 case PP_OD_FEATURE_FAN_CURVE_PWM:
1014 od_min_setting = overdrive_lowerlimits->FanLinearPwmPoints[0];
1015 od_max_setting = overdrive_upperlimits->FanLinearPwmPoints[0];
1016 break;
1017 case PP_OD_FEATURE_FAN_ACOUSTIC_LIMIT:
1018 od_min_setting = overdrive_lowerlimits->AcousticLimitRpmThreshold;
1019 od_max_setting = overdrive_upperlimits->AcousticLimitRpmThreshold;
1020 break;
1021 case PP_OD_FEATURE_FAN_ACOUSTIC_TARGET:
1022 od_min_setting = overdrive_lowerlimits->AcousticTargetRpmThreshold;
1023 od_max_setting = overdrive_upperlimits->AcousticTargetRpmThreshold;
1024 break;
1025 case PP_OD_FEATURE_FAN_TARGET_TEMPERATURE:
1026 od_min_setting = overdrive_lowerlimits->FanTargetTemperature;
1027 od_max_setting = overdrive_upperlimits->FanTargetTemperature;
1028 break;
1029 case PP_OD_FEATURE_FAN_MINIMUM_PWM:
1030 od_min_setting = overdrive_lowerlimits->FanMinimumPwm;
1031 od_max_setting = overdrive_upperlimits->FanMinimumPwm;
1032 break;
1033 case PP_OD_FEATURE_FAN_ZERO_RPM_ENABLE:
1034 od_min_setting = overdrive_lowerlimits->FanZeroRpmEnable;
1035 od_max_setting = overdrive_upperlimits->FanZeroRpmEnable;
1036 break;
1037 default:
1038 od_min_setting = od_max_setting = INT_MAX;
1039 break;
1040 }
1041
1042 if (min)
1043 *min = od_min_setting;
1044 if (max)
1045 *max = od_max_setting;
1046 }
1047
smu_v14_0_2_print_clk_levels(struct smu_context * smu,enum smu_clk_type clk_type,char * buf)1048 static int smu_v14_0_2_print_clk_levels(struct smu_context *smu,
1049 enum smu_clk_type clk_type,
1050 char *buf)
1051 {
1052 struct smu_dpm_context *smu_dpm = &smu->smu_dpm;
1053 struct smu_14_0_dpm_context *dpm_context = smu_dpm->dpm_context;
1054 OverDriveTableExternal_t *od_table =
1055 (OverDriveTableExternal_t *)smu->smu_table.overdrive_table;
1056 struct smu_14_0_dpm_table *single_dpm_table;
1057 struct smu_14_0_pcie_table *pcie_table;
1058 uint32_t gen_speed, lane_width;
1059 int i, curr_freq, size = 0;
1060 int32_t min_value, max_value;
1061 int ret = 0;
1062
1063 smu_cmn_get_sysfs_buf(&buf, &size);
1064
1065 if (amdgpu_ras_intr_triggered()) {
1066 size += sysfs_emit_at(buf, size, "unavailable\n");
1067 return size;
1068 }
1069
1070 switch (clk_type) {
1071 case SMU_SCLK:
1072 single_dpm_table = &(dpm_context->dpm_tables.gfx_table);
1073 break;
1074 case SMU_MCLK:
1075 single_dpm_table = &(dpm_context->dpm_tables.uclk_table);
1076 break;
1077 case SMU_SOCCLK:
1078 single_dpm_table = &(dpm_context->dpm_tables.soc_table);
1079 break;
1080 case SMU_FCLK:
1081 single_dpm_table = &(dpm_context->dpm_tables.fclk_table);
1082 break;
1083 case SMU_VCLK:
1084 case SMU_VCLK1:
1085 single_dpm_table = &(dpm_context->dpm_tables.vclk_table);
1086 break;
1087 case SMU_DCLK:
1088 case SMU_DCLK1:
1089 single_dpm_table = &(dpm_context->dpm_tables.dclk_table);
1090 break;
1091 case SMU_DCEFCLK:
1092 single_dpm_table = &(dpm_context->dpm_tables.dcef_table);
1093 break;
1094 default:
1095 break;
1096 }
1097
1098 switch (clk_type) {
1099 case SMU_SCLK:
1100 case SMU_MCLK:
1101 case SMU_SOCCLK:
1102 case SMU_FCLK:
1103 case SMU_VCLK:
1104 case SMU_VCLK1:
1105 case SMU_DCLK:
1106 case SMU_DCLK1:
1107 case SMU_DCEFCLK:
1108 ret = smu_v14_0_2_get_current_clk_freq_by_table(smu, clk_type, &curr_freq);
1109 if (ret) {
1110 dev_err(smu->adev->dev, "Failed to get current clock freq!");
1111 return ret;
1112 }
1113
1114 if (single_dpm_table->is_fine_grained) {
1115 /*
1116 * For fine grained dpms, there are only two dpm levels:
1117 * - level 0 -> min clock freq
1118 * - level 1 -> max clock freq
1119 * And the current clock frequency can be any value between them.
1120 * So, if the current clock frequency is not at level 0 or level 1,
1121 * we will fake it as three dpm levels:
1122 * - level 0 -> min clock freq
1123 * - level 1 -> current actual clock freq
1124 * - level 2 -> max clock freq
1125 */
1126 if ((single_dpm_table->dpm_levels[0].value != curr_freq) &&
1127 (single_dpm_table->dpm_levels[1].value != curr_freq)) {
1128 size += sysfs_emit_at(buf, size, "0: %uMhz\n",
1129 single_dpm_table->dpm_levels[0].value);
1130 size += sysfs_emit_at(buf, size, "1: %uMhz *\n",
1131 curr_freq);
1132 size += sysfs_emit_at(buf, size, "2: %uMhz\n",
1133 single_dpm_table->dpm_levels[1].value);
1134 } else {
1135 size += sysfs_emit_at(buf, size, "0: %uMhz %s\n",
1136 single_dpm_table->dpm_levels[0].value,
1137 single_dpm_table->dpm_levels[0].value == curr_freq ? "*" : "");
1138 size += sysfs_emit_at(buf, size, "1: %uMhz %s\n",
1139 single_dpm_table->dpm_levels[1].value,
1140 single_dpm_table->dpm_levels[1].value == curr_freq ? "*" : "");
1141 }
1142 } else {
1143 for (i = 0; i < single_dpm_table->count; i++)
1144 size += sysfs_emit_at(buf, size, "%d: %uMhz %s\n",
1145 i, single_dpm_table->dpm_levels[i].value,
1146 single_dpm_table->dpm_levels[i].value == curr_freq ? "*" : "");
1147 }
1148 break;
1149 case SMU_PCIE:
1150 ret = smu_v14_0_2_get_smu_metrics_data(smu,
1151 METRICS_PCIE_RATE,
1152 &gen_speed);
1153 if (ret)
1154 return ret;
1155
1156 ret = smu_v14_0_2_get_smu_metrics_data(smu,
1157 METRICS_PCIE_WIDTH,
1158 &lane_width);
1159 if (ret)
1160 return ret;
1161
1162 pcie_table = &(dpm_context->dpm_tables.pcie_table);
1163 for (i = 0; i < pcie_table->num_of_link_levels; i++)
1164 size += sysfs_emit_at(buf, size, "%d: %s %s %dMhz %s\n", i,
1165 (pcie_table->pcie_gen[i] == 0) ? "2.5GT/s," :
1166 (pcie_table->pcie_gen[i] == 1) ? "5.0GT/s," :
1167 (pcie_table->pcie_gen[i] == 2) ? "8.0GT/s," :
1168 (pcie_table->pcie_gen[i] == 3) ? "16.0GT/s," :
1169 (pcie_table->pcie_gen[i] == 4) ? "32.0GT/s," : "",
1170 (pcie_table->pcie_lane[i] == 1) ? "x1" :
1171 (pcie_table->pcie_lane[i] == 2) ? "x2" :
1172 (pcie_table->pcie_lane[i] == 3) ? "x4" :
1173 (pcie_table->pcie_lane[i] == 4) ? "x8" :
1174 (pcie_table->pcie_lane[i] == 5) ? "x12" :
1175 (pcie_table->pcie_lane[i] == 6) ? "x16" :
1176 (pcie_table->pcie_lane[i] == 7) ? "x32" : "",
1177 pcie_table->clk_freq[i],
1178 (gen_speed == DECODE_GEN_SPEED(pcie_table->pcie_gen[i])) &&
1179 (lane_width == DECODE_LANE_WIDTH(pcie_table->pcie_lane[i])) ?
1180 "*" : "");
1181 break;
1182
1183 case SMU_OD_SCLK:
1184 if (!smu_v14_0_2_is_od_feature_supported(smu,
1185 PP_OD_FEATURE_GFXCLK_BIT))
1186 break;
1187
1188 size += sysfs_emit_at(buf, size, "OD_SCLK_OFFSET:\n");
1189 size += sysfs_emit_at(buf, size, "%dMhz\n",
1190 od_table->OverDriveTable.GfxclkFoffset);
1191 break;
1192
1193 case SMU_OD_MCLK:
1194 if (!smu_v14_0_2_is_od_feature_supported(smu,
1195 PP_OD_FEATURE_UCLK_BIT))
1196 break;
1197
1198 size += sysfs_emit_at(buf, size, "OD_MCLK:\n");
1199 size += sysfs_emit_at(buf, size, "0: %uMhz\n1: %uMHz\n",
1200 od_table->OverDriveTable.UclkFmin,
1201 od_table->OverDriveTable.UclkFmax);
1202 break;
1203
1204 case SMU_OD_VDDGFX_OFFSET:
1205 if (!smu_v14_0_2_is_od_feature_supported(smu,
1206 PP_OD_FEATURE_GFX_VF_CURVE_BIT))
1207 break;
1208
1209 size += sysfs_emit_at(buf, size, "OD_VDDGFX_OFFSET:\n");
1210 size += sysfs_emit_at(buf, size, "%dmV\n",
1211 od_table->OverDriveTable.VoltageOffsetPerZoneBoundary[0]);
1212 break;
1213
1214 case SMU_OD_FAN_CURVE:
1215 if (!smu_v14_0_2_is_od_feature_supported(smu,
1216 PP_OD_FEATURE_FAN_CURVE_BIT))
1217 break;
1218
1219 size += sysfs_emit_at(buf, size, "OD_FAN_CURVE:\n");
1220 for (i = 0; i < NUM_OD_FAN_MAX_POINTS - 1; i++)
1221 size += sysfs_emit_at(buf, size, "%d: %dC %d%%\n",
1222 i,
1223 (int)od_table->OverDriveTable.FanLinearTempPoints[i],
1224 (int)od_table->OverDriveTable.FanLinearPwmPoints[i]);
1225
1226 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1227 smu_v14_0_2_get_od_setting_limits(smu,
1228 PP_OD_FEATURE_FAN_CURVE_TEMP,
1229 &min_value,
1230 &max_value);
1231 size += sysfs_emit_at(buf, size, "FAN_CURVE(hotspot temp): %uC %uC\n",
1232 min_value, max_value);
1233
1234 smu_v14_0_2_get_od_setting_limits(smu,
1235 PP_OD_FEATURE_FAN_CURVE_PWM,
1236 &min_value,
1237 &max_value);
1238 size += sysfs_emit_at(buf, size, "FAN_CURVE(fan speed): %u%% %u%%\n",
1239 min_value, max_value);
1240
1241 break;
1242
1243 case SMU_OD_ACOUSTIC_LIMIT:
1244 if (!smu_v14_0_2_is_od_feature_supported(smu,
1245 PP_OD_FEATURE_FAN_CURVE_BIT))
1246 break;
1247
1248 size += sysfs_emit_at(buf, size, "OD_ACOUSTIC_LIMIT:\n");
1249 size += sysfs_emit_at(buf, size, "%d\n",
1250 (int)od_table->OverDriveTable.AcousticLimitRpmThreshold);
1251
1252 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1253 smu_v14_0_2_get_od_setting_limits(smu,
1254 PP_OD_FEATURE_FAN_ACOUSTIC_LIMIT,
1255 &min_value,
1256 &max_value);
1257 size += sysfs_emit_at(buf, size, "ACOUSTIC_LIMIT: %u %u\n",
1258 min_value, max_value);
1259 break;
1260
1261 case SMU_OD_ACOUSTIC_TARGET:
1262 if (!smu_v14_0_2_is_od_feature_supported(smu,
1263 PP_OD_FEATURE_FAN_CURVE_BIT))
1264 break;
1265
1266 size += sysfs_emit_at(buf, size, "OD_ACOUSTIC_TARGET:\n");
1267 size += sysfs_emit_at(buf, size, "%d\n",
1268 (int)od_table->OverDriveTable.AcousticTargetRpmThreshold);
1269
1270 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1271 smu_v14_0_2_get_od_setting_limits(smu,
1272 PP_OD_FEATURE_FAN_ACOUSTIC_TARGET,
1273 &min_value,
1274 &max_value);
1275 size += sysfs_emit_at(buf, size, "ACOUSTIC_TARGET: %u %u\n",
1276 min_value, max_value);
1277 break;
1278
1279 case SMU_OD_FAN_TARGET_TEMPERATURE:
1280 if (!smu_v14_0_2_is_od_feature_supported(smu,
1281 PP_OD_FEATURE_FAN_CURVE_BIT))
1282 break;
1283
1284 size += sysfs_emit_at(buf, size, "FAN_TARGET_TEMPERATURE:\n");
1285 size += sysfs_emit_at(buf, size, "%d\n",
1286 (int)od_table->OverDriveTable.FanTargetTemperature);
1287
1288 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1289 smu_v14_0_2_get_od_setting_limits(smu,
1290 PP_OD_FEATURE_FAN_TARGET_TEMPERATURE,
1291 &min_value,
1292 &max_value);
1293 size += sysfs_emit_at(buf, size, "TARGET_TEMPERATURE: %u %u\n",
1294 min_value, max_value);
1295 break;
1296
1297 case SMU_OD_FAN_MINIMUM_PWM:
1298 if (!smu_v14_0_2_is_od_feature_supported(smu,
1299 PP_OD_FEATURE_FAN_CURVE_BIT))
1300 break;
1301
1302 size += sysfs_emit_at(buf, size, "FAN_MINIMUM_PWM:\n");
1303 size += sysfs_emit_at(buf, size, "%d\n",
1304 (int)od_table->OverDriveTable.FanMinimumPwm);
1305
1306 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1307 smu_v14_0_2_get_od_setting_limits(smu,
1308 PP_OD_FEATURE_FAN_MINIMUM_PWM,
1309 &min_value,
1310 &max_value);
1311 size += sysfs_emit_at(buf, size, "MINIMUM_PWM: %u %u\n",
1312 min_value, max_value);
1313 break;
1314
1315 case SMU_OD_FAN_ZERO_RPM_ENABLE:
1316 if (!smu_v14_0_2_is_od_feature_supported(smu,
1317 PP_OD_FEATURE_ZERO_FAN_BIT))
1318 break;
1319
1320 size += sysfs_emit_at(buf, size, "FAN_ZERO_RPM_ENABLE:\n");
1321 size += sysfs_emit_at(buf, size, "%d\n",
1322 (int)od_table->OverDriveTable.FanZeroRpmEnable);
1323
1324 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1325 smu_v14_0_2_get_od_setting_limits(smu,
1326 PP_OD_FEATURE_FAN_ZERO_RPM_ENABLE,
1327 &min_value,
1328 &max_value);
1329 size += sysfs_emit_at(buf, size, "ZERO_RPM_ENABLE: %u %u\n",
1330 min_value, max_value);
1331 break;
1332
1333 case SMU_OD_RANGE:
1334 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_GFXCLK_BIT) &&
1335 !smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_UCLK_BIT) &&
1336 !smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_GFX_VF_CURVE_BIT))
1337 break;
1338
1339 size += sysfs_emit_at(buf, size, "%s:\n", "OD_RANGE");
1340
1341 if (smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_GFXCLK_BIT)) {
1342 smu_v14_0_2_get_od_setting_limits(smu,
1343 PP_OD_FEATURE_GFXCLK_FMAX,
1344 &min_value,
1345 &max_value);
1346 size += sysfs_emit_at(buf, size, "SCLK_OFFSET: %7dMhz %10uMhz\n",
1347 min_value, max_value);
1348 }
1349
1350 if (smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_UCLK_BIT)) {
1351 smu_v14_0_2_get_od_setting_limits(smu,
1352 PP_OD_FEATURE_UCLK_FMIN,
1353 &min_value,
1354 NULL);
1355 smu_v14_0_2_get_od_setting_limits(smu,
1356 PP_OD_FEATURE_UCLK_FMAX,
1357 NULL,
1358 &max_value);
1359 size += sysfs_emit_at(buf, size, "MCLK: %7uMhz %10uMhz\n",
1360 min_value, max_value);
1361 }
1362
1363 if (smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_GFX_VF_CURVE_BIT)) {
1364 smu_v14_0_2_get_od_setting_limits(smu,
1365 PP_OD_FEATURE_GFX_VF_CURVE,
1366 &min_value,
1367 &max_value);
1368 size += sysfs_emit_at(buf, size, "VDDGFX_OFFSET: %7dmv %10dmv\n",
1369 min_value, max_value);
1370 }
1371 break;
1372
1373 default:
1374 break;
1375 }
1376
1377 return size;
1378 }
1379
smu_v14_0_2_force_clk_levels(struct smu_context * smu,enum smu_clk_type clk_type,uint32_t mask)1380 static int smu_v14_0_2_force_clk_levels(struct smu_context *smu,
1381 enum smu_clk_type clk_type,
1382 uint32_t mask)
1383 {
1384 struct smu_dpm_context *smu_dpm = &smu->smu_dpm;
1385 struct smu_14_0_dpm_context *dpm_context = smu_dpm->dpm_context;
1386 struct smu_14_0_dpm_table *single_dpm_table;
1387 uint32_t soft_min_level, soft_max_level;
1388 uint32_t min_freq, max_freq;
1389 int ret = 0;
1390
1391 soft_min_level = mask ? (ffs(mask) - 1) : 0;
1392 soft_max_level = mask ? (fls(mask) - 1) : 0;
1393
1394 switch (clk_type) {
1395 case SMU_GFXCLK:
1396 case SMU_SCLK:
1397 single_dpm_table = &(dpm_context->dpm_tables.gfx_table);
1398 break;
1399 case SMU_MCLK:
1400 case SMU_UCLK:
1401 single_dpm_table = &(dpm_context->dpm_tables.uclk_table);
1402 break;
1403 case SMU_SOCCLK:
1404 single_dpm_table = &(dpm_context->dpm_tables.soc_table);
1405 break;
1406 case SMU_FCLK:
1407 single_dpm_table = &(dpm_context->dpm_tables.fclk_table);
1408 break;
1409 case SMU_VCLK:
1410 case SMU_VCLK1:
1411 single_dpm_table = &(dpm_context->dpm_tables.vclk_table);
1412 break;
1413 case SMU_DCLK:
1414 case SMU_DCLK1:
1415 single_dpm_table = &(dpm_context->dpm_tables.dclk_table);
1416 break;
1417 default:
1418 break;
1419 }
1420
1421 switch (clk_type) {
1422 case SMU_GFXCLK:
1423 case SMU_SCLK:
1424 case SMU_MCLK:
1425 case SMU_UCLK:
1426 case SMU_SOCCLK:
1427 case SMU_FCLK:
1428 case SMU_VCLK:
1429 case SMU_VCLK1:
1430 case SMU_DCLK:
1431 case SMU_DCLK1:
1432 if (single_dpm_table->is_fine_grained) {
1433 /* There is only 2 levels for fine grained DPM */
1434 soft_max_level = (soft_max_level >= 1 ? 1 : 0);
1435 soft_min_level = (soft_min_level >= 1 ? 1 : 0);
1436 } else {
1437 if ((soft_max_level >= single_dpm_table->count) ||
1438 (soft_min_level >= single_dpm_table->count))
1439 return -EINVAL;
1440 }
1441
1442 min_freq = single_dpm_table->dpm_levels[soft_min_level].value;
1443 max_freq = single_dpm_table->dpm_levels[soft_max_level].value;
1444
1445 ret = smu_v14_0_set_soft_freq_limited_range(smu,
1446 clk_type,
1447 min_freq,
1448 max_freq,
1449 false);
1450 break;
1451 case SMU_DCEFCLK:
1452 case SMU_PCIE:
1453 default:
1454 break;
1455 }
1456
1457 return ret;
1458 }
1459
smu_v14_0_2_update_pcie_parameters(struct smu_context * smu,uint8_t pcie_gen_cap,uint8_t pcie_width_cap)1460 static int smu_v14_0_2_update_pcie_parameters(struct smu_context *smu,
1461 uint8_t pcie_gen_cap,
1462 uint8_t pcie_width_cap)
1463 {
1464 struct smu_14_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context;
1465 struct smu_14_0_pcie_table *pcie_table =
1466 &dpm_context->dpm_tables.pcie_table;
1467 int num_of_levels;
1468 uint32_t smu_pcie_arg;
1469 uint32_t link_level;
1470 struct smu_table_context *table_context = &smu->smu_table;
1471 PPTable_t *pptable = table_context->driver_pptable;
1472 SkuTable_t *skutable = &pptable->SkuTable;
1473 int ret = 0;
1474 int i;
1475
1476 pcie_table->num_of_link_levels = 0;
1477 for (link_level = 0; link_level < NUM_LINK_LEVELS; link_level++) {
1478 if (!skutable->PcieGenSpeed[link_level] &&
1479 !skutable->PcieLaneCount[link_level] &&
1480 !skutable->LclkFreq[link_level])
1481 continue;
1482
1483 pcie_table->pcie_gen[pcie_table->num_of_link_levels] =
1484 skutable->PcieGenSpeed[link_level];
1485 pcie_table->pcie_lane[pcie_table->num_of_link_levels] =
1486 skutable->PcieLaneCount[link_level];
1487 pcie_table->clk_freq[pcie_table->num_of_link_levels] =
1488 skutable->LclkFreq[link_level];
1489 pcie_table->num_of_link_levels++;
1490 }
1491 num_of_levels = pcie_table->num_of_link_levels;
1492 if (!num_of_levels)
1493 return 0;
1494
1495 if (!(smu->adev->pm.pp_feature & PP_PCIE_DPM_MASK)) {
1496 if (pcie_table->pcie_gen[num_of_levels - 1] < pcie_gen_cap)
1497 pcie_gen_cap = pcie_table->pcie_gen[num_of_levels - 1];
1498
1499 if (pcie_table->pcie_lane[num_of_levels - 1] < pcie_width_cap)
1500 pcie_width_cap = pcie_table->pcie_lane[num_of_levels - 1];
1501
1502 /* Force all levels to use the same settings */
1503 for (i = 0; i < num_of_levels; i++) {
1504 pcie_table->pcie_gen[i] = pcie_gen_cap;
1505 pcie_table->pcie_lane[i] = pcie_width_cap;
1506 smu_pcie_arg = i << 16;
1507 smu_pcie_arg |= pcie_table->pcie_gen[i] << 8;
1508 smu_pcie_arg |= pcie_table->pcie_lane[i];
1509
1510 ret = smu_cmn_send_smc_msg_with_param(smu,
1511 SMU_MSG_OverridePcieParameters,
1512 smu_pcie_arg,
1513 NULL);
1514 if (ret)
1515 break;
1516 }
1517 } else {
1518 for (i = 0; i < num_of_levels; i++) {
1519 if (pcie_table->pcie_gen[i] > pcie_gen_cap ||
1520 pcie_table->pcie_lane[i] > pcie_width_cap) {
1521 pcie_table->pcie_gen[i] = pcie_table->pcie_gen[i] > pcie_gen_cap ?
1522 pcie_gen_cap : pcie_table->pcie_gen[i];
1523 pcie_table->pcie_lane[i] = pcie_table->pcie_lane[i] > pcie_width_cap ?
1524 pcie_width_cap : pcie_table->pcie_lane[i];
1525 smu_pcie_arg = i << 16;
1526 smu_pcie_arg |= pcie_table->pcie_gen[i] << 8;
1527 smu_pcie_arg |= pcie_table->pcie_lane[i];
1528
1529 ret = smu_cmn_send_smc_msg_with_param(smu,
1530 SMU_MSG_OverridePcieParameters,
1531 smu_pcie_arg,
1532 NULL);
1533 if (ret)
1534 break;
1535 }
1536 }
1537 }
1538
1539 return ret;
1540 }
1541
1542 static const struct smu_temperature_range smu14_thermal_policy[] = {
1543 {-273150, 99000, 99000, -273150, 99000, 99000, -273150, 99000, 99000},
1544 { 120000, 120000, 120000, 120000, 120000, 120000, 120000, 120000, 120000},
1545 };
1546
smu_v14_0_2_get_thermal_temperature_range(struct smu_context * smu,struct smu_temperature_range * range)1547 static int smu_v14_0_2_get_thermal_temperature_range(struct smu_context *smu,
1548 struct smu_temperature_range *range)
1549 {
1550 struct smu_table_context *table_context = &smu->smu_table;
1551 struct smu_14_0_2_powerplay_table *powerplay_table =
1552 table_context->power_play_table;
1553 PPTable_t *pptable = smu->smu_table.driver_pptable;
1554
1555 if (amdgpu_sriov_vf(smu->adev))
1556 return 0;
1557
1558 if (!range)
1559 return -EINVAL;
1560
1561 memcpy(range, &smu14_thermal_policy[0], sizeof(struct smu_temperature_range));
1562
1563 range->max = pptable->CustomSkuTable.TemperatureLimit[TEMP_EDGE] *
1564 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1565 range->edge_emergency_max = (pptable->CustomSkuTable.TemperatureLimit[TEMP_EDGE] + CTF_OFFSET_EDGE) *
1566 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1567 range->hotspot_crit_max = pptable->CustomSkuTable.TemperatureLimit[TEMP_HOTSPOT] *
1568 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1569 range->hotspot_emergency_max = (pptable->CustomSkuTable.TemperatureLimit[TEMP_HOTSPOT] + CTF_OFFSET_HOTSPOT) *
1570 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1571 range->mem_crit_max = pptable->CustomSkuTable.TemperatureLimit[TEMP_MEM] *
1572 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1573 range->mem_emergency_max = (pptable->CustomSkuTable.TemperatureLimit[TEMP_MEM] + CTF_OFFSET_MEM)*
1574 SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1575 range->software_shutdown_temp = powerplay_table->software_shutdown_temp;
1576 range->software_shutdown_temp_offset = pptable->CustomSkuTable.FanAbnormalTempLimitOffset;
1577
1578 return 0;
1579 }
1580
smu_v14_0_2_populate_umd_state_clk(struct smu_context * smu)1581 static int smu_v14_0_2_populate_umd_state_clk(struct smu_context *smu)
1582 {
1583 struct smu_14_0_dpm_context *dpm_context =
1584 smu->smu_dpm.dpm_context;
1585 struct smu_14_0_dpm_table *gfx_table =
1586 &dpm_context->dpm_tables.gfx_table;
1587 struct smu_14_0_dpm_table *mem_table =
1588 &dpm_context->dpm_tables.uclk_table;
1589 struct smu_14_0_dpm_table *soc_table =
1590 &dpm_context->dpm_tables.soc_table;
1591 struct smu_14_0_dpm_table *vclk_table =
1592 &dpm_context->dpm_tables.vclk_table;
1593 struct smu_14_0_dpm_table *dclk_table =
1594 &dpm_context->dpm_tables.dclk_table;
1595 struct smu_14_0_dpm_table *fclk_table =
1596 &dpm_context->dpm_tables.fclk_table;
1597 struct smu_umd_pstate_table *pstate_table =
1598 &smu->pstate_table;
1599 struct smu_table_context *table_context = &smu->smu_table;
1600 PPTable_t *pptable = table_context->driver_pptable;
1601 DriverReportedClocks_t driver_clocks =
1602 pptable->SkuTable.DriverReportedClocks;
1603
1604 pstate_table->gfxclk_pstate.min = gfx_table->min;
1605 if (driver_clocks.GameClockAc &&
1606 (driver_clocks.GameClockAc < gfx_table->max))
1607 pstate_table->gfxclk_pstate.peak = driver_clocks.GameClockAc;
1608 else
1609 pstate_table->gfxclk_pstate.peak = gfx_table->max;
1610
1611 pstate_table->uclk_pstate.min = mem_table->min;
1612 pstate_table->uclk_pstate.peak = mem_table->max;
1613
1614 pstate_table->socclk_pstate.min = soc_table->min;
1615 pstate_table->socclk_pstate.peak = soc_table->max;
1616
1617 pstate_table->vclk_pstate.min = vclk_table->min;
1618 pstate_table->vclk_pstate.peak = vclk_table->max;
1619
1620 pstate_table->dclk_pstate.min = dclk_table->min;
1621 pstate_table->dclk_pstate.peak = dclk_table->max;
1622
1623 pstate_table->fclk_pstate.min = fclk_table->min;
1624 pstate_table->fclk_pstate.peak = fclk_table->max;
1625
1626 if (driver_clocks.BaseClockAc &&
1627 driver_clocks.BaseClockAc < gfx_table->max)
1628 pstate_table->gfxclk_pstate.standard = driver_clocks.BaseClockAc;
1629 else
1630 pstate_table->gfxclk_pstate.standard = gfx_table->max;
1631 pstate_table->uclk_pstate.standard = mem_table->max;
1632 pstate_table->socclk_pstate.standard = soc_table->min;
1633 pstate_table->vclk_pstate.standard = vclk_table->min;
1634 pstate_table->dclk_pstate.standard = dclk_table->min;
1635 pstate_table->fclk_pstate.standard = fclk_table->min;
1636
1637 return 0;
1638 }
1639
smu_v14_0_2_get_unique_id(struct smu_context * smu)1640 static void smu_v14_0_2_get_unique_id(struct smu_context *smu)
1641 {
1642 struct smu_table_context *smu_table = &smu->smu_table;
1643 SmuMetrics_t *metrics =
1644 &(((SmuMetricsExternal_t *)(smu_table->metrics_table))->SmuMetrics);
1645 struct amdgpu_device *adev = smu->adev;
1646 uint32_t upper32 = 0, lower32 = 0;
1647 int ret;
1648
1649 ret = smu_cmn_get_metrics_table(smu, NULL, false);
1650 if (ret)
1651 goto out;
1652
1653 upper32 = metrics->PublicSerialNumberUpper;
1654 lower32 = metrics->PublicSerialNumberLower;
1655
1656 out:
1657 adev->unique_id = ((uint64_t)upper32 << 32) | lower32;
1658 }
1659
smu_v14_0_2_get_fan_speed_pwm(struct smu_context * smu,uint32_t * speed)1660 static int smu_v14_0_2_get_fan_speed_pwm(struct smu_context *smu,
1661 uint32_t *speed)
1662 {
1663 int ret;
1664
1665 if (!speed)
1666 return -EINVAL;
1667
1668 ret = smu_v14_0_2_get_smu_metrics_data(smu,
1669 METRICS_CURR_FANPWM,
1670 speed);
1671 if (ret) {
1672 dev_err(smu->adev->dev, "Failed to get fan speed(PWM)!");
1673 return ret;
1674 }
1675
1676 /* Convert the PMFW output which is in percent to pwm(255) based */
1677 *speed = min(*speed * 255 / 100, (uint32_t)255);
1678
1679 return 0;
1680 }
1681
smu_v14_0_2_get_fan_speed_rpm(struct smu_context * smu,uint32_t * speed)1682 static int smu_v14_0_2_get_fan_speed_rpm(struct smu_context *smu,
1683 uint32_t *speed)
1684 {
1685 if (!speed)
1686 return -EINVAL;
1687
1688 return smu_v14_0_2_get_smu_metrics_data(smu,
1689 METRICS_CURR_FANSPEED,
1690 speed);
1691 }
1692
smu_v14_0_2_get_power_limit(struct smu_context * smu,uint32_t * current_power_limit,uint32_t * default_power_limit,uint32_t * max_power_limit,uint32_t * min_power_limit)1693 static int smu_v14_0_2_get_power_limit(struct smu_context *smu,
1694 uint32_t *current_power_limit,
1695 uint32_t *default_power_limit,
1696 uint32_t *max_power_limit,
1697 uint32_t *min_power_limit)
1698 {
1699 struct smu_table_context *table_context = &smu->smu_table;
1700 struct smu_14_0_2_powerplay_table *powerplay_table =
1701 table_context->power_play_table;
1702 PPTable_t *pptable = table_context->driver_pptable;
1703 CustomSkuTable_t *skutable = &pptable->CustomSkuTable;
1704 uint32_t power_limit, od_percent_upper = 0, od_percent_lower = 0;
1705 uint32_t msg_limit = pptable->SkuTable.MsgLimits.Power[PPT_THROTTLER_PPT0][POWER_SOURCE_AC];
1706
1707 if (smu_v14_0_get_current_power_limit(smu, &power_limit))
1708 power_limit = smu->adev->pm.ac_power ?
1709 skutable->SocketPowerLimitAc[PPT_THROTTLER_PPT0] :
1710 skutable->SocketPowerLimitDc[PPT_THROTTLER_PPT0];
1711
1712 if (current_power_limit)
1713 *current_power_limit = power_limit;
1714 if (default_power_limit)
1715 *default_power_limit = power_limit;
1716
1717 if (powerplay_table) {
1718 if (smu->od_enabled &&
1719 smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_PPT_BIT)) {
1720 od_percent_upper = pptable->SkuTable.OverDriveLimitsBasicMax.Ppt;
1721 od_percent_lower = pptable->SkuTable.OverDriveLimitsBasicMin.Ppt;
1722 } else if (smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_PPT_BIT)) {
1723 od_percent_upper = 0;
1724 od_percent_lower = pptable->SkuTable.OverDriveLimitsBasicMin.Ppt;
1725 }
1726 }
1727
1728 dev_dbg(smu->adev->dev, "od percent upper:%d, od percent lower:%d (default power: %d)\n",
1729 od_percent_upper, od_percent_lower, power_limit);
1730
1731 if (max_power_limit) {
1732 *max_power_limit = msg_limit * (100 + od_percent_upper);
1733 *max_power_limit /= 100;
1734 }
1735
1736 if (min_power_limit) {
1737 *min_power_limit = power_limit * (100 + od_percent_lower);
1738 *min_power_limit /= 100;
1739 }
1740
1741 return 0;
1742 }
1743
smu_v14_0_2_get_power_profile_mode(struct smu_context * smu,char * buf)1744 static int smu_v14_0_2_get_power_profile_mode(struct smu_context *smu,
1745 char *buf)
1746 {
1747 DpmActivityMonitorCoeffIntExternal_t activity_monitor_external;
1748 DpmActivityMonitorCoeffInt_t *activity_monitor =
1749 &(activity_monitor_external.DpmActivityMonitorCoeffInt);
1750 static const char *title[] = {
1751 "PROFILE_INDEX(NAME)",
1752 "CLOCK_TYPE(NAME)",
1753 "FPS",
1754 "MinActiveFreqType",
1755 "MinActiveFreq",
1756 "BoosterFreqType",
1757 "BoosterFreq",
1758 "PD_Data_limit_c",
1759 "PD_Data_error_coeff",
1760 "PD_Data_error_rate_coeff"};
1761 int16_t workload_type = 0;
1762 uint32_t i, size = 0;
1763 int result = 0;
1764
1765 if (!buf)
1766 return -EINVAL;
1767
1768 size += sysfs_emit_at(buf, size, "%16s %s %s %s %s %s %s %s %s %s\n",
1769 title[0], title[1], title[2], title[3], title[4], title[5],
1770 title[6], title[7], title[8], title[9]);
1771
1772 for (i = 0; i < PP_SMC_POWER_PROFILE_COUNT; i++) {
1773 /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */
1774 workload_type = smu_cmn_to_asic_specific_index(smu,
1775 CMN2ASIC_MAPPING_WORKLOAD,
1776 i);
1777 if (workload_type == -ENOTSUPP)
1778 continue;
1779 else if (workload_type < 0)
1780 return -EINVAL;
1781
1782 result = smu_cmn_update_table(smu,
1783 SMU_TABLE_ACTIVITY_MONITOR_COEFF,
1784 workload_type,
1785 (void *)(&activity_monitor_external),
1786 false);
1787 if (result) {
1788 dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__);
1789 return result;
1790 }
1791
1792 size += sysfs_emit_at(buf, size, "%2d %14s%s:\n",
1793 i, amdgpu_pp_profile_name[i], (i == smu->power_profile_mode) ? "*" : " ");
1794
1795 size += sysfs_emit_at(buf, size, "%19s %d(%13s) %7d %7d %7d %7d %7d %7d %7d %7d\n",
1796 " ",
1797 0,
1798 "GFXCLK",
1799 activity_monitor->Gfx_FPS,
1800 activity_monitor->Gfx_MinActiveFreqType,
1801 activity_monitor->Gfx_MinActiveFreq,
1802 activity_monitor->Gfx_BoosterFreqType,
1803 activity_monitor->Gfx_BoosterFreq,
1804 activity_monitor->Gfx_PD_Data_limit_c,
1805 activity_monitor->Gfx_PD_Data_error_coeff,
1806 activity_monitor->Gfx_PD_Data_error_rate_coeff);
1807
1808 size += sysfs_emit_at(buf, size, "%19s %d(%13s) %7d %7d %7d %7d %7d %7d %7d %7d\n",
1809 " ",
1810 1,
1811 "FCLK",
1812 activity_monitor->Fclk_FPS,
1813 activity_monitor->Fclk_MinActiveFreqType,
1814 activity_monitor->Fclk_MinActiveFreq,
1815 activity_monitor->Fclk_BoosterFreqType,
1816 activity_monitor->Fclk_BoosterFreq,
1817 activity_monitor->Fclk_PD_Data_limit_c,
1818 activity_monitor->Fclk_PD_Data_error_coeff,
1819 activity_monitor->Fclk_PD_Data_error_rate_coeff);
1820 }
1821
1822 return size;
1823 }
1824
1825 #define SMU_14_0_2_CUSTOM_PARAMS_COUNT 9
1826 #define SMU_14_0_2_CUSTOM_PARAMS_CLOCK_COUNT 2
1827 #define SMU_14_0_2_CUSTOM_PARAMS_SIZE (SMU_14_0_2_CUSTOM_PARAMS_CLOCK_COUNT * SMU_14_0_2_CUSTOM_PARAMS_COUNT * sizeof(long))
1828
smu_v14_0_2_set_power_profile_mode_coeff(struct smu_context * smu,long * input)1829 static int smu_v14_0_2_set_power_profile_mode_coeff(struct smu_context *smu,
1830 long *input)
1831 {
1832 DpmActivityMonitorCoeffIntExternal_t activity_monitor_external;
1833 DpmActivityMonitorCoeffInt_t *activity_monitor =
1834 &(activity_monitor_external.DpmActivityMonitorCoeffInt);
1835 int ret, idx;
1836
1837 ret = smu_cmn_update_table(smu,
1838 SMU_TABLE_ACTIVITY_MONITOR_COEFF,
1839 WORKLOAD_PPLIB_CUSTOM_BIT,
1840 (void *)(&activity_monitor_external),
1841 false);
1842 if (ret) {
1843 dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__);
1844 return ret;
1845 }
1846
1847 idx = 0 * SMU_14_0_2_CUSTOM_PARAMS_COUNT;
1848 if (input[idx]) {
1849 /* Gfxclk */
1850 activity_monitor->Gfx_FPS = input[idx + 1];
1851 activity_monitor->Gfx_MinActiveFreqType = input[idx + 2];
1852 activity_monitor->Gfx_MinActiveFreq = input[idx + 3];
1853 activity_monitor->Gfx_BoosterFreqType = input[idx + 4];
1854 activity_monitor->Gfx_BoosterFreq = input[idx + 5];
1855 activity_monitor->Gfx_PD_Data_limit_c = input[idx + 6];
1856 activity_monitor->Gfx_PD_Data_error_coeff = input[idx + 7];
1857 activity_monitor->Gfx_PD_Data_error_rate_coeff = input[idx + 8];
1858 }
1859 idx = 1 * SMU_14_0_2_CUSTOM_PARAMS_COUNT;
1860 if (input[idx]) {
1861 /* Fclk */
1862 activity_monitor->Fclk_FPS = input[idx + 1];
1863 activity_monitor->Fclk_MinActiveFreqType = input[idx + 2];
1864 activity_monitor->Fclk_MinActiveFreq = input[idx + 3];
1865 activity_monitor->Fclk_BoosterFreqType = input[idx + 4];
1866 activity_monitor->Fclk_BoosterFreq = input[idx + 5];
1867 activity_monitor->Fclk_PD_Data_limit_c = input[idx + 6];
1868 activity_monitor->Fclk_PD_Data_error_coeff = input[idx + 7];
1869 activity_monitor->Fclk_PD_Data_error_rate_coeff = input[idx + 8];
1870 }
1871
1872 ret = smu_cmn_update_table(smu,
1873 SMU_TABLE_ACTIVITY_MONITOR_COEFF,
1874 WORKLOAD_PPLIB_CUSTOM_BIT,
1875 (void *)(&activity_monitor_external),
1876 true);
1877 if (ret) {
1878 dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__);
1879 return ret;
1880 }
1881
1882 return ret;
1883 }
1884
smu_v14_0_2_set_power_profile_mode(struct smu_context * smu,u32 workload_mask,long * custom_params,u32 custom_params_max_idx)1885 static int smu_v14_0_2_set_power_profile_mode(struct smu_context *smu,
1886 u32 workload_mask,
1887 long *custom_params,
1888 u32 custom_params_max_idx)
1889 {
1890 u32 backend_workload_mask = 0;
1891 int ret, idx = -1, i;
1892
1893 smu_cmn_get_backend_workload_mask(smu, workload_mask,
1894 &backend_workload_mask);
1895
1896 /* disable deep sleep if compute is enabled */
1897 if (workload_mask & (1 << PP_SMC_POWER_PROFILE_COMPUTE))
1898 smu_v14_0_deep_sleep_control(smu, false);
1899 else
1900 smu_v14_0_deep_sleep_control(smu, true);
1901
1902 if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) {
1903 if (!smu->custom_profile_params) {
1904 smu->custom_profile_params =
1905 kzalloc(SMU_14_0_2_CUSTOM_PARAMS_SIZE, GFP_KERNEL);
1906 if (!smu->custom_profile_params)
1907 return -ENOMEM;
1908 }
1909 if (custom_params && custom_params_max_idx) {
1910 if (custom_params_max_idx != SMU_14_0_2_CUSTOM_PARAMS_COUNT)
1911 return -EINVAL;
1912 if (custom_params[0] >= SMU_14_0_2_CUSTOM_PARAMS_CLOCK_COUNT)
1913 return -EINVAL;
1914 idx = custom_params[0] * SMU_14_0_2_CUSTOM_PARAMS_COUNT;
1915 smu->custom_profile_params[idx] = 1;
1916 for (i = 1; i < custom_params_max_idx; i++)
1917 smu->custom_profile_params[idx + i] = custom_params[i];
1918 }
1919 ret = smu_v14_0_2_set_power_profile_mode_coeff(smu,
1920 smu->custom_profile_params);
1921 if (ret) {
1922 if (idx != -1)
1923 smu->custom_profile_params[idx] = 0;
1924 return ret;
1925 }
1926 } else if (smu->custom_profile_params) {
1927 memset(smu->custom_profile_params, 0, SMU_14_0_2_CUSTOM_PARAMS_SIZE);
1928 }
1929
1930 ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetWorkloadMask,
1931 backend_workload_mask, NULL);
1932 if (ret) {
1933 dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n",
1934 workload_mask);
1935 if (idx != -1)
1936 smu->custom_profile_params[idx] = 0;
1937 return ret;
1938 }
1939
1940 return ret;
1941 }
1942
smu_v14_0_2_baco_enter(struct smu_context * smu)1943 static int smu_v14_0_2_baco_enter(struct smu_context *smu)
1944 {
1945 struct smu_baco_context *smu_baco = &smu->smu_baco;
1946 struct amdgpu_device *adev = smu->adev;
1947
1948 if (adev->in_runpm && smu_cmn_is_audio_func_enabled(adev))
1949 return smu_v14_0_baco_set_armd3_sequence(smu,
1950 smu_baco->maco_support ? BACO_SEQ_BAMACO : BACO_SEQ_BACO);
1951 else
1952 return smu_v14_0_baco_enter(smu);
1953 }
1954
smu_v14_0_2_baco_exit(struct smu_context * smu)1955 static int smu_v14_0_2_baco_exit(struct smu_context *smu)
1956 {
1957 struct amdgpu_device *adev = smu->adev;
1958
1959 if (adev->in_runpm && smu_cmn_is_audio_func_enabled(adev)) {
1960 /* Wait for PMFW handling for the Dstate change */
1961 usleep_range(10000, 11000);
1962 return smu_v14_0_baco_set_armd3_sequence(smu, BACO_SEQ_ULPS);
1963 } else {
1964 return smu_v14_0_baco_exit(smu);
1965 }
1966 }
1967
smu_v14_0_2_is_mode1_reset_supported(struct smu_context * smu)1968 static bool smu_v14_0_2_is_mode1_reset_supported(struct smu_context *smu)
1969 {
1970 // TODO
1971
1972 return true;
1973 }
1974
smu_v14_0_2_i2c_xfer(struct i2c_adapter * i2c_adap,struct i2c_msg * msg,int num_msgs)1975 static int smu_v14_0_2_i2c_xfer(struct i2c_adapter *i2c_adap,
1976 struct i2c_msg *msg, int num_msgs)
1977 {
1978 struct amdgpu_smu_i2c_bus *smu_i2c = i2c_get_adapdata(i2c_adap);
1979 struct amdgpu_device *adev = smu_i2c->adev;
1980 struct smu_context *smu = adev->powerplay.pp_handle;
1981 struct smu_table_context *smu_table = &smu->smu_table;
1982 struct smu_table *table = &smu_table->driver_table;
1983 SwI2cRequest_t *req, *res = (SwI2cRequest_t *)table->cpu_addr;
1984 int i, j, r, c;
1985 u16 dir;
1986
1987 if (!adev->pm.dpm_enabled)
1988 return -EBUSY;
1989
1990 req = kzalloc(sizeof(*req), GFP_KERNEL);
1991 if (!req)
1992 return -ENOMEM;
1993
1994 req->I2CcontrollerPort = smu_i2c->port;
1995 req->I2CSpeed = I2C_SPEED_FAST_400K;
1996 req->SlaveAddress = msg[0].addr << 1; /* wants an 8-bit address */
1997 dir = msg[0].flags & I2C_M_RD;
1998
1999 for (c = i = 0; i < num_msgs; i++) {
2000 for (j = 0; j < msg[i].len; j++, c++) {
2001 SwI2cCmd_t *cmd = &req->SwI2cCmds[c];
2002
2003 if (!(msg[i].flags & I2C_M_RD)) {
2004 /* write */
2005 cmd->CmdConfig |= CMDCONFIG_READWRITE_MASK;
2006 cmd->ReadWriteData = msg[i].buf[j];
2007 }
2008
2009 if ((dir ^ msg[i].flags) & I2C_M_RD) {
2010 /* The direction changes.
2011 */
2012 dir = msg[i].flags & I2C_M_RD;
2013 cmd->CmdConfig |= CMDCONFIG_RESTART_MASK;
2014 }
2015
2016 req->NumCmds++;
2017
2018 /*
2019 * Insert STOP if we are at the last byte of either last
2020 * message for the transaction or the client explicitly
2021 * requires a STOP at this particular message.
2022 */
2023 if ((j == msg[i].len - 1) &&
2024 ((i == num_msgs - 1) || (msg[i].flags & I2C_M_STOP))) {
2025 cmd->CmdConfig &= ~CMDCONFIG_RESTART_MASK;
2026 cmd->CmdConfig |= CMDCONFIG_STOP_MASK;
2027 }
2028 }
2029 }
2030 mutex_lock(&adev->pm.mutex);
2031 r = smu_cmn_update_table(smu, SMU_TABLE_I2C_COMMANDS, 0, req, true);
2032 mutex_unlock(&adev->pm.mutex);
2033 if (r)
2034 goto fail;
2035
2036 for (c = i = 0; i < num_msgs; i++) {
2037 if (!(msg[i].flags & I2C_M_RD)) {
2038 c += msg[i].len;
2039 continue;
2040 }
2041 for (j = 0; j < msg[i].len; j++, c++) {
2042 SwI2cCmd_t *cmd = &res->SwI2cCmds[c];
2043
2044 msg[i].buf[j] = cmd->ReadWriteData;
2045 }
2046 }
2047 r = num_msgs;
2048 fail:
2049 kfree(req);
2050 return r;
2051 }
2052
smu_v14_0_2_i2c_func(struct i2c_adapter * adap)2053 static u32 smu_v14_0_2_i2c_func(struct i2c_adapter *adap)
2054 {
2055 return I2C_FUNC_I2C | I2C_FUNC_SMBUS_EMUL;
2056 }
2057
2058 static const struct i2c_algorithm smu_v14_0_2_i2c_algo = {
2059 .master_xfer = smu_v14_0_2_i2c_xfer,
2060 .functionality = smu_v14_0_2_i2c_func,
2061 };
2062
2063 static const struct i2c_adapter_quirks smu_v14_0_2_i2c_control_quirks = {
2064 .flags = I2C_AQ_COMB | I2C_AQ_COMB_SAME_ADDR | I2C_AQ_NO_ZERO_LEN,
2065 .max_read_len = MAX_SW_I2C_COMMANDS,
2066 .max_write_len = MAX_SW_I2C_COMMANDS,
2067 .max_comb_1st_msg_len = 2,
2068 .max_comb_2nd_msg_len = MAX_SW_I2C_COMMANDS - 2,
2069 };
2070
smu_v14_0_2_i2c_control_init(struct smu_context * smu)2071 static int smu_v14_0_2_i2c_control_init(struct smu_context *smu)
2072 {
2073 struct amdgpu_device *adev = smu->adev;
2074 int res, i;
2075
2076 for (i = 0; i < MAX_SMU_I2C_BUSES; i++) {
2077 struct amdgpu_smu_i2c_bus *smu_i2c = &adev->pm.smu_i2c[i];
2078 struct i2c_adapter *control = &smu_i2c->adapter;
2079
2080 smu_i2c->adev = adev;
2081 smu_i2c->port = i;
2082 mutex_init(&smu_i2c->mutex);
2083 control->owner = THIS_MODULE;
2084 control->dev.parent = &adev->pdev->dev;
2085 control->algo = &smu_v14_0_2_i2c_algo;
2086 snprintf(control->name, sizeof(control->name), "AMDGPU SMU %d", i);
2087 control->quirks = &smu_v14_0_2_i2c_control_quirks;
2088 i2c_set_adapdata(control, smu_i2c);
2089
2090 res = i2c_add_adapter(control);
2091 if (res) {
2092 DRM_ERROR("Failed to register hw i2c, err: %d\n", res);
2093 goto Out_err;
2094 }
2095 }
2096
2097 /* assign the buses used for the FRU EEPROM and RAS EEPROM */
2098 /* XXX ideally this would be something in a vbios data table */
2099 adev->pm.ras_eeprom_i2c_bus = &adev->pm.smu_i2c[1].adapter;
2100 adev->pm.fru_eeprom_i2c_bus = &adev->pm.smu_i2c[0].adapter;
2101
2102 return 0;
2103 Out_err:
2104 for ( ; i >= 0; i--) {
2105 struct amdgpu_smu_i2c_bus *smu_i2c = &adev->pm.smu_i2c[i];
2106 struct i2c_adapter *control = &smu_i2c->adapter;
2107
2108 i2c_del_adapter(control);
2109 }
2110 return res;
2111 }
2112
smu_v14_0_2_i2c_control_fini(struct smu_context * smu)2113 static void smu_v14_0_2_i2c_control_fini(struct smu_context *smu)
2114 {
2115 struct amdgpu_device *adev = smu->adev;
2116 int i;
2117
2118 for (i = 0; i < MAX_SMU_I2C_BUSES; i++) {
2119 struct amdgpu_smu_i2c_bus *smu_i2c = &adev->pm.smu_i2c[i];
2120 struct i2c_adapter *control = &smu_i2c->adapter;
2121
2122 i2c_del_adapter(control);
2123 }
2124 adev->pm.ras_eeprom_i2c_bus = NULL;
2125 adev->pm.fru_eeprom_i2c_bus = NULL;
2126 }
2127
smu_v14_0_2_set_mp1_state(struct smu_context * smu,enum pp_mp1_state mp1_state)2128 static int smu_v14_0_2_set_mp1_state(struct smu_context *smu,
2129 enum pp_mp1_state mp1_state)
2130 {
2131 int ret;
2132
2133 switch (mp1_state) {
2134 case PP_MP1_STATE_UNLOAD:
2135 ret = smu_cmn_set_mp1_state(smu, mp1_state);
2136 break;
2137 default:
2138 /* Ignore others */
2139 ret = 0;
2140 }
2141
2142 return ret;
2143 }
2144
smu_v14_0_2_set_df_cstate(struct smu_context * smu,enum pp_df_cstate state)2145 static int smu_v14_0_2_set_df_cstate(struct smu_context *smu,
2146 enum pp_df_cstate state)
2147 {
2148 return smu_cmn_send_smc_msg_with_param(smu,
2149 SMU_MSG_DFCstateControl,
2150 state,
2151 NULL);
2152 }
2153
smu_v14_0_2_mode1_reset(struct smu_context * smu)2154 static int smu_v14_0_2_mode1_reset(struct smu_context *smu)
2155 {
2156 int ret = 0;
2157
2158 ret = smu_cmn_send_debug_smc_msg(smu, DEBUGSMC_MSG_Mode1Reset);
2159 if (!ret) {
2160 if (amdgpu_emu_mode == 1)
2161 msleep(50000);
2162 else
2163 msleep(1000);
2164 }
2165
2166 return ret;
2167 }
2168
smu_v14_0_2_mode2_reset(struct smu_context * smu)2169 static int smu_v14_0_2_mode2_reset(struct smu_context *smu)
2170 {
2171 int ret = 0;
2172
2173 // TODO
2174
2175 return ret;
2176 }
2177
smu_v14_0_2_enable_gfx_features(struct smu_context * smu)2178 static int smu_v14_0_2_enable_gfx_features(struct smu_context *smu)
2179 {
2180 struct amdgpu_device *adev = smu->adev;
2181
2182 if (amdgpu_ip_version(adev, MP1_HWIP, 0) == IP_VERSION(14, 0, 2))
2183 return smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_EnableAllSmuFeatures,
2184 FEATURE_PWR_GFX, NULL);
2185 else
2186 return -EOPNOTSUPP;
2187 }
2188
smu_v14_0_2_set_smu_mailbox_registers(struct smu_context * smu)2189 static void smu_v14_0_2_set_smu_mailbox_registers(struct smu_context *smu)
2190 {
2191 struct amdgpu_device *adev = smu->adev;
2192
2193 smu->param_reg = SOC15_REG_OFFSET(MP1, 0, regMP1_SMN_C2PMSG_82);
2194 smu->msg_reg = SOC15_REG_OFFSET(MP1, 0, regMP1_SMN_C2PMSG_66);
2195 smu->resp_reg = SOC15_REG_OFFSET(MP1, 0, regMP1_SMN_C2PMSG_90);
2196
2197 smu->debug_param_reg = SOC15_REG_OFFSET(MP1, 0, regMP1_SMN_C2PMSG_53);
2198 smu->debug_msg_reg = SOC15_REG_OFFSET(MP1, 0, regMP1_SMN_C2PMSG_75);
2199 smu->debug_resp_reg = SOC15_REG_OFFSET(MP1, 0, regMP1_SMN_C2PMSG_54);
2200 }
2201
smu_v14_0_2_get_gpu_metrics(struct smu_context * smu,void ** table)2202 static ssize_t smu_v14_0_2_get_gpu_metrics(struct smu_context *smu,
2203 void **table)
2204 {
2205 struct smu_table_context *smu_table = &smu->smu_table;
2206 struct gpu_metrics_v1_3 *gpu_metrics =
2207 (struct gpu_metrics_v1_3 *)smu_table->gpu_metrics_table;
2208 SmuMetricsExternal_t metrics_ext;
2209 SmuMetrics_t *metrics = &metrics_ext.SmuMetrics;
2210 int ret = 0;
2211
2212 ret = smu_cmn_get_metrics_table(smu,
2213 &metrics_ext,
2214 true);
2215 if (ret)
2216 return ret;
2217
2218 smu_cmn_init_soft_gpu_metrics(gpu_metrics, 1, 3);
2219
2220 gpu_metrics->temperature_edge = metrics->AvgTemperature[TEMP_EDGE];
2221 gpu_metrics->temperature_hotspot = metrics->AvgTemperature[TEMP_HOTSPOT];
2222 gpu_metrics->temperature_mem = metrics->AvgTemperature[TEMP_MEM];
2223 gpu_metrics->temperature_vrgfx = metrics->AvgTemperature[TEMP_VR_GFX];
2224 gpu_metrics->temperature_vrsoc = metrics->AvgTemperature[TEMP_VR_SOC];
2225 gpu_metrics->temperature_vrmem = max(metrics->AvgTemperature[TEMP_VR_MEM0],
2226 metrics->AvgTemperature[TEMP_VR_MEM1]);
2227
2228 gpu_metrics->average_gfx_activity = metrics->AverageGfxActivity;
2229 gpu_metrics->average_umc_activity = metrics->AverageUclkActivity;
2230 gpu_metrics->average_mm_activity = max(metrics->AverageVcn0ActivityPercentage,
2231 metrics->Vcn1ActivityPercentage);
2232
2233 gpu_metrics->average_socket_power = metrics->AverageSocketPower;
2234 gpu_metrics->energy_accumulator = metrics->EnergyAccumulator;
2235
2236 if (metrics->AverageGfxActivity <= SMU_14_0_2_BUSY_THRESHOLD)
2237 gpu_metrics->average_gfxclk_frequency = metrics->AverageGfxclkFrequencyPostDs;
2238 else
2239 gpu_metrics->average_gfxclk_frequency = metrics->AverageGfxclkFrequencyPreDs;
2240
2241 if (metrics->AverageUclkActivity <= SMU_14_0_2_BUSY_THRESHOLD)
2242 gpu_metrics->average_uclk_frequency = metrics->AverageMemclkFrequencyPostDs;
2243 else
2244 gpu_metrics->average_uclk_frequency = metrics->AverageMemclkFrequencyPreDs;
2245
2246 gpu_metrics->average_vclk0_frequency = metrics->AverageVclk0Frequency;
2247 gpu_metrics->average_dclk0_frequency = metrics->AverageDclk0Frequency;
2248 gpu_metrics->average_vclk1_frequency = metrics->AverageVclk1Frequency;
2249 gpu_metrics->average_dclk1_frequency = metrics->AverageDclk1Frequency;
2250
2251 gpu_metrics->current_gfxclk = gpu_metrics->average_gfxclk_frequency;
2252 gpu_metrics->current_socclk = metrics->CurrClock[PPCLK_SOCCLK];
2253 gpu_metrics->current_uclk = metrics->CurrClock[PPCLK_UCLK];
2254 gpu_metrics->current_vclk0 = metrics->CurrClock[PPCLK_VCLK_0];
2255 gpu_metrics->current_dclk0 = metrics->CurrClock[PPCLK_DCLK_0];
2256 gpu_metrics->current_vclk1 = metrics->CurrClock[PPCLK_VCLK_0];
2257 gpu_metrics->current_dclk1 = metrics->CurrClock[PPCLK_DCLK_0];
2258
2259 gpu_metrics->throttle_status =
2260 smu_v14_0_2_get_throttler_status(metrics);
2261 gpu_metrics->indep_throttle_status =
2262 smu_cmn_get_indep_throttler_status(gpu_metrics->throttle_status,
2263 smu_v14_0_2_throttler_map);
2264
2265 gpu_metrics->current_fan_speed = metrics->AvgFanRpm;
2266
2267 gpu_metrics->pcie_link_width = metrics->PcieWidth;
2268 if ((metrics->PcieRate - 1) > LINK_SPEED_MAX)
2269 gpu_metrics->pcie_link_speed = pcie_gen_to_speed(1);
2270 else
2271 gpu_metrics->pcie_link_speed = pcie_gen_to_speed(metrics->PcieRate);
2272
2273 gpu_metrics->system_clock_counter = ktime_get_boottime_ns();
2274
2275 gpu_metrics->voltage_gfx = metrics->AvgVoltage[SVI_PLANE_VDD_GFX];
2276 gpu_metrics->voltage_soc = metrics->AvgVoltage[SVI_PLANE_VDD_SOC];
2277 gpu_metrics->voltage_mem = metrics->AvgVoltage[SVI_PLANE_VDDIO_MEM];
2278
2279 *table = (void *)gpu_metrics;
2280
2281 return sizeof(struct gpu_metrics_v1_3);
2282 }
2283
smu_v14_0_2_dump_od_table(struct smu_context * smu,OverDriveTableExternal_t * od_table)2284 static void smu_v14_0_2_dump_od_table(struct smu_context *smu,
2285 OverDriveTableExternal_t *od_table)
2286 {
2287 struct amdgpu_device *adev = smu->adev;
2288
2289 dev_dbg(adev->dev, "OD: Gfxclk offset: (%d)\n", od_table->OverDriveTable.GfxclkFoffset);
2290 dev_dbg(adev->dev, "OD: Uclk: (%d, %d)\n", od_table->OverDriveTable.UclkFmin,
2291 od_table->OverDriveTable.UclkFmax);
2292 }
2293
smu_v14_0_2_upload_overdrive_table(struct smu_context * smu,OverDriveTableExternal_t * od_table)2294 static int smu_v14_0_2_upload_overdrive_table(struct smu_context *smu,
2295 OverDriveTableExternal_t *od_table)
2296 {
2297 int ret;
2298 ret = smu_cmn_update_table(smu,
2299 SMU_TABLE_OVERDRIVE,
2300 0,
2301 (void *)od_table,
2302 true);
2303 if (ret)
2304 dev_err(smu->adev->dev, "Failed to upload overdrive table!\n");
2305
2306 return ret;
2307 }
2308
smu_v14_0_2_set_supported_od_feature_mask(struct smu_context * smu)2309 static void smu_v14_0_2_set_supported_od_feature_mask(struct smu_context *smu)
2310 {
2311 struct amdgpu_device *adev = smu->adev;
2312
2313 if (smu_v14_0_2_is_od_feature_supported(smu,
2314 PP_OD_FEATURE_FAN_CURVE_BIT))
2315 adev->pm.od_feature_mask |= OD_OPS_SUPPORT_FAN_CURVE_RETRIEVE |
2316 OD_OPS_SUPPORT_FAN_CURVE_SET |
2317 OD_OPS_SUPPORT_ACOUSTIC_LIMIT_THRESHOLD_RETRIEVE |
2318 OD_OPS_SUPPORT_ACOUSTIC_LIMIT_THRESHOLD_SET |
2319 OD_OPS_SUPPORT_ACOUSTIC_TARGET_THRESHOLD_RETRIEVE |
2320 OD_OPS_SUPPORT_ACOUSTIC_TARGET_THRESHOLD_SET |
2321 OD_OPS_SUPPORT_FAN_TARGET_TEMPERATURE_RETRIEVE |
2322 OD_OPS_SUPPORT_FAN_TARGET_TEMPERATURE_SET |
2323 OD_OPS_SUPPORT_FAN_MINIMUM_PWM_RETRIEVE |
2324 OD_OPS_SUPPORT_FAN_MINIMUM_PWM_SET |
2325 OD_OPS_SUPPORT_FAN_ZERO_RPM_ENABLE_RETRIEVE |
2326 OD_OPS_SUPPORT_FAN_ZERO_RPM_ENABLE_SET;
2327 }
2328
smu_v14_0_2_get_overdrive_table(struct smu_context * smu,OverDriveTableExternal_t * od_table)2329 static int smu_v14_0_2_get_overdrive_table(struct smu_context *smu,
2330 OverDriveTableExternal_t *od_table)
2331 {
2332 int ret;
2333 ret = smu_cmn_update_table(smu,
2334 SMU_TABLE_OVERDRIVE,
2335 0,
2336 (void *)od_table,
2337 false);
2338 if (ret)
2339 dev_err(smu->adev->dev, "Failed to get overdrive table!\n");
2340
2341 return ret;
2342 }
2343
smu_v14_0_2_set_default_od_settings(struct smu_context * smu)2344 static int smu_v14_0_2_set_default_od_settings(struct smu_context *smu)
2345 {
2346 OverDriveTableExternal_t *od_table =
2347 (OverDriveTableExternal_t *)smu->smu_table.overdrive_table;
2348 OverDriveTableExternal_t *boot_od_table =
2349 (OverDriveTableExternal_t *)smu->smu_table.boot_overdrive_table;
2350 OverDriveTableExternal_t *user_od_table =
2351 (OverDriveTableExternal_t *)smu->smu_table.user_overdrive_table;
2352 OverDriveTableExternal_t user_od_table_bak;
2353 int ret;
2354 int i;
2355
2356 ret = smu_v14_0_2_get_overdrive_table(smu, boot_od_table);
2357 if (ret)
2358 return ret;
2359
2360 smu_v14_0_2_dump_od_table(smu, boot_od_table);
2361
2362 memcpy(od_table,
2363 boot_od_table,
2364 sizeof(OverDriveTableExternal_t));
2365
2366 /*
2367 * For S3/S4/Runpm resume, we need to setup those overdrive tables again,
2368 * but we have to preserve user defined values in "user_od_table".
2369 */
2370 if (!smu->adev->in_suspend) {
2371 memcpy(user_od_table,
2372 boot_od_table,
2373 sizeof(OverDriveTableExternal_t));
2374 smu->user_dpm_profile.user_od = false;
2375 } else if (smu->user_dpm_profile.user_od) {
2376 memcpy(&user_od_table_bak,
2377 user_od_table,
2378 sizeof(OverDriveTableExternal_t));
2379 memcpy(user_od_table,
2380 boot_od_table,
2381 sizeof(OverDriveTableExternal_t));
2382 user_od_table->OverDriveTable.GfxclkFoffset =
2383 user_od_table_bak.OverDriveTable.GfxclkFoffset;
2384 user_od_table->OverDriveTable.UclkFmin =
2385 user_od_table_bak.OverDriveTable.UclkFmin;
2386 user_od_table->OverDriveTable.UclkFmax =
2387 user_od_table_bak.OverDriveTable.UclkFmax;
2388 for (i = 0; i < PP_NUM_OD_VF_CURVE_POINTS; i++)
2389 user_od_table->OverDriveTable.VoltageOffsetPerZoneBoundary[i] =
2390 user_od_table_bak.OverDriveTable.VoltageOffsetPerZoneBoundary[i];
2391 for (i = 0; i < NUM_OD_FAN_MAX_POINTS - 1; i++) {
2392 user_od_table->OverDriveTable.FanLinearTempPoints[i] =
2393 user_od_table_bak.OverDriveTable.FanLinearTempPoints[i];
2394 user_od_table->OverDriveTable.FanLinearPwmPoints[i] =
2395 user_od_table_bak.OverDriveTable.FanLinearPwmPoints[i];
2396 }
2397 user_od_table->OverDriveTable.AcousticLimitRpmThreshold =
2398 user_od_table_bak.OverDriveTable.AcousticLimitRpmThreshold;
2399 user_od_table->OverDriveTable.AcousticTargetRpmThreshold =
2400 user_od_table_bak.OverDriveTable.AcousticTargetRpmThreshold;
2401 user_od_table->OverDriveTable.FanTargetTemperature =
2402 user_od_table_bak.OverDriveTable.FanTargetTemperature;
2403 user_od_table->OverDriveTable.FanMinimumPwm =
2404 user_od_table_bak.OverDriveTable.FanMinimumPwm;
2405 user_od_table->OverDriveTable.FanZeroRpmEnable =
2406 user_od_table_bak.OverDriveTable.FanZeroRpmEnable;
2407 }
2408
2409 smu_v14_0_2_set_supported_od_feature_mask(smu);
2410
2411 return 0;
2412 }
2413
smu_v14_0_2_restore_user_od_settings(struct smu_context * smu)2414 static int smu_v14_0_2_restore_user_od_settings(struct smu_context *smu)
2415 {
2416 struct smu_table_context *table_context = &smu->smu_table;
2417 OverDriveTableExternal_t *od_table = table_context->overdrive_table;
2418 OverDriveTableExternal_t *user_od_table = table_context->user_overdrive_table;
2419 int res;
2420
2421 user_od_table->OverDriveTable.FeatureCtrlMask = BIT(PP_OD_FEATURE_GFXCLK_BIT) |
2422 BIT(PP_OD_FEATURE_UCLK_BIT) |
2423 BIT(PP_OD_FEATURE_GFX_VF_CURVE_BIT) |
2424 BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2425 res = smu_v14_0_2_upload_overdrive_table(smu, user_od_table);
2426 user_od_table->OverDriveTable.FeatureCtrlMask = 0;
2427 if (res == 0)
2428 memcpy(od_table, user_od_table, sizeof(OverDriveTableExternal_t));
2429
2430 return res;
2431 }
2432
smu_v14_0_2_od_restore_table_single(struct smu_context * smu,long input)2433 static int smu_v14_0_2_od_restore_table_single(struct smu_context *smu, long input)
2434 {
2435 struct smu_table_context *table_context = &smu->smu_table;
2436 OverDriveTableExternal_t *boot_overdrive_table =
2437 (OverDriveTableExternal_t *)table_context->boot_overdrive_table;
2438 OverDriveTableExternal_t *od_table =
2439 (OverDriveTableExternal_t *)table_context->overdrive_table;
2440 struct amdgpu_device *adev = smu->adev;
2441 int i;
2442
2443 switch (input) {
2444 case PP_OD_EDIT_FAN_CURVE:
2445 for (i = 0; i < NUM_OD_FAN_MAX_POINTS; i++) {
2446 od_table->OverDriveTable.FanLinearTempPoints[i] =
2447 boot_overdrive_table->OverDriveTable.FanLinearTempPoints[i];
2448 od_table->OverDriveTable.FanLinearPwmPoints[i] =
2449 boot_overdrive_table->OverDriveTable.FanLinearPwmPoints[i];
2450 }
2451 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2452 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2453 break;
2454 case PP_OD_EDIT_FAN_ZERO_RPM_ENABLE:
2455 od_table->OverDriveTable.FanZeroRpmEnable =
2456 boot_overdrive_table->OverDriveTable.FanZeroRpmEnable;
2457 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_ZERO_FAN_BIT);
2458 break;
2459 case PP_OD_EDIT_ACOUSTIC_LIMIT:
2460 od_table->OverDriveTable.AcousticLimitRpmThreshold =
2461 boot_overdrive_table->OverDriveTable.AcousticLimitRpmThreshold;
2462 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2463 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2464 break;
2465 case PP_OD_EDIT_ACOUSTIC_TARGET:
2466 od_table->OverDriveTable.AcousticTargetRpmThreshold =
2467 boot_overdrive_table->OverDriveTable.AcousticTargetRpmThreshold;
2468 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2469 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2470 break;
2471 case PP_OD_EDIT_FAN_TARGET_TEMPERATURE:
2472 od_table->OverDriveTable.FanTargetTemperature =
2473 boot_overdrive_table->OverDriveTable.FanTargetTemperature;
2474 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2475 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2476 break;
2477 case PP_OD_EDIT_FAN_MINIMUM_PWM:
2478 od_table->OverDriveTable.FanMinimumPwm =
2479 boot_overdrive_table->OverDriveTable.FanMinimumPwm;
2480 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2481 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2482 break;
2483 default:
2484 dev_info(adev->dev, "Invalid table index: %ld\n", input);
2485 return -EINVAL;
2486 }
2487
2488 return 0;
2489 }
2490
smu_v14_0_2_od_edit_dpm_table(struct smu_context * smu,enum PP_OD_DPM_TABLE_COMMAND type,long input[],uint32_t size)2491 static int smu_v14_0_2_od_edit_dpm_table(struct smu_context *smu,
2492 enum PP_OD_DPM_TABLE_COMMAND type,
2493 long input[],
2494 uint32_t size)
2495 {
2496 struct smu_table_context *table_context = &smu->smu_table;
2497 OverDriveTableExternal_t *od_table =
2498 (OverDriveTableExternal_t *)table_context->overdrive_table;
2499 struct amdgpu_device *adev = smu->adev;
2500 uint32_t offset_of_voltageoffset;
2501 int32_t minimum, maximum;
2502 uint32_t feature_ctrlmask;
2503 int i, ret = 0;
2504
2505 switch (type) {
2506 case PP_OD_EDIT_SCLK_VDDC_TABLE:
2507 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_GFXCLK_BIT)) {
2508 dev_warn(adev->dev, "GFXCLK_LIMITS setting not supported!\n");
2509 return -ENOTSUPP;
2510 }
2511
2512 if (size != 1) {
2513 dev_info(adev->dev, "invalid number of input parameters %d\n", size);
2514 return -EINVAL;
2515 }
2516
2517 smu_v14_0_2_get_od_setting_limits(smu,
2518 PP_OD_FEATURE_GFXCLK_FMAX,
2519 &minimum,
2520 &maximum);
2521 if (input[0] < minimum ||
2522 input[0] > maximum) {
2523 dev_info(adev->dev, "GfxclkFoffset must be within [%d, %u]!\n",
2524 minimum, maximum);
2525 return -EINVAL;
2526 }
2527
2528 od_table->OverDriveTable.GfxclkFoffset = input[0];
2529 od_table->OverDriveTable.FeatureCtrlMask |= 1U << PP_OD_FEATURE_GFXCLK_BIT;
2530 break;
2531
2532 case PP_OD_EDIT_MCLK_VDDC_TABLE:
2533 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_UCLK_BIT)) {
2534 dev_warn(adev->dev, "UCLK_LIMITS setting not supported!\n");
2535 return -ENOTSUPP;
2536 }
2537
2538 for (i = 0; i < size; i += 2) {
2539 if (i + 2 > size) {
2540 dev_info(adev->dev, "invalid number of input parameters %d\n", size);
2541 return -EINVAL;
2542 }
2543
2544 switch (input[i]) {
2545 case 0:
2546 smu_v14_0_2_get_od_setting_limits(smu,
2547 PP_OD_FEATURE_UCLK_FMIN,
2548 &minimum,
2549 &maximum);
2550 if (input[i + 1] < minimum ||
2551 input[i + 1] > maximum) {
2552 dev_info(adev->dev, "UclkFmin (%ld) must be within [%u, %u]!\n",
2553 input[i + 1], minimum, maximum);
2554 return -EINVAL;
2555 }
2556
2557 od_table->OverDriveTable.UclkFmin = input[i + 1];
2558 od_table->OverDriveTable.FeatureCtrlMask |= 1U << PP_OD_FEATURE_UCLK_BIT;
2559 break;
2560
2561 case 1:
2562 smu_v14_0_2_get_od_setting_limits(smu,
2563 PP_OD_FEATURE_UCLK_FMAX,
2564 &minimum,
2565 &maximum);
2566 if (input[i + 1] < minimum ||
2567 input[i + 1] > maximum) {
2568 dev_info(adev->dev, "UclkFmax (%ld) must be within [%u, %u]!\n",
2569 input[i + 1], minimum, maximum);
2570 return -EINVAL;
2571 }
2572
2573 od_table->OverDriveTable.UclkFmax = input[i + 1];
2574 od_table->OverDriveTable.FeatureCtrlMask |= 1U << PP_OD_FEATURE_UCLK_BIT;
2575 break;
2576
2577 default:
2578 dev_info(adev->dev, "Invalid MCLK_VDDC_TABLE index: %ld\n", input[i]);
2579 dev_info(adev->dev, "Supported indices: [0:min,1:max]\n");
2580 return -EINVAL;
2581 }
2582 }
2583
2584 if (od_table->OverDriveTable.UclkFmin > od_table->OverDriveTable.UclkFmax) {
2585 dev_err(adev->dev,
2586 "Invalid setting: UclkFmin(%u) is bigger than UclkFmax(%u)\n",
2587 (uint32_t)od_table->OverDriveTable.UclkFmin,
2588 (uint32_t)od_table->OverDriveTable.UclkFmax);
2589 return -EINVAL;
2590 }
2591 break;
2592
2593 case PP_OD_EDIT_VDDGFX_OFFSET:
2594 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_GFX_VF_CURVE_BIT)) {
2595 dev_warn(adev->dev, "Gfx offset setting not supported!\n");
2596 return -ENOTSUPP;
2597 }
2598
2599 smu_v14_0_2_get_od_setting_limits(smu,
2600 PP_OD_FEATURE_GFX_VF_CURVE,
2601 &minimum,
2602 &maximum);
2603 if (input[0] < minimum ||
2604 input[0] > maximum) {
2605 dev_info(adev->dev, "Voltage offset (%ld) must be within [%d, %d]!\n",
2606 input[0], minimum, maximum);
2607 return -EINVAL;
2608 }
2609
2610 for (i = 0; i < PP_NUM_OD_VF_CURVE_POINTS; i++)
2611 od_table->OverDriveTable.VoltageOffsetPerZoneBoundary[i] = input[0];
2612 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_GFX_VF_CURVE_BIT);
2613 break;
2614
2615 case PP_OD_EDIT_FAN_CURVE:
2616 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_FAN_CURVE_BIT)) {
2617 dev_warn(adev->dev, "Fan curve setting not supported!\n");
2618 return -ENOTSUPP;
2619 }
2620
2621 if (input[0] >= NUM_OD_FAN_MAX_POINTS - 1 ||
2622 input[0] < 0)
2623 return -EINVAL;
2624
2625 smu_v14_0_2_get_od_setting_limits(smu,
2626 PP_OD_FEATURE_FAN_CURVE_TEMP,
2627 &minimum,
2628 &maximum);
2629 if (input[1] < minimum ||
2630 input[1] > maximum) {
2631 dev_info(adev->dev, "Fan curve temp setting(%ld) must be within [%d, %d]!\n",
2632 input[1], minimum, maximum);
2633 return -EINVAL;
2634 }
2635
2636 smu_v14_0_2_get_od_setting_limits(smu,
2637 PP_OD_FEATURE_FAN_CURVE_PWM,
2638 &minimum,
2639 &maximum);
2640 if (input[2] < minimum ||
2641 input[2] > maximum) {
2642 dev_info(adev->dev, "Fan curve pwm setting(%ld) must be within [%d, %d]!\n",
2643 input[2], minimum, maximum);
2644 return -EINVAL;
2645 }
2646
2647 od_table->OverDriveTable.FanLinearTempPoints[input[0]] = input[1];
2648 od_table->OverDriveTable.FanLinearPwmPoints[input[0]] = input[2];
2649 od_table->OverDriveTable.FanMode = FAN_MODE_MANUAL_LINEAR;
2650 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2651 break;
2652
2653 case PP_OD_EDIT_ACOUSTIC_LIMIT:
2654 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_FAN_CURVE_BIT)) {
2655 dev_warn(adev->dev, "Fan curve setting not supported!\n");
2656 return -ENOTSUPP;
2657 }
2658
2659 smu_v14_0_2_get_od_setting_limits(smu,
2660 PP_OD_FEATURE_FAN_ACOUSTIC_LIMIT,
2661 &minimum,
2662 &maximum);
2663 if (input[0] < minimum ||
2664 input[0] > maximum) {
2665 dev_info(adev->dev, "acoustic limit threshold setting(%ld) must be within [%d, %d]!\n",
2666 input[0], minimum, maximum);
2667 return -EINVAL;
2668 }
2669
2670 od_table->OverDriveTable.AcousticLimitRpmThreshold = input[0];
2671 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2672 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2673 break;
2674
2675 case PP_OD_EDIT_ACOUSTIC_TARGET:
2676 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_FAN_CURVE_BIT)) {
2677 dev_warn(adev->dev, "Fan curve setting not supported!\n");
2678 return -ENOTSUPP;
2679 }
2680
2681 smu_v14_0_2_get_od_setting_limits(smu,
2682 PP_OD_FEATURE_FAN_ACOUSTIC_TARGET,
2683 &minimum,
2684 &maximum);
2685 if (input[0] < minimum ||
2686 input[0] > maximum) {
2687 dev_info(adev->dev, "acoustic target threshold setting(%ld) must be within [%d, %d]!\n",
2688 input[0], minimum, maximum);
2689 return -EINVAL;
2690 }
2691
2692 od_table->OverDriveTable.AcousticTargetRpmThreshold = input[0];
2693 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2694 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2695 break;
2696
2697 case PP_OD_EDIT_FAN_TARGET_TEMPERATURE:
2698 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_FAN_CURVE_BIT)) {
2699 dev_warn(adev->dev, "Fan curve setting not supported!\n");
2700 return -ENOTSUPP;
2701 }
2702
2703 smu_v14_0_2_get_od_setting_limits(smu,
2704 PP_OD_FEATURE_FAN_TARGET_TEMPERATURE,
2705 &minimum,
2706 &maximum);
2707 if (input[0] < minimum ||
2708 input[0] > maximum) {
2709 dev_info(adev->dev, "fan target temperature setting(%ld) must be within [%d, %d]!\n",
2710 input[0], minimum, maximum);
2711 return -EINVAL;
2712 }
2713
2714 od_table->OverDriveTable.FanTargetTemperature = input[0];
2715 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2716 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2717 break;
2718
2719 case PP_OD_EDIT_FAN_MINIMUM_PWM:
2720 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_FAN_CURVE_BIT)) {
2721 dev_warn(adev->dev, "Fan curve setting not supported!\n");
2722 return -ENOTSUPP;
2723 }
2724
2725 smu_v14_0_2_get_od_setting_limits(smu,
2726 PP_OD_FEATURE_FAN_MINIMUM_PWM,
2727 &minimum,
2728 &maximum);
2729 if (input[0] < minimum ||
2730 input[0] > maximum) {
2731 dev_info(adev->dev, "fan minimum pwm setting(%ld) must be within [%d, %d]!\n",
2732 input[0], minimum, maximum);
2733 return -EINVAL;
2734 }
2735
2736 od_table->OverDriveTable.FanMinimumPwm = input[0];
2737 od_table->OverDriveTable.FanMode = FAN_MODE_AUTO;
2738 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_FAN_CURVE_BIT);
2739 break;
2740
2741 case PP_OD_EDIT_FAN_ZERO_RPM_ENABLE:
2742 if (!smu_v14_0_2_is_od_feature_supported(smu, PP_OD_FEATURE_ZERO_FAN_BIT)) {
2743 dev_warn(adev->dev, "Zero RPM setting not supported!\n");
2744 return -ENOTSUPP;
2745 }
2746
2747 smu_v14_0_2_get_od_setting_limits(smu,
2748 PP_OD_FEATURE_FAN_ZERO_RPM_ENABLE,
2749 &minimum,
2750 &maximum);
2751 if (input[0] < minimum ||
2752 input[0] > maximum) {
2753 dev_info(adev->dev, "zero RPM enable setting(%ld) must be within [%d, %d]!\n",
2754 input[0], minimum, maximum);
2755 return -EINVAL;
2756 }
2757
2758 od_table->OverDriveTable.FanZeroRpmEnable = input[0];
2759 od_table->OverDriveTable.FeatureCtrlMask |= BIT(PP_OD_FEATURE_ZERO_FAN_BIT);
2760 break;
2761
2762 case PP_OD_RESTORE_DEFAULT_TABLE:
2763 if (size == 1) {
2764 ret = smu_v14_0_2_od_restore_table_single(smu, input[0]);
2765 if (ret)
2766 return ret;
2767 } else {
2768 feature_ctrlmask = od_table->OverDriveTable.FeatureCtrlMask;
2769 memcpy(od_table,
2770 table_context->boot_overdrive_table,
2771 sizeof(OverDriveTableExternal_t));
2772 od_table->OverDriveTable.FeatureCtrlMask = feature_ctrlmask;
2773 }
2774 fallthrough;
2775 case PP_OD_COMMIT_DPM_TABLE:
2776 /*
2777 * The member below instructs PMFW the settings focused in
2778 * this single operation.
2779 * `uint32_t FeatureCtrlMask;`
2780 * It does not contain actual informations about user's custom
2781 * settings. Thus we do not cache it.
2782 */
2783 offset_of_voltageoffset = offsetof(OverDriveTable_t, VoltageOffsetPerZoneBoundary);
2784 if (memcmp((u8 *)od_table + offset_of_voltageoffset,
2785 table_context->user_overdrive_table + offset_of_voltageoffset,
2786 sizeof(OverDriveTableExternal_t) - offset_of_voltageoffset)) {
2787 smu_v14_0_2_dump_od_table(smu, od_table);
2788
2789 ret = smu_v14_0_2_upload_overdrive_table(smu, od_table);
2790 if (ret) {
2791 dev_err(adev->dev, "Failed to upload overdrive table!\n");
2792 return ret;
2793 }
2794
2795 od_table->OverDriveTable.FeatureCtrlMask = 0;
2796 memcpy(table_context->user_overdrive_table + offset_of_voltageoffset,
2797 (u8 *)od_table + offset_of_voltageoffset,
2798 sizeof(OverDriveTableExternal_t) - offset_of_voltageoffset);
2799
2800 if (!memcmp(table_context->user_overdrive_table,
2801 table_context->boot_overdrive_table,
2802 sizeof(OverDriveTableExternal_t)))
2803 smu->user_dpm_profile.user_od = false;
2804 else
2805 smu->user_dpm_profile.user_od = true;
2806 }
2807 break;
2808
2809 default:
2810 return -ENOSYS;
2811 }
2812
2813 return ret;
2814 }
2815
smu_v14_0_2_set_power_limit(struct smu_context * smu,enum smu_ppt_limit_type limit_type,uint32_t limit)2816 static int smu_v14_0_2_set_power_limit(struct smu_context *smu,
2817 enum smu_ppt_limit_type limit_type,
2818 uint32_t limit)
2819 {
2820 PPTable_t *pptable = smu->smu_table.driver_pptable;
2821 uint32_t msg_limit = pptable->SkuTable.MsgLimits.Power[PPT_THROTTLER_PPT0][POWER_SOURCE_AC];
2822 struct smu_table_context *table_context = &smu->smu_table;
2823 OverDriveTableExternal_t *od_table =
2824 (OverDriveTableExternal_t *)table_context->overdrive_table;
2825 int ret = 0;
2826
2827 if (limit_type != SMU_DEFAULT_PPT_LIMIT)
2828 return -EINVAL;
2829
2830 if (limit <= msg_limit) {
2831 if (smu->current_power_limit > msg_limit) {
2832 od_table->OverDriveTable.Ppt = 0;
2833 od_table->OverDriveTable.FeatureCtrlMask |= 1U << PP_OD_FEATURE_PPT_BIT;
2834
2835 ret = smu_v14_0_2_upload_overdrive_table(smu, od_table);
2836 if (ret) {
2837 dev_err(smu->adev->dev, "Failed to upload overdrive table!\n");
2838 return ret;
2839 }
2840 }
2841 return smu_v14_0_set_power_limit(smu, limit_type, limit);
2842 } else if (smu->od_enabled) {
2843 ret = smu_v14_0_set_power_limit(smu, limit_type, msg_limit);
2844 if (ret)
2845 return ret;
2846
2847 od_table->OverDriveTable.Ppt = (limit * 100) / msg_limit - 100;
2848 od_table->OverDriveTable.FeatureCtrlMask |= 1U << PP_OD_FEATURE_PPT_BIT;
2849
2850 ret = smu_v14_0_2_upload_overdrive_table(smu, od_table);
2851 if (ret) {
2852 dev_err(smu->adev->dev, "Failed to upload overdrive table!\n");
2853 return ret;
2854 }
2855
2856 smu->current_power_limit = limit;
2857 } else {
2858 return -EINVAL;
2859 }
2860
2861 return 0;
2862 }
2863
2864 static const struct pptable_funcs smu_v14_0_2_ppt_funcs = {
2865 .get_allowed_feature_mask = smu_v14_0_2_get_allowed_feature_mask,
2866 .set_default_dpm_table = smu_v14_0_2_set_default_dpm_table,
2867 .i2c_init = smu_v14_0_2_i2c_control_init,
2868 .i2c_fini = smu_v14_0_2_i2c_control_fini,
2869 .is_dpm_running = smu_v14_0_2_is_dpm_running,
2870 .init_microcode = smu_v14_0_init_microcode,
2871 .load_microcode = smu_v14_0_load_microcode,
2872 .fini_microcode = smu_v14_0_fini_microcode,
2873 .init_smc_tables = smu_v14_0_2_init_smc_tables,
2874 .fini_smc_tables = smu_v14_0_fini_smc_tables,
2875 .init_power = smu_v14_0_init_power,
2876 .fini_power = smu_v14_0_fini_power,
2877 .check_fw_status = smu_v14_0_check_fw_status,
2878 .setup_pptable = smu_v14_0_2_setup_pptable,
2879 .check_fw_version = smu_v14_0_check_fw_version,
2880 .set_driver_table_location = smu_v14_0_set_driver_table_location,
2881 .system_features_control = smu_v14_0_system_features_control,
2882 .set_allowed_mask = smu_v14_0_set_allowed_mask,
2883 .get_enabled_mask = smu_cmn_get_enabled_mask,
2884 .dpm_set_vcn_enable = smu_v14_0_set_vcn_enable,
2885 .dpm_set_jpeg_enable = smu_v14_0_set_jpeg_enable,
2886 .get_dpm_ultimate_freq = smu_v14_0_2_get_dpm_ultimate_freq,
2887 .get_vbios_bootup_values = smu_v14_0_get_vbios_bootup_values,
2888 .read_sensor = smu_v14_0_2_read_sensor,
2889 .feature_is_enabled = smu_cmn_feature_is_enabled,
2890 .print_clk_levels = smu_v14_0_2_print_clk_levels,
2891 .force_clk_levels = smu_v14_0_2_force_clk_levels,
2892 .update_pcie_parameters = smu_v14_0_2_update_pcie_parameters,
2893 .get_thermal_temperature_range = smu_v14_0_2_get_thermal_temperature_range,
2894 .register_irq_handler = smu_v14_0_register_irq_handler,
2895 .enable_thermal_alert = smu_v14_0_enable_thermal_alert,
2896 .disable_thermal_alert = smu_v14_0_disable_thermal_alert,
2897 .notify_memory_pool_location = smu_v14_0_notify_memory_pool_location,
2898 .get_gpu_metrics = smu_v14_0_2_get_gpu_metrics,
2899 .set_soft_freq_limited_range = smu_v14_0_set_soft_freq_limited_range,
2900 .set_default_od_settings = smu_v14_0_2_set_default_od_settings,
2901 .restore_user_od_settings = smu_v14_0_2_restore_user_od_settings,
2902 .od_edit_dpm_table = smu_v14_0_2_od_edit_dpm_table,
2903 .init_pptable_microcode = smu_v14_0_init_pptable_microcode,
2904 .populate_umd_state_clk = smu_v14_0_2_populate_umd_state_clk,
2905 .set_performance_level = smu_v14_0_set_performance_level,
2906 .gfx_off_control = smu_v14_0_gfx_off_control,
2907 .get_unique_id = smu_v14_0_2_get_unique_id,
2908 .get_fan_speed_pwm = smu_v14_0_2_get_fan_speed_pwm,
2909 .get_fan_speed_rpm = smu_v14_0_2_get_fan_speed_rpm,
2910 .get_power_limit = smu_v14_0_2_get_power_limit,
2911 .set_power_limit = smu_v14_0_2_set_power_limit,
2912 .get_power_profile_mode = smu_v14_0_2_get_power_profile_mode,
2913 .set_power_profile_mode = smu_v14_0_2_set_power_profile_mode,
2914 .run_btc = smu_v14_0_run_btc,
2915 .get_pp_feature_mask = smu_cmn_get_pp_feature_mask,
2916 .set_pp_feature_mask = smu_cmn_set_pp_feature_mask,
2917 .set_tool_table_location = smu_v14_0_set_tool_table_location,
2918 .deep_sleep_control = smu_v14_0_deep_sleep_control,
2919 .gfx_ulv_control = smu_v14_0_gfx_ulv_control,
2920 .get_bamaco_support = smu_v14_0_get_bamaco_support,
2921 .baco_get_state = smu_v14_0_baco_get_state,
2922 .baco_set_state = smu_v14_0_baco_set_state,
2923 .baco_enter = smu_v14_0_2_baco_enter,
2924 .baco_exit = smu_v14_0_2_baco_exit,
2925 .mode1_reset_is_support = smu_v14_0_2_is_mode1_reset_supported,
2926 .mode1_reset = smu_v14_0_2_mode1_reset,
2927 .mode2_reset = smu_v14_0_2_mode2_reset,
2928 .enable_gfx_features = smu_v14_0_2_enable_gfx_features,
2929 .set_mp1_state = smu_v14_0_2_set_mp1_state,
2930 .set_df_cstate = smu_v14_0_2_set_df_cstate,
2931 #if 0
2932 .gpo_control = smu_v14_0_gpo_control,
2933 #endif
2934 };
2935
smu_v14_0_2_set_ppt_funcs(struct smu_context * smu)2936 void smu_v14_0_2_set_ppt_funcs(struct smu_context *smu)
2937 {
2938 smu->ppt_funcs = &smu_v14_0_2_ppt_funcs;
2939 smu->message_map = smu_v14_0_2_message_map;
2940 smu->clock_map = smu_v14_0_2_clk_map;
2941 smu->feature_map = smu_v14_0_2_feature_mask_map;
2942 smu->table_map = smu_v14_0_2_table_map;
2943 smu->pwr_src_map = smu_v14_0_2_pwr_src_map;
2944 smu->workload_map = smu_v14_0_2_workload_map;
2945 smu_v14_0_2_set_smu_mailbox_registers(smu);
2946 }
2947