Home
last modified time | relevance | path

Searched refs:smu_dpm (Results 1 – 19 of 19) sorted by relevance

/linux/drivers/gpu/drm/amd/pm/swsmu/smu13/
H A Daldebaran_ppt.c294 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in aldebaran_allocate_dpm_context() local
297 smu_dpm->dpm_context = kzalloc_obj(struct smu_13_0_dpm_context); in aldebaran_allocate_dpm_context()
298 if (!smu_dpm->dpm_context) in aldebaran_allocate_dpm_context()
300 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); in aldebaran_allocate_dpm_context()
302 smu_dpm->dpm_policies = in aldebaran_allocate_dpm_context()
305 if (!smu_dpm->dpm_policies) in aldebaran_allocate_dpm_context()
308 policy = &(smu_dpm->dpm_policies->policies[0]); in aldebaran_allocate_dpm_context()
314 smu_dpm->dpm_policies->policy_mask |= BIT(PP_PM_POLICY_XGMI_PLPD); in aldebaran_allocate_dpm_context()
345 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in aldebaran_get_dpm_ultimate_freq()
398 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in aldebaran_set_default_dpm_table()
[all …]
H A Dsmu_v13_0_6_ppt.c274 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_6_cap_set()
282 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_6_cap_clear()
290 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_6_cap_supported()
689 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_6_allocate_dpm_context() local
692 smu_dpm->dpm_context = in smu_v13_0_6_allocate_dpm_context()
694 if (!smu_dpm->dpm_context) in smu_v13_0_6_allocate_dpm_context()
696 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); in smu_v13_0_6_allocate_dpm_context()
698 smu_dpm->dpm_policies = in smu_v13_0_6_allocate_dpm_context()
700 if (!smu_dpm in smu_v13_0_6_allocate_dpm_context()
1115 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_6_set_default_dpm_table() local
1375 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_6_emit_clk_levels() local
1970 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); smu_v13_0_6_set_performance_level() local
2034 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); smu_v13_0_6_set_soft_freq_limited_range() local
2123 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); smu_v13_0_6_usr_edit_dpm_table() local
[all...]
H A Dsmu_v13_0.c508 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_fini_smc_tables() local
534 kfree(smu_dpm->dpm_policies); in smu_v13_0_fini_smc_tables()
535 kfree(smu_dpm->dpm_context); in smu_v13_0_fini_smc_tables()
536 kfree(smu_dpm->golden_dpm_context); in smu_v13_0_fini_smc_tables()
537 kfree(smu_dpm->dpm_current_power_state); in smu_v13_0_fini_smc_tables()
538 kfree(smu_dpm->dpm_request_power_state); in smu_v13_0_fini_smc_tables()
539 smu_dpm->dpm_policies = NULL; in smu_v13_0_fini_smc_tables()
540 smu_dpm->dpm_context = NULL; in smu_v13_0_fini_smc_tables()
541 smu_dpm->golden_dpm_context = NULL; in smu_v13_0_fini_smc_tables()
542 smu_dpm->dpm_context_size = 0; in smu_v13_0_fini_smc_tables()
[all …]
H A Dsmu_v13_0_7_ppt.c564 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_allocate_dpm_context()
566 smu_dpm->dpm_context = kzalloc_obj(struct smu_13_0_dpm_context); in smu_v13_0_7_allocate_dpm_context()
567 if (!smu_dpm->dpm_context) in smu_v13_0_7_allocate_dpm_context()
570 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context);
592 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_7_set_default_dpm_table()
874 smu->smu_dpm.dpm_context; in smu_v13_0_7_get_dpm_ultimate_freq()
1214 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_emit_clk_levels()
1215 struct smu_13_0_dpm_context *dpm_context = smu_dpm in smu_v13_0_7_emit_clk_levels()
560 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_7_allocate_dpm_context() local
1183 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_7_emit_clk_levels() local
1924 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_7_force_clk_levels() local
[all...]
H A Dsmu_v13_0_0_ppt.c535 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_allocate_dpm_context()
537 smu_dpm->dpm_context = kzalloc_obj(struct smu_13_0_dpm_context); in smu_v13_0_0_allocate_dpm_context()
538 if (!smu_dpm->dpm_context) in smu_v13_0_0_allocate_dpm_context()
541 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context);
563 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_0_set_default_dpm_table()
864 smu->smu_dpm.dpm_context; in smu_v13_0_0_get_dpm_ultimate_freq()
1204 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_emit_clk_levels()
1205 struct smu_13_0_dpm_context *dpm_context = smu_dpm in smu_v13_0_0_emit_clk_levels()
531 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_0_allocate_dpm_context() local
1173 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_0_emit_clk_levels() local
1914 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_0_force_clk_levels() local
[all...]
H A Dsmu_v13_0_5_ppt.c522 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in smu_v13_0_5_od_edit_dpm_table() local
526 if (smu_dpm->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) in smu_v13_0_5_od_edit_dpm_table()
H A Dyellow_carp_ppt.c653 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in yellow_carp_od_edit_dpm_table() local
657 if (smu_dpm->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) in yellow_carp_od_edit_dpm_table()
H A Dsmu_v13_0_12_ppt.c283 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_12_setup_driver_pptable()
/linux/drivers/gpu/drm/amd/pm/swsmu/smu11/
H A Darcturus_ppt.c307 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_allocate_dpm_context() local
310 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in arcturus_allocate_dpm_context()
311 if (!smu_dpm->dpm_context) in arcturus_allocate_dpm_context()
313 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in arcturus_allocate_dpm_context()
315 smu_dpm->dpm_policies = in arcturus_allocate_dpm_context()
318 if (!smu_dpm->dpm_policies) in arcturus_allocate_dpm_context()
321 policy = &(smu_dpm->dpm_policies->policies[0]); in arcturus_allocate_dpm_context()
327 smu_dpm->dpm_policies->policy_mask |= BIT(PP_PM_POLICY_XGMI_PLPD); in arcturus_allocate_dpm_context()
357 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in arcturus_set_default_dpm_table()
432 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_set_default_dpm_table() local
[all …]
H A Dsmu_v11_0.c429 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v11_0_fini_smc_tables() local
457 kfree(smu_dpm->dpm_context); in smu_v11_0_fini_smc_tables()
458 kfree(smu_dpm->golden_dpm_context); in smu_v11_0_fini_smc_tables()
459 kfree(smu_dpm->dpm_current_power_state); in smu_v11_0_fini_smc_tables()
460 kfree(smu_dpm->dpm_request_power_state); in smu_v11_0_fini_smc_tables()
461 smu_dpm->dpm_context = NULL; in smu_v11_0_fini_smc_tables()
462 smu_dpm->golden_dpm_context = NULL; in smu_v11_0_fini_smc_tables()
463 smu_dpm->dpm_context_size = 0; in smu_v11_0_fini_smc_tables()
464 smu_dpm->dpm_current_power_state = NULL; in smu_v11_0_fini_smc_tables()
465 smu_dpm->dpm_request_power_state = NULL; in smu_v11_0_fini_smc_tables()
[all …]
H A Dnavi10_ppt.c935 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_allocate_dpm_context() local
937 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in navi10_allocate_dpm_context()
938 if (!smu_dpm->dpm_context) in navi10_allocate_dpm_context()
941 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in navi10_allocate_dpm_context()
963 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in navi10_set_default_dpm_table()
1254 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_emit_clk_levels() local
1255 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in navi10_emit_clk_levels()
1460 smu->smu_dpm.dpm_context; in navi10_populate_umd_state_clk()
2190 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in navi10_update_pcie_parameters()
H A Dsienna_cichlid_ppt.c922 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in sienna_cichlid_allocate_dpm_context() local
924 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in sienna_cichlid_allocate_dpm_context()
925 if (!smu_dpm->dpm_context) in sienna_cichlid_allocate_dpm_context()
928 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in sienna_cichlid_allocate_dpm_context()
956 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in sienna_cichlid_set_default_dpm_table()
1276 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in sienna_cichlid_emit_clk_levels() local
1277 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in sienna_cichlid_emit_clk_levels()
1450 smu->smu_dpm.dpm_context; in sienna_cichlid_populate_umd_state_clk()
2102 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in sienna_cichlid_update_pcie_parameters()
H A Dvangogh_ppt.c434 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in vangogh_allocate_dpm_context() local
436 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in vangogh_allocate_dpm_context()
437 if (!smu_dpm->dpm_context) in vangogh_allocate_dpm_context()
440 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in vangogh_allocate_dpm_context()
573 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_emit_legacy_clk_levels()
2032 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_od_edit_dpm_table()
/linux/drivers/gpu/drm/amd/pm/swsmu/smu15/
H A Dsmu_v15_0.c466 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v15_0_fini_smc_tables() local
490 kfree(smu_dpm->dpm_context); in smu_v15_0_fini_smc_tables()
491 kfree(smu_dpm->golden_dpm_context); in smu_v15_0_fini_smc_tables()
492 kfree(smu_dpm->dpm_current_power_state); in smu_v15_0_fini_smc_tables()
493 kfree(smu_dpm->dpm_request_power_state); in smu_v15_0_fini_smc_tables()
494 smu_dpm->dpm_context = NULL; in smu_v15_0_fini_smc_tables()
495 smu_dpm->golden_dpm_context = NULL; in smu_v15_0_fini_smc_tables()
496 smu_dpm->dpm_context_size = 0; in smu_v15_0_fini_smc_tables()
497 smu_dpm->dpm_current_power_state = NULL; in smu_v15_0_fini_smc_tables()
498 smu_dpm->dpm_request_power_state = NULL; in smu_v15_0_fini_smc_tables()
[all …]
/linux/drivers/gpu/drm/amd/pm/swsmu/smu14/
H A Dsmu_v14_0.c497 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_fini_smc_tables() local
521 kfree(smu_dpm->dpm_context); in smu_v14_0_fini_smc_tables()
522 kfree(smu_dpm->golden_dpm_context); in smu_v14_0_fini_smc_tables()
523 kfree(smu_dpm->dpm_current_power_state); in smu_v14_0_fini_smc_tables()
524 kfree(smu_dpm->dpm_request_power_state); in smu_v14_0_fini_smc_tables()
525 smu_dpm->dpm_context = NULL; in smu_v14_0_fini_smc_tables()
526 smu_dpm->golden_dpm_context = NULL; in smu_v14_0_fini_smc_tables()
527 smu_dpm->dpm_context_size = 0; in smu_v14_0_fini_smc_tables()
528 smu_dpm->dpm_current_power_state = NULL; in smu_v14_0_fini_smc_tables()
529 smu_dpm->dpm_request_power_state = NULL; in smu_v14_0_fini_smc_tables()
[all …]
H A Dsmu_v14_0_2_ppt.c433 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_allocate_dpm_context()
435 smu_dpm->dpm_context = kzalloc_obj(struct smu_14_0_dpm_context); in smu_v14_0_2_allocate_dpm_context()
436 if (!smu_dpm->dpm_context) in smu_v14_0_2_allocate_dpm_context()
439 smu_dpm->dpm_context_size = sizeof(struct smu_14_0_dpm_context);
461 struct smu_14_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v14_0_2_set_default_dpm_table()
750 smu->smu_dpm.dpm_context; in smu_v14_0_2_get_dpm_ultimate_freq()
1033 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_emit_clk_levels()
1034 struct smu_14_0_dpm_context *dpm_context = smu_dpm in smu_v14_0_2_emit_clk_levels()
429 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v14_0_2_allocate_dpm_context() local
1002 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v14_0_2_emit_clk_levels() local
1276 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v14_0_2_force_clk_levels() local
[all...]
/linux/drivers/gpu/drm/amd/pm/swsmu/
H A Damdgpu_smu.c500 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_restore_dpm_user_profile()
950 smu->smu_dpm.dpm_level, in smu_late_init()
1292 struct smu_dpm_context *dpm_ctxt = &(smu->smu_dpm); in smu_init_xgmi_plpd_mode()
1375 smu->smu_dpm.dpm_level = AMD_DPM_FORCED_LEVEL_AUTO; in smu_sw_init()
1376 smu->smu_dpm.requested_dpm_level = AMD_DPM_FORCED_LEVEL_AUTO; in smu_sw_init()
2312 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_enable_umd_pstate()
2385 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_adjust_power_state_dynamic()
2466 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_handle_dpm_task() local
2468 return smu_handle_task(smu, smu_dpm->dpm_level, task_id); in smu_handle_dpm_task()
2477 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_switch_power_profile()
[all …]
/linux/drivers/gpu/drm/amd/pm/swsmu/smu12/
H A Drenoir_ppt.c381 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_od_edit_dpm_table()
610 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_get_current_power_state()
/linux/drivers/gpu/drm/amd/pm/swsmu/inc/
H A Damdgpu_smu.h705 struct smu_dpm_context smu_dpm; member