| /linux/drivers/gpu/drm/amd/pm/swsmu/smu13/ |
| H A D | aldebaran_ppt.c | 294 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in aldebaran_allocate_dpm_context() local 297 smu_dpm->dpm_context = kzalloc_obj(struct smu_13_0_dpm_context); in aldebaran_allocate_dpm_context() 298 if (!smu_dpm->dpm_context) in aldebaran_allocate_dpm_context() 300 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); in aldebaran_allocate_dpm_context() 302 smu_dpm->dpm_policies = in aldebaran_allocate_dpm_context() 305 if (!smu_dpm->dpm_policies) in aldebaran_allocate_dpm_context() 308 policy = &(smu_dpm->dpm_policies->policies[0]); in aldebaran_allocate_dpm_context() 314 smu_dpm->dpm_policies->policy_mask |= BIT(PP_PM_POLICY_XGMI_PLPD); in aldebaran_allocate_dpm_context() 345 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in aldebaran_get_dpm_ultimate_freq() 398 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in aldebaran_set_default_dpm_table() [all …]
|
| H A D | smu_v13_0_6_ppt.c | 274 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_6_cap_set() 282 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_6_cap_clear() 290 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_6_cap_supported() 689 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_6_allocate_dpm_context() local 692 smu_dpm->dpm_context = in smu_v13_0_6_allocate_dpm_context() 694 if (!smu_dpm->dpm_context) in smu_v13_0_6_allocate_dpm_context() 696 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); in smu_v13_0_6_allocate_dpm_context() 698 smu_dpm->dpm_policies = in smu_v13_0_6_allocate_dpm_context() 700 if (!smu_dpm in smu_v13_0_6_allocate_dpm_context() 1115 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_6_set_default_dpm_table() local 1375 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_6_emit_clk_levels() local 1970 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); smu_v13_0_6_set_performance_level() local 2034 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); smu_v13_0_6_set_soft_freq_limited_range() local 2123 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); smu_v13_0_6_usr_edit_dpm_table() local [all...] |
| H A D | smu_v13_0.c | 508 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_fini_smc_tables() local 534 kfree(smu_dpm->dpm_policies); in smu_v13_0_fini_smc_tables() 535 kfree(smu_dpm->dpm_context); in smu_v13_0_fini_smc_tables() 536 kfree(smu_dpm->golden_dpm_context); in smu_v13_0_fini_smc_tables() 537 kfree(smu_dpm->dpm_current_power_state); in smu_v13_0_fini_smc_tables() 538 kfree(smu_dpm->dpm_request_power_state); in smu_v13_0_fini_smc_tables() 539 smu_dpm->dpm_policies = NULL; in smu_v13_0_fini_smc_tables() 540 smu_dpm->dpm_context = NULL; in smu_v13_0_fini_smc_tables() 541 smu_dpm->golden_dpm_context = NULL; in smu_v13_0_fini_smc_tables() 542 smu_dpm->dpm_context_size = 0; in smu_v13_0_fini_smc_tables() [all …]
|
| H A D | smu_v13_0_7_ppt.c | 564 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_allocate_dpm_context() 566 smu_dpm->dpm_context = kzalloc_obj(struct smu_13_0_dpm_context); in smu_v13_0_7_allocate_dpm_context() 567 if (!smu_dpm->dpm_context) in smu_v13_0_7_allocate_dpm_context() 570 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); 592 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_7_set_default_dpm_table() 874 smu->smu_dpm.dpm_context; in smu_v13_0_7_get_dpm_ultimate_freq() 1214 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_emit_clk_levels() 1215 struct smu_13_0_dpm_context *dpm_context = smu_dpm in smu_v13_0_7_emit_clk_levels() 560 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_7_allocate_dpm_context() local 1183 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_7_emit_clk_levels() local 1924 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_7_force_clk_levels() local [all...] |
| H A D | smu_v13_0_0_ppt.c | 535 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_allocate_dpm_context() 537 smu_dpm->dpm_context = kzalloc_obj(struct smu_13_0_dpm_context); in smu_v13_0_0_allocate_dpm_context() 538 if (!smu_dpm->dpm_context) in smu_v13_0_0_allocate_dpm_context() 541 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); 563 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_0_set_default_dpm_table() 864 smu->smu_dpm.dpm_context; in smu_v13_0_0_get_dpm_ultimate_freq() 1204 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_emit_clk_levels() 1205 struct smu_13_0_dpm_context *dpm_context = smu_dpm in smu_v13_0_0_emit_clk_levels() 531 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_0_allocate_dpm_context() local 1173 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_0_emit_clk_levels() local 1914 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v13_0_0_force_clk_levels() local [all...] |
| H A D | smu_v13_0_5_ppt.c | 522 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in smu_v13_0_5_od_edit_dpm_table() local 526 if (smu_dpm->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) in smu_v13_0_5_od_edit_dpm_table()
|
| H A D | yellow_carp_ppt.c | 653 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in yellow_carp_od_edit_dpm_table() local 657 if (smu_dpm->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) in yellow_carp_od_edit_dpm_table()
|
| H A D | smu_v13_0_12_ppt.c | 283 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_12_setup_driver_pptable()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu11/ |
| H A D | arcturus_ppt.c | 307 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_allocate_dpm_context() local 310 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in arcturus_allocate_dpm_context() 311 if (!smu_dpm->dpm_context) in arcturus_allocate_dpm_context() 313 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in arcturus_allocate_dpm_context() 315 smu_dpm->dpm_policies = in arcturus_allocate_dpm_context() 318 if (!smu_dpm->dpm_policies) in arcturus_allocate_dpm_context() 321 policy = &(smu_dpm->dpm_policies->policies[0]); in arcturus_allocate_dpm_context() 327 smu_dpm->dpm_policies->policy_mask |= BIT(PP_PM_POLICY_XGMI_PLPD); in arcturus_allocate_dpm_context() 357 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in arcturus_set_default_dpm_table() 432 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_set_default_dpm_table() local [all …]
|
| H A D | smu_v11_0.c | 429 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v11_0_fini_smc_tables() local 457 kfree(smu_dpm->dpm_context); in smu_v11_0_fini_smc_tables() 458 kfree(smu_dpm->golden_dpm_context); in smu_v11_0_fini_smc_tables() 459 kfree(smu_dpm->dpm_current_power_state); in smu_v11_0_fini_smc_tables() 460 kfree(smu_dpm->dpm_request_power_state); in smu_v11_0_fini_smc_tables() 461 smu_dpm->dpm_context = NULL; in smu_v11_0_fini_smc_tables() 462 smu_dpm->golden_dpm_context = NULL; in smu_v11_0_fini_smc_tables() 463 smu_dpm->dpm_context_size = 0; in smu_v11_0_fini_smc_tables() 464 smu_dpm->dpm_current_power_state = NULL; in smu_v11_0_fini_smc_tables() 465 smu_dpm->dpm_request_power_state = NULL; in smu_v11_0_fini_smc_tables() [all …]
|
| H A D | navi10_ppt.c | 935 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_allocate_dpm_context() local 937 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in navi10_allocate_dpm_context() 938 if (!smu_dpm->dpm_context) in navi10_allocate_dpm_context() 941 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in navi10_allocate_dpm_context() 963 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in navi10_set_default_dpm_table() 1254 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_emit_clk_levels() local 1255 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in navi10_emit_clk_levels() 1460 smu->smu_dpm.dpm_context; in navi10_populate_umd_state_clk() 2190 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in navi10_update_pcie_parameters()
|
| H A D | sienna_cichlid_ppt.c | 922 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in sienna_cichlid_allocate_dpm_context() local 924 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in sienna_cichlid_allocate_dpm_context() 925 if (!smu_dpm->dpm_context) in sienna_cichlid_allocate_dpm_context() 928 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in sienna_cichlid_allocate_dpm_context() 956 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in sienna_cichlid_set_default_dpm_table() 1276 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in sienna_cichlid_emit_clk_levels() local 1277 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in sienna_cichlid_emit_clk_levels() 1450 smu->smu_dpm.dpm_context; in sienna_cichlid_populate_umd_state_clk() 2102 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in sienna_cichlid_update_pcie_parameters()
|
| H A D | vangogh_ppt.c | 434 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in vangogh_allocate_dpm_context() local 436 smu_dpm->dpm_context = kzalloc_obj(struct smu_11_0_dpm_context); in vangogh_allocate_dpm_context() 437 if (!smu_dpm->dpm_context) in vangogh_allocate_dpm_context() 440 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in vangogh_allocate_dpm_context() 573 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_emit_legacy_clk_levels() 2032 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_od_edit_dpm_table()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu15/ |
| H A D | smu_v15_0.c | 466 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v15_0_fini_smc_tables() local 490 kfree(smu_dpm->dpm_context); in smu_v15_0_fini_smc_tables() 491 kfree(smu_dpm->golden_dpm_context); in smu_v15_0_fini_smc_tables() 492 kfree(smu_dpm->dpm_current_power_state); in smu_v15_0_fini_smc_tables() 493 kfree(smu_dpm->dpm_request_power_state); in smu_v15_0_fini_smc_tables() 494 smu_dpm->dpm_context = NULL; in smu_v15_0_fini_smc_tables() 495 smu_dpm->golden_dpm_context = NULL; in smu_v15_0_fini_smc_tables() 496 smu_dpm->dpm_context_size = 0; in smu_v15_0_fini_smc_tables() 497 smu_dpm->dpm_current_power_state = NULL; in smu_v15_0_fini_smc_tables() 498 smu_dpm->dpm_request_power_state = NULL; in smu_v15_0_fini_smc_tables() [all …]
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu14/ |
| H A D | smu_v14_0.c | 497 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_fini_smc_tables() local 521 kfree(smu_dpm->dpm_context); in smu_v14_0_fini_smc_tables() 522 kfree(smu_dpm->golden_dpm_context); in smu_v14_0_fini_smc_tables() 523 kfree(smu_dpm->dpm_current_power_state); in smu_v14_0_fini_smc_tables() 524 kfree(smu_dpm->dpm_request_power_state); in smu_v14_0_fini_smc_tables() 525 smu_dpm->dpm_context = NULL; in smu_v14_0_fini_smc_tables() 526 smu_dpm->golden_dpm_context = NULL; in smu_v14_0_fini_smc_tables() 527 smu_dpm->dpm_context_size = 0; in smu_v14_0_fini_smc_tables() 528 smu_dpm->dpm_current_power_state = NULL; in smu_v14_0_fini_smc_tables() 529 smu_dpm->dpm_request_power_state = NULL; in smu_v14_0_fini_smc_tables() [all …]
|
| H A D | smu_v14_0_2_ppt.c | 433 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_allocate_dpm_context() 435 smu_dpm->dpm_context = kzalloc_obj(struct smu_14_0_dpm_context); in smu_v14_0_2_allocate_dpm_context() 436 if (!smu_dpm->dpm_context) in smu_v14_0_2_allocate_dpm_context() 439 smu_dpm->dpm_context_size = sizeof(struct smu_14_0_dpm_context); 461 struct smu_14_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v14_0_2_set_default_dpm_table() 750 smu->smu_dpm.dpm_context; in smu_v14_0_2_get_dpm_ultimate_freq() 1033 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_emit_clk_levels() 1034 struct smu_14_0_dpm_context *dpm_context = smu_dpm in smu_v14_0_2_emit_clk_levels() 429 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v14_0_2_allocate_dpm_context() local 1002 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v14_0_2_emit_clk_levels() local 1276 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; smu_v14_0_2_force_clk_levels() local [all...] |
| /linux/drivers/gpu/drm/amd/pm/swsmu/ |
| H A D | amdgpu_smu.c | 500 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_restore_dpm_user_profile() 950 smu->smu_dpm.dpm_level, in smu_late_init() 1292 struct smu_dpm_context *dpm_ctxt = &(smu->smu_dpm); in smu_init_xgmi_plpd_mode() 1375 smu->smu_dpm.dpm_level = AMD_DPM_FORCED_LEVEL_AUTO; in smu_sw_init() 1376 smu->smu_dpm.requested_dpm_level = AMD_DPM_FORCED_LEVEL_AUTO; in smu_sw_init() 2312 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_enable_umd_pstate() 2385 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_adjust_power_state_dynamic() 2466 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_handle_dpm_task() local 2468 return smu_handle_task(smu, smu_dpm->dpm_level, task_id); in smu_handle_dpm_task() 2477 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_switch_power_profile() [all …]
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu12/ |
| H A D | renoir_ppt.c | 381 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_od_edit_dpm_table() 610 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_get_current_power_state()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/inc/ |
| H A D | amdgpu_smu.h | 705 struct smu_dpm_context smu_dpm; member
|