Commit bc0fcffd authored by Likun Gao's avatar Likun Gao Committed by Alex Deucher

drm/amd/powerplay: Unify smu handle task function (v2)

Unify power stade adjust function into smu_handle_task by the judgment
of task_id.
Move functions which have no relationship with smu version into the file
of amdgpu_smu.
Modified the function of smu_display_config_changed into two part.
Unify some similiar function.

v2: Correct the operation of upload dpm level when force dpm limit value.
Signed-off-by: default avatarLikun Gao <Likun.Gao@amd.com>
Reviewed-by: default avatarHuang Rui <ray.huang@amd.com>
Acked-by: default avatarAlex Deucher <alexander.deucher@amd.com>
Signed-off-by: default avatarAlex Deucher <alexander.deucher@amd.com>
parent 72e91f37
...@@ -2608,6 +2608,15 @@ void amdgpu_pm_compute_clocks(struct amdgpu_device *adev) ...@@ -2608,6 +2608,15 @@ void amdgpu_pm_compute_clocks(struct amdgpu_device *adev)
amdgpu_fence_wait_empty(ring); amdgpu_fence_wait_empty(ring);
} }
if (is_support_sw_smu(adev)) {
struct smu_context *smu = &adev->smu;
struct smu_dpm_context *smu_dpm = &adev->smu.smu_dpm;
mutex_lock(&(smu->mutex));
smu_handle_task(&adev->smu,
smu_dpm->dpm_level,
AMD_PP_TASK_DISPLAY_CONFIG_CHANGE);
mutex_unlock(&(smu->mutex));
} else {
if (adev->powerplay.pp_funcs->dispatch_tasks) { if (adev->powerplay.pp_funcs->dispatch_tasks) {
if (!amdgpu_device_has_dc_support(adev)) { if (!amdgpu_device_has_dc_support(adev)) {
mutex_lock(&adev->pm.mutex); mutex_lock(&adev->pm.mutex);
...@@ -2631,6 +2640,7 @@ void amdgpu_pm_compute_clocks(struct amdgpu_device *adev) ...@@ -2631,6 +2640,7 @@ void amdgpu_pm_compute_clocks(struct amdgpu_device *adev)
amdgpu_dpm_change_power_state_locked(adev); amdgpu_dpm_change_power_state_locked(adev);
mutex_unlock(&adev->pm.mutex); mutex_unlock(&adev->pm.mutex);
} }
}
} }
/* /*
......
...@@ -1041,6 +1041,133 @@ static int smu_enable_umd_pstate(void *handle, ...@@ -1041,6 +1041,133 @@ static int smu_enable_umd_pstate(void *handle,
return 0; return 0;
} }
int smu_unforce_dpm_levels(struct smu_context *smu)
{
int ret = 0;
ret = smu_upload_dpm_level(smu, false);
if (ret) {
pr_err("Failed to upload DPM Bootup Levels!");
return ret;
}
ret = smu_upload_dpm_level(smu, true);
if (ret) {
pr_err("Failed to upload DPM Max Levels!");
return ret;
}
return ret;
}
int smu_adjust_power_state_dynamic(struct smu_context *smu,
enum amd_dpm_forced_level level,
bool skip_display_settings)
{
int ret = 0;
int index = 0;
uint32_t sclk_mask, mclk_mask, soc_mask;
long workload;
struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
if (!skip_display_settings) {
ret = smu_display_config_changed(smu);
if (ret) {
pr_err("Failed to change display config!");
return ret;
}
}
ret = smu_apply_clocks_adjust_rules(smu);
if (ret) {
pr_err("Failed to apply clocks adjust rules!");
return ret;
}
if (!skip_display_settings) {
ret = smu_notify_smc_dispaly_config(smu);
if (ret) {
pr_err("Failed to notify smc display config!");
return ret;
}
}
if (smu_dpm_ctx->dpm_level != level) {
switch (level) {
case AMD_DPM_FORCED_LEVEL_HIGH:
ret = smu_force_dpm_limit_value(smu, true);
break;
case AMD_DPM_FORCED_LEVEL_LOW:
ret = smu_force_dpm_limit_value(smu, false);
break;
case AMD_DPM_FORCED_LEVEL_AUTO:
ret = smu_unforce_dpm_levels(smu);
break;
case AMD_DPM_FORCED_LEVEL_PROFILE_STANDARD:
case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_SCLK:
case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_MCLK:
case AMD_DPM_FORCED_LEVEL_PROFILE_PEAK:
ret = smu_get_profiling_clk_mask(smu, level,
&sclk_mask,
&mclk_mask,
&soc_mask);
if (ret)
return ret;
smu_force_clk_levels(smu, PP_SCLK, 1 << sclk_mask);
smu_force_clk_levels(smu, PP_MCLK, 1 << mclk_mask);
break;
case AMD_DPM_FORCED_LEVEL_MANUAL:
case AMD_DPM_FORCED_LEVEL_PROFILE_EXIT:
default:
break;
}
if (!ret)
smu_dpm_ctx->dpm_level = level;
}
if (smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) {
index = fls(smu->workload_mask);
index = index > 0 && index <= WORKLOAD_POLICY_MAX ? index - 1 : 0;
workload = smu->workload_setting[index];
if (smu->power_profile_mode != workload)
smu_set_power_profile_mode(smu, &workload, 0);
}
return ret;
}
int smu_handle_task(struct smu_context *smu,
enum amd_dpm_forced_level level,
enum amd_pp_task task_id)
{
int ret = 0;
switch (task_id) {
case AMD_PP_TASK_DISPLAY_CONFIG_CHANGE:
ret = smu_pre_display_config_changed(smu);
if (ret)
return ret;
ret = smu_set_cpu_power_state(smu);
if (ret)
return ret;
ret = smu_adjust_power_state_dynamic(smu, level, false);
break;
case AMD_PP_TASK_COMPLETE_INIT:
case AMD_PP_TASK_READJUST_POWER_STATE:
ret = smu_adjust_power_state_dynamic(smu, level, true);
break;
default:
break;
}
return ret;
}
const struct amd_ip_funcs smu_ip_funcs = { const struct amd_ip_funcs smu_ip_funcs = {
.name = "smu", .name = "smu",
.early_init = smu_early_init, .early_init = smu_early_init,
......
...@@ -437,6 +437,18 @@ struct pptable_funcs { ...@@ -437,6 +437,18 @@ struct pptable_funcs {
int (*set_power_profile_mode)(struct smu_context *smu, long *input, uint32_t size); int (*set_power_profile_mode)(struct smu_context *smu, long *input, uint32_t size);
enum amd_dpm_forced_level (*get_performance_level)(struct smu_context *smu); enum amd_dpm_forced_level (*get_performance_level)(struct smu_context *smu);
int (*force_performance_level)(struct smu_context *smu, enum amd_dpm_forced_level level); int (*force_performance_level)(struct smu_context *smu, enum amd_dpm_forced_level level);
int (*pre_display_config_changed)(struct smu_context *smu);
int (*display_config_changed)(struct smu_context *smu);
int (*apply_clocks_adjust_rules)(struct smu_context *smu);
int (*notify_smc_dispaly_config)(struct smu_context *smu);
int (*force_dpm_limit_value)(struct smu_context *smu, bool highest);
int (*upload_dpm_level)(struct smu_context *smu, bool max);
int (*get_profiling_clk_mask)(struct smu_context *smu,
enum amd_dpm_forced_level level,
uint32_t *sclk_mask,
uint32_t *mclk_mask,
uint32_t *soc_mask);
int (*set_cpu_power_state)(struct smu_context *smu);
}; };
struct smu_funcs struct smu_funcs
...@@ -628,6 +640,22 @@ struct smu_funcs ...@@ -628,6 +640,22 @@ struct smu_funcs
((smu)->ppt_funcs->get_performance_level ? (smu)->ppt_funcs->get_performance_level((smu)) : 0) ((smu)->ppt_funcs->get_performance_level ? (smu)->ppt_funcs->get_performance_level((smu)) : 0)
#define smu_force_performance_level(smu, level) \ #define smu_force_performance_level(smu, level) \
((smu)->ppt_funcs->force_performance_level ? (smu)->ppt_funcs->force_performance_level((smu), (level)) : 0) ((smu)->ppt_funcs->force_performance_level ? (smu)->ppt_funcs->force_performance_level((smu), (level)) : 0)
#define smu_pre_display_config_changed(smu) \
((smu)->ppt_funcs->pre_display_config_changed ? (smu)->ppt_funcs->pre_display_config_changed((smu)) : 0)
#define smu_display_config_changed(smu) \
((smu)->ppt_funcs->display_config_changed ? (smu)->ppt_funcs->display_config_changed((smu)) : 0)
#define smu_apply_clocks_adjust_rules(smu) \
((smu)->ppt_funcs->apply_clocks_adjust_rules ? (smu)->ppt_funcs->apply_clocks_adjust_rules((smu)) : 0)
#define smu_notify_smc_dispaly_config(smu) \
((smu)->ppt_funcs->notify_smc_dispaly_config ? (smu)->ppt_funcs->notify_smc_dispaly_config((smu)) : 0)
#define smu_force_dpm_limit_value(smu, highest) \
((smu)->ppt_funcs->force_dpm_limit_value ? (smu)->ppt_funcs->force_dpm_limit_value((smu), (highest)) : 0)
#define smu_upload_dpm_level(smu, max) \
((smu)->ppt_funcs->upload_dpm_level ? (smu)->ppt_funcs->upload_dpm_level((smu), (max)) : 0)
#define smu_get_profiling_clk_mask(smu, level, sclk_mask, mclk_mask, soc_mask) \
((smu)->ppt_funcs->get_profiling_clk_mask ? (smu)->ppt_funcs->get_profiling_clk_mask((smu), (level), (sclk_mask), (mclk_mask), (soc_mask)) : 0)
#define smu_set_cpu_power_state(smu) \
((smu)->ppt_funcs->set_cpu_power_state ? (smu)->ppt_funcs->set_cpu_power_state((smu)) : 0)
#define smu_msg_get_index(smu, msg) \ #define smu_msg_get_index(smu, msg) \
((smu)->ppt_funcs? ((smu)->ppt_funcs->get_smu_msg_index? (smu)->ppt_funcs->get_smu_msg_index((smu), (msg)) : -EINVAL) : -EINVAL) ((smu)->ppt_funcs? ((smu)->ppt_funcs->get_smu_msg_index? (smu)->ppt_funcs->get_smu_msg_index((smu), (msg)) : -EINVAL) : -EINVAL)
...@@ -699,4 +727,7 @@ extern int smu_display_configuration_change(struct smu_context *smu, const ...@@ -699,4 +727,7 @@ extern int smu_display_configuration_change(struct smu_context *smu, const
extern int smu_get_current_clocks(struct smu_context *smu, extern int smu_get_current_clocks(struct smu_context *smu,
struct amd_pp_clock_info *clocks); struct amd_pp_clock_info *clocks);
extern int smu_dpm_set_power_gate(struct smu_context *smu,uint32_t block_type, bool gate); extern int smu_dpm_set_power_gate(struct smu_context *smu,uint32_t block_type, bool gate);
extern int smu_handle_task(struct smu_context *smu,
enum amd_dpm_forced_level level,
enum amd_pp_task task_id);
#endif #endif
...@@ -877,71 +877,39 @@ static int vega20_print_clk_levels(struct smu_context *smu, ...@@ -877,71 +877,39 @@ static int vega20_print_clk_levels(struct smu_context *smu,
return size; return size;
} }
static int vega20_upload_dpm_min_level(struct smu_context *smu) static int vega20_upload_dpm_level(struct smu_context *smu, bool max)
{ {
struct vega20_dpm_table *dpm_table; struct vega20_dpm_table *dpm_table;
struct vega20_single_dpm_table *single_dpm_table; struct vega20_single_dpm_table *single_dpm_table;
uint32_t min_freq; uint32_t freq;
int ret = 0; int ret = 0;
dpm_table = smu->smu_dpm.dpm_context; dpm_table = smu->smu_dpm.dpm_context;
if (smu_feature_is_enabled(smu, FEATURE_DPM_GFXCLK_BIT)) { if (smu_feature_is_enabled(smu, FEATURE_DPM_GFXCLK_BIT)) {
single_dpm_table = &(dpm_table->gfx_table); single_dpm_table = &(dpm_table->gfx_table);
min_freq = single_dpm_table->dpm_state.soft_min_level; freq = max ? single_dpm_table->dpm_state.soft_max_level :
single_dpm_table->dpm_state.soft_min_level;
ret = smu_send_smc_msg_with_param(smu, ret = smu_send_smc_msg_with_param(smu,
SMU_MSG_SetSoftMinByFreq, (max ? SMU_MSG_SetSoftMaxByFreq : SMU_MSG_SetSoftMinByFreq),
(PPCLK_GFXCLK << 16) | (min_freq & 0xffff)); (PPCLK_GFXCLK << 16) | (freq & 0xffff));
if (ret) { if (ret) {
pr_err("Failed to set soft min gfxclk !\n"); pr_err("Failed to set soft %s gfxclk !\n",
max ? "max" : "min");
return ret; return ret;
} }
} }
if (smu_feature_is_enabled(smu, FEATURE_DPM_UCLK_BIT)) { if (smu_feature_is_enabled(smu, FEATURE_DPM_UCLK_BIT)) {
single_dpm_table = &(dpm_table->mem_table); single_dpm_table = &(dpm_table->mem_table);
min_freq = single_dpm_table->dpm_state.soft_min_level; freq = max ? single_dpm_table->dpm_state.soft_max_level :
single_dpm_table->dpm_state.soft_min_level;
ret = smu_send_smc_msg_with_param(smu, ret = smu_send_smc_msg_with_param(smu,
SMU_MSG_SetSoftMinByFreq, (max ? SMU_MSG_SetSoftMaxByFreq : SMU_MSG_SetSoftMinByFreq),
(PPCLK_UCLK << 16) | (min_freq & 0xffff)); (PPCLK_UCLK << 16) | (freq & 0xffff));
if (ret) { if (ret) {
pr_err("Failed to set soft min memclk !\n"); pr_err("Failed to set soft %s memclk !\n",
return ret; max ? "max" : "min");
}
}
return ret;
}
static int vega20_upload_dpm_max_level(struct smu_context *smu)
{
struct vega20_dpm_table *dpm_table;
struct vega20_single_dpm_table *single_dpm_table;
uint32_t max_freq;
int ret = 0;
dpm_table = smu->smu_dpm.dpm_context;
if (smu_feature_is_enabled(smu, FEATURE_DPM_GFXCLK_BIT)) {
single_dpm_table = &(dpm_table->gfx_table);
max_freq = single_dpm_table->dpm_state.soft_max_level;
ret = smu_send_smc_msg_with_param(smu,
SMU_MSG_SetSoftMaxByFreq,
(PPCLK_GFXCLK << 16) | (max_freq & 0xffff));
if (ret) {
pr_err("Failed to set soft max gfxclk !\n");
return ret;
}
}
if (smu_feature_is_enabled(smu, FEATURE_DPM_UCLK_BIT)) {
single_dpm_table = &(dpm_table->mem_table);
max_freq = single_dpm_table->dpm_state.soft_max_level;
ret = smu_send_smc_msg_with_param(smu,
SMU_MSG_SetSoftMaxByFreq,
(PPCLK_UCLK << 16) | (max_freq & 0xffff));
if (ret) {
pr_err("Failed to set soft max memclk !\n");
return ret; return ret;
} }
} }
...@@ -986,13 +954,13 @@ static int vega20_force_clk_levels(struct smu_context *smu, ...@@ -986,13 +954,13 @@ static int vega20_force_clk_levels(struct smu_context *smu,
single_dpm_table->dpm_state.soft_max_level = single_dpm_table->dpm_state.soft_max_level =
single_dpm_table->dpm_levels[soft_max_level].value; single_dpm_table->dpm_levels[soft_max_level].value;
ret = vega20_upload_dpm_min_level(smu); ret = vega20_upload_dpm_level(smu, false);
if (ret) { if (ret) {
pr_err("Failed to upload boot level to lowest!\n"); pr_err("Failed to upload boot level to lowest!\n");
break; break;
} }
ret = vega20_upload_dpm_max_level(smu); ret = vega20_upload_dpm_level(smu, true);
if (ret) if (ret)
pr_err("Failed to upload dpm max level to highest!\n"); pr_err("Failed to upload dpm max level to highest!\n");
...@@ -1013,13 +981,13 @@ static int vega20_force_clk_levels(struct smu_context *smu, ...@@ -1013,13 +981,13 @@ static int vega20_force_clk_levels(struct smu_context *smu,
single_dpm_table->dpm_state.soft_max_level = single_dpm_table->dpm_state.soft_max_level =
single_dpm_table->dpm_levels[soft_max_level].value; single_dpm_table->dpm_levels[soft_max_level].value;
ret = vega20_upload_dpm_min_level(smu); ret = vega20_upload_dpm_level(smu, false);
if (ret) { if (ret) {
pr_err("Failed to upload boot level to lowest!\n"); pr_err("Failed to upload boot level to lowest!\n");
break; break;
} }
ret = vega20_upload_dpm_max_level(smu); ret = vega20_upload_dpm_level(smu, true);
if (ret) if (ret)
pr_err("Failed to upload dpm max level to highest!\n"); pr_err("Failed to upload dpm max level to highest!\n");
...@@ -1389,26 +1357,33 @@ vega20_set_uclk_to_highest_dpm_level(struct smu_context *smu, ...@@ -1389,26 +1357,33 @@ vega20_set_uclk_to_highest_dpm_level(struct smu_context *smu,
return ret; return ret;
} }
static int vega20_display_config_changed(struct smu_context *smu) static int vega20_pre_display_config_changed(struct smu_context *smu)
{ {
int ret = 0; int ret = 0;
struct vega20_dpm_table *dpm_table = smu->smu_dpm.dpm_context; struct vega20_dpm_table *dpm_table = smu->smu_dpm.dpm_context;
if (!smu->funcs) if (!smu->smu_dpm.dpm_context)
return -EINVAL;
if (!smu->smu_dpm.dpm_context ||
!smu->smu_table.tables ||
!smu->smu_table.tables[TABLE_WATERMARKS].cpu_addr)
return -EINVAL; return -EINVAL;
smu_send_smc_msg_with_param(smu, SMU_MSG_NumOfDisplays, 0); smu_send_smc_msg_with_param(smu, SMU_MSG_NumOfDisplays, 0);
ret = vega20_set_uclk_to_highest_dpm_level(smu, ret = vega20_set_uclk_to_highest_dpm_level(smu,
&dpm_table->mem_table); &dpm_table->mem_table);
if (ret) { if (ret)
pr_err("Failed to set uclk to highest dpm level"); pr_err("Failed to set uclk to highest dpm level");
return ret; return ret;
} }
static int vega20_display_config_changed(struct smu_context *smu)
{
int ret = 0;
if (!smu->funcs)
return -EINVAL;
if (!smu->smu_dpm.dpm_context ||
!smu->smu_table.tables ||
!smu->smu_table.tables[TABLE_WATERMARKS].cpu_addr)
return -EINVAL;
if ((smu->watermarks_bitmap & WATERMARKS_EXIST) && if ((smu->watermarks_bitmap & WATERMARKS_EXIST) &&
!(smu->watermarks_bitmap & WATERMARKS_LOADED)) { !(smu->watermarks_bitmap & WATERMARKS_LOADED)) {
...@@ -1672,85 +1647,42 @@ static uint32_t vega20_find_highest_dpm_level(struct vega20_single_dpm_table *ta ...@@ -1672,85 +1647,42 @@ static uint32_t vega20_find_highest_dpm_level(struct vega20_single_dpm_table *ta
return i; return i;
} }
static int vega20_force_dpm_highest(struct smu_context *smu) static int vega20_force_dpm_limit_value(struct smu_context *smu, bool highest)
{ {
uint32_t soft_level; uint32_t soft_level;
int ret = 0; int ret = 0;
struct vega20_dpm_table *dpm_table = (struct vega20_dpm_table *)smu->smu_dpm.dpm_context; struct vega20_dpm_table *dpm_table =
(struct vega20_dpm_table *)smu->smu_dpm.dpm_context;
if (highest)
soft_level = vega20_find_highest_dpm_level(&(dpm_table->gfx_table)); soft_level = vega20_find_highest_dpm_level(&(dpm_table->gfx_table));
else
dpm_table->gfx_table.dpm_state.soft_min_level =
dpm_table->gfx_table.dpm_state.soft_max_level =
dpm_table->gfx_table.dpm_levels[soft_level].value;
soft_level = vega20_find_highest_dpm_level(&(dpm_table->mem_table));
dpm_table->mem_table.dpm_state.soft_min_level =
dpm_table->mem_table.dpm_state.soft_max_level =
dpm_table->mem_table.dpm_levels[soft_level].value;
ret = vega20_upload_dpm_min_level(smu);
if (ret) {
pr_err("Failed to upload boot level to highest!");
return ret;
}
ret = vega20_upload_dpm_max_level(smu);
if (ret) {
pr_err("Failed to upload dpm max level to highest!");
return ret;
}
return ret;
}
static int vega20_force_dpm_lowest(struct smu_context *smu)
{
uint32_t soft_level;
int ret = 0;
struct vega20_dpm_table *dpm_table = (struct vega20_dpm_table *)smu->smu_dpm.dpm_context;
soft_level = vega20_find_lowest_dpm_level(&(dpm_table->gfx_table)); soft_level = vega20_find_lowest_dpm_level(&(dpm_table->gfx_table));
dpm_table->gfx_table.dpm_state.soft_min_level = dpm_table->gfx_table.dpm_state.soft_min_level =
dpm_table->gfx_table.dpm_state.soft_max_level = dpm_table->gfx_table.dpm_state.soft_max_level =
dpm_table->gfx_table.dpm_levels[soft_level].value; dpm_table->gfx_table.dpm_levels[soft_level].value;
if (highest)
soft_level = vega20_find_highest_dpm_level(&(dpm_table->mem_table));
else
soft_level = vega20_find_lowest_dpm_level(&(dpm_table->mem_table)); soft_level = vega20_find_lowest_dpm_level(&(dpm_table->mem_table));
dpm_table->mem_table.dpm_state.soft_min_level = dpm_table->mem_table.dpm_state.soft_min_level =
dpm_table->mem_table.dpm_state.soft_max_level = dpm_table->mem_table.dpm_state.soft_max_level =
dpm_table->mem_table.dpm_levels[soft_level].value; dpm_table->mem_table.dpm_levels[soft_level].value;
ret = vega20_upload_dpm_min_level(smu); ret = vega20_upload_dpm_level(smu, false);
if (ret) { if (ret) {
pr_err("Failed to upload boot level to lowest!"); pr_err("Failed to upload boot level to %s!\n",
highest ? "highest" : "lowest");
return ret; return ret;
} }
ret = vega20_upload_dpm_max_level(smu); ret = vega20_upload_dpm_level(smu, true);
if (ret) { if (ret) {
pr_err("Failed to upload dpm max level to lowest!"); pr_err("Failed to upload dpm max level to %s!\n!",
return ret; highest ? "highest" : "lowest");
}
return ret;
}
static int vega20_unforce_dpm_levels(struct smu_context *smu)
{
int ret = 0;
ret = vega20_upload_dpm_min_level(smu);
if (ret) {
pr_err("Failed to upload DPM Bootup Levels!");
return ret;
}
ret = vega20_upload_dpm_max_level(smu);
if (ret) {
pr_err("Failed to upload DPM Max Levels!");
return ret; return ret;
} }
...@@ -1771,78 +1703,6 @@ static enum amd_dpm_forced_level vega20_get_performance_level(struct smu_context ...@@ -1771,78 +1703,6 @@ static enum amd_dpm_forced_level vega20_get_performance_level(struct smu_context
return smu_dpm_ctx->dpm_level; return smu_dpm_ctx->dpm_level;
} }
static int vega20_adjust_power_state_dynamic(struct smu_context *smu,
enum amd_dpm_forced_level level)
{
int ret = 0;
int index = 0;
uint32_t sclk_mask, mclk_mask, soc_mask;
long workload;
struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
ret = vega20_display_config_changed(smu);
if (ret) {
pr_err("Failed to change display config!");
return ret;
}
ret = vega20_apply_clocks_adjust_rules(smu);
if (ret) {
pr_err("Failed to apply clocks adjust rules!");
return ret;
}
ret = vega20_notify_smc_dispaly_config(smu);
if (ret) {
pr_err("Failed to notify smc display config!");
return ret;
}
switch (level) {
case AMD_DPM_FORCED_LEVEL_HIGH:
ret = vega20_force_dpm_highest(smu);
break;
case AMD_DPM_FORCED_LEVEL_LOW:
ret = vega20_force_dpm_lowest(smu);
break;
case AMD_DPM_FORCED_LEVEL_AUTO:
ret = vega20_unforce_dpm_levels(smu);
break;
case AMD_DPM_FORCED_LEVEL_PROFILE_STANDARD:
case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_SCLK:
case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_MCLK:
case AMD_DPM_FORCED_LEVEL_PROFILE_PEAK:
ret = vega20_get_profiling_clk_mask(smu, level,
&sclk_mask,
&mclk_mask,
&soc_mask);
if (ret)
return ret;
vega20_force_clk_levels(smu, PP_SCLK, 1 << sclk_mask);
vega20_force_clk_levels(smu, PP_MCLK, 1 << mclk_mask);
break;
case AMD_DPM_FORCED_LEVEL_MANUAL:
case AMD_DPM_FORCED_LEVEL_PROFILE_EXIT:
default:
break;
}
if (!ret)
smu_dpm_ctx->dpm_level = level;
if (smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) {
index = fls(smu->workload_mask);
index = index > 0 && index <= WORKLOAD_POLICY_MAX ? index - 1 : 0;
workload = smu->workload_setting[index];
if (smu->power_profile_mode != workload)
smu->funcs->set_power_profile_mode(smu, &workload, 0);
}
return ret;
}
static int static int
vega20_force_performance_level(struct smu_context *smu, enum amd_dpm_forced_level level) vega20_force_performance_level(struct smu_context *smu, enum amd_dpm_forced_level level)
{ {
...@@ -1861,7 +1721,8 @@ vega20_force_performance_level(struct smu_context *smu, enum amd_dpm_forced_leve ...@@ -1861,7 +1721,8 @@ vega20_force_performance_level(struct smu_context *smu, enum amd_dpm_forced_leve
mutex_lock(&smu->mutex); mutex_lock(&smu->mutex);
smu->adev->ip_blocks[i].version->funcs->enable_umd_pstate(smu, &level); smu->adev->ip_blocks[i].version->funcs->enable_umd_pstate(smu, &level);
ret = vega20_adjust_power_state_dynamic(smu, level); ret = smu_handle_task(smu, level,
AMD_PP_TASK_READJUST_POWER_STATE);
mutex_unlock(&smu->mutex); mutex_unlock(&smu->mutex);
...@@ -2009,7 +1870,8 @@ static int vega20_set_od_percentage(struct smu_context *smu, ...@@ -2009,7 +1870,8 @@ static int vega20_set_od_percentage(struct smu_context *smu,
single_dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.gfxclk / 100; single_dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.gfxclk / 100;
} }
ret = vega20_adjust_power_state_dynamic(smu, smu_dpm->dpm_level); ret = smu_handle_task(smu, smu_dpm->dpm_level,
AMD_PP_TASK_READJUST_POWER_STATE);
set_od_failed: set_od_failed:
mutex_unlock(&(smu->mutex)); mutex_unlock(&(smu->mutex));
...@@ -2239,7 +2101,8 @@ static int vega20_odn_edit_dpm_table(struct smu_context *smu, ...@@ -2239,7 +2101,8 @@ static int vega20_odn_edit_dpm_table(struct smu_context *smu,
if (type == PP_OD_COMMIT_DPM_TABLE) { if (type == PP_OD_COMMIT_DPM_TABLE) {
mutex_lock(&(smu->mutex)); mutex_lock(&(smu->mutex));
ret = vega20_adjust_power_state_dynamic(smu, smu_dpm->dpm_level); ret = smu_handle_task(smu, smu_dpm->dpm_level,
AMD_PP_TASK_READJUST_POWER_STATE);
mutex_unlock(&(smu->mutex)); mutex_unlock(&(smu->mutex));
} }
...@@ -2268,6 +2131,13 @@ static const struct pptable_funcs vega20_ppt_funcs = { ...@@ -2268,6 +2131,13 @@ static const struct pptable_funcs vega20_ppt_funcs = {
.update_specified_od8_value = vega20_update_specified_od8_value, .update_specified_od8_value = vega20_update_specified_od8_value,
.set_od_percentage = vega20_set_od_percentage, .set_od_percentage = vega20_set_od_percentage,
.od_edit_dpm_table = vega20_odn_edit_dpm_table, .od_edit_dpm_table = vega20_odn_edit_dpm_table,
.pre_display_config_changed = vega20_pre_display_config_changed,
.display_config_changed = vega20_display_config_changed,
.apply_clocks_adjust_rules = vega20_apply_clocks_adjust_rules,
.notify_smc_dispaly_config = vega20_notify_smc_dispaly_config,
.force_dpm_limit_value = vega20_force_dpm_limit_value,
.upload_dpm_level = vega20_upload_dpm_level,
.get_profiling_clk_mask = vega20_get_profiling_clk_mask,
}; };
void vega20_set_ppt_funcs(struct smu_context *smu) void vega20_set_ppt_funcs(struct smu_context *smu)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment