Commit f3a05239 authored by Meng Li's avatar Meng Li Committed by Rafael J. Wysocki

cpufreq: amd-pstate: Enable amd-pstate preferred core support

amd-pstate driver utilizes the functions and data structures
provided by the ITMT architecture to enable the scheduler to
favor scheduling on cores which can be get a higher frequency
with lower voltage. We call it amd-pstate preferrred core.

Here sched_set_itmt_core_prio() is called to set priorities and
sched_set_itmt_support() is called to enable ITMT feature.
amd-pstate driver uses the highest performance value to indicate
the priority of CPU. The higher value has a higher priority.

The initial core rankings are set up by amd-pstate when the
system boots.

Add a variable hw_prefcore in cpudata structure. It will check
if the processor and power firmware support preferred core
feature.

Add one new early parameter `disable` to allow user to disable
the preferred core.

Only when hardware supports preferred core and user set `enabled`
in early parameter, amd pstate driver supports preferred core featue.
Tested-by: default avatarOleksandr Natalenko <oleksandr@natalenko.name>
Reviewed-by: default avatarHuang Rui <ray.huang@amd.com>
Reviewed-by: default avatarWyes Karny <wyes.karny@amd.com>
Reviewed-by: default avatarMario Limonciello <mario.limonciello@amd.com>
Co-developed-by: default avatarPerry Yuan <Perry.Yuan@amd.com>
Signed-off-by: default avatarPerry Yuan <Perry.Yuan@amd.com>
Signed-off-by: default avatarMeng Li <li.meng@amd.com>
Signed-off-by: default avatarRafael J. Wysocki <rafael.j.wysocki@intel.com>
parent 12753d71
...@@ -37,6 +37,7 @@ ...@@ -37,6 +37,7 @@
#include <linux/uaccess.h> #include <linux/uaccess.h>
#include <linux/static_call.h> #include <linux/static_call.h>
#include <linux/amd-pstate.h> #include <linux/amd-pstate.h>
#include <linux/topology.h>
#include <acpi/processor.h> #include <acpi/processor.h>
#include <acpi/cppc_acpi.h> #include <acpi/cppc_acpi.h>
...@@ -49,6 +50,7 @@ ...@@ -49,6 +50,7 @@
#define AMD_PSTATE_TRANSITION_LATENCY 20000 #define AMD_PSTATE_TRANSITION_LATENCY 20000
#define AMD_PSTATE_TRANSITION_DELAY 1000 #define AMD_PSTATE_TRANSITION_DELAY 1000
#define AMD_PSTATE_PREFCORE_THRESHOLD 166
/* /*
* TODO: We need more time to fine tune processors with shared memory solution * TODO: We need more time to fine tune processors with shared memory solution
...@@ -64,6 +66,7 @@ static struct cpufreq_driver amd_pstate_driver; ...@@ -64,6 +66,7 @@ static struct cpufreq_driver amd_pstate_driver;
static struct cpufreq_driver amd_pstate_epp_driver; static struct cpufreq_driver amd_pstate_epp_driver;
static int cppc_state = AMD_PSTATE_UNDEFINED; static int cppc_state = AMD_PSTATE_UNDEFINED;
static bool cppc_enabled; static bool cppc_enabled;
static bool amd_pstate_prefcore = true;
/* /*
* AMD Energy Preference Performance (EPP) * AMD Energy Preference Performance (EPP)
...@@ -297,13 +300,14 @@ static int pstate_init_perf(struct amd_cpudata *cpudata) ...@@ -297,13 +300,14 @@ static int pstate_init_perf(struct amd_cpudata *cpudata)
if (ret) if (ret)
return ret; return ret;
/* /* For platforms that do not support the preferred core feature, the
* TODO: Introduce AMD specific power feature. * highest_pef may be configured with 166 or 255, to avoid max frequency
* * calculated wrongly. we take the AMD_CPPC_HIGHEST_PERF(cap1) value as
* CPPC entry doesn't indicate the highest performance in some ASICs. * the default max perf.
*/ */
highest_perf = amd_get_highest_perf(); if (cpudata->hw_prefcore)
if (highest_perf > AMD_CPPC_HIGHEST_PERF(cap1)) highest_perf = AMD_PSTATE_PREFCORE_THRESHOLD;
else
highest_perf = AMD_CPPC_HIGHEST_PERF(cap1); highest_perf = AMD_CPPC_HIGHEST_PERF(cap1);
WRITE_ONCE(cpudata->highest_perf, highest_perf); WRITE_ONCE(cpudata->highest_perf, highest_perf);
...@@ -324,8 +328,9 @@ static int cppc_init_perf(struct amd_cpudata *cpudata) ...@@ -324,8 +328,9 @@ static int cppc_init_perf(struct amd_cpudata *cpudata)
if (ret) if (ret)
return ret; return ret;
highest_perf = amd_get_highest_perf(); if (cpudata->hw_prefcore)
if (highest_perf > cppc_perf.highest_perf) highest_perf = AMD_PSTATE_PREFCORE_THRESHOLD;
else
highest_perf = cppc_perf.highest_perf; highest_perf = cppc_perf.highest_perf;
WRITE_ONCE(cpudata->highest_perf, highest_perf); WRITE_ONCE(cpudata->highest_perf, highest_perf);
...@@ -706,6 +711,80 @@ static void amd_perf_ctl_reset(unsigned int cpu) ...@@ -706,6 +711,80 @@ static void amd_perf_ctl_reset(unsigned int cpu)
wrmsrl_on_cpu(cpu, MSR_AMD_PERF_CTL, 0); wrmsrl_on_cpu(cpu, MSR_AMD_PERF_CTL, 0);
} }
/*
* Set amd-pstate preferred core enable can't be done directly from cpufreq callbacks
* due to locking, so queue the work for later.
*/
static void amd_pstste_sched_prefcore_workfn(struct work_struct *work)
{
sched_set_itmt_support();
}
static DECLARE_WORK(sched_prefcore_work, amd_pstste_sched_prefcore_workfn);
/*
* Get the highest performance register value.
* @cpu: CPU from which to get highest performance.
* @highest_perf: Return address.
*
* Return: 0 for success, -EIO otherwise.
*/
static int amd_pstate_get_highest_perf(int cpu, u32 *highest_perf)
{
int ret;
if (boot_cpu_has(X86_FEATURE_CPPC)) {
u64 cap1;
ret = rdmsrl_safe_on_cpu(cpu, MSR_AMD_CPPC_CAP1, &cap1);
if (ret)
return ret;
WRITE_ONCE(*highest_perf, AMD_CPPC_HIGHEST_PERF(cap1));
} else {
u64 cppc_highest_perf;
ret = cppc_get_highest_perf(cpu, &cppc_highest_perf);
if (ret)
return ret;
WRITE_ONCE(*highest_perf, cppc_highest_perf);
}
return (ret);
}
#define CPPC_MAX_PERF U8_MAX
static void amd_pstate_init_prefcore(struct amd_cpudata *cpudata)
{
int ret, prio;
u32 highest_perf;
ret = amd_pstate_get_highest_perf(cpudata->cpu, &highest_perf);
if (ret)
return;
cpudata->hw_prefcore = true;
/* check if CPPC preferred core feature is enabled*/
if (highest_perf < CPPC_MAX_PERF)
prio = (int)highest_perf;
else {
pr_debug("AMD CPPC preferred core is unsupported!\n");
cpudata->hw_prefcore = false;
return;
}
if (!amd_pstate_prefcore)
return;
/*
* The priorities can be set regardless of whether or not
* sched_set_itmt_support(true) has been called and it is valid to
* update them at any time after it has been called.
*/
sched_set_itmt_core_prio(prio, cpudata->cpu);
schedule_work(&sched_prefcore_work);
}
static int amd_pstate_cpu_init(struct cpufreq_policy *policy) static int amd_pstate_cpu_init(struct cpufreq_policy *policy)
{ {
int min_freq, max_freq, nominal_freq, lowest_nonlinear_freq, ret; int min_freq, max_freq, nominal_freq, lowest_nonlinear_freq, ret;
...@@ -727,6 +806,8 @@ static int amd_pstate_cpu_init(struct cpufreq_policy *policy) ...@@ -727,6 +806,8 @@ static int amd_pstate_cpu_init(struct cpufreq_policy *policy)
cpudata->cpu = policy->cpu; cpudata->cpu = policy->cpu;
amd_pstate_init_prefcore(cpudata);
ret = amd_pstate_init_perf(cpudata); ret = amd_pstate_init_perf(cpudata);
if (ret) if (ret)
goto free_cpudata1; goto free_cpudata1;
...@@ -877,6 +958,17 @@ static ssize_t show_amd_pstate_highest_perf(struct cpufreq_policy *policy, ...@@ -877,6 +958,17 @@ static ssize_t show_amd_pstate_highest_perf(struct cpufreq_policy *policy,
return sysfs_emit(buf, "%u\n", perf); return sysfs_emit(buf, "%u\n", perf);
} }
static ssize_t show_amd_pstate_hw_prefcore(struct cpufreq_policy *policy,
char *buf)
{
bool hw_prefcore;
struct amd_cpudata *cpudata = policy->driver_data;
hw_prefcore = READ_ONCE(cpudata->hw_prefcore);
return sysfs_emit(buf, "%s\n", str_enabled_disabled(hw_prefcore));
}
static ssize_t show_energy_performance_available_preferences( static ssize_t show_energy_performance_available_preferences(
struct cpufreq_policy *policy, char *buf) struct cpufreq_policy *policy, char *buf)
{ {
...@@ -1074,18 +1166,27 @@ static ssize_t status_store(struct device *a, struct device_attribute *b, ...@@ -1074,18 +1166,27 @@ static ssize_t status_store(struct device *a, struct device_attribute *b,
return ret < 0 ? ret : count; return ret < 0 ? ret : count;
} }
static ssize_t prefcore_show(struct device *dev,
struct device_attribute *attr, char *buf)
{
return sysfs_emit(buf, "%s\n", str_enabled_disabled(amd_pstate_prefcore));
}
cpufreq_freq_attr_ro(amd_pstate_max_freq); cpufreq_freq_attr_ro(amd_pstate_max_freq);
cpufreq_freq_attr_ro(amd_pstate_lowest_nonlinear_freq); cpufreq_freq_attr_ro(amd_pstate_lowest_nonlinear_freq);
cpufreq_freq_attr_ro(amd_pstate_highest_perf); cpufreq_freq_attr_ro(amd_pstate_highest_perf);
cpufreq_freq_attr_ro(amd_pstate_hw_prefcore);
cpufreq_freq_attr_rw(energy_performance_preference); cpufreq_freq_attr_rw(energy_performance_preference);
cpufreq_freq_attr_ro(energy_performance_available_preferences); cpufreq_freq_attr_ro(energy_performance_available_preferences);
static DEVICE_ATTR_RW(status); static DEVICE_ATTR_RW(status);
static DEVICE_ATTR_RO(prefcore);
static struct freq_attr *amd_pstate_attr[] = { static struct freq_attr *amd_pstate_attr[] = {
&amd_pstate_max_freq, &amd_pstate_max_freq,
&amd_pstate_lowest_nonlinear_freq, &amd_pstate_lowest_nonlinear_freq,
&amd_pstate_highest_perf, &amd_pstate_highest_perf,
&amd_pstate_hw_prefcore,
NULL, NULL,
}; };
...@@ -1093,6 +1194,7 @@ static struct freq_attr *amd_pstate_epp_attr[] = { ...@@ -1093,6 +1194,7 @@ static struct freq_attr *amd_pstate_epp_attr[] = {
&amd_pstate_max_freq, &amd_pstate_max_freq,
&amd_pstate_lowest_nonlinear_freq, &amd_pstate_lowest_nonlinear_freq,
&amd_pstate_highest_perf, &amd_pstate_highest_perf,
&amd_pstate_hw_prefcore,
&energy_performance_preference, &energy_performance_preference,
&energy_performance_available_preferences, &energy_performance_available_preferences,
NULL, NULL,
...@@ -1100,6 +1202,7 @@ static struct freq_attr *amd_pstate_epp_attr[] = { ...@@ -1100,6 +1202,7 @@ static struct freq_attr *amd_pstate_epp_attr[] = {
static struct attribute *pstate_global_attributes[] = { static struct attribute *pstate_global_attributes[] = {
&dev_attr_status.attr, &dev_attr_status.attr,
&dev_attr_prefcore.attr,
NULL NULL
}; };
...@@ -1151,6 +1254,8 @@ static int amd_pstate_epp_cpu_init(struct cpufreq_policy *policy) ...@@ -1151,6 +1254,8 @@ static int amd_pstate_epp_cpu_init(struct cpufreq_policy *policy)
cpudata->cpu = policy->cpu; cpudata->cpu = policy->cpu;
cpudata->epp_policy = 0; cpudata->epp_policy = 0;
amd_pstate_init_prefcore(cpudata);
ret = amd_pstate_init_perf(cpudata); ret = amd_pstate_init_perf(cpudata);
if (ret) if (ret)
goto free_cpudata1; goto free_cpudata1;
...@@ -1567,7 +1672,17 @@ static int __init amd_pstate_param(char *str) ...@@ -1567,7 +1672,17 @@ static int __init amd_pstate_param(char *str)
return amd_pstate_set_driver(mode_idx); return amd_pstate_set_driver(mode_idx);
} }
static int __init amd_prefcore_param(char *str)
{
if (!strcmp(str, "disable"))
amd_pstate_prefcore = false;
return 0;
}
early_param("amd_pstate", amd_pstate_param); early_param("amd_pstate", amd_pstate_param);
early_param("amd_prefcore", amd_prefcore_param);
MODULE_AUTHOR("Huang Rui <ray.huang@amd.com>"); MODULE_AUTHOR("Huang Rui <ray.huang@amd.com>");
MODULE_DESCRIPTION("AMD Processor P-state Frequency Driver"); MODULE_DESCRIPTION("AMD Processor P-state Frequency Driver");
...@@ -52,6 +52,9 @@ struct amd_aperf_mperf { ...@@ -52,6 +52,9 @@ struct amd_aperf_mperf {
* @prev: Last Aperf/Mperf/tsc count value read from register * @prev: Last Aperf/Mperf/tsc count value read from register
* @freq: current cpu frequency value * @freq: current cpu frequency value
* @boost_supported: check whether the Processor or SBIOS supports boost mode * @boost_supported: check whether the Processor or SBIOS supports boost mode
* @hw_prefcore: check whether HW supports preferred core featue.
* Only when hw_prefcore and early prefcore param are true,
* AMD P-State driver supports preferred core featue.
* @epp_policy: Last saved policy used to set energy-performance preference * @epp_policy: Last saved policy used to set energy-performance preference
* @epp_cached: Cached CPPC energy-performance preference value * @epp_cached: Cached CPPC energy-performance preference value
* @policy: Cpufreq policy value * @policy: Cpufreq policy value
...@@ -85,6 +88,7 @@ struct amd_cpudata { ...@@ -85,6 +88,7 @@ struct amd_cpudata {
u64 freq; u64 freq;
bool boost_supported; bool boost_supported;
bool hw_prefcore;
/* EPP feature related attributes*/ /* EPP feature related attributes*/
s16 epp_policy; s16 epp_policy;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment