Commit 1e4230f5 authored by Rafael J. Wysocki's avatar Rafael J. Wysocki

Merge branches 'pm-sleep', 'pm-cpuidle', 'pm-cpufreq', 'pm-devfreq' and 'pm-avs'

* pm-sleep:
  ACPI: PM: s2idle: Rework ACPI events synchronization
  ACPI: EC: Rework flushing of pending work

* pm-cpuidle:
  cpuidle: minor Kconfig help text fixes
  cpuidle: Drop disabled field from struct cpuidle_state
  cpuidle: Fix Kconfig indentation

* pm-cpufreq:
  cpufreq: Fix Kconfig indentation

* pm-devfreq:
  PM / devfreq: Add missing locking while setting suspend_freq

* pm-avs:
  power: avs: Fix Kconfig indentation
...@@ -67,7 +67,7 @@ static struct cpuidle_driver cpuidle_driver = { ...@@ -67,7 +67,7 @@ static struct cpuidle_driver cpuidle_driver = {
.enter = cpuidle_sleep_enter, .enter = cpuidle_sleep_enter,
.name = "C2", .name = "C2",
.desc = "SuperH Sleep Mode [SF]", .desc = "SuperH Sleep Mode [SF]",
.disabled = true, .flags = CPUIDLE_FLAG_UNUSABLE,
}, },
{ {
.exit_latency = 2300, .exit_latency = 2300,
...@@ -76,7 +76,7 @@ static struct cpuidle_driver cpuidle_driver = { ...@@ -76,7 +76,7 @@ static struct cpuidle_driver cpuidle_driver = {
.enter = cpuidle_sleep_enter, .enter = cpuidle_sleep_enter,
.name = "C3", .name = "C3",
.desc = "SuperH Mobile Standby Mode [SF]", .desc = "SuperH Mobile Standby Mode [SF]",
.disabled = true, .flags = CPUIDLE_FLAG_UNUSABLE,
}, },
}, },
.safe_state_index = 0, .safe_state_index = 0,
...@@ -86,10 +86,10 @@ static struct cpuidle_driver cpuidle_driver = { ...@@ -86,10 +86,10 @@ static struct cpuidle_driver cpuidle_driver = {
int __init sh_mobile_setup_cpuidle(void) int __init sh_mobile_setup_cpuidle(void)
{ {
if (sh_mobile_sleep_supported & SUSP_SH_SF) if (sh_mobile_sleep_supported & SUSP_SH_SF)
cpuidle_driver.states[1].disabled = false; cpuidle_driver.states[1].flags = CPUIDLE_FLAG_NONE;
if (sh_mobile_sleep_supported & SUSP_SH_STANDBY) if (sh_mobile_sleep_supported & SUSP_SH_STANDBY)
cpuidle_driver.states[2].disabled = false; cpuidle_driver.states[2].flags = CPUIDLE_FLAG_NONE;
return cpuidle_register(&cpuidle_driver, NULL); return cpuidle_register(&cpuidle_driver, NULL);
} }
...@@ -533,26 +533,10 @@ static void acpi_ec_enable_event(struct acpi_ec *ec) ...@@ -533,26 +533,10 @@ static void acpi_ec_enable_event(struct acpi_ec *ec)
} }
#ifdef CONFIG_PM_SLEEP #ifdef CONFIG_PM_SLEEP
static bool acpi_ec_query_flushed(struct acpi_ec *ec) static void __acpi_ec_flush_work(void)
{ {
bool flushed; flush_scheduled_work(); /* flush ec->work */
unsigned long flags; flush_workqueue(ec_query_wq); /* flush queries */
spin_lock_irqsave(&ec->lock, flags);
flushed = !ec->nr_pending_queries;
spin_unlock_irqrestore(&ec->lock, flags);
return flushed;
}
static void __acpi_ec_flush_event(struct acpi_ec *ec)
{
/*
* When ec_freeze_events is true, we need to flush events in
* the proper position before entering the noirq stage.
*/
wait_event(ec->wait, acpi_ec_query_flushed(ec));
if (ec_query_wq)
flush_workqueue(ec_query_wq);
} }
static void acpi_ec_disable_event(struct acpi_ec *ec) static void acpi_ec_disable_event(struct acpi_ec *ec)
...@@ -562,15 +546,21 @@ static void acpi_ec_disable_event(struct acpi_ec *ec) ...@@ -562,15 +546,21 @@ static void acpi_ec_disable_event(struct acpi_ec *ec)
spin_lock_irqsave(&ec->lock, flags); spin_lock_irqsave(&ec->lock, flags);
__acpi_ec_disable_event(ec); __acpi_ec_disable_event(ec);
spin_unlock_irqrestore(&ec->lock, flags); spin_unlock_irqrestore(&ec->lock, flags);
__acpi_ec_flush_event(ec);
/*
* When ec_freeze_events is true, we need to flush events in
* the proper position before entering the noirq stage.
*/
__acpi_ec_flush_work();
} }
void acpi_ec_flush_work(void) void acpi_ec_flush_work(void)
{ {
if (first_ec) /* Without ec_query_wq there is nothing to flush. */
__acpi_ec_flush_event(first_ec); if (!ec_query_wq)
return;
flush_scheduled_work(); __acpi_ec_flush_work();
} }
#endif /* CONFIG_PM_SLEEP */ #endif /* CONFIG_PM_SLEEP */
......
...@@ -977,6 +977,16 @@ static int acpi_s2idle_prepare_late(void) ...@@ -977,6 +977,16 @@ static int acpi_s2idle_prepare_late(void)
return 0; return 0;
} }
static void acpi_s2idle_sync(void)
{
/*
* The EC driver uses the system workqueue and an additional special
* one, so those need to be flushed too.
*/
acpi_ec_flush_work();
acpi_os_wait_events_complete(); /* synchronize Notify handling */
}
static void acpi_s2idle_wake(void) static void acpi_s2idle_wake(void)
{ {
/* /*
...@@ -1001,13 +1011,8 @@ static void acpi_s2idle_wake(void) ...@@ -1001,13 +1011,8 @@ static void acpi_s2idle_wake(void)
* should be missed by canceling the wakeup here. * should be missed by canceling the wakeup here.
*/ */
pm_system_cancel_wakeup(); pm_system_cancel_wakeup();
/*
* The EC driver uses the system workqueue and an additional acpi_s2idle_sync();
* special one, so those need to be flushed too.
*/
acpi_os_wait_events_complete(); /* synchronize EC GPE processing */
acpi_ec_flush_work();
acpi_os_wait_events_complete(); /* synchronize Notify handling */
rearm_wake_irq(acpi_sci_irq); rearm_wake_irq(acpi_sci_irq);
} }
...@@ -1024,6 +1029,13 @@ static void acpi_s2idle_restore_early(void) ...@@ -1024,6 +1029,13 @@ static void acpi_s2idle_restore_early(void)
static void acpi_s2idle_restore(void) static void acpi_s2idle_restore(void)
{ {
/*
* Drain pending events before restoring the working-state configuration
* of GPEs.
*/
acpi_os_wait_events_complete(); /* synchronize GPE processing */
acpi_s2idle_sync();
s2idle_wakeup = false; s2idle_wakeup = false;
acpi_enable_all_runtime_gpes(); acpi_enable_all_runtime_gpes();
......
...@@ -65,21 +65,21 @@ config ARM_U8500_CPUIDLE ...@@ -65,21 +65,21 @@ config ARM_U8500_CPUIDLE
bool "Cpu Idle Driver for the ST-E u8500 processors" bool "Cpu Idle Driver for the ST-E u8500 processors"
depends on ARCH_U8500 && !ARM64 depends on ARCH_U8500 && !ARM64
help help
Select this to enable cpuidle for ST-E u8500 processors Select this to enable cpuidle for ST-E u8500 processors.
config ARM_AT91_CPUIDLE config ARM_AT91_CPUIDLE
bool "Cpu Idle Driver for the AT91 processors" bool "Cpu Idle Driver for the AT91 processors"
default y default y
depends on ARCH_AT91 && !ARM64 depends on ARCH_AT91 && !ARM64
help help
Select this to enable cpuidle for AT91 processors Select this to enable cpuidle for AT91 processors.
config ARM_EXYNOS_CPUIDLE config ARM_EXYNOS_CPUIDLE
bool "Cpu Idle Driver for the Exynos processors" bool "Cpu Idle Driver for the Exynos processors"
depends on ARCH_EXYNOS && !ARM64 depends on ARCH_EXYNOS && !ARM64
select ARCH_NEEDS_CPU_IDLE_COUPLED if SMP select ARCH_NEEDS_CPU_IDLE_COUPLED if SMP
help help
Select this to enable cpuidle for Exynos processors Select this to enable cpuidle for Exynos processors.
config ARM_MVEBU_V7_CPUIDLE config ARM_MVEBU_V7_CPUIDLE
bool "CPU Idle Driver for mvebu v7 family processors" bool "CPU Idle Driver for mvebu v7 family processors"
......
...@@ -572,7 +572,7 @@ static int __cpuidle_register_device(struct cpuidle_device *dev) ...@@ -572,7 +572,7 @@ static int __cpuidle_register_device(struct cpuidle_device *dev)
return -EINVAL; return -EINVAL;
for (i = 0; i < drv->state_count; i++) for (i = 0; i < drv->state_count; i++)
if (drv->states[i].disabled) if (drv->states[i].flags & CPUIDLE_FLAG_UNUSABLE)
dev->states_usage[i].disable |= CPUIDLE_STATE_DISABLED_BY_DRIVER; dev->states_usage[i].disable |= CPUIDLE_STATE_DISABLED_BY_DRIVER;
per_cpu(cpuidle_devices, dev->cpu) = dev; per_cpu(cpuidle_devices, dev->cpu) = dev;
......
...@@ -53,7 +53,6 @@ void cpuidle_poll_state_init(struct cpuidle_driver *drv) ...@@ -53,7 +53,6 @@ void cpuidle_poll_state_init(struct cpuidle_driver *drv)
state->target_residency_ns = 0; state->target_residency_ns = 0;
state->power_usage = -1; state->power_usage = -1;
state->enter = poll_idle; state->enter = poll_idle;
state->disabled = false;
state->flags = CPUIDLE_FLAG_POLLING; state->flags = CPUIDLE_FLAG_POLLING;
} }
EXPORT_SYMBOL_GPL(cpuidle_poll_state_init); EXPORT_SYMBOL_GPL(cpuidle_poll_state_init);
...@@ -921,7 +921,9 @@ int devfreq_suspend_device(struct devfreq *devfreq) ...@@ -921,7 +921,9 @@ int devfreq_suspend_device(struct devfreq *devfreq)
} }
if (devfreq->suspend_freq) { if (devfreq->suspend_freq) {
mutex_lock(&devfreq->lock);
ret = devfreq_set_target(devfreq, devfreq->suspend_freq, 0); ret = devfreq_set_target(devfreq, devfreq->suspend_freq, 0);
mutex_unlock(&devfreq->lock);
if (ret) if (ret)
return ret; return ret;
} }
...@@ -949,7 +951,9 @@ int devfreq_resume_device(struct devfreq *devfreq) ...@@ -949,7 +951,9 @@ int devfreq_resume_device(struct devfreq *devfreq)
return 0; return 0;
if (devfreq->resume_freq) { if (devfreq->resume_freq) {
mutex_lock(&devfreq->lock);
ret = devfreq_set_target(devfreq, devfreq->resume_freq, 0); ret = devfreq_set_target(devfreq, devfreq->resume_freq, 0);
mutex_unlock(&devfreq->lock);
if (ret) if (ret)
return ret; return ret;
} }
......
...@@ -1291,8 +1291,8 @@ static void sklh_idle_state_table_update(void) ...@@ -1291,8 +1291,8 @@ static void sklh_idle_state_table_update(void)
return; return;
} }
skl_cstates[5].disabled = 1; /* C8-SKL */ skl_cstates[5].flags |= CPUIDLE_FLAG_UNUSABLE; /* C8-SKL */
skl_cstates[6].disabled = 1; /* C9-SKL */ skl_cstates[6].flags |= CPUIDLE_FLAG_UNUSABLE; /* C9-SKL */
} }
/* /*
* intel_idle_state_table_update() * intel_idle_state_table_update()
...@@ -1355,7 +1355,7 @@ static void __init intel_idle_cpuidle_driver_init(void) ...@@ -1355,7 +1355,7 @@ static void __init intel_idle_cpuidle_driver_init(void)
continue; continue;
/* if state marked as disabled, skip it */ /* if state marked as disabled, skip it */
if (cpuidle_state_table[cstate].disabled != 0) { if (cpuidle_state_table[cstate].flags & CPUIDLE_FLAG_UNUSABLE) {
pr_debug("state %s is disabled\n", pr_debug("state %s is disabled\n",
cpuidle_state_table[cstate].name); cpuidle_state_table[cstate].name);
continue; continue;
......
...@@ -54,7 +54,6 @@ struct cpuidle_state { ...@@ -54,7 +54,6 @@ struct cpuidle_state {
unsigned int exit_latency; /* in US */ unsigned int exit_latency; /* in US */
int power_usage; /* in mW */ int power_usage; /* in mW */
unsigned int target_residency; /* in US */ unsigned int target_residency; /* in US */
bool disabled; /* disabled on all CPUs */
int (*enter) (struct cpuidle_device *dev, int (*enter) (struct cpuidle_device *dev,
struct cpuidle_driver *drv, struct cpuidle_driver *drv,
...@@ -77,6 +76,7 @@ struct cpuidle_state { ...@@ -77,6 +76,7 @@ struct cpuidle_state {
#define CPUIDLE_FLAG_POLLING BIT(0) /* polling state */ #define CPUIDLE_FLAG_POLLING BIT(0) /* polling state */
#define CPUIDLE_FLAG_COUPLED BIT(1) /* state applies to multiple cpus */ #define CPUIDLE_FLAG_COUPLED BIT(1) /* state applies to multiple cpus */
#define CPUIDLE_FLAG_TIMER_STOP BIT(2) /* timer is stopped on this state */ #define CPUIDLE_FLAG_TIMER_STOP BIT(2) /* timer is stopped on this state */
#define CPUIDLE_FLAG_UNUSABLE BIT(3) /* avoid using this state */
struct cpuidle_device_kobj; struct cpuidle_device_kobj;
struct cpuidle_state_kobj; struct cpuidle_state_kobj;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment