Commit 47935a73 authored by Linus Torvalds's avatar Linus Torvalds

Merge branches 'x86-alternatives-for-linus', 'x86-fpu-for-linus',...

Merge branches 'x86-alternatives-for-linus', 'x86-fpu-for-linus', 'x86-hwmon-for-linus', 'x86-paravirt-for-linus', 'core-locking-for-linus' and 'irq-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip

* 'x86-alternatives-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  x86, suspend: Avoid unnecessary smp alternatives switch during suspend/resume

* 'x86-fpu-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  x86-64, asm: Use fxsaveq/fxrestorq in more places

* 'x86-hwmon-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  x86, hwmon: Add core threshold notification to therm_throt.c

* 'x86-paravirt-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  x86, paravirt: Use native_halt on a halt, not native_safe_halt

* 'core-locking-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  locking, lockdep: Convert sprintf_symbol to %pS

* 'irq-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  irq: Better struct irqaction layout
...@@ -66,6 +66,7 @@ extern void alternatives_smp_module_add(struct module *mod, char *name, ...@@ -66,6 +66,7 @@ extern void alternatives_smp_module_add(struct module *mod, char *name,
extern void alternatives_smp_module_del(struct module *mod); extern void alternatives_smp_module_del(struct module *mod);
extern void alternatives_smp_switch(int smp); extern void alternatives_smp_switch(int smp);
extern int alternatives_text_reserved(void *start, void *end); extern int alternatives_text_reserved(void *start, void *end);
extern bool skip_smp_alternatives;
#else #else
static inline void alternatives_smp_module_add(struct module *mod, char *name, static inline void alternatives_smp_module_add(struct module *mod, char *name,
void *locks, void *locks_end, void *locks, void *locks_end,
......
...@@ -93,6 +93,17 @@ static inline int fxrstor_checking(struct i387_fxsave_struct *fx) ...@@ -93,6 +93,17 @@ static inline int fxrstor_checking(struct i387_fxsave_struct *fx)
int err; int err;
/* See comment in fxsave() below. */ /* See comment in fxsave() below. */
#ifdef CONFIG_AS_FXSAVEQ
asm volatile("1: fxrstorq %[fx]\n\t"
"2:\n"
".section .fixup,\"ax\"\n"
"3: movl $-1,%[err]\n"
" jmp 2b\n"
".previous\n"
_ASM_EXTABLE(1b, 3b)
: [err] "=r" (err)
: [fx] "m" (*fx), "0" (0));
#else
asm volatile("1: rex64/fxrstor (%[fx])\n\t" asm volatile("1: rex64/fxrstor (%[fx])\n\t"
"2:\n" "2:\n"
".section .fixup,\"ax\"\n" ".section .fixup,\"ax\"\n"
...@@ -102,6 +113,7 @@ static inline int fxrstor_checking(struct i387_fxsave_struct *fx) ...@@ -102,6 +113,7 @@ static inline int fxrstor_checking(struct i387_fxsave_struct *fx)
_ASM_EXTABLE(1b, 3b) _ASM_EXTABLE(1b, 3b)
: [err] "=r" (err) : [err] "=r" (err)
: [fx] "R" (fx), "m" (*fx), "0" (0)); : [fx] "R" (fx), "m" (*fx), "0" (0));
#endif
return err; return err;
} }
...@@ -119,6 +131,17 @@ static inline int fxsave_user(struct i387_fxsave_struct __user *fx) ...@@ -119,6 +131,17 @@ static inline int fxsave_user(struct i387_fxsave_struct __user *fx)
return -EFAULT; return -EFAULT;
/* See comment in fxsave() below. */ /* See comment in fxsave() below. */
#ifdef CONFIG_AS_FXSAVEQ
asm volatile("1: fxsaveq %[fx]\n\t"
"2:\n"
".section .fixup,\"ax\"\n"
"3: movl $-1,%[err]\n"
" jmp 2b\n"
".previous\n"
_ASM_EXTABLE(1b, 3b)
: [err] "=r" (err), [fx] "=m" (*fx)
: "0" (0));
#else
asm volatile("1: rex64/fxsave (%[fx])\n\t" asm volatile("1: rex64/fxsave (%[fx])\n\t"
"2:\n" "2:\n"
".section .fixup,\"ax\"\n" ".section .fixup,\"ax\"\n"
...@@ -128,6 +151,7 @@ static inline int fxsave_user(struct i387_fxsave_struct __user *fx) ...@@ -128,6 +151,7 @@ static inline int fxsave_user(struct i387_fxsave_struct __user *fx)
_ASM_EXTABLE(1b, 3b) _ASM_EXTABLE(1b, 3b)
: [err] "=r" (err), "=m" (*fx) : [err] "=r" (err), "=m" (*fx)
: [fx] "R" (fx), "0" (0)); : [fx] "R" (fx), "0" (0));
#endif
if (unlikely(err) && if (unlikely(err) &&
__clear_user(fx, sizeof(struct i387_fxsave_struct))) __clear_user(fx, sizeof(struct i387_fxsave_struct)))
err = -EFAULT; err = -EFAULT;
......
...@@ -223,6 +223,9 @@ void intel_init_thermal(struct cpuinfo_x86 *c); ...@@ -223,6 +223,9 @@ void intel_init_thermal(struct cpuinfo_x86 *c);
void mce_log_therm_throt_event(__u64 status); void mce_log_therm_throt_event(__u64 status);
/* Interrupt Handler for core thermal thresholds */
extern int (*platform_thermal_notify)(__u64 msr_val);
#ifdef CONFIG_X86_THERMAL_VECTOR #ifdef CONFIG_X86_THERMAL_VECTOR
extern void mcheck_intel_therm_init(void); extern void mcheck_intel_therm_init(void);
#else #else
......
...@@ -257,6 +257,18 @@ ...@@ -257,6 +257,18 @@
#define PACKAGE_THERM_INT_LOW_ENABLE (1 << 1) #define PACKAGE_THERM_INT_LOW_ENABLE (1 << 1)
#define PACKAGE_THERM_INT_PLN_ENABLE (1 << 24) #define PACKAGE_THERM_INT_PLN_ENABLE (1 << 24)
/* Thermal Thresholds Support */
#define THERM_INT_THRESHOLD0_ENABLE (1 << 15)
#define THERM_SHIFT_THRESHOLD0 8
#define THERM_MASK_THRESHOLD0 (0x7f << THERM_SHIFT_THRESHOLD0)
#define THERM_INT_THRESHOLD1_ENABLE (1 << 23)
#define THERM_SHIFT_THRESHOLD1 16
#define THERM_MASK_THRESHOLD1 (0x7f << THERM_SHIFT_THRESHOLD1)
#define THERM_STATUS_THRESHOLD0 (1 << 6)
#define THERM_LOG_THRESHOLD0 (1 << 7)
#define THERM_STATUS_THRESHOLD1 (1 << 8)
#define THERM_LOG_THRESHOLD1 (1 << 9)
/* MISC_ENABLE bits: architectural */ /* MISC_ENABLE bits: architectural */
#define MSR_IA32_MISC_ENABLE_FAST_STRING (1ULL << 0) #define MSR_IA32_MISC_ENABLE_FAST_STRING (1ULL << 0)
#define MSR_IA32_MISC_ENABLE_TCC (1ULL << 1) #define MSR_IA32_MISC_ENABLE_TCC (1ULL << 1)
......
...@@ -112,7 +112,7 @@ static inline void arch_safe_halt(void) ...@@ -112,7 +112,7 @@ static inline void arch_safe_halt(void)
static inline void halt(void) static inline void halt(void)
{ {
PVOP_VCALL0(pv_irq_ops.safe_halt); PVOP_VCALL0(pv_irq_ops.halt);
} }
static inline void wbinvd(void) static inline void wbinvd(void)
......
...@@ -353,6 +353,7 @@ void __init_or_module alternatives_smp_module_del(struct module *mod) ...@@ -353,6 +353,7 @@ void __init_or_module alternatives_smp_module_del(struct module *mod)
mutex_unlock(&smp_alt); mutex_unlock(&smp_alt);
} }
bool skip_smp_alternatives;
void alternatives_smp_switch(int smp) void alternatives_smp_switch(int smp)
{ {
struct smp_alt_module *mod; struct smp_alt_module *mod;
...@@ -368,7 +369,7 @@ void alternatives_smp_switch(int smp) ...@@ -368,7 +369,7 @@ void alternatives_smp_switch(int smp)
printk("lockdep: fixing up alternatives.\n"); printk("lockdep: fixing up alternatives.\n");
#endif #endif
if (noreplace_smp || smp_alt_once) if (noreplace_smp || smp_alt_once || skip_smp_alternatives)
return; return;
BUG_ON(!smp && (num_online_cpus() > 1)); BUG_ON(!smp && (num_online_cpus() > 1));
......
...@@ -53,8 +53,13 @@ struct thermal_state { ...@@ -53,8 +53,13 @@ struct thermal_state {
struct _thermal_state core_power_limit; struct _thermal_state core_power_limit;
struct _thermal_state package_throttle; struct _thermal_state package_throttle;
struct _thermal_state package_power_limit; struct _thermal_state package_power_limit;
struct _thermal_state core_thresh0;
struct _thermal_state core_thresh1;
}; };
/* Callback to handle core threshold interrupts */
int (*platform_thermal_notify)(__u64 msr_val);
static DEFINE_PER_CPU(struct thermal_state, thermal_state); static DEFINE_PER_CPU(struct thermal_state, thermal_state);
static atomic_t therm_throt_en = ATOMIC_INIT(0); static atomic_t therm_throt_en = ATOMIC_INIT(0);
...@@ -200,6 +205,22 @@ static int therm_throt_process(bool new_event, int event, int level) ...@@ -200,6 +205,22 @@ static int therm_throt_process(bool new_event, int event, int level)
return 0; return 0;
} }
static int thresh_event_valid(int event)
{
struct _thermal_state *state;
unsigned int this_cpu = smp_processor_id();
struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu);
u64 now = get_jiffies_64();
state = (event == 0) ? &pstate->core_thresh0 : &pstate->core_thresh1;
if (time_before64(now, state->next_check))
return 0;
state->next_check = now + CHECK_INTERVAL;
return 1;
}
#ifdef CONFIG_SYSFS #ifdef CONFIG_SYSFS
/* Add/Remove thermal_throttle interface for CPU device: */ /* Add/Remove thermal_throttle interface for CPU device: */
static __cpuinit int thermal_throttle_add_dev(struct sys_device *sys_dev, static __cpuinit int thermal_throttle_add_dev(struct sys_device *sys_dev,
...@@ -313,6 +334,22 @@ device_initcall(thermal_throttle_init_device); ...@@ -313,6 +334,22 @@ device_initcall(thermal_throttle_init_device);
#define PACKAGE_THROTTLED ((__u64)2 << 62) #define PACKAGE_THROTTLED ((__u64)2 << 62)
#define PACKAGE_POWER_LIMIT ((__u64)3 << 62) #define PACKAGE_POWER_LIMIT ((__u64)3 << 62)
static void notify_thresholds(__u64 msr_val)
{
/* check whether the interrupt handler is defined;
* otherwise simply return
*/
if (!platform_thermal_notify)
return;
/* lower threshold reached */
if ((msr_val & THERM_LOG_THRESHOLD0) && thresh_event_valid(0))
platform_thermal_notify(msr_val);
/* higher threshold reached */
if ((msr_val & THERM_LOG_THRESHOLD1) && thresh_event_valid(1))
platform_thermal_notify(msr_val);
}
/* Thermal transition interrupt handler */ /* Thermal transition interrupt handler */
static void intel_thermal_interrupt(void) static void intel_thermal_interrupt(void)
{ {
...@@ -321,6 +358,9 @@ static void intel_thermal_interrupt(void) ...@@ -321,6 +358,9 @@ static void intel_thermal_interrupt(void)
rdmsrl(MSR_IA32_THERM_STATUS, msr_val); rdmsrl(MSR_IA32_THERM_STATUS, msr_val);
/* Check for violation of core thermal thresholds*/
notify_thresholds(msr_val);
if (therm_throt_process(msr_val & THERM_STATUS_PROCHOT, if (therm_throt_process(msr_val & THERM_STATUS_PROCHOT,
THERMAL_THROTTLING_EVENT, THERMAL_THROTTLING_EVENT,
CORE_LEVEL) != 0) CORE_LEVEL) != 0)
......
...@@ -1161,6 +1161,20 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus) ...@@ -1161,6 +1161,20 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus)
preempt_enable(); preempt_enable();
} }
void arch_disable_nonboot_cpus_begin(void)
{
/*
* Avoid the smp alternatives switch during the disable_nonboot_cpus().
* In the suspend path, we will be back in the SMP mode shortly anyways.
*/
skip_smp_alternatives = true;
}
void arch_disable_nonboot_cpus_end(void)
{
skip_smp_alternatives = false;
}
void arch_enable_nonboot_cpus_begin(void) void arch_enable_nonboot_cpus_begin(void)
{ {
set_mtrr_aps_delayed_init(); set_mtrr_aps_delayed_init();
......
...@@ -114,15 +114,15 @@ typedef irqreturn_t (*irq_handler_t)(int, void *); ...@@ -114,15 +114,15 @@ typedef irqreturn_t (*irq_handler_t)(int, void *);
struct irqaction { struct irqaction {
irq_handler_t handler; irq_handler_t handler;
unsigned long flags; unsigned long flags;
const char *name;
void *dev_id; void *dev_id;
struct irqaction *next; struct irqaction *next;
int irq; int irq;
struct proc_dir_entry *dir;
irq_handler_t thread_fn; irq_handler_t thread_fn;
struct task_struct *thread; struct task_struct *thread;
unsigned long thread_flags; unsigned long thread_flags;
}; const char *name;
struct proc_dir_entry *dir;
} ____cacheline_internodealigned_in_smp;
extern irqreturn_t no_action(int cpl, void *dev_id); extern irqreturn_t no_action(int cpl, void *dev_id);
......
...@@ -384,6 +384,14 @@ int __cpuinit cpu_up(unsigned int cpu) ...@@ -384,6 +384,14 @@ int __cpuinit cpu_up(unsigned int cpu)
#ifdef CONFIG_PM_SLEEP_SMP #ifdef CONFIG_PM_SLEEP_SMP
static cpumask_var_t frozen_cpus; static cpumask_var_t frozen_cpus;
void __weak arch_disable_nonboot_cpus_begin(void)
{
}
void __weak arch_disable_nonboot_cpus_end(void)
{
}
int disable_nonboot_cpus(void) int disable_nonboot_cpus(void)
{ {
int cpu, first_cpu, error = 0; int cpu, first_cpu, error = 0;
...@@ -395,6 +403,7 @@ int disable_nonboot_cpus(void) ...@@ -395,6 +403,7 @@ int disable_nonboot_cpus(void)
* with the userspace trying to use the CPU hotplug at the same time * with the userspace trying to use the CPU hotplug at the same time
*/ */
cpumask_clear(frozen_cpus); cpumask_clear(frozen_cpus);
arch_disable_nonboot_cpus_begin();
printk("Disabling non-boot CPUs ...\n"); printk("Disabling non-boot CPUs ...\n");
for_each_online_cpu(cpu) { for_each_online_cpu(cpu) {
...@@ -410,6 +419,8 @@ int disable_nonboot_cpus(void) ...@@ -410,6 +419,8 @@ int disable_nonboot_cpus(void)
} }
} }
arch_disable_nonboot_cpus_end();
if (!error) { if (!error) {
BUG_ON(num_online_cpus() > 1); BUG_ON(num_online_cpus() > 1);
/* Make sure the CPUs won't be enabled by someone else */ /* Make sure the CPUs won't be enabled by someone else */
......
...@@ -494,7 +494,6 @@ static void seq_stats(struct seq_file *m, struct lock_stat_data *data) ...@@ -494,7 +494,6 @@ static void seq_stats(struct seq_file *m, struct lock_stat_data *data)
namelen += 2; namelen += 2;
for (i = 0; i < LOCKSTAT_POINTS; i++) { for (i = 0; i < LOCKSTAT_POINTS; i++) {
char sym[KSYM_SYMBOL_LEN];
char ip[32]; char ip[32];
if (class->contention_point[i] == 0) if (class->contention_point[i] == 0)
...@@ -503,15 +502,13 @@ static void seq_stats(struct seq_file *m, struct lock_stat_data *data) ...@@ -503,15 +502,13 @@ static void seq_stats(struct seq_file *m, struct lock_stat_data *data)
if (!i) if (!i)
seq_line(m, '-', 40-namelen, namelen); seq_line(m, '-', 40-namelen, namelen);
sprint_symbol(sym, class->contention_point[i]);
snprintf(ip, sizeof(ip), "[<%p>]", snprintf(ip, sizeof(ip), "[<%p>]",
(void *)class->contention_point[i]); (void *)class->contention_point[i]);
seq_printf(m, "%40s %14lu %29s %s\n", name, seq_printf(m, "%40s %14lu %29s %pS\n",
stats->contention_point[i], name, stats->contention_point[i],
ip, sym); ip, (void *)class->contention_point[i]);
} }
for (i = 0; i < LOCKSTAT_POINTS; i++) { for (i = 0; i < LOCKSTAT_POINTS; i++) {
char sym[KSYM_SYMBOL_LEN];
char ip[32]; char ip[32];
if (class->contending_point[i] == 0) if (class->contending_point[i] == 0)
...@@ -520,12 +517,11 @@ static void seq_stats(struct seq_file *m, struct lock_stat_data *data) ...@@ -520,12 +517,11 @@ static void seq_stats(struct seq_file *m, struct lock_stat_data *data)
if (!i) if (!i)
seq_line(m, '-', 40-namelen, namelen); seq_line(m, '-', 40-namelen, namelen);
sprint_symbol(sym, class->contending_point[i]);
snprintf(ip, sizeof(ip), "[<%p>]", snprintf(ip, sizeof(ip), "[<%p>]",
(void *)class->contending_point[i]); (void *)class->contending_point[i]);
seq_printf(m, "%40s %14lu %29s %s\n", name, seq_printf(m, "%40s %14lu %29s %pS\n",
stats->contending_point[i], name, stats->contending_point[i],
ip, sym); ip, (void *)class->contending_point[i]);
} }
if (i) { if (i) {
seq_puts(m, "\n"); seq_puts(m, "\n");
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment