Commit c5861658 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip

Pull x86 fixes from Ingo Molnar:
 "Misc fixes from all around the place:

   - a KASLR related revert where we ran out of time to get a fix - this
     represents a substantial portion of the diffstat,

   - two FPU fixes,

   - two x86 platform fixes: an ACPI reduced-hw fix and a NumaChip fix,

   - an entry code fix,

   - and a VDSO build fix"

* 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
  Revert "x86/mm/ASLR: Propagate base load address calculation"
  x86/fpu: Drop_fpu() should not assume that tsk equals current
  x86/fpu: Avoid math_state_restore() without used_math() in __restore_xstate_sig()
  x86/apic/numachip: Fix sibling map with NumaChip
  x86/platform, acpi: Bypass legacy PIC and PIT in ACPI hardware reduced mode
  x86/asm/entry/32: Fix user_mode() misuses
  x86/vdso: Fix the build on GCC5
parents 13326e5a 69797daf
...@@ -14,13 +14,6 @@ ...@@ -14,13 +14,6 @@
static const char build_str[] = UTS_RELEASE " (" LINUX_COMPILE_BY "@" static const char build_str[] = UTS_RELEASE " (" LINUX_COMPILE_BY "@"
LINUX_COMPILE_HOST ") (" LINUX_COMPILER ") " UTS_VERSION; LINUX_COMPILE_HOST ") (" LINUX_COMPILER ") " UTS_VERSION;
struct kaslr_setup_data {
__u64 next;
__u32 type;
__u32 len;
__u8 data[1];
} kaslr_setup_data;
#define I8254_PORT_CONTROL 0x43 #define I8254_PORT_CONTROL 0x43
#define I8254_PORT_COUNTER0 0x40 #define I8254_PORT_COUNTER0 0x40
#define I8254_CMD_READBACK 0xC0 #define I8254_CMD_READBACK 0xC0
...@@ -302,29 +295,7 @@ static unsigned long find_random_addr(unsigned long minimum, ...@@ -302,29 +295,7 @@ static unsigned long find_random_addr(unsigned long minimum,
return slots_fetch_random(); return slots_fetch_random();
} }
static void add_kaslr_setup_data(struct boot_params *params, __u8 enabled) unsigned char *choose_kernel_location(unsigned char *input,
{
struct setup_data *data;
kaslr_setup_data.type = SETUP_KASLR;
kaslr_setup_data.len = 1;
kaslr_setup_data.next = 0;
kaslr_setup_data.data[0] = enabled;
data = (struct setup_data *)(unsigned long)params->hdr.setup_data;
while (data && data->next)
data = (struct setup_data *)(unsigned long)data->next;
if (data)
data->next = (unsigned long)&kaslr_setup_data;
else
params->hdr.setup_data = (unsigned long)&kaslr_setup_data;
}
unsigned char *choose_kernel_location(struct boot_params *params,
unsigned char *input,
unsigned long input_size, unsigned long input_size,
unsigned char *output, unsigned char *output,
unsigned long output_size) unsigned long output_size)
...@@ -335,17 +306,14 @@ unsigned char *choose_kernel_location(struct boot_params *params, ...@@ -335,17 +306,14 @@ unsigned char *choose_kernel_location(struct boot_params *params,
#ifdef CONFIG_HIBERNATION #ifdef CONFIG_HIBERNATION
if (!cmdline_find_option_bool("kaslr")) { if (!cmdline_find_option_bool("kaslr")) {
debug_putstr("KASLR disabled by default...\n"); debug_putstr("KASLR disabled by default...\n");
add_kaslr_setup_data(params, 0);
goto out; goto out;
} }
#else #else
if (cmdline_find_option_bool("nokaslr")) { if (cmdline_find_option_bool("nokaslr")) {
debug_putstr("KASLR disabled by cmdline...\n"); debug_putstr("KASLR disabled by cmdline...\n");
add_kaslr_setup_data(params, 0);
goto out; goto out;
} }
#endif #endif
add_kaslr_setup_data(params, 1);
/* Record the various known unsafe memory ranges. */ /* Record the various known unsafe memory ranges. */
mem_avoid_init((unsigned long)input, input_size, mem_avoid_init((unsigned long)input, input_size,
......
...@@ -401,8 +401,7 @@ asmlinkage __visible void *decompress_kernel(void *rmode, memptr heap, ...@@ -401,8 +401,7 @@ asmlinkage __visible void *decompress_kernel(void *rmode, memptr heap,
* the entire decompressed kernel plus relocation table, or the * the entire decompressed kernel plus relocation table, or the
* entire decompressed kernel plus .bss and .brk sections. * entire decompressed kernel plus .bss and .brk sections.
*/ */
output = choose_kernel_location(real_mode, input_data, input_len, output = choose_kernel_location(input_data, input_len, output,
output,
output_len > run_size ? output_len output_len > run_size ? output_len
: run_size); : run_size);
......
...@@ -57,8 +57,7 @@ int cmdline_find_option_bool(const char *option); ...@@ -57,8 +57,7 @@ int cmdline_find_option_bool(const char *option);
#if CONFIG_RANDOMIZE_BASE #if CONFIG_RANDOMIZE_BASE
/* aslr.c */ /* aslr.c */
unsigned char *choose_kernel_location(struct boot_params *params, unsigned char *choose_kernel_location(unsigned char *input,
unsigned char *input,
unsigned long input_size, unsigned long input_size,
unsigned char *output, unsigned char *output,
unsigned long output_size); unsigned long output_size);
...@@ -66,8 +65,7 @@ unsigned char *choose_kernel_location(struct boot_params *params, ...@@ -66,8 +65,7 @@ unsigned char *choose_kernel_location(struct boot_params *params,
bool has_cpuflag(int flag); bool has_cpuflag(int flag);
#else #else
static inline static inline
unsigned char *choose_kernel_location(struct boot_params *params, unsigned char *choose_kernel_location(unsigned char *input,
unsigned char *input,
unsigned long input_size, unsigned long input_size,
unsigned char *output, unsigned char *output,
unsigned long output_size) unsigned long output_size)
......
...@@ -370,7 +370,7 @@ static inline void drop_fpu(struct task_struct *tsk) ...@@ -370,7 +370,7 @@ static inline void drop_fpu(struct task_struct *tsk)
preempt_disable(); preempt_disable();
tsk->thread.fpu_counter = 0; tsk->thread.fpu_counter = 0;
__drop_fpu(tsk); __drop_fpu(tsk);
clear_used_math(); clear_stopped_child_used_math(tsk);
preempt_enable(); preempt_enable();
} }
......
...@@ -51,8 +51,6 @@ extern int devmem_is_allowed(unsigned long pagenr); ...@@ -51,8 +51,6 @@ extern int devmem_is_allowed(unsigned long pagenr);
extern unsigned long max_low_pfn_mapped; extern unsigned long max_low_pfn_mapped;
extern unsigned long max_pfn_mapped; extern unsigned long max_pfn_mapped;
extern bool kaslr_enabled;
static inline phys_addr_t get_max_mapped(void) static inline phys_addr_t get_max_mapped(void)
{ {
return (phys_addr_t)max_pfn_mapped << PAGE_SHIFT; return (phys_addr_t)max_pfn_mapped << PAGE_SHIFT;
......
...@@ -7,7 +7,6 @@ ...@@ -7,7 +7,6 @@
#define SETUP_DTB 2 #define SETUP_DTB 2
#define SETUP_PCI 3 #define SETUP_PCI 3
#define SETUP_EFI 4 #define SETUP_EFI 4
#define SETUP_KASLR 5
/* ram_size flags */ /* ram_size flags */
#define RAMDISK_IMAGE_START_MASK 0x07FF #define RAMDISK_IMAGE_START_MASK 0x07FF
......
...@@ -1337,6 +1337,26 @@ static int __init dmi_ignore_irq0_timer_override(const struct dmi_system_id *d) ...@@ -1337,6 +1337,26 @@ static int __init dmi_ignore_irq0_timer_override(const struct dmi_system_id *d)
return 0; return 0;
} }
/*
* ACPI offers an alternative platform interface model that removes
* ACPI hardware requirements for platforms that do not implement
* the PC Architecture.
*
* We initialize the Hardware-reduced ACPI model here:
*/
static void __init acpi_reduced_hw_init(void)
{
if (acpi_gbl_reduced_hardware) {
/*
* Override x86_init functions and bypass legacy pic
* in Hardware-reduced ACPI mode
*/
x86_init.timers.timer_init = x86_init_noop;
x86_init.irqs.pre_vector_init = x86_init_noop;
legacy_pic = &null_legacy_pic;
}
}
/* /*
* If your system is blacklisted here, but you find that acpi=force * If your system is blacklisted here, but you find that acpi=force
* works for you, please contact linux-acpi@vger.kernel.org * works for you, please contact linux-acpi@vger.kernel.org
...@@ -1536,6 +1556,11 @@ int __init early_acpi_boot_init(void) ...@@ -1536,6 +1556,11 @@ int __init early_acpi_boot_init(void)
*/ */
early_acpi_process_madt(); early_acpi_process_madt();
/*
* Hardware-reduced ACPI mode initialization:
*/
acpi_reduced_hw_init();
return 0; return 0;
} }
......
...@@ -37,10 +37,12 @@ static const struct apic apic_numachip; ...@@ -37,10 +37,12 @@ static const struct apic apic_numachip;
static unsigned int get_apic_id(unsigned long x) static unsigned int get_apic_id(unsigned long x)
{ {
unsigned long value; unsigned long value;
unsigned int id; unsigned int id = (x >> 24) & 0xff;
if (static_cpu_has_safe(X86_FEATURE_NODEID_MSR)) {
rdmsrl(MSR_FAM10H_NODE_ID, value); rdmsrl(MSR_FAM10H_NODE_ID, value);
id = ((x >> 24) & 0xffU) | ((value << 2) & 0xff00U); id |= (value << 2) & 0xff00;
}
return id; return id;
} }
...@@ -155,10 +157,18 @@ static int __init numachip_probe(void) ...@@ -155,10 +157,18 @@ static int __init numachip_probe(void)
static void fixup_cpu_id(struct cpuinfo_x86 *c, int node) static void fixup_cpu_id(struct cpuinfo_x86 *c, int node)
{ {
if (c->phys_proc_id != node) { u64 val;
c->phys_proc_id = node; u32 nodes = 1;
per_cpu(cpu_llc_id, smp_processor_id()) = node;
this_cpu_write(cpu_llc_id, node);
/* Account for nodes per socket in multi-core-module processors */
if (static_cpu_has_safe(X86_FEATURE_NODEID_MSR)) {
rdmsrl(MSR_FAM10H_NODE_ID, val);
nodes = ((val >> 3) & 7) + 1;
} }
c->phys_proc_id = node / nodes;
} }
static int __init numachip_system_init(void) static int __init numachip_system_init(void)
......
...@@ -47,13 +47,21 @@ do { \ ...@@ -47,13 +47,21 @@ do { \
#ifdef CONFIG_RANDOMIZE_BASE #ifdef CONFIG_RANDOMIZE_BASE
static unsigned long module_load_offset; static unsigned long module_load_offset;
static int randomize_modules = 1;
/* Mutex protects the module_load_offset. */ /* Mutex protects the module_load_offset. */
static DEFINE_MUTEX(module_kaslr_mutex); static DEFINE_MUTEX(module_kaslr_mutex);
static int __init parse_nokaslr(char *p)
{
randomize_modules = 0;
return 0;
}
early_param("nokaslr", parse_nokaslr);
static unsigned long int get_module_load_offset(void) static unsigned long int get_module_load_offset(void)
{ {
if (kaslr_enabled) { if (randomize_modules) {
mutex_lock(&module_kaslr_mutex); mutex_lock(&module_kaslr_mutex);
/* /*
* Calculate the module_load_offset the first time this * Calculate the module_load_offset the first time this
......
...@@ -122,8 +122,6 @@ ...@@ -122,8 +122,6 @@
unsigned long max_low_pfn_mapped; unsigned long max_low_pfn_mapped;
unsigned long max_pfn_mapped; unsigned long max_pfn_mapped;
bool __read_mostly kaslr_enabled = false;
#ifdef CONFIG_DMI #ifdef CONFIG_DMI
RESERVE_BRK(dmi_alloc, 65536); RESERVE_BRK(dmi_alloc, 65536);
#endif #endif
...@@ -427,11 +425,6 @@ static void __init reserve_initrd(void) ...@@ -427,11 +425,6 @@ static void __init reserve_initrd(void)
} }
#endif /* CONFIG_BLK_DEV_INITRD */ #endif /* CONFIG_BLK_DEV_INITRD */
static void __init parse_kaslr_setup(u64 pa_data, u32 data_len)
{
kaslr_enabled = (bool)(pa_data + sizeof(struct setup_data));
}
static void __init parse_setup_data(void) static void __init parse_setup_data(void)
{ {
struct setup_data *data; struct setup_data *data;
...@@ -457,9 +450,6 @@ static void __init parse_setup_data(void) ...@@ -457,9 +450,6 @@ static void __init parse_setup_data(void)
case SETUP_EFI: case SETUP_EFI:
parse_efi_setup(pa_data, data_len); parse_efi_setup(pa_data, data_len);
break; break;
case SETUP_KASLR:
parse_kaslr_setup(pa_data, data_len);
break;
default: default:
break; break;
} }
...@@ -842,14 +832,10 @@ static void __init trim_low_memory_range(void) ...@@ -842,14 +832,10 @@ static void __init trim_low_memory_range(void)
static int static int
dump_kernel_offset(struct notifier_block *self, unsigned long v, void *p) dump_kernel_offset(struct notifier_block *self, unsigned long v, void *p)
{ {
if (kaslr_enabled) pr_emerg("Kernel Offset: 0x%lx from 0x%lx "
pr_emerg("Kernel Offset: 0x%lx from 0x%lx (relocation range: 0x%lx-0x%lx)\n", "(relocation range: 0x%lx-0x%lx)\n",
(unsigned long)&_text - __START_KERNEL, (unsigned long)&_text - __START_KERNEL, __START_KERNEL,
__START_KERNEL, __START_KERNEL_map, MODULES_VADDR-1);
__START_KERNEL_map,
MODULES_VADDR-1);
else
pr_emerg("Kernel Offset: disabled\n");
return 0; return 0;
} }
......
...@@ -384,7 +384,7 @@ dotraplinkage void do_bounds(struct pt_regs *regs, long error_code) ...@@ -384,7 +384,7 @@ dotraplinkage void do_bounds(struct pt_regs *regs, long error_code)
goto exit; goto exit;
conditional_sti(regs); conditional_sti(regs);
if (!user_mode(regs)) if (!user_mode_vm(regs))
die("bounds", regs, error_code); die("bounds", regs, error_code);
if (!cpu_feature_enabled(X86_FEATURE_MPX)) { if (!cpu_feature_enabled(X86_FEATURE_MPX)) {
...@@ -637,7 +637,7 @@ dotraplinkage void do_debug(struct pt_regs *regs, long error_code) ...@@ -637,7 +637,7 @@ dotraplinkage void do_debug(struct pt_regs *regs, long error_code)
* then it's very likely the result of an icebp/int01 trap. * then it's very likely the result of an icebp/int01 trap.
* User wants a sigtrap for that. * User wants a sigtrap for that.
*/ */
if (!dr6 && user_mode(regs)) if (!dr6 && user_mode_vm(regs))
user_icebp = 1; user_icebp = 1;
/* Catch kmemcheck conditions first of all! */ /* Catch kmemcheck conditions first of all! */
......
...@@ -379,7 +379,7 @@ int __restore_xstate_sig(void __user *buf, void __user *buf_fx, int size) ...@@ -379,7 +379,7 @@ int __restore_xstate_sig(void __user *buf, void __user *buf_fx, int size)
* thread's fpu state, reconstruct fxstate from the fsave * thread's fpu state, reconstruct fxstate from the fsave
* header. Sanitize the copied state etc. * header. Sanitize the copied state etc.
*/ */
struct xsave_struct *xsave = &tsk->thread.fpu.state->xsave; struct fpu *fpu = &tsk->thread.fpu;
struct user_i387_ia32_struct env; struct user_i387_ia32_struct env;
int err = 0; int err = 0;
...@@ -393,14 +393,15 @@ int __restore_xstate_sig(void __user *buf, void __user *buf_fx, int size) ...@@ -393,14 +393,15 @@ int __restore_xstate_sig(void __user *buf, void __user *buf_fx, int size)
*/ */
drop_fpu(tsk); drop_fpu(tsk);
if (__copy_from_user(xsave, buf_fx, state_size) || if (__copy_from_user(&fpu->state->xsave, buf_fx, state_size) ||
__copy_from_user(&env, buf, sizeof(env))) { __copy_from_user(&env, buf, sizeof(env))) {
fpu_finit(fpu);
err = -1; err = -1;
} else { } else {
sanitize_restored_xstate(tsk, &env, xstate_bv, fx_only); sanitize_restored_xstate(tsk, &env, xstate_bv, fx_only);
set_used_math();
} }
set_used_math();
if (use_eager_fpu()) { if (use_eager_fpu()) {
preempt_disable(); preempt_disable();
math_state_restore(); math_state_restore();
......
...@@ -17,6 +17,7 @@ ...@@ -17,6 +17,7 @@
.text .text
.globl __kernel_sigreturn .globl __kernel_sigreturn
.type __kernel_sigreturn,@function .type __kernel_sigreturn,@function
nop /* this guy is needed for .LSTARTFDEDLSI1 below (watch for HACK) */
ALIGN ALIGN
__kernel_sigreturn: __kernel_sigreturn:
.LSTART_sigreturn: .LSTART_sigreturn:
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment