Commit 93946a33 authored by Ashok Raj's avatar Ashok Raj Committed by Borislav Petkov

x86/microcode: Update late microcode in parallel

Microcode update was changed to be serialized due to restrictions after
Spectre days. Updating serially on a large multi-socket system can be
painful since it is being done on one CPU at a time.

Cloud customers have expressed discontent as services disappear for
a prolonged time. The restriction is that only one core (or only one
thread of a core in the case of an SMT system) goes through the update
while other cores (or respectively, SMT threads) are quiesced.

Do the microcode update only on the first thread of each core while
other siblings simply wait for this to complete.

 [ bp: Simplify, massage, cleanup comments. ]
Signed-off-by: default avatarAshok Raj <ashok.raj@intel.com>
Signed-off-by: default avatarMihai Carabas <mihai.carabas@oracle.com>
Signed-off-by: default avatarBorislav Petkov <bp@suse.de>
Cc: Boris Ostrovsky <boris.ostrovsky@oracle.com>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Jon Grimm <Jon.Grimm@amd.com>
Cc: kanth.ghatraju@oracle.com
Cc: konrad.wilk@oracle.com
Cc: patrick.colp@oracle.com
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Tom Lendacky <thomas.lendacky@amd.com>
Cc: x86-ml <x86@kernel.org>
Link: https://lkml.kernel.org/r/1566506627-16536-2-git-send-email-mihai.carabas@oracle.com
parent 2b730952
...@@ -63,11 +63,6 @@ LIST_HEAD(microcode_cache); ...@@ -63,11 +63,6 @@ LIST_HEAD(microcode_cache);
*/ */
static DEFINE_MUTEX(microcode_mutex); static DEFINE_MUTEX(microcode_mutex);
/*
* Serialize late loading so that CPUs get updated one-by-one.
*/
static DEFINE_RAW_SPINLOCK(update_lock);
struct ucode_cpu_info ucode_cpu_info[NR_CPUS]; struct ucode_cpu_info ucode_cpu_info[NR_CPUS];
struct cpu_info_ctx { struct cpu_info_ctx {
...@@ -566,11 +561,18 @@ static int __reload_late(void *info) ...@@ -566,11 +561,18 @@ static int __reload_late(void *info)
if (__wait_for_cpus(&late_cpus_in, NSEC_PER_SEC)) if (__wait_for_cpus(&late_cpus_in, NSEC_PER_SEC))
return -1; return -1;
raw_spin_lock(&update_lock); /*
apply_microcode_local(&err); * On an SMT system, it suffices to load the microcode on one sibling of
raw_spin_unlock(&update_lock); * the core because the microcode engine is shared between the threads.
* Synchronization still needs to take place so that no concurrent
* loading attempts happen on multiple threads of an SMT core. See
* below.
*/
if (cpumask_first(topology_sibling_cpumask(cpu)) == cpu)
apply_microcode_local(&err);
else
goto wait_for_siblings;
/* siblings return UCODE_OK because their engine got updated already */
if (err > UCODE_NFOUND) { if (err > UCODE_NFOUND) {
pr_warn("Error reloading microcode on CPU %d\n", cpu); pr_warn("Error reloading microcode on CPU %d\n", cpu);
ret = -1; ret = -1;
...@@ -578,14 +580,18 @@ static int __reload_late(void *info) ...@@ -578,14 +580,18 @@ static int __reload_late(void *info)
ret = 1; ret = 1;
} }
wait_for_siblings:
if (__wait_for_cpus(&late_cpus_out, NSEC_PER_SEC))
panic("Timeout during microcode update!\n");
/* /*
* Increase the wait timeout to a safe value here since we're * At least one thread has completed update on each core.
* serializing the microcode update and that could take a while on a * For others, simply call the update to make sure the
* large number of CPUs. And that is fine as the *actual* timeout will * per-cpu cpuinfo can be updated with right microcode
* be determined by the last CPU finished updating and thus cut short. * revision.
*/ */
if (__wait_for_cpus(&late_cpus_out, NSEC_PER_SEC * num_online_cpus())) if (cpumask_first(topology_sibling_cpumask(cpu)) != cpu)
panic("Timeout during microcode update!\n"); apply_microcode_local(&err);
return ret; return ret;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment