[patch V2 10/37] x86/microcode/intel: Rip out mixed stepping support for Intel CPUs
From: Thomas Gleixner
Date: Sat Aug 12 2023 - 16:01:22 EST
From: Ashok Raj <ashok.raj@xxxxxxxxx>
Mixed steppings aren't supported on Intel CPUs. Only one patch is required
for the entire system. The caching of micro code blobs which match the
family and model is therefore pointless and in fact it is disfunctional as
CPU hotplug updates use only a single microcode blob, i.e. the one where
*intel_ucode_patch points to.
Remove the microcode cache and make it an AMD local feature.
[ tglx: Save only at the end. Otherwise random microcode ends up in the
pointer for early loading ]
Originally-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
Signed-off-by: Ashok Raj <ashok.raj@xxxxxxxxx>
Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
---
V2: Fix the bogus condition - Borislav
---
arch/x86/kernel/cpu/microcode/amd.c | 10 ++
arch/x86/kernel/cpu/microcode/core.c | 2
arch/x86/kernel/cpu/microcode/intel.c | 135 +++++--------------------------
arch/x86/kernel/cpu/microcode/internal.h | 10 --
4 files changed, 36 insertions(+), 121 deletions(-)
---
--- a/arch/x86/kernel/cpu/microcode/amd.c
+++ b/arch/x86/kernel/cpu/microcode/amd.c
@@ -37,6 +37,16 @@
#include "internal.h"
+struct ucode_patch {
+ struct list_head plist;
+ void *data;
+ unsigned int size;
+ u32 patch_id;
+ u16 equiv_cpu;
+};
+
+static LIST_HEAD(microcode_cache);
+
#define UCODE_MAGIC 0x00414d44
#define UCODE_EQUIV_CPU_TABLE_TYPE 0x00000000
#define UCODE_UCODE_TYPE 0x00000001
--- a/arch/x86/kernel/cpu/microcode/core.c
+++ b/arch/x86/kernel/cpu/microcode/core.c
@@ -46,8 +46,6 @@ static bool dis_ucode_ldr = true;
bool initrd_gone;
-LIST_HEAD(microcode_cache);
-
/*
* Synchronization.
*
--- a/arch/x86/kernel/cpu/microcode/intel.c
+++ b/arch/x86/kernel/cpu/microcode/intel.c
@@ -33,10 +33,10 @@
static const char ucode_path[] = "kernel/x86/microcode/GenuineIntel.bin";
/* Current microcode patch used in early patching on the APs. */
-static struct microcode_intel *intel_ucode_patch;
+static struct microcode_intel *intel_ucode_patch __read_mostly;
/* last level cache size per core */
-static int llc_size_per_core;
+static int llc_size_per_core __ro_after_init;
/* microcode format is extended from prescott processors */
struct extended_signature {
@@ -253,81 +253,26 @@ static int has_newer_microcode(void *mc,
return intel_find_matching_signature(mc, csig, cpf);
}
-static struct ucode_patch *memdup_patch(void *data, unsigned int size)
+static void save_microcode_patch(void *data, unsigned int size)
{
- struct ucode_patch *p;
+ struct microcode_header_intel *p;
- p = kzalloc(sizeof(struct ucode_patch), GFP_KERNEL);
- if (!p)
- return NULL;
-
- p->data = kmemdup(data, size, GFP_KERNEL);
- if (!p->data) {
- kfree(p);
- return NULL;
- }
-
- return p;
-}
-
-static void save_microcode_patch(struct ucode_cpu_info *uci, void *data, unsigned int size)
-{
- struct microcode_header_intel *mc_hdr, *mc_saved_hdr;
- struct ucode_patch *iter, *tmp, *p = NULL;
- bool prev_found = false;
- unsigned int sig, pf;
-
- mc_hdr = (struct microcode_header_intel *)data;
-
- list_for_each_entry_safe(iter, tmp, µcode_cache, plist) {
- mc_saved_hdr = (struct microcode_header_intel *)iter->data;
- sig = mc_saved_hdr->sig;
- pf = mc_saved_hdr->pf;
-
- if (intel_find_matching_signature(data, sig, pf)) {
- prev_found = true;
-
- if (mc_hdr->rev <= mc_saved_hdr->rev)
- continue;
-
- p = memdup_patch(data, size);
- if (!p)
- pr_err("Error allocating buffer %p\n", data);
- else {
- list_replace(&iter->plist, &p->plist);
- kfree(iter->data);
- kfree(iter);
- }
- }
- }
-
- /*
- * There weren't any previous patches found in the list cache; save the
- * newly found.
- */
- if (!prev_found) {
- p = memdup_patch(data, size);
- if (!p)
- pr_err("Error allocating buffer for %p\n", data);
- else
- list_add_tail(&p->plist, µcode_cache);
- }
+ kfree(intel_ucode_patch);
+ intel_ucode_patch = NULL;
+ p = kmemdup(data, size, GFP_KERNEL);
if (!p)
return;
- if (!intel_find_matching_signature(p->data, uci->cpu_sig.sig, uci->cpu_sig.pf))
- return;
-
/*
* Save for early loading. On 32-bit, that needs to be a physical
* address as the APs are running from physical addresses, before
* paging has been enabled.
*/
if (IS_ENABLED(CONFIG_X86_32))
- intel_ucode_patch = (struct microcode_intel *)__pa_nodebug(p->data);
+ intel_ucode_patch = (struct microcode_intel *)__pa_nodebug(p);
else
- intel_ucode_patch = p->data;
+ intel_ucode_patch = (struct microcode_intel *)p;
}
/*
@@ -339,6 +284,7 @@ scan_microcode(void *data, size_t size,
{
struct microcode_header_intel *mc_header;
struct microcode_intel *patch = NULL;
+ u32 cur_rev = uci->cpu_sig.rev;
unsigned int mc_size;
while (size) {
@@ -348,8 +294,7 @@ scan_microcode(void *data, size_t size,
mc_header = (struct microcode_header_intel *)data;
mc_size = get_totalsize(mc_header);
- if (!mc_size ||
- mc_size > size ||
+ if (!mc_size || mc_size > size ||
intel_microcode_sanity_check(data, false, MC_HEADER_TYPE_MICROCODE) < 0)
break;
@@ -361,31 +306,16 @@ scan_microcode(void *data, size_t size,
continue;
}
- if (save) {
- save_microcode_patch(uci, data, mc_size);
+ /* BSP scan: Check whether there is newer microcode */
+ if (!save && cur_rev >= mc_header->rev)
goto next;
- }
-
- if (!patch) {
- if (!has_newer_microcode(data,
- uci->cpu_sig.sig,
- uci->cpu_sig.pf,
- uci->cpu_sig.rev))
- goto next;
-
- } else {
- struct microcode_header_intel *phdr = &patch->hdr;
-
- if (!has_newer_microcode(data,
- phdr->sig,
- phdr->pf,
- phdr->rev))
- goto next;
- }
+ /* Save scan: Check whether there is newer or matching microcode */
+ if (save && cur_rev != mc_header->rev)
+ goto next;
- /* We have a newer patch, save it. */
patch = data;
+ cur_rev = mc_header->rev;
next:
data += mc_size;
@@ -394,6 +324,9 @@ scan_microcode(void *data, size_t size,
if (size)
return NULL;
+ if (save && patch)
+ save_microcode_patch(patch, mc_size);
+
return patch;
}
@@ -612,26 +545,10 @@ void load_ucode_intel_ap(void)
apply_microcode_early(&uci, true);
}
-static struct microcode_intel *find_patch(struct ucode_cpu_info *uci)
+/* Accessor for microcode pointer */
+static struct microcode_intel *ucode_get_patch(void)
{
- struct microcode_header_intel *phdr;
- struct ucode_patch *iter, *tmp;
-
- list_for_each_entry_safe(iter, tmp, µcode_cache, plist) {
-
- phdr = (struct microcode_header_intel *)iter->data;
-
- if (phdr->rev <= uci->cpu_sig.rev)
- continue;
-
- if (!intel_find_matching_signature(phdr,
- uci->cpu_sig.sig,
- uci->cpu_sig.pf))
- continue;
-
- return iter->data;
- }
- return NULL;
+ return intel_ucode_patch;
}
void reload_ucode_intel(void)
@@ -641,7 +558,7 @@ void reload_ucode_intel(void)
intel_cpu_collect_info(&uci);
- p = find_patch(&uci);
+ p = ucode_get_patch();
if (!p)
return;
@@ -685,7 +602,7 @@ static enum ucode_state apply_microcode_
return UCODE_ERROR;
/* Look for a newer patch in our cache: */
- mc = find_patch(uci);
+ mc = ucode_get_patch();
if (!mc) {
mc = uci->mc;
if (!mc)
@@ -814,7 +731,7 @@ static enum ucode_state generic_load_mic
uci->mc = (struct microcode_intel *)new_mc;
/* Save for CPU hotplug */
- save_microcode_patch(uci, new_mc, new_mc_size);
+ save_microcode_patch(new_mc, new_mc_size);
pr_debug("CPU%d found a matching microcode update with version 0x%x (current=0x%x)\n",
cpu, new_rev, uci->cpu_sig.rev);
--- a/arch/x86/kernel/cpu/microcode/internal.h
+++ b/arch/x86/kernel/cpu/microcode/internal.h
@@ -8,16 +8,6 @@
#include <asm/cpu.h>
#include <asm/microcode.h>
-struct ucode_patch {
- struct list_head plist;
- void *data; /* Intel uses only this one */
- unsigned int size;
- u32 patch_id;
- u16 equiv_cpu;
-};
-
-extern struct list_head microcode_cache;
-
struct device;
enum ucode_state {