Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Intel CPU Microcode Update Driver for Linux
4 *
5 * Copyright (C) 2000-2006 Tigran Aivazian <aivazian.tigran@gmail.com>
6 * 2006 Shaohua Li <shaohua.li@intel.com>
7 *
8 * Intel CPU microcode early update for Linux
9 *
10 * Copyright (C) 2012 Fenghua Yu <fenghua.yu@intel.com>
11 * H Peter Anvin" <hpa@zytor.com>
12 */
13#define pr_fmt(fmt) "microcode: " fmt
14#include <linux/earlycpio.h>
15#include <linux/firmware.h>
16#include <linux/uaccess.h>
17#include <linux/initrd.h>
18#include <linux/kernel.h>
19#include <linux/slab.h>
20#include <linux/cpu.h>
21#include <linux/uio.h>
22#include <linux/mm.h>
23
24#include <asm/cpu_device_id.h>
25#include <asm/processor.h>
26#include <asm/tlbflush.h>
27#include <asm/setup.h>
28#include <asm/msr.h>
29
30#include "internal.h"
31
32static const char ucode_path[] = "kernel/x86/microcode/GenuineIntel.bin";
33
34#define UCODE_BSP_LOADED ((struct microcode_intel *)0x1UL)
35
36/* Current microcode patch used in early patching on the APs. */
37static struct microcode_intel *ucode_patch_va __read_mostly;
38static struct microcode_intel *ucode_patch_late __read_mostly;
39
40/* last level cache size per core */
41static unsigned int llc_size_per_core __ro_after_init;
42
43/* microcode format is extended from prescott processors */
44struct extended_signature {
45 unsigned int sig;
46 unsigned int pf;
47 unsigned int cksum;
48};
49
50struct extended_sigtable {
51 unsigned int count;
52 unsigned int cksum;
53 unsigned int reserved[3];
54 struct extended_signature sigs[];
55};
56
57#define DEFAULT_UCODE_TOTALSIZE (DEFAULT_UCODE_DATASIZE + MC_HEADER_SIZE)
58#define EXT_HEADER_SIZE (sizeof(struct extended_sigtable))
59#define EXT_SIGNATURE_SIZE (sizeof(struct extended_signature))
60
61static inline unsigned int get_totalsize(struct microcode_header_intel *hdr)
62{
63 return hdr->datasize ? hdr->totalsize : DEFAULT_UCODE_TOTALSIZE;
64}
65
66static inline unsigned int exttable_size(struct extended_sigtable *et)
67{
68 return et->count * EXT_SIGNATURE_SIZE + EXT_HEADER_SIZE;
69}
70
71void intel_collect_cpu_info(struct cpu_signature *sig)
72{
73 sig->sig = cpuid_eax(1);
74 sig->pf = 0;
75 sig->rev = intel_get_microcode_revision();
76
77 if (x86_model(sig->sig) >= 5 || x86_family(sig->sig) > 6) {
78 unsigned int val[2];
79
80 /* get processor flags from MSR 0x17 */
81 native_rdmsr(MSR_IA32_PLATFORM_ID, val[0], val[1]);
82 sig->pf = 1 << ((val[1] >> 18) & 7);
83 }
84}
85EXPORT_SYMBOL_GPL(intel_collect_cpu_info);
86
87static inline bool cpu_signatures_match(struct cpu_signature *s1, unsigned int sig2,
88 unsigned int pf2)
89{
90 if (s1->sig != sig2)
91 return false;
92
93 /* Processor flags are either both 0 or they intersect. */
94 return ((!s1->pf && !pf2) || (s1->pf & pf2));
95}
96
97bool intel_find_matching_signature(void *mc, struct cpu_signature *sig)
98{
99 struct microcode_header_intel *mc_hdr = mc;
100 struct extended_signature *ext_sig;
101 struct extended_sigtable *ext_hdr;
102 int i;
103
104 if (cpu_signatures_match(sig, mc_hdr->sig, mc_hdr->pf))
105 return true;
106
107 /* Look for ext. headers: */
108 if (get_totalsize(mc_hdr) <= intel_microcode_get_datasize(mc_hdr) + MC_HEADER_SIZE)
109 return false;
110
111 ext_hdr = mc + intel_microcode_get_datasize(mc_hdr) + MC_HEADER_SIZE;
112 ext_sig = (void *)ext_hdr + EXT_HEADER_SIZE;
113
114 for (i = 0; i < ext_hdr->count; i++) {
115 if (cpu_signatures_match(sig, ext_sig->sig, ext_sig->pf))
116 return true;
117 ext_sig++;
118 }
119 return 0;
120}
121EXPORT_SYMBOL_GPL(intel_find_matching_signature);
122
123/**
124 * intel_microcode_sanity_check() - Sanity check microcode file.
125 * @mc: Pointer to the microcode file contents.
126 * @print_err: Display failure reason if true, silent if false.
127 * @hdr_type: Type of file, i.e. normal microcode file or In Field Scan file.
128 * Validate if the microcode header type matches with the type
129 * specified here.
130 *
131 * Validate certain header fields and verify if computed checksum matches
132 * with the one specified in the header.
133 *
134 * Return: 0 if the file passes all the checks, -EINVAL if any of the checks
135 * fail.
136 */
137int intel_microcode_sanity_check(void *mc, bool print_err, int hdr_type)
138{
139 unsigned long total_size, data_size, ext_table_size;
140 struct microcode_header_intel *mc_header = mc;
141 struct extended_sigtable *ext_header = NULL;
142 u32 sum, orig_sum, ext_sigcount = 0, i;
143 struct extended_signature *ext_sig;
144
145 total_size = get_totalsize(mc_header);
146 data_size = intel_microcode_get_datasize(mc_header);
147
148 if (data_size + MC_HEADER_SIZE > total_size) {
149 if (print_err)
150 pr_err("Error: bad microcode data file size.\n");
151 return -EINVAL;
152 }
153
154 if (mc_header->ldrver != 1 || mc_header->hdrver != hdr_type) {
155 if (print_err)
156 pr_err("Error: invalid/unknown microcode update format. Header type %d\n",
157 mc_header->hdrver);
158 return -EINVAL;
159 }
160
161 ext_table_size = total_size - (MC_HEADER_SIZE + data_size);
162 if (ext_table_size) {
163 u32 ext_table_sum = 0;
164 u32 *ext_tablep;
165
166 if (ext_table_size < EXT_HEADER_SIZE ||
167 ((ext_table_size - EXT_HEADER_SIZE) % EXT_SIGNATURE_SIZE)) {
168 if (print_err)
169 pr_err("Error: truncated extended signature table.\n");
170 return -EINVAL;
171 }
172
173 ext_header = mc + MC_HEADER_SIZE + data_size;
174 if (ext_table_size != exttable_size(ext_header)) {
175 if (print_err)
176 pr_err("Error: extended signature table size mismatch.\n");
177 return -EFAULT;
178 }
179
180 ext_sigcount = ext_header->count;
181
182 /*
183 * Check extended table checksum: the sum of all dwords that
184 * comprise a valid table must be 0.
185 */
186 ext_tablep = (u32 *)ext_header;
187
188 i = ext_table_size / sizeof(u32);
189 while (i--)
190 ext_table_sum += ext_tablep[i];
191
192 if (ext_table_sum) {
193 if (print_err)
194 pr_warn("Bad extended signature table checksum, aborting.\n");
195 return -EINVAL;
196 }
197 }
198
199 /*
200 * Calculate the checksum of update data and header. The checksum of
201 * valid update data and header including the extended signature table
202 * must be 0.
203 */
204 orig_sum = 0;
205 i = (MC_HEADER_SIZE + data_size) / sizeof(u32);
206 while (i--)
207 orig_sum += ((u32 *)mc)[i];
208
209 if (orig_sum) {
210 if (print_err)
211 pr_err("Bad microcode data checksum, aborting.\n");
212 return -EINVAL;
213 }
214
215 if (!ext_table_size)
216 return 0;
217
218 /*
219 * Check extended signature checksum: 0 => valid.
220 */
221 for (i = 0; i < ext_sigcount; i++) {
222 ext_sig = (void *)ext_header + EXT_HEADER_SIZE +
223 EXT_SIGNATURE_SIZE * i;
224
225 sum = (mc_header->sig + mc_header->pf + mc_header->cksum) -
226 (ext_sig->sig + ext_sig->pf + ext_sig->cksum);
227 if (sum) {
228 if (print_err)
229 pr_err("Bad extended signature checksum, aborting.\n");
230 return -EINVAL;
231 }
232 }
233 return 0;
234}
235EXPORT_SYMBOL_GPL(intel_microcode_sanity_check);
236
237static void update_ucode_pointer(struct microcode_intel *mc)
238{
239 kvfree(ucode_patch_va);
240
241 /*
242 * Save the virtual address for early loading and for eventual free
243 * on late loading.
244 */
245 ucode_patch_va = mc;
246}
247
248static void save_microcode_patch(struct microcode_intel *patch)
249{
250 unsigned int size = get_totalsize(&patch->hdr);
251 struct microcode_intel *mc;
252
253 mc = kvmemdup(patch, size, GFP_KERNEL);
254 if (mc)
255 update_ucode_pointer(mc);
256 else
257 pr_err("Unable to allocate microcode memory size: %u\n", size);
258}
259
260/* Scan blob for microcode matching the boot CPUs family, model, stepping */
261static __init struct microcode_intel *scan_microcode(void *data, size_t size,
262 struct ucode_cpu_info *uci,
263 bool save)
264{
265 struct microcode_header_intel *mc_header;
266 struct microcode_intel *patch = NULL;
267 u32 cur_rev = uci->cpu_sig.rev;
268 unsigned int mc_size;
269
270 for (; size >= sizeof(struct microcode_header_intel); size -= mc_size, data += mc_size) {
271 mc_header = (struct microcode_header_intel *)data;
272
273 mc_size = get_totalsize(mc_header);
274 if (!mc_size || mc_size > size ||
275 intel_microcode_sanity_check(data, false, MC_HEADER_TYPE_MICROCODE) < 0)
276 break;
277
278 if (!intel_find_matching_signature(data, &uci->cpu_sig))
279 continue;
280
281 /*
282 * For saving the early microcode, find the matching revision which
283 * was loaded on the BSP.
284 *
285 * On the BSP during early boot, find a newer revision than
286 * actually loaded in the CPU.
287 */
288 if (save) {
289 if (cur_rev != mc_header->rev)
290 continue;
291 } else if (cur_rev >= mc_header->rev) {
292 continue;
293 }
294
295 patch = data;
296 cur_rev = mc_header->rev;
297 }
298
299 return size ? NULL : patch;
300}
301
302static enum ucode_state __apply_microcode(struct ucode_cpu_info *uci,
303 struct microcode_intel *mc,
304 u32 *cur_rev)
305{
306 u32 rev;
307
308 if (!mc)
309 return UCODE_NFOUND;
310
311 /*
312 * Save us the MSR write below - which is a particular expensive
313 * operation - when the other hyperthread has updated the microcode
314 * already.
315 */
316 *cur_rev = intel_get_microcode_revision();
317 if (*cur_rev >= mc->hdr.rev) {
318 uci->cpu_sig.rev = *cur_rev;
319 return UCODE_OK;
320 }
321
322 /* write microcode via MSR 0x79 */
323 native_wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
324
325 rev = intel_get_microcode_revision();
326 if (rev != mc->hdr.rev)
327 return UCODE_ERROR;
328
329 uci->cpu_sig.rev = rev;
330 return UCODE_UPDATED;
331}
332
333static enum ucode_state apply_microcode_early(struct ucode_cpu_info *uci)
334{
335 struct microcode_intel *mc = uci->mc;
336 u32 cur_rev;
337
338 return __apply_microcode(uci, mc, &cur_rev);
339}
340
341static __init bool load_builtin_intel_microcode(struct cpio_data *cp)
342{
343 unsigned int eax = 1, ebx, ecx = 0, edx;
344 struct firmware fw;
345 char name[30];
346
347 if (IS_ENABLED(CONFIG_X86_32))
348 return false;
349
350 native_cpuid(&eax, &ebx, &ecx, &edx);
351
352 sprintf(name, "intel-ucode/%02x-%02x-%02x",
353 x86_family(eax), x86_model(eax), x86_stepping(eax));
354
355 if (firmware_request_builtin(&fw, name)) {
356 cp->size = fw.size;
357 cp->data = (void *)fw.data;
358 return true;
359 }
360 return false;
361}
362
363static __init struct microcode_intel *get_microcode_blob(struct ucode_cpu_info *uci, bool save)
364{
365 struct cpio_data cp;
366
367 intel_collect_cpu_info(&uci->cpu_sig);
368
369 if (!load_builtin_intel_microcode(&cp))
370 cp = find_microcode_in_initrd(ucode_path);
371
372 if (!(cp.data && cp.size))
373 return NULL;
374
375 return scan_microcode(cp.data, cp.size, uci, save);
376}
377
378/*
379 * Invoked from an early init call to save the microcode blob which was
380 * selected during early boot when mm was not usable. The microcode must be
381 * saved because initrd is going away. It's an early init call so the APs
382 * just can use the pointer and do not have to scan initrd/builtin firmware
383 * again.
384 */
385static int __init save_builtin_microcode(void)
386{
387 struct ucode_cpu_info uci;
388
389 if (xchg(&ucode_patch_va, NULL) != UCODE_BSP_LOADED)
390 return 0;
391
392 if (dis_ucode_ldr || boot_cpu_data.x86_vendor != X86_VENDOR_INTEL)
393 return 0;
394
395 uci.mc = get_microcode_blob(&uci, true);
396 if (uci.mc)
397 save_microcode_patch(uci.mc);
398 return 0;
399}
400early_initcall(save_builtin_microcode);
401
402/* Load microcode on BSP from initrd or builtin blobs */
403void __init load_ucode_intel_bsp(struct early_load_data *ed)
404{
405 struct ucode_cpu_info uci;
406
407 uci.mc = get_microcode_blob(&uci, false);
408 ed->old_rev = uci.cpu_sig.rev;
409
410 if (uci.mc && apply_microcode_early(&uci) == UCODE_UPDATED) {
411 ucode_patch_va = UCODE_BSP_LOADED;
412 ed->new_rev = uci.cpu_sig.rev;
413 }
414}
415
416void load_ucode_intel_ap(void)
417{
418 struct ucode_cpu_info uci;
419
420 uci.mc = ucode_patch_va;
421 if (uci.mc)
422 apply_microcode_early(&uci);
423}
424
425/* Reload microcode on resume */
426void reload_ucode_intel(void)
427{
428 struct ucode_cpu_info uci = { .mc = ucode_patch_va, };
429
430 if (uci.mc)
431 apply_microcode_early(&uci);
432}
433
434static int collect_cpu_info(int cpu_num, struct cpu_signature *csig)
435{
436 intel_collect_cpu_info(csig);
437 return 0;
438}
439
440static enum ucode_state apply_microcode_late(int cpu)
441{
442 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
443 struct microcode_intel *mc = ucode_patch_late;
444 enum ucode_state ret;
445 u32 cur_rev;
446
447 if (WARN_ON_ONCE(smp_processor_id() != cpu))
448 return UCODE_ERROR;
449
450 ret = __apply_microcode(uci, mc, &cur_rev);
451 if (ret != UCODE_UPDATED && ret != UCODE_OK)
452 return ret;
453
454 cpu_data(cpu).microcode = uci->cpu_sig.rev;
455 if (!cpu)
456 boot_cpu_data.microcode = uci->cpu_sig.rev;
457
458 return ret;
459}
460
461static bool ucode_validate_minrev(struct microcode_header_intel *mc_header)
462{
463 int cur_rev = boot_cpu_data.microcode;
464
465 /*
466 * When late-loading, ensure the header declares a minimum revision
467 * required to perform a late-load. The previously reserved field
468 * is 0 in older microcode blobs.
469 */
470 if (!mc_header->min_req_ver) {
471 pr_info("Unsafe microcode update: Microcode header does not specify a required min version\n");
472 return false;
473 }
474
475 /*
476 * Check whether the current revision is either greater or equal to
477 * to the minimum revision specified in the header.
478 */
479 if (cur_rev < mc_header->min_req_ver) {
480 pr_info("Unsafe microcode update: Current revision 0x%x too old\n", cur_rev);
481 pr_info("Current should be at 0x%x or higher. Use early loading instead\n", mc_header->min_req_ver);
482 return false;
483 }
484 return true;
485}
486
487static enum ucode_state parse_microcode_blobs(int cpu, struct iov_iter *iter)
488{
489 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
490 bool is_safe, new_is_safe = false;
491 int cur_rev = uci->cpu_sig.rev;
492 unsigned int curr_mc_size = 0;
493 u8 *new_mc = NULL, *mc = NULL;
494
495 while (iov_iter_count(iter)) {
496 struct microcode_header_intel mc_header;
497 unsigned int mc_size, data_size;
498 u8 *data;
499
500 if (!copy_from_iter_full(&mc_header, sizeof(mc_header), iter)) {
501 pr_err("error! Truncated or inaccessible header in microcode data file\n");
502 goto fail;
503 }
504
505 mc_size = get_totalsize(&mc_header);
506 if (mc_size < sizeof(mc_header)) {
507 pr_err("error! Bad data in microcode data file (totalsize too small)\n");
508 goto fail;
509 }
510 data_size = mc_size - sizeof(mc_header);
511 if (data_size > iov_iter_count(iter)) {
512 pr_err("error! Bad data in microcode data file (truncated file?)\n");
513 goto fail;
514 }
515
516 /* For performance reasons, reuse mc area when possible */
517 if (!mc || mc_size > curr_mc_size) {
518 kvfree(mc);
519 mc = kvmalloc(mc_size, GFP_KERNEL);
520 if (!mc)
521 goto fail;
522 curr_mc_size = mc_size;
523 }
524
525 memcpy(mc, &mc_header, sizeof(mc_header));
526 data = mc + sizeof(mc_header);
527 if (!copy_from_iter_full(data, data_size, iter) ||
528 intel_microcode_sanity_check(mc, true, MC_HEADER_TYPE_MICROCODE) < 0)
529 goto fail;
530
531 if (cur_rev >= mc_header.rev)
532 continue;
533
534 if (!intel_find_matching_signature(mc, &uci->cpu_sig))
535 continue;
536
537 is_safe = ucode_validate_minrev(&mc_header);
538 if (force_minrev && !is_safe)
539 continue;
540
541 kvfree(new_mc);
542 cur_rev = mc_header.rev;
543 new_mc = mc;
544 new_is_safe = is_safe;
545 mc = NULL;
546 }
547
548 if (iov_iter_count(iter))
549 goto fail;
550
551 kvfree(mc);
552 if (!new_mc)
553 return UCODE_NFOUND;
554
555 ucode_patch_late = (struct microcode_intel *)new_mc;
556 return new_is_safe ? UCODE_NEW_SAFE : UCODE_NEW;
557
558fail:
559 kvfree(mc);
560 kvfree(new_mc);
561 return UCODE_ERROR;
562}
563
564static bool is_blacklisted(unsigned int cpu)
565{
566 struct cpuinfo_x86 *c = &cpu_data(cpu);
567
568 /*
569 * Late loading on model 79 with microcode revision less than 0x0b000021
570 * and LLC size per core bigger than 2.5MB may result in a system hang.
571 * This behavior is documented in item BDX90, #334165 (Intel Xeon
572 * Processor E7-8800/4800 v4 Product Family).
573 */
574 if (c->x86_vfm == INTEL_BROADWELL_X &&
575 c->x86_stepping == 0x01 &&
576 llc_size_per_core > 2621440 &&
577 c->microcode < 0x0b000021) {
578 pr_err_once("Erratum BDX90: late loading with revision < 0x0b000021 (0x%x) disabled.\n", c->microcode);
579 pr_err_once("Please consider either early loading through initrd/built-in or a potential BIOS update.\n");
580 return true;
581 }
582
583 return false;
584}
585
586static enum ucode_state request_microcode_fw(int cpu, struct device *device)
587{
588 struct cpuinfo_x86 *c = &cpu_data(cpu);
589 const struct firmware *firmware;
590 struct iov_iter iter;
591 enum ucode_state ret;
592 struct kvec kvec;
593 char name[30];
594
595 if (is_blacklisted(cpu))
596 return UCODE_NFOUND;
597
598 sprintf(name, "intel-ucode/%02x-%02x-%02x",
599 c->x86, c->x86_model, c->x86_stepping);
600
601 if (request_firmware_direct(&firmware, name, device)) {
602 pr_debug("data file %s load failed\n", name);
603 return UCODE_NFOUND;
604 }
605
606 kvec.iov_base = (void *)firmware->data;
607 kvec.iov_len = firmware->size;
608 iov_iter_kvec(&iter, ITER_SOURCE, &kvec, 1, firmware->size);
609 ret = parse_microcode_blobs(cpu, &iter);
610
611 release_firmware(firmware);
612
613 return ret;
614}
615
616static void finalize_late_load(int result)
617{
618 if (!result)
619 update_ucode_pointer(ucode_patch_late);
620 else
621 kvfree(ucode_patch_late);
622 ucode_patch_late = NULL;
623}
624
625static struct microcode_ops microcode_intel_ops = {
626 .request_microcode_fw = request_microcode_fw,
627 .collect_cpu_info = collect_cpu_info,
628 .apply_microcode = apply_microcode_late,
629 .finalize_late_load = finalize_late_load,
630 .use_nmi = IS_ENABLED(CONFIG_X86_64),
631};
632
633static __init void calc_llc_size_per_core(struct cpuinfo_x86 *c)
634{
635 u64 llc_size = c->x86_cache_size * 1024ULL;
636
637 do_div(llc_size, topology_num_cores_per_package());
638 llc_size_per_core = (unsigned int)llc_size;
639}
640
641struct microcode_ops * __init init_intel_microcode(void)
642{
643 struct cpuinfo_x86 *c = &boot_cpu_data;
644
645 if (c->x86_vendor != X86_VENDOR_INTEL || c->x86 < 6 ||
646 cpu_has(c, X86_FEATURE_IA64)) {
647 pr_err("Intel CPU family 0x%x not supported\n", c->x86);
648 return NULL;
649 }
650
651 calc_llc_size_per_core(c);
652
653 return µcode_intel_ops;
654}
1/*
2 * Intel CPU Microcode Update Driver for Linux
3 *
4 * Copyright (C) 2000-2006 Tigran Aivazian <tigran@aivazian.fsnet.co.uk>
5 * 2006 Shaohua Li <shaohua.li@intel.com>
6 *
7 * Intel CPU microcode early update for Linux
8 *
9 * Copyright (C) 2012 Fenghua Yu <fenghua.yu@intel.com>
10 * H Peter Anvin" <hpa@zytor.com>
11 *
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
16 */
17
18/*
19 * This needs to be before all headers so that pr_debug in printk.h doesn't turn
20 * printk calls into no_printk().
21 *
22 *#define DEBUG
23 */
24#define pr_fmt(fmt) "microcode: " fmt
25
26#include <linux/earlycpio.h>
27#include <linux/firmware.h>
28#include <linux/uaccess.h>
29#include <linux/vmalloc.h>
30#include <linux/initrd.h>
31#include <linux/kernel.h>
32#include <linux/slab.h>
33#include <linux/cpu.h>
34#include <linux/mm.h>
35
36#include <asm/microcode_intel.h>
37#include <asm/processor.h>
38#include <asm/tlbflush.h>
39#include <asm/setup.h>
40#include <asm/msr.h>
41
42/*
43 * Temporary microcode blobs pointers storage. We note here the pointers to
44 * microcode blobs we've got from whatever storage (detached initrd, builtin).
45 * Later on, we put those into final storage mc_saved_data.mc_saved.
46 */
47static unsigned long mc_tmp_ptrs[MAX_UCODE_COUNT];
48
49static struct mc_saved_data {
50 unsigned int num_saved;
51 struct microcode_intel **mc_saved;
52} mc_saved_data;
53
54static enum ucode_state
55load_microcode_early(struct microcode_intel **saved,
56 unsigned int num_saved, struct ucode_cpu_info *uci)
57{
58 struct microcode_intel *ucode_ptr, *new_mc = NULL;
59 struct microcode_header_intel *mc_hdr;
60 int new_rev, ret, i;
61
62 new_rev = uci->cpu_sig.rev;
63
64 for (i = 0; i < num_saved; i++) {
65 ucode_ptr = saved[i];
66 mc_hdr = (struct microcode_header_intel *)ucode_ptr;
67
68 ret = has_newer_microcode(ucode_ptr,
69 uci->cpu_sig.sig,
70 uci->cpu_sig.pf,
71 new_rev);
72 if (!ret)
73 continue;
74
75 new_rev = mc_hdr->rev;
76 new_mc = ucode_ptr;
77 }
78
79 if (!new_mc)
80 return UCODE_NFOUND;
81
82 uci->mc = (struct microcode_intel *)new_mc;
83 return UCODE_OK;
84}
85
86static inline void
87copy_ptrs(struct microcode_intel **mc_saved, unsigned long *mc_ptrs,
88 unsigned long off, int num_saved)
89{
90 int i;
91
92 for (i = 0; i < num_saved; i++)
93 mc_saved[i] = (struct microcode_intel *)(mc_ptrs[i] + off);
94}
95
96#ifdef CONFIG_X86_32
97static void
98microcode_phys(struct microcode_intel **mc_saved_tmp, struct mc_saved_data *mcs)
99{
100 int i;
101 struct microcode_intel ***mc_saved;
102
103 mc_saved = (struct microcode_intel ***)__pa_nodebug(&mcs->mc_saved);
104
105 for (i = 0; i < mcs->num_saved; i++) {
106 struct microcode_intel *p;
107
108 p = *(struct microcode_intel **)__pa_nodebug(mcs->mc_saved + i);
109 mc_saved_tmp[i] = (struct microcode_intel *)__pa_nodebug(p);
110 }
111}
112#endif
113
114static enum ucode_state
115load_microcode(struct mc_saved_data *mcs, unsigned long *mc_ptrs,
116 unsigned long offset, struct ucode_cpu_info *uci)
117{
118 struct microcode_intel *mc_saved_tmp[MAX_UCODE_COUNT];
119 unsigned int count = mcs->num_saved;
120
121 if (!mcs->mc_saved) {
122 copy_ptrs(mc_saved_tmp, mc_ptrs, offset, count);
123
124 return load_microcode_early(mc_saved_tmp, count, uci);
125 } else {
126#ifdef CONFIG_X86_32
127 microcode_phys(mc_saved_tmp, mcs);
128 return load_microcode_early(mc_saved_tmp, count, uci);
129#else
130 return load_microcode_early(mcs->mc_saved, count, uci);
131#endif
132 }
133}
134
135/*
136 * Given CPU signature and a microcode patch, this function finds if the
137 * microcode patch has matching family and model with the CPU.
138 */
139static enum ucode_state
140matching_model_microcode(struct microcode_header_intel *mc_header,
141 unsigned long sig)
142{
143 unsigned int fam, model;
144 unsigned int fam_ucode, model_ucode;
145 struct extended_sigtable *ext_header;
146 unsigned long total_size = get_totalsize(mc_header);
147 unsigned long data_size = get_datasize(mc_header);
148 int ext_sigcount, i;
149 struct extended_signature *ext_sig;
150
151 fam = x86_family(sig);
152 model = x86_model(sig);
153
154 fam_ucode = x86_family(mc_header->sig);
155 model_ucode = x86_model(mc_header->sig);
156
157 if (fam == fam_ucode && model == model_ucode)
158 return UCODE_OK;
159
160 /* Look for ext. headers: */
161 if (total_size <= data_size + MC_HEADER_SIZE)
162 return UCODE_NFOUND;
163
164 ext_header = (void *) mc_header + data_size + MC_HEADER_SIZE;
165 ext_sig = (void *)ext_header + EXT_HEADER_SIZE;
166 ext_sigcount = ext_header->count;
167
168 for (i = 0; i < ext_sigcount; i++) {
169 fam_ucode = x86_family(ext_sig->sig);
170 model_ucode = x86_model(ext_sig->sig);
171
172 if (fam == fam_ucode && model == model_ucode)
173 return UCODE_OK;
174
175 ext_sig++;
176 }
177 return UCODE_NFOUND;
178}
179
180static int
181save_microcode(struct mc_saved_data *mcs,
182 struct microcode_intel **mc_saved_src,
183 unsigned int num_saved)
184{
185 int i, j;
186 struct microcode_intel **saved_ptr;
187 int ret;
188
189 if (!num_saved)
190 return -EINVAL;
191
192 /*
193 * Copy new microcode data.
194 */
195 saved_ptr = kcalloc(num_saved, sizeof(struct microcode_intel *), GFP_KERNEL);
196 if (!saved_ptr)
197 return -ENOMEM;
198
199 for (i = 0; i < num_saved; i++) {
200 struct microcode_header_intel *mc_hdr;
201 struct microcode_intel *mc;
202 unsigned long size;
203
204 if (!mc_saved_src[i]) {
205 ret = -EINVAL;
206 goto err;
207 }
208
209 mc = mc_saved_src[i];
210 mc_hdr = &mc->hdr;
211 size = get_totalsize(mc_hdr);
212
213 saved_ptr[i] = kmemdup(mc, size, GFP_KERNEL);
214 if (!saved_ptr[i]) {
215 ret = -ENOMEM;
216 goto err;
217 }
218 }
219
220 /*
221 * Point to newly saved microcode.
222 */
223 mcs->mc_saved = saved_ptr;
224 mcs->num_saved = num_saved;
225
226 return 0;
227
228err:
229 for (j = 0; j <= i; j++)
230 kfree(saved_ptr[j]);
231 kfree(saved_ptr);
232
233 return ret;
234}
235
236/*
237 * A microcode patch in ucode_ptr is saved into mc_saved
238 * - if it has matching signature and newer revision compared to an existing
239 * patch mc_saved.
240 * - or if it is a newly discovered microcode patch.
241 *
242 * The microcode patch should have matching model with CPU.
243 *
244 * Returns: The updated number @num_saved of saved microcode patches.
245 */
246static unsigned int _save_mc(struct microcode_intel **mc_saved,
247 u8 *ucode_ptr, unsigned int num_saved)
248{
249 struct microcode_header_intel *mc_hdr, *mc_saved_hdr;
250 unsigned int sig, pf;
251 int found = 0, i;
252
253 mc_hdr = (struct microcode_header_intel *)ucode_ptr;
254
255 for (i = 0; i < num_saved; i++) {
256 mc_saved_hdr = (struct microcode_header_intel *)mc_saved[i];
257 sig = mc_saved_hdr->sig;
258 pf = mc_saved_hdr->pf;
259
260 if (!find_matching_signature(ucode_ptr, sig, pf))
261 continue;
262
263 found = 1;
264
265 if (mc_hdr->rev <= mc_saved_hdr->rev)
266 continue;
267
268 /*
269 * Found an older ucode saved earlier. Replace it with
270 * this newer one.
271 */
272 mc_saved[i] = (struct microcode_intel *)ucode_ptr;
273 break;
274 }
275
276 /* Newly detected microcode, save it to memory. */
277 if (i >= num_saved && !found)
278 mc_saved[num_saved++] = (struct microcode_intel *)ucode_ptr;
279
280 return num_saved;
281}
282
283/*
284 * Get microcode matching with BSP's model. Only CPUs with the same model as
285 * BSP can stay in the platform.
286 */
287static enum ucode_state __init
288get_matching_model_microcode(unsigned long start, void *data, size_t size,
289 struct mc_saved_data *mcs, unsigned long *mc_ptrs,
290 struct ucode_cpu_info *uci)
291{
292 struct microcode_intel *mc_saved_tmp[MAX_UCODE_COUNT];
293 struct microcode_header_intel *mc_header;
294 unsigned int num_saved = mcs->num_saved;
295 enum ucode_state state = UCODE_OK;
296 unsigned int leftover = size;
297 u8 *ucode_ptr = data;
298 unsigned int mc_size;
299 int i;
300
301 while (leftover && num_saved < ARRAY_SIZE(mc_saved_tmp)) {
302
303 if (leftover < sizeof(mc_header))
304 break;
305
306 mc_header = (struct microcode_header_intel *)ucode_ptr;
307
308 mc_size = get_totalsize(mc_header);
309 if (!mc_size || mc_size > leftover ||
310 microcode_sanity_check(ucode_ptr, 0) < 0)
311 break;
312
313 leftover -= mc_size;
314
315 /*
316 * Since APs with same family and model as the BSP may boot in
317 * the platform, we need to find and save microcode patches
318 * with the same family and model as the BSP.
319 */
320 if (matching_model_microcode(mc_header, uci->cpu_sig.sig) != UCODE_OK) {
321 ucode_ptr += mc_size;
322 continue;
323 }
324
325 num_saved = _save_mc(mc_saved_tmp, ucode_ptr, num_saved);
326
327 ucode_ptr += mc_size;
328 }
329
330 if (leftover) {
331 state = UCODE_ERROR;
332 return state;
333 }
334
335 if (!num_saved) {
336 state = UCODE_NFOUND;
337 return state;
338 }
339
340 for (i = 0; i < num_saved; i++)
341 mc_ptrs[i] = (unsigned long)mc_saved_tmp[i] - start;
342
343 mcs->num_saved = num_saved;
344
345 return state;
346}
347
348static int collect_cpu_info_early(struct ucode_cpu_info *uci)
349{
350 unsigned int val[2];
351 unsigned int family, model;
352 struct cpu_signature csig;
353 unsigned int eax, ebx, ecx, edx;
354
355 csig.sig = 0;
356 csig.pf = 0;
357 csig.rev = 0;
358
359 memset(uci, 0, sizeof(*uci));
360
361 eax = 0x00000001;
362 ecx = 0;
363 native_cpuid(&eax, &ebx, &ecx, &edx);
364 csig.sig = eax;
365
366 family = x86_family(csig.sig);
367 model = x86_model(csig.sig);
368
369 if ((model >= 5) || (family > 6)) {
370 /* get processor flags from MSR 0x17 */
371 native_rdmsr(MSR_IA32_PLATFORM_ID, val[0], val[1]);
372 csig.pf = 1 << ((val[1] >> 18) & 7);
373 }
374 native_wrmsrl(MSR_IA32_UCODE_REV, 0);
375
376 /* As documented in the SDM: Do a CPUID 1 here */
377 sync_core();
378
379 /* get the current revision from MSR 0x8B */
380 native_rdmsr(MSR_IA32_UCODE_REV, val[0], val[1]);
381
382 csig.rev = val[1];
383
384 uci->cpu_sig = csig;
385 uci->valid = 1;
386
387 return 0;
388}
389
390static void show_saved_mc(void)
391{
392#ifdef DEBUG
393 int i, j;
394 unsigned int sig, pf, rev, total_size, data_size, date;
395 struct ucode_cpu_info uci;
396
397 if (!mc_saved_data.num_saved) {
398 pr_debug("no microcode data saved.\n");
399 return;
400 }
401 pr_debug("Total microcode saved: %d\n", mc_saved_data.num_saved);
402
403 collect_cpu_info_early(&uci);
404
405 sig = uci.cpu_sig.sig;
406 pf = uci.cpu_sig.pf;
407 rev = uci.cpu_sig.rev;
408 pr_debug("CPU: sig=0x%x, pf=0x%x, rev=0x%x\n", sig, pf, rev);
409
410 for (i = 0; i < mc_saved_data.num_saved; i++) {
411 struct microcode_header_intel *mc_saved_header;
412 struct extended_sigtable *ext_header;
413 int ext_sigcount;
414 struct extended_signature *ext_sig;
415
416 mc_saved_header = (struct microcode_header_intel *)
417 mc_saved_data.mc_saved[i];
418 sig = mc_saved_header->sig;
419 pf = mc_saved_header->pf;
420 rev = mc_saved_header->rev;
421 total_size = get_totalsize(mc_saved_header);
422 data_size = get_datasize(mc_saved_header);
423 date = mc_saved_header->date;
424
425 pr_debug("mc_saved[%d]: sig=0x%x, pf=0x%x, rev=0x%x, toal size=0x%x, date = %04x-%02x-%02x\n",
426 i, sig, pf, rev, total_size,
427 date & 0xffff,
428 date >> 24,
429 (date >> 16) & 0xff);
430
431 /* Look for ext. headers: */
432 if (total_size <= data_size + MC_HEADER_SIZE)
433 continue;
434
435 ext_header = (void *) mc_saved_header + data_size + MC_HEADER_SIZE;
436 ext_sigcount = ext_header->count;
437 ext_sig = (void *)ext_header + EXT_HEADER_SIZE;
438
439 for (j = 0; j < ext_sigcount; j++) {
440 sig = ext_sig->sig;
441 pf = ext_sig->pf;
442
443 pr_debug("\tExtended[%d]: sig=0x%x, pf=0x%x\n",
444 j, sig, pf);
445
446 ext_sig++;
447 }
448
449 }
450#endif
451}
452
453#ifdef CONFIG_HOTPLUG_CPU
454static DEFINE_MUTEX(x86_cpu_microcode_mutex);
455/*
456 * Save this mc into mc_saved_data. So it will be loaded early when a CPU is
457 * hot added or resumes.
458 *
459 * Please make sure this mc should be a valid microcode patch before calling
460 * this function.
461 */
462int save_mc_for_early(u8 *mc)
463{
464 struct microcode_intel *mc_saved_tmp[MAX_UCODE_COUNT];
465 unsigned int mc_saved_count_init;
466 unsigned int num_saved;
467 struct microcode_intel **mc_saved;
468 int ret = 0;
469 int i;
470
471 /*
472 * Hold hotplug lock so mc_saved_data is not accessed by a CPU in
473 * hotplug.
474 */
475 mutex_lock(&x86_cpu_microcode_mutex);
476
477 mc_saved_count_init = mc_saved_data.num_saved;
478 num_saved = mc_saved_data.num_saved;
479 mc_saved = mc_saved_data.mc_saved;
480
481 if (mc_saved && num_saved)
482 memcpy(mc_saved_tmp, mc_saved,
483 num_saved * sizeof(struct microcode_intel *));
484 /*
485 * Save the microcode patch mc in mc_save_tmp structure if it's a newer
486 * version.
487 */
488 num_saved = _save_mc(mc_saved_tmp, mc, num_saved);
489
490 /*
491 * Save the mc_save_tmp in global mc_saved_data.
492 */
493 ret = save_microcode(&mc_saved_data, mc_saved_tmp, num_saved);
494 if (ret) {
495 pr_err("Cannot save microcode patch.\n");
496 goto out;
497 }
498
499 show_saved_mc();
500
501 /*
502 * Free old saved microcode data.
503 */
504 if (mc_saved) {
505 for (i = 0; i < mc_saved_count_init; i++)
506 kfree(mc_saved[i]);
507 kfree(mc_saved);
508 }
509
510out:
511 mutex_unlock(&x86_cpu_microcode_mutex);
512
513 return ret;
514}
515EXPORT_SYMBOL_GPL(save_mc_for_early);
516#endif
517
518static bool __init load_builtin_intel_microcode(struct cpio_data *cp)
519{
520#ifdef CONFIG_X86_64
521 unsigned int eax = 0x00000001, ebx, ecx = 0, edx;
522 char name[30];
523
524 native_cpuid(&eax, &ebx, &ecx, &edx);
525
526 sprintf(name, "intel-ucode/%02x-%02x-%02x",
527 x86_family(eax), x86_model(eax), x86_stepping(eax));
528
529 return get_builtin_firmware(cp, name);
530#else
531 return false;
532#endif
533}
534
535static __initdata char ucode_name[] = "kernel/x86/microcode/GenuineIntel.bin";
536static __init enum ucode_state
537scan_microcode(struct mc_saved_data *mcs, unsigned long *mc_ptrs,
538 unsigned long start, unsigned long size,
539 struct ucode_cpu_info *uci)
540{
541 struct cpio_data cd;
542 long offset = 0;
543#ifdef CONFIG_X86_32
544 char *p = (char *)__pa_nodebug(ucode_name);
545#else
546 char *p = ucode_name;
547#endif
548
549 cd.data = NULL;
550 cd.size = 0;
551
552 /* try built-in microcode if no initrd */
553 if (!size) {
554 if (!load_builtin_intel_microcode(&cd))
555 return UCODE_ERROR;
556 } else {
557 cd = find_cpio_data(p, (void *)start, size, &offset);
558 if (!cd.data)
559 return UCODE_ERROR;
560 }
561
562 return get_matching_model_microcode(start, cd.data, cd.size,
563 mcs, mc_ptrs, uci);
564}
565
566/*
567 * Print ucode update info.
568 */
569static void
570print_ucode_info(struct ucode_cpu_info *uci, unsigned int date)
571{
572 pr_info_once("microcode updated early to revision 0x%x, date = %04x-%02x-%02x\n",
573 uci->cpu_sig.rev,
574 date & 0xffff,
575 date >> 24,
576 (date >> 16) & 0xff);
577}
578
579#ifdef CONFIG_X86_32
580
581static int delay_ucode_info;
582static int current_mc_date;
583
584/*
585 * Print early updated ucode info after printk works. This is delayed info dump.
586 */
587void show_ucode_info_early(void)
588{
589 struct ucode_cpu_info uci;
590
591 if (delay_ucode_info) {
592 collect_cpu_info_early(&uci);
593 print_ucode_info(&uci, current_mc_date);
594 delay_ucode_info = 0;
595 }
596}
597
598/*
599 * At this point, we can not call printk() yet. Keep microcode patch number in
600 * mc_saved_data.mc_saved and delay printing microcode info in
601 * show_ucode_info_early() until printk() works.
602 */
603static void print_ucode(struct ucode_cpu_info *uci)
604{
605 struct microcode_intel *mc;
606 int *delay_ucode_info_p;
607 int *current_mc_date_p;
608
609 mc = uci->mc;
610 if (!mc)
611 return;
612
613 delay_ucode_info_p = (int *)__pa_nodebug(&delay_ucode_info);
614 current_mc_date_p = (int *)__pa_nodebug(¤t_mc_date);
615
616 *delay_ucode_info_p = 1;
617 *current_mc_date_p = mc->hdr.date;
618}
619#else
620
621/*
622 * Flush global tlb. We only do this in x86_64 where paging has been enabled
623 * already and PGE should be enabled as well.
624 */
625static inline void flush_tlb_early(void)
626{
627 __native_flush_tlb_global_irq_disabled();
628}
629
630static inline void print_ucode(struct ucode_cpu_info *uci)
631{
632 struct microcode_intel *mc;
633
634 mc = uci->mc;
635 if (!mc)
636 return;
637
638 print_ucode_info(uci, mc->hdr.date);
639}
640#endif
641
642static int apply_microcode_early(struct ucode_cpu_info *uci, bool early)
643{
644 struct microcode_intel *mc;
645 unsigned int val[2];
646
647 mc = uci->mc;
648 if (!mc)
649 return 0;
650
651 /* write microcode via MSR 0x79 */
652 native_wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
653 native_wrmsrl(MSR_IA32_UCODE_REV, 0);
654
655 /* As documented in the SDM: Do a CPUID 1 here */
656 sync_core();
657
658 /* get the current revision from MSR 0x8B */
659 native_rdmsr(MSR_IA32_UCODE_REV, val[0], val[1]);
660 if (val[1] != mc->hdr.rev)
661 return -1;
662
663#ifdef CONFIG_X86_64
664 /* Flush global tlb. This is precaution. */
665 flush_tlb_early();
666#endif
667 uci->cpu_sig.rev = val[1];
668
669 if (early)
670 print_ucode(uci);
671 else
672 print_ucode_info(uci, mc->hdr.date);
673
674 return 0;
675}
676
677/*
678 * This function converts microcode patch offsets previously stored in
679 * mc_tmp_ptrs to pointers and stores the pointers in mc_saved_data.
680 */
681int __init save_microcode_in_initrd_intel(void)
682{
683 unsigned int count = mc_saved_data.num_saved;
684 struct microcode_intel *mc_saved[MAX_UCODE_COUNT];
685 int ret = 0;
686
687 if (!count)
688 return ret;
689
690 copy_ptrs(mc_saved, mc_tmp_ptrs, get_initrd_start(), count);
691
692 ret = save_microcode(&mc_saved_data, mc_saved, count);
693 if (ret)
694 pr_err("Cannot save microcode patches from initrd.\n");
695
696 show_saved_mc();
697
698 return ret;
699}
700
701static void __init
702_load_ucode_intel_bsp(struct mc_saved_data *mcs, unsigned long *mc_ptrs,
703 unsigned long start, unsigned long size)
704{
705 struct ucode_cpu_info uci;
706 enum ucode_state ret;
707
708 collect_cpu_info_early(&uci);
709
710 ret = scan_microcode(mcs, mc_ptrs, start, size, &uci);
711 if (ret != UCODE_OK)
712 return;
713
714 ret = load_microcode(mcs, mc_ptrs, start, &uci);
715 if (ret != UCODE_OK)
716 return;
717
718 apply_microcode_early(&uci, true);
719}
720
721void __init load_ucode_intel_bsp(void)
722{
723 u64 start, size;
724#ifdef CONFIG_X86_32
725 struct boot_params *p;
726
727 p = (struct boot_params *)__pa_nodebug(&boot_params);
728 size = p->hdr.ramdisk_size;
729
730 /*
731 * Set start only if we have an initrd image. We cannot use initrd_start
732 * because it is not set that early yet.
733 */
734 start = (size ? p->hdr.ramdisk_image : 0);
735
736 _load_ucode_intel_bsp((struct mc_saved_data *)__pa_nodebug(&mc_saved_data),
737 (unsigned long *)__pa_nodebug(&mc_tmp_ptrs),
738 start, size);
739#else
740 size = boot_params.hdr.ramdisk_size;
741 start = (size ? boot_params.hdr.ramdisk_image + PAGE_OFFSET : 0);
742
743 _load_ucode_intel_bsp(&mc_saved_data, mc_tmp_ptrs, start, size);
744#endif
745}
746
747void load_ucode_intel_ap(void)
748{
749 unsigned long *mcs_tmp_p;
750 struct mc_saved_data *mcs_p;
751 struct ucode_cpu_info uci;
752 enum ucode_state ret;
753#ifdef CONFIG_X86_32
754
755 mcs_tmp_p = (unsigned long *)__pa_nodebug(mc_tmp_ptrs);
756 mcs_p = (struct mc_saved_data *)__pa_nodebug(&mc_saved_data);
757#else
758 mcs_tmp_p = mc_tmp_ptrs;
759 mcs_p = &mc_saved_data;
760#endif
761
762 /*
763 * If there is no valid ucode previously saved in memory, no need to
764 * update ucode on this AP.
765 */
766 if (!mcs_p->num_saved)
767 return;
768
769 collect_cpu_info_early(&uci);
770 ret = load_microcode(mcs_p, mcs_tmp_p, get_initrd_start_addr(), &uci);
771 if (ret != UCODE_OK)
772 return;
773
774 apply_microcode_early(&uci, true);
775}
776
777void reload_ucode_intel(void)
778{
779 struct ucode_cpu_info uci;
780 enum ucode_state ret;
781
782 if (!mc_saved_data.num_saved)
783 return;
784
785 collect_cpu_info_early(&uci);
786
787 ret = load_microcode_early(mc_saved_data.mc_saved,
788 mc_saved_data.num_saved, &uci);
789 if (ret != UCODE_OK)
790 return;
791
792 apply_microcode_early(&uci, false);
793}
794
795static int collect_cpu_info(int cpu_num, struct cpu_signature *csig)
796{
797 struct cpuinfo_x86 *c = &cpu_data(cpu_num);
798 unsigned int val[2];
799
800 memset(csig, 0, sizeof(*csig));
801
802 csig->sig = cpuid_eax(0x00000001);
803
804 if ((c->x86_model >= 5) || (c->x86 > 6)) {
805 /* get processor flags from MSR 0x17 */
806 rdmsr(MSR_IA32_PLATFORM_ID, val[0], val[1]);
807 csig->pf = 1 << ((val[1] >> 18) & 7);
808 }
809
810 csig->rev = c->microcode;
811 pr_info("CPU%d sig=0x%x, pf=0x%x, revision=0x%x\n",
812 cpu_num, csig->sig, csig->pf, csig->rev);
813
814 return 0;
815}
816
817/*
818 * return 0 - no update found
819 * return 1 - found update
820 */
821static int get_matching_mc(struct microcode_intel *mc, int cpu)
822{
823 struct cpu_signature cpu_sig;
824 unsigned int csig, cpf, crev;
825
826 collect_cpu_info(cpu, &cpu_sig);
827
828 csig = cpu_sig.sig;
829 cpf = cpu_sig.pf;
830 crev = cpu_sig.rev;
831
832 return has_newer_microcode(mc, csig, cpf, crev);
833}
834
835static int apply_microcode_intel(int cpu)
836{
837 struct microcode_intel *mc;
838 struct ucode_cpu_info *uci;
839 struct cpuinfo_x86 *c;
840 unsigned int val[2];
841
842 /* We should bind the task to the CPU */
843 if (WARN_ON(raw_smp_processor_id() != cpu))
844 return -1;
845
846 uci = ucode_cpu_info + cpu;
847 mc = uci->mc;
848 if (!mc)
849 return 0;
850
851 /*
852 * Microcode on this CPU could be updated earlier. Only apply the
853 * microcode patch in mc when it is newer than the one on this
854 * CPU.
855 */
856 if (!get_matching_mc(mc, cpu))
857 return 0;
858
859 /* write microcode via MSR 0x79 */
860 wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
861 wrmsrl(MSR_IA32_UCODE_REV, 0);
862
863 /* As documented in the SDM: Do a CPUID 1 here */
864 sync_core();
865
866 /* get the current revision from MSR 0x8B */
867 rdmsr(MSR_IA32_UCODE_REV, val[0], val[1]);
868
869 if (val[1] != mc->hdr.rev) {
870 pr_err("CPU%d update to revision 0x%x failed\n",
871 cpu, mc->hdr.rev);
872 return -1;
873 }
874
875 pr_info("CPU%d updated to revision 0x%x, date = %04x-%02x-%02x\n",
876 cpu, val[1],
877 mc->hdr.date & 0xffff,
878 mc->hdr.date >> 24,
879 (mc->hdr.date >> 16) & 0xff);
880
881 c = &cpu_data(cpu);
882
883 uci->cpu_sig.rev = val[1];
884 c->microcode = val[1];
885
886 return 0;
887}
888
889static enum ucode_state generic_load_microcode(int cpu, void *data, size_t size,
890 int (*get_ucode_data)(void *, const void *, size_t))
891{
892 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
893 u8 *ucode_ptr = data, *new_mc = NULL, *mc = NULL;
894 int new_rev = uci->cpu_sig.rev;
895 unsigned int leftover = size;
896 enum ucode_state state = UCODE_OK;
897 unsigned int curr_mc_size = 0;
898 unsigned int csig, cpf;
899
900 while (leftover) {
901 struct microcode_header_intel mc_header;
902 unsigned int mc_size;
903
904 if (leftover < sizeof(mc_header)) {
905 pr_err("error! Truncated header in microcode data file\n");
906 break;
907 }
908
909 if (get_ucode_data(&mc_header, ucode_ptr, sizeof(mc_header)))
910 break;
911
912 mc_size = get_totalsize(&mc_header);
913 if (!mc_size || mc_size > leftover) {
914 pr_err("error! Bad data in microcode data file\n");
915 break;
916 }
917
918 /* For performance reasons, reuse mc area when possible */
919 if (!mc || mc_size > curr_mc_size) {
920 vfree(mc);
921 mc = vmalloc(mc_size);
922 if (!mc)
923 break;
924 curr_mc_size = mc_size;
925 }
926
927 if (get_ucode_data(mc, ucode_ptr, mc_size) ||
928 microcode_sanity_check(mc, 1) < 0) {
929 break;
930 }
931
932 csig = uci->cpu_sig.sig;
933 cpf = uci->cpu_sig.pf;
934 if (has_newer_microcode(mc, csig, cpf, new_rev)) {
935 vfree(new_mc);
936 new_rev = mc_header.rev;
937 new_mc = mc;
938 mc = NULL; /* trigger new vmalloc */
939 }
940
941 ucode_ptr += mc_size;
942 leftover -= mc_size;
943 }
944
945 vfree(mc);
946
947 if (leftover) {
948 vfree(new_mc);
949 state = UCODE_ERROR;
950 goto out;
951 }
952
953 if (!new_mc) {
954 state = UCODE_NFOUND;
955 goto out;
956 }
957
958 vfree(uci->mc);
959 uci->mc = (struct microcode_intel *)new_mc;
960
961 /*
962 * If early loading microcode is supported, save this mc into
963 * permanent memory. So it will be loaded early when a CPU is hot added
964 * or resumes.
965 */
966 save_mc_for_early(new_mc);
967
968 pr_debug("CPU%d found a matching microcode update with version 0x%x (current=0x%x)\n",
969 cpu, new_rev, uci->cpu_sig.rev);
970out:
971 return state;
972}
973
974static int get_ucode_fw(void *to, const void *from, size_t n)
975{
976 memcpy(to, from, n);
977 return 0;
978}
979
980static enum ucode_state request_microcode_fw(int cpu, struct device *device,
981 bool refresh_fw)
982{
983 char name[30];
984 struct cpuinfo_x86 *c = &cpu_data(cpu);
985 const struct firmware *firmware;
986 enum ucode_state ret;
987
988 sprintf(name, "intel-ucode/%02x-%02x-%02x",
989 c->x86, c->x86_model, c->x86_mask);
990
991 if (request_firmware_direct(&firmware, name, device)) {
992 pr_debug("data file %s load failed\n", name);
993 return UCODE_NFOUND;
994 }
995
996 ret = generic_load_microcode(cpu, (void *)firmware->data,
997 firmware->size, &get_ucode_fw);
998
999 release_firmware(firmware);
1000
1001 return ret;
1002}
1003
1004static int get_ucode_user(void *to, const void *from, size_t n)
1005{
1006 return copy_from_user(to, from, n);
1007}
1008
1009static enum ucode_state
1010request_microcode_user(int cpu, const void __user *buf, size_t size)
1011{
1012 return generic_load_microcode(cpu, (void *)buf, size, &get_ucode_user);
1013}
1014
1015static void microcode_fini_cpu(int cpu)
1016{
1017 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
1018
1019 vfree(uci->mc);
1020 uci->mc = NULL;
1021}
1022
1023static struct microcode_ops microcode_intel_ops = {
1024 .request_microcode_user = request_microcode_user,
1025 .request_microcode_fw = request_microcode_fw,
1026 .collect_cpu_info = collect_cpu_info,
1027 .apply_microcode = apply_microcode_intel,
1028 .microcode_fini_cpu = microcode_fini_cpu,
1029};
1030
1031struct microcode_ops * __init init_intel_microcode(void)
1032{
1033 struct cpuinfo_x86 *c = &boot_cpu_data;
1034
1035 if (c->x86_vendor != X86_VENDOR_INTEL || c->x86 < 6 ||
1036 cpu_has(c, X86_FEATURE_IA64)) {
1037 pr_err("Intel CPU family 0x%x not supported\n", c->x86);
1038 return NULL;
1039 }
1040
1041 return µcode_intel_ops;
1042}
1043