Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Intel CPU Microcode Update Driver for Linux
4 *
5 * Copyright (C) 2000-2006 Tigran Aivazian <aivazian.tigran@gmail.com>
6 * 2006 Shaohua Li <shaohua.li@intel.com>
7 *
8 * Intel CPU microcode early update for Linux
9 *
10 * Copyright (C) 2012 Fenghua Yu <fenghua.yu@intel.com>
11 * H Peter Anvin" <hpa@zytor.com>
12 */
13
14/*
15 * This needs to be before all headers so that pr_debug in printk.h doesn't turn
16 * printk calls into no_printk().
17 *
18 *#define DEBUG
19 */
20#define pr_fmt(fmt) "microcode: " fmt
21
22#include <linux/earlycpio.h>
23#include <linux/firmware.h>
24#include <linux/uaccess.h>
25#include <linux/vmalloc.h>
26#include <linux/initrd.h>
27#include <linux/kernel.h>
28#include <linux/slab.h>
29#include <linux/cpu.h>
30#include <linux/uio.h>
31#include <linux/mm.h>
32
33#include <asm/microcode_intel.h>
34#include <asm/intel-family.h>
35#include <asm/processor.h>
36#include <asm/tlbflush.h>
37#include <asm/setup.h>
38#include <asm/msr.h>
39
40static const char ucode_path[] = "kernel/x86/microcode/GenuineIntel.bin";
41
42/* Current microcode patch used in early patching on the APs. */
43static struct microcode_intel *intel_ucode_patch;
44
45/* last level cache size per core */
46static int llc_size_per_core;
47
48/*
49 * Returns 1 if update has been found, 0 otherwise.
50 */
51static int has_newer_microcode(void *mc, unsigned int csig, int cpf, int new_rev)
52{
53 struct microcode_header_intel *mc_hdr = mc;
54
55 if (mc_hdr->rev <= new_rev)
56 return 0;
57
58 return intel_find_matching_signature(mc, csig, cpf);
59}
60
61static struct ucode_patch *memdup_patch(void *data, unsigned int size)
62{
63 struct ucode_patch *p;
64
65 p = kzalloc(sizeof(struct ucode_patch), GFP_KERNEL);
66 if (!p)
67 return NULL;
68
69 p->data = kmemdup(data, size, GFP_KERNEL);
70 if (!p->data) {
71 kfree(p);
72 return NULL;
73 }
74
75 return p;
76}
77
78static void save_microcode_patch(struct ucode_cpu_info *uci, void *data, unsigned int size)
79{
80 struct microcode_header_intel *mc_hdr, *mc_saved_hdr;
81 struct ucode_patch *iter, *tmp, *p = NULL;
82 bool prev_found = false;
83 unsigned int sig, pf;
84
85 mc_hdr = (struct microcode_header_intel *)data;
86
87 list_for_each_entry_safe(iter, tmp, µcode_cache, plist) {
88 mc_saved_hdr = (struct microcode_header_intel *)iter->data;
89 sig = mc_saved_hdr->sig;
90 pf = mc_saved_hdr->pf;
91
92 if (intel_find_matching_signature(data, sig, pf)) {
93 prev_found = true;
94
95 if (mc_hdr->rev <= mc_saved_hdr->rev)
96 continue;
97
98 p = memdup_patch(data, size);
99 if (!p)
100 pr_err("Error allocating buffer %p\n", data);
101 else {
102 list_replace(&iter->plist, &p->plist);
103 kfree(iter->data);
104 kfree(iter);
105 }
106 }
107 }
108
109 /*
110 * There weren't any previous patches found in the list cache; save the
111 * newly found.
112 */
113 if (!prev_found) {
114 p = memdup_patch(data, size);
115 if (!p)
116 pr_err("Error allocating buffer for %p\n", data);
117 else
118 list_add_tail(&p->plist, µcode_cache);
119 }
120
121 if (!p)
122 return;
123
124 if (!intel_find_matching_signature(p->data, uci->cpu_sig.sig, uci->cpu_sig.pf))
125 return;
126
127 /*
128 * Save for early loading. On 32-bit, that needs to be a physical
129 * address as the APs are running from physical addresses, before
130 * paging has been enabled.
131 */
132 if (IS_ENABLED(CONFIG_X86_32))
133 intel_ucode_patch = (struct microcode_intel *)__pa_nodebug(p->data);
134 else
135 intel_ucode_patch = p->data;
136}
137
138/*
139 * Get microcode matching with BSP's model. Only CPUs with the same model as
140 * BSP can stay in the platform.
141 */
142static struct microcode_intel *
143scan_microcode(void *data, size_t size, struct ucode_cpu_info *uci, bool save)
144{
145 struct microcode_header_intel *mc_header;
146 struct microcode_intel *patch = NULL;
147 unsigned int mc_size;
148
149 while (size) {
150 if (size < sizeof(struct microcode_header_intel))
151 break;
152
153 mc_header = (struct microcode_header_intel *)data;
154
155 mc_size = get_totalsize(mc_header);
156 if (!mc_size ||
157 mc_size > size ||
158 intel_microcode_sanity_check(data, false, MC_HEADER_TYPE_MICROCODE) < 0)
159 break;
160
161 size -= mc_size;
162
163 if (!intel_find_matching_signature(data, uci->cpu_sig.sig,
164 uci->cpu_sig.pf)) {
165 data += mc_size;
166 continue;
167 }
168
169 if (save) {
170 save_microcode_patch(uci, data, mc_size);
171 goto next;
172 }
173
174
175 if (!patch) {
176 if (!has_newer_microcode(data,
177 uci->cpu_sig.sig,
178 uci->cpu_sig.pf,
179 uci->cpu_sig.rev))
180 goto next;
181
182 } else {
183 struct microcode_header_intel *phdr = &patch->hdr;
184
185 if (!has_newer_microcode(data,
186 phdr->sig,
187 phdr->pf,
188 phdr->rev))
189 goto next;
190 }
191
192 /* We have a newer patch, save it. */
193 patch = data;
194
195next:
196 data += mc_size;
197 }
198
199 if (size)
200 return NULL;
201
202 return patch;
203}
204
205static void show_saved_mc(void)
206{
207#ifdef DEBUG
208 int i = 0, j;
209 unsigned int sig, pf, rev, total_size, data_size, date;
210 struct ucode_cpu_info uci;
211 struct ucode_patch *p;
212
213 if (list_empty(µcode_cache)) {
214 pr_debug("no microcode data saved.\n");
215 return;
216 }
217
218 intel_cpu_collect_info(&uci);
219
220 sig = uci.cpu_sig.sig;
221 pf = uci.cpu_sig.pf;
222 rev = uci.cpu_sig.rev;
223 pr_debug("CPU: sig=0x%x, pf=0x%x, rev=0x%x\n", sig, pf, rev);
224
225 list_for_each_entry(p, µcode_cache, plist) {
226 struct microcode_header_intel *mc_saved_header;
227 struct extended_sigtable *ext_header;
228 struct extended_signature *ext_sig;
229 int ext_sigcount;
230
231 mc_saved_header = (struct microcode_header_intel *)p->data;
232
233 sig = mc_saved_header->sig;
234 pf = mc_saved_header->pf;
235 rev = mc_saved_header->rev;
236 date = mc_saved_header->date;
237
238 total_size = get_totalsize(mc_saved_header);
239 data_size = get_datasize(mc_saved_header);
240
241 pr_debug("mc_saved[%d]: sig=0x%x, pf=0x%x, rev=0x%x, total size=0x%x, date = %04x-%02x-%02x\n",
242 i++, sig, pf, rev, total_size,
243 date & 0xffff,
244 date >> 24,
245 (date >> 16) & 0xff);
246
247 /* Look for ext. headers: */
248 if (total_size <= data_size + MC_HEADER_SIZE)
249 continue;
250
251 ext_header = (void *)mc_saved_header + data_size + MC_HEADER_SIZE;
252 ext_sigcount = ext_header->count;
253 ext_sig = (void *)ext_header + EXT_HEADER_SIZE;
254
255 for (j = 0; j < ext_sigcount; j++) {
256 sig = ext_sig->sig;
257 pf = ext_sig->pf;
258
259 pr_debug("\tExtended[%d]: sig=0x%x, pf=0x%x\n",
260 j, sig, pf);
261
262 ext_sig++;
263 }
264 }
265#endif
266}
267
268/*
269 * Save this microcode patch. It will be loaded early when a CPU is
270 * hot-added or resumes.
271 */
272static void save_mc_for_early(struct ucode_cpu_info *uci, u8 *mc, unsigned int size)
273{
274 /* Synchronization during CPU hotplug. */
275 static DEFINE_MUTEX(x86_cpu_microcode_mutex);
276
277 mutex_lock(&x86_cpu_microcode_mutex);
278
279 save_microcode_patch(uci, mc, size);
280 show_saved_mc();
281
282 mutex_unlock(&x86_cpu_microcode_mutex);
283}
284
285static bool load_builtin_intel_microcode(struct cpio_data *cp)
286{
287 unsigned int eax = 1, ebx, ecx = 0, edx;
288 struct firmware fw;
289 char name[30];
290
291 if (IS_ENABLED(CONFIG_X86_32))
292 return false;
293
294 native_cpuid(&eax, &ebx, &ecx, &edx);
295
296 sprintf(name, "intel-ucode/%02x-%02x-%02x",
297 x86_family(eax), x86_model(eax), x86_stepping(eax));
298
299 if (firmware_request_builtin(&fw, name)) {
300 cp->size = fw.size;
301 cp->data = (void *)fw.data;
302 return true;
303 }
304
305 return false;
306}
307
308/*
309 * Print ucode update info.
310 */
311static void
312print_ucode_info(struct ucode_cpu_info *uci, unsigned int date)
313{
314 pr_info_once("microcode updated early to revision 0x%x, date = %04x-%02x-%02x\n",
315 uci->cpu_sig.rev,
316 date & 0xffff,
317 date >> 24,
318 (date >> 16) & 0xff);
319}
320
321#ifdef CONFIG_X86_32
322
323static int delay_ucode_info;
324static int current_mc_date;
325
326/*
327 * Print early updated ucode info after printk works. This is delayed info dump.
328 */
329void show_ucode_info_early(void)
330{
331 struct ucode_cpu_info uci;
332
333 if (delay_ucode_info) {
334 intel_cpu_collect_info(&uci);
335 print_ucode_info(&uci, current_mc_date);
336 delay_ucode_info = 0;
337 }
338}
339
340/*
341 * At this point, we can not call printk() yet. Delay printing microcode info in
342 * show_ucode_info_early() until printk() works.
343 */
344static void print_ucode(struct ucode_cpu_info *uci)
345{
346 struct microcode_intel *mc;
347 int *delay_ucode_info_p;
348 int *current_mc_date_p;
349
350 mc = uci->mc;
351 if (!mc)
352 return;
353
354 delay_ucode_info_p = (int *)__pa_nodebug(&delay_ucode_info);
355 current_mc_date_p = (int *)__pa_nodebug(¤t_mc_date);
356
357 *delay_ucode_info_p = 1;
358 *current_mc_date_p = mc->hdr.date;
359}
360#else
361
362static inline void print_ucode(struct ucode_cpu_info *uci)
363{
364 struct microcode_intel *mc;
365
366 mc = uci->mc;
367 if (!mc)
368 return;
369
370 print_ucode_info(uci, mc->hdr.date);
371}
372#endif
373
374static int apply_microcode_early(struct ucode_cpu_info *uci, bool early)
375{
376 struct microcode_intel *mc;
377 u32 rev;
378
379 mc = uci->mc;
380 if (!mc)
381 return 0;
382
383 /*
384 * Save us the MSR write below - which is a particular expensive
385 * operation - when the other hyperthread has updated the microcode
386 * already.
387 */
388 rev = intel_get_microcode_revision();
389 if (rev >= mc->hdr.rev) {
390 uci->cpu_sig.rev = rev;
391 return UCODE_OK;
392 }
393
394 /*
395 * Writeback and invalidate caches before updating microcode to avoid
396 * internal issues depending on what the microcode is updating.
397 */
398 native_wbinvd();
399
400 /* write microcode via MSR 0x79 */
401 native_wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
402
403 rev = intel_get_microcode_revision();
404 if (rev != mc->hdr.rev)
405 return -1;
406
407 uci->cpu_sig.rev = rev;
408
409 if (early)
410 print_ucode(uci);
411 else
412 print_ucode_info(uci, mc->hdr.date);
413
414 return 0;
415}
416
417int __init save_microcode_in_initrd_intel(void)
418{
419 struct ucode_cpu_info uci;
420 struct cpio_data cp;
421
422 /*
423 * initrd is going away, clear patch ptr. We will scan the microcode one
424 * last time before jettisoning and save a patch, if found. Then we will
425 * update that pointer too, with a stable patch address to use when
426 * resuming the cores.
427 */
428 intel_ucode_patch = NULL;
429
430 if (!load_builtin_intel_microcode(&cp))
431 cp = find_microcode_in_initrd(ucode_path, false);
432
433 if (!(cp.data && cp.size))
434 return 0;
435
436 intel_cpu_collect_info(&uci);
437
438 scan_microcode(cp.data, cp.size, &uci, true);
439
440 show_saved_mc();
441
442 return 0;
443}
444
445/*
446 * @res_patch, output: a pointer to the patch we found.
447 */
448static struct microcode_intel *__load_ucode_intel(struct ucode_cpu_info *uci)
449{
450 static const char *path;
451 struct cpio_data cp;
452 bool use_pa;
453
454 if (IS_ENABLED(CONFIG_X86_32)) {
455 path = (const char *)__pa_nodebug(ucode_path);
456 use_pa = true;
457 } else {
458 path = ucode_path;
459 use_pa = false;
460 }
461
462 /* try built-in microcode first */
463 if (!load_builtin_intel_microcode(&cp))
464 cp = find_microcode_in_initrd(path, use_pa);
465
466 if (!(cp.data && cp.size))
467 return NULL;
468
469 intel_cpu_collect_info(uci);
470
471 return scan_microcode(cp.data, cp.size, uci, false);
472}
473
474void __init load_ucode_intel_bsp(void)
475{
476 struct microcode_intel *patch;
477 struct ucode_cpu_info uci;
478
479 patch = __load_ucode_intel(&uci);
480 if (!patch)
481 return;
482
483 uci.mc = patch;
484
485 apply_microcode_early(&uci, true);
486}
487
488void load_ucode_intel_ap(void)
489{
490 struct microcode_intel *patch, **iup;
491 struct ucode_cpu_info uci;
492
493 if (IS_ENABLED(CONFIG_X86_32))
494 iup = (struct microcode_intel **) __pa_nodebug(&intel_ucode_patch);
495 else
496 iup = &intel_ucode_patch;
497
498 if (!*iup) {
499 patch = __load_ucode_intel(&uci);
500 if (!patch)
501 return;
502
503 *iup = patch;
504 }
505
506 uci.mc = *iup;
507
508 apply_microcode_early(&uci, true);
509}
510
511static struct microcode_intel *find_patch(struct ucode_cpu_info *uci)
512{
513 struct microcode_header_intel *phdr;
514 struct ucode_patch *iter, *tmp;
515
516 list_for_each_entry_safe(iter, tmp, µcode_cache, plist) {
517
518 phdr = (struct microcode_header_intel *)iter->data;
519
520 if (phdr->rev <= uci->cpu_sig.rev)
521 continue;
522
523 if (!intel_find_matching_signature(phdr,
524 uci->cpu_sig.sig,
525 uci->cpu_sig.pf))
526 continue;
527
528 return iter->data;
529 }
530 return NULL;
531}
532
533void reload_ucode_intel(void)
534{
535 struct microcode_intel *p;
536 struct ucode_cpu_info uci;
537
538 intel_cpu_collect_info(&uci);
539
540 p = find_patch(&uci);
541 if (!p)
542 return;
543
544 uci.mc = p;
545
546 apply_microcode_early(&uci, false);
547}
548
549static int collect_cpu_info(int cpu_num, struct cpu_signature *csig)
550{
551 struct cpuinfo_x86 *c = &cpu_data(cpu_num);
552 unsigned int val[2];
553
554 memset(csig, 0, sizeof(*csig));
555
556 csig->sig = cpuid_eax(0x00000001);
557
558 if ((c->x86_model >= 5) || (c->x86 > 6)) {
559 /* get processor flags from MSR 0x17 */
560 rdmsr(MSR_IA32_PLATFORM_ID, val[0], val[1]);
561 csig->pf = 1 << ((val[1] >> 18) & 7);
562 }
563
564 csig->rev = c->microcode;
565
566 return 0;
567}
568
569static enum ucode_state apply_microcode_intel(int cpu)
570{
571 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
572 struct cpuinfo_x86 *c = &cpu_data(cpu);
573 bool bsp = c->cpu_index == boot_cpu_data.cpu_index;
574 struct microcode_intel *mc;
575 enum ucode_state ret;
576 static int prev_rev;
577 u32 rev;
578
579 /* We should bind the task to the CPU */
580 if (WARN_ON(raw_smp_processor_id() != cpu))
581 return UCODE_ERROR;
582
583 /* Look for a newer patch in our cache: */
584 mc = find_patch(uci);
585 if (!mc) {
586 mc = uci->mc;
587 if (!mc)
588 return UCODE_NFOUND;
589 }
590
591 /*
592 * Save us the MSR write below - which is a particular expensive
593 * operation - when the other hyperthread has updated the microcode
594 * already.
595 */
596 rev = intel_get_microcode_revision();
597 if (rev >= mc->hdr.rev) {
598 ret = UCODE_OK;
599 goto out;
600 }
601
602 /*
603 * Writeback and invalidate caches before updating microcode to avoid
604 * internal issues depending on what the microcode is updating.
605 */
606 native_wbinvd();
607
608 /* write microcode via MSR 0x79 */
609 wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
610
611 rev = intel_get_microcode_revision();
612
613 if (rev != mc->hdr.rev) {
614 pr_err("CPU%d update to revision 0x%x failed\n",
615 cpu, mc->hdr.rev);
616 return UCODE_ERROR;
617 }
618
619 if (bsp && rev != prev_rev) {
620 pr_info("updated to revision 0x%x, date = %04x-%02x-%02x\n",
621 rev,
622 mc->hdr.date & 0xffff,
623 mc->hdr.date >> 24,
624 (mc->hdr.date >> 16) & 0xff);
625 prev_rev = rev;
626 }
627
628 ret = UCODE_UPDATED;
629
630out:
631 uci->cpu_sig.rev = rev;
632 c->microcode = rev;
633
634 /* Update boot_cpu_data's revision too, if we're on the BSP: */
635 if (bsp)
636 boot_cpu_data.microcode = rev;
637
638 return ret;
639}
640
641static enum ucode_state generic_load_microcode(int cpu, struct iov_iter *iter)
642{
643 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
644 unsigned int curr_mc_size = 0, new_mc_size = 0;
645 enum ucode_state ret = UCODE_OK;
646 int new_rev = uci->cpu_sig.rev;
647 u8 *new_mc = NULL, *mc = NULL;
648 unsigned int csig, cpf;
649
650 while (iov_iter_count(iter)) {
651 struct microcode_header_intel mc_header;
652 unsigned int mc_size, data_size;
653 u8 *data;
654
655 if (!copy_from_iter_full(&mc_header, sizeof(mc_header), iter)) {
656 pr_err("error! Truncated or inaccessible header in microcode data file\n");
657 break;
658 }
659
660 mc_size = get_totalsize(&mc_header);
661 if (mc_size < sizeof(mc_header)) {
662 pr_err("error! Bad data in microcode data file (totalsize too small)\n");
663 break;
664 }
665 data_size = mc_size - sizeof(mc_header);
666 if (data_size > iov_iter_count(iter)) {
667 pr_err("error! Bad data in microcode data file (truncated file?)\n");
668 break;
669 }
670
671 /* For performance reasons, reuse mc area when possible */
672 if (!mc || mc_size > curr_mc_size) {
673 vfree(mc);
674 mc = vmalloc(mc_size);
675 if (!mc)
676 break;
677 curr_mc_size = mc_size;
678 }
679
680 memcpy(mc, &mc_header, sizeof(mc_header));
681 data = mc + sizeof(mc_header);
682 if (!copy_from_iter_full(data, data_size, iter) ||
683 intel_microcode_sanity_check(mc, true, MC_HEADER_TYPE_MICROCODE) < 0) {
684 break;
685 }
686
687 csig = uci->cpu_sig.sig;
688 cpf = uci->cpu_sig.pf;
689 if (has_newer_microcode(mc, csig, cpf, new_rev)) {
690 vfree(new_mc);
691 new_rev = mc_header.rev;
692 new_mc = mc;
693 new_mc_size = mc_size;
694 mc = NULL; /* trigger new vmalloc */
695 ret = UCODE_NEW;
696 }
697 }
698
699 vfree(mc);
700
701 if (iov_iter_count(iter)) {
702 vfree(new_mc);
703 return UCODE_ERROR;
704 }
705
706 if (!new_mc)
707 return UCODE_NFOUND;
708
709 vfree(uci->mc);
710 uci->mc = (struct microcode_intel *)new_mc;
711
712 /*
713 * If early loading microcode is supported, save this mc into
714 * permanent memory. So it will be loaded early when a CPU is hot added
715 * or resumes.
716 */
717 save_mc_for_early(uci, new_mc, new_mc_size);
718
719 pr_debug("CPU%d found a matching microcode update with version 0x%x (current=0x%x)\n",
720 cpu, new_rev, uci->cpu_sig.rev);
721
722 return ret;
723}
724
725static bool is_blacklisted(unsigned int cpu)
726{
727 struct cpuinfo_x86 *c = &cpu_data(cpu);
728
729 /*
730 * Late loading on model 79 with microcode revision less than 0x0b000021
731 * and LLC size per core bigger than 2.5MB may result in a system hang.
732 * This behavior is documented in item BDF90, #334165 (Intel Xeon
733 * Processor E7-8800/4800 v4 Product Family).
734 */
735 if (c->x86 == 6 &&
736 c->x86_model == INTEL_FAM6_BROADWELL_X &&
737 c->x86_stepping == 0x01 &&
738 llc_size_per_core > 2621440 &&
739 c->microcode < 0x0b000021) {
740 pr_err_once("Erratum BDF90: late loading with revision < 0x0b000021 (0x%x) disabled.\n", c->microcode);
741 pr_err_once("Please consider either early loading through initrd/built-in or a potential BIOS update.\n");
742 return true;
743 }
744
745 return false;
746}
747
748static enum ucode_state request_microcode_fw(int cpu, struct device *device)
749{
750 struct cpuinfo_x86 *c = &cpu_data(cpu);
751 const struct firmware *firmware;
752 struct iov_iter iter;
753 enum ucode_state ret;
754 struct kvec kvec;
755 char name[30];
756
757 if (is_blacklisted(cpu))
758 return UCODE_NFOUND;
759
760 sprintf(name, "intel-ucode/%02x-%02x-%02x",
761 c->x86, c->x86_model, c->x86_stepping);
762
763 if (request_firmware_direct(&firmware, name, device)) {
764 pr_debug("data file %s load failed\n", name);
765 return UCODE_NFOUND;
766 }
767
768 kvec.iov_base = (void *)firmware->data;
769 kvec.iov_len = firmware->size;
770 iov_iter_kvec(&iter, ITER_SOURCE, &kvec, 1, firmware->size);
771 ret = generic_load_microcode(cpu, &iter);
772
773 release_firmware(firmware);
774
775 return ret;
776}
777
778static struct microcode_ops microcode_intel_ops = {
779 .request_microcode_fw = request_microcode_fw,
780 .collect_cpu_info = collect_cpu_info,
781 .apply_microcode = apply_microcode_intel,
782};
783
784static int __init calc_llc_size_per_core(struct cpuinfo_x86 *c)
785{
786 u64 llc_size = c->x86_cache_size * 1024ULL;
787
788 do_div(llc_size, c->x86_max_cores);
789
790 return (int)llc_size;
791}
792
793struct microcode_ops * __init init_intel_microcode(void)
794{
795 struct cpuinfo_x86 *c = &boot_cpu_data;
796
797 if (c->x86_vendor != X86_VENDOR_INTEL || c->x86 < 6 ||
798 cpu_has(c, X86_FEATURE_IA64)) {
799 pr_err("Intel CPU family 0x%x not supported\n", c->x86);
800 return NULL;
801 }
802
803 llc_size_per_core = calc_llc_size_per_core(c);
804
805 return µcode_intel_ops;
806}
1/*
2 * Intel CPU Microcode Update Driver for Linux
3 *
4 * Copyright (C) 2000-2006 Tigran Aivazian <tigran@aivazian.fsnet.co.uk>
5 * 2006 Shaohua Li <shaohua.li@intel.com>
6 *
7 * Intel CPU microcode early update for Linux
8 *
9 * Copyright (C) 2012 Fenghua Yu <fenghua.yu@intel.com>
10 * H Peter Anvin" <hpa@zytor.com>
11 *
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
16 */
17
18/*
19 * This needs to be before all headers so that pr_debug in printk.h doesn't turn
20 * printk calls into no_printk().
21 *
22 *#define DEBUG
23 */
24#define pr_fmt(fmt) "microcode: " fmt
25
26#include <linux/earlycpio.h>
27#include <linux/firmware.h>
28#include <linux/uaccess.h>
29#include <linux/vmalloc.h>
30#include <linux/initrd.h>
31#include <linux/kernel.h>
32#include <linux/slab.h>
33#include <linux/cpu.h>
34#include <linux/mm.h>
35
36#include <asm/microcode_intel.h>
37#include <asm/processor.h>
38#include <asm/tlbflush.h>
39#include <asm/setup.h>
40#include <asm/msr.h>
41
42/*
43 * Temporary microcode blobs pointers storage. We note here the pointers to
44 * microcode blobs we've got from whatever storage (detached initrd, builtin).
45 * Later on, we put those into final storage mc_saved_data.mc_saved.
46 */
47static unsigned long mc_tmp_ptrs[MAX_UCODE_COUNT];
48
49static struct mc_saved_data {
50 unsigned int num_saved;
51 struct microcode_intel **mc_saved;
52} mc_saved_data;
53
54static enum ucode_state
55load_microcode_early(struct microcode_intel **saved,
56 unsigned int num_saved, struct ucode_cpu_info *uci)
57{
58 struct microcode_intel *ucode_ptr, *new_mc = NULL;
59 struct microcode_header_intel *mc_hdr;
60 int new_rev, ret, i;
61
62 new_rev = uci->cpu_sig.rev;
63
64 for (i = 0; i < num_saved; i++) {
65 ucode_ptr = saved[i];
66 mc_hdr = (struct microcode_header_intel *)ucode_ptr;
67
68 ret = has_newer_microcode(ucode_ptr,
69 uci->cpu_sig.sig,
70 uci->cpu_sig.pf,
71 new_rev);
72 if (!ret)
73 continue;
74
75 new_rev = mc_hdr->rev;
76 new_mc = ucode_ptr;
77 }
78
79 if (!new_mc)
80 return UCODE_NFOUND;
81
82 uci->mc = (struct microcode_intel *)new_mc;
83 return UCODE_OK;
84}
85
86static inline void
87copy_ptrs(struct microcode_intel **mc_saved, unsigned long *mc_ptrs,
88 unsigned long off, int num_saved)
89{
90 int i;
91
92 for (i = 0; i < num_saved; i++)
93 mc_saved[i] = (struct microcode_intel *)(mc_ptrs[i] + off);
94}
95
96#ifdef CONFIG_X86_32
97static void
98microcode_phys(struct microcode_intel **mc_saved_tmp, struct mc_saved_data *mcs)
99{
100 int i;
101 struct microcode_intel ***mc_saved;
102
103 mc_saved = (struct microcode_intel ***)__pa_nodebug(&mcs->mc_saved);
104
105 for (i = 0; i < mcs->num_saved; i++) {
106 struct microcode_intel *p;
107
108 p = *(struct microcode_intel **)__pa_nodebug(mcs->mc_saved + i);
109 mc_saved_tmp[i] = (struct microcode_intel *)__pa_nodebug(p);
110 }
111}
112#endif
113
114static enum ucode_state
115load_microcode(struct mc_saved_data *mcs, unsigned long *mc_ptrs,
116 unsigned long offset, struct ucode_cpu_info *uci)
117{
118 struct microcode_intel *mc_saved_tmp[MAX_UCODE_COUNT];
119 unsigned int count = mcs->num_saved;
120
121 if (!mcs->mc_saved) {
122 copy_ptrs(mc_saved_tmp, mc_ptrs, offset, count);
123
124 return load_microcode_early(mc_saved_tmp, count, uci);
125 } else {
126#ifdef CONFIG_X86_32
127 microcode_phys(mc_saved_tmp, mcs);
128 return load_microcode_early(mc_saved_tmp, count, uci);
129#else
130 return load_microcode_early(mcs->mc_saved, count, uci);
131#endif
132 }
133}
134
135/*
136 * Given CPU signature and a microcode patch, this function finds if the
137 * microcode patch has matching family and model with the CPU.
138 */
139static enum ucode_state
140matching_model_microcode(struct microcode_header_intel *mc_header,
141 unsigned long sig)
142{
143 unsigned int fam, model;
144 unsigned int fam_ucode, model_ucode;
145 struct extended_sigtable *ext_header;
146 unsigned long total_size = get_totalsize(mc_header);
147 unsigned long data_size = get_datasize(mc_header);
148 int ext_sigcount, i;
149 struct extended_signature *ext_sig;
150
151 fam = x86_family(sig);
152 model = x86_model(sig);
153
154 fam_ucode = x86_family(mc_header->sig);
155 model_ucode = x86_model(mc_header->sig);
156
157 if (fam == fam_ucode && model == model_ucode)
158 return UCODE_OK;
159
160 /* Look for ext. headers: */
161 if (total_size <= data_size + MC_HEADER_SIZE)
162 return UCODE_NFOUND;
163
164 ext_header = (void *) mc_header + data_size + MC_HEADER_SIZE;
165 ext_sig = (void *)ext_header + EXT_HEADER_SIZE;
166 ext_sigcount = ext_header->count;
167
168 for (i = 0; i < ext_sigcount; i++) {
169 fam_ucode = x86_family(ext_sig->sig);
170 model_ucode = x86_model(ext_sig->sig);
171
172 if (fam == fam_ucode && model == model_ucode)
173 return UCODE_OK;
174
175 ext_sig++;
176 }
177 return UCODE_NFOUND;
178}
179
180static int
181save_microcode(struct mc_saved_data *mcs,
182 struct microcode_intel **mc_saved_src,
183 unsigned int num_saved)
184{
185 int i, j;
186 struct microcode_intel **saved_ptr;
187 int ret;
188
189 if (!num_saved)
190 return -EINVAL;
191
192 /*
193 * Copy new microcode data.
194 */
195 saved_ptr = kcalloc(num_saved, sizeof(struct microcode_intel *), GFP_KERNEL);
196 if (!saved_ptr)
197 return -ENOMEM;
198
199 for (i = 0; i < num_saved; i++) {
200 struct microcode_header_intel *mc_hdr;
201 struct microcode_intel *mc;
202 unsigned long size;
203
204 if (!mc_saved_src[i]) {
205 ret = -EINVAL;
206 goto err;
207 }
208
209 mc = mc_saved_src[i];
210 mc_hdr = &mc->hdr;
211 size = get_totalsize(mc_hdr);
212
213 saved_ptr[i] = kmemdup(mc, size, GFP_KERNEL);
214 if (!saved_ptr[i]) {
215 ret = -ENOMEM;
216 goto err;
217 }
218 }
219
220 /*
221 * Point to newly saved microcode.
222 */
223 mcs->mc_saved = saved_ptr;
224 mcs->num_saved = num_saved;
225
226 return 0;
227
228err:
229 for (j = 0; j <= i; j++)
230 kfree(saved_ptr[j]);
231 kfree(saved_ptr);
232
233 return ret;
234}
235
236/*
237 * A microcode patch in ucode_ptr is saved into mc_saved
238 * - if it has matching signature and newer revision compared to an existing
239 * patch mc_saved.
240 * - or if it is a newly discovered microcode patch.
241 *
242 * The microcode patch should have matching model with CPU.
243 *
244 * Returns: The updated number @num_saved of saved microcode patches.
245 */
246static unsigned int _save_mc(struct microcode_intel **mc_saved,
247 u8 *ucode_ptr, unsigned int num_saved)
248{
249 struct microcode_header_intel *mc_hdr, *mc_saved_hdr;
250 unsigned int sig, pf;
251 int found = 0, i;
252
253 mc_hdr = (struct microcode_header_intel *)ucode_ptr;
254
255 for (i = 0; i < num_saved; i++) {
256 mc_saved_hdr = (struct microcode_header_intel *)mc_saved[i];
257 sig = mc_saved_hdr->sig;
258 pf = mc_saved_hdr->pf;
259
260 if (!find_matching_signature(ucode_ptr, sig, pf))
261 continue;
262
263 found = 1;
264
265 if (mc_hdr->rev <= mc_saved_hdr->rev)
266 continue;
267
268 /*
269 * Found an older ucode saved earlier. Replace it with
270 * this newer one.
271 */
272 mc_saved[i] = (struct microcode_intel *)ucode_ptr;
273 break;
274 }
275
276 /* Newly detected microcode, save it to memory. */
277 if (i >= num_saved && !found)
278 mc_saved[num_saved++] = (struct microcode_intel *)ucode_ptr;
279
280 return num_saved;
281}
282
283/*
284 * Get microcode matching with BSP's model. Only CPUs with the same model as
285 * BSP can stay in the platform.
286 */
287static enum ucode_state __init
288get_matching_model_microcode(unsigned long start, void *data, size_t size,
289 struct mc_saved_data *mcs, unsigned long *mc_ptrs,
290 struct ucode_cpu_info *uci)
291{
292 struct microcode_intel *mc_saved_tmp[MAX_UCODE_COUNT];
293 struct microcode_header_intel *mc_header;
294 unsigned int num_saved = mcs->num_saved;
295 enum ucode_state state = UCODE_OK;
296 unsigned int leftover = size;
297 u8 *ucode_ptr = data;
298 unsigned int mc_size;
299 int i;
300
301 while (leftover && num_saved < ARRAY_SIZE(mc_saved_tmp)) {
302
303 if (leftover < sizeof(mc_header))
304 break;
305
306 mc_header = (struct microcode_header_intel *)ucode_ptr;
307
308 mc_size = get_totalsize(mc_header);
309 if (!mc_size || mc_size > leftover ||
310 microcode_sanity_check(ucode_ptr, 0) < 0)
311 break;
312
313 leftover -= mc_size;
314
315 /*
316 * Since APs with same family and model as the BSP may boot in
317 * the platform, we need to find and save microcode patches
318 * with the same family and model as the BSP.
319 */
320 if (matching_model_microcode(mc_header, uci->cpu_sig.sig) != UCODE_OK) {
321 ucode_ptr += mc_size;
322 continue;
323 }
324
325 num_saved = _save_mc(mc_saved_tmp, ucode_ptr, num_saved);
326
327 ucode_ptr += mc_size;
328 }
329
330 if (leftover) {
331 state = UCODE_ERROR;
332 return state;
333 }
334
335 if (!num_saved) {
336 state = UCODE_NFOUND;
337 return state;
338 }
339
340 for (i = 0; i < num_saved; i++)
341 mc_ptrs[i] = (unsigned long)mc_saved_tmp[i] - start;
342
343 mcs->num_saved = num_saved;
344
345 return state;
346}
347
348static int collect_cpu_info_early(struct ucode_cpu_info *uci)
349{
350 unsigned int val[2];
351 unsigned int family, model;
352 struct cpu_signature csig;
353 unsigned int eax, ebx, ecx, edx;
354
355 csig.sig = 0;
356 csig.pf = 0;
357 csig.rev = 0;
358
359 memset(uci, 0, sizeof(*uci));
360
361 eax = 0x00000001;
362 ecx = 0;
363 native_cpuid(&eax, &ebx, &ecx, &edx);
364 csig.sig = eax;
365
366 family = x86_family(csig.sig);
367 model = x86_model(csig.sig);
368
369 if ((model >= 5) || (family > 6)) {
370 /* get processor flags from MSR 0x17 */
371 native_rdmsr(MSR_IA32_PLATFORM_ID, val[0], val[1]);
372 csig.pf = 1 << ((val[1] >> 18) & 7);
373 }
374 native_wrmsrl(MSR_IA32_UCODE_REV, 0);
375
376 /* As documented in the SDM: Do a CPUID 1 here */
377 sync_core();
378
379 /* get the current revision from MSR 0x8B */
380 native_rdmsr(MSR_IA32_UCODE_REV, val[0], val[1]);
381
382 csig.rev = val[1];
383
384 uci->cpu_sig = csig;
385 uci->valid = 1;
386
387 return 0;
388}
389
390static void show_saved_mc(void)
391{
392#ifdef DEBUG
393 int i, j;
394 unsigned int sig, pf, rev, total_size, data_size, date;
395 struct ucode_cpu_info uci;
396
397 if (!mc_saved_data.num_saved) {
398 pr_debug("no microcode data saved.\n");
399 return;
400 }
401 pr_debug("Total microcode saved: %d\n", mc_saved_data.num_saved);
402
403 collect_cpu_info_early(&uci);
404
405 sig = uci.cpu_sig.sig;
406 pf = uci.cpu_sig.pf;
407 rev = uci.cpu_sig.rev;
408 pr_debug("CPU: sig=0x%x, pf=0x%x, rev=0x%x\n", sig, pf, rev);
409
410 for (i = 0; i < mc_saved_data.num_saved; i++) {
411 struct microcode_header_intel *mc_saved_header;
412 struct extended_sigtable *ext_header;
413 int ext_sigcount;
414 struct extended_signature *ext_sig;
415
416 mc_saved_header = (struct microcode_header_intel *)
417 mc_saved_data.mc_saved[i];
418 sig = mc_saved_header->sig;
419 pf = mc_saved_header->pf;
420 rev = mc_saved_header->rev;
421 total_size = get_totalsize(mc_saved_header);
422 data_size = get_datasize(mc_saved_header);
423 date = mc_saved_header->date;
424
425 pr_debug("mc_saved[%d]: sig=0x%x, pf=0x%x, rev=0x%x, toal size=0x%x, date = %04x-%02x-%02x\n",
426 i, sig, pf, rev, total_size,
427 date & 0xffff,
428 date >> 24,
429 (date >> 16) & 0xff);
430
431 /* Look for ext. headers: */
432 if (total_size <= data_size + MC_HEADER_SIZE)
433 continue;
434
435 ext_header = (void *) mc_saved_header + data_size + MC_HEADER_SIZE;
436 ext_sigcount = ext_header->count;
437 ext_sig = (void *)ext_header + EXT_HEADER_SIZE;
438
439 for (j = 0; j < ext_sigcount; j++) {
440 sig = ext_sig->sig;
441 pf = ext_sig->pf;
442
443 pr_debug("\tExtended[%d]: sig=0x%x, pf=0x%x\n",
444 j, sig, pf);
445
446 ext_sig++;
447 }
448
449 }
450#endif
451}
452
453#ifdef CONFIG_HOTPLUG_CPU
454static DEFINE_MUTEX(x86_cpu_microcode_mutex);
455/*
456 * Save this mc into mc_saved_data. So it will be loaded early when a CPU is
457 * hot added or resumes.
458 *
459 * Please make sure this mc should be a valid microcode patch before calling
460 * this function.
461 */
462int save_mc_for_early(u8 *mc)
463{
464 struct microcode_intel *mc_saved_tmp[MAX_UCODE_COUNT];
465 unsigned int mc_saved_count_init;
466 unsigned int num_saved;
467 struct microcode_intel **mc_saved;
468 int ret = 0;
469 int i;
470
471 /*
472 * Hold hotplug lock so mc_saved_data is not accessed by a CPU in
473 * hotplug.
474 */
475 mutex_lock(&x86_cpu_microcode_mutex);
476
477 mc_saved_count_init = mc_saved_data.num_saved;
478 num_saved = mc_saved_data.num_saved;
479 mc_saved = mc_saved_data.mc_saved;
480
481 if (mc_saved && num_saved)
482 memcpy(mc_saved_tmp, mc_saved,
483 num_saved * sizeof(struct microcode_intel *));
484 /*
485 * Save the microcode patch mc in mc_save_tmp structure if it's a newer
486 * version.
487 */
488 num_saved = _save_mc(mc_saved_tmp, mc, num_saved);
489
490 /*
491 * Save the mc_save_tmp in global mc_saved_data.
492 */
493 ret = save_microcode(&mc_saved_data, mc_saved_tmp, num_saved);
494 if (ret) {
495 pr_err("Cannot save microcode patch.\n");
496 goto out;
497 }
498
499 show_saved_mc();
500
501 /*
502 * Free old saved microcode data.
503 */
504 if (mc_saved) {
505 for (i = 0; i < mc_saved_count_init; i++)
506 kfree(mc_saved[i]);
507 kfree(mc_saved);
508 }
509
510out:
511 mutex_unlock(&x86_cpu_microcode_mutex);
512
513 return ret;
514}
515EXPORT_SYMBOL_GPL(save_mc_for_early);
516#endif
517
518static bool __init load_builtin_intel_microcode(struct cpio_data *cp)
519{
520#ifdef CONFIG_X86_64
521 unsigned int eax = 0x00000001, ebx, ecx = 0, edx;
522 char name[30];
523
524 native_cpuid(&eax, &ebx, &ecx, &edx);
525
526 sprintf(name, "intel-ucode/%02x-%02x-%02x",
527 x86_family(eax), x86_model(eax), x86_stepping(eax));
528
529 return get_builtin_firmware(cp, name);
530#else
531 return false;
532#endif
533}
534
535static __initdata char ucode_name[] = "kernel/x86/microcode/GenuineIntel.bin";
536static __init enum ucode_state
537scan_microcode(struct mc_saved_data *mcs, unsigned long *mc_ptrs,
538 unsigned long start, unsigned long size,
539 struct ucode_cpu_info *uci)
540{
541 struct cpio_data cd;
542 long offset = 0;
543#ifdef CONFIG_X86_32
544 char *p = (char *)__pa_nodebug(ucode_name);
545#else
546 char *p = ucode_name;
547#endif
548
549 cd.data = NULL;
550 cd.size = 0;
551
552 /* try built-in microcode if no initrd */
553 if (!size) {
554 if (!load_builtin_intel_microcode(&cd))
555 return UCODE_ERROR;
556 } else {
557 cd = find_cpio_data(p, (void *)start, size, &offset);
558 if (!cd.data)
559 return UCODE_ERROR;
560 }
561
562 return get_matching_model_microcode(start, cd.data, cd.size,
563 mcs, mc_ptrs, uci);
564}
565
566/*
567 * Print ucode update info.
568 */
569static void
570print_ucode_info(struct ucode_cpu_info *uci, unsigned int date)
571{
572 pr_info_once("microcode updated early to revision 0x%x, date = %04x-%02x-%02x\n",
573 uci->cpu_sig.rev,
574 date & 0xffff,
575 date >> 24,
576 (date >> 16) & 0xff);
577}
578
579#ifdef CONFIG_X86_32
580
581static int delay_ucode_info;
582static int current_mc_date;
583
584/*
585 * Print early updated ucode info after printk works. This is delayed info dump.
586 */
587void show_ucode_info_early(void)
588{
589 struct ucode_cpu_info uci;
590
591 if (delay_ucode_info) {
592 collect_cpu_info_early(&uci);
593 print_ucode_info(&uci, current_mc_date);
594 delay_ucode_info = 0;
595 }
596}
597
598/*
599 * At this point, we can not call printk() yet. Keep microcode patch number in
600 * mc_saved_data.mc_saved and delay printing microcode info in
601 * show_ucode_info_early() until printk() works.
602 */
603static void print_ucode(struct ucode_cpu_info *uci)
604{
605 struct microcode_intel *mc;
606 int *delay_ucode_info_p;
607 int *current_mc_date_p;
608
609 mc = uci->mc;
610 if (!mc)
611 return;
612
613 delay_ucode_info_p = (int *)__pa_nodebug(&delay_ucode_info);
614 current_mc_date_p = (int *)__pa_nodebug(¤t_mc_date);
615
616 *delay_ucode_info_p = 1;
617 *current_mc_date_p = mc->hdr.date;
618}
619#else
620
621/*
622 * Flush global tlb. We only do this in x86_64 where paging has been enabled
623 * already and PGE should be enabled as well.
624 */
625static inline void flush_tlb_early(void)
626{
627 __native_flush_tlb_global_irq_disabled();
628}
629
630static inline void print_ucode(struct ucode_cpu_info *uci)
631{
632 struct microcode_intel *mc;
633
634 mc = uci->mc;
635 if (!mc)
636 return;
637
638 print_ucode_info(uci, mc->hdr.date);
639}
640#endif
641
642static int apply_microcode_early(struct ucode_cpu_info *uci, bool early)
643{
644 struct microcode_intel *mc;
645 unsigned int val[2];
646
647 mc = uci->mc;
648 if (!mc)
649 return 0;
650
651 /* write microcode via MSR 0x79 */
652 native_wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
653 native_wrmsrl(MSR_IA32_UCODE_REV, 0);
654
655 /* As documented in the SDM: Do a CPUID 1 here */
656 sync_core();
657
658 /* get the current revision from MSR 0x8B */
659 native_rdmsr(MSR_IA32_UCODE_REV, val[0], val[1]);
660 if (val[1] != mc->hdr.rev)
661 return -1;
662
663#ifdef CONFIG_X86_64
664 /* Flush global tlb. This is precaution. */
665 flush_tlb_early();
666#endif
667 uci->cpu_sig.rev = val[1];
668
669 if (early)
670 print_ucode(uci);
671 else
672 print_ucode_info(uci, mc->hdr.date);
673
674 return 0;
675}
676
677/*
678 * This function converts microcode patch offsets previously stored in
679 * mc_tmp_ptrs to pointers and stores the pointers in mc_saved_data.
680 */
681int __init save_microcode_in_initrd_intel(void)
682{
683 unsigned int count = mc_saved_data.num_saved;
684 struct microcode_intel *mc_saved[MAX_UCODE_COUNT];
685 int ret = 0;
686
687 if (!count)
688 return ret;
689
690 copy_ptrs(mc_saved, mc_tmp_ptrs, get_initrd_start(), count);
691
692 ret = save_microcode(&mc_saved_data, mc_saved, count);
693 if (ret)
694 pr_err("Cannot save microcode patches from initrd.\n");
695
696 show_saved_mc();
697
698 return ret;
699}
700
701static void __init
702_load_ucode_intel_bsp(struct mc_saved_data *mcs, unsigned long *mc_ptrs,
703 unsigned long start, unsigned long size)
704{
705 struct ucode_cpu_info uci;
706 enum ucode_state ret;
707
708 collect_cpu_info_early(&uci);
709
710 ret = scan_microcode(mcs, mc_ptrs, start, size, &uci);
711 if (ret != UCODE_OK)
712 return;
713
714 ret = load_microcode(mcs, mc_ptrs, start, &uci);
715 if (ret != UCODE_OK)
716 return;
717
718 apply_microcode_early(&uci, true);
719}
720
721void __init load_ucode_intel_bsp(void)
722{
723 u64 start, size;
724#ifdef CONFIG_X86_32
725 struct boot_params *p;
726
727 p = (struct boot_params *)__pa_nodebug(&boot_params);
728 size = p->hdr.ramdisk_size;
729
730 /*
731 * Set start only if we have an initrd image. We cannot use initrd_start
732 * because it is not set that early yet.
733 */
734 start = (size ? p->hdr.ramdisk_image : 0);
735
736 _load_ucode_intel_bsp((struct mc_saved_data *)__pa_nodebug(&mc_saved_data),
737 (unsigned long *)__pa_nodebug(&mc_tmp_ptrs),
738 start, size);
739#else
740 size = boot_params.hdr.ramdisk_size;
741 start = (size ? boot_params.hdr.ramdisk_image + PAGE_OFFSET : 0);
742
743 _load_ucode_intel_bsp(&mc_saved_data, mc_tmp_ptrs, start, size);
744#endif
745}
746
747void load_ucode_intel_ap(void)
748{
749 unsigned long *mcs_tmp_p;
750 struct mc_saved_data *mcs_p;
751 struct ucode_cpu_info uci;
752 enum ucode_state ret;
753#ifdef CONFIG_X86_32
754
755 mcs_tmp_p = (unsigned long *)__pa_nodebug(mc_tmp_ptrs);
756 mcs_p = (struct mc_saved_data *)__pa_nodebug(&mc_saved_data);
757#else
758 mcs_tmp_p = mc_tmp_ptrs;
759 mcs_p = &mc_saved_data;
760#endif
761
762 /*
763 * If there is no valid ucode previously saved in memory, no need to
764 * update ucode on this AP.
765 */
766 if (!mcs_p->num_saved)
767 return;
768
769 collect_cpu_info_early(&uci);
770 ret = load_microcode(mcs_p, mcs_tmp_p, get_initrd_start_addr(), &uci);
771 if (ret != UCODE_OK)
772 return;
773
774 apply_microcode_early(&uci, true);
775}
776
777void reload_ucode_intel(void)
778{
779 struct ucode_cpu_info uci;
780 enum ucode_state ret;
781
782 if (!mc_saved_data.num_saved)
783 return;
784
785 collect_cpu_info_early(&uci);
786
787 ret = load_microcode_early(mc_saved_data.mc_saved,
788 mc_saved_data.num_saved, &uci);
789 if (ret != UCODE_OK)
790 return;
791
792 apply_microcode_early(&uci, false);
793}
794
795static int collect_cpu_info(int cpu_num, struct cpu_signature *csig)
796{
797 struct cpuinfo_x86 *c = &cpu_data(cpu_num);
798 unsigned int val[2];
799
800 memset(csig, 0, sizeof(*csig));
801
802 csig->sig = cpuid_eax(0x00000001);
803
804 if ((c->x86_model >= 5) || (c->x86 > 6)) {
805 /* get processor flags from MSR 0x17 */
806 rdmsr(MSR_IA32_PLATFORM_ID, val[0], val[1]);
807 csig->pf = 1 << ((val[1] >> 18) & 7);
808 }
809
810 csig->rev = c->microcode;
811 pr_info("CPU%d sig=0x%x, pf=0x%x, revision=0x%x\n",
812 cpu_num, csig->sig, csig->pf, csig->rev);
813
814 return 0;
815}
816
817/*
818 * return 0 - no update found
819 * return 1 - found update
820 */
821static int get_matching_mc(struct microcode_intel *mc, int cpu)
822{
823 struct cpu_signature cpu_sig;
824 unsigned int csig, cpf, crev;
825
826 collect_cpu_info(cpu, &cpu_sig);
827
828 csig = cpu_sig.sig;
829 cpf = cpu_sig.pf;
830 crev = cpu_sig.rev;
831
832 return has_newer_microcode(mc, csig, cpf, crev);
833}
834
835static int apply_microcode_intel(int cpu)
836{
837 struct microcode_intel *mc;
838 struct ucode_cpu_info *uci;
839 struct cpuinfo_x86 *c;
840 unsigned int val[2];
841
842 /* We should bind the task to the CPU */
843 if (WARN_ON(raw_smp_processor_id() != cpu))
844 return -1;
845
846 uci = ucode_cpu_info + cpu;
847 mc = uci->mc;
848 if (!mc)
849 return 0;
850
851 /*
852 * Microcode on this CPU could be updated earlier. Only apply the
853 * microcode patch in mc when it is newer than the one on this
854 * CPU.
855 */
856 if (!get_matching_mc(mc, cpu))
857 return 0;
858
859 /* write microcode via MSR 0x79 */
860 wrmsrl(MSR_IA32_UCODE_WRITE, (unsigned long)mc->bits);
861 wrmsrl(MSR_IA32_UCODE_REV, 0);
862
863 /* As documented in the SDM: Do a CPUID 1 here */
864 sync_core();
865
866 /* get the current revision from MSR 0x8B */
867 rdmsr(MSR_IA32_UCODE_REV, val[0], val[1]);
868
869 if (val[1] != mc->hdr.rev) {
870 pr_err("CPU%d update to revision 0x%x failed\n",
871 cpu, mc->hdr.rev);
872 return -1;
873 }
874
875 pr_info("CPU%d updated to revision 0x%x, date = %04x-%02x-%02x\n",
876 cpu, val[1],
877 mc->hdr.date & 0xffff,
878 mc->hdr.date >> 24,
879 (mc->hdr.date >> 16) & 0xff);
880
881 c = &cpu_data(cpu);
882
883 uci->cpu_sig.rev = val[1];
884 c->microcode = val[1];
885
886 return 0;
887}
888
889static enum ucode_state generic_load_microcode(int cpu, void *data, size_t size,
890 int (*get_ucode_data)(void *, const void *, size_t))
891{
892 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
893 u8 *ucode_ptr = data, *new_mc = NULL, *mc = NULL;
894 int new_rev = uci->cpu_sig.rev;
895 unsigned int leftover = size;
896 enum ucode_state state = UCODE_OK;
897 unsigned int curr_mc_size = 0;
898 unsigned int csig, cpf;
899
900 while (leftover) {
901 struct microcode_header_intel mc_header;
902 unsigned int mc_size;
903
904 if (leftover < sizeof(mc_header)) {
905 pr_err("error! Truncated header in microcode data file\n");
906 break;
907 }
908
909 if (get_ucode_data(&mc_header, ucode_ptr, sizeof(mc_header)))
910 break;
911
912 mc_size = get_totalsize(&mc_header);
913 if (!mc_size || mc_size > leftover) {
914 pr_err("error! Bad data in microcode data file\n");
915 break;
916 }
917
918 /* For performance reasons, reuse mc area when possible */
919 if (!mc || mc_size > curr_mc_size) {
920 vfree(mc);
921 mc = vmalloc(mc_size);
922 if (!mc)
923 break;
924 curr_mc_size = mc_size;
925 }
926
927 if (get_ucode_data(mc, ucode_ptr, mc_size) ||
928 microcode_sanity_check(mc, 1) < 0) {
929 break;
930 }
931
932 csig = uci->cpu_sig.sig;
933 cpf = uci->cpu_sig.pf;
934 if (has_newer_microcode(mc, csig, cpf, new_rev)) {
935 vfree(new_mc);
936 new_rev = mc_header.rev;
937 new_mc = mc;
938 mc = NULL; /* trigger new vmalloc */
939 }
940
941 ucode_ptr += mc_size;
942 leftover -= mc_size;
943 }
944
945 vfree(mc);
946
947 if (leftover) {
948 vfree(new_mc);
949 state = UCODE_ERROR;
950 goto out;
951 }
952
953 if (!new_mc) {
954 state = UCODE_NFOUND;
955 goto out;
956 }
957
958 vfree(uci->mc);
959 uci->mc = (struct microcode_intel *)new_mc;
960
961 /*
962 * If early loading microcode is supported, save this mc into
963 * permanent memory. So it will be loaded early when a CPU is hot added
964 * or resumes.
965 */
966 save_mc_for_early(new_mc);
967
968 pr_debug("CPU%d found a matching microcode update with version 0x%x (current=0x%x)\n",
969 cpu, new_rev, uci->cpu_sig.rev);
970out:
971 return state;
972}
973
974static int get_ucode_fw(void *to, const void *from, size_t n)
975{
976 memcpy(to, from, n);
977 return 0;
978}
979
980static enum ucode_state request_microcode_fw(int cpu, struct device *device,
981 bool refresh_fw)
982{
983 char name[30];
984 struct cpuinfo_x86 *c = &cpu_data(cpu);
985 const struct firmware *firmware;
986 enum ucode_state ret;
987
988 sprintf(name, "intel-ucode/%02x-%02x-%02x",
989 c->x86, c->x86_model, c->x86_mask);
990
991 if (request_firmware_direct(&firmware, name, device)) {
992 pr_debug("data file %s load failed\n", name);
993 return UCODE_NFOUND;
994 }
995
996 ret = generic_load_microcode(cpu, (void *)firmware->data,
997 firmware->size, &get_ucode_fw);
998
999 release_firmware(firmware);
1000
1001 return ret;
1002}
1003
1004static int get_ucode_user(void *to, const void *from, size_t n)
1005{
1006 return copy_from_user(to, from, n);
1007}
1008
1009static enum ucode_state
1010request_microcode_user(int cpu, const void __user *buf, size_t size)
1011{
1012 return generic_load_microcode(cpu, (void *)buf, size, &get_ucode_user);
1013}
1014
1015static void microcode_fini_cpu(int cpu)
1016{
1017 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
1018
1019 vfree(uci->mc);
1020 uci->mc = NULL;
1021}
1022
1023static struct microcode_ops microcode_intel_ops = {
1024 .request_microcode_user = request_microcode_user,
1025 .request_microcode_fw = request_microcode_fw,
1026 .collect_cpu_info = collect_cpu_info,
1027 .apply_microcode = apply_microcode_intel,
1028 .microcode_fini_cpu = microcode_fini_cpu,
1029};
1030
1031struct microcode_ops * __init init_intel_microcode(void)
1032{
1033 struct cpuinfo_x86 *c = &boot_cpu_data;
1034
1035 if (c->x86_vendor != X86_VENDOR_INTEL || c->x86 < 6 ||
1036 cpu_has(c, X86_FEATURE_IA64)) {
1037 pr_err("Intel CPU family 0x%x not supported\n", c->x86);
1038 return NULL;
1039 }
1040
1041 return µcode_intel_ops;
1042}
1043