Linux Audio

Check our new training course

Loading...
v5.4
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 *  linux/arch/i386/kernel/head32.c -- prepare to run common code
  4 *
  5 *  Copyright (C) 2000 Andrea Arcangeli <andrea@suse.de> SuSE
  6 *  Copyright (C) 2007 Eric Biederman <ebiederm@xmission.com>
  7 */
  8
  9#include <linux/init.h>
 10#include <linux/start_kernel.h>
 11#include <linux/mm.h>
 12#include <linux/memblock.h>
 13
 14#include <asm/desc.h>
 15#include <asm/setup.h>
 16#include <asm/sections.h>
 17#include <asm/e820/api.h>
 18#include <asm/page.h>
 19#include <asm/apic.h>
 20#include <asm/io_apic.h>
 21#include <asm/bios_ebda.h>
 
 22#include <asm/tlbflush.h>
 23#include <asm/bootparam_utils.h>
 24
 25static void __init i386_default_early_setup(void)
 26{
 27	/* Initialize 32bit specific setup functions */
 28	x86_init.resources.reserve_resources = i386_reserve_resources;
 29	x86_init.mpparse.setup_ioapic_ids = setup_ioapic_ids_from_mpc;
 30}
 31
 32asmlinkage __visible void __init i386_start_kernel(void)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 33{
 34	/* Make sure IDT is set up before any exception happens */
 35	idt_setup_early_handler();
 36
 
 
 
 37	cr4_init_shadow();
 38
 39	sanitize_boot_params(&boot_params);
 40
 41	x86_early_init_platform_quirks();
 42
 43	/* Call the subarch specific early setup function */
 44	switch (boot_params.hdr.hardware_subarch) {
 45	case X86_SUBARCH_INTEL_MID:
 46		x86_intel_mid_early_setup();
 47		break;
 48	case X86_SUBARCH_CE4100:
 49		x86_ce4100_early_setup();
 50		break;
 51	default:
 52		i386_default_early_setup();
 53		break;
 54	}
 55
 56	start_kernel();
 57}
 58
 59/*
 60 * Initialize page tables.  This creates a PDE and a set of page
 61 * tables, which are located immediately beyond __brk_base.  The variable
 62 * _brk_end is set up to point to the first "safe" location.
 63 * Mappings are created both at virtual address 0 (identity mapping)
 64 * and PAGE_OFFSET for up to _end.
 65 *
 66 * In PAE mode initial_page_table is statically defined to contain
 67 * enough entries to cover the VMSPLIT option (that is the top 1, 2 or 3
 68 * entries). The identity mapping is handled by pointing two PGD entries
 69 * to the first kernel PMD. Note the upper half of each PMD or PTE are
 70 * always zero at this stage.
 71 */
 72void __init mk_early_pgtbl_32(void)
 73{
 74#ifdef __pa
 75#undef __pa
 76#endif
 77#define __pa(x)  ((unsigned long)(x) - PAGE_OFFSET)
 78	pte_t pte, *ptep;
 79	int i;
 80	unsigned long *ptr;
 81	/* Enough space to fit pagetables for the low memory linear map */
 82	const unsigned long limit = __pa(_end) +
 83		(PAGE_TABLE_SIZE(LOWMEM_PAGES) << PAGE_SHIFT);
 84#ifdef CONFIG_X86_PAE
 85	pmd_t pl2, *pl2p = (pmd_t *)__pa(initial_pg_pmd);
 86#define SET_PL2(pl2, val)    { (pl2).pmd = (val); }
 
 87#else
 88	pgd_t pl2, *pl2p = (pgd_t *)__pa(initial_page_table);
 89#define SET_PL2(pl2, val)   { (pl2).pgd = (val); }
 
 90#endif
 91
 92	ptep = (pte_t *)__pa(__brk_base);
 93	pte.pte = PTE_IDENT_ATTR;
 94
 95	while ((pte.pte & PTE_PFN_MASK) < limit) {
 
 
 
 
 
 
 
 
 96
 97		SET_PL2(pl2, (unsigned long)ptep | PDE_IDENT_ATTR);
 98		*pl2p = pl2;
 99#ifndef CONFIG_X86_PAE
100		/* Kernel PDE entry */
101		*(pl2p +  ((PAGE_OFFSET >> PGDIR_SHIFT))) = pl2;
102#endif
103		for (i = 0; i < PTRS_PER_PTE; i++) {
104			*ptep = pte;
105			pte.pte += PAGE_SIZE;
106			ptep++;
107		}
108
109		pl2p++;
110	}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
111
112	ptr = (unsigned long *)__pa(&max_pfn_mapped);
113	/* Can't use pte_pfn() since it's a call with CONFIG_PARAVIRT */
114	*ptr = (pte.pte & PTE_PFN_MASK) >> PAGE_SHIFT;
115
116	ptr = (unsigned long *)__pa(&_brk_end);
117	*ptr = (unsigned long)ptep + PAGE_OFFSET;
118}
119
v6.8
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 *  linux/arch/i386/kernel/head32.c -- prepare to run common code
  4 *
  5 *  Copyright (C) 2000 Andrea Arcangeli <andrea@suse.de> SuSE
  6 *  Copyright (C) 2007 Eric Biederman <ebiederm@xmission.com>
  7 */
  8
  9#include <linux/init.h>
 10#include <linux/start_kernel.h>
 11#include <linux/mm.h>
 12#include <linux/memblock.h>
 13
 14#include <asm/desc.h>
 15#include <asm/setup.h>
 16#include <asm/sections.h>
 17#include <asm/e820/api.h>
 18#include <asm/page.h>
 19#include <asm/apic.h>
 20#include <asm/io_apic.h>
 21#include <asm/bios_ebda.h>
 22#include <asm/microcode.h>
 23#include <asm/tlbflush.h>
 24#include <asm/bootparam_utils.h>
 25
 26static void __init i386_default_early_setup(void)
 27{
 28	/* Initialize 32bit specific setup functions */
 29	x86_init.resources.reserve_resources = i386_reserve_resources;
 30	x86_init.mpparse.setup_ioapic_ids = setup_ioapic_ids_from_mpc;
 31}
 32
 33#ifdef CONFIG_MICROCODE_INITRD32
 34unsigned long __initdata initrd_start_early;
 35static pte_t __initdata *initrd_pl2p_start, *initrd_pl2p_end;
 36
 37static void zap_early_initrd_mapping(void)
 38{
 39	pte_t *pl2p = initrd_pl2p_start;
 40
 41	for (; pl2p < initrd_pl2p_end; pl2p++) {
 42		*pl2p = (pte_t){ .pte = 0 };
 43
 44		if (!IS_ENABLED(CONFIG_X86_PAE))
 45			*(pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = (pte_t) {.pte = 0};
 46	}
 47}
 48#else
 49static inline void zap_early_initrd_mapping(void) { }
 50#endif
 51
 52asmlinkage __visible void __init __noreturn i386_start_kernel(void)
 53{
 54	/* Make sure IDT is set up before any exception happens */
 55	idt_setup_early_handler();
 56
 57	load_ucode_bsp();
 58	zap_early_initrd_mapping();
 59
 60	cr4_init_shadow();
 61
 62	sanitize_boot_params(&boot_params);
 63
 64	x86_early_init_platform_quirks();
 65
 66	/* Call the subarch specific early setup function */
 67	switch (boot_params.hdr.hardware_subarch) {
 68	case X86_SUBARCH_INTEL_MID:
 69		x86_intel_mid_early_setup();
 70		break;
 71	case X86_SUBARCH_CE4100:
 72		x86_ce4100_early_setup();
 73		break;
 74	default:
 75		i386_default_early_setup();
 76		break;
 77	}
 78
 79	start_kernel();
 80}
 81
 82/*
 83 * Initialize page tables.  This creates a PDE and a set of page
 84 * tables, which are located immediately beyond __brk_base.  The variable
 85 * _brk_end is set up to point to the first "safe" location.
 86 * Mappings are created both at virtual address 0 (identity mapping)
 87 * and PAGE_OFFSET for up to _end.
 88 *
 89 * In PAE mode initial_page_table is statically defined to contain
 90 * enough entries to cover the VMSPLIT option (that is the top 1, 2 or 3
 91 * entries). The identity mapping is handled by pointing two PGD entries
 92 * to the first kernel PMD. Note the upper half of each PMD or PTE are
 93 * always zero at this stage.
 94 */
 
 
 
 
 
 
 
 
 
 
 
 
 95#ifdef CONFIG_X86_PAE
 96typedef pmd_t			pl2_t;
 97#define pl2_base		initial_pg_pmd
 98#define SET_PL2(val)		{ .pmd = (val), }
 99#else
100typedef pgd_t			pl2_t;
101#define pl2_base		initial_page_table
102#define SET_PL2(val)		{ .pgd = (val), }
103#endif
104
105static __init __no_stack_protector pte_t init_map(pte_t pte, pte_t **ptep, pl2_t **pl2p,
106						  const unsigned long limit)
107{
108	while ((pte.pte & PTE_PFN_MASK) < limit) {
109		pl2_t pl2 = SET_PL2((unsigned long)*ptep | PDE_IDENT_ATTR);
110		int i;
111
112		**pl2p = pl2;
113		if (!IS_ENABLED(CONFIG_X86_PAE)) {
114			/* Kernel PDE entry */
115			*(*pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = pl2;
116		}
117
 
 
 
 
 
 
118		for (i = 0; i < PTRS_PER_PTE; i++) {
119			**ptep = pte;
120			pte.pte += PAGE_SIZE;
121			(*ptep)++;
122		}
123		(*pl2p)++;
 
124	}
125	return pte;
126}
127
128void __init __no_stack_protector mk_early_pgtbl_32(void)
129{
130	/* Enough space to fit pagetables for the low memory linear map */
131	unsigned long limit = __pa_nodebug(_end) + (PAGE_TABLE_SIZE(LOWMEM_PAGES) << PAGE_SHIFT);
132	pte_t pte, *ptep = (pte_t *)__pa_nodebug(__brk_base);
133	struct boot_params __maybe_unused *params;
134	pl2_t *pl2p = (pl2_t *)__pa_nodebug(pl2_base);
135	unsigned long *ptr;
136
137	pte.pte = PTE_IDENT_ATTR;
138	pte = init_map(pte, &ptep, &pl2p, limit);
139
140	ptr = (unsigned long *)__pa_nodebug(&max_pfn_mapped);
141	/* Can't use pte_pfn() since it's a call with CONFIG_PARAVIRT */
142	*ptr = (pte.pte & PTE_PFN_MASK) >> PAGE_SHIFT;
143
144	ptr = (unsigned long *)__pa_nodebug(&_brk_end);
145	*ptr = (unsigned long)ptep + PAGE_OFFSET;
 
146
147#ifdef CONFIG_MICROCODE_INITRD32
148	/* Running on a hypervisor? */
149	if (native_cpuid_ecx(1) & BIT(31))
150		return;
151
152	params = (struct boot_params *)__pa_nodebug(&boot_params);
153	if (!params->hdr.ramdisk_size || !params->hdr.ramdisk_image)
154		return;
155
156	/* Save the virtual start address */
157	ptr = (unsigned long *)__pa_nodebug(&initrd_start_early);
158	*ptr = (pte.pte & PTE_PFN_MASK) + PAGE_OFFSET;
159	*ptr += ((unsigned long)params->hdr.ramdisk_image) & ~PAGE_MASK;
160
161	/* Save PLP2 for cleanup */
162	ptr = (unsigned long *)__pa_nodebug(&initrd_pl2p_start);
163	*ptr = (unsigned long)pl2p + PAGE_OFFSET;
164
165	limit = (unsigned long)params->hdr.ramdisk_image;
166	pte.pte = PTE_IDENT_ATTR | PFN_ALIGN(limit);
167	limit = (unsigned long)params->hdr.ramdisk_image + params->hdr.ramdisk_size;
168
169	init_map(pte, &ptep, &pl2p, limit);
170
171	ptr = (unsigned long *)__pa_nodebug(&initrd_pl2p_end);
172	*ptr = (unsigned long)pl2p + PAGE_OFFSET;
173#endif
174}