Linux Audio

Check our new training course

Loading...
v4.6
 
 1/*
 2 * Copyright (C) 2015 Linaro Ltd <ard.biesheuvel@linaro.org>
 3 *
 4 * This program is free software; you can redistribute it and/or modify
 5 * it under the terms of the GNU General Public License version 2 as
 6 * published by the Free Software Foundation.
 7 */
 8
 9#include <linux/efi.h>
 
10#include <asm/efi.h>
11#include <asm/mach/map.h>
12#include <asm/mmu_context.h>
13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
15{
16	struct map_desc desc = {
17		.virtual	= md->virt_addr,
18		.pfn		= __phys_to_pfn(md->phys_addr),
19		.length		= md->num_pages * EFI_PAGE_SIZE,
20	};
21
22	/*
23	 * Order is important here: memory regions may have all of the
24	 * bits below set (and usually do), so we check them in order of
25	 * preference.
26	 */
27	if (md->attribute & EFI_MEMORY_WB)
28		desc.type = MT_MEMORY_RWX;
29	else if (md->attribute & EFI_MEMORY_WT)
30		desc.type = MT_MEMORY_RWX_NONCACHED;
31	else if (md->attribute & EFI_MEMORY_WC)
32		desc.type = MT_DEVICE_WC;
33	else
34		desc.type = MT_DEVICE;
35
36	create_mapping_late(mm, &desc, true);
 
 
 
 
 
 
37	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38}
v6.2
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Copyright (C) 2015 Linaro Ltd <ard.biesheuvel@linaro.org>
 
 
 
 
  4 */
  5
  6#include <linux/efi.h>
  7#include <linux/memblock.h>
  8#include <asm/efi.h>
  9#include <asm/mach/map.h>
 10#include <asm/mmu_context.h>
 11
 12static int __init set_permissions(pte_t *ptep, unsigned long addr, void *data)
 13{
 14	efi_memory_desc_t *md = data;
 15	pte_t pte = *ptep;
 16
 17	if (md->attribute & EFI_MEMORY_RO)
 18		pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
 19	if (md->attribute & EFI_MEMORY_XP)
 20		pte = set_pte_bit(pte, __pgprot(L_PTE_XN));
 21	set_pte_ext(ptep, pte, PTE_EXT_NG);
 22	return 0;
 23}
 24
 25int __init efi_set_mapping_permissions(struct mm_struct *mm,
 26				       efi_memory_desc_t *md)
 27{
 28	unsigned long base, size;
 29
 30	base = md->virt_addr;
 31	size = md->num_pages << EFI_PAGE_SHIFT;
 32
 33	/*
 34	 * We can only use apply_to_page_range() if we can guarantee that the
 35	 * entire region was mapped using pages. This should be the case if the
 36	 * region does not cover any naturally aligned SECTION_SIZE sized
 37	 * blocks.
 38	 */
 39	if (round_down(base + size, SECTION_SIZE) <
 40	    round_up(base, SECTION_SIZE) + SECTION_SIZE)
 41		return apply_to_page_range(mm, base, size, set_permissions, md);
 42
 43	return 0;
 44}
 45
 46int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
 47{
 48	struct map_desc desc = {
 49		.virtual	= md->virt_addr,
 50		.pfn		= __phys_to_pfn(md->phys_addr),
 51		.length		= md->num_pages * EFI_PAGE_SIZE,
 52	};
 53
 54	/*
 55	 * Order is important here: memory regions may have all of the
 56	 * bits below set (and usually do), so we check them in order of
 57	 * preference.
 58	 */
 59	if (md->attribute & EFI_MEMORY_WB)
 60		desc.type = MT_MEMORY_RWX;
 61	else if (md->attribute & EFI_MEMORY_WT)
 62		desc.type = MT_MEMORY_RWX_NONCACHED;
 63	else if (md->attribute & EFI_MEMORY_WC)
 64		desc.type = MT_DEVICE_WC;
 65	else
 66		desc.type = MT_DEVICE;
 67
 68	create_mapping_late(mm, &desc, true);
 69
 70	/*
 71	 * If stricter permissions were specified, apply them now.
 72	 */
 73	if (md->attribute & (EFI_MEMORY_RO | EFI_MEMORY_XP))
 74		return efi_set_mapping_permissions(mm, md);
 75	return 0;
 76}
 77
 78static unsigned long __initdata cpu_state_table = EFI_INVALID_TABLE_ADDR;
 79
 80const efi_config_table_type_t efi_arch_tables[] __initconst = {
 81	{LINUX_EFI_ARM_CPU_STATE_TABLE_GUID, &cpu_state_table},
 82	{}
 83};
 84
 85static void __init load_cpu_state_table(void)
 86{
 87	if (cpu_state_table != EFI_INVALID_TABLE_ADDR) {
 88		struct efi_arm_entry_state *state;
 89		bool dump_state = true;
 90
 91		state = early_memremap_ro(cpu_state_table,
 92					  sizeof(struct efi_arm_entry_state));
 93		if (state == NULL) {
 94			pr_warn("Unable to map CPU entry state table.\n");
 95			return;
 96		}
 97
 98		if ((state->sctlr_before_ebs & 1) == 0)
 99			pr_warn(FW_BUG "EFI stub was entered with MMU and Dcache disabled, please fix your firmware!\n");
100		else if ((state->sctlr_after_ebs & 1) == 0)
101			pr_warn(FW_BUG "ExitBootServices() returned with MMU and Dcache disabled, please fix your firmware!\n");
102		else
103			dump_state = false;
104
105		if (dump_state || efi_enabled(EFI_DBG)) {
106			pr_info("CPSR at EFI stub entry        : 0x%08x\n",
107				state->cpsr_before_ebs);
108			pr_info("SCTLR at EFI stub entry       : 0x%08x\n",
109				state->sctlr_before_ebs);
110			pr_info("CPSR after ExitBootServices() : 0x%08x\n",
111				state->cpsr_after_ebs);
112			pr_info("SCTLR after ExitBootServices(): 0x%08x\n",
113				state->sctlr_after_ebs);
114		}
115		early_memunmap(state, sizeof(struct efi_arm_entry_state));
116	}
117}
118
119void __init arm_efi_init(void)
120{
121	efi_init();
122
123	if (screen_info.orig_video_isVGA == VIDEO_TYPE_EFI) {
124		/* dummycon on ARM needs non-zero values for columns/lines */
125		screen_info.orig_video_cols = 80;
126		screen_info.orig_video_lines = 25;
127	}
128
129	/* ARM does not permit early mappings to persist across paging_init() */
130	efi_memmap_unmap();
131
132	load_cpu_state_table();
133}