Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright (C) 2015 Linaro Ltd <ard.biesheuvel@linaro.org>
4 */
5
6#include <linux/efi.h>
7#include <linux/memblock.h>
8#include <linux/screen_info.h>
9
10#include <asm/efi.h>
11#include <asm/mach/map.h>
12#include <asm/mmu_context.h>
13
14static int __init set_permissions(pte_t *ptep, unsigned long addr, void *data)
15{
16 efi_memory_desc_t *md = data;
17 pte_t pte = *ptep;
18
19 if (md->attribute & EFI_MEMORY_RO)
20 pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
21 if (md->attribute & EFI_MEMORY_XP)
22 pte = set_pte_bit(pte, __pgprot(L_PTE_XN));
23 set_pte_ext(ptep, pte, PTE_EXT_NG);
24 return 0;
25}
26
27int __init efi_set_mapping_permissions(struct mm_struct *mm,
28 efi_memory_desc_t *md,
29 bool ignored)
30{
31 unsigned long base, size;
32
33 base = md->virt_addr;
34 size = md->num_pages << EFI_PAGE_SHIFT;
35
36 /*
37 * We can only use apply_to_page_range() if we can guarantee that the
38 * entire region was mapped using pages. This should be the case if the
39 * region does not cover any naturally aligned SECTION_SIZE sized
40 * blocks.
41 */
42 if (round_down(base + size, SECTION_SIZE) <
43 round_up(base, SECTION_SIZE) + SECTION_SIZE)
44 return apply_to_page_range(mm, base, size, set_permissions, md);
45
46 return 0;
47}
48
49int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
50{
51 struct map_desc desc = {
52 .virtual = md->virt_addr,
53 .pfn = __phys_to_pfn(md->phys_addr),
54 .length = md->num_pages * EFI_PAGE_SIZE,
55 };
56
57 /*
58 * Order is important here: memory regions may have all of the
59 * bits below set (and usually do), so we check them in order of
60 * preference.
61 */
62 if (md->attribute & EFI_MEMORY_WB)
63 desc.type = MT_MEMORY_RWX;
64 else if (md->attribute & EFI_MEMORY_WT)
65 desc.type = MT_MEMORY_RWX_NONCACHED;
66 else if (md->attribute & EFI_MEMORY_WC)
67 desc.type = MT_DEVICE_WC;
68 else
69 desc.type = MT_DEVICE;
70
71 create_mapping_late(mm, &desc, true);
72
73 /*
74 * If stricter permissions were specified, apply them now.
75 */
76 if (md->attribute & (EFI_MEMORY_RO | EFI_MEMORY_XP))
77 return efi_set_mapping_permissions(mm, md, false);
78 return 0;
79}
80
81static unsigned long __initdata cpu_state_table = EFI_INVALID_TABLE_ADDR;
82
83const efi_config_table_type_t efi_arch_tables[] __initconst = {
84 {LINUX_EFI_ARM_CPU_STATE_TABLE_GUID, &cpu_state_table},
85 {}
86};
87
88static void __init load_cpu_state_table(void)
89{
90 if (cpu_state_table != EFI_INVALID_TABLE_ADDR) {
91 struct efi_arm_entry_state *state;
92 bool dump_state = true;
93
94 state = early_memremap_ro(cpu_state_table,
95 sizeof(struct efi_arm_entry_state));
96 if (state == NULL) {
97 pr_warn("Unable to map CPU entry state table.\n");
98 return;
99 }
100
101 if ((state->sctlr_before_ebs & 1) == 0)
102 pr_warn(FW_BUG "EFI stub was entered with MMU and Dcache disabled, please fix your firmware!\n");
103 else if ((state->sctlr_after_ebs & 1) == 0)
104 pr_warn(FW_BUG "ExitBootServices() returned with MMU and Dcache disabled, please fix your firmware!\n");
105 else
106 dump_state = false;
107
108 if (dump_state || efi_enabled(EFI_DBG)) {
109 pr_info("CPSR at EFI stub entry : 0x%08x\n",
110 state->cpsr_before_ebs);
111 pr_info("SCTLR at EFI stub entry : 0x%08x\n",
112 state->sctlr_before_ebs);
113 pr_info("CPSR after ExitBootServices() : 0x%08x\n",
114 state->cpsr_after_ebs);
115 pr_info("SCTLR after ExitBootServices(): 0x%08x\n",
116 state->sctlr_after_ebs);
117 }
118 early_memunmap(state, sizeof(struct efi_arm_entry_state));
119 }
120}
121
122void __init arm_efi_init(void)
123{
124 efi_init();
125
126 /* ARM does not permit early mappings to persist across paging_init() */
127 efi_memmap_unmap();
128
129 load_cpu_state_table();
130}
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright (C) 2015 Linaro Ltd <ard.biesheuvel@linaro.org>
4 */
5
6#include <linux/efi.h>
7#include <linux/memblock.h>
8#include <asm/efi.h>
9#include <asm/mach/map.h>
10#include <asm/mmu_context.h>
11
12static int __init set_permissions(pte_t *ptep, unsigned long addr, void *data)
13{
14 efi_memory_desc_t *md = data;
15 pte_t pte = *ptep;
16
17 if (md->attribute & EFI_MEMORY_RO)
18 pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
19 if (md->attribute & EFI_MEMORY_XP)
20 pte = set_pte_bit(pte, __pgprot(L_PTE_XN));
21 set_pte_ext(ptep, pte, PTE_EXT_NG);
22 return 0;
23}
24
25int __init efi_set_mapping_permissions(struct mm_struct *mm,
26 efi_memory_desc_t *md)
27{
28 unsigned long base, size;
29
30 base = md->virt_addr;
31 size = md->num_pages << EFI_PAGE_SHIFT;
32
33 /*
34 * We can only use apply_to_page_range() if we can guarantee that the
35 * entire region was mapped using pages. This should be the case if the
36 * region does not cover any naturally aligned SECTION_SIZE sized
37 * blocks.
38 */
39 if (round_down(base + size, SECTION_SIZE) <
40 round_up(base, SECTION_SIZE) + SECTION_SIZE)
41 return apply_to_page_range(mm, base, size, set_permissions, md);
42
43 return 0;
44}
45
46int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
47{
48 struct map_desc desc = {
49 .virtual = md->virt_addr,
50 .pfn = __phys_to_pfn(md->phys_addr),
51 .length = md->num_pages * EFI_PAGE_SIZE,
52 };
53
54 /*
55 * Order is important here: memory regions may have all of the
56 * bits below set (and usually do), so we check them in order of
57 * preference.
58 */
59 if (md->attribute & EFI_MEMORY_WB)
60 desc.type = MT_MEMORY_RWX;
61 else if (md->attribute & EFI_MEMORY_WT)
62 desc.type = MT_MEMORY_RWX_NONCACHED;
63 else if (md->attribute & EFI_MEMORY_WC)
64 desc.type = MT_DEVICE_WC;
65 else
66 desc.type = MT_DEVICE;
67
68 create_mapping_late(mm, &desc, true);
69
70 /*
71 * If stricter permissions were specified, apply them now.
72 */
73 if (md->attribute & (EFI_MEMORY_RO | EFI_MEMORY_XP))
74 return efi_set_mapping_permissions(mm, md);
75 return 0;
76}
77
78static unsigned long __initdata cpu_state_table = EFI_INVALID_TABLE_ADDR;
79
80const efi_config_table_type_t efi_arch_tables[] __initconst = {
81 {LINUX_EFI_ARM_CPU_STATE_TABLE_GUID, &cpu_state_table},
82 {}
83};
84
85static void __init load_cpu_state_table(void)
86{
87 if (cpu_state_table != EFI_INVALID_TABLE_ADDR) {
88 struct efi_arm_entry_state *state;
89 bool dump_state = true;
90
91 state = early_memremap_ro(cpu_state_table,
92 sizeof(struct efi_arm_entry_state));
93 if (state == NULL) {
94 pr_warn("Unable to map CPU entry state table.\n");
95 return;
96 }
97
98 if ((state->sctlr_before_ebs & 1) == 0)
99 pr_warn(FW_BUG "EFI stub was entered with MMU and Dcache disabled, please fix your firmware!\n");
100 else if ((state->sctlr_after_ebs & 1) == 0)
101 pr_warn(FW_BUG "ExitBootServices() returned with MMU and Dcache disabled, please fix your firmware!\n");
102 else
103 dump_state = false;
104
105 if (dump_state || efi_enabled(EFI_DBG)) {
106 pr_info("CPSR at EFI stub entry : 0x%08x\n",
107 state->cpsr_before_ebs);
108 pr_info("SCTLR at EFI stub entry : 0x%08x\n",
109 state->sctlr_before_ebs);
110 pr_info("CPSR after ExitBootServices() : 0x%08x\n",
111 state->cpsr_after_ebs);
112 pr_info("SCTLR after ExitBootServices(): 0x%08x\n",
113 state->sctlr_after_ebs);
114 }
115 early_memunmap(state, sizeof(struct efi_arm_entry_state));
116 }
117}
118
119void __init arm_efi_init(void)
120{
121 efi_init();
122
123 if (screen_info.orig_video_isVGA == VIDEO_TYPE_EFI) {
124 /* dummycon on ARM needs non-zero values for columns/lines */
125 screen_info.orig_video_cols = 80;
126 screen_info.orig_video_lines = 25;
127 }
128
129 /* ARM does not permit early mappings to persist across paging_init() */
130 efi_memmap_unmap();
131
132 load_cpu_state_table();
133}