Loading...
1#ifndef _ASM_X86_PAGE_64_H
2#define _ASM_X86_PAGE_64_H
3
4#include <asm/page_64_types.h>
5
6#ifndef __ASSEMBLY__
7
8/* duplicated to the one in bootmem.h */
9extern unsigned long max_pfn;
10extern unsigned long phys_base;
11
12static inline unsigned long __phys_addr_nodebug(unsigned long x)
13{
14 unsigned long y = x - __START_KERNEL_map;
15
16 /* use the carry flag to determine if x was < __START_KERNEL_map */
17 x = y + ((x > y) ? phys_base : (__START_KERNEL_map - PAGE_OFFSET));
18
19 return x;
20}
21
22#ifdef CONFIG_DEBUG_VIRTUAL
23extern unsigned long __phys_addr(unsigned long);
24extern unsigned long __phys_addr_symbol(unsigned long);
25#else
26#define __phys_addr(x) __phys_addr_nodebug(x)
27#define __phys_addr_symbol(x) \
28 ((unsigned long)(x) - __START_KERNEL_map + phys_base)
29#endif
30
31#define __phys_reloc_hide(x) (x)
32
33#ifdef CONFIG_FLATMEM
34#define pfn_valid(pfn) ((pfn) < max_pfn)
35#endif
36
37void clear_page(void *page);
38void copy_page(void *to, void *from);
39
40#endif /* !__ASSEMBLY__ */
41
42#ifdef CONFIG_X86_VSYSCALL_EMULATION
43# define __HAVE_ARCH_GATE_AREA 1
44#endif
45
46#endif /* _ASM_X86_PAGE_64_H */
1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef _ASM_X86_PAGE_64_H
3#define _ASM_X86_PAGE_64_H
4
5#include <asm/page_64_types.h>
6
7#ifndef __ASSEMBLY__
8#include <asm/cpufeatures.h>
9#include <asm/alternative.h>
10
11#include <linux/kmsan-checks.h>
12
13/* duplicated to the one in bootmem.h */
14extern unsigned long max_pfn;
15extern unsigned long phys_base;
16
17extern unsigned long page_offset_base;
18extern unsigned long vmalloc_base;
19extern unsigned long vmemmap_base;
20
21static __always_inline unsigned long __phys_addr_nodebug(unsigned long x)
22{
23 unsigned long y = x - __START_KERNEL_map;
24
25 /* use the carry flag to determine if x was < __START_KERNEL_map */
26 x = y + ((x > y) ? phys_base : (__START_KERNEL_map - PAGE_OFFSET));
27
28 return x;
29}
30
31#ifdef CONFIG_DEBUG_VIRTUAL
32extern unsigned long __phys_addr(unsigned long);
33extern unsigned long __phys_addr_symbol(unsigned long);
34#else
35#define __phys_addr(x) __phys_addr_nodebug(x)
36#define __phys_addr_symbol(x) \
37 ((unsigned long)(x) - __START_KERNEL_map + phys_base)
38#endif
39
40#define __phys_reloc_hide(x) (x)
41
42#ifdef CONFIG_FLATMEM
43#define pfn_valid(pfn) ((pfn) < max_pfn)
44#endif
45
46void clear_page_orig(void *page);
47void clear_page_rep(void *page);
48void clear_page_erms(void *page);
49
50static inline void clear_page(void *page)
51{
52 /*
53 * Clean up KMSAN metadata for the page being cleared. The assembly call
54 * below clobbers @page, so we perform unpoisoning before it.
55 */
56 kmsan_unpoison_memory(page, PAGE_SIZE);
57 alternative_call_2(clear_page_orig,
58 clear_page_rep, X86_FEATURE_REP_GOOD,
59 clear_page_erms, X86_FEATURE_ERMS,
60 "=D" (page),
61 "0" (page)
62 : "cc", "memory", "rax", "rcx");
63}
64
65void copy_page(void *to, void *from);
66
67#ifdef CONFIG_X86_5LEVEL
68/*
69 * User space process size. This is the first address outside the user range.
70 * There are a few constraints that determine this:
71 *
72 * On Intel CPUs, if a SYSCALL instruction is at the highest canonical
73 * address, then that syscall will enter the kernel with a
74 * non-canonical return address, and SYSRET will explode dangerously.
75 * We avoid this particular problem by preventing anything
76 * from being mapped at the maximum canonical address.
77 *
78 * On AMD CPUs in the Ryzen family, there's a nasty bug in which the
79 * CPUs malfunction if they execute code from the highest canonical page.
80 * They'll speculate right off the end of the canonical space, and
81 * bad things happen. This is worked around in the same way as the
82 * Intel problem.
83 *
84 * With page table isolation enabled, we map the LDT in ... [stay tuned]
85 */
86static __always_inline unsigned long task_size_max(void)
87{
88 unsigned long ret;
89
90 alternative_io("movq %[small],%0","movq %[large],%0",
91 X86_FEATURE_LA57,
92 "=r" (ret),
93 [small] "i" ((1ul << 47)-PAGE_SIZE),
94 [large] "i" ((1ul << 56)-PAGE_SIZE));
95
96 return ret;
97}
98#endif /* CONFIG_X86_5LEVEL */
99
100#endif /* !__ASSEMBLY__ */
101
102#ifdef CONFIG_X86_VSYSCALL_EMULATION
103# define __HAVE_ARCH_GATE_AREA 1
104#endif
105
106#endif /* _ASM_X86_PAGE_64_H */