Loading...
1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * Copyright 2003 PathScale Inc
4 * Derived from include/asm-i386/pgtable.h
5 */
6
7#ifndef __UM_PGTABLE_3LEVEL_H
8#define __UM_PGTABLE_3LEVEL_H
9
10#include <asm-generic/pgtable-nopud.h>
11
12/* PGDIR_SHIFT determines what a third-level page table entry can map */
13
14#ifdef CONFIG_64BIT
15#define PGDIR_SHIFT 30
16#else
17#define PGDIR_SHIFT 31
18#endif
19#define PGDIR_SIZE (1UL << PGDIR_SHIFT)
20#define PGDIR_MASK (~(PGDIR_SIZE-1))
21
22/* PMD_SHIFT determines the size of the area a second-level page table can
23 * map
24 */
25
26#define PMD_SHIFT 21
27#define PMD_SIZE (1UL << PMD_SHIFT)
28#define PMD_MASK (~(PMD_SIZE-1))
29
30/*
31 * entries per page directory level
32 */
33
34#define PTRS_PER_PTE 512
35#ifdef CONFIG_64BIT
36#define PTRS_PER_PMD 512
37#define PTRS_PER_PGD 512
38#else
39#define PTRS_PER_PMD 1024
40#define PTRS_PER_PGD 1024
41#endif
42
43#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
44
45#define pte_ERROR(e) \
46 printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \
47 pte_val(e))
48#define pmd_ERROR(e) \
49 printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
50 pmd_val(e))
51#define pgd_ERROR(e) \
52 printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
53 pgd_val(e))
54
55#define pud_none(x) (!(pud_val(x) & ~_PAGE_NEWPAGE))
56#define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
57#define pud_present(x) (pud_val(x) & _PAGE_PRESENT)
58#define pud_populate(mm, pud, pmd) \
59 set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
60
61#define set_pud(pudptr, pudval) (*(pudptr) = (pudval))
62
63static inline int pgd_newpage(pgd_t pgd)
64{
65 return(pgd_val(pgd) & _PAGE_NEWPAGE);
66}
67
68static inline void pgd_mkuptodate(pgd_t pgd) { pgd_val(pgd) &= ~_PAGE_NEWPAGE; }
69
70#define set_pmd(pmdptr, pmdval) (*(pmdptr) = (pmdval))
71
72static inline void pud_clear (pud_t *pud)
73{
74 set_pud(pud, __pud(_PAGE_NEWPAGE));
75}
76
77#define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK)
78#define pud_pgtable(pud) ((pmd_t *) __va(pud_val(pud) & PAGE_MASK))
79
80static inline unsigned long pte_pfn(pte_t pte)
81{
82 return phys_to_pfn(pte_val(pte));
83}
84
85static inline pte_t pfn_pte(unsigned long page_nr, pgprot_t pgprot)
86{
87 pte_t pte;
88 phys_t phys = pfn_to_phys(page_nr);
89
90 pte_set_val(pte, phys, pgprot);
91 return pte;
92}
93
94static inline pmd_t pfn_pmd(unsigned long page_nr, pgprot_t pgprot)
95{
96 return __pmd((page_nr << PAGE_SHIFT) | pgprot_val(pgprot));
97}
98
99#endif
100
1/*
2 * Copyright 2003 PathScale Inc
3 * Derived from include/asm-i386/pgtable.h
4 * Licensed under the GPL
5 */
6
7#ifndef __UM_PGTABLE_3LEVEL_H
8#define __UM_PGTABLE_3LEVEL_H
9
10#define __ARCH_USE_5LEVEL_HACK
11#include <asm-generic/pgtable-nopud.h>
12
13/* PGDIR_SHIFT determines what a third-level page table entry can map */
14
15#ifdef CONFIG_64BIT
16#define PGDIR_SHIFT 30
17#else
18#define PGDIR_SHIFT 31
19#endif
20#define PGDIR_SIZE (1UL << PGDIR_SHIFT)
21#define PGDIR_MASK (~(PGDIR_SIZE-1))
22
23/* PMD_SHIFT determines the size of the area a second-level page table can
24 * map
25 */
26
27#define PMD_SHIFT 21
28#define PMD_SIZE (1UL << PMD_SHIFT)
29#define PMD_MASK (~(PMD_SIZE-1))
30
31/*
32 * entries per page directory level
33 */
34
35#define PTRS_PER_PTE 512
36#ifdef CONFIG_64BIT
37#define PTRS_PER_PMD 512
38#define PTRS_PER_PGD 512
39#else
40#define PTRS_PER_PMD 1024
41#define PTRS_PER_PGD 1024
42#endif
43
44#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
45#define FIRST_USER_ADDRESS 0UL
46
47#define pte_ERROR(e) \
48 printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \
49 pte_val(e))
50#define pmd_ERROR(e) \
51 printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
52 pmd_val(e))
53#define pgd_ERROR(e) \
54 printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
55 pgd_val(e))
56
57#define pud_none(x) (!(pud_val(x) & ~_PAGE_NEWPAGE))
58#define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
59#define pud_present(x) (pud_val(x) & _PAGE_PRESENT)
60#define pud_populate(mm, pud, pmd) \
61 set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
62
63#ifdef CONFIG_64BIT
64#define set_pud(pudptr, pudval) set_64bit((u64 *) (pudptr), pud_val(pudval))
65#else
66#define set_pud(pudptr, pudval) (*(pudptr) = (pudval))
67#endif
68
69static inline int pgd_newpage(pgd_t pgd)
70{
71 return(pgd_val(pgd) & _PAGE_NEWPAGE);
72}
73
74static inline void pgd_mkuptodate(pgd_t pgd) { pgd_val(pgd) &= ~_PAGE_NEWPAGE; }
75
76#ifdef CONFIG_64BIT
77#define set_pmd(pmdptr, pmdval) set_64bit((u64 *) (pmdptr), pmd_val(pmdval))
78#else
79#define set_pmd(pmdptr, pmdval) (*(pmdptr) = (pmdval))
80#endif
81
82struct mm_struct;
83extern pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address);
84
85static inline void pud_clear (pud_t *pud)
86{
87 set_pud(pud, __pud(_PAGE_NEWPAGE));
88}
89
90#define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK)
91#define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PAGE_MASK))
92
93/* Find an entry in the second-level page table.. */
94#define pmd_offset(pud, address) ((pmd_t *) pud_page_vaddr(*(pud)) + \
95 pmd_index(address))
96
97static inline unsigned long pte_pfn(pte_t pte)
98{
99 return phys_to_pfn(pte_val(pte));
100}
101
102static inline pte_t pfn_pte(unsigned long page_nr, pgprot_t pgprot)
103{
104 pte_t pte;
105 phys_t phys = pfn_to_phys(page_nr);
106
107 pte_set_val(pte, phys, pgprot);
108 return pte;
109}
110
111static inline pmd_t pfn_pmd(unsigned long page_nr, pgprot_t pgprot)
112{
113 return __pmd((page_nr << PAGE_SHIFT) | pgprot_val(pgprot));
114}
115
116#endif
117