Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 | /* * The SH64 TLB miss. * * Original code from fault.c * Copyright (C) 2000, 2001 Paolo Alberelli * * Fast PTE->TLB refill path * Copyright (C) 2003 Richard.Curnow@superh.com * * IMPORTANT NOTES : * The do_fast_page_fault function is called from a context in entry.S * where very few registers have been saved. In particular, the code in * this file must be compiled not to use ANY caller-save registers that * are not part of the restricted save set. Also, it means that code in * this file must not make calls to functions elsewhere in the kernel, or * else the excepting context will see corruption in its caller-save * registers. Plus, the entry.S save area is non-reentrant, so this code * has to run with SR.BL==1, i.e. no interrupts taken inside it and panic * on any exception. * * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. */ #include <linux/signal.h> #include <linux/sched.h> #include <linux/kernel.h> #include <linux/errno.h> #include <linux/string.h> #include <linux/types.h> #include <linux/ptrace.h> #include <linux/mman.h> #include <linux/mm.h> #include <linux/smp.h> #include <linux/interrupt.h> #include <linux/kprobes.h> #include <asm/tlb.h> #include <asm/io.h> #include <asm/uaccess.h> #include <asm/pgalloc.h> #include <asm/mmu_context.h> static int handle_tlbmiss(unsigned long long protection_flags, unsigned long address) { pgd_t *pgd; pud_t *pud; pmd_t *pmd; pte_t *pte; pte_t entry; if (is_vmalloc_addr((void *)address)) { pgd = pgd_offset_k(address); } else { if (unlikely(address >= TASK_SIZE || !current->mm)) return 1; pgd = pgd_offset(current->mm, address); } pud = pud_offset(pgd, address); if (pud_none(*pud) || !pud_present(*pud)) return 1; pmd = pmd_offset(pud, address); if (pmd_none(*pmd) || !pmd_present(*pmd)) return 1; pte = pte_offset_kernel(pmd, address); entry = *pte; if (pte_none(entry) || !pte_present(entry)) return 1; /* * If the page doesn't have sufficient protection bits set to * service the kind of fault being handled, there's not much * point doing the TLB refill. Punt the fault to the general * handler. */ if ((pte_val(entry) & protection_flags) != protection_flags) return 1; update_mmu_cache(NULL, address, pte); return 0; } /* * Put all this information into one structure so that everything is just * arithmetic relative to a single base address. This reduces the number * of movi/shori pairs needed just to load addresses of static data. */ struct expevt_lookup { unsigned short protection_flags[8]; unsigned char is_text_access[8]; unsigned char is_write_access[8]; }; #define PRU (1<<9) #define PRW (1<<8) #define PRX (1<<7) #define PRR (1<<6) /* Sized as 8 rather than 4 to allow checking the PTE's PRU bit against whether the fault happened in user mode or privileged mode. */ static struct expevt_lookup expevt_lookup_table = { .protection_flags = {PRX, PRX, 0, 0, PRR, PRR, PRW, PRW}, .is_text_access = {1, 1, 0, 0, 0, 0, 0, 0} }; static inline unsigned int expevt_to_fault_code(unsigned long expevt) { if (expevt == 0xa40) return FAULT_CODE_ITLB; else if (expevt == 0x060) return FAULT_CODE_WRITE; return 0; } /* This routine handles page faults that can be serviced just by refilling a TLB entry from an existing page table entry. (This case represents a very large majority of page faults.) Return 1 if the fault was successfully handled. Return 0 if the fault could not be handled. (This leads into the general fault handling in fault.c which deals with mapping file-backed pages, stack growth, segmentation faults, swapping etc etc) */ asmlinkage int __kprobes do_fast_page_fault(unsigned long long ssr_md, unsigned long long expevt, unsigned long address) { unsigned long long protection_flags; unsigned long long index; unsigned long long expevt4; unsigned int fault_code; /* The next few lines implement a way of hashing EXPEVT into a * small array index which can be used to lookup parameters * specific to the type of TLBMISS being handled. * * Note: * ITLBMISS has EXPEVT==0xa40 * RTLBMISS has EXPEVT==0x040 * WTLBMISS has EXPEVT==0x060 */ expevt4 = (expevt >> 4); /* TODO : xor ssr_md into this expression too. Then we can check * that PRU is set when it needs to be. */ index = expevt4 ^ (expevt4 >> 5); index &= 7; fault_code = expevt_to_fault_code(expevt); protection_flags = expevt_lookup_table.protection_flags[index]; if (expevt_lookup_table.is_text_access[index]) fault_code |= FAULT_CODE_ITLB; if (!ssr_md) fault_code |= FAULT_CODE_USER; set_thread_fault_code(fault_code); return handle_tlbmiss(protection_flags, address); } |