Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Kernel unwinding support
4 *
5 * (c) 2002-2004 Randolph Chung <tausq@debian.org>
6 *
7 * Derived partially from the IA64 implementation. The PA-RISC
8 * Runtime Architecture Document is also a useful reference to
9 * understand what is happening here
10 */
11
12#include <linux/kernel.h>
13#include <linux/init.h>
14#include <linux/sched.h>
15#include <linux/slab.h>
16#include <linux/sort.h>
17
18#include <linux/uaccess.h>
19#include <asm/assembly.h>
20#include <asm/asm-offsets.h>
21#include <asm/ptrace.h>
22
23#include <asm/unwind.h>
24
25/* #define DEBUG 1 */
26#ifdef DEBUG
27#define dbg(x...) pr_debug(x)
28#else
29#define dbg(x...)
30#endif
31
32#define KERNEL_START (KERNEL_BINARY_TEXT_START)
33
34extern struct unwind_table_entry __start___unwind[];
35extern struct unwind_table_entry __stop___unwind[];
36
37static DEFINE_SPINLOCK(unwind_lock);
38/*
39 * the kernel unwind block is not dynamically allocated so that
40 * we can call unwind_init as early in the bootup process as
41 * possible (before the slab allocator is initialized)
42 */
43static struct unwind_table kernel_unwind_table __ro_after_init;
44static LIST_HEAD(unwind_tables);
45
46static inline const struct unwind_table_entry *
47find_unwind_entry_in_table(const struct unwind_table *table, unsigned long addr)
48{
49 const struct unwind_table_entry *e = NULL;
50 unsigned long lo, hi, mid;
51
52 lo = 0;
53 hi = table->length - 1;
54
55 while (lo <= hi) {
56 mid = (hi - lo) / 2 + lo;
57 e = &table->table[mid];
58 if (addr < e->region_start)
59 hi = mid - 1;
60 else if (addr > e->region_end)
61 lo = mid + 1;
62 else
63 return e;
64 }
65
66 return NULL;
67}
68
69static const struct unwind_table_entry *
70find_unwind_entry(unsigned long addr)
71{
72 struct unwind_table *table;
73 const struct unwind_table_entry *e = NULL;
74
75 if (addr >= kernel_unwind_table.start &&
76 addr <= kernel_unwind_table.end)
77 e = find_unwind_entry_in_table(&kernel_unwind_table, addr);
78 else {
79 unsigned long flags;
80
81 spin_lock_irqsave(&unwind_lock, flags);
82 list_for_each_entry(table, &unwind_tables, list) {
83 if (addr >= table->start &&
84 addr <= table->end)
85 e = find_unwind_entry_in_table(table, addr);
86 if (e) {
87 /* Move-to-front to exploit common traces */
88 list_move(&table->list, &unwind_tables);
89 break;
90 }
91 }
92 spin_unlock_irqrestore(&unwind_lock, flags);
93 }
94
95 return e;
96}
97
98static void
99unwind_table_init(struct unwind_table *table, const char *name,
100 unsigned long base_addr, unsigned long gp,
101 void *table_start, void *table_end)
102{
103 struct unwind_table_entry *start = table_start;
104 struct unwind_table_entry *end =
105 (struct unwind_table_entry *)table_end - 1;
106
107 table->name = name;
108 table->base_addr = base_addr;
109 table->gp = gp;
110 table->start = base_addr + start->region_start;
111 table->end = base_addr + end->region_end;
112 table->table = (struct unwind_table_entry *)table_start;
113 table->length = end - start + 1;
114 INIT_LIST_HEAD(&table->list);
115
116 for (; start <= end; start++) {
117 if (start < end &&
118 start->region_end > (start+1)->region_start) {
119 pr_warn("Out of order unwind entry! %px and %px\n",
120 start, start+1);
121 }
122
123 start->region_start += base_addr;
124 start->region_end += base_addr;
125 }
126}
127
128static int cmp_unwind_table_entry(const void *a, const void *b)
129{
130 return ((const struct unwind_table_entry *)a)->region_start
131 - ((const struct unwind_table_entry *)b)->region_start;
132}
133
134static void
135unwind_table_sort(struct unwind_table_entry *start,
136 struct unwind_table_entry *finish)
137{
138 sort(start, finish - start, sizeof(struct unwind_table_entry),
139 cmp_unwind_table_entry, NULL);
140}
141
142struct unwind_table *
143unwind_table_add(const char *name, unsigned long base_addr,
144 unsigned long gp,
145 void *start, void *end)
146{
147 struct unwind_table *table;
148 unsigned long flags;
149 struct unwind_table_entry *s = (struct unwind_table_entry *)start;
150 struct unwind_table_entry *e = (struct unwind_table_entry *)end;
151
152 unwind_table_sort(s, e);
153
154 table = kmalloc(sizeof(struct unwind_table), GFP_USER);
155 if (table == NULL)
156 return NULL;
157 unwind_table_init(table, name, base_addr, gp, start, end);
158 spin_lock_irqsave(&unwind_lock, flags);
159 list_add_tail(&table->list, &unwind_tables);
160 spin_unlock_irqrestore(&unwind_lock, flags);
161
162 return table;
163}
164
165void unwind_table_remove(struct unwind_table *table)
166{
167 unsigned long flags;
168
169 spin_lock_irqsave(&unwind_lock, flags);
170 list_del(&table->list);
171 spin_unlock_irqrestore(&unwind_lock, flags);
172
173 kfree(table);
174}
175
176/* Called from setup_arch to import the kernel unwind info */
177int __init unwind_init(void)
178{
179 long start, stop;
180 register unsigned long gp __asm__ ("r27");
181
182 start = (long)&__start___unwind[0];
183 stop = (long)&__stop___unwind[0];
184
185 dbg("unwind_init: start = 0x%lx, end = 0x%lx, entries = %lu\n",
186 start, stop,
187 (stop - start) / sizeof(struct unwind_table_entry));
188
189 unwind_table_init(&kernel_unwind_table, "kernel", KERNEL_START,
190 gp,
191 &__start___unwind[0], &__stop___unwind[0]);
192#if 0
193 {
194 int i;
195 for (i = 0; i < 10; i++)
196 {
197 printk("region 0x%x-0x%x\n",
198 __start___unwind[i].region_start,
199 __start___unwind[i].region_end);
200 }
201 }
202#endif
203 return 0;
204}
205
206static int unwind_special(struct unwind_frame_info *info, unsigned long pc, int frame_size)
207{
208 /*
209 * We have to use void * instead of a function pointer, because
210 * function pointers aren't a pointer to the function on 64-bit.
211 * Make them const so the compiler knows they live in .text
212 * Note: We could use dereference_kernel_function_descriptor()
213 * instead but we want to keep it simple here.
214 */
215 extern void * const handle_interruption;
216 extern void * const ret_from_kernel_thread;
217 extern void * const syscall_exit;
218 extern void * const intr_return;
219 extern void * const _switch_to_ret;
220#ifdef CONFIG_IRQSTACKS
221 extern void * const _call_on_stack;
222#endif /* CONFIG_IRQSTACKS */
223
224 if (pc == (unsigned long) &handle_interruption) {
225 struct pt_regs *regs = (struct pt_regs *)(info->sp - frame_size - PT_SZ_ALGN);
226 dbg("Unwinding through handle_interruption()\n");
227 info->prev_sp = regs->gr[30];
228 info->prev_ip = regs->iaoq[0];
229 return 1;
230 }
231
232 if (pc == (unsigned long) &ret_from_kernel_thread ||
233 pc == (unsigned long) &syscall_exit) {
234 info->prev_sp = info->prev_ip = 0;
235 return 1;
236 }
237
238 if (pc == (unsigned long) &intr_return) {
239 struct pt_regs *regs;
240
241 dbg("Found intr_return()\n");
242 regs = (struct pt_regs *)(info->sp - PT_SZ_ALGN);
243 info->prev_sp = regs->gr[30];
244 info->prev_ip = regs->iaoq[0];
245 info->rp = regs->gr[2];
246 return 1;
247 }
248
249 if (pc == (unsigned long) &_switch_to_ret) {
250 info->prev_sp = info->sp - CALLEE_SAVE_FRAME_SIZE;
251 info->prev_ip = *(unsigned long *)(info->prev_sp - RP_OFFSET);
252 return 1;
253 }
254
255#ifdef CONFIG_IRQSTACKS
256 if (pc == (unsigned long) &_call_on_stack) {
257 info->prev_sp = *(unsigned long *)(info->sp - FRAME_SIZE - REG_SZ);
258 info->prev_ip = *(unsigned long *)(info->sp - FRAME_SIZE - RP_OFFSET);
259 return 1;
260 }
261#endif
262
263 return 0;
264}
265
266static void unwind_frame_regs(struct unwind_frame_info *info)
267{
268 const struct unwind_table_entry *e;
269 unsigned long npc;
270 unsigned int insn;
271 long frame_size = 0;
272 int looking_for_rp, rpoffset = 0;
273
274 e = find_unwind_entry(info->ip);
275 if (e == NULL) {
276 unsigned long sp;
277
278 dbg("Cannot find unwind entry for %pS; forced unwinding\n",
279 (void *) info->ip);
280
281 /* Since we are doing the unwinding blind, we don't know if
282 we are adjusting the stack correctly or extracting the rp
283 correctly. The rp is checked to see if it belongs to the
284 kernel text section, if not we assume we don't have a
285 correct stack frame and we continue to unwind the stack.
286 This is not quite correct, and will fail for loadable
287 modules. */
288 sp = info->sp & ~63;
289 do {
290 unsigned long tmp;
291
292 info->prev_sp = sp - 64;
293 info->prev_ip = 0;
294
295 /* The stack is at the end inside the thread_union
296 * struct. If we reach data, we have reached the
297 * beginning of the stack and should stop unwinding. */
298 if (info->prev_sp >= (unsigned long) task_thread_info(info->t) &&
299 info->prev_sp < ((unsigned long) task_thread_info(info->t)
300 + THREAD_SZ_ALGN)) {
301 info->prev_sp = 0;
302 break;
303 }
304
305 if (get_user(tmp, (unsigned long *)(info->prev_sp - RP_OFFSET)))
306 break;
307 info->prev_ip = tmp;
308 sp = info->prev_sp;
309 } while (!kernel_text_address(info->prev_ip));
310
311 info->rp = 0;
312
313 dbg("analyzing func @ %lx with no unwind info, setting "
314 "prev_sp=%lx prev_ip=%lx\n", info->ip,
315 info->prev_sp, info->prev_ip);
316 } else {
317 dbg("e->start = 0x%x, e->end = 0x%x, Save_SP = %d, "
318 "Save_RP = %d, Millicode = %d size = %u\n",
319 e->region_start, e->region_end, e->Save_SP, e->Save_RP,
320 e->Millicode, e->Total_frame_size);
321
322 looking_for_rp = e->Save_RP;
323
324 for (npc = e->region_start;
325 (frame_size < (e->Total_frame_size << 3) ||
326 looking_for_rp) &&
327 npc < info->ip;
328 npc += 4) {
329
330 insn = *(unsigned int *)npc;
331
332 if ((insn & 0xffffc001) == 0x37de0000 ||
333 (insn & 0xffe00001) == 0x6fc00000) {
334 /* ldo X(sp), sp, or stwm X,D(sp) */
335 frame_size += (insn & 0x3fff) >> 1;
336 dbg("analyzing func @ %lx, insn=%08x @ "
337 "%lx, frame_size = %ld\n", info->ip,
338 insn, npc, frame_size);
339 } else if ((insn & 0xffe00009) == 0x73c00008) {
340 /* std,ma X,D(sp) */
341 frame_size += ((insn >> 4) & 0x3ff) << 3;
342 dbg("analyzing func @ %lx, insn=%08x @ "
343 "%lx, frame_size = %ld\n", info->ip,
344 insn, npc, frame_size);
345 } else if (insn == 0x6bc23fd9) {
346 /* stw rp,-20(sp) */
347 rpoffset = 20;
348 looking_for_rp = 0;
349 dbg("analyzing func @ %lx, insn=stw rp,"
350 "-20(sp) @ %lx\n", info->ip, npc);
351 } else if (insn == 0x0fc212c1) {
352 /* std rp,-16(sr0,sp) */
353 rpoffset = 16;
354 looking_for_rp = 0;
355 dbg("analyzing func @ %lx, insn=std rp,"
356 "-16(sp) @ %lx\n", info->ip, npc);
357 }
358 }
359
360 if (frame_size > e->Total_frame_size << 3)
361 frame_size = e->Total_frame_size << 3;
362
363 if (!unwind_special(info, e->region_start, frame_size)) {
364 info->prev_sp = info->sp - frame_size;
365 if (e->Millicode)
366 info->rp = info->r31;
367 else if (rpoffset)
368 info->rp = *(unsigned long *)(info->prev_sp - rpoffset);
369 info->prev_ip = info->rp;
370 info->rp = 0;
371 }
372
373 dbg("analyzing func @ %lx, setting prev_sp=%lx "
374 "prev_ip=%lx npc=%lx\n", info->ip, info->prev_sp,
375 info->prev_ip, npc);
376 }
377}
378
379void unwind_frame_init(struct unwind_frame_info *info, struct task_struct *t,
380 struct pt_regs *regs)
381{
382 memset(info, 0, sizeof(struct unwind_frame_info));
383 info->t = t;
384 info->sp = regs->gr[30];
385 info->ip = regs->iaoq[0];
386 info->rp = regs->gr[2];
387 info->r31 = regs->gr[31];
388
389 dbg("(%d) Start unwind from sp=%08lx ip=%08lx\n",
390 t ? (int)t->pid : -1, info->sp, info->ip);
391}
392
393void unwind_frame_init_from_blocked_task(struct unwind_frame_info *info, struct task_struct *t)
394{
395 struct pt_regs *r = &t->thread.regs;
396 struct pt_regs *r2;
397
398 r2 = kmalloc(sizeof(struct pt_regs), GFP_ATOMIC);
399 if (!r2)
400 return;
401 *r2 = *r;
402 r2->gr[30] = r->ksp;
403 r2->iaoq[0] = r->kpc;
404 unwind_frame_init(info, t, r2);
405 kfree(r2);
406}
407
408#define get_parisc_stackpointer() ({ \
409 unsigned long sp; \
410 __asm__("copy %%r30, %0" : "=r"(sp)); \
411 (sp); \
412})
413
414void unwind_frame_init_task(struct unwind_frame_info *info,
415 struct task_struct *task, struct pt_regs *regs)
416{
417 task = task ? task : current;
418
419 if (task == current) {
420 struct pt_regs r;
421
422 if (!regs) {
423 memset(&r, 0, sizeof(r));
424 r.iaoq[0] = _THIS_IP_;
425 r.gr[2] = _RET_IP_;
426 r.gr[30] = get_parisc_stackpointer();
427 regs = &r;
428 }
429 unwind_frame_init(info, task, regs);
430 } else {
431 unwind_frame_init_from_blocked_task(info, task);
432 }
433}
434
435int unwind_once(struct unwind_frame_info *next_frame)
436{
437 unwind_frame_regs(next_frame);
438
439 if (next_frame->prev_sp == 0 ||
440 next_frame->prev_ip == 0)
441 return -1;
442
443 next_frame->sp = next_frame->prev_sp;
444 next_frame->ip = next_frame->prev_ip;
445 next_frame->prev_sp = 0;
446 next_frame->prev_ip = 0;
447
448 dbg("(%d) Continue unwind to sp=%08lx ip=%08lx\n",
449 next_frame->t ? (int)next_frame->t->pid : -1,
450 next_frame->sp, next_frame->ip);
451
452 return 0;
453}
454
455int unwind_to_user(struct unwind_frame_info *info)
456{
457 int ret;
458
459 do {
460 ret = unwind_once(info);
461 } while (!ret && !(info->ip & 3));
462
463 return ret;
464}
465
466unsigned long return_address(unsigned int level)
467{
468 struct unwind_frame_info info;
469
470 /* initialize unwind info */
471 unwind_frame_init_task(&info, current, NULL);
472
473 /* unwind stack */
474 level += 2;
475 do {
476 if (unwind_once(&info) < 0 || info.ip == 0)
477 return 0;
478 if (!kernel_text_address(info.ip))
479 return 0;
480 } while (info.ip && level--);
481
482 return info.ip;
483}
1/*
2 * Kernel unwinding support
3 *
4 * (c) 2002-2004 Randolph Chung <tausq@debian.org>
5 *
6 * Derived partially from the IA64 implementation. The PA-RISC
7 * Runtime Architecture Document is also a useful reference to
8 * understand what is happening here
9 */
10
11#include <linux/kernel.h>
12#include <linux/init.h>
13#include <linux/sched.h>
14#include <linux/slab.h>
15#include <linux/kallsyms.h>
16#include <linux/sort.h>
17
18#include <asm/uaccess.h>
19#include <asm/assembly.h>
20#include <asm/asm-offsets.h>
21#include <asm/ptrace.h>
22
23#include <asm/unwind.h>
24
25/* #define DEBUG 1 */
26#ifdef DEBUG
27#define dbg(x...) printk(x)
28#else
29#define dbg(x...)
30#endif
31
32#define KERNEL_START (KERNEL_BINARY_TEXT_START)
33
34extern struct unwind_table_entry __start___unwind[];
35extern struct unwind_table_entry __stop___unwind[];
36
37static spinlock_t unwind_lock;
38/*
39 * the kernel unwind block is not dynamically allocated so that
40 * we can call unwind_init as early in the bootup process as
41 * possible (before the slab allocator is initialized)
42 */
43static struct unwind_table kernel_unwind_table __read_mostly;
44static LIST_HEAD(unwind_tables);
45
46static inline const struct unwind_table_entry *
47find_unwind_entry_in_table(const struct unwind_table *table, unsigned long addr)
48{
49 const struct unwind_table_entry *e = NULL;
50 unsigned long lo, hi, mid;
51
52 lo = 0;
53 hi = table->length - 1;
54
55 while (lo <= hi) {
56 mid = (hi - lo) / 2 + lo;
57 e = &table->table[mid];
58 if (addr < e->region_start)
59 hi = mid - 1;
60 else if (addr > e->region_end)
61 lo = mid + 1;
62 else
63 return e;
64 }
65
66 return NULL;
67}
68
69static const struct unwind_table_entry *
70find_unwind_entry(unsigned long addr)
71{
72 struct unwind_table *table;
73 const struct unwind_table_entry *e = NULL;
74
75 if (addr >= kernel_unwind_table.start &&
76 addr <= kernel_unwind_table.end)
77 e = find_unwind_entry_in_table(&kernel_unwind_table, addr);
78 else
79 list_for_each_entry(table, &unwind_tables, list) {
80 if (addr >= table->start &&
81 addr <= table->end)
82 e = find_unwind_entry_in_table(table, addr);
83 if (e) {
84 /* Move-to-front to exploit common traces */
85 list_move(&table->list, &unwind_tables);
86 break;
87 }
88 }
89
90 return e;
91}
92
93static void
94unwind_table_init(struct unwind_table *table, const char *name,
95 unsigned long base_addr, unsigned long gp,
96 void *table_start, void *table_end)
97{
98 struct unwind_table_entry *start = table_start;
99 struct unwind_table_entry *end =
100 (struct unwind_table_entry *)table_end - 1;
101
102 table->name = name;
103 table->base_addr = base_addr;
104 table->gp = gp;
105 table->start = base_addr + start->region_start;
106 table->end = base_addr + end->region_end;
107 table->table = (struct unwind_table_entry *)table_start;
108 table->length = end - start + 1;
109 INIT_LIST_HEAD(&table->list);
110
111 for (; start <= end; start++) {
112 if (start < end &&
113 start->region_end > (start+1)->region_start) {
114 printk("WARNING: Out of order unwind entry! %p and %p\n", start, start+1);
115 }
116
117 start->region_start += base_addr;
118 start->region_end += base_addr;
119 }
120}
121
122static int cmp_unwind_table_entry(const void *a, const void *b)
123{
124 return ((const struct unwind_table_entry *)a)->region_start
125 - ((const struct unwind_table_entry *)b)->region_start;
126}
127
128static void
129unwind_table_sort(struct unwind_table_entry *start,
130 struct unwind_table_entry *finish)
131{
132 sort(start, finish - start, sizeof(struct unwind_table_entry),
133 cmp_unwind_table_entry, NULL);
134}
135
136struct unwind_table *
137unwind_table_add(const char *name, unsigned long base_addr,
138 unsigned long gp,
139 void *start, void *end)
140{
141 struct unwind_table *table;
142 unsigned long flags;
143 struct unwind_table_entry *s = (struct unwind_table_entry *)start;
144 struct unwind_table_entry *e = (struct unwind_table_entry *)end;
145
146 unwind_table_sort(s, e);
147
148 table = kmalloc(sizeof(struct unwind_table), GFP_USER);
149 if (table == NULL)
150 return NULL;
151 unwind_table_init(table, name, base_addr, gp, start, end);
152 spin_lock_irqsave(&unwind_lock, flags);
153 list_add_tail(&table->list, &unwind_tables);
154 spin_unlock_irqrestore(&unwind_lock, flags);
155
156 return table;
157}
158
159void unwind_table_remove(struct unwind_table *table)
160{
161 unsigned long flags;
162
163 spin_lock_irqsave(&unwind_lock, flags);
164 list_del(&table->list);
165 spin_unlock_irqrestore(&unwind_lock, flags);
166
167 kfree(table);
168}
169
170/* Called from setup_arch to import the kernel unwind info */
171int unwind_init(void)
172{
173 long start, stop;
174 register unsigned long gp __asm__ ("r27");
175
176 start = (long)&__start___unwind[0];
177 stop = (long)&__stop___unwind[0];
178
179 spin_lock_init(&unwind_lock);
180
181 printk("unwind_init: start = 0x%lx, end = 0x%lx, entries = %lu\n",
182 start, stop,
183 (stop - start) / sizeof(struct unwind_table_entry));
184
185 unwind_table_init(&kernel_unwind_table, "kernel", KERNEL_START,
186 gp,
187 &__start___unwind[0], &__stop___unwind[0]);
188#if 0
189 {
190 int i;
191 for (i = 0; i < 10; i++)
192 {
193 printk("region 0x%x-0x%x\n",
194 __start___unwind[i].region_start,
195 __start___unwind[i].region_end);
196 }
197 }
198#endif
199 return 0;
200}
201
202#ifdef CONFIG_64BIT
203#define get_func_addr(fptr) fptr[2]
204#else
205#define get_func_addr(fptr) fptr[0]
206#endif
207
208static int unwind_special(struct unwind_frame_info *info, unsigned long pc, int frame_size)
209{
210 extern void handle_interruption(int, struct pt_regs *);
211 static unsigned long *hi = (unsigned long *)&handle_interruption;
212
213 if (pc == get_func_addr(hi)) {
214 struct pt_regs *regs = (struct pt_regs *)(info->sp - frame_size - PT_SZ_ALGN);
215 dbg("Unwinding through handle_interruption()\n");
216 info->prev_sp = regs->gr[30];
217 info->prev_ip = regs->iaoq[0];
218
219 return 1;
220 }
221
222 return 0;
223}
224
225static void unwind_frame_regs(struct unwind_frame_info *info)
226{
227 const struct unwind_table_entry *e;
228 unsigned long npc;
229 unsigned int insn;
230 long frame_size = 0;
231 int looking_for_rp, rpoffset = 0;
232
233 e = find_unwind_entry(info->ip);
234 if (e == NULL) {
235 unsigned long sp;
236 extern char _stext[], _etext[];
237
238 dbg("Cannot find unwind entry for 0x%lx; forced unwinding\n", info->ip);
239
240#ifdef CONFIG_KALLSYMS
241 /* Handle some frequent special cases.... */
242 {
243 char symname[KSYM_NAME_LEN];
244 char *modname;
245
246 kallsyms_lookup(info->ip, NULL, NULL, &modname,
247 symname);
248
249 dbg("info->ip = 0x%lx, name = %s\n", info->ip, symname);
250
251 if (strcmp(symname, "_switch_to_ret") == 0) {
252 info->prev_sp = info->sp - CALLEE_SAVE_FRAME_SIZE;
253 info->prev_ip = *(unsigned long *)(info->prev_sp - RP_OFFSET);
254 dbg("_switch_to_ret @ %lx - setting "
255 "prev_sp=%lx prev_ip=%lx\n",
256 info->ip, info->prev_sp,
257 info->prev_ip);
258 return;
259 } else if (strcmp(symname, "ret_from_kernel_thread") == 0 ||
260 strcmp(symname, "syscall_exit") == 0) {
261 info->prev_ip = info->prev_sp = 0;
262 return;
263 }
264 }
265#endif
266
267 /* Since we are doing the unwinding blind, we don't know if
268 we are adjusting the stack correctly or extracting the rp
269 correctly. The rp is checked to see if it belongs to the
270 kernel text section, if not we assume we don't have a
271 correct stack frame and we continue to unwind the stack.
272 This is not quite correct, and will fail for loadable
273 modules. */
274 sp = info->sp & ~63;
275 do {
276 unsigned long tmp;
277
278 info->prev_sp = sp - 64;
279 info->prev_ip = 0;
280 if (get_user(tmp, (unsigned long *)(info->prev_sp - RP_OFFSET)))
281 break;
282 info->prev_ip = tmp;
283 sp = info->prev_sp;
284 } while (info->prev_ip < (unsigned long)_stext ||
285 info->prev_ip > (unsigned long)_etext);
286
287 info->rp = 0;
288
289 dbg("analyzing func @ %lx with no unwind info, setting "
290 "prev_sp=%lx prev_ip=%lx\n", info->ip,
291 info->prev_sp, info->prev_ip);
292 } else {
293 dbg("e->start = 0x%x, e->end = 0x%x, Save_SP = %d, "
294 "Save_RP = %d, Millicode = %d size = %u\n",
295 e->region_start, e->region_end, e->Save_SP, e->Save_RP,
296 e->Millicode, e->Total_frame_size);
297
298 looking_for_rp = e->Save_RP;
299
300 for (npc = e->region_start;
301 (frame_size < (e->Total_frame_size << 3) ||
302 looking_for_rp) &&
303 npc < info->ip;
304 npc += 4) {
305
306 insn = *(unsigned int *)npc;
307
308 if ((insn & 0xffffc000) == 0x37de0000 ||
309 (insn & 0xffe00000) == 0x6fc00000) {
310 /* ldo X(sp), sp, or stwm X,D(sp) */
311 frame_size += (insn & 0x1 ? -1 << 13 : 0) |
312 ((insn & 0x3fff) >> 1);
313 dbg("analyzing func @ %lx, insn=%08x @ "
314 "%lx, frame_size = %ld\n", info->ip,
315 insn, npc, frame_size);
316 } else if ((insn & 0xffe00008) == 0x73c00008) {
317 /* std,ma X,D(sp) */
318 frame_size += (insn & 0x1 ? -1 << 13 : 0) |
319 (((insn >> 4) & 0x3ff) << 3);
320 dbg("analyzing func @ %lx, insn=%08x @ "
321 "%lx, frame_size = %ld\n", info->ip,
322 insn, npc, frame_size);
323 } else if (insn == 0x6bc23fd9) {
324 /* stw rp,-20(sp) */
325 rpoffset = 20;
326 looking_for_rp = 0;
327 dbg("analyzing func @ %lx, insn=stw rp,"
328 "-20(sp) @ %lx\n", info->ip, npc);
329 } else if (insn == 0x0fc212c1) {
330 /* std rp,-16(sr0,sp) */
331 rpoffset = 16;
332 looking_for_rp = 0;
333 dbg("analyzing func @ %lx, insn=std rp,"
334 "-16(sp) @ %lx\n", info->ip, npc);
335 }
336 }
337
338 if (!unwind_special(info, e->region_start, frame_size)) {
339 info->prev_sp = info->sp - frame_size;
340 if (e->Millicode)
341 info->rp = info->r31;
342 else if (rpoffset)
343 info->rp = *(unsigned long *)(info->prev_sp - rpoffset);
344 info->prev_ip = info->rp;
345 info->rp = 0;
346 }
347
348 dbg("analyzing func @ %lx, setting prev_sp=%lx "
349 "prev_ip=%lx npc=%lx\n", info->ip, info->prev_sp,
350 info->prev_ip, npc);
351 }
352}
353
354void unwind_frame_init(struct unwind_frame_info *info, struct task_struct *t,
355 struct pt_regs *regs)
356{
357 memset(info, 0, sizeof(struct unwind_frame_info));
358 info->t = t;
359 info->sp = regs->gr[30];
360 info->ip = regs->iaoq[0];
361 info->rp = regs->gr[2];
362 info->r31 = regs->gr[31];
363
364 dbg("(%d) Start unwind from sp=%08lx ip=%08lx\n",
365 t ? (int)t->pid : -1, info->sp, info->ip);
366}
367
368void unwind_frame_init_from_blocked_task(struct unwind_frame_info *info, struct task_struct *t)
369{
370 struct pt_regs *r = &t->thread.regs;
371 struct pt_regs *r2;
372
373 r2 = kmalloc(sizeof(struct pt_regs), GFP_ATOMIC);
374 if (!r2)
375 return;
376 *r2 = *r;
377 r2->gr[30] = r->ksp;
378 r2->iaoq[0] = r->kpc;
379 unwind_frame_init(info, t, r2);
380 kfree(r2);
381}
382
383void unwind_frame_init_running(struct unwind_frame_info *info, struct pt_regs *regs)
384{
385 unwind_frame_init(info, current, regs);
386}
387
388int unwind_once(struct unwind_frame_info *next_frame)
389{
390 unwind_frame_regs(next_frame);
391
392 if (next_frame->prev_sp == 0 ||
393 next_frame->prev_ip == 0)
394 return -1;
395
396 next_frame->sp = next_frame->prev_sp;
397 next_frame->ip = next_frame->prev_ip;
398 next_frame->prev_sp = 0;
399 next_frame->prev_ip = 0;
400
401 dbg("(%d) Continue unwind to sp=%08lx ip=%08lx\n",
402 next_frame->t ? (int)next_frame->t->pid : -1,
403 next_frame->sp, next_frame->ip);
404
405 return 0;
406}
407
408int unwind_to_user(struct unwind_frame_info *info)
409{
410 int ret;
411
412 do {
413 ret = unwind_once(info);
414 } while (!ret && !(info->ip & 3));
415
416 return ret;
417}
418
419unsigned long return_address(unsigned int level)
420{
421 struct unwind_frame_info info;
422 struct pt_regs r;
423 unsigned long sp;
424
425 /* initialize unwind info */
426 asm volatile ("copy %%r30, %0" : "=r"(sp));
427 memset(&r, 0, sizeof(struct pt_regs));
428 r.iaoq[0] = (unsigned long) current_text_addr();
429 r.gr[2] = (unsigned long) __builtin_return_address(0);
430 r.gr[30] = sp;
431 unwind_frame_init(&info, current, &r);
432
433 /* unwind stack */
434 ++level;
435 do {
436 if (unwind_once(&info) < 0 || info.ip == 0)
437 return 0;
438 if (!__kernel_text_address(info.ip)) {
439 return 0;
440 }
441 } while (info.ip && level--);
442
443 return info.ip;
444}