Linux Audio

Check our new training course

Loading...
v5.14.15
  1/* SPDX-License-Identifier: GPL-2.0-only */
  2/*
  3 * Early kernel startup code for Hexagon
  4 *
  5 * Copyright (c) 2010-2013, The Linux Foundation. All rights reserved.
  6 */
  7
  8#include <linux/linkage.h>
  9#include <linux/init.h>
 10#include <asm/asm-offsets.h>
 11#include <asm/mem-layout.h>
 12#include <asm/vm_mmu.h>
 13#include <asm/page.h>
 14#include <asm/hexagon_vm.h>
 15
 16#define SEGTABLE_ENTRIES #0x0e0
 17
 18	__INIT
 19ENTRY(stext)
 20	/*
 21	 * VMM will already have set up true vector page, MMU, etc.
 22	 * To set up initial kernel identity map, we have to pass
 23	 * the VMM a pointer to some canonical page tables. In
 24	 * this implementation, we're assuming that we've got
 25	 * them precompiled. Generate value in R24, as we'll need
 26	 * it again shortly.
 27	 */
 28	r24.L = #LO(swapper_pg_dir)
 29	r24.H = #HI(swapper_pg_dir)
 30
 31	/*
 32	 * Symbol is kernel segment address, but we need
 33	 * the logical/physical address.
 34	 */
 35	r25 = pc;
 36	r2.h = #0xffc0;
 37	r2.l = #0x0000;
 38	r25 = and(r2,r25);	/*  R25 holds PHYS_OFFSET now  */
 39	r1.h = #HI(PAGE_OFFSET);
 40	r1.l = #LO(PAGE_OFFSET);
 41	r24 = sub(r24,r1);	/* swapper_pg_dir - PAGE_OFFSET */
 42	r24 = add(r24,r25);	/* + PHYS_OFFSET */
 43
 44	r0 = r24;  /* aka __pa(swapper_pg_dir)  */
 45
 46	/*
 47	 * Initialize page dir to make the virtual and physical
 48	 * addresses where the kernel was loaded be identical.
 49	 * Done in 4MB chunks.
 50	 */
 51#define PTE_BITS ( __HVM_PTE_R | __HVM_PTE_W | __HVM_PTE_X	\
 52		  | __HEXAGON_C_WB_L2 << 6			\
 53		  | __HVM_PDE_S_4MB)
 54
 55	/*
 56	 * Get number of VA=PA entries; only really needed for jump
 57	 * to hyperspace; gets blown away immediately after
 58	 */
 59
 60	{
 61		r1.l = #LO(_end);
 62		r2.l = #LO(stext);
 63		r3 = #1;
 64	}
 65	{
 66		r1.h = #HI(_end);
 67		r2.h = #HI(stext);
 68		r3 = asl(r3, #22);
 69	}
 70	{
 71		r1 = sub(r1, r2);
 72		r3 = add(r3, #-1);
 73	}  /* r1 =  _end - stext  */
 74	r1 = add(r1, r3);  /*  + (4M-1) */
 75	r26 = lsr(r1, #22); /*  / 4M = # of entries */
 76
 77	r1 = r25;
 78	r2.h = #0xffc0;
 79	r2.l = #0x0000;		/* round back down to 4MB boundary  */
 80	r1 = and(r1,r2);
 81	r2 = lsr(r1, #22)	/* 4MB page number		*/
 82	r2 = asl(r2, #2)	/* times sizeof(PTE) (4bytes)	*/
 83	r0 = add(r0,r2)		/* r0 = address of correct PTE	*/
 84	r2 = #PTE_BITS
 85	r1 = add(r1,r2)		/* r1 = 4MB PTE for the first entry	*/
 86	r2.h = #0x0040
 87	r2.l = #0x0000		/* 4MB increments */
 88	loop0(1f,r26);
 891:
 90	memw(r0 ++ #4) = r1
 91	{ r1 = add(r1, r2); } :endloop0
 92
 93	/*  Also need to overwrite the initial 0xc0000000 entries  */
 94	/*  PAGE_OFFSET >> (4MB shift - 4 bytes per entry shift)  */
 95	R1.H = #HI(PAGE_OFFSET >> (22 - 2))
 96	R1.L = #LO(PAGE_OFFSET >> (22 - 2))
 97
 98	r0 = add(r1, r24);	/* advance to 0xc0000000 entry */
 99	r1 = r25;
100	r2.h = #0xffc0;
101	r2.l = #0x0000;		/* round back down to 4MB boundary  */
102	r1 = and(r1,r2);	/* for huge page */
103	r2 = #PTE_BITS
104	r1 = add(r1,r2);
105	r2.h = #0x0040
106	r2.l = #0x0000		/* 4MB increments */
107
108	loop0(1f,SEGTABLE_ENTRIES);
1091:
110	memw(r0 ++ #4) = r1;
111	{ r1 = add(r1,r2); } :endloop0
112
113	r0 = r24;
114
115	/*
116	 * The subroutine wrapper around the virtual instruction touches
117	 * no memory, so we should be able to use it even here.
118	 * Note that in this version, R1 and R2 get "clobbered"; see
119	 * vm_ops.S
120	 */
121	r1 = #VM_TRANS_TYPE_TABLE
122	call	__vmnewmap;
123
124	/*  Jump into virtual address range.  */
125
126	r31.h = #hi(__head_s_vaddr_target)
127	r31.l = #lo(__head_s_vaddr_target)
128	jumpr r31
129
130	/*  Insert trippy space effects.  */
131
132__head_s_vaddr_target:
133	/*
134	 * Tear down VA=PA translation now that we are running
135	 * in kernel virtual space.
136	 */
137	r0 = #__HVM_PDE_S_INVALID
138
139	r1.h = #0xffc0;
140	r1.l = #0x0000;
141	r2 = r25;		/* phys_offset */
142	r2 = and(r1,r2);
143
144	r1.l = #lo(swapper_pg_dir)
145	r1.h = #hi(swapper_pg_dir)
146	r2 = lsr(r2, #22)	/* 4MB page number		*/
147	r2 = asl(r2, #2)	/* times sizeof(PTE) (4bytes)	*/
148	r1 = add(r1,r2);
149	loop0(1f,r26)
150
1511:
152	{
153		memw(R1 ++ #4) = R0
154	}:endloop0
155
156	r0 = r24
157	r1 = #VM_TRANS_TYPE_TABLE
158	call __vmnewmap
159
160	/*  Go ahead and install the trap0 return so angel calls work  */
161	r0.h = #hi(_K_provisional_vec)
162	r0.l = #lo(_K_provisional_vec)
163	call __vmsetvec
164
165	/*
166	 * OK, at this point we should start to be much more careful,
167	 * we're going to enter C code and start touching memory
168	 * in all sorts of places.
169	 * This means:
170	 *      SGP needs to be OK
171	 *	Need to lock shared resources
172	 *	A bunch of other things that will cause
173	 * 	all kinds of painful bugs
174	 */
175
176	/*
177	 * Stack pointer should be pointed at the init task's
178	 * thread stack, which should have been declared in arch/init_task.c.
179	 * So uhhhhh...
180	 * It's accessible via the init_thread_union, which is a union
181	 * of a thread_info struct and a stack; of course, the top
182	 * of the stack is not for you.  The end of the stack
183	 * is simply init_thread_union + THREAD_SIZE.
184	 */
185
186	{r29.H = #HI(init_thread_union); r0.H = #HI(_THREAD_SIZE); }
187	{r29.L = #LO(init_thread_union); r0.L = #LO(_THREAD_SIZE); }
188
189	/*  initialize the register used to point to current_thread_info */
190	/*  Fixme:  THREADINFO_REG can't be R2 because of that memset thing. */
191	{r29 = add(r29,r0); THREADINFO_REG = r29; }
192
193	/*  Hack:  zero bss; */
194	{ r0.L = #LO(__bss_start);  r1 = #0; r2.l = #LO(__bss_stop); }
195	{ r0.H = #HI(__bss_start);           r2.h = #HI(__bss_stop); }
196
197	r2 = sub(r2,r0);
198	call memset;
199
200	/*  Set PHYS_OFFSET; should be in R25 */
201#ifdef CONFIG_HEXAGON_PHYS_OFFSET
202	r0.l = #LO(__phys_offset);
203	r0.h = #HI(__phys_offset);
204	memw(r0) = r25;
205#endif
206
207	/* Time to make the doughnuts.   */
208	call start_kernel
209
210	/*
211	 * Should not reach here.
212	 */
2131:
214	jump 1b
215
216.p2align PAGE_SHIFT
217ENTRY(external_cmdline_buffer)
218        .fill _PAGE_SIZE,1,0
219
220.data
221.p2align PAGE_SHIFT
222ENTRY(empty_zero_page)
223        .fill _PAGE_SIZE,1,0
v5.14.15
  1/* SPDX-License-Identifier: GPL-2.0-only */
  2/*
  3 * Early kernel startup code for Hexagon
  4 *
  5 * Copyright (c) 2010-2013, The Linux Foundation. All rights reserved.
  6 */
  7
  8#include <linux/linkage.h>
  9#include <linux/init.h>
 10#include <asm/asm-offsets.h>
 11#include <asm/mem-layout.h>
 12#include <asm/vm_mmu.h>
 13#include <asm/page.h>
 14#include <asm/hexagon_vm.h>
 15
 16#define SEGTABLE_ENTRIES #0x0e0
 17
 18	__INIT
 19ENTRY(stext)
 20	/*
 21	 * VMM will already have set up true vector page, MMU, etc.
 22	 * To set up initial kernel identity map, we have to pass
 23	 * the VMM a pointer to some canonical page tables. In
 24	 * this implementation, we're assuming that we've got
 25	 * them precompiled. Generate value in R24, as we'll need
 26	 * it again shortly.
 27	 */
 28	r24.L = #LO(swapper_pg_dir)
 29	r24.H = #HI(swapper_pg_dir)
 30
 31	/*
 32	 * Symbol is kernel segment address, but we need
 33	 * the logical/physical address.
 34	 */
 35	r25 = pc;
 36	r2.h = #0xffc0;
 37	r2.l = #0x0000;
 38	r25 = and(r2,r25);	/*  R25 holds PHYS_OFFSET now  */
 39	r1.h = #HI(PAGE_OFFSET);
 40	r1.l = #LO(PAGE_OFFSET);
 41	r24 = sub(r24,r1);	/* swapper_pg_dir - PAGE_OFFSET */
 42	r24 = add(r24,r25);	/* + PHYS_OFFSET */
 43
 44	r0 = r24;  /* aka __pa(swapper_pg_dir)  */
 45
 46	/*
 47	 * Initialize page dir to make the virtual and physical
 48	 * addresses where the kernel was loaded be identical.
 49	 * Done in 4MB chunks.
 50	 */
 51#define PTE_BITS ( __HVM_PTE_R | __HVM_PTE_W | __HVM_PTE_X	\
 52		  | __HEXAGON_C_WB_L2 << 6			\
 53		  | __HVM_PDE_S_4MB)
 54
 55	/*
 56	 * Get number of VA=PA entries; only really needed for jump
 57	 * to hyperspace; gets blown away immediately after
 58	 */
 59
 60	{
 61		r1.l = #LO(_end);
 62		r2.l = #LO(stext);
 63		r3 = #1;
 64	}
 65	{
 66		r1.h = #HI(_end);
 67		r2.h = #HI(stext);
 68		r3 = asl(r3, #22);
 69	}
 70	{
 71		r1 = sub(r1, r2);
 72		r3 = add(r3, #-1);
 73	}  /* r1 =  _end - stext  */
 74	r1 = add(r1, r3);  /*  + (4M-1) */
 75	r26 = lsr(r1, #22); /*  / 4M = # of entries */
 76
 77	r1 = r25;
 78	r2.h = #0xffc0;
 79	r2.l = #0x0000;		/* round back down to 4MB boundary  */
 80	r1 = and(r1,r2);
 81	r2 = lsr(r1, #22)	/* 4MB page number		*/
 82	r2 = asl(r2, #2)	/* times sizeof(PTE) (4bytes)	*/
 83	r0 = add(r0,r2)		/* r0 = address of correct PTE	*/
 84	r2 = #PTE_BITS
 85	r1 = add(r1,r2)		/* r1 = 4MB PTE for the first entry	*/
 86	r2.h = #0x0040
 87	r2.l = #0x0000		/* 4MB increments */
 88	loop0(1f,r26);
 891:
 90	memw(r0 ++ #4) = r1
 91	{ r1 = add(r1, r2); } :endloop0
 92
 93	/*  Also need to overwrite the initial 0xc0000000 entries  */
 94	/*  PAGE_OFFSET >> (4MB shift - 4 bytes per entry shift)  */
 95	R1.H = #HI(PAGE_OFFSET >> (22 - 2))
 96	R1.L = #LO(PAGE_OFFSET >> (22 - 2))
 97
 98	r0 = add(r1, r24);	/* advance to 0xc0000000 entry */
 99	r1 = r25;
100	r2.h = #0xffc0;
101	r2.l = #0x0000;		/* round back down to 4MB boundary  */
102	r1 = and(r1,r2);	/* for huge page */
103	r2 = #PTE_BITS
104	r1 = add(r1,r2);
105	r2.h = #0x0040
106	r2.l = #0x0000		/* 4MB increments */
107
108	loop0(1f,SEGTABLE_ENTRIES);
1091:
110	memw(r0 ++ #4) = r1;
111	{ r1 = add(r1,r2); } :endloop0
112
113	r0 = r24;
114
115	/*
116	 * The subroutine wrapper around the virtual instruction touches
117	 * no memory, so we should be able to use it even here.
118	 * Note that in this version, R1 and R2 get "clobbered"; see
119	 * vm_ops.S
120	 */
121	r1 = #VM_TRANS_TYPE_TABLE
122	call	__vmnewmap;
123
124	/*  Jump into virtual address range.  */
125
126	r31.h = #hi(__head_s_vaddr_target)
127	r31.l = #lo(__head_s_vaddr_target)
128	jumpr r31
129
130	/*  Insert trippy space effects.  */
131
132__head_s_vaddr_target:
133	/*
134	 * Tear down VA=PA translation now that we are running
135	 * in kernel virtual space.
136	 */
137	r0 = #__HVM_PDE_S_INVALID
138
139	r1.h = #0xffc0;
140	r1.l = #0x0000;
141	r2 = r25;		/* phys_offset */
142	r2 = and(r1,r2);
143
144	r1.l = #lo(swapper_pg_dir)
145	r1.h = #hi(swapper_pg_dir)
146	r2 = lsr(r2, #22)	/* 4MB page number		*/
147	r2 = asl(r2, #2)	/* times sizeof(PTE) (4bytes)	*/
148	r1 = add(r1,r2);
149	loop0(1f,r26)
150
1511:
152	{
153		memw(R1 ++ #4) = R0
154	}:endloop0
155
156	r0 = r24
157	r1 = #VM_TRANS_TYPE_TABLE
158	call __vmnewmap
159
160	/*  Go ahead and install the trap0 return so angel calls work  */
161	r0.h = #hi(_K_provisional_vec)
162	r0.l = #lo(_K_provisional_vec)
163	call __vmsetvec
164
165	/*
166	 * OK, at this point we should start to be much more careful,
167	 * we're going to enter C code and start touching memory
168	 * in all sorts of places.
169	 * This means:
170	 *      SGP needs to be OK
171	 *	Need to lock shared resources
172	 *	A bunch of other things that will cause
173	 * 	all kinds of painful bugs
174	 */
175
176	/*
177	 * Stack pointer should be pointed at the init task's
178	 * thread stack, which should have been declared in arch/init_task.c.
179	 * So uhhhhh...
180	 * It's accessible via the init_thread_union, which is a union
181	 * of a thread_info struct and a stack; of course, the top
182	 * of the stack is not for you.  The end of the stack
183	 * is simply init_thread_union + THREAD_SIZE.
184	 */
185
186	{r29.H = #HI(init_thread_union); r0.H = #HI(_THREAD_SIZE); }
187	{r29.L = #LO(init_thread_union); r0.L = #LO(_THREAD_SIZE); }
188
189	/*  initialize the register used to point to current_thread_info */
190	/*  Fixme:  THREADINFO_REG can't be R2 because of that memset thing. */
191	{r29 = add(r29,r0); THREADINFO_REG = r29; }
192
193	/*  Hack:  zero bss; */
194	{ r0.L = #LO(__bss_start);  r1 = #0; r2.l = #LO(__bss_stop); }
195	{ r0.H = #HI(__bss_start);           r2.h = #HI(__bss_stop); }
196
197	r2 = sub(r2,r0);
198	call memset;
199
200	/*  Set PHYS_OFFSET; should be in R25 */
201#ifdef CONFIG_HEXAGON_PHYS_OFFSET
202	r0.l = #LO(__phys_offset);
203	r0.h = #HI(__phys_offset);
204	memw(r0) = r25;
205#endif
206
207	/* Time to make the doughnuts.   */
208	call start_kernel
209
210	/*
211	 * Should not reach here.
212	 */
2131:
214	jump 1b
215
216.p2align PAGE_SHIFT
217ENTRY(external_cmdline_buffer)
218        .fill _PAGE_SIZE,1,0
219
220.data
221.p2align PAGE_SHIFT
222ENTRY(empty_zero_page)
223        .fill _PAGE_SIZE,1,0