4 * Linux architectural port borrowing liberally from similar works of
5 * others. All original copyrights apply as per the original source
8 * Modifications for the OpenRISC architecture:
9 * Copyright (C) 2003 Matjaz Breskvar <phoenix@bsemi.com>
10 * Copyright (C) 2010-2011 Jonas Bonn <jonas@southpole.se>
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
18 #include <linux/linkage.h>
19 #include <linux/threads.h>
20 #include <linux/errno.h>
21 #include <linux/init.h>
22 #include <linux/serial_reg.h>
23 #include <asm/processor.h>
26 #include <asm/pgtable.h>
27 #include <asm/cache.h>
28 #include <asm/spr_defs.h>
29 #include <asm/asm-offsets.h>
30 #include <linux/of_fdt.h>
32 #define tophys(rd,rs) \
33 l.movhi rd,hi(-KERNELBASE) ;\
36 #define CLEAR_GPR(gpr) \
39 #define LOAD_SYMBOL_2_GPR(gpr,symbol) \
40 l.movhi gpr,hi(symbol) ;\
41 l.ori gpr,gpr,lo(symbol)
44 #define UART_BASE_ADD 0x90000000
46 #define EXCEPTION_SR (SPR_SR_DME | SPR_SR_IME | SPR_SR_DCE | SPR_SR_ICE | SPR_SR_SM)
47 #define SYSCALL_SR (SPR_SR_DME | SPR_SR_IME | SPR_SR_DCE | SPR_SR_ICE | SPR_SR_IEE | SPR_SR_TEE | SPR_SR_SM)
49 /* ============================================[ tmp store locations ]=== */
52 * emergency_print temporary stores
54 #define EMERGENCY_PRINT_STORE_GPR4 l.sw 0x20(r0),r4
55 #define EMERGENCY_PRINT_LOAD_GPR4 l.lwz r4,0x20(r0)
57 #define EMERGENCY_PRINT_STORE_GPR5 l.sw 0x24(r0),r5
58 #define EMERGENCY_PRINT_LOAD_GPR5 l.lwz r5,0x24(r0)
60 #define EMERGENCY_PRINT_STORE_GPR6 l.sw 0x28(r0),r6
61 #define EMERGENCY_PRINT_LOAD_GPR6 l.lwz r6,0x28(r0)
63 #define EMERGENCY_PRINT_STORE_GPR7 l.sw 0x2c(r0),r7
64 #define EMERGENCY_PRINT_LOAD_GPR7 l.lwz r7,0x2c(r0)
66 #define EMERGENCY_PRINT_STORE_GPR8 l.sw 0x30(r0),r8
67 #define EMERGENCY_PRINT_LOAD_GPR8 l.lwz r8,0x30(r0)
69 #define EMERGENCY_PRINT_STORE_GPR9 l.sw 0x34(r0),r9
70 #define EMERGENCY_PRINT_LOAD_GPR9 l.lwz r9,0x34(r0)
74 * TLB miss handlers temorary stores
76 #define EXCEPTION_STORE_GPR9 l.sw 0x10(r0),r9
77 #define EXCEPTION_LOAD_GPR9 l.lwz r9,0x10(r0)
79 #define EXCEPTION_STORE_GPR2 l.sw 0x64(r0),r2
80 #define EXCEPTION_LOAD_GPR2 l.lwz r2,0x64(r0)
82 #define EXCEPTION_STORE_GPR3 l.sw 0x68(r0),r3
83 #define EXCEPTION_LOAD_GPR3 l.lwz r3,0x68(r0)
85 #define EXCEPTION_STORE_GPR4 l.sw 0x6c(r0),r4
86 #define EXCEPTION_LOAD_GPR4 l.lwz r4,0x6c(r0)
88 #define EXCEPTION_STORE_GPR5 l.sw 0x70(r0),r5
89 #define EXCEPTION_LOAD_GPR5 l.lwz r5,0x70(r0)
91 #define EXCEPTION_STORE_GPR6 l.sw 0x74(r0),r6
92 #define EXCEPTION_LOAD_GPR6 l.lwz r6,0x74(r0)
96 * EXCEPTION_HANDLE temporary stores
99 #define EXCEPTION_T_STORE_GPR30 l.sw 0x78(r0),r30
100 #define EXCEPTION_T_LOAD_GPR30(reg) l.lwz reg,0x78(r0)
102 #define EXCEPTION_T_STORE_GPR10 l.sw 0x7c(r0),r10
103 #define EXCEPTION_T_LOAD_GPR10(reg) l.lwz reg,0x7c(r0)
105 #define EXCEPTION_T_STORE_SP l.sw 0x80(r0),r1
106 #define EXCEPTION_T_LOAD_SP(reg) l.lwz reg,0x80(r0)
109 * For UNHANLDED_EXCEPTION
112 #define EXCEPTION_T_STORE_GPR31 l.sw 0x84(r0),r31
113 #define EXCEPTION_T_LOAD_GPR31(reg) l.lwz reg,0x84(r0)
115 /* =========================================================[ macros ]=== */
118 #define GET_CURRENT_PGD(reg,t1) \
119 LOAD_SYMBOL_2_GPR(reg,current_pgd) ;\
125 * DSCR: this is a common hook for handling exceptions. it will save
126 * the needed registers, set up stack and pointer to current
127 * then jump to the handler while enabling MMU
129 * PRMS: handler - a function to jump to. it has to save the
130 * remaining registers to kernel stack, call
131 * appropriate arch-independant exception handler
132 * and finaly jump to ret_from_except
134 * PREQ: unchanged state from the time exception happened
136 * POST: SAVED the following registers original value
137 * to the new created exception frame pointed to by r1
139 * r1 - ksp pointing to the new (exception) frame
140 * r4 - EEAR exception EA
141 * r10 - current pointing to current_thread_info struct
142 * r12 - syscall 0, since we didn't come from syscall
143 * r13 - temp it actually contains new SR, not needed anymore
144 * r31 - handler address of the handler we'll jump to
146 * handler has to save remaining registers to the exception
147 * ksp frame *before* tainting them!
149 * NOTE: this function is not reentrant per se. reentrancy is guaranteed
150 * by processor disabling all exceptions/interrupts when exception
153 * OPTM: no need to make it so wasteful to extract ksp when in user mode
156 #define EXCEPTION_HANDLE(handler) \
157 EXCEPTION_T_STORE_GPR30 ;\
158 l.mfspr r30,r0,SPR_ESR_BASE ;\
159 l.andi r30,r30,SPR_SR_SM ;\
161 EXCEPTION_T_STORE_GPR10 ;\
162 l.bnf 2f /* kernel_mode */ ;\
163 EXCEPTION_T_STORE_SP /* delay slot */ ;\
164 1: /* user_mode: */ ;\
165 LOAD_SYMBOL_2_GPR(r1,current_thread_info_set) ;\
167 /* r10: current_thread_info */ ;\
170 l.lwz r1,(TI_KSP)(r30) ;\
171 /* fall through */ ;\
172 2: /* kernel_mode: */ ;\
173 /* create new stack frame, save only needed gprs */ ;\
174 /* r1: KSP, r10: current, r4: EEAR, r31: __pa(KSP) */ ;\
175 /* r12: temp, syscall indicator */ ;\
176 l.addi r1,r1,-(INT_FRAME_SIZE) ;\
177 /* r1 is KSP, r30 is __pa(KSP) */ ;\
179 l.sw PT_GPR12(r30),r12 ;\
180 l.mfspr r12,r0,SPR_EPCR_BASE ;\
181 l.sw PT_PC(r30),r12 ;\
182 l.mfspr r12,r0,SPR_ESR_BASE ;\
183 l.sw PT_SR(r30),r12 ;\
185 EXCEPTION_T_LOAD_GPR30(r12) ;\
186 l.sw PT_GPR30(r30),r12 ;\
187 /* save r10 as was prior to exception */ ;\
188 EXCEPTION_T_LOAD_GPR10(r12) ;\
189 l.sw PT_GPR10(r30),r12 ;\
190 /* save PT_SP as was prior to exception */ ;\
191 EXCEPTION_T_LOAD_SP(r12) ;\
192 l.sw PT_SP(r30),r12 ;\
193 /* save exception r4, set r4 = EA */ ;\
194 l.sw PT_GPR4(r30),r4 ;\
195 l.mfspr r4,r0,SPR_EEAR_BASE ;\
196 /* r12 == 1 if we come from syscall */ ;\
198 /* ----- turn on MMU ----- */ ;\
199 l.ori r30,r0,(EXCEPTION_SR) ;\
200 l.mtspr r0,r30,SPR_ESR_BASE ;\
201 /* r30: EA address of handler */ ;\
202 LOAD_SYMBOL_2_GPR(r30,handler) ;\
203 l.mtspr r0,r30,SPR_EPCR_BASE ;\
210 * #ifdef CONFIG_JUMP_UPON_UNHANDLED_EXCEPTION
211 * #define UNHANDLED_EXCEPTION(handler) \
213 * l.mtspr r0,r3,SPR_SR ;\
214 * l.movhi r3,hi(0xf0000100) ;\
215 * l.ori r3,r3,lo(0xf0000100) ;\
222 /* DSCR: this is the same as EXCEPTION_HANDLE(), we are just
223 * a bit more carefull (if we have a PT_SP or current pointer
224 * corruption) and set them up from 'current_set'
227 #define UNHANDLED_EXCEPTION(handler) \
228 EXCEPTION_T_STORE_GPR31 ;\
229 EXCEPTION_T_STORE_GPR10 ;\
230 EXCEPTION_T_STORE_SP ;\
231 /* temporary store r3, r9 into r1, r10 */ ;\
234 /* the string referenced by r3 must be low enough */ ;\
235 l.jal _emergency_print ;\
236 l.ori r3,r0,lo(_string_unhandled_exception) ;\
237 l.mfspr r3,r0,SPR_NPC ;\
238 l.jal _emergency_print_nr ;\
239 l.andi r3,r3,0x1f00 ;\
240 /* the string referenced by r3 must be low enough */ ;\
241 l.jal _emergency_print ;\
242 l.ori r3,r0,lo(_string_epc_prefix) ;\
243 l.jal _emergency_print_nr ;\
244 l.mfspr r3,r0,SPR_EPCR_BASE ;\
245 l.jal _emergency_print ;\
246 l.ori r3,r0,lo(_string_nl) ;\
247 /* end of printing */ ;\
250 /* extract current, ksp from current_set */ ;\
251 LOAD_SYMBOL_2_GPR(r1,_unhandled_stack_top) ;\
252 LOAD_SYMBOL_2_GPR(r10,init_thread_union) ;\
253 /* create new stack frame, save only needed gprs */ ;\
254 /* r1: KSP, r10: current, r31: __pa(KSP) */ ;\
255 /* r12: temp, syscall indicator, r13 temp */ ;\
256 l.addi r1,r1,-(INT_FRAME_SIZE) ;\
257 /* r1 is KSP, r31 is __pa(KSP) */ ;\
259 l.sw PT_GPR12(r31),r12 ;\
260 l.mfspr r12,r0,SPR_EPCR_BASE ;\
261 l.sw PT_PC(r31),r12 ;\
262 l.mfspr r12,r0,SPR_ESR_BASE ;\
263 l.sw PT_SR(r31),r12 ;\
265 EXCEPTION_T_LOAD_GPR31(r12) ;\
266 l.sw PT_GPR31(r31),r12 ;\
267 /* save r10 as was prior to exception */ ;\
268 EXCEPTION_T_LOAD_GPR10(r12) ;\
269 l.sw PT_GPR10(r31),r12 ;\
270 /* save PT_SP as was prior to exception */ ;\
271 EXCEPTION_T_LOAD_SP(r12) ;\
272 l.sw PT_SP(r31),r12 ;\
273 l.sw PT_GPR13(r31),r13 ;\
275 /* save exception r4, set r4 = EA */ ;\
276 l.sw PT_GPR4(r31),r4 ;\
277 l.mfspr r4,r0,SPR_EEAR_BASE ;\
278 /* r12 == 1 if we come from syscall */ ;\
280 /* ----- play a MMU trick ----- */ ;\
281 l.ori r31,r0,(EXCEPTION_SR) ;\
282 l.mtspr r0,r31,SPR_ESR_BASE ;\
283 /* r31: EA address of handler */ ;\
284 LOAD_SYMBOL_2_GPR(r31,handler) ;\
285 l.mtspr r0,r31,SPR_EPCR_BASE ;\
288 /* =====================================================[ exceptions] === */
290 /* ---[ 0x100: RESET exception ]----------------------------------------- */
292 /* Jump to .init code at _start which lives in the .head section
293 * and will be discarded after boot.
295 LOAD_SYMBOL_2_GPR(r15, _start)
296 tophys (r13,r15) /* MMU disabled */
300 /* ---[ 0x200: BUS exception ]------------------------------------------- */
303 EXCEPTION_HANDLE(_bus_fault_handler)
305 /* ---[ 0x300: Data Page Fault exception ]------------------------------- */
307 _dispatch_do_dpage_fault:
308 // totaly disable timer interrupt
309 // l.mtspr r0,r0,SPR_TTMR
310 // DEBUG_TLB_PROBE(0x300)
311 // EXCEPTION_DEBUG_VALUE_ER_ENABLED(0x300)
312 EXCEPTION_HANDLE(_data_page_fault_handler)
314 /* ---[ 0x400: Insn Page Fault exception ]------------------------------- */
316 _dispatch_do_ipage_fault:
317 // totaly disable timer interrupt
318 // l.mtspr r0,r0,SPR_TTMR
319 // DEBUG_TLB_PROBE(0x400)
320 // EXCEPTION_DEBUG_VALUE_ER_ENABLED(0x400)
321 EXCEPTION_HANDLE(_insn_page_fault_handler)
323 /* ---[ 0x500: Timer exception ]----------------------------------------- */
325 EXCEPTION_HANDLE(_timer_handler)
327 /* ---[ 0x600: Aligment exception ]-------------------------------------- */
329 EXCEPTION_HANDLE(_alignment_handler)
331 /* ---[ 0x700: Illegal insn exception ]---------------------------------- */
333 EXCEPTION_HANDLE(_illegal_instruction_handler)
335 /* ---[ 0x800: External interrupt exception ]---------------------------- */
337 EXCEPTION_HANDLE(_external_irq_handler)
339 /* ---[ 0x900: DTLB miss exception ]------------------------------------- */
341 l.j boot_dtlb_miss_handler
344 /* ---[ 0xa00: ITLB miss exception ]------------------------------------- */
346 l.j boot_itlb_miss_handler
349 /* ---[ 0xb00: Range exception ]----------------------------------------- */
351 UNHANDLED_EXCEPTION(_vector_0xb00)
353 /* ---[ 0xc00: Syscall exception ]--------------------------------------- */
355 EXCEPTION_HANDLE(_sys_call_handler)
357 /* ---[ 0xd00: Trap exception ]------------------------------------------ */
359 UNHANDLED_EXCEPTION(_vector_0xd00)
361 /* ---[ 0xe00: Trap exception ]------------------------------------------ */
363 // UNHANDLED_EXCEPTION(_vector_0xe00)
364 EXCEPTION_HANDLE(_trap_handler)
366 /* ---[ 0xf00: Reserved exception ]-------------------------------------- */
368 UNHANDLED_EXCEPTION(_vector_0xf00)
370 /* ---[ 0x1000: Reserved exception ]------------------------------------- */
372 UNHANDLED_EXCEPTION(_vector_0x1000)
374 /* ---[ 0x1100: Reserved exception ]------------------------------------- */
376 UNHANDLED_EXCEPTION(_vector_0x1100)
378 /* ---[ 0x1200: Reserved exception ]------------------------------------- */
380 UNHANDLED_EXCEPTION(_vector_0x1200)
382 /* ---[ 0x1300: Reserved exception ]------------------------------------- */
384 UNHANDLED_EXCEPTION(_vector_0x1300)
386 /* ---[ 0x1400: Reserved exception ]------------------------------------- */
388 UNHANDLED_EXCEPTION(_vector_0x1400)
390 /* ---[ 0x1500: Reserved exception ]------------------------------------- */
392 UNHANDLED_EXCEPTION(_vector_0x1500)
394 /* ---[ 0x1600: Reserved exception ]------------------------------------- */
396 UNHANDLED_EXCEPTION(_vector_0x1600)
398 /* ---[ 0x1700: Reserved exception ]------------------------------------- */
400 UNHANDLED_EXCEPTION(_vector_0x1700)
402 /* ---[ 0x1800: Reserved exception ]------------------------------------- */
404 UNHANDLED_EXCEPTION(_vector_0x1800)
406 /* ---[ 0x1900: Reserved exception ]------------------------------------- */
408 UNHANDLED_EXCEPTION(_vector_0x1900)
410 /* ---[ 0x1a00: Reserved exception ]------------------------------------- */
412 UNHANDLED_EXCEPTION(_vector_0x1a00)
414 /* ---[ 0x1b00: Reserved exception ]------------------------------------- */
416 UNHANDLED_EXCEPTION(_vector_0x1b00)
418 /* ---[ 0x1c00: Reserved exception ]------------------------------------- */
420 UNHANDLED_EXCEPTION(_vector_0x1c00)
422 /* ---[ 0x1d00: Reserved exception ]------------------------------------- */
424 UNHANDLED_EXCEPTION(_vector_0x1d00)
426 /* ---[ 0x1e00: Reserved exception ]------------------------------------- */
428 UNHANDLED_EXCEPTION(_vector_0x1e00)
430 /* ---[ 0x1f00: Reserved exception ]------------------------------------- */
432 UNHANDLED_EXCEPTION(_vector_0x1f00)
435 /* ===================================================[ kernel start ]=== */
439 /* This early stuff belongs in HEAD, but some of the functions below definitely
445 /* save kernel parameters */
446 l.or r25,r0,r3 /* pointer to fdt */
449 * ensure a deterministic start
456 * Start the TTCR as early as possible, so that the RNG can make use of
457 * measurements of boot time from the earliest opportunity. Especially
458 * important is that the TTCR does not return zero by the time we reach
461 l.movhi r3,hi(SPR_TTMR_CR)
462 l.mtspr r0,r3,SPR_TTMR
496 * set up initial ksp and current
498 LOAD_SYMBOL_2_GPR(r1,init_thread_union+0x2000) // setup kernel stack
499 LOAD_SYMBOL_2_GPR(r10,init_thread_union) // setup current
507 * .data contains initialized data,
508 * .bss contains uninitialized data - clear it up
511 LOAD_SYMBOL_2_GPR(r24, __bss_start)
512 LOAD_SYMBOL_2_GPR(r26, _end)
533 * I N V A L I D A T E T L B e n t r i e s
535 LOAD_SYMBOL_2_GPR(r5,SPR_DTLBMR_BASE(0))
536 LOAD_SYMBOL_2_GPR(r6,SPR_ITLBMR_BASE(0))
537 l.addi r7,r0,128 /* Maximum number of sets */
549 /* The MMU needs to be enabled before or32_early_setup is called */
554 * SR[5] = 0, SR[6] = 0, 6th and 7th bit of SR set to 0
556 l.mfspr r30,r0,SPR_SR
557 l.movhi r28,hi(SPR_SR_DME | SPR_SR_IME)
558 l.ori r28,r28,lo(SPR_SR_DME | SPR_SR_IME)
560 l.mtspr r0,r30,SPR_SR
578 // reset the simulation counters
581 /* check fdt header magic word */
582 l.lwz r3,0(r25) /* load magic from fdt into r3 */
583 l.movhi r4,hi(OF_DT_HEADER)
584 l.ori r4,r4,lo(OF_DT_HEADER)
588 /* magic number mismatch, set fdt pointer to null */
591 /* pass fdt pointer to or32_early_setup in r3 */
593 LOAD_SYMBOL_2_GPR(r24, or32_early_setup)
599 * clear all GPRS to increase determinism
633 * jump to kernel entry (start_kernel)
635 LOAD_SYMBOL_2_GPR(r30, start_kernel)
639 /* ========================================[ cache ]=== */
641 /* aligment here so we don't change memory offsets with
642 * memory controler defined
647 /* Check if IC present and skip enabling otherwise */
648 l.mfspr r24,r0,SPR_UPR
649 l.andi r26,r24,SPR_UPR_ICP
657 l.xori r5,r5,SPR_SR_ICE
661 /* Establish cache block size
664 r14 contain block size
666 l.mfspr r24,r0,SPR_ICCFGR
667 l.andi r26,r24,SPR_ICCFGR_CBS
672 /* Establish number of cache sets
673 r16 contains number of cache sets
674 r28 contains log(# of cache sets)
676 l.andi r26,r24,SPR_ICCFGR_NCS
686 // l.addi r5,r0,IC_SIZE
688 l.mtspr r0,r6,SPR_ICBIR
692 // l.addi r6,r6,IC_LINE
696 l.ori r6,r6,SPR_SR_ICE
713 /* Check if DC present and skip enabling otherwise */
714 l.mfspr r24,r0,SPR_UPR
715 l.andi r26,r24,SPR_UPR_DCP
723 l.xori r5,r5,SPR_SR_DCE
727 /* Establish cache block size
730 r14 contain block size
732 l.mfspr r24,r0,SPR_DCCFGR
733 l.andi r26,r24,SPR_DCCFGR_CBS
738 /* Establish number of cache sets
739 r16 contains number of cache sets
740 r28 contains log(# of cache sets)
742 l.andi r26,r24,SPR_DCCFGR_NCS
751 l.mtspr r0,r6,SPR_DCBIR
758 l.ori r6,r6,SPR_SR_DCE
764 /* ===============================================[ page table masks ]=== */
766 #define DTLB_UP_CONVERT_MASK 0x3fa
767 #define ITLB_UP_CONVERT_MASK 0x3a
769 /* for SMP we'd have (this is a bit subtle, CC must be always set
770 * for SMP, but since we have _PAGE_PRESENT bit always defined
771 * we can just modify the mask)
773 #define DTLB_SMP_CONVERT_MASK 0x3fb
774 #define ITLB_SMP_CONVERT_MASK 0x3b
776 /* ---[ boot dtlb miss handler ]----------------------------------------- */
778 boot_dtlb_miss_handler:
780 /* mask for DTLB_MR register: - (0) sets V (valid) bit,
781 * - (31-12) sets bits belonging to VPN (31-12)
783 #define DTLB_MR_MASK 0xfffff001
785 /* mask for DTLB_TR register: - (2) sets CI (cache inhibit) bit,
786 * - (4) sets A (access) bit,
787 * - (5) sets D (dirty) bit,
788 * - (8) sets SRE (superuser read) bit
789 * - (9) sets SWE (superuser write) bit
790 * - (31-12) sets bits belonging to VPN (31-12)
792 #define DTLB_TR_MASK 0xfffff332
794 /* These are for masking out the VPN/PPN value from the MR/TR registers...
795 * it's not the same as the PFN */
796 #define VPN_MASK 0xfffff000
797 #define PPN_MASK 0xfffff000
803 l.mfspr r6,r0,SPR_ESR_BASE //
804 l.andi r6,r6,SPR_SR_SM // are we in kernel mode ?
805 l.sfeqi r6,0 // r6 == 0x1 --> SM
806 l.bf exit_with_no_dtranslation //
810 /* this could be optimized by moving storing of
811 * non r6 registers here, and jumping r6 restore
812 * if not in supervisor mode
820 l.mfspr r4,r0,SPR_EEAR_BASE // get the offending EA
822 immediate_translation:
825 l.srli r3,r4,0xd // r3 <- r4 / 8192 (sets are relative to page size (8Kb) NOT VPN size (4Kb)
827 l.mfspr r6, r0, SPR_DMMUCFGR
828 l.andi r6, r6, SPR_DMMUCFGR_NTS
829 l.srli r6, r6, SPR_DMMUCFGR_NTS_OFF
831 l.sll r5, r5, r6 // r5 = number DMMU sets
832 l.addi r6, r5, -1 // r6 = nsets mask
833 l.and r2, r3, r6 // r2 <- r3 % NSETS_MASK
835 l.or r6,r6,r4 // r6 <- r4
836 l.ori r6,r6,~(VPN_MASK) // r6 <- VPN :VPN .xfff - clear up lo(r6) to 0x**** *fff
837 l.movhi r5,hi(DTLB_MR_MASK) // r5 <- ffff:0000.x000
838 l.ori r5,r5,lo(DTLB_MR_MASK) // r5 <- ffff:1111.x001 - apply DTLB_MR_MASK
839 l.and r5,r5,r6 // r5 <- VPN :VPN .x001 - we have DTLBMR entry
840 l.mtspr r2,r5,SPR_DTLBMR_BASE(0) // set DTLBMR
842 /* set up DTLB with no translation for EA <= 0xbfffffff */
843 LOAD_SYMBOL_2_GPR(r6,0xbfffffff)
844 l.sfgeu r6,r4 // flag if r6 >= r4 (if 0xbfffffff >= EA)
846 l.and r3,r4,r4 // delay slot :: 24 <- r4 (if flag==1)
848 tophys(r3,r4) // r3 <- PA
850 l.ori r3,r3,~(PPN_MASK) // r3 <- PPN :PPN .xfff - clear up lo(r6) to 0x**** *fff
851 l.movhi r5,hi(DTLB_TR_MASK) // r5 <- ffff:0000.x000
852 l.ori r5,r5,lo(DTLB_TR_MASK) // r5 <- ffff:1111.x330 - apply DTLB_MR_MASK
853 l.and r5,r5,r3 // r5 <- PPN :PPN .x330 - we have DTLBTR entry
854 l.mtspr r2,r5,SPR_DTLBTR_BASE(0) // set DTLBTR
862 l.rfe // SR <- ESR, PC <- EPC
864 exit_with_no_dtranslation:
865 /* EA out of memory or not in supervisor mode */
868 l.j _dispatch_bus_fault
870 /* ---[ boot itlb miss handler ]----------------------------------------- */
872 boot_itlb_miss_handler:
874 /* mask for ITLB_MR register: - sets V (valid) bit,
875 * - sets bits belonging to VPN (15-12)
877 #define ITLB_MR_MASK 0xfffff001
879 /* mask for ITLB_TR register: - sets A (access) bit,
880 * - sets SXE (superuser execute) bit
881 * - sets bits belonging to VPN (15-12)
883 #define ITLB_TR_MASK 0xfffff050
886 #define VPN_MASK 0xffffe000
887 #define PPN_MASK 0xffffe000
899 l.mfspr r6,r0,SPR_ESR_BASE //
900 l.andi r6,r6,SPR_SR_SM // are we in kernel mode ?
901 l.sfeqi r6,0 // r6 == 0x1 --> SM
902 l.bf exit_with_no_itranslation
907 l.mfspr r4,r0,SPR_EEAR_BASE // get the offending EA
912 l.srli r3,r4,0xd // r3 <- r4 / 8192 (sets are relative to page size (8Kb) NOT VPN size (4Kb)
914 l.mfspr r6, r0, SPR_IMMUCFGR
915 l.andi r6, r6, SPR_IMMUCFGR_NTS
916 l.srli r6, r6, SPR_IMMUCFGR_NTS_OFF
918 l.sll r5, r5, r6 // r5 = number IMMU sets from IMMUCFGR
919 l.addi r6, r5, -1 // r6 = nsets mask
920 l.and r2, r3, r6 // r2 <- r3 % NSETS_MASK
922 l.or r6,r6,r4 // r6 <- r4
923 l.ori r6,r6,~(VPN_MASK) // r6 <- VPN :VPN .xfff - clear up lo(r6) to 0x**** *fff
924 l.movhi r5,hi(ITLB_MR_MASK) // r5 <- ffff:0000.x000
925 l.ori r5,r5,lo(ITLB_MR_MASK) // r5 <- ffff:1111.x001 - apply ITLB_MR_MASK
926 l.and r5,r5,r6 // r5 <- VPN :VPN .x001 - we have ITLBMR entry
927 l.mtspr r2,r5,SPR_ITLBMR_BASE(0) // set ITLBMR
930 * set up ITLB with no translation for EA <= 0x0fffffff
932 * we need this for head.S mapping (EA = PA). if we move all functions
933 * which run with mmu enabled into entry.S, we might be able to eliminate this.
936 LOAD_SYMBOL_2_GPR(r6,0x0fffffff)
937 l.sfgeu r6,r4 // flag if r6 >= r4 (if 0xb0ffffff >= EA)
939 l.and r3,r4,r4 // delay slot :: 24 <- r4 (if flag==1)
941 tophys(r3,r4) // r3 <- PA
943 l.ori r3,r3,~(PPN_MASK) // r3 <- PPN :PPN .xfff - clear up lo(r6) to 0x**** *fff
944 l.movhi r5,hi(ITLB_TR_MASK) // r5 <- ffff:0000.x000
945 l.ori r5,r5,lo(ITLB_TR_MASK) // r5 <- ffff:1111.x050 - apply ITLB_MR_MASK
946 l.and r5,r5,r3 // r5 <- PPN :PPN .x050 - we have ITLBTR entry
947 l.mtspr r2,r5,SPR_ITLBTR_BASE(0) // set ITLBTR
955 l.rfe // SR <- ESR, PC <- EPC
957 exit_with_no_itranslation:
960 l.j _dispatch_bus_fault
963 /* ====================================================================== */
965 * Stuff below here shouldn't go into .head section... maybe this stuff
966 * can be moved to entry.S ???
969 /* ==============================================[ DTLB miss handler ]=== */
973 * Exception handlers are entered with MMU off so the following handler
974 * needs to use physical addressing
979 ENTRY(dtlb_miss_handler)
988 l.mfspr r2,r0,SPR_EEAR_BASE
990 * pmd = (pmd_t *)(current_pgd + pgd_index(daddr));
992 GET_CURRENT_PGD(r3,r5) // r3 is current_pgd, r5 is temp
993 l.srli r4,r2,0x18 // >> PAGE_SHIFT + (PAGE_SHIFT - 2)
994 l.slli r4,r4,0x2 // to get address << 2
995 l.add r5,r4,r3 // r4 is pgd_index(daddr)
997 * if (pmd_none(*pmd))
1001 l.lwz r3,0x0(r4) // get *pmd value
1004 l.andi r3,r3,~PAGE_MASK //0x1fff // ~PAGE_MASK
1006 * if (pmd_bad(*pmd))
1010 // l.sfeq r3,r0 // check *pmd value
1012 l.addi r3,r0,0xffffe000 // PAGE_MASK
1014 // l.sw 0x0(r4),r0 // clear pmd
1017 * pte = *pte_offset(pmd, daddr);
1019 l.lwz r4,0x0(r4) // get **pmd value
1020 l.and r4,r4,r3 // & PAGE_MASK
1021 l.srli r5,r2,0xd // >> PAGE_SHIFT, r2 == EEAR
1022 l.andi r3,r5,0x7ff // (1UL << PAGE_SHIFT - 2) - 1
1023 l.slli r3,r3,0x2 // to get address << 2
1025 l.lwz r2,0x0(r3) // this is pte at last
1027 * if (!pte_present(pte))
1030 l.sfne r4,r0 // is pte present
1031 l.bnf d_pte_not_present
1032 l.addi r3,r0,0xffffe3fa // PAGE_MASK | DTLB_UP_CONVERT_MASK
1034 * fill DTLB TR register
1036 l.and r4,r2,r3 // apply the mask
1037 // Determine number of DMMU sets
1038 l.mfspr r6, r0, SPR_DMMUCFGR
1039 l.andi r6, r6, SPR_DMMUCFGR_NTS
1040 l.srli r6, r6, SPR_DMMUCFGR_NTS_OFF
1042 l.sll r3, r3, r6 // r3 = number DMMU sets DMMUCFGR
1043 l.addi r6, r3, -1 // r6 = nsets mask
1044 l.and r5, r5, r6 // calc offset: & (NUM_TLB_ENTRIES-1)
1046 l.mtspr r5,r4,SPR_DTLBTR_BASE(0)
1048 * fill DTLB MR register
1050 l.mfspr r2,r0,SPR_EEAR_BASE
1051 l.addi r3,r0,0xffffe000 // PAGE_MASK
1052 l.and r4,r2,r3 // apply PAGE_MASK to EA (__PHX__ do we really need this?)
1053 l.ori r4,r4,0x1 // set hardware valid bit: DTBL_MR entry
1054 l.mtspr r5,r4,SPR_DTLBMR_BASE(0)
1077 EXCEPTION_HANDLE(_dtlb_miss_page_fault_handler)
1079 /* ==============================================[ ITLB miss handler ]=== */
1080 ENTRY(itlb_miss_handler)
1081 EXCEPTION_STORE_GPR2
1082 EXCEPTION_STORE_GPR3
1083 EXCEPTION_STORE_GPR4
1084 EXCEPTION_STORE_GPR5
1085 EXCEPTION_STORE_GPR6
1087 * get EA of the miss
1089 l.mfspr r2,r0,SPR_EEAR_BASE
1092 * pmd = (pmd_t *)(current_pgd + pgd_index(daddr));
1095 GET_CURRENT_PGD(r3,r5) // r3 is current_pgd, r5 is temp
1096 l.srli r4,r2,0x18 // >> PAGE_SHIFT + (PAGE_SHIFT - 2)
1097 l.slli r4,r4,0x2 // to get address << 2
1098 l.add r5,r4,r3 // r4 is pgd_index(daddr)
1100 * if (pmd_none(*pmd))
1104 l.lwz r3,0x0(r4) // get *pmd value
1107 l.andi r3,r3,0x1fff // ~PAGE_MASK
1109 * if (pmd_bad(*pmd))
1114 // l.sfeq r3,r0 // check *pmd value
1116 l.addi r3,r0,0xffffe000 // PAGE_MASK
1118 // l.sw 0x0(r4),r0 // clear pmd
1122 * pte = *pte_offset(pmd, iaddr);
1125 l.lwz r4,0x0(r4) // get **pmd value
1126 l.and r4,r4,r3 // & PAGE_MASK
1127 l.srli r5,r2,0xd // >> PAGE_SHIFT, r2 == EEAR
1128 l.andi r3,r5,0x7ff // (1UL << PAGE_SHIFT - 2) - 1
1129 l.slli r3,r3,0x2 // to get address << 2
1131 l.lwz r2,0x0(r3) // this is pte at last
1133 * if (!pte_present(pte))
1137 l.sfne r4,r0 // is pte present
1138 l.bnf i_pte_not_present
1139 l.addi r3,r0,0xffffe03a // PAGE_MASK | ITLB_UP_CONVERT_MASK
1141 * fill ITLB TR register
1143 l.and r4,r2,r3 // apply the mask
1144 l.andi r3,r2,0x7c0 // _PAGE_EXEC | _PAGE_SRE | _PAGE_SWE | _PAGE_URE | _PAGE_UWE
1145 // l.andi r3,r2,0x400 // _PAGE_EXEC
1147 l.bf itlb_tr_fill //_workaround
1148 // Determine number of IMMU sets
1149 l.mfspr r6, r0, SPR_IMMUCFGR
1150 l.andi r6, r6, SPR_IMMUCFGR_NTS
1151 l.srli r6, r6, SPR_IMMUCFGR_NTS_OFF
1153 l.sll r3, r3, r6 // r3 = number IMMU sets IMMUCFGR
1154 l.addi r6, r3, -1 // r6 = nsets mask
1155 l.and r5, r5, r6 // calc offset: & (NUM_TLB_ENTRIES-1)
1159 * we should not just blindly set executable flags,
1160 * but it does help with ping. the clean way would be to find out
1161 * (and fix it) why stack doesn't have execution permissions
1164 itlb_tr_fill_workaround:
1165 l.ori r4,r4,0xc0 // | (SPR_ITLBTR_UXE | ITLBTR_SXE)
1167 l.mtspr r5,r4,SPR_ITLBTR_BASE(0)
1169 * fill DTLB MR register
1171 l.mfspr r2,r0,SPR_EEAR_BASE
1172 l.addi r3,r0,0xffffe000 // PAGE_MASK
1173 l.and r4,r2,r3 // apply PAGE_MASK to EA (__PHX__ do we really need this?)
1174 l.ori r4,r4,0x1 // set hardware valid bit: DTBL_MR entry
1175 l.mtspr r5,r4,SPR_ITLBMR_BASE(0)
1199 EXCEPTION_HANDLE(_itlb_miss_page_fault_handler)
1201 /* ==============================================[ boot tlb handlers ]=== */
1204 /* =================================================[ debugging aids ]=== */
1209 _immu_trampoline_top:
1211 #define TRAMP_SLOT_0 (0x0)
1212 #define TRAMP_SLOT_1 (0x4)
1213 #define TRAMP_SLOT_2 (0x8)
1214 #define TRAMP_SLOT_3 (0xc)
1215 #define TRAMP_SLOT_4 (0x10)
1216 #define TRAMP_SLOT_5 (0x14)
1217 #define TRAMP_FRAME_SIZE (0x18)
1219 ENTRY(_immu_trampoline_workaround)
1221 // r6 is physical EEA
1224 LOAD_SYMBOL_2_GPR(r5,_immu_trampoline)
1225 tophys (r3,r5) // r3 is trampoline (physical)
1227 LOAD_SYMBOL_2_GPR(r4,0x15000000)
1228 l.sw TRAMP_SLOT_0(r3),r4
1229 l.sw TRAMP_SLOT_1(r3),r4
1230 l.sw TRAMP_SLOT_4(r3),r4
1231 l.sw TRAMP_SLOT_5(r3),r4
1234 l.lwz r4,0x0(r6) // load op @ EEA + 0x0 (fc address)
1235 l.sw TRAMP_SLOT_3(r3),r4 // store it to _immu_trampoline_data
1236 l.lwz r4,-0x4(r6) // load op @ EEA - 0x4 (f8 address)
1237 l.sw TRAMP_SLOT_2(r3),r4 // store it to _immu_trampoline_data
1239 l.srli r5,r4,26 // check opcode for write access
1242 l.sfeqi r5,0x11 // l.jr
1244 l.sfeqi r5,1 // l.jal
1246 l.sfeqi r5,0x12 // l.jalr
1248 l.sfeqi r5,3 // l.bnf
1250 l.sfeqi r5,4 // l.bf
1254 l.j 99b // should never happen
1258 // r3 is trampoline address (physical)
1259 // r4 is instruction
1260 // r6 is physical(EEA)
1266 /* 19 20 aa aa l.movhi r9,0xaaaa
1267 * a9 29 bb bb l.ori r9,0xbbbb
1269 * where 0xaaaabbbb is EEA + 0x4 shifted right 2
1272 l.addi r6,r2,0x4 // this is 0xaaaabbbb
1274 // l.movhi r9,0xaaaa
1275 l.ori r5,r0,0x1920 // 0x1920 == l.movhi r9
1276 l.sh (TRAMP_SLOT_0+0x0)(r3),r5
1278 l.sh (TRAMP_SLOT_0+0x2)(r3),r5
1281 l.ori r5,r0,0xa929 // 0xa929 == l.ori r9
1282 l.sh (TRAMP_SLOT_1+0x0)(r3),r5
1284 l.sh (TRAMP_SLOT_1+0x2)(r3),r5
1286 /* falthrough, need to set up new jump offset */
1290 l.slli r6,r4,6 // original offset shifted left 6 - 2
1291 // l.srli r6,r6,6 // original offset shifted right 2
1293 l.slli r4,r2,4 // old jump position: EEA shifted left 4
1294 // l.srli r4,r4,6 // old jump position: shifted right 2
1296 l.addi r5,r3,0xc // new jump position (physical)
1297 l.slli r5,r5,4 // new jump position: shifted left 4
1299 // calculate new jump offset
1300 // new_off = old_off + (old_jump - new_jump)
1302 l.sub r5,r4,r5 // old_jump - new_jump
1303 l.add r5,r6,r5 // orig_off + (old_jump - new_jump)
1304 l.srli r5,r5,6 // new offset shifted right 2
1306 // r5 is new jump offset
1307 // l.j has opcode 0x0...
1308 l.sw TRAMP_SLOT_2(r3),r5 // write it back
1313 /* ----------------------------- */
1317 /* 19 20 aa aa l.movhi r9,0xaaaa
1318 * a9 29 bb bb l.ori r9,0xbbbb
1320 * where 0xaaaabbbb is EEA + 0x4 shifted right 2
1323 l.addi r6,r2,0x4 // this is 0xaaaabbbb
1325 // l.movhi r9,0xaaaa
1326 l.ori r5,r0,0x1920 // 0x1920 == l.movhi r9
1327 l.sh (TRAMP_SLOT_0+0x0)(r3),r5
1329 l.sh (TRAMP_SLOT_0+0x2)(r3),r5
1332 l.ori r5,r0,0xa929 // 0xa929 == l.ori r9
1333 l.sh (TRAMP_SLOT_1+0x0)(r3),r5
1335 l.sh (TRAMP_SLOT_1+0x2)(r3),r5
1337 l.lhz r5,(TRAMP_SLOT_2+0x0)(r3) // load hi part of jump instruction
1338 l.andi r5,r5,0x3ff // clear out opcode part
1339 l.ori r5,r5,0x4400 // opcode changed from l.jalr -> l.jr
1340 l.sh (TRAMP_SLOT_2+0x0)(r3),r5 // write it back
1348 /* ----------------------------- */
1352 l.slli r6,r4,6 // original offset shifted left 6 - 2
1353 // l.srli r6,r6,6 // original offset shifted right 2
1355 l.slli r4,r2,4 // old jump position: EEA shifted left 4
1356 // l.srli r4,r4,6 // old jump position: shifted right 2
1358 l.addi r5,r3,0xc // new jump position (physical)
1359 l.slli r5,r5,4 // new jump position: shifted left 4
1361 // calculate new jump offset
1362 // new_off = old_off + (old_jump - new_jump)
1364 l.add r6,r6,r4 // (orig_off + old_jump)
1365 l.sub r6,r6,r5 // (orig_off + old_jump) - new_jump
1366 l.srli r6,r6,6 // new offset shifted right 2
1368 // r6 is new jump offset
1369 l.lwz r4,(TRAMP_SLOT_2+0x0)(r3) // load jump instruction
1371 l.andi r4,r4,0xfc00 // get opcode part
1373 l.or r6,r4,r6 // l.b(n)f new offset
1374 l.sw TRAMP_SLOT_2(r3),r6 // write it back
1376 /* we need to add l.j to EEA + 0x8 */
1377 tophys (r4,r2) // may not be needed (due to shifts down_
1378 l.addi r4,r4,(0x8 - 0x8) // jump target = r2 + 0x8 (compensate for 0x8)
1379 // jump position = r5 + 0x8 (0x8 compensated)
1380 l.sub r4,r4,r5 // jump offset = target - new_position + 0x8
1382 l.slli r4,r4,4 // the amount of info in imediate of jump
1383 l.srli r4,r4,6 // jump instruction with offset
1384 l.sw TRAMP_SLOT_4(r3),r4 // write it to 4th slot
1389 // set up new EPC to point to our trampoline code
1390 LOAD_SYMBOL_2_GPR(r5,_immu_trampoline)
1391 l.mtspr r0,r5,SPR_EPCR_BASE
1393 // immu_trampoline is (4x) CACHE_LINE aligned
1394 // and only 6 instructions long,
1395 // so we need to invalidate only 2 lines
1397 /* Establish cache block size
1400 r14 contain block size
1402 l.mfspr r21,r0,SPR_ICCFGR
1403 l.andi r21,r21,SPR_ICCFGR_CBS
1408 l.mtspr r0,r5,SPR_ICBIR
1410 l.mtspr r0,r5,SPR_ICBIR
1417 * DSCR: prints a string referenced by r3.
1419 * PRMS: r3 - address of the first character of null
1420 * terminated string to be printed
1422 * PREQ: UART at UART_BASE_ADD has to be initialized
1424 * POST: caller should be aware that r3, r9 are changed
1426 ENTRY(_emergency_print)
1427 EMERGENCY_PRINT_STORE_GPR4
1428 EMERGENCY_PRINT_STORE_GPR5
1429 EMERGENCY_PRINT_STORE_GPR6
1430 EMERGENCY_PRINT_STORE_GPR7
1438 l.movhi r4,hi(UART_BASE_ADD)
1456 /* next character */
1461 EMERGENCY_PRINT_LOAD_GPR7
1462 EMERGENCY_PRINT_LOAD_GPR6
1463 EMERGENCY_PRINT_LOAD_GPR5
1464 EMERGENCY_PRINT_LOAD_GPR4
1468 ENTRY(_emergency_print_nr)
1469 EMERGENCY_PRINT_STORE_GPR4
1470 EMERGENCY_PRINT_STORE_GPR5
1471 EMERGENCY_PRINT_STORE_GPR6
1472 EMERGENCY_PRINT_STORE_GPR7
1473 EMERGENCY_PRINT_STORE_GPR8
1475 l.addi r8,r0,32 // shift register
1477 1: /* remove leading zeros */
1482 /* don't skip the last zero if number == 0x0 */
1506 l.movhi r4,hi(UART_BASE_ADD)
1524 /* next character */
1529 EMERGENCY_PRINT_LOAD_GPR8
1530 EMERGENCY_PRINT_LOAD_GPR7
1531 EMERGENCY_PRINT_LOAD_GPR6
1532 EMERGENCY_PRINT_LOAD_GPR5
1533 EMERGENCY_PRINT_LOAD_GPR4
1539 * This should be used for debugging only.
1540 * It messes up the Linux early serial output
1541 * somehow, so use it sparingly and essentially
1542 * only if you need to debug something that goes wrong
1543 * before Linux gets the early serial going.
1545 * Furthermore, you'll have to make sure you set the
1546 * UART_DEVISOR correctly according to the system
1554 #define SYS_CLK 20000000
1555 //#define SYS_CLK 1843200
1556 #define OR32_CONSOLE_BAUD 115200
1557 #define UART_DIVISOR SYS_CLK/(16*OR32_CONSOLE_BAUD)
1559 ENTRY(_early_uart_init)
1560 l.movhi r3,hi(UART_BASE_ADD)
1574 l.addi r4,r0,((UART_DIVISOR>>8) & 0x000000ff)
1575 l.sb UART_DLM(r3),r4
1576 l.addi r4,r0,((UART_DIVISOR) & 0x000000ff)
1577 l.sb UART_DLL(r3),r4
1583 _string_copying_linux:
1584 .string "\n\n\n\n\n\rCopying Linux... \0"
1587 .string "Ok, booting the kernel.\n\r\0"
1589 _string_unhandled_exception:
1590 .string "\n\rRunarunaround: Unhandled exception 0x\0"
1593 .string ": EPC=0x\0"
1598 .global _string_esr_irq_bug
1599 _string_esr_irq_bug:
1600 .string "\n\rESR external interrupt bug, for details look into entry.S\n\r\0"
1604 /* ========================================[ page aligned structures ]=== */
1607 * .data section should be page aligned
1608 * (look into arch/openrisc/kernel/vmlinux.lds.S)
1612 .global empty_zero_page
1616 .global swapper_pg_dir
1620 .global _unhandled_stack
1623 _unhandled_stack_top:
1625 /* ============================================================[ EOF ]=== */