xref: /linux/arch/powerpc/kernel/head_book3s_32.S (revision 55d0969c451159cff86949b38c39171cab962069)
1/* SPDX-License-Identifier: GPL-2.0-or-later */
2/*
3 *  PowerPC version
4 *    Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org)
5 *
6 *  Rewritten by Cort Dougan (cort@cs.nmt.edu) for PReP
7 *    Copyright (C) 1996 Cort Dougan <cort@cs.nmt.edu>
8 *  Adapted for Power Macintosh by Paul Mackerras.
9 *  Low-level exception handlers and MMU support
10 *  rewritten by Paul Mackerras.
11 *    Copyright (C) 1996 Paul Mackerras.
12 *  MPC8xx modifications Copyright (C) 1997 Dan Malek (dmalek@jlc.net).
13 *
14 *  This file contains the low-level support and setup for the
15 *  PowerPC platform, including trap and interrupt dispatch.
16 *  (The PPC 8xx embedded CPUs use head_8xx.S instead.)
17 */
18
19#include <linux/init.h>
20#include <linux/pgtable.h>
21#include <linux/linkage.h>
22
23#include <asm/reg.h>
24#include <asm/page.h>
25#include <asm/mmu.h>
26#include <asm/cputable.h>
27#include <asm/cache.h>
28#include <asm/thread_info.h>
29#include <asm/ppc_asm.h>
30#include <asm/asm-offsets.h>
31#include <asm/ptrace.h>
32#include <asm/bug.h>
33#include <asm/kvm_book3s_asm.h>
34#include <asm/feature-fixups.h>
35#include <asm/interrupt.h>
36
37#include "head_32.h"
38
39#define LOAD_BAT(n, reg, RA, RB)	\
40	/* see the comment for clear_bats() -- Cort */ \
41	li	RA,0;			\
42	mtspr	SPRN_IBAT##n##U,RA;	\
43	mtspr	SPRN_DBAT##n##U,RA;	\
44	lwz	RA,(n*16)+0(reg);	\
45	lwz	RB,(n*16)+4(reg);	\
46	mtspr	SPRN_IBAT##n##U,RA;	\
47	mtspr	SPRN_IBAT##n##L,RB;	\
48	lwz	RA,(n*16)+8(reg);	\
49	lwz	RB,(n*16)+12(reg);	\
50	mtspr	SPRN_DBAT##n##U,RA;	\
51	mtspr	SPRN_DBAT##n##L,RB
52
53	__HEAD
54_GLOBAL(_stext);
55
56/*
57 * _start is defined this way because the XCOFF loader in the OpenFirmware
58 * on the powermac expects the entry point to be a procedure descriptor.
59 */
60_GLOBAL(_start);
61	/*
62	 * These are here for legacy reasons, the kernel used to
63	 * need to look like a coff function entry for the pmac
64	 * but we're always started by some kind of bootloader now.
65	 *  -- Cort
66	 */
67	nop	/* used by __secondary_hold on prep (mtx) and chrp smp */
68	nop	/* used by __secondary_hold on prep (mtx) and chrp smp */
69	nop
70
71/* PMAC
72 * Enter here with the kernel text, data and bss loaded starting at
73 * 0, running with virtual == physical mapping.
74 * r5 points to the prom entry point (the client interface handler
75 * address).  Address translation is turned on, with the prom
76 * managing the hash table.  Interrupts are disabled.  The stack
77 * pointer (r1) points to just below the end of the half-meg region
78 * from 0x380000 - 0x400000, which is mapped in already.
79 *
80 * If we are booted from MacOS via BootX, we enter with the kernel
81 * image loaded somewhere, and the following values in registers:
82 *  r3: 'BooX' (0x426f6f58)
83 *  r4: virtual address of boot_infos_t
84 *  r5: 0
85 *
86 * PREP
87 * This is jumped to on prep systems right after the kernel is relocated
88 * to its proper place in memory by the boot loader.  The expected layout
89 * of the regs is:
90 *   r3: ptr to residual data
91 *   r4: initrd_start or if no initrd then 0
92 *   r5: initrd_end - unused if r4 is 0
93 *   r6: Start of command line string
94 *   r7: End of command line string
95 *
96 * This just gets a minimal mmu environment setup so we can call
97 * start_here() to do the real work.
98 * -- Cort
99 */
100
101	.globl	__start
102__start:
103/*
104 * We have to do any OF calls before we map ourselves to KERNELBASE,
105 * because OF may have I/O devices mapped into that area
106 * (particularly on CHRP).
107 */
108	cmpwi	0,r5,0
109	beq	1f
110
111#ifdef CONFIG_PPC_OF_BOOT_TRAMPOLINE
112	/* find out where we are now */
113	bcl	20,31,$+4
1140:	mflr	r8			/* r8 = runtime addr here */
115	addis	r8,r8,(_stext - 0b)@ha
116	addi	r8,r8,(_stext - 0b)@l	/* current runtime base addr */
117	bl	prom_init
118#endif /* CONFIG_PPC_OF_BOOT_TRAMPOLINE */
119
120	/* We never return. We also hit that trap if trying to boot
121	 * from OF while CONFIG_PPC_OF_BOOT_TRAMPOLINE isn't selected */
122	trap
123
124/*
125 * Check for BootX signature when supporting PowerMac and branch to
126 * appropriate trampoline if it's present
127 */
128#ifdef CONFIG_PPC_PMAC
1291:	lis	r31,0x426f
130	ori	r31,r31,0x6f58
131	cmpw	0,r3,r31
132	bne	1f
133	bl	bootx_init
134	trap
135#endif /* CONFIG_PPC_PMAC */
136
1371:	mr	r31,r3			/* save device tree ptr */
138	li	r24,0			/* cpu # */
139
140/*
141 * early_init() does the early machine identification and does
142 * the necessary low-level setup and clears the BSS
143 *  -- Cort <cort@fsmlabs.com>
144 */
145	bl	early_init
146
147/* Switch MMU off, clear BATs and flush TLB. At this point, r3 contains
148 * the physical address we are running at, returned by early_init()
149 */
150 	bl	mmu_off
151__after_mmu_off:
152	bl	clear_bats
153	bl	flush_tlbs
154
155	bl	initial_bats
156	bl	load_segment_registers
157	bl	reloc_offset
158	bl	early_hash_table
159#if defined(CONFIG_BOOTX_TEXT)
160	bl	setup_disp_bat
161#endif
162#ifdef CONFIG_PPC_EARLY_DEBUG_CPM
163	bl	setup_cpm_bat
164#endif
165#ifdef CONFIG_PPC_EARLY_DEBUG_USBGECKO
166	bl	setup_usbgecko_bat
167#endif
168
169/*
170 * Call setup_cpu for CPU 0 and initialize 6xx Idle
171 */
172	bl	reloc_offset
173	li	r24,0			/* cpu# */
174	bl	call_setup_cpu		/* Call setup_cpu for this CPU */
175	bl	reloc_offset
176	bl	init_idle_6xx
177
178
179/*
180 * We need to run with _start at physical address 0.
181 * On CHRP, we are loaded at 0x10000 since OF on CHRP uses
182 * the exception vectors at 0 (and therefore this copy
183 * overwrites OF's exception vectors with our own).
184 * The MMU is off at this point.
185 */
186	bl	reloc_offset
187	mr	r26,r3
188	addis	r4,r3,KERNELBASE@h	/* current address of _start */
189	lis	r5,PHYSICAL_START@h
190	cmplw	0,r4,r5			/* already running at PHYSICAL_START? */
191	bne	relocate_kernel
192/*
193 * we now have the 1st 16M of ram mapped with the bats.
194 * prep needs the mmu to be turned on here, but pmac already has it on.
195 * this shouldn't bother the pmac since it just gets turned on again
196 * as we jump to our code at KERNELBASE. -- Cort
197 * Actually no, pmac doesn't have it on any more. BootX enters with MMU
198 * off, and in other cases, we now turn it off before changing BATs above.
199 */
200turn_on_mmu:
201	mfmsr	r0
202	ori	r0,r0,MSR_DR|MSR_IR|MSR_RI
203	mtspr	SPRN_SRR1,r0
204	lis	r0,start_here@h
205	ori	r0,r0,start_here@l
206	mtspr	SPRN_SRR0,r0
207	rfi				/* enables MMU */
208
209/*
210 * We need __secondary_hold as a place to hold the other cpus on
211 * an SMP machine, even when we are running a UP kernel.
212 */
213	. = 0xc0			/* for prep bootloader */
214	li	r3,1			/* MTX only has 1 cpu */
215	.globl	__secondary_hold
216__secondary_hold:
217	/* tell the master we're here */
218	stw	r3,__secondary_hold_acknowledge@l(0)
219#ifdef CONFIG_SMP
220100:	lwz	r4,0(0)
221	/* wait until we're told to start */
222	cmpw	0,r4,r3
223	bne	100b
224	/* our cpu # was at addr 0 - go */
225	mr	r24,r3			/* cpu # */
226	b	__secondary_start
227#else
228	b	.
229#endif /* CONFIG_SMP */
230
231	.globl	__secondary_hold_spinloop
232__secondary_hold_spinloop:
233	.long	0
234	.globl	__secondary_hold_acknowledge
235__secondary_hold_acknowledge:
236	.long	-1
237
238/* System reset */
239/* core99 pmac starts the seconary here by changing the vector, and
240   putting it back to what it was (unknown_async_exception) when done.  */
241	EXCEPTION(INTERRUPT_SYSTEM_RESET, Reset, unknown_async_exception)
242
243/* Machine check */
244/*
245 * On CHRP, this is complicated by the fact that we could get a
246 * machine check inside RTAS, and we have no guarantee that certain
247 * critical registers will have the values we expect.  The set of
248 * registers that might have bad values includes all the GPRs
249 * and all the BATs.  We indicate that we are in RTAS by putting
250 * a non-zero value, the address of the exception frame to use,
251 * in thread.rtas_sp.  The machine check handler checks thread.rtas_sp
252 * and uses its value if it is non-zero.
253 * (Other exception handlers assume that r1 is a valid kernel stack
254 * pointer when we take an exception from supervisor mode.)
255 *	-- paulus.
256 */
257	START_EXCEPTION(INTERRUPT_MACHINE_CHECK, MachineCheck)
258	EXCEPTION_PROLOG_0
259#ifdef CONFIG_PPC_CHRP
260	mtspr	SPRN_SPRG_SCRATCH2,r1
261	mfspr	r1, SPRN_SPRG_THREAD
262	lwz	r1, RTAS_SP(r1)
263	cmpwi	cr1, r1, 0
264	bne	cr1, 7f
265	mfspr	r1, SPRN_SPRG_SCRATCH2
266#endif /* CONFIG_PPC_CHRP */
267	EXCEPTION_PROLOG_1
2687:	EXCEPTION_PROLOG_2 0x200 MachineCheck
269#ifdef CONFIG_PPC_CHRP
270	beq	cr1, 1f
271	twi	31, 0, 0
272#endif
2731:	prepare_transfer_to_handler
274	bl	machine_check_exception
275	b	interrupt_return
276
277/* Data access exception. */
278	START_EXCEPTION(INTERRUPT_DATA_STORAGE, DataAccess)
279#ifdef CONFIG_PPC_BOOK3S_604
280BEGIN_MMU_FTR_SECTION
281	mtspr	SPRN_SPRG_SCRATCH2,r10
282	mfspr	r10, SPRN_SPRG_THREAD
283	stw	r11, THR11(r10)
284	mfspr	r10, SPRN_DSISR
285	mfcr	r11
286	andis.	r10, r10, (DSISR_BAD_FAULT_32S | DSISR_DABRMATCH)@h
287	mfspr	r10, SPRN_SPRG_THREAD
288	beq	hash_page_dsi
289.Lhash_page_dsi_cont:
290	mtcr	r11
291	lwz	r11, THR11(r10)
292	mfspr	r10, SPRN_SPRG_SCRATCH2
293MMU_FTR_SECTION_ELSE
294	b	1f
295ALT_MMU_FTR_SECTION_END_IFSET(MMU_FTR_HPTE_TABLE)
296#endif
2971:	EXCEPTION_PROLOG_0 handle_dar_dsisr=1
298	EXCEPTION_PROLOG_1
299	EXCEPTION_PROLOG_2 INTERRUPT_DATA_STORAGE DataAccess handle_dar_dsisr=1
300	prepare_transfer_to_handler
301	lwz	r5, _DSISR(r1)
302	andis.	r0, r5, DSISR_DABRMATCH@h
303	bne-	1f
304	bl	do_page_fault
305	b	interrupt_return
3061:	bl	do_break
307	REST_NVGPRS(r1)
308	b	interrupt_return
309
310
311/* Instruction access exception. */
312	START_EXCEPTION(INTERRUPT_INST_STORAGE, InstructionAccess)
313	mtspr	SPRN_SPRG_SCRATCH0,r10
314	mtspr	SPRN_SPRG_SCRATCH1,r11
315	mfspr	r10, SPRN_SPRG_THREAD
316	mfspr	r11, SPRN_SRR0
317	stw	r11, SRR0(r10)
318	mfspr	r11, SPRN_SRR1		/* check whether user or kernel */
319	stw	r11, SRR1(r10)
320	mfcr	r10
321#ifdef CONFIG_PPC_BOOK3S_604
322BEGIN_MMU_FTR_SECTION
323	andis.	r11, r11, SRR1_ISI_NOPT@h	/* no pte found? */
324	bne	hash_page_isi
325.Lhash_page_isi_cont:
326	mfspr	r11, SPRN_SRR1		/* check whether user or kernel */
327END_MMU_FTR_SECTION_IFSET(MMU_FTR_HPTE_TABLE)
328#endif
329	andi.	r11, r11, MSR_PR
330
331	EXCEPTION_PROLOG_1
332	EXCEPTION_PROLOG_2 INTERRUPT_INST_STORAGE InstructionAccess
333	andis.	r5,r9,DSISR_SRR1_MATCH_32S@h /* Filter relevant SRR1 bits */
334	stw	r5, _DSISR(r11)
335	stw	r12, _DAR(r11)
336	prepare_transfer_to_handler
337	bl	do_page_fault
338	b	interrupt_return
339
340/* External interrupt */
341	EXCEPTION(INTERRUPT_EXTERNAL, HardwareInterrupt, do_IRQ)
342
343/* Alignment exception */
344	START_EXCEPTION(INTERRUPT_ALIGNMENT, Alignment)
345	EXCEPTION_PROLOG INTERRUPT_ALIGNMENT Alignment handle_dar_dsisr=1
346	prepare_transfer_to_handler
347	bl	alignment_exception
348	REST_NVGPRS(r1)
349	b	interrupt_return
350
351/* Program check exception */
352	START_EXCEPTION(INTERRUPT_PROGRAM, ProgramCheck)
353	EXCEPTION_PROLOG INTERRUPT_PROGRAM ProgramCheck
354	prepare_transfer_to_handler
355	bl	program_check_exception
356	REST_NVGPRS(r1)
357	b	interrupt_return
358
359/* Floating-point unavailable */
360	START_EXCEPTION(0x800, FPUnavailable)
361#ifdef CONFIG_PPC_FPU
362BEGIN_FTR_SECTION
363/*
364 * Certain Freescale cores don't have a FPU and treat fp instructions
365 * as a FP Unavailable exception.  Redirect to illegal/emulation handling.
366 */
367	b 	ProgramCheck
368END_FTR_SECTION_IFSET(CPU_FTR_FPU_UNAVAILABLE)
369	EXCEPTION_PROLOG INTERRUPT_FP_UNAVAIL FPUnavailable
370	beq	1f
371	bl	load_up_fpu		/* if from user, just load it up */
372	b	fast_exception_return
3731:	prepare_transfer_to_handler
374	bl	kernel_fp_unavailable_exception
375	b	interrupt_return
376#else
377	b 	ProgramCheck
378#endif
379
380/* Decrementer */
381	EXCEPTION(INTERRUPT_DECREMENTER, Decrementer, timer_interrupt)
382
383	EXCEPTION(0xa00, Trap_0a, unknown_exception)
384	EXCEPTION(0xb00, Trap_0b, unknown_exception)
385
386/* System call */
387	START_EXCEPTION(INTERRUPT_SYSCALL, SystemCall)
388	SYSCALL_ENTRY	INTERRUPT_SYSCALL
389
390	EXCEPTION(INTERRUPT_TRACE, SingleStep, single_step_exception)
391	EXCEPTION(0xe00, Trap_0e, unknown_exception)
392
393/*
394 * The Altivec unavailable trap is at 0x0f20.  Foo.
395 * We effectively remap it to 0x3000.
396 * We include an altivec unavailable exception vector even if
397 * not configured for Altivec, so that you can't panic a
398 * non-altivec kernel running on a machine with altivec just
399 * by executing an altivec instruction.
400 */
401	START_EXCEPTION(INTERRUPT_PERFMON, PerformanceMonitorTrap)
402	b	PerformanceMonitor
403
404	START_EXCEPTION(INTERRUPT_ALTIVEC_UNAVAIL, AltiVecUnavailableTrap)
405	b	AltiVecUnavailable
406
407	__HEAD
408/*
409 * Handle TLB miss for instruction on 603/603e.
410 * Note: we get an alternate set of r0 - r3 to use automatically.
411 */
412	. = INTERRUPT_INST_TLB_MISS_603
413InstructionTLBMiss:
414	/* Get PTE (linux-style) and check access */
415	mfspr	r0,SPRN_IMISS
416	mfspr	r2, SPRN_SDR1
417	li	r1,_PAGE_PRESENT | _PAGE_ACCESSED | _PAGE_EXEC
418	rlwinm	r2, r2, 28, 0xfffff000
419	rlwimi	r2,r0,12,20,29		/* insert top 10 bits of address */
420	lwz	r2,0(r2)		/* get pmd entry */
421#ifdef CONFIG_EXECMEM
422	rlwinm	r3, r0, 4, 0xf
423	subi	r3, r3, (TASK_SIZE >> 28) & 0xf
424#endif
425	rlwinm.	r2,r2,0,0,19		/* extract address of pte page */
426	beq-	InstructionAddressInvalid	/* return if no mapping */
427	rlwimi	r2,r0,22,20,29		/* insert next 10 bits of address */
428	lwz	r2,0(r2)		/* get linux-style pte */
429	andc.	r1,r1,r2		/* check access & ~permission */
430	bne-	InstructionAddressInvalid /* return if access not permitted */
431	/* Convert linux-style PTE to low word of PPC-style PTE */
432#ifdef CONFIG_EXECMEM
433	rlwimi	r2, r3, 1, 31, 31	/* userspace ? -> PP lsb */
434#endif
435	ori	r1, r1, 0xe06		/* clear out reserved bits */
436	andc	r1, r2, r1		/* PP = user? 1 : 0 */
437BEGIN_FTR_SECTION
438	rlwinm	r1,r1,0,~_PAGE_COHERENT	/* clear M (coherence not required) */
439END_FTR_SECTION_IFCLR(CPU_FTR_NEED_COHERENT)
440	mtspr	SPRN_RPA,r1
441	tlbli	r0
442	mfspr	r3,SPRN_SRR1		/* Need to restore CR0 */
443	mtcrf	0x80,r3
444	rfi
445InstructionAddressInvalid:
446	mfspr	r3,SPRN_SRR1
447	rlwinm	r1,r3,9,6,6	/* Get load/store bit */
448
449	addis	r1,r1,0x2000
450	mtspr	SPRN_DSISR,r1	/* (shouldn't be needed) */
451	andi.	r2,r3,0xFFFF	/* Clear upper bits of SRR1 */
452	or	r2,r2,r1
453	mtspr	SPRN_SRR1,r2
454	mfspr	r1,SPRN_IMISS	/* Get failing address */
455	rlwinm.	r2,r2,0,31,31	/* Check for little endian access */
456	rlwimi	r2,r2,1,30,30	/* change 1 -> 3 */
457	xor	r1,r1,r2
458	mtspr	SPRN_DAR,r1	/* Set fault address */
459	mfmsr	r0		/* Restore "normal" registers */
460	xoris	r0,r0,MSR_TGPR>>16
461	mtcrf	0x80,r3		/* Restore CR0 */
462	mtmsr	r0
463	b	InstructionAccess
464
465/*
466 * Handle TLB miss for DATA Load operation on 603/603e
467 */
468	. = INTERRUPT_DATA_LOAD_TLB_MISS_603
469DataLoadTLBMiss:
470	/* Get PTE (linux-style) and check access */
471	mfspr	r0,SPRN_DMISS
472	mfspr	r2, SPRN_SDR1
473	rlwinm	r1, r2, 28, 0xfffff000
474	rlwimi	r1,r0,12,20,29		/* insert top 10 bits of address */
475	lwz	r2,0(r1)		/* get pmd entry */
476	rlwinm	r3, r0, 4, 0xf
477	rlwinm.	r2,r2,0,0,19		/* extract address of pte page */
478	subi	r3, r3, (TASK_SIZE >> 28) & 0xf
479	beq-	2f			/* bail if no mapping */
4801:	rlwimi	r2,r0,22,20,29		/* insert next 10 bits of address */
481	lwz	r2,0(r2)		/* get linux-style pte */
482	li	r1, _PAGE_PRESENT | _PAGE_ACCESSED | _PAGE_READ
483	andc.	r1,r1,r2		/* check access & ~permission */
484	bne-	DataAddressInvalid	/* return if access not permitted */
485	/* Convert linux-style PTE to low word of PPC-style PTE */
486	rlwinm	r1,r2,32-9,30,30	/* _PAGE_WRITE -> PP msb */
487	rlwimi	r2,r3,2,30,31		/* userspace ? -> PP */
488	rlwimi	r1,r2,32-3,24,24	/* _PAGE_WRITE -> _PAGE_DIRTY */
489	xori	r1,r1,_PAGE_DIRTY	/* clear dirty when not rw */
490	ori	r1,r1,0xe04		/* clear out reserved bits */
491	andc	r1,r2,r1		/* PP = user? rw? 1: 3: 0 */
492BEGIN_FTR_SECTION
493	rlwinm	r1,r1,0,~_PAGE_COHERENT	/* clear M (coherence not required) */
494END_FTR_SECTION_IFCLR(CPU_FTR_NEED_COHERENT)
495	mtspr	SPRN_RPA,r1
496BEGIN_MMU_FTR_SECTION
497	li	r3,1
498	mfspr	r1,SPRN_SPRG_603_LRU
499	rlwinm	r2,r0,20,27,31		/* Get Address bits 15:19 */
500	slw	r3,r3,r2
501	xor	r1,r3,r1
502	srw	r3,r1,r2
503	mtspr   SPRN_SPRG_603_LRU,r1
504	mfspr	r2,SPRN_SRR1
505	rlwimi	r2,r3,31-14,14,14
506	mtspr   SPRN_SRR1,r2
507	mtcrf	0x80,r2
508	tlbld	r0
509	rfi
510MMU_FTR_SECTION_ELSE
511	mfspr	r2,SPRN_SRR1		/* Need to restore CR0 */
512	mtcrf	0x80,r2
513	tlbld	r0
514	rfi
515ALT_MMU_FTR_SECTION_END_IFSET(MMU_FTR_NEED_DTLB_SW_LRU)
516
5172:	lis     r2, (swapper_pg_dir - PAGE_OFFSET)@ha
518	addi    r2, r2, (swapper_pg_dir - PAGE_OFFSET)@l        /* kernel page table */
519	rlwimi	r2,r0,12,20,29		/* insert top 10 bits of address */
520	lwz	r2,0(r2)		/* get pmd entry */
521	cmpwi	cr0,r2,0
522	beq-	DataAddressInvalid	/* return if no mapping */
523	stw	r2,0(r1)
524	rlwinm.	r2,r2,0,0,19		/* extract address of pte page */
525	b	1b
526DataAddressInvalid:
527	mfspr	r3,SPRN_SRR1
528	rlwinm	r1,r3,9,6,6	/* Get load/store bit */
529	addis	r1,r1,0x2000
530	mtspr	SPRN_DSISR,r1
531	andi.	r2,r3,0xFFFF	/* Clear upper bits of SRR1 */
532	mtspr	SPRN_SRR1,r2
533	mfspr	r1,SPRN_DMISS	/* Get failing address */
534	rlwinm.	r2,r2,0,31,31	/* Check for little endian access */
535	beq	20f		/* Jump if big endian */
536	xori	r1,r1,3
53720:	mtspr	SPRN_DAR,r1	/* Set fault address */
538	mfmsr	r0		/* Restore "normal" registers */
539	xoris	r0,r0,MSR_TGPR>>16
540	mtcrf	0x80,r3		/* Restore CR0 */
541	mtmsr	r0
542	b	DataAccess
543
544/*
545 * Handle TLB miss for DATA Store on 603/603e
546 */
547	. = INTERRUPT_DATA_STORE_TLB_MISS_603
548DataStoreTLBMiss:
549	/* Get PTE (linux-style) and check access */
550	mfspr	r0,SPRN_DMISS
551	mfspr	r2, SPRN_SDR1
552	rlwinm	r1, r2, 28, 0xfffff000
553	rlwimi	r1,r0,12,20,29		/* insert top 10 bits of address */
554	lwz	r2,0(r1)		/* get pmd entry */
555	rlwinm	r3, r0, 4, 0xf
556	rlwinm.	r2,r2,0,0,19		/* extract address of pte page */
557	subi	r3, r3, (TASK_SIZE >> 28) & 0xf
558	beq-	2f			/* bail if no mapping */
5591:
560	rlwimi	r2,r0,22,20,29		/* insert next 10 bits of address */
561	lwz	r2,0(r2)		/* get linux-style pte */
562	li	r1, _PAGE_RW | _PAGE_DIRTY | _PAGE_PRESENT | _PAGE_ACCESSED
563	andc.	r1,r1,r2		/* check access & ~permission */
564	bne-	DataAddressInvalid	/* return if access not permitted */
565	/* Convert linux-style PTE to low word of PPC-style PTE */
566	rlwimi	r2,r3,1,31,31		/* userspace ? -> PP lsb */
567	li	r1,0xe06		/* clear out reserved bits & PP msb */
568	andc	r1,r2,r1		/* PP = user? 1: 0 */
569BEGIN_FTR_SECTION
570	rlwinm	r1,r1,0,~_PAGE_COHERENT	/* clear M (coherence not required) */
571END_FTR_SECTION_IFCLR(CPU_FTR_NEED_COHERENT)
572	mtspr	SPRN_RPA,r1
573	mfspr	r2,SPRN_SRR1		/* Need to restore CR0 */
574	mtcrf	0x80,r2
575BEGIN_MMU_FTR_SECTION
576	li	r3,1
577	mfspr	r1,SPRN_SPRG_603_LRU
578	rlwinm	r2,r0,20,27,31		/* Get Address bits 15:19 */
579	slw	r3,r3,r2
580	xor	r1,r3,r1
581	srw	r3,r1,r2
582	mtspr   SPRN_SPRG_603_LRU,r1
583	mfspr	r2,SPRN_SRR1
584	rlwimi	r2,r3,31-14,14,14
585	mtspr   SPRN_SRR1,r2
586	mtcrf	0x80,r2
587	tlbld	r0
588	rfi
589MMU_FTR_SECTION_ELSE
590	mfspr	r2,SPRN_SRR1		/* Need to restore CR0 */
591	mtcrf	0x80,r2
592	tlbld	r0
593	rfi
594ALT_MMU_FTR_SECTION_END_IFSET(MMU_FTR_NEED_DTLB_SW_LRU)
595
5962:	lis     r2, (swapper_pg_dir - PAGE_OFFSET)@ha
597	addi    r2, r2, (swapper_pg_dir - PAGE_OFFSET)@l        /* kernel page table */
598	rlwimi	r2,r0,12,20,29		/* insert top 10 bits of address */
599	lwz	r2,0(r2)		/* get pmd entry */
600	cmpwi	cr0,r2,0
601	beq-	DataAddressInvalid	/* return if no mapping */
602	stw	r2,0(r1)
603	rlwinm	r2,r2,0,0,19		/* extract address of pte page */
604	b	1b
605
606#ifndef CONFIG_ALTIVEC
607#define altivec_assist_exception	unknown_exception
608#endif
609
610#ifndef CONFIG_TAU_INT
611#define TAUException	unknown_async_exception
612#endif
613
614	EXCEPTION(0x1300, Trap_13, instruction_breakpoint_exception)
615	EXCEPTION(0x1400, SMI, SMIException)
616	EXCEPTION(0x1500, Trap_15, unknown_exception)
617	EXCEPTION(0x1600, Trap_16, altivec_assist_exception)
618	EXCEPTION(0x1700, Trap_17, TAUException)
619	EXCEPTION(0x1800, Trap_18, unknown_exception)
620	EXCEPTION(0x1900, Trap_19, unknown_exception)
621	EXCEPTION(0x1a00, Trap_1a, unknown_exception)
622	EXCEPTION(0x1b00, Trap_1b, unknown_exception)
623	EXCEPTION(0x1c00, Trap_1c, unknown_exception)
624	EXCEPTION(0x1d00, Trap_1d, unknown_exception)
625	EXCEPTION(0x1e00, Trap_1e, unknown_exception)
626	EXCEPTION(0x1f00, Trap_1f, unknown_exception)
627	EXCEPTION(0x2000, RunMode, RunModeException)
628	EXCEPTION(0x2100, Trap_21, unknown_exception)
629	EXCEPTION(0x2200, Trap_22, unknown_exception)
630	EXCEPTION(0x2300, Trap_23, unknown_exception)
631	EXCEPTION(0x2400, Trap_24, unknown_exception)
632	EXCEPTION(0x2500, Trap_25, unknown_exception)
633	EXCEPTION(0x2600, Trap_26, unknown_exception)
634	EXCEPTION(0x2700, Trap_27, unknown_exception)
635	EXCEPTION(0x2800, Trap_28, unknown_exception)
636	EXCEPTION(0x2900, Trap_29, unknown_exception)
637	EXCEPTION(0x2a00, Trap_2a, unknown_exception)
638	EXCEPTION(0x2b00, Trap_2b, unknown_exception)
639	EXCEPTION(0x2c00, Trap_2c, unknown_exception)
640	EXCEPTION(0x2d00, Trap_2d, unknown_exception)
641	EXCEPTION(0x2e00, Trap_2e, unknown_exception)
642	EXCEPTION(0x2f00, Trap_2f, unknown_exception)
643
644	__HEAD
645	. = 0x3000
646
647#ifdef CONFIG_PPC_BOOK3S_604
648.macro save_regs_thread		thread
649	stw	r0, THR0(\thread)
650	stw	r3, THR3(\thread)
651	stw	r4, THR4(\thread)
652	stw	r5, THR5(\thread)
653	stw	r6, THR6(\thread)
654	stw	r8, THR8(\thread)
655	stw	r9, THR9(\thread)
656	mflr	r0
657	stw	r0, THLR(\thread)
658	mfctr	r0
659	stw	r0, THCTR(\thread)
660.endm
661
662.macro restore_regs_thread	thread
663	lwz	r0, THLR(\thread)
664	mtlr	r0
665	lwz	r0, THCTR(\thread)
666	mtctr	r0
667	lwz	r0, THR0(\thread)
668	lwz	r3, THR3(\thread)
669	lwz	r4, THR4(\thread)
670	lwz	r5, THR5(\thread)
671	lwz	r6, THR6(\thread)
672	lwz	r8, THR8(\thread)
673	lwz	r9, THR9(\thread)
674.endm
675
676hash_page_dsi:
677	save_regs_thread	r10
678	mfdsisr	r3
679	mfdar	r4
680	mfsrr0	r5
681	mfsrr1	r9
682	rlwinm	r3, r3, 32 - 15, _PAGE_WRITE	/* DSISR_STORE -> _PAGE_WRITE */
683	ori	r3, r3, _PAGE_PRESENT | _PAGE_READ
684	bl	hash_page
685	mfspr	r10, SPRN_SPRG_THREAD
686	restore_regs_thread r10
687	b	.Lhash_page_dsi_cont
688
689hash_page_isi:
690	mr	r11, r10
691	mfspr	r10, SPRN_SPRG_THREAD
692	save_regs_thread	r10
693	li	r3, _PAGE_PRESENT | _PAGE_EXEC
694	lwz	r4, SRR0(r10)
695	lwz	r9, SRR1(r10)
696	bl	hash_page
697	mfspr	r10, SPRN_SPRG_THREAD
698	restore_regs_thread r10
699	mr	r10, r11
700	b	.Lhash_page_isi_cont
701
702	.globl fast_hash_page_return
703fast_hash_page_return:
704	andis.	r10, r9, SRR1_ISI_NOPT@h	/* Set on ISI, cleared on DSI */
705	mfspr	r10, SPRN_SPRG_THREAD
706	restore_regs_thread r10
707	bne	1f
708
709	/* DSI */
710	mtcr	r11
711	lwz	r11, THR11(r10)
712	mfspr	r10, SPRN_SPRG_SCRATCH2
713	rfi
714
7151:	/* ISI */
716	mtcr	r11
717	mfspr	r11, SPRN_SPRG_SCRATCH1
718	mfspr	r10, SPRN_SPRG_SCRATCH0
719	rfi
720#endif /* CONFIG_PPC_BOOK3S_604 */
721
722#ifdef CONFIG_VMAP_STACK
723	vmap_stack_overflow_exception
724#endif
725
726	__HEAD
727AltiVecUnavailable:
728	EXCEPTION_PROLOG 0xf20 AltiVecUnavailable
729#ifdef CONFIG_ALTIVEC
730	beq	1f
731	bl	load_up_altivec		/* if from user, just load it up */
732	b	fast_exception_return
733#endif /* CONFIG_ALTIVEC */
7341:	prepare_transfer_to_handler
735	bl	altivec_unavailable_exception
736	b	interrupt_return
737
738	__HEAD
739PerformanceMonitor:
740	EXCEPTION_PROLOG 0xf00 PerformanceMonitor
741	prepare_transfer_to_handler
742	bl	performance_monitor_exception
743	b	interrupt_return
744
745
746	__HEAD
747/*
748 * This code is jumped to from the startup code to copy
749 * the kernel image to physical address PHYSICAL_START.
750 */
751relocate_kernel:
752	lis	r3,PHYSICAL_START@h	/* Destination base address */
753	li	r6,0			/* Destination offset */
754	li	r5,0x4000		/* # bytes of memory to copy */
755	bl	copy_and_flush		/* copy the first 0x4000 bytes */
756	addi	r0,r3,4f@l		/* jump to the address of 4f */
757	mtctr	r0			/* in copy and do the rest. */
758	bctr				/* jump to the copy */
7594:	lis	r5,_end-KERNELBASE@h
760	ori	r5,r5,_end-KERNELBASE@l
761	bl	copy_and_flush		/* copy the rest */
762	b	turn_on_mmu
763
764/*
765 * Copy routine used to copy the kernel to start at physical address 0
766 * and flush and invalidate the caches as needed.
767 * r3 = dest addr, r4 = source addr, r5 = copy limit, r6 = start offset
768 * on exit, r3, r4, r5 are unchanged, r6 is updated to be >= r5.
769 */
770_GLOBAL(copy_and_flush)
771	addi	r5,r5,-4
772	addi	r6,r6,-4
7734:	li	r0,L1_CACHE_BYTES/4
774	mtctr	r0
7753:	addi	r6,r6,4			/* copy a cache line */
776	lwzx	r0,r6,r4
777	stwx	r0,r6,r3
778	bdnz	3b
779	dcbst	r6,r3			/* write it to memory */
780	sync
781	icbi	r6,r3			/* flush the icache line */
782	cmplw	0,r6,r5
783	blt	4b
784	sync				/* additional sync needed on g4 */
785	isync
786	addi	r5,r5,4
787	addi	r6,r6,4
788	blr
789
790#ifdef CONFIG_SMP
791	.globl __secondary_start_mpc86xx
792__secondary_start_mpc86xx:
793	mfspr	r3, SPRN_PIR
794	stw	r3, __secondary_hold_acknowledge@l(0)
795	mr	r24, r3			/* cpu # */
796	b	__secondary_start
797
798	.globl	__secondary_start_pmac_0
799__secondary_start_pmac_0:
800	/* NB the entries for cpus 0, 1, 2 must each occupy 8 bytes. */
801	li	r24,0
802	b	1f
803	li	r24,1
804	b	1f
805	li	r24,2
806	b	1f
807	li	r24,3
8081:
809	/* on powersurge, we come in here with IR=0 and DR=1, and DBAT 0
810	   set to map the 0xf0000000 - 0xffffffff region */
811	mfmsr	r0
812	rlwinm	r0,r0,0,28,26		/* clear DR (0x10) */
813	mtmsr	r0
814	isync
815
816	.globl	__secondary_start
817__secondary_start:
818	/* Copy some CPU settings from CPU 0 */
819	bl	__restore_cpu_setup
820
821	lis	r3,-KERNELBASE@h
822	mr	r4,r24
823	bl	call_setup_cpu		/* Call setup_cpu for this CPU */
824	lis	r3,-KERNELBASE@h
825	bl	init_idle_6xx
826
827	/* get current's stack and current */
828	lis	r2,secondary_current@ha
829	tophys(r2,r2)
830	lwz	r2,secondary_current@l(r2)
831	tophys(r1,r2)
832	lwz	r1,TASK_STACK(r1)
833
834	/* stack */
835	addi	r1,r1,THREAD_SIZE-STACK_FRAME_MIN_SIZE
836	li	r0,0
837	tophys(r3,r1)
838	stw	r0,0(r3)
839
840	/* load up the MMU */
841	bl	load_segment_registers
842	bl	load_up_mmu
843
844	/* ptr to phys current thread */
845	tophys(r4,r2)
846	addi	r4,r4,THREAD	/* phys address of our thread_struct */
847	mtspr	SPRN_SPRG_THREAD,r4
848BEGIN_MMU_FTR_SECTION
849	lis	r4, (swapper_pg_dir - PAGE_OFFSET)@h
850	ori	r4, r4, (swapper_pg_dir - PAGE_OFFSET)@l
851	rlwinm	r4, r4, 4, 0xffff01ff
852	mtspr	SPRN_SDR1, r4
853END_MMU_FTR_SECTION_IFCLR(MMU_FTR_HPTE_TABLE)
854
855	/* enable MMU and jump to start_secondary */
856	li	r4,MSR_KERNEL
857	lis	r3,start_secondary@h
858	ori	r3,r3,start_secondary@l
859	mtspr	SPRN_SRR0,r3
860	mtspr	SPRN_SRR1,r4
861	rfi
862#endif /* CONFIG_SMP */
863
864#ifdef CONFIG_KVM_BOOK3S_HANDLER
865#include "../kvm/book3s_rmhandlers.S"
866#endif
867
868/*
869 * Load stuff into the MMU.  Intended to be called with
870 * IR=0 and DR=0.
871 */
872SYM_FUNC_START_LOCAL(early_hash_table)
873	sync			/* Force all PTE updates to finish */
874	isync
875	tlbia			/* Clear all TLB entries */
876	sync			/* wait for tlbia/tlbie to finish */
877	TLBSYNC			/* ... on all CPUs */
878	/* Load the SDR1 register (hash table base & size) */
879	lis	r6, early_hash - PAGE_OFFSET@h
880	ori	r6, r6, 3	/* 256kB table */
881	mtspr	SPRN_SDR1, r6
882	blr
883SYM_FUNC_END(early_hash_table)
884
885SYM_FUNC_START_LOCAL(load_up_mmu)
886	sync			/* Force all PTE updates to finish */
887	isync
888	tlbia			/* Clear all TLB entries */
889	sync			/* wait for tlbia/tlbie to finish */
890	TLBSYNC			/* ... on all CPUs */
891BEGIN_MMU_FTR_SECTION
892	/* Load the SDR1 register (hash table base & size) */
893	lis	r6,_SDR1@ha
894	tophys(r6,r6)
895	lwz	r6,_SDR1@l(r6)
896	mtspr	SPRN_SDR1,r6
897END_MMU_FTR_SECTION_IFSET(MMU_FTR_HPTE_TABLE)
898
899/* Load the BAT registers with the values set up by MMU_init. */
900	lis	r3,BATS@ha
901	addi	r3,r3,BATS@l
902	tophys(r3,r3)
903	LOAD_BAT(0,r3,r4,r5)
904	LOAD_BAT(1,r3,r4,r5)
905	LOAD_BAT(2,r3,r4,r5)
906	LOAD_BAT(3,r3,r4,r5)
907BEGIN_MMU_FTR_SECTION
908	LOAD_BAT(4,r3,r4,r5)
909	LOAD_BAT(5,r3,r4,r5)
910	LOAD_BAT(6,r3,r4,r5)
911	LOAD_BAT(7,r3,r4,r5)
912END_MMU_FTR_SECTION_IFSET(MMU_FTR_USE_HIGH_BATS)
913	blr
914SYM_FUNC_END(load_up_mmu)
915
916_GLOBAL(load_segment_registers)
917	li	r0, NUM_USER_SEGMENTS /* load up user segment register values */
918	mtctr	r0		/* for context 0 */
919#ifdef CONFIG_PPC_KUEP
920	lis	r3, SR_NX@h	/* Kp = 0, Ks = 0, VSID = 0 */
921#else
922	li	r3, 0		/* Kp = 0, Ks = 0, VSID = 0 */
923#endif
924	li	r4, 0
9253:	mtsrin	r3, r4
926	addi	r3, r3, 0x111	/* increment VSID */
927	addis	r4, r4, 0x1000	/* address of next segment */
928	bdnz	3b
929	li	r0, 16 - NUM_USER_SEGMENTS /* load up kernel segment registers */
930	mtctr	r0			/* for context 0 */
931	rlwinm	r3, r3, 0, ~SR_NX	/* Nx = 0 */
932	rlwinm	r3, r3, 0, ~SR_KS	/* Ks = 0 */
933	oris	r3, r3, SR_KP@h		/* Kp = 1 */
9343:	mtsrin	r3, r4
935	addi	r3, r3, 0x111	/* increment VSID */
936	addis	r4, r4, 0x1000	/* address of next segment */
937	bdnz	3b
938	blr
939
940/*
941 * This is where the main kernel code starts.
942 */
943start_here:
944	/* ptr to current */
945	lis	r2,init_task@h
946	ori	r2,r2,init_task@l
947	/* Set up for using our exception vectors */
948	/* ptr to phys current thread */
949	tophys(r4,r2)
950	addi	r4,r4,THREAD	/* init task's THREAD */
951	mtspr	SPRN_SPRG_THREAD,r4
952BEGIN_MMU_FTR_SECTION
953	lis	r4, (swapper_pg_dir - PAGE_OFFSET)@h
954	ori	r4, r4, (swapper_pg_dir - PAGE_OFFSET)@l
955	rlwinm	r4, r4, 4, 0xffff01ff
956	mtspr	SPRN_SDR1, r4
957END_MMU_FTR_SECTION_IFCLR(MMU_FTR_HPTE_TABLE)
958
959	/* stack */
960	lis	r1,init_thread_union@ha
961	addi	r1,r1,init_thread_union@l
962	li	r0,0
963	stwu	r0,THREAD_SIZE-STACK_FRAME_MIN_SIZE(r1)
964/*
965 * Do early platform-specific initialization,
966 * and set up the MMU.
967 */
968#ifdef CONFIG_KASAN
969	bl	kasan_early_init
970#endif
971	li	r3,0
972	mr	r4,r31
973	bl	machine_init
974	bl	__save_cpu_setup
975	bl	MMU_init
976	bl	MMU_init_hw_patch
977
978/*
979 * Go back to running unmapped so we can load up new values
980 * for SDR1 (hash table pointer) and the segment registers
981 * and change to using our exception vectors.
982 */
983	lis	r4,2f@h
984	ori	r4,r4,2f@l
985	tophys(r4,r4)
986	li	r3,MSR_KERNEL & ~(MSR_IR|MSR_DR)
987
988	.align	4
989	mtspr	SPRN_SRR0,r4
990	mtspr	SPRN_SRR1,r3
991	rfi
992/* Load up the kernel context */
9932:	bl	load_up_mmu
994
995#ifdef CONFIG_BDI_SWITCH
996	/* Add helper information for the Abatron bdiGDB debugger.
997	 * We do this here because we know the mmu is disabled, and
998	 * will be enabled for real in just a few instructions.
999	 */
1000	lis	r5, abatron_pteptrs@h
1001	ori	r5, r5, abatron_pteptrs@l
1002	stw	r5, 0xf0(0)	/* This much match your Abatron config */
1003	lis	r6, swapper_pg_dir@h
1004	ori	r6, r6, swapper_pg_dir@l
1005	tophys(r5, r5)
1006	stw	r6, 0(r5)
1007#endif /* CONFIG_BDI_SWITCH */
1008
1009/* Now turn on the MMU for real! */
1010	li	r4,MSR_KERNEL
1011	lis	r3,start_kernel@h
1012	ori	r3,r3,start_kernel@l
1013	mtspr	SPRN_SRR0,r3
1014	mtspr	SPRN_SRR1,r4
1015	rfi
1016
1017/*
1018 * An undocumented "feature" of 604e requires that the v bit
1019 * be cleared before changing BAT values.
1020 *
1021 * Also, newer IBM firmware does not clear bat3 and 4 so
1022 * this makes sure it's done.
1023 *  -- Cort
1024 */
1025SYM_FUNC_START_LOCAL(clear_bats)
1026	li	r10,0
1027
1028	mtspr	SPRN_DBAT0U,r10
1029	mtspr	SPRN_DBAT0L,r10
1030	mtspr	SPRN_DBAT1U,r10
1031	mtspr	SPRN_DBAT1L,r10
1032	mtspr	SPRN_DBAT2U,r10
1033	mtspr	SPRN_DBAT2L,r10
1034	mtspr	SPRN_DBAT3U,r10
1035	mtspr	SPRN_DBAT3L,r10
1036	mtspr	SPRN_IBAT0U,r10
1037	mtspr	SPRN_IBAT0L,r10
1038	mtspr	SPRN_IBAT1U,r10
1039	mtspr	SPRN_IBAT1L,r10
1040	mtspr	SPRN_IBAT2U,r10
1041	mtspr	SPRN_IBAT2L,r10
1042	mtspr	SPRN_IBAT3U,r10
1043	mtspr	SPRN_IBAT3L,r10
1044BEGIN_MMU_FTR_SECTION
1045	/* Here's a tweak: at this point, CPU setup have
1046	 * not been called yet, so HIGH_BAT_EN may not be
1047	 * set in HID0 for the 745x processors. However, it
1048	 * seems that doesn't affect our ability to actually
1049	 * write to these SPRs.
1050	 */
1051	mtspr	SPRN_DBAT4U,r10
1052	mtspr	SPRN_DBAT4L,r10
1053	mtspr	SPRN_DBAT5U,r10
1054	mtspr	SPRN_DBAT5L,r10
1055	mtspr	SPRN_DBAT6U,r10
1056	mtspr	SPRN_DBAT6L,r10
1057	mtspr	SPRN_DBAT7U,r10
1058	mtspr	SPRN_DBAT7L,r10
1059	mtspr	SPRN_IBAT4U,r10
1060	mtspr	SPRN_IBAT4L,r10
1061	mtspr	SPRN_IBAT5U,r10
1062	mtspr	SPRN_IBAT5L,r10
1063	mtspr	SPRN_IBAT6U,r10
1064	mtspr	SPRN_IBAT6L,r10
1065	mtspr	SPRN_IBAT7U,r10
1066	mtspr	SPRN_IBAT7L,r10
1067END_MMU_FTR_SECTION_IFSET(MMU_FTR_USE_HIGH_BATS)
1068	blr
1069SYM_FUNC_END(clear_bats)
1070
1071_GLOBAL(update_bats)
1072	lis	r4, 1f@h
1073	ori	r4, r4, 1f@l
1074	tophys(r4, r4)
1075	mfmsr	r6
1076	mflr	r7
1077	li	r3, MSR_KERNEL & ~(MSR_IR | MSR_DR)
1078	rlwinm	r0, r6, 0, ~MSR_RI
1079	rlwinm	r0, r0, 0, ~MSR_EE
1080	mtmsr	r0
1081
1082	.align	4
1083	mtspr	SPRN_SRR0, r4
1084	mtspr	SPRN_SRR1, r3
1085	rfi
10861:	bl	clear_bats
1087	lis	r3, BATS@ha
1088	addi	r3, r3, BATS@l
1089	tophys(r3, r3)
1090	LOAD_BAT(0, r3, r4, r5)
1091	LOAD_BAT(1, r3, r4, r5)
1092	LOAD_BAT(2, r3, r4, r5)
1093	LOAD_BAT(3, r3, r4, r5)
1094BEGIN_MMU_FTR_SECTION
1095	LOAD_BAT(4, r3, r4, r5)
1096	LOAD_BAT(5, r3, r4, r5)
1097	LOAD_BAT(6, r3, r4, r5)
1098	LOAD_BAT(7, r3, r4, r5)
1099END_MMU_FTR_SECTION_IFSET(MMU_FTR_USE_HIGH_BATS)
1100	li	r3, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI)
1101	mtmsr	r3
1102	mtspr	SPRN_SRR0, r7
1103	mtspr	SPRN_SRR1, r6
1104	rfi
1105
1106SYM_FUNC_START_LOCAL(flush_tlbs)
1107	lis	r10, 0x40
11081:	addic.	r10, r10, -0x1000
1109	tlbie	r10
1110	bgt	1b
1111	sync
1112	blr
1113SYM_FUNC_END(flush_tlbs)
1114
1115SYM_FUNC_START_LOCAL(mmu_off)
1116 	addi	r4, r3, __after_mmu_off - _start
1117	mfmsr	r3
1118	andi.	r0,r3,MSR_DR|MSR_IR		/* MMU enabled? */
1119	beqlr
1120	andc	r3,r3,r0
1121
1122	.align	4
1123	mtspr	SPRN_SRR0,r4
1124	mtspr	SPRN_SRR1,r3
1125	sync
1126	rfi
1127SYM_FUNC_END(mmu_off)
1128
1129/* We use one BAT to map up to 256M of RAM at _PAGE_OFFSET */
1130SYM_FUNC_START_LOCAL(initial_bats)
1131	lis	r11,PAGE_OFFSET@h
1132	tophys(r8,r11)
1133#ifdef CONFIG_SMP
1134	ori	r8,r8,0x12		/* R/W access, M=1 */
1135#else
1136	ori	r8,r8,2			/* R/W access */
1137#endif /* CONFIG_SMP */
1138	ori	r11,r11,BL_256M<<2|0x2	/* set up BAT registers for 604 */
1139
1140	mtspr	SPRN_DBAT0L,r8		/* N.B. 6xx have valid */
1141	mtspr	SPRN_DBAT0U,r11		/* bit in upper BAT register */
1142	mtspr	SPRN_IBAT0L,r8
1143	mtspr	SPRN_IBAT0U,r11
1144	isync
1145	blr
1146SYM_FUNC_END(initial_bats)
1147
1148#ifdef CONFIG_BOOTX_TEXT
1149SYM_FUNC_START_LOCAL(setup_disp_bat)
1150	/*
1151	 * setup the display bat prepared for us in prom.c
1152	 */
1153	mflr	r8
1154	bl	reloc_offset
1155	mtlr	r8
1156	addis	r8,r3,disp_BAT@ha
1157	addi	r8,r8,disp_BAT@l
1158	cmpwi	cr0,r8,0
1159	beqlr
1160	lwz	r11,0(r8)
1161	lwz	r8,4(r8)
1162	mtspr	SPRN_DBAT3L,r8
1163	mtspr	SPRN_DBAT3U,r11
1164	blr
1165SYM_FUNC_END(setup_disp_bat)
1166#endif /* CONFIG_BOOTX_TEXT */
1167
1168#ifdef CONFIG_PPC_EARLY_DEBUG_CPM
1169SYM_FUNC_START_LOCAL(setup_cpm_bat)
1170	lis	r8, 0xf000
1171	ori	r8, r8,	0x002a
1172	mtspr	SPRN_DBAT1L, r8
1173
1174	lis	r11, 0xf000
1175	ori	r11, r11, (BL_1M << 2) | 2
1176	mtspr	SPRN_DBAT1U, r11
1177
1178	blr
1179SYM_FUNC_END(setup_cpm_bat)
1180#endif
1181
1182#ifdef CONFIG_PPC_EARLY_DEBUG_USBGECKO
1183SYM_FUNC_START_LOCAL(setup_usbgecko_bat)
1184	/* prepare a BAT for early io */
1185#if defined(CONFIG_GAMECUBE)
1186	lis	r8, 0x0c00
1187#elif defined(CONFIG_WII)
1188	lis	r8, 0x0d00
1189#else
1190#error Invalid platform for USB Gecko based early debugging.
1191#endif
1192	/*
1193	 * The virtual address used must match the virtual address
1194	 * associated to the fixmap entry FIX_EARLY_DEBUG_BASE.
1195	 */
1196	lis	r11, 0xfffe	/* top 128K */
1197	ori	r8, r8, 0x002a	/* uncached, guarded ,rw */
1198	ori	r11, r11, 0x2	/* 128K, Vs=1, Vp=0 */
1199	mtspr	SPRN_DBAT1L, r8
1200	mtspr	SPRN_DBAT1U, r11
1201	blr
1202SYM_FUNC_END(setup_usbgecko_bat)
1203#endif
1204
1205	.data
1206