xref: /linux/arch/mips/kernel/r4k_fpu.S (revision cea0f76a483d1270ac6f6513964e3e75193dda48)
1/*
2 * This file is subject to the terms and conditions of the GNU General Public
3 * License.  See the file "COPYING" in the main directory of this archive
4 * for more details.
5 *
6 * Copyright (C) 1996, 98, 99, 2000, 01 Ralf Baechle
7 *
8 * Multi-arch abstraction and asm macros for easier reading:
9 * Copyright (C) 1996 David S. Miller (davem@davemloft.net)
10 *
11 * Carsten Langgaard, carstenl@mips.com
12 * Copyright (C) 2000 MIPS Technologies, Inc.
13 * Copyright (C) 1999, 2001 Silicon Graphics, Inc.
14 */
15#include <asm/asm.h>
16#include <asm/asmmacro.h>
17#include <asm/errno.h>
18#include <asm/export.h>
19#include <asm/fpregdef.h>
20#include <asm/mipsregs.h>
21#include <asm/asm-offsets.h>
22#include <asm/regdef.h>
23
24/* preprocessor replaces the fp in ".set fp=64" with $30 otherwise */
25#undef fp
26
27	.macro	EX insn, reg, src
28	.set	push
29	SET_HARDFLOAT
30	.set	nomacro
31.ex\@:	\insn	\reg, \src
32	.set	pop
33	.section __ex_table,"a"
34	PTR	.ex\@, fault
35	.previous
36	.endm
37
38/*
39 * Save a thread's fp context.
40 */
41LEAF(_save_fp)
42EXPORT_SYMBOL(_save_fp)
43#if defined(CONFIG_64BIT) || defined(CONFIG_CPU_MIPSR2) || \
44    defined(CONFIG_CPU_MIPSR5) || defined(CONFIG_CPU_MIPSR6)
45	mfc0	t0, CP0_STATUS
46#endif
47	fpu_save_double a0 t0 t1		# clobbers t1
48	jr	ra
49	END(_save_fp)
50
51/*
52 * Restore a thread's fp context.
53 */
54LEAF(_restore_fp)
55#if defined(CONFIG_64BIT) || defined(CONFIG_CPU_MIPSR2) || \
56    defined(CONFIG_CPU_MIPSR5) || defined(CONFIG_CPU_MIPSR6)
57	mfc0	t0, CP0_STATUS
58#endif
59	fpu_restore_double a0 t0 t1		# clobbers t1
60	jr	ra
61	END(_restore_fp)
62
63#ifdef CONFIG_CPU_HAS_MSA
64
65/*
66 * Save a thread's MSA vector context.
67 */
68LEAF(_save_msa)
69EXPORT_SYMBOL(_save_msa)
70	msa_save_all	a0
71	jr	ra
72	END(_save_msa)
73
74/*
75 * Restore a thread's MSA vector context.
76 */
77LEAF(_restore_msa)
78	msa_restore_all	a0
79	jr	ra
80	END(_restore_msa)
81
82LEAF(_init_msa_upper)
83	msa_init_all_upper
84	jr	ra
85	END(_init_msa_upper)
86
87#endif
88
89	.set	noreorder
90
91/**
92 * _save_fp_context() - save FP context from the FPU
93 * @a0 - pointer to fpregs field of sigcontext
94 * @a1 - pointer to fpc_csr field of sigcontext
95 *
96 * Save FP context, including the 32 FP data registers and the FP
97 * control & status register, from the FPU to signal context.
98 */
99LEAF(_save_fp_context)
100	.set	push
101	SET_HARDFLOAT
102	cfc1	t1, fcr31
103	.set	pop
104
105#if defined(CONFIG_64BIT) || defined(CONFIG_CPU_MIPSR2) || \
106    defined(CONFIG_CPU_MIPSR5) || defined(CONFIG_CPU_MIPSR6)
107	.set	push
108	SET_HARDFLOAT
109#if defined(CONFIG_CPU_MIPSR2) || defined(CONFIG_CPU_MIPSR5)
110	.set	mips32r2
111	.set	fp=64
112	mfc0	t0, CP0_STATUS
113	sll	t0, t0, 5
114	bgez	t0, 1f			# skip storing odd if FR=0
115	 nop
116#endif
117	/* Store the 16 odd double precision registers */
118	EX	sdc1 $f1, 8(a0)
119	EX	sdc1 $f3, 24(a0)
120	EX	sdc1 $f5, 40(a0)
121	EX	sdc1 $f7, 56(a0)
122	EX	sdc1 $f9, 72(a0)
123	EX	sdc1 $f11, 88(a0)
124	EX	sdc1 $f13, 104(a0)
125	EX	sdc1 $f15, 120(a0)
126	EX	sdc1 $f17, 136(a0)
127	EX	sdc1 $f19, 152(a0)
128	EX	sdc1 $f21, 168(a0)
129	EX	sdc1 $f23, 184(a0)
130	EX	sdc1 $f25, 200(a0)
131	EX	sdc1 $f27, 216(a0)
132	EX	sdc1 $f29, 232(a0)
133	EX	sdc1 $f31, 248(a0)
1341:	.set	pop
135#endif
136
137	.set push
138	SET_HARDFLOAT
139	/* Store the 16 even double precision registers */
140	EX	sdc1 $f0, 0(a0)
141	EX	sdc1 $f2, 16(a0)
142	EX	sdc1 $f4, 32(a0)
143	EX	sdc1 $f6, 48(a0)
144	EX	sdc1 $f8, 64(a0)
145	EX	sdc1 $f10, 80(a0)
146	EX	sdc1 $f12, 96(a0)
147	EX	sdc1 $f14, 112(a0)
148	EX	sdc1 $f16, 128(a0)
149	EX	sdc1 $f18, 144(a0)
150	EX	sdc1 $f20, 160(a0)
151	EX	sdc1 $f22, 176(a0)
152	EX	sdc1 $f24, 192(a0)
153	EX	sdc1 $f26, 208(a0)
154	EX	sdc1 $f28, 224(a0)
155	EX	sdc1 $f30, 240(a0)
156	EX	sw t1, 0(a1)
157	jr	ra
158	 li	v0, 0					# success
159	.set pop
160	END(_save_fp_context)
161
162/**
163 * _restore_fp_context() - restore FP context to the FPU
164 * @a0 - pointer to fpregs field of sigcontext
165 * @a1 - pointer to fpc_csr field of sigcontext
166 *
167 * Restore FP context, including the 32 FP data registers and the FP
168 * control & status register, from signal context to the FPU.
169 */
170LEAF(_restore_fp_context)
171	EX	lw t1, 0(a1)
172
173#if defined(CONFIG_64BIT) || defined(CONFIG_CPU_MIPSR2) || \
174    defined(CONFIG_CPU_MIPSR5) || defined(CONFIG_CPU_MIPSR6)
175	.set	push
176	SET_HARDFLOAT
177#if defined(CONFIG_CPU_MIPSR2) || defined(CONFIG_CPU_MIPSR5)
178	.set	mips32r2
179	.set	fp=64
180	mfc0	t0, CP0_STATUS
181	sll	t0, t0, 5
182	bgez	t0, 1f			# skip loading odd if FR=0
183	 nop
184#endif
185	EX	ldc1 $f1, 8(a0)
186	EX	ldc1 $f3, 24(a0)
187	EX	ldc1 $f5, 40(a0)
188	EX	ldc1 $f7, 56(a0)
189	EX	ldc1 $f9, 72(a0)
190	EX	ldc1 $f11, 88(a0)
191	EX	ldc1 $f13, 104(a0)
192	EX	ldc1 $f15, 120(a0)
193	EX	ldc1 $f17, 136(a0)
194	EX	ldc1 $f19, 152(a0)
195	EX	ldc1 $f21, 168(a0)
196	EX	ldc1 $f23, 184(a0)
197	EX	ldc1 $f25, 200(a0)
198	EX	ldc1 $f27, 216(a0)
199	EX	ldc1 $f29, 232(a0)
200	EX	ldc1 $f31, 248(a0)
2011:	.set pop
202#endif
203	.set push
204	SET_HARDFLOAT
205	EX	ldc1 $f0, 0(a0)
206	EX	ldc1 $f2, 16(a0)
207	EX	ldc1 $f4, 32(a0)
208	EX	ldc1 $f6, 48(a0)
209	EX	ldc1 $f8, 64(a0)
210	EX	ldc1 $f10, 80(a0)
211	EX	ldc1 $f12, 96(a0)
212	EX	ldc1 $f14, 112(a0)
213	EX	ldc1 $f16, 128(a0)
214	EX	ldc1 $f18, 144(a0)
215	EX	ldc1 $f20, 160(a0)
216	EX	ldc1 $f22, 176(a0)
217	EX	ldc1 $f24, 192(a0)
218	EX	ldc1 $f26, 208(a0)
219	EX	ldc1 $f28, 224(a0)
220	EX	ldc1 $f30, 240(a0)
221	ctc1	t1, fcr31
222	.set pop
223	jr	ra
224	 li	v0, 0					# success
225	END(_restore_fp_context)
226
227#ifdef CONFIG_CPU_HAS_MSA
228
229	.macro	op_one_wr	op, idx, base
230	.align	4
231\idx:	\op	\idx, 0, \base
232	jr	ra
233	 nop
234	.endm
235
236	.macro	op_msa_wr	name, op
237LEAF(\name)
238	.set		push
239	.set		noreorder
240	sll		t0, a0, 4
241	PTR_LA		t1, 0f
242	PTR_ADDU	t0, t0, t1
243	jr		t0
244	  nop
245	op_one_wr	\op, 0, a1
246	op_one_wr	\op, 1, a1
247	op_one_wr	\op, 2, a1
248	op_one_wr	\op, 3, a1
249	op_one_wr	\op, 4, a1
250	op_one_wr	\op, 5, a1
251	op_one_wr	\op, 6, a1
252	op_one_wr	\op, 7, a1
253	op_one_wr	\op, 8, a1
254	op_one_wr	\op, 9, a1
255	op_one_wr	\op, 10, a1
256	op_one_wr	\op, 11, a1
257	op_one_wr	\op, 12, a1
258	op_one_wr	\op, 13, a1
259	op_one_wr	\op, 14, a1
260	op_one_wr	\op, 15, a1
261	op_one_wr	\op, 16, a1
262	op_one_wr	\op, 17, a1
263	op_one_wr	\op, 18, a1
264	op_one_wr	\op, 19, a1
265	op_one_wr	\op, 20, a1
266	op_one_wr	\op, 21, a1
267	op_one_wr	\op, 22, a1
268	op_one_wr	\op, 23, a1
269	op_one_wr	\op, 24, a1
270	op_one_wr	\op, 25, a1
271	op_one_wr	\op, 26, a1
272	op_one_wr	\op, 27, a1
273	op_one_wr	\op, 28, a1
274	op_one_wr	\op, 29, a1
275	op_one_wr	\op, 30, a1
276	op_one_wr	\op, 31, a1
277	.set		pop
278	END(\name)
279	.endm
280
281	op_msa_wr	read_msa_wr_b, st_b
282	op_msa_wr	read_msa_wr_h, st_h
283	op_msa_wr	read_msa_wr_w, st_w
284	op_msa_wr	read_msa_wr_d, st_d
285
286	op_msa_wr	write_msa_wr_b, ld_b
287	op_msa_wr	write_msa_wr_h, ld_h
288	op_msa_wr	write_msa_wr_w, ld_w
289	op_msa_wr	write_msa_wr_d, ld_d
290
291#endif /* CONFIG_CPU_HAS_MSA */
292
293#ifdef CONFIG_CPU_HAS_MSA
294
295	.macro	save_msa_upper	wr, off, base
296	.set	push
297	.set	noat
298#ifdef CONFIG_64BIT
299	copy_s_d \wr, 1
300	EX sd	$1, \off(\base)
301#elif defined(CONFIG_CPU_LITTLE_ENDIAN)
302	copy_s_w \wr, 2
303	EX sw	$1, \off(\base)
304	copy_s_w \wr, 3
305	EX sw	$1, (\off+4)(\base)
306#else /* CONFIG_CPU_BIG_ENDIAN */
307	copy_s_w \wr, 2
308	EX sw	$1, (\off+4)(\base)
309	copy_s_w \wr, 3
310	EX sw	$1, \off(\base)
311#endif
312	.set	pop
313	.endm
314
315LEAF(_save_msa_all_upper)
316	save_msa_upper	0, 0x00, a0
317	save_msa_upper	1, 0x08, a0
318	save_msa_upper	2, 0x10, a0
319	save_msa_upper	3, 0x18, a0
320	save_msa_upper	4, 0x20, a0
321	save_msa_upper	5, 0x28, a0
322	save_msa_upper	6, 0x30, a0
323	save_msa_upper	7, 0x38, a0
324	save_msa_upper	8, 0x40, a0
325	save_msa_upper	9, 0x48, a0
326	save_msa_upper	10, 0x50, a0
327	save_msa_upper	11, 0x58, a0
328	save_msa_upper	12, 0x60, a0
329	save_msa_upper	13, 0x68, a0
330	save_msa_upper	14, 0x70, a0
331	save_msa_upper	15, 0x78, a0
332	save_msa_upper	16, 0x80, a0
333	save_msa_upper	17, 0x88, a0
334	save_msa_upper	18, 0x90, a0
335	save_msa_upper	19, 0x98, a0
336	save_msa_upper	20, 0xa0, a0
337	save_msa_upper	21, 0xa8, a0
338	save_msa_upper	22, 0xb0, a0
339	save_msa_upper	23, 0xb8, a0
340	save_msa_upper	24, 0xc0, a0
341	save_msa_upper	25, 0xc8, a0
342	save_msa_upper	26, 0xd0, a0
343	save_msa_upper	27, 0xd8, a0
344	save_msa_upper	28, 0xe0, a0
345	save_msa_upper	29, 0xe8, a0
346	save_msa_upper	30, 0xf0, a0
347	save_msa_upper	31, 0xf8, a0
348	jr	ra
349	 li	v0, 0
350	END(_save_msa_all_upper)
351
352	.macro	restore_msa_upper	wr, off, base
353	.set	push
354	.set	noat
355#ifdef CONFIG_64BIT
356	EX ld	$1, \off(\base)
357	insert_d \wr, 1
358#elif defined(CONFIG_CPU_LITTLE_ENDIAN)
359	EX lw	$1, \off(\base)
360	insert_w \wr, 2
361	EX lw	$1, (\off+4)(\base)
362	insert_w \wr, 3
363#else /* CONFIG_CPU_BIG_ENDIAN */
364	EX lw	$1, (\off+4)(\base)
365	insert_w \wr, 2
366	EX lw	$1, \off(\base)
367	insert_w \wr, 3
368#endif
369	.set	pop
370	.endm
371
372LEAF(_restore_msa_all_upper)
373	restore_msa_upper	0, 0x00, a0
374	restore_msa_upper	1, 0x08, a0
375	restore_msa_upper	2, 0x10, a0
376	restore_msa_upper	3, 0x18, a0
377	restore_msa_upper	4, 0x20, a0
378	restore_msa_upper	5, 0x28, a0
379	restore_msa_upper	6, 0x30, a0
380	restore_msa_upper	7, 0x38, a0
381	restore_msa_upper	8, 0x40, a0
382	restore_msa_upper	9, 0x48, a0
383	restore_msa_upper	10, 0x50, a0
384	restore_msa_upper	11, 0x58, a0
385	restore_msa_upper	12, 0x60, a0
386	restore_msa_upper	13, 0x68, a0
387	restore_msa_upper	14, 0x70, a0
388	restore_msa_upper	15, 0x78, a0
389	restore_msa_upper	16, 0x80, a0
390	restore_msa_upper	17, 0x88, a0
391	restore_msa_upper	18, 0x90, a0
392	restore_msa_upper	19, 0x98, a0
393	restore_msa_upper	20, 0xa0, a0
394	restore_msa_upper	21, 0xa8, a0
395	restore_msa_upper	22, 0xb0, a0
396	restore_msa_upper	23, 0xb8, a0
397	restore_msa_upper	24, 0xc0, a0
398	restore_msa_upper	25, 0xc8, a0
399	restore_msa_upper	26, 0xd0, a0
400	restore_msa_upper	27, 0xd8, a0
401	restore_msa_upper	28, 0xe0, a0
402	restore_msa_upper	29, 0xe8, a0
403	restore_msa_upper	30, 0xf0, a0
404	restore_msa_upper	31, 0xf8, a0
405	jr	ra
406	 li	v0, 0
407	END(_restore_msa_all_upper)
408
409#endif /* CONFIG_CPU_HAS_MSA */
410
411	.set	reorder
412
413	.type	fault, @function
414	.ent	fault
415fault:	li	v0, -EFAULT				# failure
416	jr	ra
417	.end	fault
418