xref: /titanic_41/usr/src/uts/intel/ia32/os/fpu.c (revision 672986541be54a7a471bb088e60780c37e371d7e)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*	Copyright (c) 1990, 1991 UNIX System Laboratories, Inc. */
27 /*	Copyright (c) 1984, 1986, 1987, 1988, 1989, 1990 AT&T   */
28 /*		All Rights Reserved				*/
29 
30 /*	Copyright (c) 1987, 1988 Microsoft Corporation		*/
31 /*		All Rights Reserved				*/
32 
33 #pragma ident	"%Z%%M%	%I%	%E% SMI"
34 
35 #include <sys/types.h>
36 #include <sys/param.h>
37 #include <sys/signal.h>
38 #include <sys/regset.h>
39 #include <sys/privregs.h>
40 #include <sys/psw.h>
41 #include <sys/trap.h>
42 #include <sys/fault.h>
43 #include <sys/systm.h>
44 #include <sys/user.h>
45 #include <sys/file.h>
46 #include <sys/proc.h>
47 #include <sys/pcb.h>
48 #include <sys/lwp.h>
49 #include <sys/cpuvar.h>
50 #include <sys/thread.h>
51 #include <sys/disp.h>
52 #include <sys/fp.h>
53 #include <sys/siginfo.h>
54 #include <sys/archsystm.h>
55 #include <sys/kmem.h>
56 #include <sys/debug.h>
57 #include <sys/x86_archext.h>
58 #include <sys/sysmacros.h>
59 
60 /*CSTYLED*/
61 #pragma	align 16 (sse_initial)
62 
63 /*
64  * Initial kfpu state for SSE/SSE2 used by fpinit()
65  */
66 const struct fxsave_state sse_initial = {
67 	FPU_CW_INIT,	/* fx_fcw */
68 	0,		/* fx_fsw */
69 	0,		/* fx_fctw */
70 	0,		/* fx_fop */
71 #if defined(__amd64)
72 	0,		/* fx_rip */
73 	0,		/* fx_rdp */
74 #else
75 	0,		/* fx_eip */
76 	0,		/* fx_cs */
77 	0,		/* __fx_ign0 */
78 	0,		/* fx_dp */
79 	0,		/* fx_ds */
80 	0,		/* __fx_ign1 */
81 #endif /* __amd64 */
82 	SSE_MXCSR_INIT	/* fx_mxcsr */
83 	/* rest of structure is zero */
84 };
85 
86 /*
87  * mxcsr_mask value (possibly reset in fpu_probe); used to avoid
88  * the #gp exception caused by setting unsupported bits in the
89  * MXCSR register
90  */
91 uint32_t sse_mxcsr_mask = SSE_MXCSR_MASK_DEFAULT;
92 
93 /*
94  * Initial kfpu state for x87 used by fpinit()
95  */
96 const struct fnsave_state x87_initial = {
97 	FPU_CW_INIT,	/* f_fcw */
98 	0,		/* __f_ign0 */
99 	0,		/* f_fsw */
100 	0,		/* __f_ign1 */
101 	0xffff,		/* f_ftw */
102 	/* rest of structure is zero */
103 };
104 
105 #if defined(__amd64)
106 #define	fpsave_ctxt	fpxsave_ctxt
107 #elif defined(__i386)
108 /*
109  * This vector is patched to fpxsave_ctxt() if we discover
110  * we have an SSE-capable chip in fpu_probe().
111  */
112 void (*fpsave_ctxt)(void *) = fpnsave_ctxt;
113 #endif
114 
115 static int fpe_sicode(uint_t);
116 static int fpe_simd_sicode(uint_t);
117 
118 /*
119  * Copy the state of parent lwp's floating point context into the new lwp.
120  * Invoked for both fork() and lwp_create().
121  *
122  * Note that we inherit -only- the control state (e.g. exception masks,
123  * rounding, precision control, etc.); the FPU registers are otherwise
124  * reset to their initial state.
125  */
126 static void
127 fp_new_lwp(kthread_id_t t, kthread_id_t ct)
128 {
129 	struct fpu_ctx *fp;		/* parent fpu context */
130 	struct fpu_ctx *cfp;		/* new fpu context */
131 	struct fxsave_state *fx, *cfx;
132 
133 	ASSERT(fp_kind != FP_NO);
134 
135 	fp = &t->t_lwp->lwp_pcb.pcb_fpu;
136 	cfp = &ct->t_lwp->lwp_pcb.pcb_fpu;
137 
138 	/*
139 	 * If the parent FPU state is still in the FPU hw then save it;
140 	 * conveniently, fp_save() already does this for us nicely.
141 	 */
142 	fp_save(fp);
143 
144 	cfp->fpu_flags = FPU_EN | FPU_VALID;
145 	cfp->fpu_regs.kfpu_status = 0;
146 	cfp->fpu_regs.kfpu_xstatus = 0;
147 
148 #if defined(__amd64)
149 	fx = &fp->fpu_regs.kfpu_u.kfpu_fx;
150 	cfx = &cfp->fpu_regs.kfpu_u.kfpu_fx;
151 	bcopy(&sse_initial, cfx, sizeof (*cfx));
152 	cfx->fx_mxcsr = fx->fx_mxcsr & ~SSE_MXCSR_EFLAGS;
153 	cfx->fx_fcw = fx->fx_fcw;
154 #else
155 	if (fp_kind == __FP_SSE) {
156 		fx = &fp->fpu_regs.kfpu_u.kfpu_fx;
157 		cfx = &cfp->fpu_regs.kfpu_u.kfpu_fx;
158 		bcopy(&sse_initial, cfx, sizeof (*cfx));
159 		cfx->fx_mxcsr = fx->fx_mxcsr & ~SSE_MXCSR_EFLAGS;
160 		cfx->fx_fcw = fx->fx_fcw;
161 	} else {
162 		struct fnsave_state *fn = &fp->fpu_regs.kfpu_u.kfpu_fn;
163 		struct fnsave_state *cfn = &cfp->fpu_regs.kfpu_u.kfpu_fn;
164 
165 		bcopy(&x87_initial, cfn, sizeof (*cfn));
166 		cfn->f_fcw = fn->f_fcw;
167 	}
168 #endif
169 	installctx(ct, cfp,
170 	    fpsave_ctxt, NULL, fp_new_lwp, fp_new_lwp, NULL, fp_free);
171 	/*
172 	 * Now, when the new lwp starts running, it will take a trap
173 	 * that will be handled inline in the trap table to cause
174 	 * the appropriate f*rstor instruction to load the save area we
175 	 * constructed above directly into the hardware.
176 	 */
177 }
178 
179 /*
180  * Free any state associated with floating point context.
181  * Fp_free can be called in three cases:
182  * 1) from reaper -> thread_free -> ctxfree -> fp_free
183  *	fp context belongs to a thread on deathrow
184  *	nothing to do,  thread will never be resumed
185  *	thread calling ctxfree is reaper
186  *
187  * 2) from exec -> ctxfree -> fp_free
188  *	fp context belongs to the current thread
189  *	must disable fpu, thread calling ctxfree is curthread
190  *
191  * 3) from restorecontext -> setfpregs -> fp_free
192  *	we have a modified context in the memory (lwp->pcb_fpu)
193  *	disable fpu and release the fp context for the CPU
194  *
195  */
196 /*ARGSUSED*/
197 void
198 fp_free(struct fpu_ctx *fp, int isexec)
199 {
200 	ASSERT(fp_kind != FP_NO);
201 
202 	if (fp->fpu_flags & FPU_VALID)
203 		return;
204 
205 	kpreempt_disable();
206 	/*
207 	 * We want to do fpsave rather than fpdisable so that we can
208 	 * keep the fpu_flags as FPU_VALID tracking the CR0_TS bit
209 	 */
210 	fp->fpu_flags |= FPU_VALID;
211 	/* If for current thread disable FP to track FPU_VALID */
212 	if (curthread->t_lwp && fp == &curthread->t_lwp->lwp_pcb.pcb_fpu) {
213 		/* Clear errors if any to prevent frstor from complaining */
214 		(void) fperr_reset();
215 		if (fp_kind == __FP_SSE)
216 			(void) fpxerr_reset();
217 		fpdisable();
218 	}
219 	kpreempt_enable();
220 }
221 
222 /*
223  * Store the floating point state and disable the floating point unit.
224  */
225 void
226 fp_save(struct fpu_ctx *fp)
227 {
228 	ASSERT(fp_kind != FP_NO);
229 
230 	kpreempt_disable();
231 	if (!fp || fp->fpu_flags & FPU_VALID) {
232 		kpreempt_enable();
233 		return;
234 	}
235 	ASSERT(curthread->t_lwp && fp == &curthread->t_lwp->lwp_pcb.pcb_fpu);
236 
237 #if defined(__amd64)
238 	fpxsave(&fp->fpu_regs.kfpu_u.kfpu_fx);
239 #else
240 	switch (fp_kind) {
241 	case __FP_SSE:
242 		fpxsave(&fp->fpu_regs.kfpu_u.kfpu_fx);
243 		break;
244 	default:
245 		fpsave(&fp->fpu_regs.kfpu_u.kfpu_fn);
246 		break;
247 	}
248 #endif
249 	fp->fpu_flags |= FPU_VALID;
250 	kpreempt_enable();
251 }
252 
253 /*
254  * Restore the FPU context for the thread:
255  * The possibilities are:
256  *	1. No active FPU context: Load the new context into the FPU hw
257  *	   and enable the FPU.
258  */
259 void
260 fp_restore(struct fpu_ctx *fp)
261 {
262 #if defined(__amd64)
263 	fpxrestore(&fp->fpu_regs.kfpu_u.kfpu_fx);
264 #else
265 	/* case 2 */
266 	if (fp_kind == __FP_SSE)
267 		fpxrestore(&fp->fpu_regs.kfpu_u.kfpu_fx);
268 	else
269 		fprestore(&fp->fpu_regs.kfpu_u.kfpu_fn);
270 #endif
271 	fp->fpu_flags &= ~FPU_VALID;
272 }
273 
274 
275 /*
276  * Seeds the initial state for the current thread.  The possibilities are:
277  *      1. Another process has modified the FPU state before we have done any
278  *         initialization: Load the FPU state from the LWP state.
279  *      2. The FPU state has not been externally modified:  Load a clean state.
280  */
281 static void
282 fp_seed(void)
283 {
284 	struct fpu_ctx *fp = &ttolwp(curthread)->lwp_pcb.pcb_fpu;
285 
286 	ASSERT(curthread->t_preempt >= 1);
287 	ASSERT((fp->fpu_flags & FPU_EN) == 0);
288 
289 	/*
290 	 * Always initialize a new context and initialize the hardware.
291 	 */
292 	installctx(curthread, fp,
293 	    fpsave_ctxt, NULL, fp_new_lwp, fp_new_lwp, NULL, fp_free);
294 	fpinit();
295 
296 	/*
297 	 * If FPU_VALID is set, it means someone has modified registers via
298 	 * /proc.  In this case, restore the current lwp's state.
299 	 */
300 	if (fp->fpu_flags & FPU_VALID)
301 		fp_restore(fp);
302 
303 	ASSERT((fp->fpu_flags & FPU_VALID) == 0);
304 	fp->fpu_flags = FPU_EN;
305 }
306 
307 /*
308  * This routine is called from trap() when User thread takes No Extension
309  * Fault. The possiblities are:
310  *	1. User thread has executed a FP instruction for the first time.
311  *	   Save current FPU context if any. Initialize FPU, setup FPU
312  *	   context for the thread and enable FP hw.
313  *	2. Thread's pcb has a valid FPU state: Restore the FPU state and
314  *	   enable FP hw.
315  *
316  * Note that case #2 is inlined in the trap table.
317  */
318 int
319 fpnoextflt(struct regs *rp)
320 {
321 	struct fpu_ctx *fp = &ttolwp(curthread)->lwp_pcb.pcb_fpu;
322 
323 #if !defined(__lint)
324 	ASSERT(sizeof (struct fxsave_state) == 512 &&
325 	    sizeof (struct fnsave_state) == 108);
326 	ASSERT((offsetof(struct fxsave_state, fx_xmm[0]) & 0xf) == 0);
327 #if defined(__i386)
328 	ASSERT(sizeof (struct fpu) == sizeof (struct __old_fpu));
329 #endif	/* __i386 */
330 #endif	/* !__lint */
331 
332 	/*
333 	 * save area MUST be 16-byte aligned, else will page fault
334 	 */
335 	ASSERT(((uintptr_t)(&fp->fpu_regs.kfpu_u.kfpu_fx) & 0xf) == 0);
336 
337 	kpreempt_disable();
338 	/*
339 	 * Now we can enable the interrupts.
340 	 * (NOTE: fp-no-coprocessor comes thru interrupt gate)
341 	 */
342 	sti();
343 
344 	if (!fpu_exists) { /* check for FPU hw exists */
345 		if (fp_kind == FP_NO) {
346 			uint32_t inst;
347 
348 			/*
349 			 * When the system has no floating point support,
350 			 * i.e. no FP hardware and no emulator, skip the
351 			 * two kinds of FP instruction that occur in
352 			 * fpstart.  Allows processes that do no real FP
353 			 * to run normally.
354 			 */
355 			if (fuword32((void *)rp->r_pc, &inst) != -1 &&
356 			    ((inst & 0xFFFF) == 0x7dd9 ||
357 			    (inst & 0xFFFF) == 0x6dd9)) {
358 				rp->r_pc += 3;
359 				kpreempt_enable();
360 				return (0);
361 			}
362 		}
363 
364 		/*
365 		 * If we have neither a processor extension nor
366 		 * an emulator, kill the process OR panic the kernel.
367 		 */
368 		kpreempt_enable();
369 		return (1); /* error */
370 	}
371 
372 	/*
373 	 * A paranoid cross-check: for the SSE case, ensure that %cr4 is
374 	 * configured to enable fully fledged (%xmm) fxsave/fxrestor on
375 	 * this CPU.  For the non-SSE case, ensure that it isn't.
376 	 */
377 	ASSERT((fp_kind == __FP_SSE && (getcr4() & CR4_OSFXSR) == CR4_OSFXSR) ||
378 	    (fp_kind != __FP_SSE &&
379 	    (getcr4() & (CR4_OSXMMEXCPT|CR4_OSFXSR)) == 0));
380 
381 	if (fp->fpu_flags & FPU_EN) {
382 		/* case 2 */
383 		fp_restore(fp);
384 	} else {
385 		/* case 1 */
386 		fp_seed();
387 	}
388 	kpreempt_enable();
389 	return (0);
390 }
391 
392 
393 /*
394  * Handle a processor extension overrun fault
395  * Returns non zero for error.
396  *
397  * XXX	Shouldn't this just be abolished given that we're not supporting
398  *	anything prior to Pentium?
399  */
400 
401 /* ARGSUSED */
402 int
403 fpextovrflt(struct regs *rp)
404 {
405 	ulong_t cur_cr0;
406 
407 	ASSERT(fp_kind != FP_NO);
408 
409 	cur_cr0 = getcr0();
410 	fpinit();		/* initialize the FPU hardware */
411 	setcr0(cur_cr0);
412 	sti();
413 	return (1); 		/* error, send SIGSEGV signal to the thread */
414 }
415 
416 /*
417  * Handle a processor extension error fault
418  * Returns non zero for error.
419  */
420 
421 /*ARGSUSED*/
422 int
423 fpexterrflt(struct regs *rp)
424 {
425 	uint32_t fpcwsw;
426 	fpu_ctx_t *fp = &ttolwp(curthread)->lwp_pcb.pcb_fpu;
427 
428 	ASSERT(fp_kind != FP_NO);
429 
430 	fpcwsw = fpgetcwsw();
431 	/*
432 	 * Now we can enable the interrupts.
433 	 * (NOTE: x87 fp exceptions come thru interrupt gate)
434 	 */
435 	sti();
436 
437 	if ((fpcwsw & FPS_ES) == 0)
438 		return (0);	/* No exception */
439 
440 	if (fpu_exists) {
441 		fp_save(fp);
442 		/* clear exception flags in saved state, as if by fnclex */
443 #if defined(__amd64)
444 		fp->fpu_regs.kfpu_u.kfpu_fx.fx_fsw &= ~FPS_SW_EFLAGS;
445 #else
446 		switch (fp_kind) {
447 		case __FP_SSE:
448 			fp->fpu_regs.kfpu_u.kfpu_fx.fx_fsw &= ~FPS_SW_EFLAGS;
449 			break;
450 		default:
451 			fp->fpu_regs.kfpu_u.kfpu_fn.f_fsw &= ~FPS_SW_EFLAGS;
452 			break;
453 		}
454 #endif
455 	}
456 	fp->fpu_regs.kfpu_status = fpcwsw & 0xffff;
457 	/*
458 	 * "and" the exception flags with the complement of the mask
459 	 * bits to determine which exception occurred
460 	 */
461 	return (fpe_sicode(fpcwsw & ~(fpcwsw >> 16) & 0x3f));
462 }
463 
464 /*
465  * Handle an SSE/SSE2 precise exception.
466  * Returns a non-zero sicode for error.
467  */
468 /*ARGSUSED*/
469 int
470 fpsimderrflt(struct regs *rp)
471 {
472 	uint32_t mxcsr, xmask;
473 	fpu_ctx_t *fp = &ttolwp(curthread)->lwp_pcb.pcb_fpu;
474 
475 	ASSERT(fp_kind == __FP_SSE);
476 
477 	mxcsr = fpgetmxcsr();
478 	if (fpu_exists) {
479 		fp_save(fp); 		/* save the FPU state */
480 		fp->fpu_regs.kfpu_status = fp->fpu_regs.kfpu_u.kfpu_fx.fx_fsw;
481 	} else {
482 		fp->fpu_regs.kfpu_status = fpgetcwsw() & 0xffff;
483 	}
484 	fp->fpu_regs.kfpu_xstatus = mxcsr;
485 
486 	/*
487 	 * compute the mask that determines which conditions can cause
488 	 * a #xm exception, and use this to clean the status bits so that
489 	 * we can identify the true cause of this one.
490 	 */
491 	xmask = (mxcsr >> 7) & SSE_MXCSR_EFLAGS;
492 	return (fpe_simd_sicode((mxcsr & SSE_MXCSR_EFLAGS) & ~xmask));
493 }
494 
495 /*
496  * In the unlikely event that someone is relying on this subcode being
497  * FPE_FLTILL for denormalize exceptions, it can always be patched back
498  * again to restore old behaviour.
499  */
500 int fpe_fltden = FPE_FLTDEN;
501 
502 /*
503  * Map from the FPU status word to the FP exception si_code.
504  */
505 static int
506 fpe_sicode(uint_t sw)
507 {
508 	if (sw & FPS_IE)
509 		return (FPE_FLTINV);
510 	if (sw & FPS_ZE)
511 		return (FPE_FLTDIV);
512 	if (sw & FPS_DE)
513 		return (fpe_fltden);
514 	if (sw & FPS_OE)
515 		return (FPE_FLTOVF);
516 	if (sw & FPS_UE)
517 		return (FPE_FLTUND);
518 	if (sw & FPS_PE)
519 		return (FPE_FLTRES);
520 	return (FPE_FLTINV);	/* default si_code for other exceptions */
521 }
522 
523 /*
524  * Map from the SSE status word to the FP exception si_code.
525  */
526 static int
527 fpe_simd_sicode(uint_t sw)
528 {
529 	if (sw & SSE_IE)
530 		return (FPE_FLTINV);
531 	if (sw & SSE_ZE)
532 		return (FPE_FLTDIV);
533 	if (sw & SSE_DE)
534 		return (FPE_FLTDEN);
535 	if (sw & SSE_OE)
536 		return (FPE_FLTOVF);
537 	if (sw & SSE_UE)
538 		return (FPE_FLTUND);
539 	if (sw & SSE_PE)
540 		return (FPE_FLTRES);
541 	return (FPE_FLTINV);	/* default si_code for other exceptions */
542 }
543 
544 /*
545  * This routine is invoked as part of libc's __fpstart implementation
546  * via sysi86(2).
547  *
548  * It may be called -before- any context has been assigned in which case
549  * we try and avoid touching the hardware.  Or it may be invoked well
550  * after the context has been assigned and fiddled with, in which case
551  * just tweak it directly.
552  */
553 void
554 fpsetcw(uint16_t fcw, uint32_t mxcsr)
555 {
556 	struct fpu_ctx *fp = &curthread->t_lwp->lwp_pcb.pcb_fpu;
557 	struct fxsave_state *fx;
558 
559 	if (!fpu_exists || fp_kind == FP_NO)
560 		return;
561 
562 	if ((fp->fpu_flags & FPU_EN) == 0) {
563 		if (fcw == FPU_CW_INIT && mxcsr == SSE_MXCSR_INIT) {
564 			/*
565 			 * Common case.  Floating point unit not yet
566 			 * enabled, and kernel already intends to initialize
567 			 * the hardware the way the caller wants.
568 			 */
569 			return;
570 		}
571 		/*
572 		 * Hmm.  Userland wants a different default.
573 		 * Do a fake "first trap" to establish the context, then
574 		 * handle as if we already had a context before we came in.
575 		 */
576 		kpreempt_disable();
577 		fp_seed();
578 		kpreempt_enable();
579 	}
580 
581 	/*
582 	 * Ensure that the current hardware state is flushed back to the
583 	 * pcb, then modify that copy.  Next use of the fp will
584 	 * restore the context.
585 	 */
586 	fp_save(fp);
587 
588 #if defined(__amd64)
589 	fx = &fp->fpu_regs.kfpu_u.kfpu_fx;
590 	fx->fx_fcw = fcw;
591 	fx->fx_mxcsr = sse_mxcsr_mask & mxcsr;
592 #else
593 	switch (fp_kind) {
594 	case __FP_SSE:
595 		fx = &fp->fpu_regs.kfpu_u.kfpu_fx;
596 		fx->fx_fcw = fcw;
597 		fx->fx_mxcsr = sse_mxcsr_mask & mxcsr;
598 		break;
599 	default:
600 		fp->fpu_regs.kfpu_u.kfpu_fn.f_fcw = fcw;
601 		break;
602 	}
603 #endif
604 }
605