165e60ab1SDavid Schultz /*- 277927540SDavid Schultz * Copyright (c) 2005-2011 David Schultz <das@FreeBSD.ORG> 365e60ab1SDavid Schultz * All rights reserved. 465e60ab1SDavid Schultz * 565e60ab1SDavid Schultz * Redistribution and use in source and binary forms, with or without 665e60ab1SDavid Schultz * modification, are permitted provided that the following conditions 765e60ab1SDavid Schultz * are met: 865e60ab1SDavid Schultz * 1. Redistributions of source code must retain the above copyright 965e60ab1SDavid Schultz * notice, this list of conditions and the following disclaimer. 1065e60ab1SDavid Schultz * 2. Redistributions in binary form must reproduce the above copyright 1165e60ab1SDavid Schultz * notice, this list of conditions and the following disclaimer in the 1265e60ab1SDavid Schultz * documentation and/or other materials provided with the distribution. 1365e60ab1SDavid Schultz * 1465e60ab1SDavid Schultz * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 1565e60ab1SDavid Schultz * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 1665e60ab1SDavid Schultz * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 1765e60ab1SDavid Schultz * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 1865e60ab1SDavid Schultz * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 1965e60ab1SDavid Schultz * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 2065e60ab1SDavid Schultz * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 2165e60ab1SDavid Schultz * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 2265e60ab1SDavid Schultz * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 2365e60ab1SDavid Schultz * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 2465e60ab1SDavid Schultz * SUCH DAMAGE. 2565e60ab1SDavid Schultz */ 2665e60ab1SDavid Schultz 2765e60ab1SDavid Schultz #include <sys/cdefs.h> 2865e60ab1SDavid Schultz __FBSDID("$FreeBSD$"); 2965e60ab1SDavid Schultz 3065e60ab1SDavid Schultz #include <fenv.h> 3165e60ab1SDavid Schultz #include <float.h> 3265e60ab1SDavid Schultz #include <math.h> 3365e60ab1SDavid Schultz 34*ec950830SDavid Schultz #include "fpmath.h" 35*ec950830SDavid Schultz 3665e60ab1SDavid Schultz /* 3777927540SDavid Schultz * A struct dd represents a floating-point number with twice the precision 3877927540SDavid Schultz * of a long double. We maintain the invariant that "hi" stores the high-order 3977927540SDavid Schultz * bits of the result. 4077927540SDavid Schultz */ 4177927540SDavid Schultz struct dd { 4277927540SDavid Schultz long double hi; 4377927540SDavid Schultz long double lo; 4477927540SDavid Schultz }; 4577927540SDavid Schultz 4677927540SDavid Schultz /* 4777927540SDavid Schultz * Compute a+b exactly, returning the exact result in a struct dd. We assume 4877927540SDavid Schultz * that both a and b are finite, but make no assumptions about their relative 4977927540SDavid Schultz * magnitudes. 5077927540SDavid Schultz */ 5177927540SDavid Schultz static inline struct dd 5277927540SDavid Schultz dd_add(long double a, long double b) 5377927540SDavid Schultz { 5477927540SDavid Schultz struct dd ret; 5577927540SDavid Schultz long double s; 5677927540SDavid Schultz 5777927540SDavid Schultz ret.hi = a + b; 5877927540SDavid Schultz s = ret.hi - a; 5977927540SDavid Schultz ret.lo = (a - (ret.hi - s)) + (b - s); 6077927540SDavid Schultz return (ret); 6177927540SDavid Schultz } 6277927540SDavid Schultz 6377927540SDavid Schultz /* 64*ec950830SDavid Schultz * Compute a+b, with a small tweak: The least significant bit of the 65*ec950830SDavid Schultz * result is adjusted into a sticky bit summarizing all the bits that 66*ec950830SDavid Schultz * were lost to rounding. This adjustment negates the effects of double 67*ec950830SDavid Schultz * rounding when the result is added to another number with a higher 68*ec950830SDavid Schultz * exponent. For an explanation of round and sticky bits, see any reference 69*ec950830SDavid Schultz * on FPU design, e.g., 70*ec950830SDavid Schultz * 71*ec950830SDavid Schultz * J. Coonen. An Implementation Guide to a Proposed Standard for 72*ec950830SDavid Schultz * Floating-Point Arithmetic. Computer, vol. 13, no. 1, Jan 1980. 73*ec950830SDavid Schultz */ 74*ec950830SDavid Schultz static inline long double 75*ec950830SDavid Schultz add_adjusted(long double a, long double b) 76*ec950830SDavid Schultz { 77*ec950830SDavid Schultz struct dd sum; 78*ec950830SDavid Schultz union IEEEl2bits u; 79*ec950830SDavid Schultz 80*ec950830SDavid Schultz sum = dd_add(a, b); 81*ec950830SDavid Schultz if (sum.lo != 0) { 82*ec950830SDavid Schultz u.e = sum.hi; 83*ec950830SDavid Schultz if ((u.bits.manl & 1) == 0) 84*ec950830SDavid Schultz sum.hi = nextafterl(sum.hi, INFINITY * sum.lo); 85*ec950830SDavid Schultz } 86*ec950830SDavid Schultz return (sum.hi); 87*ec950830SDavid Schultz } 88*ec950830SDavid Schultz 89*ec950830SDavid Schultz /* 90*ec950830SDavid Schultz * Compute ldexp(a+b, scale) with a single rounding error. It is assumed 91*ec950830SDavid Schultz * that the result will be subnormal, and care is taken to ensure that 92*ec950830SDavid Schultz * double rounding does not occur. 93*ec950830SDavid Schultz */ 94*ec950830SDavid Schultz static inline long double 95*ec950830SDavid Schultz add_and_denormalize(long double a, long double b, int scale) 96*ec950830SDavid Schultz { 97*ec950830SDavid Schultz struct dd sum; 98*ec950830SDavid Schultz int bits_lost; 99*ec950830SDavid Schultz union IEEEl2bits u; 100*ec950830SDavid Schultz 101*ec950830SDavid Schultz sum = dd_add(a, b); 102*ec950830SDavid Schultz 103*ec950830SDavid Schultz /* 104*ec950830SDavid Schultz * If we are losing at least two bits of accuracy to denormalization, 105*ec950830SDavid Schultz * then the first lost bit becomes a round bit, and we adjust the 106*ec950830SDavid Schultz * lowest bit of sum.hi to make it a sticky bit summarizing all the 107*ec950830SDavid Schultz * bits in sum.lo. With the sticky bit adjusted, the hardware will 108*ec950830SDavid Schultz * break any ties in the correct direction. 109*ec950830SDavid Schultz * 110*ec950830SDavid Schultz * If we are losing only one bit to denormalization, however, we must 111*ec950830SDavid Schultz * break the ties manually. 112*ec950830SDavid Schultz */ 113*ec950830SDavid Schultz if (sum.lo != 0) { 114*ec950830SDavid Schultz u.e = sum.hi; 115*ec950830SDavid Schultz bits_lost = -u.bits.exp - scale + 1; 116*ec950830SDavid Schultz if (bits_lost != 1 ^ (int)(u.bits.manl & 1)) 117*ec950830SDavid Schultz sum.hi = nextafterl(sum.hi, INFINITY * sum.lo); 118*ec950830SDavid Schultz } 119*ec950830SDavid Schultz return (ldexp(sum.hi, scale)); 120*ec950830SDavid Schultz } 121*ec950830SDavid Schultz 122*ec950830SDavid Schultz /* 12377927540SDavid Schultz * Compute a*b exactly, returning the exact result in a struct dd. We assume 12477927540SDavid Schultz * that both a and b are normalized, so no underflow or overflow will occur. 12577927540SDavid Schultz * The current rounding mode must be round-to-nearest. 12677927540SDavid Schultz */ 12777927540SDavid Schultz static inline struct dd 12877927540SDavid Schultz dd_mul(long double a, long double b) 12977927540SDavid Schultz { 13077927540SDavid Schultz #if LDBL_MANT_DIG == 64 13177927540SDavid Schultz static const long double split = 0x1p32L + 1.0; 13277927540SDavid Schultz #elif LDBL_MANT_DIG == 113 13377927540SDavid Schultz static const long double split = 0x1p57L + 1.0; 13477927540SDavid Schultz #endif 13577927540SDavid Schultz struct dd ret; 13677927540SDavid Schultz long double ha, hb, la, lb, p, q; 13777927540SDavid Schultz 13877927540SDavid Schultz p = a * split; 13977927540SDavid Schultz ha = a - p; 14077927540SDavid Schultz ha += p; 14177927540SDavid Schultz la = a - ha; 14277927540SDavid Schultz 14377927540SDavid Schultz p = b * split; 14477927540SDavid Schultz hb = b - p; 14577927540SDavid Schultz hb += p; 14677927540SDavid Schultz lb = b - hb; 14777927540SDavid Schultz 14877927540SDavid Schultz p = ha * hb; 14977927540SDavid Schultz q = ha * lb + la * hb; 15077927540SDavid Schultz 15177927540SDavid Schultz ret.hi = p + q; 15277927540SDavid Schultz ret.lo = p - ret.hi + q + la * lb; 15377927540SDavid Schultz return (ret); 15477927540SDavid Schultz } 15577927540SDavid Schultz 15677927540SDavid Schultz /* 15765e60ab1SDavid Schultz * Fused multiply-add: Compute x * y + z with a single rounding error. 15865e60ab1SDavid Schultz * 15965e60ab1SDavid Schultz * We use scaling to avoid overflow/underflow, along with the 16065e60ab1SDavid Schultz * canonical precision-doubling technique adapted from: 16165e60ab1SDavid Schultz * 16265e60ab1SDavid Schultz * Dekker, T. A Floating-Point Technique for Extending the 16365e60ab1SDavid Schultz * Available Precision. Numer. Math. 18, 224-242 (1971). 16465e60ab1SDavid Schultz */ 16565e60ab1SDavid Schultz long double 16665e60ab1SDavid Schultz fmal(long double x, long double y, long double z) 16765e60ab1SDavid Schultz { 168*ec950830SDavid Schultz long double xs, ys, zs, adj; 169*ec950830SDavid Schultz struct dd xy, r; 17065e60ab1SDavid Schultz int oround; 17165e60ab1SDavid Schultz int ex, ey, ez; 17265e60ab1SDavid Schultz int spread; 17365e60ab1SDavid Schultz 17492a1a6e1SDavid Schultz /* 17592a1a6e1SDavid Schultz * Handle special cases. The order of operations and the particular 17692a1a6e1SDavid Schultz * return values here are crucial in handling special cases involving 17792a1a6e1SDavid Schultz * infinities, NaNs, overflows, and signed zeroes correctly. 17892a1a6e1SDavid Schultz */ 17965e60ab1SDavid Schultz if (x == 0.0 || y == 0.0) 18065e60ab1SDavid Schultz return (x * y + z); 18192a1a6e1SDavid Schultz if (z == 0.0) 18292a1a6e1SDavid Schultz return (x * y); 18392a1a6e1SDavid Schultz if (!isfinite(x) || !isfinite(y)) 18465e60ab1SDavid Schultz return (x * y + z); 18592a1a6e1SDavid Schultz if (!isfinite(z)) 18692a1a6e1SDavid Schultz return (z); 18765e60ab1SDavid Schultz 18865e60ab1SDavid Schultz xs = frexpl(x, &ex); 18965e60ab1SDavid Schultz ys = frexpl(y, &ey); 19065e60ab1SDavid Schultz zs = frexpl(z, &ez); 19165e60ab1SDavid Schultz oround = fegetround(); 19265e60ab1SDavid Schultz spread = ex + ey - ez; 19365e60ab1SDavid Schultz 19465e60ab1SDavid Schultz /* 19565e60ab1SDavid Schultz * If x * y and z are many orders of magnitude apart, the scaling 19665e60ab1SDavid Schultz * will overflow, so we handle these cases specially. Rounding 19765e60ab1SDavid Schultz * modes other than FE_TONEAREST are painful. 19865e60ab1SDavid Schultz */ 19965e60ab1SDavid Schultz if (spread < -LDBL_MANT_DIG) { 20065e60ab1SDavid Schultz feraiseexcept(FE_INEXACT); 20165e60ab1SDavid Schultz if (!isnormal(z)) 20265e60ab1SDavid Schultz feraiseexcept(FE_UNDERFLOW); 20365e60ab1SDavid Schultz switch (oround) { 20465e60ab1SDavid Schultz case FE_TONEAREST: 20565e60ab1SDavid Schultz return (z); 20665e60ab1SDavid Schultz case FE_TOWARDZERO: 20765e60ab1SDavid Schultz if (x > 0.0 ^ y < 0.0 ^ z < 0.0) 20865e60ab1SDavid Schultz return (z); 20965e60ab1SDavid Schultz else 21065e60ab1SDavid Schultz return (nextafterl(z, 0)); 21165e60ab1SDavid Schultz case FE_DOWNWARD: 21265e60ab1SDavid Schultz if (x > 0.0 ^ y < 0.0) 21365e60ab1SDavid Schultz return (z); 21465e60ab1SDavid Schultz else 21565e60ab1SDavid Schultz return (nextafterl(z, -INFINITY)); 21665e60ab1SDavid Schultz default: /* FE_UPWARD */ 21765e60ab1SDavid Schultz if (x > 0.0 ^ y < 0.0) 21865e60ab1SDavid Schultz return (nextafterl(z, INFINITY)); 21965e60ab1SDavid Schultz else 22065e60ab1SDavid Schultz return (z); 22165e60ab1SDavid Schultz } 22265e60ab1SDavid Schultz } 223*ec950830SDavid Schultz if (spread <= LDBL_MANT_DIG * 2) 224*ec950830SDavid Schultz zs = ldexpl(zs, -spread); 225*ec950830SDavid Schultz else 226*ec950830SDavid Schultz zs = copysignl(LDBL_MIN, zs); 22765e60ab1SDavid Schultz 22865e60ab1SDavid Schultz fesetround(FE_TONEAREST); 22965e60ab1SDavid Schultz 230*ec950830SDavid Schultz /* 231*ec950830SDavid Schultz * Basic approach for round-to-nearest: 232*ec950830SDavid Schultz * 233*ec950830SDavid Schultz * (xy.hi, xy.lo) = x * y (exact) 234*ec950830SDavid Schultz * (r.hi, r.lo) = xy.hi + z (exact) 235*ec950830SDavid Schultz * adj = xy.lo + r.lo (inexact; low bit is sticky) 236*ec950830SDavid Schultz * result = r.hi + adj (correctly rounded) 237*ec950830SDavid Schultz */ 23877927540SDavid Schultz xy = dd_mul(xs, ys); 23977927540SDavid Schultz r = dd_add(xy.hi, zs); 240*ec950830SDavid Schultz 241*ec950830SDavid Schultz if (r.hi == 0.0) { 242*ec950830SDavid Schultz /* 243*ec950830SDavid Schultz * When the addends cancel to 0, ensure that the result has 244*ec950830SDavid Schultz * the correct sign. 245*ec950830SDavid Schultz */ 246*ec950830SDavid Schultz fesetround(oround); 247*ec950830SDavid Schultz volatile long double vzs = zs; /* XXX gcc CSE bug workaround */ 248*ec950830SDavid Schultz return (xy.hi + vzs); 249*ec950830SDavid Schultz } 25065e60ab1SDavid Schultz 2512c243582SDavid Schultz spread = ex + ey; 252*ec950830SDavid Schultz 253*ec950830SDavid Schultz if (oround != FE_TONEAREST) { 2542c243582SDavid Schultz /* 255*ec950830SDavid Schultz * There is no need to worry about double rounding in directed 256*ec950830SDavid Schultz * rounding modes. 2572c243582SDavid Schultz */ 2582c243582SDavid Schultz fesetround(oround); 259*ec950830SDavid Schultz adj = r.lo + xy.lo; 260*ec950830SDavid Schultz return (ldexpl(r.hi + adj, spread)); 2612c243582SDavid Schultz } 262*ec950830SDavid Schultz 263*ec950830SDavid Schultz adj = add_adjusted(r.lo, xy.lo); 264*ec950830SDavid Schultz if (spread + ilogbl(r.hi) > -16383) 265*ec950830SDavid Schultz return (ldexpl(r.hi + adj, spread)); 266*ec950830SDavid Schultz else 267*ec950830SDavid Schultz return (add_and_denormalize(r.hi, adj, spread)); 26865e60ab1SDavid Schultz } 269