/freebsd/contrib/arm-optimized-routines/math/ |
H A D | log2f.c | 31 double_t z, r, r2, p, y, y0, invc, logc; in log2f() local 64 logc = T[i].logc; in log2f() 69 y0 = logc + (double_t) k; in log2f()
|
H A D | logf.c | 32 double_t z, r, r2, y, y0, invc, logc; in logf() local 64 logc = T[i].logc; in logf() 69 y0 = logc + (double_t) k * Ln2; in logf()
|
H A D | log10f.c | 41 double_t z, r, r2, y, y0, invc, logc; in log10f() local 73 logc = T[i].logc; in log10f() 78 y0 = logc + (double_t) k * Ln2; in log10f()
|
H A D | math_config.h | 418 double invc, logc; member 431 double invc, logc; member 448 double invc, logc; member 494 struct {double invc, logc;} tab[1 << LOG_TABLE_BITS]; member 509 struct {double invc, logc;} tab[1 << LOG2_TABLE_BITS]; member 523 struct {double invc, pad, logc, logctail;} tab[1 << POW_LOG_TABLE_BITS]; member 563 double invc, logc; member 575 double logc[V_POWF_LOG2_N]; member 594 double logc[1 << V_POW_LOG_TABLE_BITS]; member 743 double invc, logc; member
|
H A D | log2.c | 35 double_t z, r, r2, r4, y, invc, logc, kd, hi, lo, t1, t2, t3, p; in log2() local 99 logc = T[i].logc; in log2() 121 t3 = kd + logc; in log2()
|
H A D | powf.c | 34 double_t z, r, r2, r4, p, q, y, y0, invc, logc; in log2_inline() local 47 logc = T[i].logc; in log2_inline() 52 y0 = logc + (double_t) k; in log2_inline()
|
H A D | log.c | 35 double_t w, z, r, r2, r3, y, invc, logc, kd, hi, lo; in log() local 123 logc = T[i].logc; in log() 138 w = kd * Ln2hi + logc; in log()
|
H A D | pow.c | 41 double_t z, r, y, invc, logc, logctail, kd, hi, t1, t2, lo, lo1, lo2, p; in log_inline() local 57 logc = T[i].logc; in log_inline() 74 t1 = kd * Ln2hi + logc; in log_inline()
|
/freebsd/contrib/arm-optimized-routines/math/aarch64/experimental/ |
H A D | log10_2u.c | 44 double_t w, z, r, r2, r3, y, invc, logc, kd, hi, lo; in log10() local 103 logc = T[i].logc; in log10() 118 w = kd * Ln2hi + logc; in log10()
|
/freebsd/contrib/arm-optimized-routines/math/aarch64/sve/ |
H A D | sv_log_inline.h | 67 svfloat64_t logc = svld1_gather_index (pg, &__v_log_data.table[0].logc, i); in sv_log_inline() local 73 svfloat64_t hi = svmla_x (pg, svadd_x (pg, logc, r), kd, __v_log_data.ln2); in sv_log_inline()
|
H A D | log.c | 64 svfloat64_t logc = svld1_gather_index (pg, &__v_log_data.table[0].logc, i); in SV_NAME_D1() local 71 svfloat64_t hi = svmla_lane_f64 (logc, kd, ln2_and_c4, 0); in SV_NAME_D1()
|
H A D | asinh.c | 77 svfloat64_t logc = svld1_gather_index (pg, &__v_log_data.table[0].logc, i); in __sv_log_inline() local 86 svfloat64_t hi = svmla_lane (svadd_x (pg, logc, r), kd, ln2_p3, 0); in __sv_log_inline()
|
H A D | log10.c | 66 svfloat64_t logc in SV_NAME_D1() local 76 svfloat64_t w = svmla_lane_f64 (logc, r, invln10_log10_2, 0); in SV_NAME_D1()
|
H A D | powf.c | 15 #define Tlogc __v_powf_data.logc 169 svfloat64_t logc = svld1_gather_index (pg, Tlogc, i); in sv_powf_core_ext() local 173 svfloat64_t y0 = svadd_x (pg, logc, svcvt_f64_x (pg, k)); in sv_powf_core_ext()
|
H A D | pow.c | 209 svfloat64_t logc = svld1_gather_index (pg, __v_pow_log_data.logc, i); in sv_log_inline() local 218 svfloat64_t t1 = svmla_lane_f64 (logc, kd, ln2_hilo, 0); in sv_log_inline()
|
/freebsd/contrib/arm-optimized-routines/math/aarch64/advsimd/ |
H A D | v_log_inline.h | 52 float64x2_t logc; member 65 e.logc = vuzp2q_f64 (e0, e1); in log_lookup() 93 hi = vfmaq_f64 (vaddq_f64 (e.logc, r), kd, d->ln2); in v_log_inline()
|
H A D | log.c | 41 float64x2_t logc; member 54 e.logc = vuzp2q_f64 (e0, e1); in lookup() 98 float64x2_t hi = vfmaq_laneq_f64 (vaddq_f64 (e.logc, r), kd, ln2_and_c4, 0); in V_NAME_D1()
|
H A D | powf.c | 32 double invc, logc; member 99 float64x2_t invc, float64x2_t logc, float64x2_t y) in ylogx_core() argument 104 float64x2_t y0 = vaddq_f64 (logc, k); in ylogx_core()
|
H A D | finite_pow.h | 62 double logc = __v_pow_log_data.logc[i]; in log_inline() local 70 double t1 = kd * __v_pow_log_data.ln2_hi + logc; in log_inline()
|
H A D | asinh.c | 80 float64x2_t logc; member 93 e.logc = vuzp2q_f64 (e0, e1); in lookup() 116 float64x2_t hi = vfmaq_laneq_f64 (vaddq_f64 (e.logc, r), kd, ln2_and_lc4, 0); in log_inline()
|
H A D | pow.c | 100 float64x2_t logc = v_masked_lookup_f64 (__v_pow_log_data.logc, tmp); in v_log_inline() local 107 float64x2_t t1 = vfmaq_laneq_f64 (logc, kd, ln2, 1); in v_log_inline()
|
/freebsd/contrib/arm-optimized-routines/math/aarch64/ |
H A D | v_powf_data.c | 44 .logc = { -0x1.e960f97b22702p+3,
|
H A D | v_pow_log_data.c | 86 .logc
|