| /src/contrib/arm-optimized-routines/math/aarch64/advsimd/ |
| H A D | v_expm1_inline.h | 20 double ln2[2]; member 31 .ln2 = { 0x1.62e42fefa39efp-1, 0x1.abc9e3b39803fp-56 }, \ 41 float64x2_t ln2 = vld1q_f64 (&d->ln2[0]); in expm1_inline() local 50 float64x2_t f = vfmsq_laneq_f64 (x, n, ln2, 0); in expm1_inline() 51 f = vfmsq_laneq_f64 (f, n, ln2, 1); in expm1_inline()
|
| H A D | cosh.c | 16 double ln2[2]; member 25 .ln2 = {-0x1.62e42fefa39efp-9, -0x1.abc9e3b39803f3p-64}, 53 float64x2_t ln2 = vld1q_f64 (d->ln2); in exp_inline() local 54 float64x2_t r = vfmaq_laneq_f64 (x, n, ln2, 0); in exp_inline() 55 r = vfmaq_laneq_f64 (r, n, ln2, 1); in exp_inline()
|
| H A D | v_log1p_inline.h | 19 double ln2[2]; member 35 .ln2 = { 0x1.62e42fefa3800p-1, 0x1.ef35793c76730p-45 }, \ 113 float64x2_t ln2 = vld1q_f64 (&d->ln2[0]); in log1p_inline() local 114 float64x2_t ylo = vfmaq_laneq_f64 (cm, k, ln2, 1); in log1p_inline() 115 float64x2_t yhi = vfmaq_laneq_f64 (f, k, ln2, 0); in log1p_inline()
|
| H A D | v_expm1f_inline.h | 37 float32x2_t ln2 = vld1_f32 (&d->ln2_hi); in expm1f_inline() local 43 float32x4_t f = vfmsq_lane_f32 (x, j, ln2, 0); in expm1f_inline() 44 f = vfmsq_lane_f32 (f, j, ln2, 1); in expm1f_inline()
|
| H A D | v_log1pf_inline.h | 20 float32x4_t c4, c6, c1, c2, ln2; member 31 .ln2 = V4 (0x1.62e43p-1f), .four = V4 (0x40800000), \ 91 return vfmaq_f32 (p, scale_back, d->ln2); in log1pf_inline()
|
| H A D | logf.c | 13 float32x4_t c2, c4, c6, ln2; member 26 .ln2 = V4 (0x1.62e43p-1f), 74 p = vfmaq_f32 (r, d->ln2, n); in log()
|
| H A D | v_log_inline.h | 35 float64x2_t ln2; member 41 .poly = POLY, .ln2 = V2 (0x1.62e42fefa39efp-1), \ 93 hi = vfmaq_f64 (vaddq_f64 (e.logc, r), kd, d->ln2); in v_log_inline()
|
| H A D | log10f.c | 14 float32x4_t c0, c2, c4, c6, inv_ln10, ln2; member 30 .ln2 = V4 (0x1.62e43p-1f), 89 float32x4_t y = vfmaq_f32 (r, d->ln2, n); in log10()
|
| /src/contrib/arm-optimized-routines/math/tools/ |
| H A D | log2_abs.sollya | 11 ln2 = evaluate(log(2),0); 12 invln2hi = double(1/ln2 + 0x1p21) - 0x1p21; // round away last 21 bits 13 invln2lo = double(1/ln2 - invln2hi); 16 f = log(1+x)/ln2; 37 //for i from 0 to 60 do { g = g + (-x)^i/(i+1)/ln2; };
|
| H A D | log2.sollya | 11 ln2 = evaluate(log(2),0); 12 invln2hi = double(1/ln2 + 0x1p21) - 0x1p21; // round away last 21 bits 13 invln2lo = double(1/ln2 - invln2hi); 22 f = f/ln2;
|
| /src/contrib/arm-optimized-routines/math/aarch64/sve/ |
| H A D | sv_log1p_inline.h | 16 double poly[19], ln2[2]; member 31 .ln2 = { 0x1.62e42fefa3800p-1, 0x1.ef35793c76730p-45 }, 91 svfloat64_t ylo = svmla_x (pg, cm, k, d->ln2[0]); in sv_log1p_inline() 92 svfloat64_t yhi = svmla_x (pg, f, k, d->ln2[1]); in sv_log1p_inline()
|
| H A D | sv_log1pf_inline.h | 18 float ln2, exp_bias, quarter; member 26 .c6 = 0x1.abcb6p-4f, .c7 = -0x1.6f0d5ep-5f, .ln2 = 0x1.62e43p-1f, 58 svfloat32_t fconst = svld1rq_f32 (svptrue_b32 (), &d->ln2); in sv_log1pf_inline()
|
| H A D | sv_log_inline.h | 34 double ln2; member 40 .poly = POLY, .ln2 = 0x1.62e42fefa39efp-1, \ 73 svfloat64_t hi = svmla_x (pg, svadd_x (pg, logc, r), kd, __v_log_data.ln2); in sv_log_inline()
|
| H A D | expm1.c | 61 svfloat64_t ln2 = svld1rq (svptrue_b64 (), &d->ln2_hi); in expm1() local 62 svfloat64_t f = svmls_lane (x, n, ln2, 0); in expm1() 63 f = svmls_lane (f, n, ln2, 1); in expm1()
|
| H A D | logf.c | 16 float ln2; member 26 .ln2 = 0x1.62e43p-1f, 78 p = svmla_x (pg, r, n, d->ln2); in log()
|
| H A D | log10f.c | 16 float ln2, inv_ln10; member 27 .ln2 = 0x1.62e43p-1f, 85 svfloat32_t hi = svmla_x (pg, r, n, d->ln2); in log10()
|
| H A D | exp.c | 100 svfloat64_t ln2 = svld1rq (svptrue_b64 (), &d->ln2_hi); in exp() local 101 svfloat64_t r = svmls_lane (x, n, ln2, 0); in exp() 102 r = svmls_lane (r, n, ln2, 1); in exp()
|
| /src/lib/msun/src/ |
| H A D | e_acoshl.c | 53 #define ln2 u_ln2.e macro 56 ln2 = 6.93147180559945309417232121458176568e-1L; /* 0x162e42fefa39ef35793c7673007e6.0p-113 */ variable 75 RETURNI(logl(x)+ln2); /* acosh(huge)=log(2x), or misnormal */ in acoshl()
|
| H A D | s_asinhl.c | 56 #define ln2 u_ln2.e macro 59 ln2 = 6.93147180559945309417232121458176568e-1L; /* 0x162e42fefa39ef35793c7673007e6.0p-113 */ variable 78 w = logl(fabsl(x))+ln2; in asinhl()
|
| H A D | e_acoshf.c | 21 ln2 = 6.9314718246e-01; /* 0x3f317218 */ variable 35 return logf(x)+ln2; /* acosh(huge)=log(2x) */ in acoshf()
|
| H A D | e_acosh.c | 35 ln2 = 6.93147180559945286227e-01; /* 0x3FE62E42, 0xFEFA39EF */ variable 50 return log(x)+ln2; /* acosh(huge)=log(2x) */ in acosh()
|
| H A D | s_asinhf.c | 21 ln2 = 6.9314718246e-01, /* 0x3f317218 */ variable 36 w = logf(fabsf(x))+ln2; in asinhf()
|
| H A D | s_asinh.c | 30 ln2 = 6.93147180559945286227e-01, /* 0x3FE62E42, 0xFEFA39EF */ variable 45 w = log(fabs(x))+ln2; in asinh()
|
| /src/contrib/arm-optimized-routines/math/aarch64/experimental/advsimd/ |
| H A D | v_logf_inline.h | 13 float32x4_t ln2; member 23 .ln2 = V4 (0x1.62e43p-1f), .off = V4 (0x3f2aaaab), \ 54 p = vfmaq_f32 (r, d->ln2, n); in v_logf_inline()
|
| /src/contrib/arm-optimized-routines/math/aarch64/experimental/sve/ |
| H A D | sv_logf_inline.h | 13 float ln2; member 21 .p6 = -0x1.3e737cp-3f, .ln2 = 0x1.62e43p-1f, .off = 0x3f2aaaab, \ 48 p = svmla_x (pg, r, n, d->ln2); in sv_logf_inline()
|