Home
last modified time | relevance | path

Searched refs:ln2 (Results 1 – 25 of 44) sorted by relevance

12

/src/contrib/arm-optimized-routines/math/aarch64/advsimd/
H A Dv_expm1_inline.h20 double ln2[2]; member
31 .ln2 = { 0x1.62e42fefa39efp-1, 0x1.abc9e3b39803fp-56 }, \
41 float64x2_t ln2 = vld1q_f64 (&d->ln2[0]); in expm1_inline() local
50 float64x2_t f = vfmsq_laneq_f64 (x, n, ln2, 0); in expm1_inline()
51 f = vfmsq_laneq_f64 (f, n, ln2, 1); in expm1_inline()
H A Dcosh.c16 double ln2[2]; member
25 .ln2 = {-0x1.62e42fefa39efp-9, -0x1.abc9e3b39803f3p-64},
53 float64x2_t ln2 = vld1q_f64 (d->ln2); in exp_inline() local
54 float64x2_t r = vfmaq_laneq_f64 (x, n, ln2, 0); in exp_inline()
55 r = vfmaq_laneq_f64 (r, n, ln2, 1); in exp_inline()
H A Dv_log1p_inline.h19 double ln2[2]; member
35 .ln2 = { 0x1.62e42fefa3800p-1, 0x1.ef35793c76730p-45 }, \
113 float64x2_t ln2 = vld1q_f64 (&d->ln2[0]); in log1p_inline() local
114 float64x2_t ylo = vfmaq_laneq_f64 (cm, k, ln2, 1); in log1p_inline()
115 float64x2_t yhi = vfmaq_laneq_f64 (f, k, ln2, 0); in log1p_inline()
H A Dv_expm1f_inline.h37 float32x2_t ln2 = vld1_f32 (&d->ln2_hi); in expm1f_inline() local
43 float32x4_t f = vfmsq_lane_f32 (x, j, ln2, 0); in expm1f_inline()
44 f = vfmsq_lane_f32 (f, j, ln2, 1); in expm1f_inline()
H A Dv_log1pf_inline.h20 float32x4_t c4, c6, c1, c2, ln2; member
31 .ln2 = V4 (0x1.62e43p-1f), .four = V4 (0x40800000), \
91 return vfmaq_f32 (p, scale_back, d->ln2); in log1pf_inline()
H A Dlogf.c13 float32x4_t c2, c4, c6, ln2; member
26 .ln2 = V4 (0x1.62e43p-1f),
74 p = vfmaq_f32 (r, d->ln2, n); in log()
H A Dv_log_inline.h35 float64x2_t ln2; member
41 .poly = POLY, .ln2 = V2 (0x1.62e42fefa39efp-1), \
93 hi = vfmaq_f64 (vaddq_f64 (e.logc, r), kd, d->ln2); in v_log_inline()
H A Dlog10f.c14 float32x4_t c0, c2, c4, c6, inv_ln10, ln2; member
30 .ln2 = V4 (0x1.62e43p-1f),
89 float32x4_t y = vfmaq_f32 (r, d->ln2, n); in log10()
/src/contrib/arm-optimized-routines/math/tools/
H A Dlog2_abs.sollya11 ln2 = evaluate(log(2),0);
12 invln2hi = double(1/ln2 + 0x1p21) - 0x1p21; // round away last 21 bits
13 invln2lo = double(1/ln2 - invln2hi);
16 f = log(1+x)/ln2;
37 //for i from 0 to 60 do { g = g + (-x)^i/(i+1)/ln2; };
H A Dlog2.sollya11 ln2 = evaluate(log(2),0);
12 invln2hi = double(1/ln2 + 0x1p21) - 0x1p21; // round away last 21 bits
13 invln2lo = double(1/ln2 - invln2hi);
22 f = f/ln2;
/src/contrib/arm-optimized-routines/math/aarch64/sve/
H A Dsv_log1p_inline.h16 double poly[19], ln2[2]; member
31 .ln2 = { 0x1.62e42fefa3800p-1, 0x1.ef35793c76730p-45 },
91 svfloat64_t ylo = svmla_x (pg, cm, k, d->ln2[0]); in sv_log1p_inline()
92 svfloat64_t yhi = svmla_x (pg, f, k, d->ln2[1]); in sv_log1p_inline()
H A Dsv_log1pf_inline.h18 float ln2, exp_bias, quarter; member
26 .c6 = 0x1.abcb6p-4f, .c7 = -0x1.6f0d5ep-5f, .ln2 = 0x1.62e43p-1f,
58 svfloat32_t fconst = svld1rq_f32 (svptrue_b32 (), &d->ln2); in sv_log1pf_inline()
H A Dsv_log_inline.h34 double ln2; member
40 .poly = POLY, .ln2 = 0x1.62e42fefa39efp-1, \
73 svfloat64_t hi = svmla_x (pg, svadd_x (pg, logc, r), kd, __v_log_data.ln2); in sv_log_inline()
H A Dexpm1.c61 svfloat64_t ln2 = svld1rq (svptrue_b64 (), &d->ln2_hi); in expm1() local
62 svfloat64_t f = svmls_lane (x, n, ln2, 0); in expm1()
63 f = svmls_lane (f, n, ln2, 1); in expm1()
H A Dlogf.c16 float ln2; member
26 .ln2 = 0x1.62e43p-1f,
78 p = svmla_x (pg, r, n, d->ln2); in log()
H A Dlog10f.c16 float ln2, inv_ln10; member
27 .ln2 = 0x1.62e43p-1f,
85 svfloat32_t hi = svmla_x (pg, r, n, d->ln2); in log10()
H A Dexp.c100 svfloat64_t ln2 = svld1rq (svptrue_b64 (), &d->ln2_hi); in exp() local
101 svfloat64_t r = svmls_lane (x, n, ln2, 0); in exp()
102 r = svmls_lane (r, n, ln2, 1); in exp()
/src/lib/msun/src/
H A De_acoshl.c53 #define ln2 u_ln2.e macro
56 ln2 = 6.93147180559945309417232121458176568e-1L; /* 0x162e42fefa39ef35793c7673007e6.0p-113 */ variable
75 RETURNI(logl(x)+ln2); /* acosh(huge)=log(2x), or misnormal */ in acoshl()
H A Ds_asinhl.c56 #define ln2 u_ln2.e macro
59 ln2 = 6.93147180559945309417232121458176568e-1L; /* 0x162e42fefa39ef35793c7673007e6.0p-113 */ variable
78 w = logl(fabsl(x))+ln2; in asinhl()
H A De_acoshf.c21 ln2 = 6.9314718246e-01; /* 0x3f317218 */ variable
35 return logf(x)+ln2; /* acosh(huge)=log(2x) */ in acoshf()
H A De_acosh.c35 ln2 = 6.93147180559945286227e-01; /* 0x3FE62E42, 0xFEFA39EF */ variable
50 return log(x)+ln2; /* acosh(huge)=log(2x) */ in acosh()
H A Ds_asinhf.c21 ln2 = 6.9314718246e-01, /* 0x3f317218 */ variable
36 w = logf(fabsf(x))+ln2; in asinhf()
H A Ds_asinh.c30 ln2 = 6.93147180559945286227e-01, /* 0x3FE62E42, 0xFEFA39EF */ variable
45 w = log(fabs(x))+ln2; in asinh()
/src/contrib/arm-optimized-routines/math/aarch64/experimental/advsimd/
H A Dv_logf_inline.h13 float32x4_t ln2; member
23 .ln2 = V4 (0x1.62e43p-1f), .off = V4 (0x3f2aaaab), \
54 p = vfmaq_f32 (r, d->ln2, n); in v_logf_inline()
/src/contrib/arm-optimized-routines/math/aarch64/experimental/sve/
H A Dsv_logf_inline.h13 float ln2; member
21 .p6 = -0x1.3e737cp-3f, .ln2 = 0x1.62e43p-1f, .off = 0x3f2aaaab, \
48 p = svmla_x (pg, r, n, d->ln2); in sv_logf_inline()

12