| /src/contrib/arm-optimized-routines/math/aarch64/sve/ |
| H A D | logf.c | 38 special_case (svuint32_t u_off, svfloat32_t p, svfloat32_t r2, svfloat32_t y, in special_case() argument 42 logf, svreinterpret_f32 (svadd_x (svptrue_b32 (), u_off, data.off)), in special_case() 54 svuint32_t u_off = svreinterpret_u32 (x); in log() local 56 u_off = svsub_x (pg, u_off, d->off); in log() 57 svbool_t cmp = svcmpge (pg, svsub_x (pg, u_off, d->lower), Thresh); in log() 61 pg, svasr_x (pg, svreinterpret_s32 (u_off), 23)); /* Sign-extend. */ in log() 63 svuint32_t u = svand_x (pg, u_off, Mask); in log() 81 return special_case (u_off, p, r2, y, cmp); in log()
|
| H A D | log2f.c | 37 special_case (svuint32_t u_off, svfloat32_t p, svfloat32_t r2, svfloat32_t y, in special_case() argument 41 log2f, svreinterpret_f32 (svadd_x (svptrue_b32 (), u_off, data.off)), in special_case() 54 svuint32_t u_off = svreinterpret_u32 (x); in log2() local 56 u_off = svsub_x (pg, u_off, d->off); in log2() 57 svbool_t special = svcmpge (pg, svsub_x (pg, u_off, d->lower), Thresh); in log2() 61 pg, svasr_x (pg, svreinterpret_s32 (u_off), 23)); /* Sign-extend. */ in log2() 62 svuint32_t u = svand_x (pg, u_off, MantissaMask); in log2() 81 return special_case (u_off, n, r, y, special); in log2()
|
| H A D | log10f.c | 40 special_case (svuint32_t u_off, svfloat32_t p, svfloat32_t r2, svfloat32_t y, in special_case() argument 44 log10f, svreinterpret_f32 (svadd_x (svptrue_b32 (), u_off, data.off)), in special_case() 57 svuint32_t u_off = svreinterpret_u32 (x); in log10() local 59 u_off = svsub_x (pg, u_off, d->off); in log10() 60 svbool_t special = svcmpge (pg, svsub_x (pg, u_off, d->lower), Thres); in log10() 64 pg, svasr_x (pg, svreinterpret_s32 (u_off), 23)); /* signextend. */ in log10() 65 svuint32_t ix = svand_x (pg, u_off, MantissaMask); in log10() 89 return special_case (u_off, hi, r2, y, special); in log10()
|
| /src/contrib/arm-optimized-routines/math/aarch64/advsimd/ |
| H A D | log2f.c | 42 special_case (float32x4_t n, uint32x4_t u_off, float32x4_t p, float32x4_t r, in special_case() argument 46 return v_call_f32 (log2f, vreinterpretq_f32_u32 (vaddq_u32 (u_off, d->off)), in special_case() 62 uint32x4_t u_off = vreinterpretq_u32_f32 (x); in log2() local 65 u_off = vsubq_u32 (u_off, d->off); in log2() 67 vshrq_n_s32 (vreinterpretq_s32_u32 (u_off), 23)); /* signextend. */ in log2() 69 uint16x4_t special = vcge_u16 (vsubhn_u32 (u_off, d->offset_lower_bound), in log2() 72 uint32x4_t u = vaddq_u32 (vandq_u32 (u_off, d->mantissa_mask), d->off); in log2() 89 return special_case (n, u_off, p, r, special, d); in log2()
|
| H A D | log10f.c | 42 special_case (float32x4_t y, uint32x4_t u_off, float32x4_t p, float32x4_t r2, in special_case() argument 46 return v_call_f32 (log10f, vreinterpretq_f32_u32 (vaddq_u32 (u_off, d->off)), in special_case() 63 uint32x4_t u_off = vreinterpretq_u32_f32 (x); in log10() local 66 u_off = vsubq_u32 (u_off, d->off); in log10() 68 vshrq_n_s32 (vreinterpretq_s32_u32 (u_off), 23)); /* signextend. */ in log10() 70 uint16x4_t special = vcge_u16 (vsubhn_u32 (u_off, d->offset_lower_bound), in log10() 73 uint32x4_t u = vaddq_u32 (vandq_u32 (u_off, d->mantissa_mask), d->off); in log10() 93 return special_case (y, u_off, poly, r2, special, d); in log10()
|
| H A D | log.c | 59 special_case (float64x2_t hi, uint64x2_t u_off, float64x2_t y, float64x2_t r2, in special_case() argument 62 float64x2_t x = vreinterpretq_f64_u64 (vaddq_u64 (u_off, d->off)); in special_case() 78 uint64x2_t u_off = vsubq_u64 (u, d->off); in log() local 83 int64x2_t k = vshrq_n_s64 (vreinterpretq_s64_u64 (u_off), 52); in log() 84 uint64x2_t iz = vsubq_u64 (u, vandq_u64 (u_off, d->sign_exp_mask)); in log() 87 struct entry e = lookup (u_off); in log() 89 uint32x2_t special = vcge_u32 (vsubhn_u64 (u_off, d->offset_lower_bound), in log() 109 return special_case (hi, u_off, y, r2, special, d); in log()
|
| H A D | log10.c | 63 special_case (float64x2_t hi, uint64x2_t u_off, float64x2_t y, float64x2_t r2, in special_case() argument 66 float64x2_t x = vreinterpretq_f64_u64 (vaddq_u64 (u_off, d->off)); in special_case() 84 uint64x2_t u_off = vsubq_u64 (u, d->off); in log10() local 89 int64x2_t k = vshrq_n_s64 (vreinterpretq_s64_u64 (u_off), 52); in log10() 90 uint64x2_t iz = vsubq_u64 (u, vandq_u64 (u_off, d->sign_exp_mask)); in log10() 93 struct entry e = lookup (u_off); in log10() 95 uint32x2_t special = vcge_u32 (vsubhn_u64 (u_off, d->offset_lower_bound), in log10() 120 return special_case (hi, u_off, y, r2, special, d); in log10()
|
| H A D | log2.c | 62 special_case (float64x2_t hi, uint64x2_t u_off, float64x2_t y, float64x2_t r2, in special_case() argument 65 float64x2_t x = vreinterpretq_f64_u64 (vaddq_u64 (u_off, d->off)); in special_case() 82 uint64x2_t u_off = vsubq_u64 (u, d->off); in log2() local 87 int64x2_t k = vshrq_n_s64 (vreinterpretq_s64_u64 (u_off), 52); in log2() 88 uint64x2_t iz = vsubq_u64 (u, vandq_u64 (u_off, d->sign_exp_mask)); in log2() 91 struct entry e = lookup (u_off); in log2() 93 uint32x2_t special = vcge_u32 (vsubhn_u64 (u_off, d->offset_lower_bound), in log2() 112 return special_case (hi, u_off, y, r2, special, d); in log2()
|
| H A D | logf.c | 37 special_case (float32x4_t p, uint32x4_t u_off, float32x4_t y, float32x4_t r2, in special_case() argument 41 return v_call_f32 (logf, vreinterpretq_f32_u32 (vaddq_u32 (u_off, d->off)), in special_case() 53 uint32x4_t u_off = vsubq_u32 (vreinterpretq_u32_f32 (x), d->off); in log() local 57 vshrq_n_s32 (vreinterpretq_s32_u32 (u_off), 23)); /* signextend. */ in log() 58 uint16x4_t cmp = vcge_u16 (vsubhn_u32 (u_off, d->offset_lower_bound), in log() 61 uint32x4_t u = vaddq_u32 (vandq_u32 (u_off, d->mantissa_mask), d->off); in log() 77 return special_case (p, u_off, y, r2, cmp, d); in log()
|
| H A D | asinh.c | 102 uint64x2_t u_off = vsubq_u64 (u, d->off); in log_inline() local 104 int64x2_t k = vshrq_n_s64 (vreinterpretq_s64_u64 (u_off), 52); in log_inline() 105 uint64x2_t iz = vsubq_u64 (u, vandq_u64 (u_off, d->mask)); in log_inline() 108 struct entry e = lookup (u_off); in log_inline()
|