Lines Matching refs:f128

1623     t.f128 = float128_add(xa->f128, xb->f128, &tstat);  in helper_xsaddqp()
1630 helper_compute_fprf_float128(env, t.f128); in helper_xsaddqp()
1695 t.f128 = float128_mul(xa->f128, xb->f128, &tstat); in helper_xsmulqp()
1701 helper_compute_fprf_float128(env, t.f128); in helper_xsmulqp()
1769 t.f128 = float128_div(xa->f128, xb->f128, &tstat); in helper_xsdivqp()
1779 helper_compute_fprf_float128(env, t.f128); in helper_xsdivqp()
2100 t.f128 = float128_muladd(s1->f128, s3->f128, s2->f128, maddflgs, &tstat); \
2108 helper_compute_fprf_float128(env, t.f128); \
2165 VSX_SCALAR_CMP(XSCMPEQQP, float128, eq, f128, 0)
2166 VSX_SCALAR_CMP(XSCMPGEQP, float128, le, f128, 1)
2167 VSX_SCALAR_CMP(XSCMPGTQP, float128, lt, f128, 1)
2207 if (unlikely(float128_is_any_nan(xa->f128) || in helper_xscmpexpqp()
2208 float128_is_any_nan(xb->f128))) { in helper_xscmpexpqp()
2300 switch (float128_compare(xa->f128, xb->f128, &env->fp_status)) { in do_scalar_cmpq()
2313 if (float128_is_signaling_nan(xa->f128, &env->fp_status) || in do_scalar_cmpq()
2314 float128_is_signaling_nan(xb->f128, &env->fp_status)) { in do_scalar_cmpq()
2319 } else if (float128_is_quiet_nan(xa->f128, &env->fp_status) || in do_scalar_cmpq()
2320 float128_is_quiet_nan(xb->f128, &env->fp_status)) { in do_scalar_cmpq()
2420 VSX_MAX_MINC(XSMAXCQP, true, float128, f128);
2421 VSX_MAX_MINC(XSMINCQP, false, float128, f128);
2637 VSX_CVT_FP_TO_FP_VECTOR(xscvdpqp, 1, float64, float128, VsrD(0), f128, 1)
2712 t.VsrD(0) = float128_to_float64(xb->f128, &tstat); in helper_XSCVQPDP()
2714 if (unlikely(float128_is_signaling_nan(xb->f128, &tstat))) { in helper_XSCVQPDP()
2825 t.s128 = float128_to_##tp##_round_to_zero(xb->f128, &env->fp_status); \
2906 VSX_CVT_FP_TO_INT_VECTOR(xscvqpsdz, float128, int64, f128, VsrD(0), \
2908 VSX_CVT_FP_TO_INT_VECTOR(xscvqpswz, float128, int32, f128, VsrD(0), \
2910 VSX_CVT_FP_TO_INT_VECTOR(xscvqpudz, float128, uint64, f128, VsrD(0), 0x0ULL)
2911 VSX_CVT_FP_TO_INT_VECTOR(xscvqpuwz, float128, uint32, f128, VsrD(0), 0x0ULL)
2979 xt->f128 = tp##_to_float128(xb->s128, &env->fp_status); \
2980 helper_compute_fprf_float128(env, xt->f128); \
3009 VSX_CVT_INT_TO_FP_VECTOR(xscvsdqp, int64, float128, VsrD(0), f128)
3010 VSX_CVT_INT_TO_FP_VECTOR(xscvudqp, uint64, float128, VsrD(0), f128)
3177 VSX_XS_TSTDC(XSTSTDCQP, f128, float128) in VSX_XS_TSTDC() argument
3231 t.f128 = float128_round_to_int(xb->f128, &tstat); in helper_xsrqpi()
3242 helper_compute_fprf_float128(env, t.f128); in helper_xsrqpi()
3285 round_res = float128_to_floatx80(xb->f128, &tstat); in helper_xsrqpxp()
3286 t.f128 = floatx80_to_float128(round_res, &tstat); in helper_xsrqpxp()
3291 t.f128 = float128_snan_to_qnan(t.f128); in helper_xsrqpxp()
3294 helper_compute_fprf_float128(env, t.f128); in helper_xsrqpxp()
3313 t.f128 = float128_sqrt(xb->f128, &tstat); in helper_xssqrtqp()
3320 helper_compute_fprf_float128(env, t.f128); in helper_xssqrtqp()
3339 t.f128 = float128_sub(xa->f128, xb->f128, &tstat); in helper_xssubqp()
3346 helper_compute_fprf_float128(env, t.f128); in helper_xssubqp()