diff options
author | Joseph Myers <joseph@codesourcery.com> | 2015-09-23 18:14:57 +0000 |
---|---|---|
committer | Joseph Myers <joseph@codesourcery.com> | 2015-09-23 18:14:57 +0000 |
commit | 54142c44e963f410262fa868c159b6df858f3c53 (patch) | |
tree | cc38bd53701afabc893a378e24fefd25d2062d86 /sysdeps/ieee754/dbl-64/k_rem_pio2.c | |
parent | 5df386a5492e00348c0ba4c6520d27783c253816 (diff) | |
download | glibc-54142c44e963f410262fa868c159b6df858f3c53.tar.gz glibc-54142c44e963f410262fa868c159b6df858f3c53.tar.xz glibc-54142c44e963f410262fa868c159b6df858f3c53.zip |
Use math_narrow_eval more consistently.
Where glibc code needs to avoid excess range and precision in floating-point arithmetic, code variously uses either asms or volatile to force the results of that arithmetic to memory; mostly this is conditional on FLT_EVAL_METHOD, but in the case of lrint / llrint functions some use of volatile is unconditional (and is present unnecessarily in versions for long double). This patch make such code use the recently-added math_narrow_eval macro consistently, removing the unnecessary uses of volatile in long double lrint / llrint implementations completely. Tested for x86_64, x86, mips64 and powerpc. * math/s_nexttowardf.c (__nexttowardf): Use math_narrow_eval. * stdlib/strtod_l.c: Include <math_private.h>. (overflow_value): Use math_narrow_eval. (underflow_value): Likewise. * sysdeps/i386/fpu/s_nexttoward.c (__nexttoward): Likewise. * sysdeps/i386/fpu/s_nexttowardf.c (__nexttowardf): Likewise. * sysdeps/ieee754/dbl-64/e_gamma_r.c (gamma_positive): Likewise. (__ieee754_gamma_r): Likewise. * sysdeps/ieee754/dbl-64/gamma_productf.c (__gamma_productf): Likewise. * sysdeps/ieee754/dbl-64/k_rem_pio2.c (__kernel_rem_pio2): Likewise. * sysdeps/ieee754/dbl-64/lgamma_neg.c (__lgamma_neg): Likewise. * sysdeps/ieee754/dbl-64/s_erf.c (__erfc): Likewise. * sysdeps/ieee754/dbl-64/s_llrint.c (__llrint): Likewise. * sysdeps/ieee754/dbl-64/s_lrint.c (__lrint): Likewise. * sysdeps/ieee754/flt-32/e_gammaf_r.c (gammaf_positive): Likewise. (__ieee754_gammaf_r): Likewise. * sysdeps/ieee754/flt-32/k_rem_pio2f.c (__kernel_rem_pio2f): Likewise. * sysdeps/ieee754/flt-32/lgamma_negf.c (__lgamma_negf): Likewise. * sysdeps/ieee754/flt-32/s_erff.c (__erfcf): Likewise. * sysdeps/ieee754/flt-32/s_llrintf.c (__llrintf): Likewise. * sysdeps/ieee754/flt-32/s_lrintf.c (__lrintf): Likewise. * sysdeps/ieee754/ldbl-128/s_llrintl.c (__llrintl): Do not use volatile. * sysdeps/ieee754/ldbl-128/s_lrintl.c (__lrintl): Likewise. * sysdeps/ieee754/ldbl-128/s_nexttoward.c (__nexttoward): Use math_narrow_eval. * sysdeps/ieee754/ldbl-128ibm/s_nexttoward.c (__nexttoward): Likewise. * sysdeps/ieee754/ldbl-128ibm/s_nexttowardf.c (__nexttowardf): Likewise. * sysdeps/ieee754/ldbl-96/gamma_product.c (__gamma_product): Likewise. * sysdeps/ieee754/ldbl-96/s_llrintl.c (__llrintl): Do not use volatile. * sysdeps/ieee754/ldbl-96/s_lrintl.c (__lrintl): Likewise. * sysdeps/ieee754/ldbl-96/s_nexttoward.c (__nexttoward): Use math_narrow_eval. * sysdeps/ieee754/ldbl-96/s_nexttowardf.c (__nexttowardf): Likewise. * sysdeps/ieee754/ldbl-opt/s_nexttowardfd.c (__nldbl_nexttowardf): Likewise.
Diffstat (limited to 'sysdeps/ieee754/dbl-64/k_rem_pio2.c')
-rw-r--r-- | sysdeps/ieee754/dbl-64/k_rem_pio2.c | 19 |
1 files changed, 5 insertions, 14 deletions
diff --git a/sysdeps/ieee754/dbl-64/k_rem_pio2.c b/sysdeps/ieee754/dbl-64/k_rem_pio2.c index 047c6c2886..e58c9e854c 100644 --- a/sysdeps/ieee754/dbl-64/k_rem_pio2.c +++ b/sysdeps/ieee754/dbl-64/k_rem_pio2.c @@ -315,34 +315,25 @@ recompute: break; case 1: case 2:; -#if __FLT_EVAL_METHOD__ != 0 - volatile -#endif double fv = 0.0; for (i = jz; i >= 0; i--) - fv += fq[i]; + fv = math_narrow_eval (fv + fq[i]); y[0] = (ih == 0) ? fv : -fv; - fv = fq[0] - fv; + fv = math_narrow_eval (fq[0] - fv); for (i = 1; i <= jz; i++) - fv += fq[i]; + fv = math_narrow_eval (fv + fq[i]); y[1] = (ih == 0) ? fv : -fv; break; case 3: /* painful */ for (i = jz; i > 0; i--) { -#if __FLT_EVAL_METHOD__ != 0 - volatile -#endif - double fv = (double) (fq[i - 1] + fq[i]); + double fv = math_narrow_eval (fq[i - 1] + fq[i]); fq[i] += fq[i - 1] - fv; fq[i - 1] = fv; } for (i = jz; i > 1; i--) { -#if __FLT_EVAL_METHOD__ != 0 - volatile -#endif - double fv = (double) (fq[i - 1] + fq[i]); + double fv = math_narrow_eval (fq[i - 1] + fq[i]); fq[i] += fq[i - 1] - fv; fq[i - 1] = fv; } |