diff options
author | Noah Goldstein <goldstein.w.n@gmail.com> | 2022-06-20 13:02:10 -0700 |
---|---|---|
committer | Noah Goldstein <goldstein.w.n@gmail.com> | 2022-06-22 19:42:17 -0700 |
commit | 3079f652d7cc34456aefb412677c01e758922527 (patch) | |
tree | 5d3f749db7fcfdeb17e3765858894a15c1267e26 /sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S | |
parent | 3edda6a0f013736ca9554a95e553739a41dbd4b7 (diff) | |
download | glibc-3079f652d7cc34456aefb412677c01e758922527.tar.gz glibc-3079f652d7cc34456aefb412677c01e758922527.tar.xz glibc-3079f652d7cc34456aefb412677c01e758922527.zip |
x86: Replace all sse instructions with vex equivilent in avx+ files
Most of these don't really matter as there was no dirty upper state but we should generally avoid stray sse when its not needed. The one case that really matters is in svml_d_tanh4_core_avx2.S: blendvps %xmm0, %xmm8, %xmm7 When there was a dirty upper state. Tested on x86_64-linux
Diffstat (limited to 'sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S')
-rw-r--r-- | sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S | 6 |
1 files changed, 3 insertions, 3 deletions
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S index ae5738c1b7..5c693d132e 100644 --- a/sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_hypot4_core_avx2.S @@ -231,12 +231,12 @@ L(SPECIAL_VALUES_LOOP): L(SCALAR_MATH_CALL): movl %r12d, %r14d - movsd 32(%rsp, %r14, 8), %xmm0 - movsd 64(%rsp, %r14, 8), %xmm1 + vmovsd 32(%rsp, %r14, 8), %xmm0 + vmovsd 64(%rsp, %r14, 8), %xmm1 call hypot@PLT # LOE rbx r14 r15 r12d r13d xmm0 - movsd %xmm0, 96(%rsp, %r14, 8) + vmovsd %xmm0, 96(%rsp, %r14, 8) /* Process special inputs in loop */ jmp L(SPECIAL_VALUES_LOOP) |