I am checking in this as an obvious fix.

H.J.
---
Since XMM BF16 tests only require SSE2, replace vmovdqu with movdqu in
BF16 XMM ABI tests to support SSE2 machines without AVX.

Tested on x86-64 machines with and without AVX.

        * gcc.target/x86_64/abi/bf16/asm-support.S: Replace vmovdqu with
        movdqu.
---
 .../gcc.target/x86_64/abi/bf16/asm-support.S  | 36 +++++++++----------
 1 file changed, 18 insertions(+), 18 deletions(-)

diff --git a/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S 
b/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S
index a8165d86317..7559aa910c4 100644
--- a/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S
+++ b/gcc/testsuite/gcc.target/x86_64/abi/bf16/asm-support.S
@@ -20,22 +20,22 @@ snapshot:
        movq    %r13, r13(%rip)
        movq    %r14, r14(%rip)
        movq    %r15, r15(%rip)
-       vmovdqu %xmm0, xmm_regs+0(%rip)
-       vmovdqu %xmm1, xmm_regs+16(%rip)
-       vmovdqu %xmm2, xmm_regs+32(%rip)
-       vmovdqu %xmm3, xmm_regs+48(%rip)
-       vmovdqu %xmm4, xmm_regs+64(%rip)
-       vmovdqu %xmm5, xmm_regs+80(%rip)
-       vmovdqu %xmm6, xmm_regs+96(%rip)
-       vmovdqu %xmm7, xmm_regs+112(%rip)
-       vmovdqu %xmm8, xmm_regs+128(%rip)
-       vmovdqu %xmm9, xmm_regs+144(%rip)
-       vmovdqu %xmm10, xmm_regs+160(%rip)
-       vmovdqu %xmm11, xmm_regs+176(%rip)
-       vmovdqu %xmm12, xmm_regs+192(%rip)
-       vmovdqu %xmm13, xmm_regs+208(%rip)
-       vmovdqu %xmm14, xmm_regs+224(%rip)
-       vmovdqu %xmm15, xmm_regs+240(%rip)
+       movdqu  %xmm0, xmm_regs+0(%rip)
+       movdqu  %xmm1, xmm_regs+16(%rip)
+       movdqu  %xmm2, xmm_regs+32(%rip)
+       movdqu  %xmm3, xmm_regs+48(%rip)
+       movdqu  %xmm4, xmm_regs+64(%rip)
+       movdqu  %xmm5, xmm_regs+80(%rip)
+       movdqu  %xmm6, xmm_regs+96(%rip)
+       movdqu  %xmm7, xmm_regs+112(%rip)
+       movdqu  %xmm8, xmm_regs+128(%rip)
+       movdqu  %xmm9, xmm_regs+144(%rip)
+       movdqu  %xmm10, xmm_regs+160(%rip)
+       movdqu  %xmm11, xmm_regs+176(%rip)
+       movdqu  %xmm12, xmm_regs+192(%rip)
+       movdqu  %xmm13, xmm_regs+208(%rip)
+       movdqu  %xmm14, xmm_regs+224(%rip)
+       movdqu  %xmm15, xmm_regs+240(%rip)
        jmp     *callthis(%rip)
 .LFE3:
        .size   snapshot, .-snapshot
@@ -50,8 +50,8 @@ snapshot_ret:
        addq    $8, %rsp
        movq    %rax, rax(%rip)
        movq    %rdx, rdx(%rip)
-       vmovdqu %xmm0, xmm_regs+0(%rip)
-       vmovdqu %xmm1, xmm_regs+16(%rip)
+       movdqu  %xmm0, xmm_regs+0(%rip)
+       movdqu  %xmm1, xmm_regs+16(%rip)
        fstpt   x87_regs(%rip)
        fstpt   x87_regs+16(%rip)
        fldt    x87_regs+16(%rip)
-- 
2.37.2

Reply via email to