diff options
author | Sanjay Patel <spatel@rotateright.com> | 2018-11-11 14:57:26 +0000 |
---|---|---|
committer | Sanjay Patel <spatel@rotateright.com> | 2018-11-11 14:57:26 +0000 |
commit | 89fcd8b878977c9c467cb5d6e33a3404d2996822 (patch) | |
tree | 725d55e938e6eb622d38fa34eb4804bf5455ebc4 /test | |
parent | ac4b70f74612eccc851a1aa9ff389e2104a6c747 (diff) |
[x86] auto-generate complete checks; NFC
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@346609 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'test')
-rw-r--r-- | test/CodeGen/X86/i64-mem-copy.ll | 107 |
1 files changed, 103 insertions, 4 deletions
diff --git a/test/CodeGen/X86/i64-mem-copy.ll b/test/CodeGen/X86/i64-mem-copy.ll index e14293797e8..4dd6065189f 100644 --- a/test/CodeGen/X86/i64-mem-copy.ll +++ b/test/CodeGen/X86/i64-mem-copy.ll @@ -1,3 +1,4 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=sse2 | FileCheck %s --check-prefix=X64 ; RUN: llc < %s -mtriple=i386-unknown-unknown -mattr=sse2 | FileCheck %s --check-prefix=X32 ; RUN: llc < %s -mtriple=i386-unknown-unknown -mattr=avx2 | FileCheck %s --check-prefix=X32AVX @@ -19,6 +20,14 @@ define void @foo(i64* %x, i64* %y) { ; X32-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero ; X32-NEXT: movsd %xmm0, (%eax) ; X32-NEXT: retl +; +; X32AVX-LABEL: foo: +; X32AVX: # %bb.0: +; X32AVX-NEXT: movl {{[0-9]+}}(%esp), %eax +; X32AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X32AVX-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero +; X32AVX-NEXT: vmovsd %xmm0, (%eax) +; X32AVX-NEXT: retl %tmp1 = load i64, i64* %y, align 8 store i64 %tmp1, i64* %x, align 8 ret void @@ -40,6 +49,13 @@ define void @store_i64_from_vector(<8 x i16> %x, <8 x i16> %y, i64* %i) { ; X32-NEXT: paddw %xmm1, %xmm0 ; X32-NEXT: movq %xmm0, (%eax) ; X32-NEXT: retl +; +; X32AVX-LABEL: store_i64_from_vector: +; X32AVX: # %bb.0: +; X32AVX-NEXT: movl {{[0-9]+}}(%esp), %eax +; X32AVX-NEXT: vpaddw %xmm1, %xmm0, %xmm0 +; X32AVX-NEXT: vmovq %xmm0, (%eax) +; X32AVX-NEXT: retl %z = add <8 x i16> %x, %y ; force execution domain %bc = bitcast <8 x i16> %z to <2 x i64> %vecext = extractelement <2 x i64> %bc, i32 0 @@ -48,6 +64,29 @@ define void @store_i64_from_vector(<8 x i16> %x, <8 x i16> %y, i64* %i) { } define void @store_i64_from_vector256(<16 x i16> %x, <16 x i16> %y, i64* %i) { +; X64-LABEL: store_i64_from_vector256: +; X64: # %bb.0: +; X64-NEXT: paddw %xmm3, %xmm1 +; X64-NEXT: movq %xmm1, (%rdi) +; X64-NEXT: retq +; +; X32-LABEL: store_i64_from_vector256: +; X32: # %bb.0: +; X32-NEXT: pushl %ebp +; X32-NEXT: .cfi_def_cfa_offset 8 +; X32-NEXT: .cfi_offset %ebp, -8 +; X32-NEXT: movl %esp, %ebp +; X32-NEXT: .cfi_def_cfa_register %ebp +; X32-NEXT: andl $-16, %esp +; X32-NEXT: subl $16, %esp +; X32-NEXT: movl 24(%ebp), %eax +; X32-NEXT: paddw 8(%ebp), %xmm1 +; X32-NEXT: movq %xmm1, (%eax) +; X32-NEXT: movl %ebp, %esp +; X32-NEXT: popl %ebp +; X32-NEXT: .cfi_def_cfa %esp, 4 +; X32-NEXT: retl +; ; X32AVX-LABEL: store_i64_from_vector256: ; X32AVX: # %bb.0: ; X32AVX-NEXT: movl {{[0-9]+}}(%esp), %eax @@ -66,13 +105,73 @@ define void @store_i64_from_vector256(<16 x i16> %x, <16 x i16> %y, i64* %i) { ; PR23476 ; Handle extraction from a non-simple / pre-legalization type. -define void @PR23476(<5 x i64> %in, i64* %out, i32 %index) { +define void @PR23476(<5 x i64> %in, i64* %out, i32 %index) nounwind { +; X64-LABEL: PR23476: +; X64: # %bb.0: +; X64-NEXT: pushq %rbp +; X64-NEXT: movq %rsp, %rbp +; X64-NEXT: andq $-64, %rsp +; X64-NEXT: subq $128, %rsp +; X64-NEXT: movq %rsi, %xmm0 +; X64-NEXT: movq %rdi, %xmm1 +; X64-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] +; X64-NEXT: movq %rcx, %xmm0 +; X64-NEXT: movq %rdx, %xmm2 +; X64-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm0[0] +; X64-NEXT: movl 16(%rbp), %eax +; X64-NEXT: andl $7, %eax +; X64-NEXT: movq %r8, %xmm0 +; X64-NEXT: movdqa %xmm0, {{[0-9]+}}(%rsp) +; X64-NEXT: movdqa %xmm2, {{[0-9]+}}(%rsp) +; X64-NEXT: movdqa %xmm1, (%rsp) +; X64-NEXT: movq (%rsp,%rax,8), %rax +; X64-NEXT: movq %rax, (%r9) +; X64-NEXT: movq %rbp, %rsp +; X64-NEXT: popq %rbp +; X64-NEXT: retq +; ; X32-LABEL: PR23476: -; X32: andl $7, %eax -; X32: movsd {{.*#+}} xmm0 = mem[0],zero -; X32: movsd {{.*#+}} xmm0 = mem[0],zero +; X32: # %bb.0: +; X32-NEXT: pushl %ebp +; X32-NEXT: movl %esp, %ebp +; X32-NEXT: andl $-64, %esp +; X32-NEXT: subl $128, %esp +; X32-NEXT: movl 52(%ebp), %eax +; X32-NEXT: andl $7, %eax +; X32-NEXT: movl 48(%ebp), %ecx +; X32-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero +; X32-NEXT: movups 8(%ebp), %xmm1 +; X32-NEXT: movups 24(%ebp), %xmm2 +; X32-NEXT: movaps %xmm2, {{[0-9]+}}(%esp) +; X32-NEXT: movaps %xmm1, (%esp) +; X32-NEXT: movaps %xmm0, {{[0-9]+}}(%esp) +; X32-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero ; X32-NEXT: movsd %xmm0, (%ecx) +; X32-NEXT: movl %ebp, %esp +; X32-NEXT: popl %ebp +; X32-NEXT: retl +; +; X32AVX-LABEL: PR23476: +; X32AVX: # %bb.0: +; X32AVX-NEXT: pushl %ebp +; X32AVX-NEXT: movl %esp, %ebp +; X32AVX-NEXT: andl $-64, %esp +; X32AVX-NEXT: subl $128, %esp +; X32AVX-NEXT: movl 52(%ebp), %eax +; X32AVX-NEXT: andl $7, %eax +; X32AVX-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero +; X32AVX-NEXT: movl 48(%ebp), %ecx +; X32AVX-NEXT: vmovups 8(%ebp), %ymm1 +; X32AVX-NEXT: vmovaps %ymm1, (%esp) +; X32AVX-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) +; X32AVX-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero +; X32AVX-NEXT: vmovsd %xmm0, (%ecx) +; X32AVX-NEXT: movl %ebp, %esp +; X32AVX-NEXT: popl %ebp +; X32AVX-NEXT: vzeroupper +; X32AVX-NEXT: retl %ext = extractelement <5 x i64> %in, i32 %index store i64 %ext, i64* %out, align 8 ret void } + |