From: Bruno Cardoso Lopes Date: Thu, 11 Aug 2011 21:50:35 +0000 (+0000) Subject: Fix the test added by Nadav in r137308. Make it more strict: X-Git-Url: http://demsky.eecs.uci.edu/git/?a=commitdiff_plain;h=ec91640997f1df46fce723dfa630ae6942a1dd00;p=oota-llvm.git Fix the test added by Nadav in r137308. Make it more strict: 1) check for the "v" version of movaps 2) add a couple of CHECK-NOT to guarantee the behavior 3) move to a more appropriate test file git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@137361 91177308-0d34-0410-b5e6-96231b3b80d8 --- diff --git a/test/CodeGen/X86/avx-insert.ll b/test/CodeGen/X86/avx-insert.ll deleted file mode 100644 index d9eae03eee7..00000000000 --- a/test/CodeGen/X86/avx-insert.ll +++ /dev/null @@ -1,17 +0,0 @@ -; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7-avx -mattr=+avx | FileCheck %s - -; It is faster to make two saves, if the data is already in XMM registers. For -; example, after making an integer operation. -define void @double_save(<4 x i32>* %Ap, <4 x i32>* %Bp, <8 x i32>* %P) nounwind ssp { -entry: - ; CHECK: movaps - ; CHECK: movaps - ; CHECK: movaps - ; CHECK: movaps - %A = load <4 x i32>* %Ap - %B = load <4 x i32>* %Bp - %Z = shufflevector <4 x i32>%A, <4 x i32>%B, <8 x i32> - store <8 x i32> %Z, <8 x i32>* %P, align 16 - ret void -} - diff --git a/test/CodeGen/X86/avx-load-store.ll b/test/CodeGen/X86/avx-load-store.ll index d191af88638..f70291b96e1 100644 --- a/test/CodeGen/X86/avx-load-store.ll +++ b/test/CodeGen/X86/avx-load-store.ll @@ -64,3 +64,17 @@ define void @storev32i8_01(<32 x i8> %a) nounwind { unreachable } +; It is faster to make two saves, if the data is already in XMM registers. For +; example, after making an integer operation. +; CHECK: _double_save +; CHECK-NOT: vinsertf128 $1 +; CHECK-NOT: vinsertf128 $0 +; CHECK: vmovaps %xmm +; CHECK: vmovaps %xmm +define void @double_save(<4 x i32> %A, <4 x i32> %B, <8 x i32>* %P) nounwind ssp { +entry: + %Z = shufflevector <4 x i32>%A, <4 x i32>%B, <8 x i32> + store <8 x i32> %Z, <8 x i32>* %P, align 16 + ret void +} +