-; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=core-avx2 -mattr=+avx2 | FileCheck %s\r
-\r
-; CHECK: trunc4\r
-; CHECK: vpermd\r
-; CHECK-NOT: vinsert\r
-; CHECK: ret\r
-define <4 x i32> @trunc4(<4 x i64> %A) nounwind {\r
- %B = trunc <4 x i64> %A to <4 x i32>\r
- ret <4 x i32>%B\r
-}\r
-\r
-; CHECK: trunc8\r
-; CHECK: vpshufb\r
-; CHECK-NOT: vinsert\r
-; CHECK: ret\r
-\r
-define <8 x i16> @trunc8(<8 x i32> %A) nounwind {\r
- %B = trunc <8 x i32> %A to <8 x i16>\r
- ret <8 x i16>%B\r
-}\r
-\r
-; CHECK: sext4\r
-; CHECK: vpmovsxdq\r
-; CHECK-NOT: vinsert\r
-; CHECK: ret\r
-define <4 x i64> @sext4(<4 x i32> %A) nounwind {\r
- %B = sext <4 x i32> %A to <4 x i64>\r
- ret <4 x i64>%B\r
-}\r
-\r
-; CHECK: sext8\r
-; CHECK: vpmovsxwd\r
-; CHECK-NOT: vinsert\r
-; CHECK: ret\r
-define <8 x i32> @sext8(<8 x i16> %A) nounwind {\r
- %B = sext <8 x i16> %A to <8 x i32>\r
- ret <8 x i32>%B\r
-}\r
-\r
-; CHECK: zext4\r
-; CHECK: vpmovzxdq\r
-; CHECK-NOT: vinsert\r
-; CHECK: ret\r
-define <4 x i64> @zext4(<4 x i32> %A) nounwind {\r
- %B = zext <4 x i32> %A to <4 x i64>\r
- ret <4 x i64>%B\r
-}\r
-\r
-; CHECK: zext8\r
-; CHECK: vpmovzxwd\r
-; CHECK-NOT: vinsert\r
-; CHECK: ret\r
-define <8 x i32> @zext8(<8 x i16> %A) nounwind {\r
- %B = zext <8 x i16> %A to <8 x i32>\r
- ret <8 x i32>%B\r
-}\r
-; CHECK: zext_8i8_8i32\r
-; CHECK: vpmovzxwd\r
-; CHECK: vpand\r
-; CHECK: ret\r
-define <8 x i32> @zext_8i8_8i32(<8 x i8> %A) nounwind {\r
- %B = zext <8 x i8> %A to <8 x i32> \r
- ret <8 x i32>%B\r
-}\r
-\r
-\r
-\r
-\r
+; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=core-avx2 -mattr=+avx2 | FileCheck %s
+
+; CHECK: trunc4
+; CHECK: vpermd
+; CHECK-NOT: vinsert
+; CHECK: ret
+define <4 x i32> @trunc4(<4 x i64> %A) nounwind {
+ %B = trunc <4 x i64> %A to <4 x i32>
+ ret <4 x i32>%B
+}
+
+; CHECK: trunc8
+; CHECK: vpshufb
+; CHECK-NOT: vinsert
+; CHECK: ret
+
+define <8 x i16> @trunc8(<8 x i32> %A) nounwind {
+ %B = trunc <8 x i32> %A to <8 x i16>
+ ret <8 x i16>%B
+}
+
+; CHECK: sext4
+; CHECK: vpmovsxdq
+; CHECK-NOT: vinsert
+; CHECK: ret
+define <4 x i64> @sext4(<4 x i32> %A) nounwind {
+ %B = sext <4 x i32> %A to <4 x i64>
+ ret <4 x i64>%B
+}
+
+; CHECK: sext8
+; CHECK: vpmovsxwd
+; CHECK-NOT: vinsert
+; CHECK: ret
+define <8 x i32> @sext8(<8 x i16> %A) nounwind {
+ %B = sext <8 x i16> %A to <8 x i32>
+ ret <8 x i32>%B
+}
+
+; CHECK: zext4
+; CHECK: vpmovzxdq
+; CHECK-NOT: vinsert
+; CHECK: ret
+define <4 x i64> @zext4(<4 x i32> %A) nounwind {
+ %B = zext <4 x i32> %A to <4 x i64>
+ ret <4 x i64>%B
+}
+
+; CHECK: zext8
+; CHECK: vpmovzxwd
+; CHECK-NOT: vinsert
+; CHECK: ret
+define <8 x i32> @zext8(<8 x i16> %A) nounwind {
+ %B = zext <8 x i16> %A to <8 x i32>
+ ret <8 x i32>%B
+}
+; CHECK: zext_8i8_8i32
+; CHECK: vpmovzxwd
+; CHECK: vpand
+; CHECK: ret
+define <8 x i32> @zext_8i8_8i32(<8 x i8> %A) nounwind {
+ %B = zext <8 x i8> %A to <8 x i32>
+ ret <8 x i32>%B
+}
+
+
+
+