X-Git-Url: http://demsky.eecs.uci.edu/git/?a=blobdiff_plain;f=test%2FCodeGen%2FPowerPC%2Fvec_shuffle.ll;h=a942dd1c41c99812954982de9aa89191d03d4177;hb=77a0728d96eebf324044e7e5b60d2a244b5b6a96;hp=cfb7396f82f5a1ccbb8356645b04f4551e8fbf78;hpb=0ad91f516037066b6544f80e15977cc47698dfd8;p=oota-llvm.git diff --git a/test/CodeGen/PowerPC/vec_shuffle.ll b/test/CodeGen/PowerPC/vec_shuffle.ll index cfb7396f82f..a942dd1c41c 100644 --- a/test/CodeGen/PowerPC/vec_shuffle.ll +++ b/test/CodeGen/PowerPC/vec_shuffle.ll @@ -1,504 +1,504 @@ -; RUN: llvm-as < %s | opt -instcombine | llc -march=ppc32 -mcpu=g5 | not grep vperm && -; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 | grep vsldoi | wc -l | grep 2 && -; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 | grep vmrgh | wc -l | grep 7 && -; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 | grep vmrgl | wc -l | grep 6 && -; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 | grep vpkuhum | wc -l | grep 1 && -; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 | grep vpkuwum | wc -l | grep 1 +; RUN: opt < %s -instcombine | \ +; RUN: llc -march=ppc32 -mcpu=g5 | not grep vperm +; RUN: llc < %s -march=ppc32 -mcpu=g5 > %t +; RUN: grep vsldoi %t | count 2 +; RUN: grep vmrgh %t | count 7 +; RUN: grep vmrgl %t | count 6 +; RUN: grep vpkuhum %t | count 1 +; RUN: grep vpkuwum %t | count 1 -void %VSLDOI_xy(<8 x short>* %A, <8 x short>* %B) { +define void @VSLDOI_xy(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=1] - %tmp2 = load <8 x short>* %B ; <<8 x short>> [#uses=1] - %tmp = cast <8 x short> %tmp to <16 x sbyte> ; <<16 x sbyte>> [#uses=11] - %tmp2 = cast <8 x short> %tmp2 to <16 x sbyte> ; <<16 x sbyte>> [#uses=5] - %tmp = extractelement <16 x sbyte> %tmp, uint 5 ; [#uses=1] - %tmp3 = extractelement <16 x sbyte> %tmp, uint 6 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 7 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp, uint 8 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 9 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp, uint 10 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 11 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp, uint 12 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 13 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp, uint 14 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp, uint 15 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp2, uint 0 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp2, uint 1 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp2, uint 2 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp2, uint 3 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp2, uint 4 ; [#uses=1] - %tmp18 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp19 = insertelement <16 x sbyte> %tmp18, sbyte %tmp3, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 15 ; <<16 x sbyte>> [#uses=1] - %tmp33 = cast <16 x sbyte> %tmp33 to <8 x short> ; <<8 x short>> [#uses=1] - store <8 x short> %tmp33, <8 x short>* %A + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=1] + %tmp2 = load <8 x i16>, <8 x i16>* %B ; <<8 x i16>> [#uses=1] + %tmp.upgrd.1 = bitcast <8 x i16> %tmp to <16 x i8> ; <<16 x i8>> [#uses=11] + %tmp2.upgrd.2 = bitcast <8 x i16> %tmp2 to <16 x i8> ; <<16 x i8>> [#uses=5] + %tmp.upgrd.3 = extractelement <16 x i8> %tmp.upgrd.1, i32 5 ; [#uses=1] + %tmp3 = extractelement <16 x i8> %tmp.upgrd.1, i32 6 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp.upgrd.1, i32 7 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp.upgrd.1, i32 8 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp.upgrd.1, i32 9 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp.upgrd.1, i32 10 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp.upgrd.1, i32 11 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp.upgrd.1, i32 12 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp.upgrd.1, i32 13 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp.upgrd.1, i32 14 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp.upgrd.1, i32 15 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp2.upgrd.2, i32 0 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp2.upgrd.2, i32 1 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp2.upgrd.2, i32 2 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp2.upgrd.2, i32 3 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp2.upgrd.2, i32 4 ; [#uses=1] + %tmp18 = insertelement <16 x i8> undef, i8 %tmp.upgrd.3, i32 0 ; <<16 x i8>> [#uses=1] + %tmp19 = insertelement <16 x i8> %tmp18, i8 %tmp3, i32 1 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 2 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 3 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 4 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 5 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 6 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 7 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 8 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 9 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 10 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 11 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 12 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 13 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 14 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 15 ; <<16 x i8>> [#uses=1] + %tmp33.upgrd.4 = bitcast <16 x i8> %tmp33 to <8 x i16> ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp33.upgrd.4, <8 x i16>* %A ret void } -void %VSLDOI_xx(<8 x short>* %A, <8 x short>* %B) { - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=1] - %tmp2 = load <8 x short>* %A ; <<8 x short>> [#uses=1] - %tmp = cast <8 x short> %tmp to <16 x sbyte> ; <<16 x sbyte>> [#uses=11] - %tmp2 = cast <8 x short> %tmp2 to <16 x sbyte> ; <<16 x sbyte>> [#uses=5] - %tmp = extractelement <16 x sbyte> %tmp, uint 5 ; [#uses=1] - %tmp3 = extractelement <16 x sbyte> %tmp, uint 6 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 7 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp, uint 8 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 9 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp, uint 10 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 11 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp, uint 12 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 13 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp, uint 14 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp, uint 15 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp2, uint 0 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp2, uint 1 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp2, uint 2 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp2, uint 3 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp2, uint 4 ; [#uses=1] - %tmp18 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp19 = insertelement <16 x sbyte> %tmp18, sbyte %tmp3, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 15 ; <<16 x sbyte>> [#uses=1] - %tmp33 = cast <16 x sbyte> %tmp33 to <8 x short> ; <<8 x short>> [#uses=1] - store <8 x short> %tmp33, <8 x short>* %A +define void @VSLDOI_xx(<8 x i16>* %A, <8 x i16>* %B) { + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=1] + %tmp2 = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=1] + %tmp.upgrd.5 = bitcast <8 x i16> %tmp to <16 x i8> ; <<16 x i8>> [#uses=11] + %tmp2.upgrd.6 = bitcast <8 x i16> %tmp2 to <16 x i8> ; <<16 x i8>> [#uses=5] + %tmp.upgrd.7 = extractelement <16 x i8> %tmp.upgrd.5, i32 5 ; [#uses=1] + %tmp3 = extractelement <16 x i8> %tmp.upgrd.5, i32 6 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp.upgrd.5, i32 7 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp.upgrd.5, i32 8 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp.upgrd.5, i32 9 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp.upgrd.5, i32 10 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp.upgrd.5, i32 11 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp.upgrd.5, i32 12 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp.upgrd.5, i32 13 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp.upgrd.5, i32 14 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp.upgrd.5, i32 15 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp2.upgrd.6, i32 0 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp2.upgrd.6, i32 1 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp2.upgrd.6, i32 2 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp2.upgrd.6, i32 3 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp2.upgrd.6, i32 4 ; [#uses=1] + %tmp18 = insertelement <16 x i8> undef, i8 %tmp.upgrd.7, i32 0 ; <<16 x i8>> [#uses=1] + %tmp19 = insertelement <16 x i8> %tmp18, i8 %tmp3, i32 1 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 2 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 3 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 4 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 5 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 6 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 7 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 8 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 9 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 10 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 11 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 12 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 13 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 14 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 15 ; <<16 x i8>> [#uses=1] + %tmp33.upgrd.8 = bitcast <16 x i8> %tmp33 to <8 x i16> ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp33.upgrd.8, <8 x i16>* %A ret void } -void %VPERM_promote(<8 x short>* %A, <8 x short>* %B) { +define void @VPERM_promote(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=1] - %tmp = cast <8 x short> %tmp to <4 x int> ; <<4 x int>> [#uses=1] - %tmp2 = load <8 x short>* %B ; <<8 x short>> [#uses=1] - %tmp2 = cast <8 x short> %tmp2 to <4 x int> ; <<4 x int>> [#uses=1] - %tmp3 = call <4 x int> %llvm.ppc.altivec.vperm( <4 x int> %tmp, <4 x int> %tmp2, <16 x sbyte> < sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14, sbyte 14 > ) ; <<4 x int>> [#uses=1] - %tmp3 = cast <4 x int> %tmp3 to <8 x short> ; <<8 x short>> [#uses=1] - store <8 x short> %tmp3, <8 x short>* %A - ret void + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=1] + %tmp.upgrd.9 = bitcast <8 x i16> %tmp to <4 x i32> ; <<4 x i32>> [#uses=1] + %tmp2 = load <8 x i16>, <8 x i16>* %B ; <<8 x i16>> [#uses=1] + %tmp2.upgrd.10 = bitcast <8 x i16> %tmp2 to <4 x i32> ; <<4 x i32>> [#uses=1] + %tmp3 = call <4 x i32> @llvm.ppc.altivec.vperm( <4 x i32> %tmp.upgrd.9, <4 x i32> %tmp2.upgrd.10, <16 x i8> < i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14, i8 14 > ) ; <<4 x i32>> [#uses=1] + %tmp3.upgrd.11 = bitcast <4 x i32> %tmp3 to <8 x i16> ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp3.upgrd.11, <8 x i16>* %A + ret void } -declare <4 x int> %llvm.ppc.altivec.vperm(<4 x int>, <4 x int>, <16 x sbyte>) - +declare <4 x i32> @llvm.ppc.altivec.vperm(<4 x i32>, <4 x i32>, <16 x i8>) -void %tb_l(<16 x sbyte>* %A, <16 x sbyte>* %B) { +define void @tb_l(<16 x i8>* %A, <16 x i8>* %B) { entry: - %tmp = load <16 x sbyte>* %A ; <<16 x sbyte>> [#uses=8] - %tmp2 = load <16 x sbyte>* %B ; <<16 x sbyte>> [#uses=8] - %tmp = extractelement <16 x sbyte> %tmp, uint 8 ; [#uses=1] - %tmp3 = extractelement <16 x sbyte> %tmp2, uint 8 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 9 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp2, uint 9 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 10 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp2, uint 10 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 11 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp2, uint 11 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 12 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp2, uint 12 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp, uint 13 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp2, uint 13 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp, uint 14 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp2, uint 14 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp, uint 15 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp2, uint 15 ; [#uses=1] - %tmp18 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp19 = insertelement <16 x sbyte> %tmp18, sbyte %tmp3, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 15 ; <<16 x sbyte>> [#uses=1] - store <16 x sbyte> %tmp33, <16 x sbyte>* %A + %tmp = load <16 x i8>, <16 x i8>* %A ; <<16 x i8>> [#uses=8] + %tmp2 = load <16 x i8>, <16 x i8>* %B ; <<16 x i8>> [#uses=8] + %tmp.upgrd.12 = extractelement <16 x i8> %tmp, i32 8 ; [#uses=1] + %tmp3 = extractelement <16 x i8> %tmp2, i32 8 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp, i32 9 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp2, i32 9 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp, i32 10 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp2, i32 10 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp, i32 11 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp2, i32 11 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp, i32 12 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp2, i32 12 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp, i32 13 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp2, i32 13 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp, i32 14 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp2, i32 14 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp, i32 15 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp2, i32 15 ; [#uses=1] + %tmp18 = insertelement <16 x i8> undef, i8 %tmp.upgrd.12, i32 0 ; <<16 x i8>> [#uses=1] + %tmp19 = insertelement <16 x i8> %tmp18, i8 %tmp3, i32 1 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 2 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 3 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 4 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 5 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 6 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 7 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 8 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 9 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 10 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 11 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 12 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 13 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 14 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 15 ; <<16 x i8>> [#uses=1] + store <16 x i8> %tmp33, <16 x i8>* %A ret void } -void %th_l(<8 x short>* %A, <8 x short>* %B) { +define void @th_l(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=4] - %tmp2 = load <8 x short>* %B ; <<8 x short>> [#uses=4] - %tmp = extractelement <8 x short> %tmp, uint 4 ; [#uses=1] - %tmp3 = extractelement <8 x short> %tmp2, uint 4 ; [#uses=1] - %tmp4 = extractelement <8 x short> %tmp, uint 5 ; [#uses=1] - %tmp5 = extractelement <8 x short> %tmp2, uint 5 ; [#uses=1] - %tmp6 = extractelement <8 x short> %tmp, uint 6 ; [#uses=1] - %tmp7 = extractelement <8 x short> %tmp2, uint 6 ; [#uses=1] - %tmp8 = extractelement <8 x short> %tmp, uint 7 ; [#uses=1] - %tmp9 = extractelement <8 x short> %tmp2, uint 7 ; [#uses=1] - %tmp10 = insertelement <8 x short> undef, short %tmp, uint 0 ; <<8 x short>> [#uses=1] - %tmp11 = insertelement <8 x short> %tmp10, short %tmp3, uint 1 ; <<8 x short>> [#uses=1] - %tmp12 = insertelement <8 x short> %tmp11, short %tmp4, uint 2 ; <<8 x short>> [#uses=1] - %tmp13 = insertelement <8 x short> %tmp12, short %tmp5, uint 3 ; <<8 x short>> [#uses=1] - %tmp14 = insertelement <8 x short> %tmp13, short %tmp6, uint 4 ; <<8 x short>> [#uses=1] - %tmp15 = insertelement <8 x short> %tmp14, short %tmp7, uint 5 ; <<8 x short>> [#uses=1] - %tmp16 = insertelement <8 x short> %tmp15, short %tmp8, uint 6 ; <<8 x short>> [#uses=1] - %tmp17 = insertelement <8 x short> %tmp16, short %tmp9, uint 7 ; <<8 x short>> [#uses=1] - store <8 x short> %tmp17, <8 x short>* %A + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=4] + %tmp2 = load <8 x i16>, <8 x i16>* %B ; <<8 x i16>> [#uses=4] + %tmp.upgrd.13 = extractelement <8 x i16> %tmp, i32 4 ; [#uses=1] + %tmp3 = extractelement <8 x i16> %tmp2, i32 4 ; [#uses=1] + %tmp4 = extractelement <8 x i16> %tmp, i32 5 ; [#uses=1] + %tmp5 = extractelement <8 x i16> %tmp2, i32 5 ; [#uses=1] + %tmp6 = extractelement <8 x i16> %tmp, i32 6 ; [#uses=1] + %tmp7 = extractelement <8 x i16> %tmp2, i32 6 ; [#uses=1] + %tmp8 = extractelement <8 x i16> %tmp, i32 7 ; [#uses=1] + %tmp9 = extractelement <8 x i16> %tmp2, i32 7 ; [#uses=1] + %tmp10 = insertelement <8 x i16> undef, i16 %tmp.upgrd.13, i32 0 ; <<8 x i16>> [#uses=1] + %tmp11 = insertelement <8 x i16> %tmp10, i16 %tmp3, i32 1 ; <<8 x i16>> [#uses=1] + %tmp12 = insertelement <8 x i16> %tmp11, i16 %tmp4, i32 2 ; <<8 x i16>> [#uses=1] + %tmp13 = insertelement <8 x i16> %tmp12, i16 %tmp5, i32 3 ; <<8 x i16>> [#uses=1] + %tmp14 = insertelement <8 x i16> %tmp13, i16 %tmp6, i32 4 ; <<8 x i16>> [#uses=1] + %tmp15 = insertelement <8 x i16> %tmp14, i16 %tmp7, i32 5 ; <<8 x i16>> [#uses=1] + %tmp16 = insertelement <8 x i16> %tmp15, i16 %tmp8, i32 6 ; <<8 x i16>> [#uses=1] + %tmp17 = insertelement <8 x i16> %tmp16, i16 %tmp9, i32 7 ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp17, <8 x i16>* %A ret void } -void %tw_l(<4 x int>* %A, <4 x int>* %B) { +define void @tw_l(<4 x i32>* %A, <4 x i32>* %B) { entry: - %tmp = load <4 x int>* %A ; <<4 x int>> [#uses=2] - %tmp2 = load <4 x int>* %B ; <<4 x int>> [#uses=2] - %tmp = extractelement <4 x int> %tmp, uint 2 ; [#uses=1] - %tmp3 = extractelement <4 x int> %tmp2, uint 2 ; [#uses=1] - %tmp4 = extractelement <4 x int> %tmp, uint 3 ; [#uses=1] - %tmp5 = extractelement <4 x int> %tmp2, uint 3 ; [#uses=1] - %tmp6 = insertelement <4 x int> undef, int %tmp, uint 0 ; <<4 x int>> [#uses=1] - %tmp7 = insertelement <4 x int> %tmp6, int %tmp3, uint 1 ; <<4 x int>> [#uses=1] - %tmp8 = insertelement <4 x int> %tmp7, int %tmp4, uint 2 ; <<4 x int>> [#uses=1] - %tmp9 = insertelement <4 x int> %tmp8, int %tmp5, uint 3 ; <<4 x int>> [#uses=1] - store <4 x int> %tmp9, <4 x int>* %A + %tmp = load <4 x i32>, <4 x i32>* %A ; <<4 x i32>> [#uses=2] + %tmp2 = load <4 x i32>, <4 x i32>* %B ; <<4 x i32>> [#uses=2] + %tmp.upgrd.14 = extractelement <4 x i32> %tmp, i32 2 ; [#uses=1] + %tmp3 = extractelement <4 x i32> %tmp2, i32 2 ; [#uses=1] + %tmp4 = extractelement <4 x i32> %tmp, i32 3 ; [#uses=1] + %tmp5 = extractelement <4 x i32> %tmp2, i32 3 ; [#uses=1] + %tmp6 = insertelement <4 x i32> undef, i32 %tmp.upgrd.14, i32 0 ; <<4 x i32>> [#uses=1] + %tmp7 = insertelement <4 x i32> %tmp6, i32 %tmp3, i32 1 ; <<4 x i32>> [#uses=1] + %tmp8 = insertelement <4 x i32> %tmp7, i32 %tmp4, i32 2 ; <<4 x i32>> [#uses=1] + %tmp9 = insertelement <4 x i32> %tmp8, i32 %tmp5, i32 3 ; <<4 x i32>> [#uses=1] + store <4 x i32> %tmp9, <4 x i32>* %A ret void } -void %tb_h(<16 x sbyte>* %A, <16 x sbyte>* %B) { +define void @tb_h(<16 x i8>* %A, <16 x i8>* %B) { entry: - %tmp = load <16 x sbyte>* %A ; <<16 x sbyte>> [#uses=8] - %tmp2 = load <16 x sbyte>* %B ; <<16 x sbyte>> [#uses=8] - %tmp = extractelement <16 x sbyte> %tmp, uint 0 ; [#uses=1] - %tmp3 = extractelement <16 x sbyte> %tmp2, uint 0 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 1 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp2, uint 1 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 2 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp2, uint 2 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 3 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp2, uint 3 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 4 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp2, uint 4 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp, uint 5 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp2, uint 5 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp, uint 6 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp2, uint 6 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp, uint 7 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp2, uint 7 ; [#uses=1] - %tmp18 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp19 = insertelement <16 x sbyte> %tmp18, sbyte %tmp3, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 15 ; <<16 x sbyte>> [#uses=1] - store <16 x sbyte> %tmp33, <16 x sbyte>* %A + %tmp = load <16 x i8>, <16 x i8>* %A ; <<16 x i8>> [#uses=8] + %tmp2 = load <16 x i8>, <16 x i8>* %B ; <<16 x i8>> [#uses=8] + %tmp.upgrd.15 = extractelement <16 x i8> %tmp, i32 0 ; [#uses=1] + %tmp3 = extractelement <16 x i8> %tmp2, i32 0 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp, i32 1 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp2, i32 1 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp, i32 2 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp2, i32 2 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp, i32 3 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp2, i32 3 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp, i32 4 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp2, i32 4 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp, i32 5 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp2, i32 5 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp, i32 6 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp2, i32 6 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp, i32 7 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp2, i32 7 ; [#uses=1] + %tmp18 = insertelement <16 x i8> undef, i8 %tmp.upgrd.15, i32 0 ; <<16 x i8>> [#uses=1] + %tmp19 = insertelement <16 x i8> %tmp18, i8 %tmp3, i32 1 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 2 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 3 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 4 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 5 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 6 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 7 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 8 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 9 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 10 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 11 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 12 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 13 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 14 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 15 ; <<16 x i8>> [#uses=1] + store <16 x i8> %tmp33, <16 x i8>* %A ret void } -void %th_h(<8 x short>* %A, <8 x short>* %B) { +define void @th_h(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=4] - %tmp2 = load <8 x short>* %B ; <<8 x short>> [#uses=4] - %tmp = extractelement <8 x short> %tmp, uint 0 ; [#uses=1] - %tmp3 = extractelement <8 x short> %tmp2, uint 0 ; [#uses=1] - %tmp4 = extractelement <8 x short> %tmp, uint 1 ; [#uses=1] - %tmp5 = extractelement <8 x short> %tmp2, uint 1 ; [#uses=1] - %tmp6 = extractelement <8 x short> %tmp, uint 2 ; [#uses=1] - %tmp7 = extractelement <8 x short> %tmp2, uint 2 ; [#uses=1] - %tmp8 = extractelement <8 x short> %tmp, uint 3 ; [#uses=1] - %tmp9 = extractelement <8 x short> %tmp2, uint 3 ; [#uses=1] - %tmp10 = insertelement <8 x short> undef, short %tmp, uint 0 ; <<8 x short>> [#uses=1] - %tmp11 = insertelement <8 x short> %tmp10, short %tmp3, uint 1 ; <<8 x short>> [#uses=1] - %tmp12 = insertelement <8 x short> %tmp11, short %tmp4, uint 2 ; <<8 x short>> [#uses=1] - %tmp13 = insertelement <8 x short> %tmp12, short %tmp5, uint 3 ; <<8 x short>> [#uses=1] - %tmp14 = insertelement <8 x short> %tmp13, short %tmp6, uint 4 ; <<8 x short>> [#uses=1] - %tmp15 = insertelement <8 x short> %tmp14, short %tmp7, uint 5 ; <<8 x short>> [#uses=1] - %tmp16 = insertelement <8 x short> %tmp15, short %tmp8, uint 6 ; <<8 x short>> [#uses=1] - %tmp17 = insertelement <8 x short> %tmp16, short %tmp9, uint 7 ; <<8 x short>> [#uses=1] - store <8 x short> %tmp17, <8 x short>* %A + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=4] + %tmp2 = load <8 x i16>, <8 x i16>* %B ; <<8 x i16>> [#uses=4] + %tmp.upgrd.16 = extractelement <8 x i16> %tmp, i32 0 ; [#uses=1] + %tmp3 = extractelement <8 x i16> %tmp2, i32 0 ; [#uses=1] + %tmp4 = extractelement <8 x i16> %tmp, i32 1 ; [#uses=1] + %tmp5 = extractelement <8 x i16> %tmp2, i32 1 ; [#uses=1] + %tmp6 = extractelement <8 x i16> %tmp, i32 2 ; [#uses=1] + %tmp7 = extractelement <8 x i16> %tmp2, i32 2 ; [#uses=1] + %tmp8 = extractelement <8 x i16> %tmp, i32 3 ; [#uses=1] + %tmp9 = extractelement <8 x i16> %tmp2, i32 3 ; [#uses=1] + %tmp10 = insertelement <8 x i16> undef, i16 %tmp.upgrd.16, i32 0 ; <<8 x i16>> [#uses=1] + %tmp11 = insertelement <8 x i16> %tmp10, i16 %tmp3, i32 1 ; <<8 x i16>> [#uses=1] + %tmp12 = insertelement <8 x i16> %tmp11, i16 %tmp4, i32 2 ; <<8 x i16>> [#uses=1] + %tmp13 = insertelement <8 x i16> %tmp12, i16 %tmp5, i32 3 ; <<8 x i16>> [#uses=1] + %tmp14 = insertelement <8 x i16> %tmp13, i16 %tmp6, i32 4 ; <<8 x i16>> [#uses=1] + %tmp15 = insertelement <8 x i16> %tmp14, i16 %tmp7, i32 5 ; <<8 x i16>> [#uses=1] + %tmp16 = insertelement <8 x i16> %tmp15, i16 %tmp8, i32 6 ; <<8 x i16>> [#uses=1] + %tmp17 = insertelement <8 x i16> %tmp16, i16 %tmp9, i32 7 ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp17, <8 x i16>* %A ret void } -void %tw_h(<4 x int>* %A, <4 x int>* %B) { +define void @tw_h(<4 x i32>* %A, <4 x i32>* %B) { entry: - %tmp = load <4 x int>* %A ; <<4 x int>> [#uses=2] - %tmp2 = load <4 x int>* %B ; <<4 x int>> [#uses=2] - %tmp = extractelement <4 x int> %tmp2, uint 0 ; [#uses=1] - %tmp3 = extractelement <4 x int> %tmp, uint 0 ; [#uses=1] - %tmp4 = extractelement <4 x int> %tmp2, uint 1 ; [#uses=1] - %tmp5 = extractelement <4 x int> %tmp, uint 1 ; [#uses=1] - %tmp6 = insertelement <4 x int> undef, int %tmp, uint 0 ; <<4 x int>> [#uses=1] - %tmp7 = insertelement <4 x int> %tmp6, int %tmp3, uint 1 ; <<4 x int>> [#uses=1] - %tmp8 = insertelement <4 x int> %tmp7, int %tmp4, uint 2 ; <<4 x int>> [#uses=1] - %tmp9 = insertelement <4 x int> %tmp8, int %tmp5, uint 3 ; <<4 x int>> [#uses=1] - store <4 x int> %tmp9, <4 x int>* %A + %tmp = load <4 x i32>, <4 x i32>* %A ; <<4 x i32>> [#uses=2] + %tmp2 = load <4 x i32>, <4 x i32>* %B ; <<4 x i32>> [#uses=2] + %tmp.upgrd.17 = extractelement <4 x i32> %tmp2, i32 0 ; [#uses=1] + %tmp3 = extractelement <4 x i32> %tmp, i32 0 ; [#uses=1] + %tmp4 = extractelement <4 x i32> %tmp2, i32 1 ; [#uses=1] + %tmp5 = extractelement <4 x i32> %tmp, i32 1 ; [#uses=1] + %tmp6 = insertelement <4 x i32> undef, i32 %tmp.upgrd.17, i32 0 ; <<4 x i32>> [#uses=1] + %tmp7 = insertelement <4 x i32> %tmp6, i32 %tmp3, i32 1 ; <<4 x i32>> [#uses=1] + %tmp8 = insertelement <4 x i32> %tmp7, i32 %tmp4, i32 2 ; <<4 x i32>> [#uses=1] + %tmp9 = insertelement <4 x i32> %tmp8, i32 %tmp5, i32 3 ; <<4 x i32>> [#uses=1] + store <4 x i32> %tmp9, <4 x i32>* %A ret void } -void %tw_h_flop(<4 x int>* %A, <4 x int>* %B) { - %tmp = load <4 x int>* %A ; <<4 x int>> [#uses=2] - %tmp2 = load <4 x int>* %B ; <<4 x int>> [#uses=2] - %tmp = extractelement <4 x int> %tmp, uint 0 ; [#uses=1] - %tmp3 = extractelement <4 x int> %tmp2, uint 0 ; [#uses=1] - %tmp4 = extractelement <4 x int> %tmp, uint 1 ; [#uses=1] - %tmp5 = extractelement <4 x int> %tmp2, uint 1 ; [#uses=1] - %tmp6 = insertelement <4 x int> undef, int %tmp, uint 0 ; <<4 x int>> [#uses=1] - %tmp7 = insertelement <4 x int> %tmp6, int %tmp3, uint 1 ; <<4 x int>> [#uses=1] - %tmp8 = insertelement <4 x int> %tmp7, int %tmp4, uint 2 ; <<4 x int>> [#uses=1] - %tmp9 = insertelement <4 x int> %tmp8, int %tmp5, uint 3 ; <<4 x int>> [#uses=1] - store <4 x int> %tmp9, <4 x int>* %A +define void @tw_h_flop(<4 x i32>* %A, <4 x i32>* %B) { + %tmp = load <4 x i32>, <4 x i32>* %A ; <<4 x i32>> [#uses=2] + %tmp2 = load <4 x i32>, <4 x i32>* %B ; <<4 x i32>> [#uses=2] + %tmp.upgrd.18 = extractelement <4 x i32> %tmp, i32 0 ; [#uses=1] + %tmp3 = extractelement <4 x i32> %tmp2, i32 0 ; [#uses=1] + %tmp4 = extractelement <4 x i32> %tmp, i32 1 ; [#uses=1] + %tmp5 = extractelement <4 x i32> %tmp2, i32 1 ; [#uses=1] + %tmp6 = insertelement <4 x i32> undef, i32 %tmp.upgrd.18, i32 0 ; <<4 x i32>> [#uses=1] + %tmp7 = insertelement <4 x i32> %tmp6, i32 %tmp3, i32 1 ; <<4 x i32>> [#uses=1] + %tmp8 = insertelement <4 x i32> %tmp7, i32 %tmp4, i32 2 ; <<4 x i32>> [#uses=1] + %tmp9 = insertelement <4 x i32> %tmp8, i32 %tmp5, i32 3 ; <<4 x i32>> [#uses=1] + store <4 x i32> %tmp9, <4 x i32>* %A ret void } - -void %VMRG_UNARY_tb_l(<16 x sbyte>* %A, <16 x sbyte>* %B) { +define void @VMRG_UNARY_tb_l(<16 x i8>* %A, <16 x i8>* %B) { entry: - %tmp = load <16 x sbyte>* %A ; <<16 x sbyte>> [#uses=16] - %tmp = extractelement <16 x sbyte> %tmp, uint 8 ; [#uses=1] - %tmp3 = extractelement <16 x sbyte> %tmp, uint 8 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 9 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp, uint 9 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 10 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp, uint 10 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 11 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp, uint 11 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 12 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp, uint 12 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp, uint 13 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp, uint 13 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp, uint 14 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp, uint 14 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp, uint 15 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp, uint 15 ; [#uses=1] - %tmp18 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp19 = insertelement <16 x sbyte> %tmp18, sbyte %tmp3, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 15 ; <<16 x sbyte>> [#uses=1] - store <16 x sbyte> %tmp33, <16 x sbyte>* %A + %tmp = load <16 x i8>, <16 x i8>* %A ; <<16 x i8>> [#uses=16] + %tmp.upgrd.19 = extractelement <16 x i8> %tmp, i32 8 ; [#uses=1] + %tmp3 = extractelement <16 x i8> %tmp, i32 8 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp, i32 9 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp, i32 9 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp, i32 10 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp, i32 10 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp, i32 11 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp, i32 11 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp, i32 12 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp, i32 12 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp, i32 13 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp, i32 13 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp, i32 14 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp, i32 14 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp, i32 15 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp, i32 15 ; [#uses=1] + %tmp18 = insertelement <16 x i8> undef, i8 %tmp.upgrd.19, i32 0 ; <<16 x i8>> [#uses=1] + %tmp19 = insertelement <16 x i8> %tmp18, i8 %tmp3, i32 1 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 2 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 3 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 4 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 5 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 6 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 7 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 8 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 9 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 10 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 11 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 12 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 13 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 14 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 15 ; <<16 x i8>> [#uses=1] + store <16 x i8> %tmp33, <16 x i8>* %A ret void } -void %VMRG_UNARY_th_l(<8 x short>* %A, <8 x short>* %B) { +define void @VMRG_UNARY_th_l(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=8] - %tmp = extractelement <8 x short> %tmp, uint 4 ; [#uses=1] - %tmp3 = extractelement <8 x short> %tmp, uint 4 ; [#uses=1] - %tmp4 = extractelement <8 x short> %tmp, uint 5 ; [#uses=1] - %tmp5 = extractelement <8 x short> %tmp, uint 5 ; [#uses=1] - %tmp6 = extractelement <8 x short> %tmp, uint 6 ; [#uses=1] - %tmp7 = extractelement <8 x short> %tmp, uint 6 ; [#uses=1] - %tmp8 = extractelement <8 x short> %tmp, uint 7 ; [#uses=1] - %tmp9 = extractelement <8 x short> %tmp, uint 7 ; [#uses=1] - %tmp10 = insertelement <8 x short> undef, short %tmp, uint 0 ; <<8 x short>> [#uses=1] - %tmp11 = insertelement <8 x short> %tmp10, short %tmp3, uint 1 ; <<8 x short>> [#uses=1] - %tmp12 = insertelement <8 x short> %tmp11, short %tmp4, uint 2 ; <<8 x short>> [#uses=1] - %tmp13 = insertelement <8 x short> %tmp12, short %tmp5, uint 3 ; <<8 x short>> [#uses=1] - %tmp14 = insertelement <8 x short> %tmp13, short %tmp6, uint 4 ; <<8 x short>> [#uses=1] - %tmp15 = insertelement <8 x short> %tmp14, short %tmp7, uint 5 ; <<8 x short>> [#uses=1] - %tmp16 = insertelement <8 x short> %tmp15, short %tmp8, uint 6 ; <<8 x short>> [#uses=1] - %tmp17 = insertelement <8 x short> %tmp16, short %tmp9, uint 7 ; <<8 x short>> [#uses=1] - store <8 x short> %tmp17, <8 x short>* %A + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=8] + %tmp.upgrd.20 = extractelement <8 x i16> %tmp, i32 4 ; [#uses=1] + %tmp3 = extractelement <8 x i16> %tmp, i32 4 ; [#uses=1] + %tmp4 = extractelement <8 x i16> %tmp, i32 5 ; [#uses=1] + %tmp5 = extractelement <8 x i16> %tmp, i32 5 ; [#uses=1] + %tmp6 = extractelement <8 x i16> %tmp, i32 6 ; [#uses=1] + %tmp7 = extractelement <8 x i16> %tmp, i32 6 ; [#uses=1] + %tmp8 = extractelement <8 x i16> %tmp, i32 7 ; [#uses=1] + %tmp9 = extractelement <8 x i16> %tmp, i32 7 ; [#uses=1] + %tmp10 = insertelement <8 x i16> undef, i16 %tmp.upgrd.20, i32 0 ; <<8 x i16>> [#uses=1] + %tmp11 = insertelement <8 x i16> %tmp10, i16 %tmp3, i32 1 ; <<8 x i16>> [#uses=1] + %tmp12 = insertelement <8 x i16> %tmp11, i16 %tmp4, i32 2 ; <<8 x i16>> [#uses=1] + %tmp13 = insertelement <8 x i16> %tmp12, i16 %tmp5, i32 3 ; <<8 x i16>> [#uses=1] + %tmp14 = insertelement <8 x i16> %tmp13, i16 %tmp6, i32 4 ; <<8 x i16>> [#uses=1] + %tmp15 = insertelement <8 x i16> %tmp14, i16 %tmp7, i32 5 ; <<8 x i16>> [#uses=1] + %tmp16 = insertelement <8 x i16> %tmp15, i16 %tmp8, i32 6 ; <<8 x i16>> [#uses=1] + %tmp17 = insertelement <8 x i16> %tmp16, i16 %tmp9, i32 7 ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp17, <8 x i16>* %A ret void } -void %VMRG_UNARY_tw_l(<4 x int>* %A, <4 x int>* %B) { +define void @VMRG_UNARY_tw_l(<4 x i32>* %A, <4 x i32>* %B) { entry: - %tmp = load <4 x int>* %A ; <<4 x int>> [#uses=4] - %tmp = extractelement <4 x int> %tmp, uint 2 ; [#uses=1] - %tmp3 = extractelement <4 x int> %tmp, uint 2 ; [#uses=1] - %tmp4 = extractelement <4 x int> %tmp, uint 3 ; [#uses=1] - %tmp5 = extractelement <4 x int> %tmp, uint 3 ; [#uses=1] - %tmp6 = insertelement <4 x int> undef, int %tmp, uint 0 ; <<4 x int>> [#uses=1] - %tmp7 = insertelement <4 x int> %tmp6, int %tmp3, uint 1 ; <<4 x int>> [#uses=1] - %tmp8 = insertelement <4 x int> %tmp7, int %tmp4, uint 2 ; <<4 x int>> [#uses=1] - %tmp9 = insertelement <4 x int> %tmp8, int %tmp5, uint 3 ; <<4 x int>> [#uses=1] - store <4 x int> %tmp9, <4 x int>* %A + %tmp = load <4 x i32>, <4 x i32>* %A ; <<4 x i32>> [#uses=4] + %tmp.upgrd.21 = extractelement <4 x i32> %tmp, i32 2 ; [#uses=1] + %tmp3 = extractelement <4 x i32> %tmp, i32 2 ; [#uses=1] + %tmp4 = extractelement <4 x i32> %tmp, i32 3 ; [#uses=1] + %tmp5 = extractelement <4 x i32> %tmp, i32 3 ; [#uses=1] + %tmp6 = insertelement <4 x i32> undef, i32 %tmp.upgrd.21, i32 0 ; <<4 x i32>> [#uses=1] + %tmp7 = insertelement <4 x i32> %tmp6, i32 %tmp3, i32 1 ; <<4 x i32>> [#uses=1] + %tmp8 = insertelement <4 x i32> %tmp7, i32 %tmp4, i32 2 ; <<4 x i32>> [#uses=1] + %tmp9 = insertelement <4 x i32> %tmp8, i32 %tmp5, i32 3 ; <<4 x i32>> [#uses=1] + store <4 x i32> %tmp9, <4 x i32>* %A ret void } -void %VMRG_UNARY_tb_h(<16 x sbyte>* %A, <16 x sbyte>* %B) { +define void @VMRG_UNARY_tb_h(<16 x i8>* %A, <16 x i8>* %B) { entry: - %tmp = load <16 x sbyte>* %A ; <<16 x sbyte>> [#uses=16] - %tmp = extractelement <16 x sbyte> %tmp, uint 0 ; [#uses=1] - %tmp3 = extractelement <16 x sbyte> %tmp, uint 0 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 1 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp, uint 1 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 2 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp, uint 2 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 3 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp, uint 3 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 4 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp, uint 4 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp, uint 5 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp, uint 5 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp, uint 6 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp, uint 6 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp, uint 7 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp, uint 7 ; [#uses=1] - %tmp18 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp19 = insertelement <16 x sbyte> %tmp18, sbyte %tmp3, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 15 ; <<16 x sbyte>> [#uses=1] - store <16 x sbyte> %tmp33, <16 x sbyte>* %A + %tmp = load <16 x i8>, <16 x i8>* %A ; <<16 x i8>> [#uses=16] + %tmp.upgrd.22 = extractelement <16 x i8> %tmp, i32 0 ; [#uses=1] + %tmp3 = extractelement <16 x i8> %tmp, i32 0 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp, i32 1 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp, i32 1 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp, i32 2 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp, i32 2 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp, i32 3 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp, i32 3 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp, i32 4 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp, i32 4 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp, i32 5 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp, i32 5 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp, i32 6 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp, i32 6 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp, i32 7 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp, i32 7 ; [#uses=1] + %tmp18 = insertelement <16 x i8> undef, i8 %tmp.upgrd.22, i32 0 ; <<16 x i8>> [#uses=1] + %tmp19 = insertelement <16 x i8> %tmp18, i8 %tmp3, i32 1 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 2 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 3 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 4 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 5 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 6 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 7 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 8 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 9 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 10 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 11 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 12 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 13 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 14 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 15 ; <<16 x i8>> [#uses=1] + store <16 x i8> %tmp33, <16 x i8>* %A ret void } -void %VMRG_UNARY_th_h(<8 x short>* %A, <8 x short>* %B) { +define void @VMRG_UNARY_th_h(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=8] - %tmp = extractelement <8 x short> %tmp, uint 0 ; [#uses=1] - %tmp3 = extractelement <8 x short> %tmp, uint 0 ; [#uses=1] - %tmp4 = extractelement <8 x short> %tmp, uint 1 ; [#uses=1] - %tmp5 = extractelement <8 x short> %tmp, uint 1 ; [#uses=1] - %tmp6 = extractelement <8 x short> %tmp, uint 2 ; [#uses=1] - %tmp7 = extractelement <8 x short> %tmp, uint 2 ; [#uses=1] - %tmp8 = extractelement <8 x short> %tmp, uint 3 ; [#uses=1] - %tmp9 = extractelement <8 x short> %tmp, uint 3 ; [#uses=1] - %tmp10 = insertelement <8 x short> undef, short %tmp, uint 0 ; <<8 x short>> [#uses=1] - %tmp11 = insertelement <8 x short> %tmp10, short %tmp3, uint 1 ; <<8 x short>> [#uses=1] - %tmp12 = insertelement <8 x short> %tmp11, short %tmp4, uint 2 ; <<8 x short>> [#uses=1] - %tmp13 = insertelement <8 x short> %tmp12, short %tmp5, uint 3 ; <<8 x short>> [#uses=1] - %tmp14 = insertelement <8 x short> %tmp13, short %tmp6, uint 4 ; <<8 x short>> [#uses=1] - %tmp15 = insertelement <8 x short> %tmp14, short %tmp7, uint 5 ; <<8 x short>> [#uses=1] - %tmp16 = insertelement <8 x short> %tmp15, short %tmp8, uint 6 ; <<8 x short>> [#uses=1] - %tmp17 = insertelement <8 x short> %tmp16, short %tmp9, uint 7 ; <<8 x short>> [#uses=1] - store <8 x short> %tmp17, <8 x short>* %A + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=8] + %tmp.upgrd.23 = extractelement <8 x i16> %tmp, i32 0 ; [#uses=1] + %tmp3 = extractelement <8 x i16> %tmp, i32 0 ; [#uses=1] + %tmp4 = extractelement <8 x i16> %tmp, i32 1 ; [#uses=1] + %tmp5 = extractelement <8 x i16> %tmp, i32 1 ; [#uses=1] + %tmp6 = extractelement <8 x i16> %tmp, i32 2 ; [#uses=1] + %tmp7 = extractelement <8 x i16> %tmp, i32 2 ; [#uses=1] + %tmp8 = extractelement <8 x i16> %tmp, i32 3 ; [#uses=1] + %tmp9 = extractelement <8 x i16> %tmp, i32 3 ; [#uses=1] + %tmp10 = insertelement <8 x i16> undef, i16 %tmp.upgrd.23, i32 0 ; <<8 x i16>> [#uses=1] + %tmp11 = insertelement <8 x i16> %tmp10, i16 %tmp3, i32 1 ; <<8 x i16>> [#uses=1] + %tmp12 = insertelement <8 x i16> %tmp11, i16 %tmp4, i32 2 ; <<8 x i16>> [#uses=1] + %tmp13 = insertelement <8 x i16> %tmp12, i16 %tmp5, i32 3 ; <<8 x i16>> [#uses=1] + %tmp14 = insertelement <8 x i16> %tmp13, i16 %tmp6, i32 4 ; <<8 x i16>> [#uses=1] + %tmp15 = insertelement <8 x i16> %tmp14, i16 %tmp7, i32 5 ; <<8 x i16>> [#uses=1] + %tmp16 = insertelement <8 x i16> %tmp15, i16 %tmp8, i32 6 ; <<8 x i16>> [#uses=1] + %tmp17 = insertelement <8 x i16> %tmp16, i16 %tmp9, i32 7 ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp17, <8 x i16>* %A ret void } -void %VMRG_UNARY_tw_h(<4 x int>* %A, <4 x int>* %B) { +define void @VMRG_UNARY_tw_h(<4 x i32>* %A, <4 x i32>* %B) { entry: - %tmp = load <4 x int>* %A ; <<4 x int>> [#uses=4] - %tmp = extractelement <4 x int> %tmp, uint 0 ; [#uses=1] - %tmp3 = extractelement <4 x int> %tmp, uint 0 ; [#uses=1] - %tmp4 = extractelement <4 x int> %tmp, uint 1 ; [#uses=1] - %tmp5 = extractelement <4 x int> %tmp, uint 1 ; [#uses=1] - %tmp6 = insertelement <4 x int> undef, int %tmp, uint 0 ; <<4 x int>> [#uses=1] - %tmp7 = insertelement <4 x int> %tmp6, int %tmp3, uint 1 ; <<4 x int>> [#uses=1] - %tmp8 = insertelement <4 x int> %tmp7, int %tmp4, uint 2 ; <<4 x int>> [#uses=1] - %tmp9 = insertelement <4 x int> %tmp8, int %tmp5, uint 3 ; <<4 x int>> [#uses=1] - store <4 x int> %tmp9, <4 x int>* %A + %tmp = load <4 x i32>, <4 x i32>* %A ; <<4 x i32>> [#uses=4] + %tmp.upgrd.24 = extractelement <4 x i32> %tmp, i32 0 ; [#uses=1] + %tmp3 = extractelement <4 x i32> %tmp, i32 0 ; [#uses=1] + %tmp4 = extractelement <4 x i32> %tmp, i32 1 ; [#uses=1] + %tmp5 = extractelement <4 x i32> %tmp, i32 1 ; [#uses=1] + %tmp6 = insertelement <4 x i32> undef, i32 %tmp.upgrd.24, i32 0 ; <<4 x i32>> [#uses=1] + %tmp7 = insertelement <4 x i32> %tmp6, i32 %tmp3, i32 1 ; <<4 x i32>> [#uses=1] + %tmp8 = insertelement <4 x i32> %tmp7, i32 %tmp4, i32 2 ; <<4 x i32>> [#uses=1] + %tmp9 = insertelement <4 x i32> %tmp8, i32 %tmp5, i32 3 ; <<4 x i32>> [#uses=1] + store <4 x i32> %tmp9, <4 x i32>* %A ret void } -void %VPCKUHUM_unary(<8 x short>* %A, <8 x short>* %B) { +define void @VPCKUHUM_unary(<8 x i16>* %A, <8 x i16>* %B) { entry: - %tmp = load <8 x short>* %A ; <<8 x short>> [#uses=2] - %tmp = cast <8 x short> %tmp to <16 x sbyte> ; <<16 x sbyte>> [#uses=8] - %tmp3 = cast <8 x short> %tmp to <16 x sbyte> ; <<16 x sbyte>> [#uses=8] - %tmp = extractelement <16 x sbyte> %tmp, uint 1 ; [#uses=1] - %tmp4 = extractelement <16 x sbyte> %tmp, uint 3 ; [#uses=1] - %tmp5 = extractelement <16 x sbyte> %tmp, uint 5 ; [#uses=1] - %tmp6 = extractelement <16 x sbyte> %tmp, uint 7 ; [#uses=1] - %tmp7 = extractelement <16 x sbyte> %tmp, uint 9 ; [#uses=1] - %tmp8 = extractelement <16 x sbyte> %tmp, uint 11 ; [#uses=1] - %tmp9 = extractelement <16 x sbyte> %tmp, uint 13 ; [#uses=1] - %tmp10 = extractelement <16 x sbyte> %tmp, uint 15 ; [#uses=1] - %tmp11 = extractelement <16 x sbyte> %tmp3, uint 1 ; [#uses=1] - %tmp12 = extractelement <16 x sbyte> %tmp3, uint 3 ; [#uses=1] - %tmp13 = extractelement <16 x sbyte> %tmp3, uint 5 ; [#uses=1] - %tmp14 = extractelement <16 x sbyte> %tmp3, uint 7 ; [#uses=1] - %tmp15 = extractelement <16 x sbyte> %tmp3, uint 9 ; [#uses=1] - %tmp16 = extractelement <16 x sbyte> %tmp3, uint 11 ; [#uses=1] - %tmp17 = extractelement <16 x sbyte> %tmp3, uint 13 ; [#uses=1] - %tmp18 = extractelement <16 x sbyte> %tmp3, uint 15 ; [#uses=1] - %tmp19 = insertelement <16 x sbyte> undef, sbyte %tmp, uint 0 ; <<16 x sbyte>> [#uses=1] - %tmp20 = insertelement <16 x sbyte> %tmp19, sbyte %tmp4, uint 1 ; <<16 x sbyte>> [#uses=1] - %tmp21 = insertelement <16 x sbyte> %tmp20, sbyte %tmp5, uint 2 ; <<16 x sbyte>> [#uses=1] - %tmp22 = insertelement <16 x sbyte> %tmp21, sbyte %tmp6, uint 3 ; <<16 x sbyte>> [#uses=1] - %tmp23 = insertelement <16 x sbyte> %tmp22, sbyte %tmp7, uint 4 ; <<16 x sbyte>> [#uses=1] - %tmp24 = insertelement <16 x sbyte> %tmp23, sbyte %tmp8, uint 5 ; <<16 x sbyte>> [#uses=1] - %tmp25 = insertelement <16 x sbyte> %tmp24, sbyte %tmp9, uint 6 ; <<16 x sbyte>> [#uses=1] - %tmp26 = insertelement <16 x sbyte> %tmp25, sbyte %tmp10, uint 7 ; <<16 x sbyte>> [#uses=1] - %tmp27 = insertelement <16 x sbyte> %tmp26, sbyte %tmp11, uint 8 ; <<16 x sbyte>> [#uses=1] - %tmp28 = insertelement <16 x sbyte> %tmp27, sbyte %tmp12, uint 9 ; <<16 x sbyte>> [#uses=1] - %tmp29 = insertelement <16 x sbyte> %tmp28, sbyte %tmp13, uint 10 ; <<16 x sbyte>> [#uses=1] - %tmp30 = insertelement <16 x sbyte> %tmp29, sbyte %tmp14, uint 11 ; <<16 x sbyte>> [#uses=1] - %tmp31 = insertelement <16 x sbyte> %tmp30, sbyte %tmp15, uint 12 ; <<16 x sbyte>> [#uses=1] - %tmp32 = insertelement <16 x sbyte> %tmp31, sbyte %tmp16, uint 13 ; <<16 x sbyte>> [#uses=1] - %tmp33 = insertelement <16 x sbyte> %tmp32, sbyte %tmp17, uint 14 ; <<16 x sbyte>> [#uses=1] - %tmp34 = insertelement <16 x sbyte> %tmp33, sbyte %tmp18, uint 15 ; <<16 x sbyte>> [#uses=1] - %tmp34 = cast <16 x sbyte> %tmp34 to <8 x short> ; <<8 x short>> [#uses=1] - store <8 x short> %tmp34, <8 x short>* %A + %tmp = load <8 x i16>, <8 x i16>* %A ; <<8 x i16>> [#uses=2] + %tmp.upgrd.25 = bitcast <8 x i16> %tmp to <16 x i8> ; <<16 x i8>> [#uses=8] + %tmp3 = bitcast <8 x i16> %tmp to <16 x i8> ; <<16 x i8>> [#uses=8] + %tmp.upgrd.26 = extractelement <16 x i8> %tmp.upgrd.25, i32 1 ; [#uses=1] + %tmp4 = extractelement <16 x i8> %tmp.upgrd.25, i32 3 ; [#uses=1] + %tmp5 = extractelement <16 x i8> %tmp.upgrd.25, i32 5 ; [#uses=1] + %tmp6 = extractelement <16 x i8> %tmp.upgrd.25, i32 7 ; [#uses=1] + %tmp7 = extractelement <16 x i8> %tmp.upgrd.25, i32 9 ; [#uses=1] + %tmp8 = extractelement <16 x i8> %tmp.upgrd.25, i32 11 ; [#uses=1] + %tmp9 = extractelement <16 x i8> %tmp.upgrd.25, i32 13 ; [#uses=1] + %tmp10 = extractelement <16 x i8> %tmp.upgrd.25, i32 15 ; [#uses=1] + %tmp11 = extractelement <16 x i8> %tmp3, i32 1 ; [#uses=1] + %tmp12 = extractelement <16 x i8> %tmp3, i32 3 ; [#uses=1] + %tmp13 = extractelement <16 x i8> %tmp3, i32 5 ; [#uses=1] + %tmp14 = extractelement <16 x i8> %tmp3, i32 7 ; [#uses=1] + %tmp15 = extractelement <16 x i8> %tmp3, i32 9 ; [#uses=1] + %tmp16 = extractelement <16 x i8> %tmp3, i32 11 ; [#uses=1] + %tmp17 = extractelement <16 x i8> %tmp3, i32 13 ; [#uses=1] + %tmp18 = extractelement <16 x i8> %tmp3, i32 15 ; [#uses=1] + %tmp19 = insertelement <16 x i8> undef, i8 %tmp.upgrd.26, i32 0 ; <<16 x i8>> [#uses=1] + %tmp20 = insertelement <16 x i8> %tmp19, i8 %tmp4, i32 1 ; <<16 x i8>> [#uses=1] + %tmp21 = insertelement <16 x i8> %tmp20, i8 %tmp5, i32 2 ; <<16 x i8>> [#uses=1] + %tmp22 = insertelement <16 x i8> %tmp21, i8 %tmp6, i32 3 ; <<16 x i8>> [#uses=1] + %tmp23 = insertelement <16 x i8> %tmp22, i8 %tmp7, i32 4 ; <<16 x i8>> [#uses=1] + %tmp24 = insertelement <16 x i8> %tmp23, i8 %tmp8, i32 5 ; <<16 x i8>> [#uses=1] + %tmp25 = insertelement <16 x i8> %tmp24, i8 %tmp9, i32 6 ; <<16 x i8>> [#uses=1] + %tmp26 = insertelement <16 x i8> %tmp25, i8 %tmp10, i32 7 ; <<16 x i8>> [#uses=1] + %tmp27 = insertelement <16 x i8> %tmp26, i8 %tmp11, i32 8 ; <<16 x i8>> [#uses=1] + %tmp28 = insertelement <16 x i8> %tmp27, i8 %tmp12, i32 9 ; <<16 x i8>> [#uses=1] + %tmp29 = insertelement <16 x i8> %tmp28, i8 %tmp13, i32 10 ; <<16 x i8>> [#uses=1] + %tmp30 = insertelement <16 x i8> %tmp29, i8 %tmp14, i32 11 ; <<16 x i8>> [#uses=1] + %tmp31 = insertelement <16 x i8> %tmp30, i8 %tmp15, i32 12 ; <<16 x i8>> [#uses=1] + %tmp32 = insertelement <16 x i8> %tmp31, i8 %tmp16, i32 13 ; <<16 x i8>> [#uses=1] + %tmp33 = insertelement <16 x i8> %tmp32, i8 %tmp17, i32 14 ; <<16 x i8>> [#uses=1] + %tmp34 = insertelement <16 x i8> %tmp33, i8 %tmp18, i32 15 ; <<16 x i8>> [#uses=1] + %tmp34.upgrd.27 = bitcast <16 x i8> %tmp34 to <8 x i16> ; <<8 x i16>> [#uses=1] + store <8 x i16> %tmp34.upgrd.27, <8 x i16>* %A ret void } -void %VPCKUWUM_unary(<4 x int>* %A, <4 x int>* %B) { +define void @VPCKUWUM_unary(<4 x i32>* %A, <4 x i32>* %B) { entry: - %tmp = load <4 x int>* %A ; <<4 x int>> [#uses=2] - %tmp = cast <4 x int> %tmp to <8 x short> ; <<8 x short>> [#uses=4] - %tmp3 = cast <4 x int> %tmp to <8 x short> ; <<8 x short>> [#uses=4] - %tmp = extractelement <8 x short> %tmp, uint 1 ; [#uses=1] - %tmp4 = extractelement <8 x short> %tmp, uint 3 ; [#uses=1] - %tmp5 = extractelement <8 x short> %tmp, uint 5 ; [#uses=1] - %tmp6 = extractelement <8 x short> %tmp, uint 7 ; [#uses=1] - %tmp7 = extractelement <8 x short> %tmp3, uint 1 ; [#uses=1] - %tmp8 = extractelement <8 x short> %tmp3, uint 3 ; [#uses=1] - %tmp9 = extractelement <8 x short> %tmp3, uint 5 ; [#uses=1] - %tmp10 = extractelement <8 x short> %tmp3, uint 7 ; [#uses=1] - %tmp11 = insertelement <8 x short> undef, short %tmp, uint 0 ; <<8 x short>> [#uses=1] - %tmp12 = insertelement <8 x short> %tmp11, short %tmp4, uint 1 ; <<8 x short>> [#uses=1] - %tmp13 = insertelement <8 x short> %tmp12, short %tmp5, uint 2 ; <<8 x short>> [#uses=1] - %tmp14 = insertelement <8 x short> %tmp13, short %tmp6, uint 3 ; <<8 x short>> [#uses=1] - %tmp15 = insertelement <8 x short> %tmp14, short %tmp7, uint 4 ; <<8 x short>> [#uses=1] - %tmp16 = insertelement <8 x short> %tmp15, short %tmp8, uint 5 ; <<8 x short>> [#uses=1] - %tmp17 = insertelement <8 x short> %tmp16, short %tmp9, uint 6 ; <<8 x short>> [#uses=1] - %tmp18 = insertelement <8 x short> %tmp17, short %tmp10, uint 7 ; <<8 x short>> [#uses=1] - %tmp18 = cast <8 x short> %tmp18 to <4 x int> ; <<4 x int>> [#uses=1] - store <4 x int> %tmp18, <4 x int>* %A + %tmp = load <4 x i32>, <4 x i32>* %A ; <<4 x i32>> [#uses=2] + %tmp.upgrd.28 = bitcast <4 x i32> %tmp to <8 x i16> ; <<8 x i16>> [#uses=4] + %tmp3 = bitcast <4 x i32> %tmp to <8 x i16> ; <<8 x i16>> [#uses=4] + %tmp.upgrd.29 = extractelement <8 x i16> %tmp.upgrd.28, i32 1 ; [#uses=1] + %tmp4 = extractelement <8 x i16> %tmp.upgrd.28, i32 3 ; [#uses=1] + %tmp5 = extractelement <8 x i16> %tmp.upgrd.28, i32 5 ; [#uses=1] + %tmp6 = extractelement <8 x i16> %tmp.upgrd.28, i32 7 ; [#uses=1] + %tmp7 = extractelement <8 x i16> %tmp3, i32 1 ; [#uses=1] + %tmp8 = extractelement <8 x i16> %tmp3, i32 3 ; [#uses=1] + %tmp9 = extractelement <8 x i16> %tmp3, i32 5 ; [#uses=1] + %tmp10 = extractelement <8 x i16> %tmp3, i32 7 ; [#uses=1] + %tmp11 = insertelement <8 x i16> undef, i16 %tmp.upgrd.29, i32 0 ; <<8 x i16>> [#uses=1] + %tmp12 = insertelement <8 x i16> %tmp11, i16 %tmp4, i32 1 ; <<8 x i16>> [#uses=1] + %tmp13 = insertelement <8 x i16> %tmp12, i16 %tmp5, i32 2 ; <<8 x i16>> [#uses=1] + %tmp14 = insertelement <8 x i16> %tmp13, i16 %tmp6, i32 3 ; <<8 x i16>> [#uses=1] + %tmp15 = insertelement <8 x i16> %tmp14, i16 %tmp7, i32 4 ; <<8 x i16>> [#uses=1] + %tmp16 = insertelement <8 x i16> %tmp15, i16 %tmp8, i32 5 ; <<8 x i16>> [#uses=1] + %tmp17 = insertelement <8 x i16> %tmp16, i16 %tmp9, i32 6 ; <<8 x i16>> [#uses=1] + %tmp18 = insertelement <8 x i16> %tmp17, i16 %tmp10, i32 7 ; <<8 x i16>> [#uses=1] + %tmp18.upgrd.30 = bitcast <8 x i16> %tmp18 to <4 x i32> ; <<4 x i32>> [#uses=1] + store <4 x i32> %tmp18.upgrd.30, <4 x i32>* %A ret void }