From 0e5665bf0314b609bfa08bb64bad834e7678c8a6 Mon Sep 17 00:00:00 2001 From: Scott Michel Date: Wed, 19 Dec 2007 21:17:42 +0000 Subject: [PATCH] CellSPU testcase, extract_elt.ll: extract vector element. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@45219 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/Target/CellSPU/SPUISelLowering.cpp | 7 +- test/CodeGen/CellSPU/extract_elt.ll | 175 +++++++++++++++++++++++++ 2 files changed, 180 insertions(+), 2 deletions(-) create mode 100644 test/CodeGen/CellSPU/extract_elt.ll diff --git a/lib/Target/CellSPU/SPUISelLowering.cpp b/lib/Target/CellSPU/SPUISelLowering.cpp index 7d221877512..2ab4841c64e 100644 --- a/lib/Target/CellSPU/SPUISelLowering.cpp +++ b/lib/Target/CellSPU/SPUISelLowering.cpp @@ -2101,7 +2101,7 @@ static SDOperand LowerEXTRACT_VECTOR_ELT(SDOperand Op, SelectionDAG &DAG) { } // Need to generate shuffle mask and extract: - int prefslot_begin, prefslot_end; + int prefslot_begin = -1, prefslot_end = -1; int elt_byte = EltNo * MVT::getSizeInBits(VT) / 8; switch (VT) { @@ -2123,6 +2123,9 @@ static SDOperand LowerEXTRACT_VECTOR_ELT(SDOperand Op, SelectionDAG &DAG) { } } + assert(prefslot_begin != -1 && prefslot_end != -1 && + "LowerEXTRACT_VECTOR_ELT: preferred slots uninitialized"); + for (int i = 0; i < 16; ++i) { // zero fill uppper part of preferred slot, don't care about the // other slots: @@ -2134,7 +2137,7 @@ static SDOperand LowerEXTRACT_VECTOR_ELT(SDOperand Op, SelectionDAG &DAG) { ? 0x80 : elt_byte + (i - prefslot_begin)); - ShufMask[i] = DAG.getConstant(mask_val, MVT::i16); + ShufMask[i] = DAG.getConstant(mask_val, MVT::i8); } else ShufMask[i] = ShufMask[i % (prefslot_end + 1)]; } diff --git a/test/CodeGen/CellSPU/extract_elt.ll b/test/CodeGen/CellSPU/extract_elt.ll new file mode 100644 index 00000000000..ab485a81fd3 --- /dev/null +++ b/test/CodeGen/CellSPU/extract_elt.ll @@ -0,0 +1,175 @@ +; RUN: llvm-as -o - %s | llc -march=cellspu > %t1.s +; RUN: llvm-as -o - %s | llc -march=cellspu -mattr=large_mem > %t2.s +; RUN: grep shufb %t1.s | count 27 && +; RUN: grep lqa %t1.s | count 27 && +; RUN: grep lqx %t2.s | count 27 && +; RUN: grep space %t1.s | count 8 && +; RUN: grep byte %t1.s | count 424 + +define i32 @i32_extract_0(<4 x i32> %v) { +entry: + %a = extractelement <4 x i32> %v, i32 0 + ret i32 %a +} + +define i32 @i32_extract_1(<4 x i32> %v) { +entry: + %a = extractelement <4 x i32> %v, i32 1 + ret i32 %a +} + +define i32 @i32_extract_2(<4 x i32> %v) { +entry: + %a = extractelement <4 x i32> %v, i32 2 + ret i32 %a +} + +define i32 @i32_extract_3(<4 x i32> %v) { +entry: + %a = extractelement <4 x i32> %v, i32 3 + ret i32 %a +} + +define i16 @i16_extract_0(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 0 + ret i16 %a +} + +define i16 @i16_extract_1(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 1 + ret i16 %a +} + +define i16 @i16_extract_2(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 2 + ret i16 %a +} + +define i16 @i16_extract_3(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 3 + ret i16 %a +} + +define i16 @i16_extract_4(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 4 + ret i16 %a +} + +define i16 @i16_extract_5(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 5 + ret i16 %a +} + +define i16 @i16_extract_6(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 6 + ret i16 %a +} + +define i16 @i16_extract_7(<8 x i16> %v) { +entry: + %a = extractelement <8 x i16> %v, i32 7 + ret i16 %a +} + +define i8 @i8_extract_0(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 0 + ret i8 %a +} + +define i8 @i8_extract_1(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 1 + ret i8 %a +} + +define i8 @i8_extract_2(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 2 + ret i8 %a +} + +define i8 @i8_extract_3(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 3 + ret i8 %a +} + +define i8 @i8_extract_4(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 4 + ret i8 %a +} + +define i8 @i8_extract_5(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 5 + ret i8 %a +} + +define i8 @i8_extract_6(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 6 + ret i8 %a +} + +define i8 @i8_extract_7(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 7 + ret i8 %a +} + +define i8 @i8_extract_8(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 8 + ret i8 %a +} + +define i8 @i8_extract_9(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 9 + ret i8 %a +} + +define i8 @i8_extract_10(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 10 + ret i8 %a +} + +define i8 @i8_extract_11(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 11 + ret i8 %a +} + +define i8 @i8_extract_12(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 12 + ret i8 %a +} + +define i8 @i8_extract_13(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 13 + ret i8 %a +} + +define i8 @i8_extract_14(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 14 + ret i8 %a +} + +define i8 @i8_extract_15(<16 x i8> %v) { +entry: + %a = extractelement <16 x i8> %v, i32 15 + ret i8 %a +} -- 2.34.1