1 //===-- ARMTargetTransformInfo.cpp - ARM specific TTI pass ----------------===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 /// This file implements a TargetTransformInfo analysis pass specific to the
11 /// ARM target machine. It uses the target's detailed information to provide
12 /// more precise answers to certain TTI queries, while letting the target
13 /// independent and default TTI implementations handle the rest.
15 //===----------------------------------------------------------------------===//
17 #define DEBUG_TYPE "armtti"
19 #include "ARMTargetMachine.h"
20 #include "llvm/Analysis/TargetTransformInfo.h"
21 #include "llvm/Support/Debug.h"
22 #include "llvm/Target/TargetLowering.h"
23 #include "llvm/Target/CostTable.h"
26 // Declare the pass initialization routine locally as target-specific passes
27 // don't havve a target-wide initialization entry point, and so we rely on the
28 // pass constructor initialization.
30 void initializeARMTTIPass(PassRegistry &);
35 class ARMTTI : public ImmutablePass, public TargetTransformInfo {
36 const ARMBaseTargetMachine *TM;
37 const ARMSubtarget *ST;
38 const ARMTargetLowering *TLI;
40 /// Estimate the overhead of scalarizing an instruction. Insert and Extract
41 /// are set if the result needs to be inserted and/or extracted from vectors.
42 unsigned getScalarizationOverhead(Type *Ty, bool Insert, bool Extract) const;
45 ARMTTI() : ImmutablePass(ID), TM(0), ST(0), TLI(0) {
46 llvm_unreachable("This pass cannot be directly constructed");
49 ARMTTI(const ARMBaseTargetMachine *TM)
50 : ImmutablePass(ID), TM(TM), ST(TM->getSubtargetImpl()),
51 TLI(TM->getTargetLowering()) {
52 initializeARMTTIPass(*PassRegistry::getPassRegistry());
55 virtual void initializePass() {
59 virtual void finalizePass() {
63 virtual void getAnalysisUsage(AnalysisUsage &AU) const {
64 TargetTransformInfo::getAnalysisUsage(AU);
67 /// Pass identification.
70 /// Provide necessary pointer adjustments for the two base classes.
71 virtual void *getAdjustedAnalysisPointer(const void *ID) {
72 if (ID == &TargetTransformInfo::ID)
73 return (TargetTransformInfo*)this;
77 /// \name Scalar TTI Implementations
80 virtual unsigned getIntImmCost(const APInt &Imm, Type *Ty) const;
85 /// \name Vector TTI Implementations
88 unsigned getNumberOfRegisters(bool Vector) const {
95 if (ST->isThumb1Only())
100 unsigned getRegisterBitWidth(bool Vector) const {
110 unsigned getMaximumUnrollFactor() const {
111 // These are out of order CPUs:
112 if (ST->isCortexA15() || ST->isSwift())
117 unsigned getShuffleCost(ShuffleKind Kind, Type *Tp,
118 int Index, Type *SubTp) const;
120 unsigned getCastInstrCost(unsigned Opcode, Type *Dst,
123 unsigned getCmpSelInstrCost(unsigned Opcode, Type *ValTy, Type *CondTy) const;
125 unsigned getVectorInstrCost(unsigned Opcode, Type *Val, unsigned Index) const;
127 unsigned getAddressComputationCost(Type *Val, bool IsComplex) const;
129 unsigned getArithmeticInstrCost(unsigned Opcode, Type *Ty,
130 OperandValueKind Op1Info = OK_AnyValue,
131 OperandValueKind Op2Info = OK_AnyValue) const;
135 } // end anonymous namespace
137 INITIALIZE_AG_PASS(ARMTTI, TargetTransformInfo, "armtti",
138 "ARM Target Transform Info", true, true, false)
142 llvm::createARMTargetTransformInfoPass(const ARMBaseTargetMachine *TM) {
143 return new ARMTTI(TM);
147 unsigned ARMTTI::getIntImmCost(const APInt &Imm, Type *Ty) const {
148 assert(Ty->isIntegerTy());
150 unsigned Bits = Ty->getPrimitiveSizeInBits();
151 if (Bits == 0 || Bits > 32)
154 int32_t SImmVal = Imm.getSExtValue();
155 uint32_t ZImmVal = Imm.getZExtValue();
156 if (!ST->isThumb()) {
157 if ((SImmVal >= 0 && SImmVal < 65536) ||
158 (ARM_AM::getSOImmVal(ZImmVal) != -1) ||
159 (ARM_AM::getSOImmVal(~ZImmVal) != -1))
161 return ST->hasV6T2Ops() ? 2 : 3;
162 } else if (ST->isThumb2()) {
163 if ((SImmVal >= 0 && SImmVal < 65536) ||
164 (ARM_AM::getT2SOImmVal(ZImmVal) != -1) ||
165 (ARM_AM::getT2SOImmVal(~ZImmVal) != -1))
167 return ST->hasV6T2Ops() ? 2 : 3;
169 if (SImmVal >= 0 && SImmVal < 256)
171 if ((~ZImmVal < 256) || ARM_AM::isThumbImmShiftedVal(ZImmVal))
173 // Load from constantpool.
179 unsigned ARMTTI::getCastInstrCost(unsigned Opcode, Type *Dst,
181 int ISD = TLI->InstructionOpcodeToISD(Opcode);
182 assert(ISD && "Invalid opcode");
184 // Single to/from double precision conversions.
185 static const CostTblEntry<MVT::SimpleValueType> NEONFltDblTbl[] = {
186 // Vector fptrunc/fpext conversions.
187 { ISD::FP_ROUND, MVT::v2f64, 2 },
188 { ISD::FP_EXTEND, MVT::v2f32, 2 },
189 { ISD::FP_EXTEND, MVT::v4f32, 4 }
192 if (Src->isVectorTy() && ST->hasNEON() && (ISD == ISD::FP_ROUND ||
193 ISD == ISD::FP_EXTEND)) {
194 std::pair<unsigned, MVT> LT = TLI->getTypeLegalizationCost(Src);
195 int Idx = CostTableLookup(NEONFltDblTbl, ISD, LT.second);
197 return LT.first * NEONFltDblTbl[Idx].Cost;
200 EVT SrcTy = TLI->getValueType(Src);
201 EVT DstTy = TLI->getValueType(Dst);
203 if (!SrcTy.isSimple() || !DstTy.isSimple())
204 return TargetTransformInfo::getCastInstrCost(Opcode, Dst, Src);
206 // Some arithmetic, load and store operations have specific instructions
207 // to cast up/down their types automatically at no extra cost.
208 // TODO: Get these tables to know at least what the related operations are.
209 static const TypeConversionCostTblEntry<MVT::SimpleValueType>
210 NEONVectorConversionTbl[] = {
211 { ISD::SIGN_EXTEND, MVT::v4i32, MVT::v4i16, 0 },
212 { ISD::ZERO_EXTEND, MVT::v4i32, MVT::v4i16, 0 },
213 { ISD::SIGN_EXTEND, MVT::v2i64, MVT::v2i32, 1 },
214 { ISD::ZERO_EXTEND, MVT::v2i64, MVT::v2i32, 1 },
215 { ISD::TRUNCATE, MVT::v4i32, MVT::v4i64, 0 },
216 { ISD::TRUNCATE, MVT::v4i16, MVT::v4i32, 1 },
218 // The number of vmovl instructions for the extension.
219 { ISD::SIGN_EXTEND, MVT::v4i64, MVT::v4i16, 3 },
220 { ISD::ZERO_EXTEND, MVT::v4i64, MVT::v4i16, 3 },
221 { ISD::SIGN_EXTEND, MVT::v8i32, MVT::v8i8, 3 },
222 { ISD::ZERO_EXTEND, MVT::v8i32, MVT::v8i8, 3 },
223 { ISD::SIGN_EXTEND, MVT::v8i64, MVT::v8i8, 7 },
224 { ISD::ZERO_EXTEND, MVT::v8i64, MVT::v8i8, 7 },
225 { ISD::SIGN_EXTEND, MVT::v8i64, MVT::v8i16, 6 },
226 { ISD::ZERO_EXTEND, MVT::v8i64, MVT::v8i16, 6 },
227 { ISD::SIGN_EXTEND, MVT::v16i32, MVT::v16i8, 6 },
228 { ISD::ZERO_EXTEND, MVT::v16i32, MVT::v16i8, 6 },
230 // Operations that we legalize using splitting.
231 { ISD::TRUNCATE, MVT::v16i8, MVT::v16i32, 6 },
232 { ISD::TRUNCATE, MVT::v8i8, MVT::v8i32, 3 },
234 // Vector float <-> i32 conversions.
235 { ISD::SINT_TO_FP, MVT::v4f32, MVT::v4i32, 1 },
236 { ISD::UINT_TO_FP, MVT::v4f32, MVT::v4i32, 1 },
238 { ISD::SINT_TO_FP, MVT::v2f32, MVT::v2i8, 3 },
239 { ISD::UINT_TO_FP, MVT::v2f32, MVT::v2i8, 3 },
240 { ISD::SINT_TO_FP, MVT::v2f32, MVT::v2i16, 2 },
241 { ISD::UINT_TO_FP, MVT::v2f32, MVT::v2i16, 2 },
242 { ISD::SINT_TO_FP, MVT::v2f32, MVT::v2i32, 1 },
243 { ISD::UINT_TO_FP, MVT::v2f32, MVT::v2i32, 1 },
244 { ISD::SINT_TO_FP, MVT::v4f32, MVT::v4i1, 3 },
245 { ISD::UINT_TO_FP, MVT::v4f32, MVT::v4i1, 3 },
246 { ISD::SINT_TO_FP, MVT::v4f32, MVT::v4i8, 3 },
247 { ISD::UINT_TO_FP, MVT::v4f32, MVT::v4i8, 3 },
248 { ISD::SINT_TO_FP, MVT::v4f32, MVT::v4i16, 2 },
249 { ISD::UINT_TO_FP, MVT::v4f32, MVT::v4i16, 2 },
250 { ISD::SINT_TO_FP, MVT::v8f32, MVT::v8i16, 4 },
251 { ISD::UINT_TO_FP, MVT::v8f32, MVT::v8i16, 4 },
252 { ISD::SINT_TO_FP, MVT::v8f32, MVT::v8i32, 2 },
253 { ISD::UINT_TO_FP, MVT::v8f32, MVT::v8i32, 2 },
254 { ISD::SINT_TO_FP, MVT::v16f32, MVT::v16i16, 8 },
255 { ISD::UINT_TO_FP, MVT::v16f32, MVT::v16i16, 8 },
256 { ISD::SINT_TO_FP, MVT::v16f32, MVT::v16i32, 4 },
257 { ISD::UINT_TO_FP, MVT::v16f32, MVT::v16i32, 4 },
259 { ISD::FP_TO_SINT, MVT::v4i32, MVT::v4f32, 1 },
260 { ISD::FP_TO_UINT, MVT::v4i32, MVT::v4f32, 1 },
261 { ISD::FP_TO_SINT, MVT::v4i8, MVT::v4f32, 3 },
262 { ISD::FP_TO_UINT, MVT::v4i8, MVT::v4f32, 3 },
263 { ISD::FP_TO_SINT, MVT::v4i16, MVT::v4f32, 2 },
264 { ISD::FP_TO_UINT, MVT::v4i16, MVT::v4f32, 2 },
266 // Vector double <-> i32 conversions.
267 { ISD::SINT_TO_FP, MVT::v2f64, MVT::v2i32, 2 },
268 { ISD::UINT_TO_FP, MVT::v2f64, MVT::v2i32, 2 },
270 { ISD::SINT_TO_FP, MVT::v2f64, MVT::v2i8, 4 },
271 { ISD::UINT_TO_FP, MVT::v2f64, MVT::v2i8, 4 },
272 { ISD::SINT_TO_FP, MVT::v2f64, MVT::v2i16, 3 },
273 { ISD::UINT_TO_FP, MVT::v2f64, MVT::v2i16, 3 },
274 { ISD::SINT_TO_FP, MVT::v2f64, MVT::v2i32, 2 },
275 { ISD::UINT_TO_FP, MVT::v2f64, MVT::v2i32, 2 },
277 { ISD::FP_TO_SINT, MVT::v2i32, MVT::v2f64, 2 },
278 { ISD::FP_TO_UINT, MVT::v2i32, MVT::v2f64, 2 },
279 { ISD::FP_TO_SINT, MVT::v8i16, MVT::v8f32, 4 },
280 { ISD::FP_TO_UINT, MVT::v8i16, MVT::v8f32, 4 },
281 { ISD::FP_TO_SINT, MVT::v16i16, MVT::v16f32, 8 },
282 { ISD::FP_TO_UINT, MVT::v16i16, MVT::v16f32, 8 }
285 if (SrcTy.isVector() && ST->hasNEON()) {
286 int Idx = ConvertCostTableLookup(NEONVectorConversionTbl, ISD,
287 DstTy.getSimpleVT(), SrcTy.getSimpleVT());
289 return NEONVectorConversionTbl[Idx].Cost;
292 // Scalar float to integer conversions.
293 static const TypeConversionCostTblEntry<MVT::SimpleValueType>
294 NEONFloatConversionTbl[] = {
295 { ISD::FP_TO_SINT, MVT::i1, MVT::f32, 2 },
296 { ISD::FP_TO_UINT, MVT::i1, MVT::f32, 2 },
297 { ISD::FP_TO_SINT, MVT::i1, MVT::f64, 2 },
298 { ISD::FP_TO_UINT, MVT::i1, MVT::f64, 2 },
299 { ISD::FP_TO_SINT, MVT::i8, MVT::f32, 2 },
300 { ISD::FP_TO_UINT, MVT::i8, MVT::f32, 2 },
301 { ISD::FP_TO_SINT, MVT::i8, MVT::f64, 2 },
302 { ISD::FP_TO_UINT, MVT::i8, MVT::f64, 2 },
303 { ISD::FP_TO_SINT, MVT::i16, MVT::f32, 2 },
304 { ISD::FP_TO_UINT, MVT::i16, MVT::f32, 2 },
305 { ISD::FP_TO_SINT, MVT::i16, MVT::f64, 2 },
306 { ISD::FP_TO_UINT, MVT::i16, MVT::f64, 2 },
307 { ISD::FP_TO_SINT, MVT::i32, MVT::f32, 2 },
308 { ISD::FP_TO_UINT, MVT::i32, MVT::f32, 2 },
309 { ISD::FP_TO_SINT, MVT::i32, MVT::f64, 2 },
310 { ISD::FP_TO_UINT, MVT::i32, MVT::f64, 2 },
311 { ISD::FP_TO_SINT, MVT::i64, MVT::f32, 10 },
312 { ISD::FP_TO_UINT, MVT::i64, MVT::f32, 10 },
313 { ISD::FP_TO_SINT, MVT::i64, MVT::f64, 10 },
314 { ISD::FP_TO_UINT, MVT::i64, MVT::f64, 10 }
316 if (SrcTy.isFloatingPoint() && ST->hasNEON()) {
317 int Idx = ConvertCostTableLookup(NEONFloatConversionTbl, ISD,
318 DstTy.getSimpleVT(), SrcTy.getSimpleVT());
320 return NEONFloatConversionTbl[Idx].Cost;
323 // Scalar integer to float conversions.
324 static const TypeConversionCostTblEntry<MVT::SimpleValueType>
325 NEONIntegerConversionTbl[] = {
326 { ISD::SINT_TO_FP, MVT::f32, MVT::i1, 2 },
327 { ISD::UINT_TO_FP, MVT::f32, MVT::i1, 2 },
328 { ISD::SINT_TO_FP, MVT::f64, MVT::i1, 2 },
329 { ISD::UINT_TO_FP, MVT::f64, MVT::i1, 2 },
330 { ISD::SINT_TO_FP, MVT::f32, MVT::i8, 2 },
331 { ISD::UINT_TO_FP, MVT::f32, MVT::i8, 2 },
332 { ISD::SINT_TO_FP, MVT::f64, MVT::i8, 2 },
333 { ISD::UINT_TO_FP, MVT::f64, MVT::i8, 2 },
334 { ISD::SINT_TO_FP, MVT::f32, MVT::i16, 2 },
335 { ISD::UINT_TO_FP, MVT::f32, MVT::i16, 2 },
336 { ISD::SINT_TO_FP, MVT::f64, MVT::i16, 2 },
337 { ISD::UINT_TO_FP, MVT::f64, MVT::i16, 2 },
338 { ISD::SINT_TO_FP, MVT::f32, MVT::i32, 2 },
339 { ISD::UINT_TO_FP, MVT::f32, MVT::i32, 2 },
340 { ISD::SINT_TO_FP, MVT::f64, MVT::i32, 2 },
341 { ISD::UINT_TO_FP, MVT::f64, MVT::i32, 2 },
342 { ISD::SINT_TO_FP, MVT::f32, MVT::i64, 10 },
343 { ISD::UINT_TO_FP, MVT::f32, MVT::i64, 10 },
344 { ISD::SINT_TO_FP, MVT::f64, MVT::i64, 10 },
345 { ISD::UINT_TO_FP, MVT::f64, MVT::i64, 10 }
348 if (SrcTy.isInteger() && ST->hasNEON()) {
349 int Idx = ConvertCostTableLookup(NEONIntegerConversionTbl, ISD,
350 DstTy.getSimpleVT(), SrcTy.getSimpleVT());
352 return NEONIntegerConversionTbl[Idx].Cost;
355 // Scalar integer conversion costs.
356 static const TypeConversionCostTblEntry<MVT::SimpleValueType>
357 ARMIntegerConversionTbl[] = {
358 // i16 -> i64 requires two dependent operations.
359 { ISD::SIGN_EXTEND, MVT::i64, MVT::i16, 2 },
361 // Truncates on i64 are assumed to be free.
362 { ISD::TRUNCATE, MVT::i32, MVT::i64, 0 },
363 { ISD::TRUNCATE, MVT::i16, MVT::i64, 0 },
364 { ISD::TRUNCATE, MVT::i8, MVT::i64, 0 },
365 { ISD::TRUNCATE, MVT::i1, MVT::i64, 0 }
368 if (SrcTy.isInteger()) {
369 int Idx = ConvertCostTableLookup(ARMIntegerConversionTbl, ISD,
370 DstTy.getSimpleVT(), SrcTy.getSimpleVT());
372 return ARMIntegerConversionTbl[Idx].Cost;
375 return TargetTransformInfo::getCastInstrCost(Opcode, Dst, Src);
378 unsigned ARMTTI::getVectorInstrCost(unsigned Opcode, Type *ValTy,
379 unsigned Index) const {
380 // Penalize inserting into an D-subregister. We end up with a three times
381 // lower estimated throughput on swift.
383 Opcode == Instruction::InsertElement &&
384 ValTy->isVectorTy() &&
385 ValTy->getScalarSizeInBits() <= 32)
388 return TargetTransformInfo::getVectorInstrCost(Opcode, ValTy, Index);
391 unsigned ARMTTI::getCmpSelInstrCost(unsigned Opcode, Type *ValTy,
392 Type *CondTy) const {
394 int ISD = TLI->InstructionOpcodeToISD(Opcode);
395 // On NEON a a vector select gets lowered to vbsl.
396 if (ST->hasNEON() && ValTy->isVectorTy() && ISD == ISD::SELECT) {
397 // Lowering of some vector selects is currently far from perfect.
398 static const TypeConversionCostTblEntry<MVT::SimpleValueType>
399 NEONVectorSelectTbl[] = {
400 { ISD::SELECT, MVT::v16i1, MVT::v16i16, 2*16 + 1 + 3*1 + 4*1 },
401 { ISD::SELECT, MVT::v8i1, MVT::v8i32, 4*8 + 1*3 + 1*4 + 1*2 },
402 { ISD::SELECT, MVT::v16i1, MVT::v16i32, 4*16 + 1*6 + 1*8 + 1*4 },
403 { ISD::SELECT, MVT::v4i1, MVT::v4i64, 4*4 + 1*2 + 1 },
404 { ISD::SELECT, MVT::v8i1, MVT::v8i64, 50 },
405 { ISD::SELECT, MVT::v16i1, MVT::v16i64, 100 }
408 EVT SelCondTy = TLI->getValueType(CondTy);
409 EVT SelValTy = TLI->getValueType(ValTy);
410 if (SelCondTy.isSimple() && SelValTy.isSimple()) {
411 int Idx = ConvertCostTableLookup(NEONVectorSelectTbl, ISD,
412 SelCondTy.getSimpleVT(),
413 SelValTy.getSimpleVT());
415 return NEONVectorSelectTbl[Idx].Cost;
418 std::pair<unsigned, MVT> LT = TLI->getTypeLegalizationCost(ValTy);
422 return TargetTransformInfo::getCmpSelInstrCost(Opcode, ValTy, CondTy);
425 unsigned ARMTTI::getAddressComputationCost(Type *Ty, bool IsComplex) const {
426 // Address computations in vectorized code with non-consecutive addresses will
427 // likely result in more instructions compared to scalar code where the
428 // computation can more often be merged into the index mode. The resulting
429 // extra micro-ops can significantly decrease throughput.
430 unsigned NumVectorInstToHideOverhead = 10;
432 if (Ty->isVectorTy() && IsComplex)
433 return NumVectorInstToHideOverhead;
435 // In many cases the address computation is not merged into the instruction
440 unsigned ARMTTI::getShuffleCost(ShuffleKind Kind, Type *Tp, int Index,
442 // We only handle costs of reverse shuffles for now.
443 if (Kind != SK_Reverse)
444 return TargetTransformInfo::getShuffleCost(Kind, Tp, Index, SubTp);
446 static const CostTblEntry<MVT::SimpleValueType> NEONShuffleTbl[] = {
447 // Reverse shuffle cost one instruction if we are shuffling within a double
448 // word (vrev) or two if we shuffle a quad word (vrev, vext).
449 { ISD::VECTOR_SHUFFLE, MVT::v2i32, 1 },
450 { ISD::VECTOR_SHUFFLE, MVT::v2f32, 1 },
451 { ISD::VECTOR_SHUFFLE, MVT::v2i64, 1 },
452 { ISD::VECTOR_SHUFFLE, MVT::v2f64, 1 },
454 { ISD::VECTOR_SHUFFLE, MVT::v4i32, 2 },
455 { ISD::VECTOR_SHUFFLE, MVT::v4f32, 2 },
456 { ISD::VECTOR_SHUFFLE, MVT::v8i16, 2 },
457 { ISD::VECTOR_SHUFFLE, MVT::v16i8, 2 }
460 std::pair<unsigned, MVT> LT = TLI->getTypeLegalizationCost(Tp);
462 int Idx = CostTableLookup(NEONShuffleTbl, ISD::VECTOR_SHUFFLE, LT.second);
464 return TargetTransformInfo::getShuffleCost(Kind, Tp, Index, SubTp);
466 return LT.first * NEONShuffleTbl[Idx].Cost;
469 unsigned ARMTTI::getArithmeticInstrCost(unsigned Opcode, Type *Ty, OperandValueKind Op1Info,
470 OperandValueKind Op2Info) const {
472 int ISDOpcode = TLI->InstructionOpcodeToISD(Opcode);
473 std::pair<unsigned, MVT> LT = TLI->getTypeLegalizationCost(Ty);
475 const unsigned FunctionCallDivCost = 20;
476 const unsigned ReciprocalDivCost = 10;
477 static const CostTblEntry<MVT::SimpleValueType> CostTbl[] = {
479 // These costs are somewhat random. Choose a cost of 20 to indicate that
480 // vectorizing devision (added function call) is going to be very expensive.
481 // Double registers types.
482 { ISD::SDIV, MVT::v1i64, 1 * FunctionCallDivCost},
483 { ISD::UDIV, MVT::v1i64, 1 * FunctionCallDivCost},
484 { ISD::SREM, MVT::v1i64, 1 * FunctionCallDivCost},
485 { ISD::UREM, MVT::v1i64, 1 * FunctionCallDivCost},
486 { ISD::SDIV, MVT::v2i32, 2 * FunctionCallDivCost},
487 { ISD::UDIV, MVT::v2i32, 2 * FunctionCallDivCost},
488 { ISD::SREM, MVT::v2i32, 2 * FunctionCallDivCost},
489 { ISD::UREM, MVT::v2i32, 2 * FunctionCallDivCost},
490 { ISD::SDIV, MVT::v4i16, ReciprocalDivCost},
491 { ISD::UDIV, MVT::v4i16, ReciprocalDivCost},
492 { ISD::SREM, MVT::v4i16, 4 * FunctionCallDivCost},
493 { ISD::UREM, MVT::v4i16, 4 * FunctionCallDivCost},
494 { ISD::SDIV, MVT::v8i8, ReciprocalDivCost},
495 { ISD::UDIV, MVT::v8i8, ReciprocalDivCost},
496 { ISD::SREM, MVT::v8i8, 8 * FunctionCallDivCost},
497 { ISD::UREM, MVT::v8i8, 8 * FunctionCallDivCost},
498 // Quad register types.
499 { ISD::SDIV, MVT::v2i64, 2 * FunctionCallDivCost},
500 { ISD::UDIV, MVT::v2i64, 2 * FunctionCallDivCost},
501 { ISD::SREM, MVT::v2i64, 2 * FunctionCallDivCost},
502 { ISD::UREM, MVT::v2i64, 2 * FunctionCallDivCost},
503 { ISD::SDIV, MVT::v4i32, 4 * FunctionCallDivCost},
504 { ISD::UDIV, MVT::v4i32, 4 * FunctionCallDivCost},
505 { ISD::SREM, MVT::v4i32, 4 * FunctionCallDivCost},
506 { ISD::UREM, MVT::v4i32, 4 * FunctionCallDivCost},
507 { ISD::SDIV, MVT::v8i16, 8 * FunctionCallDivCost},
508 { ISD::UDIV, MVT::v8i16, 8 * FunctionCallDivCost},
509 { ISD::SREM, MVT::v8i16, 8 * FunctionCallDivCost},
510 { ISD::UREM, MVT::v8i16, 8 * FunctionCallDivCost},
511 { ISD::SDIV, MVT::v16i8, 16 * FunctionCallDivCost},
512 { ISD::UDIV, MVT::v16i8, 16 * FunctionCallDivCost},
513 { ISD::SREM, MVT::v16i8, 16 * FunctionCallDivCost},
514 { ISD::UREM, MVT::v16i8, 16 * FunctionCallDivCost},
521 Idx = CostTableLookup(CostTbl, ISDOpcode, LT.second);
524 return LT.first * CostTbl[Idx].Cost;
527 TargetTransformInfo::getArithmeticInstrCost(Opcode, Ty, Op1Info, Op2Info);
529 // This is somewhat of a hack. The problem that we are facing is that SROA
530 // creates a sequence of shift, and, or instructions to construct values.
531 // These sequences are recognized by the ISel and have zero-cost. Not so for
532 // the vectorized code. Because we have support for v2i64 but not i64 those
533 // sequences look particularily beneficial to vectorize.
534 // To work around this we increase the cost of v2i64 operations to make them
535 // seem less beneficial.
536 if (LT.second == MVT::v2i64 &&
537 Op2Info == TargetTransformInfo::OK_UniformConstantValue)