//
// The LLVM Compiler Infrastructure
//
-// This file was developed by the LLVM research group and is distributed under
-// the University of Illinois Open Source License. See LICENSE.TXT for details.
+// This file is distributed under the University of Illinois Open Source
+// License. See LICENSE.TXT for details.
//
//===----------------------------------------------------------------------===//
//
#include "llvm/ADT/GraphTraits.h"
#include "llvm/ADT/iterator"
#include "llvm/ADT/APFloat.h"
+#include "llvm/ADT/APInt.h"
#include "llvm/CodeGen/ValueTypes.h"
+#include "llvm/CodeGen/MemOperand.h"
#include "llvm/Support/DataTypes.h"
#include <cassert>
// CopyFromReg - This node indicates that the input value is a virtual or
// physical register that is defined outside of the scope of this
- // SelectionDAG. The register is available from the RegSDNode object.
+ // SelectionDAG. The register is available from the RegisterSDNode object.
CopyFromReg,
// UNDEF - An undefined node
// Simple integer binary arithmetic operators.
ADD, SUB, MUL, SDIV, UDIV, SREM, UREM,
+
+ // SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing
+ // a signed/unsigned value of type i[2*N], and return the full value as
+ // two results, each of type iN.
+ SMUL_LOHI, UMUL_LOHI,
+
+ // SDIVREM/UDIVREM - Divide two integers and produce both a quotient and
+ // remainder result.
+ SDIVREM, UDIVREM,
// CARRY_FALSE - This node is used when folding other nodes,
// like ADDC/SUBC, which indicate the carry result is always false.
// FCOPYSIGN(f32, f64) is allowed.
FCOPYSIGN,
+ // INT = FGETSIGN(FP) - Return the sign bit of the specified floating point
+ // value as an integer 0/1 value.
+ FGETSIGN,
+
/// BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a vector
/// with the specified, possibly variable, elements. The number of elements
/// is required to be a power of two.
BUILD_VECTOR,
/// INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element
- /// at IDX replaced with VAL.
+ /// at IDX replaced with VAL. If the type of VAL is larger than the vector
+ /// element type then VAL is truncated before replacement.
INSERT_VECTOR_ELT,
/// EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR
/// vector value) starting with the (potentially variable) element number
/// IDX, which must be a multiple of the result vector length.
EXTRACT_SUBVECTOR,
-
+
/// VECTOR_SHUFFLE(VEC1, VEC2, SHUFFLEVEC) - Returns a vector, of the same
/// type as VEC1/VEC2. SHUFFLEVEC is a BUILD_VECTOR of constant int values
- /// (regardless of whether its datatype is legal or not) that indicate
- /// which value each result element will get. The elements of VEC1/VEC2 are
- /// enumerated in order. This is quite similar to the Altivec 'vperm'
- /// instruction, except that the indices must be constants and are in terms
- /// of the element size of VEC1/VEC2, not in terms of bytes.
+ /// (maybe of an illegal datatype) or undef that indicate which value each
+ /// result element will get. The elements of VEC1/VEC2 are enumerated in
+ /// order. This is quite similar to the Altivec 'vperm' instruction, except
+ /// that the indices must be constants and are in terms of the element size
+ /// of VEC1/VEC2, not in terms of bytes.
VECTOR_SHUFFLE,
-
+
/// SCALAR_TO_VECTOR(VAL) - This represents the operation of loading a
- /// scalar value into the low element of the resultant vector type. The top
- /// elements of the vector are undefined.
+ /// scalar value into element 0 of the resultant vector type. The top
+ /// elements 1 to N-1 of the N-element vector are undefined.
SCALAR_TO_VECTOR,
// EXTRACT_SUBREG - This node is used to extract a sub-register value.
INSERT_SUBREG,
// MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing
- // an unsigned/signed value of type i[2*n], then return the top part.
+ // an unsigned/signed value of type i[2*N], then return the top part.
MULHU, MULHS,
// Bitwise operators - logical and, logical or, logical xor, shift left,
// operand, a ValueType node.
SIGN_EXTEND_INREG,
- // FP_TO_[US]INT - Convert a floating point value to a signed or unsigned
- // integer.
+ /// FP_TO_[US]INT - Convert a floating point value to a signed or unsigned
+ /// integer.
FP_TO_SINT,
FP_TO_UINT,
- // FP_ROUND - Perform a rounding operation from the current
- // precision down to the specified precision (currently always 64->32).
+ /// X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type
+ /// down to the precision of the destination VT. TRUNC is a flag, which is
+ /// always an integer that is zero or one. If TRUNC is 0, this is a
+ /// normal rounding, if it is 1, this FP_ROUND is known to not change the
+ /// value of Y.
+ ///
+ /// The TRUNC = 1 case is used in cases where we know that the value will
+ /// not be modified by the node, because Y is not using any of the extra
+ /// precision of source type. This allows certain transformations like
+ /// FP_EXTEND(FP_ROUND(X,1)) -> X which are not safe for
+ /// FP_EXTEND(FP_ROUND(X,0)) because the extra bits aren't removed.
FP_ROUND,
-
- // FP_ROUND_INREG - This operator takes a floating point register, and
- // rounds it to a floating point value. It then promotes it and returns it
- // in a register of the same size. This operation effectively just discards
- // excess precision. The type to round down to is specified by the 1th
- // operation, a VTSDNode (currently always 64->32->64).
+
+ // FLT_ROUNDS_ - Returns current rounding mode:
+ // -1 Undefined
+ // 0 Round to 0
+ // 1 Round to nearest
+ // 2 Round to +inf
+ // 3 Round to -inf
+ FLT_ROUNDS_,
+
+ /// X = FP_ROUND_INREG(Y, VT) - This operator takes an FP register, and
+ /// rounds it to a floating point value. It then promotes it and returns it
+ /// in a register of the same size. This operation effectively just
+ /// discards excess precision. The type to round down to is specified by
+ /// the VT operand, a VTSDNode.
FP_ROUND_INREG,
- // FP_EXTEND - Extend a smaller FP type into a larger FP type.
+ /// X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
FP_EXTEND,
// BIT_CONVERT - Theis operator converts between integer and FP values, as
// conversions, but that is a noop, deleted by getNode().
BIT_CONVERT,
- // FNEG, FABS, FSQRT, FSIN, FCOS, FPOWI - Perform unary floating point
- // negation, absolute value, square root, sine and cosine, and powi
+ // FNEG, FABS, FSQRT, FSIN, FCOS, FPOWI, FPOW - Perform unary floating point
+ // negation, absolute value, square root, sine and cosine, powi, and pow
// operations.
- FNEG, FABS, FSQRT, FSIN, FCOS, FPOWI,
+ FNEG, FABS, FSQRT, FSIN, FCOS, FPOWI, FPOW,
// LOAD and STORE have token chains as their first operand, then the same
// operands as an LLVM load/store instruction, then an offset node that
// is added / subtracted from the base pointer to form the address (for
// indexed memory ops).
LOAD, STORE,
-
- // TRUNCSTORE - This operators truncates (for integer) or rounds (for FP) a
- // value and stores it to memory in one operation. This can be used for
- // either integer or floating point operands. The first four operands of
- // this are the same as a standard store. The fifth is the ValueType to
- // store it as (which will be smaller than the source value).
- TRUNCSTORE,
// DYNAMIC_STACKALLOC - Allocate some number of bytes on the stack aligned
// to a specified boundary. This node always has two return values: a new
// returns a chain.
// Operand #0 : input chain.
// Operand #1 : module unique number use to identify the label.
+ // Operand #2 : 0 indicates a debug label (e.g. stoppoint), 1 indicates
+ // a EH label, 2 indicates unknown label type.
LABEL,
+
+ // DECLARE - Represents a llvm.dbg.declare intrinsic. It's used to track
+ // local variable declarations for debugging information. First operand is
+ // a chain, while the next two operands are first two arguments (address
+ // and variable) of a llvm.dbg.declare instruction.
+ DECLARE,
// STACKSAVE - STACKSAVE has one operand, an input chain. It produces a
// value, the same type as the pointer type for the system, and an output
// it returns an output chain.
STACKRESTORE,
- // MEMSET/MEMCPY/MEMMOVE - The first operand is the chain, and the rest
- // correspond to the operands of the LLVM intrinsic functions. The only
- // result is a token chain. The alignment argument is guaranteed to be a
- // Constant node.
+ // MEMSET/MEMCPY/MEMMOVE - The first operand is the chain. The following
+ // correspond to the operands of the LLVM intrinsic functions and the last
+ // one is AlwaysInline. The only result is a token chain. The alignment
+ // argument is guaranteed to be a Constant node.
MEMSET,
MEMMOVE,
MEMCPY,
// pointer, and a SRCVALUE.
VAEND, VASTART,
- // SRCVALUE - This corresponds to a Value*, and is used to associate memory
- // locations with their value. This allows one use alias analysis
- // information in the backend.
+ // SRCVALUE - This is a node type that holds a Value* that is used to
+ // make reference to a value in the LLVM IR.
SRCVALUE,
+ // MEMOPERAND - This is a node that contains a MemOperand which records
+ // information about a memory reference. This is used to make AliasAnalysis
+ // queries from the backend.
+ MEMOPERAND,
+
// PCMARKER - This corresponds to the pcmarker intrinsic.
PCMARKER,
// chain as output.
TRAMPOLINE,
+ // TRAP - Trapping instruction
+ TRAP,
+
+ // OUTCHAIN = MEMBARRIER(INCHAIN, load-load, load-store, store-load,
+ // store-store, device)
+ // This corresponds to the memory.barrier intrinsic.
+ // it takes an input chain, 4 operands to specify the type of barrier, an
+ // operand specifying if the barrier applies to device and uncached memory
+ // and produces an output chain.
+ MEMBARRIER,
+
+ // Val, OUTCHAIN = ATOMIC_LCS(INCHAIN, ptr, cmp, swap)
+ // this corresponds to the atomic.lcs intrinsic.
+ // cmp is compared to *ptr, and if equal, swap is stored in *ptr.
+ // the return is always the original value in *ptr
+ ATOMIC_LCS,
+
+ // Val, OUTCHAIN = ATOMIC_LAS(INCHAIN, ptr, amt)
+ // this corresponds to the atomic.las intrinsic.
+ // *ptr + amt is stored to *ptr atomically.
+ // the return is always the original value in *ptr
+ ATOMIC_LAS,
+
+ // Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt)
+ // this corresponds to the atomic.swap intrinsic.
+ // amt is stored to *ptr atomically.
+ // the return is always the original value in *ptr
+ ATOMIC_SWAP,
+
// BUILTIN_OP_END - This must be the last enum value in this list.
BUILTIN_OP_END
};
/// isBuildVectorAllZeros - Return true if the specified node is a
/// BUILD_VECTOR where all of the elements are 0 or undef.
bool isBuildVectorAllZeros(const SDNode *N);
+
+ /// isScalarToVector - Return true if the specified node is a
+ /// ISD::SCALAR_TO_VECTOR node or a BUILD_VECTOR node where only the low
+ /// element is not an undef.
+ bool isScalarToVector(const SDNode *N);
+
+ /// isDebugLabel - Return true if the specified node represents a debug
+ /// label (i.e. ISD::LABEL or TargetInstrInfo::LABEL node and third operand
+ /// is 0).
+ bool isDebugLabel(const SDNode *N);
//===--------------------------------------------------------------------===//
/// MemIndexedMode enum - This enum defines the load / store indexed
return SDOperand(Val, R);
}
- // isOperand - Return true if this node is an operand of N.
- bool isOperand(SDNode *N) const;
+ // isOperandOf - Return true if this node is an operand of N.
+ bool isOperandOf(SDNode *N) const;
/// getValueType - Return the ValueType of the referenced return value.
///
inline MVT::ValueType getValueType() const;
+ /// getValueSizeInBits - Returns MVT::getSizeInBits(getValueType()).
+ ///
+ unsigned getValueSizeInBits() const {
+ return MVT::getSizeInBits(getValueType());
+ }
+
// Forwarding methods - These forward to the corresponding methods in SDNode.
inline unsigned getOpcode() const;
inline unsigned getNumOperands() const;
inline bool isTargetOpcode() const;
inline unsigned getTargetOpcode() const;
+
+ /// reachesChainWithoutSideEffects - Return true if this operand (which must
+ /// be a chain) reaches the specified operand without crossing any
+ /// side-effecting instructions. In practice, this looks through token
+ /// factors and non-volatile loads. In order to remain efficient, this only
+ /// looks a couple of nodes in, it does not do an exhaustive search.
+ bool reachesChainWithoutSideEffects(SDOperand Dest, unsigned Depth = 2) const;
+
/// hasOneUse - Return true if there is exactly one operation using this
/// result value of the defining operator.
inline bool hasOneUse() const;
static inline SDOperand getEmptyKey() { return SDOperand((SDNode*)-1, -1U); }
static inline SDOperand getTombstoneKey() { return SDOperand((SDNode*)-1, 0);}
static unsigned getHashValue(const SDOperand &Val) {
- return (unsigned)((uintptr_t)Val.Val >> 4) ^
- (unsigned)((uintptr_t)Val.Val >> 9) + Val.ResNo;
+ return ((unsigned)((uintptr_t)Val.Val >> 4) ^
+ (unsigned)((uintptr_t)Val.Val >> 9)) + Val.ResNo;
}
static bool isEqual(const SDOperand &LHS, const SDOperand &RHS) {
return LHS == RHS;
///
int getNodeId() const { return NodeId; }
+ /// setNodeId - Set unique node id.
+ void setNodeId(int Id) { NodeId = Id; }
+
typedef SmallVector<SDNode*,3>::const_iterator use_iterator;
use_iterator use_begin() const { return Uses.begin(); }
use_iterator use_end() const { return Uses.end(); }
/// value. This method ignores uses of other values defined by this operation.
bool hasAnyUseOfValue(unsigned Value) const;
- /// isOnlyUse - Return true if this node is the only use of N.
+ /// isOnlyUseOf - Return true if this node is the only use of N.
///
- bool isOnlyUse(SDNode *N) const;
+ bool isOnlyUseOf(SDNode *N) const;
- /// isOperand - Return true if this node is an operand of N.
+ /// isOperandOf - Return true if this node is an operand of N.
///
- bool isOperand(SDNode *N) const;
+ bool isOperandOf(SDNode *N) const;
- /// isPredecessor - Return true if this node is a predecessor of N. This node
- /// is either an operand of N or it can be reached by recursively traversing
- /// up the operands.
+ /// isPredecessorOf - Return true if this node is a predecessor of N. This
+ /// node is either an operand of N or it can be reached by recursively
+ /// traversing up the operands.
/// NOTE: this is an expensive method. Use it carefully.
- bool isPredecessor(SDNode *N) const;
+ bool isPredecessorOf(SDNode *N) const;
/// getNumOperands - Return the number of values used by this operation.
///
return ValueList[ResNo];
}
+ /// getValueSizeInBits - Returns MVT::getSizeInBits(getValueType(ResNo)).
+ ///
+ unsigned getValueSizeInBits(unsigned ResNo) const {
+ return MVT::getSizeInBits(getValueType(ResNo));
+ }
+
typedef const MVT::ValueType* value_iterator;
value_iterator value_begin() const { return ValueList; }
value_iterator value_end() const { return ValueList+NumValues; }
/// getValueTypeList - Return a pointer to the specified value type.
///
- static MVT::ValueType *getValueTypeList(MVT::ValueType VT);
+ static const MVT::ValueType *getValueTypeList(MVT::ValueType VT);
static SDVTList getSDVTList(MVT::ValueType VT) {
SDVTList Ret = { getValueTypeList(VT), 1 };
return Ret;
}
}
}
-
- void setNodeId(int Id) {
- NodeId = Id;
- }
};
SDOperand getValue() const { return Op; }
};
+class AtomicSDNode : public SDNode {
+ virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
+ SDOperand Ops[4];
+ MVT::ValueType OrigVT;
+public:
+ AtomicSDNode(unsigned Opc, SDVTList VTL, SDOperand Chain, SDOperand Ptr,
+ SDOperand Cmp, SDOperand Swp, MVT::ValueType VT)
+ : SDNode(Opc, VTL) {
+ Ops[0] = Chain;
+ Ops[1] = Ptr;
+ Ops[2] = Swp;
+ Ops[3] = Cmp;
+ InitOperands(Ops, 4);
+ OrigVT=VT;
+ }
+ AtomicSDNode(unsigned Opc, SDVTList VTL, SDOperand Chain, SDOperand Ptr,
+ SDOperand Val, MVT::ValueType VT)
+ : SDNode(Opc, VTL) {
+ Ops[0] = Chain;
+ Ops[1] = Ptr;
+ Ops[2] = Val;
+ InitOperands(Ops, 3);
+ OrigVT=VT;
+ }
+ MVT::ValueType getVT() const { return OrigVT; }
+ bool isCompareAndSwap() const { return getOpcode() == ISD::ATOMIC_LCS; }
+};
+
class StringSDNode : public SDNode {
std::string Value;
virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
};
class ConstantSDNode : public SDNode {
- uint64_t Value;
+ APInt Value;
virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
protected:
friend class SelectionDAG;
- ConstantSDNode(bool isTarget, uint64_t val, MVT::ValueType VT)
+ ConstantSDNode(bool isTarget, const APInt &val, MVT::ValueType VT)
: SDNode(isTarget ? ISD::TargetConstant : ISD::Constant, getSDVTList(VT)),
Value(val) {
}
public:
- uint64_t getValue() const { return Value; }
+ const APInt &getAPIntValue() const { return Value; }
+ uint64_t getValue() const { return Value.getZExtValue(); }
int64_t getSignExtended() const {
unsigned Bits = MVT::getSizeInBits(getValueType(0));
- return ((int64_t)Value << (64-Bits)) >> (64-Bits);
+ return ((int64_t)Value.getZExtValue() << (64-Bits)) >> (64-Bits);
}
bool isNullValue() const { return Value == 0; }
class ConstantFPSDNode : public SDNode {
APFloat Value;
virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
- // Longterm plan: replace all uses of getValue with getValueAPF, remove
- // getValue, rename getValueAPF to getValue.
protected:
friend class SelectionDAG;
ConstantFPSDNode(bool isTarget, const APFloat& val, MVT::ValueType VT)
/// convenient to write "2.0" and the like. Without this function we'd
/// have to duplicate its logic everywhere it's called.
bool isExactlyValue(double V) const {
- if (getValueType(0)==MVT::f64)
- return isExactlyValue(APFloat(V));
- else
- return isExactlyValue(APFloat((float)V));
+ APFloat Tmp(V);
+ Tmp.convert(Value.getSemantics(), APFloat::rmNearestTiesToEven);
+ return isExactlyValue(Tmp);
}
bool isExactlyValue(const APFloat& V) const;
}
public:
- int getIndex() const { return JTI; }
+ int getIndex() const { return JTI; }
static bool classof(const JumpTableSDNode *) { return true; }
static bool classof(const SDNode *N) {
}
};
+/// SrcValueSDNode - An SDNode that holds an arbitrary LLVM IR Value. This is
+/// used when the SelectionDAG needs to make a simple reference to something
+/// in the LLVM IR representation.
+///
+/// Note that this is not used for carrying alias information; that is done
+/// with MemOperandSDNode, which includes a Value which is required to be a
+/// pointer, and several other fields specific to memory references.
+///
class SrcValueSDNode : public SDNode {
const Value *V;
- int offset;
virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
protected:
friend class SelectionDAG;
- SrcValueSDNode(const Value* v, int o)
- : SDNode(ISD::SRCVALUE, getSDVTList(MVT::Other)), V(v), offset(o) {
- }
+ /// Create a SrcValue for a general value.
+ explicit SrcValueSDNode(const Value *v)
+ : SDNode(ISD::SRCVALUE, getSDVTList(MVT::Other)), V(v) {}
public:
+ /// getValue - return the contained Value.
const Value *getValue() const { return V; }
- int getOffset() const { return offset; }
static bool classof(const SrcValueSDNode *) { return true; }
static bool classof(const SDNode *N) {
};
+/// MemOperandSDNode - An SDNode that holds a MemOperand. This is
+/// used to represent a reference to memory after ISD::LOAD
+/// and ISD::STORE have been lowered.
+///
+class MemOperandSDNode : public SDNode {
+ virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
+protected:
+ friend class SelectionDAG;
+ /// Create a MemOperand node
+ explicit MemOperandSDNode(const MemOperand &mo)
+ : SDNode(ISD::MEMOPERAND, getSDVTList(MVT::Other)), MO(mo) {}
+
+public:
+ /// MO - The contained MemOperand.
+ const MemOperand MO;
+
+ static bool classof(const MemOperandSDNode *) { return true; }
+ static bool classof(const SDNode *N) {
+ return N->getOpcode() == ISD::MEMOPERAND;
+ }
+};
+
+
class RegisterSDNode : public SDNode {
unsigned Reg;
virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
}
};
-/// LoadSDNode - This class is used to represent ISD::LOAD nodes.
+/// LSBaseSDNode - Base class for LoadSDNode and StoreSDNode
///
-class LoadSDNode : public SDNode {
- virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
- SDOperand Ops[3];
-
+class LSBaseSDNode : public SDNode {
+private:
// AddrMode - unindexed, pre-indexed, post-indexed.
ISD::MemIndexedMode AddrMode;
- // ExtType - non-ext, anyext, sext, zext.
- ISD::LoadExtType ExtType;
-
- // LoadedVT - VT of loaded value before extension.
- MVT::ValueType LoadedVT;
+ // MemoryVT - VT of in-memory value.
+ MVT::ValueType MemoryVT;
- // SrcValue - Memory location for alias analysis.
+ //! SrcValue - Memory location for alias analysis.
const Value *SrcValue;
- // SVOffset - Memory location offset.
+ //! SVOffset - Memory location offset.
int SVOffset;
- // Alignment - Alignment of memory location in bytes.
+ //! Alignment - Alignment of memory location in bytes.
unsigned Alignment;
- // IsVolatile - True if the load is volatile.
+ //! IsVolatile - True if the store is volatile.
bool IsVolatile;
protected:
- friend class SelectionDAG;
- LoadSDNode(SDOperand *ChainPtrOff, SDVTList VTs,
- ISD::MemIndexedMode AM, ISD::LoadExtType ETy, MVT::ValueType LVT,
- const Value *SV, int O=0, unsigned Align=0, bool Vol=false)
- : SDNode(ISD::LOAD, VTs),
- AddrMode(AM), ExtType(ETy), LoadedVT(LVT), SrcValue(SV), SVOffset(O),
- Alignment(Align), IsVolatile(Vol) {
- Ops[0] = ChainPtrOff[0]; // Chain
- Ops[1] = ChainPtrOff[1]; // Ptr
- Ops[2] = ChainPtrOff[2]; // Off
- InitOperands(Ops, 3);
- assert(Align != 0 && "Loads should have non-zero aligment");
- assert((getOffset().getOpcode() == ISD::UNDEF ||
- AddrMode != ISD::UNINDEXED) &&
- "Only indexed load has a non-undef offset operand");
- }
+ //! Operand array for load and store
+ /*!
+ \note Moving this array to the base class captures more
+ common functionality shared between LoadSDNode and
+ StoreSDNode
+ */
+ SDOperand Ops[4];
public:
+ LSBaseSDNode(ISD::NodeType NodeTy, SDOperand *Operands, unsigned NumOperands,
+ SDVTList VTs, ISD::MemIndexedMode AM, MVT::ValueType VT,
+ const Value *SV, int SVO, unsigned Align, bool Vol)
+ : SDNode(NodeTy, VTs),
+ AddrMode(AM), MemoryVT(VT),
+ SrcValue(SV), SVOffset(SVO), Alignment(Align), IsVolatile(Vol) {
+ for (unsigned i = 0; i != NumOperands; ++i)
+ Ops[i] = Operands[i];
+ InitOperands(Ops, NumOperands);
+ assert(Align != 0 && "Loads and stores should have non-zero aligment");
+ assert((getOffset().getOpcode() == ISD::UNDEF || isIndexed()) &&
+ "Only indexed loads and stores have a non-undef offset operand");
+ }
+
+ const SDOperand &getChain() const { return getOperand(0); }
+ const SDOperand &getBasePtr() const {
+ return getOperand(getOpcode() == ISD::LOAD ? 1 : 2);
+ }
+ const SDOperand &getOffset() const {
+ return getOperand(getOpcode() == ISD::LOAD ? 2 : 3);
+ }
- const SDOperand getChain() const { return getOperand(0); }
- const SDOperand getBasePtr() const { return getOperand(1); }
- const SDOperand getOffset() const { return getOperand(2); }
- ISD::MemIndexedMode getAddressingMode() const { return AddrMode; }
- ISD::LoadExtType getExtensionType() const { return ExtType; }
- MVT::ValueType getLoadedVT() const { return LoadedVT; }
const Value *getSrcValue() const { return SrcValue; }
int getSrcValueOffset() const { return SVOffset; }
unsigned getAlignment() const { return Alignment; }
+ MVT::ValueType getMemoryVT() const { return MemoryVT; }
bool isVolatile() const { return IsVolatile; }
+ ISD::MemIndexedMode getAddressingMode() const { return AddrMode; }
+
+ /// isIndexed - Return true if this is a pre/post inc/dec load/store.
+ bool isIndexed() const { return AddrMode != ISD::UNINDEXED; }
+
+ /// isUnindexed - Return true if this is NOT a pre/post inc/dec load/store.
+ bool isUnindexed() const { return AddrMode == ISD::UNINDEXED; }
+
+ /// getMemOperand - Return a MemOperand object describing the memory
+ /// reference performed by this load or store.
+ MemOperand getMemOperand() const;
+
+ static bool classof(const LSBaseSDNode *N) { return true; }
+ static bool classof(const SDNode *N) {
+ return N->getOpcode() == ISD::LOAD ||
+ N->getOpcode() == ISD::STORE;
+ }
+};
+
+/// LoadSDNode - This class is used to represent ISD::LOAD nodes.
+///
+class LoadSDNode : public LSBaseSDNode {
+ virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
+
+ // ExtType - non-ext, anyext, sext, zext.
+ ISD::LoadExtType ExtType;
+
+protected:
+ friend class SelectionDAG;
+ LoadSDNode(SDOperand *ChainPtrOff, SDVTList VTs,
+ ISD::MemIndexedMode AM, ISD::LoadExtType ETy, MVT::ValueType LVT,
+ const Value *SV, int O=0, unsigned Align=0, bool Vol=false)
+ : LSBaseSDNode(ISD::LOAD, ChainPtrOff, 3,
+ VTs, AM, LVT, SV, O, Align, Vol),
+ ExtType(ETy) {}
+public:
+
+ ISD::LoadExtType getExtensionType() const { return ExtType; }
+ const SDOperand &getBasePtr() const { return getOperand(1); }
+ const SDOperand &getOffset() const { return getOperand(2); }
+
static bool classof(const LoadSDNode *) { return true; }
static bool classof(const SDNode *N) {
return N->getOpcode() == ISD::LOAD;
/// StoreSDNode - This class is used to represent ISD::STORE nodes.
///
-class StoreSDNode : public SDNode {
+class StoreSDNode : public LSBaseSDNode {
virtual void ANCHOR(); // Out-of-line virtual method to give class a home.
- SDOperand Ops[4];
- // AddrMode - unindexed, pre-indexed, post-indexed.
- ISD::MemIndexedMode AddrMode;
-
- // IsTruncStore - True is the op does a truncation before store.
+ // IsTruncStore - True if the op does a truncation before store.
bool IsTruncStore;
-
- // StoredVT - VT of the value after truncation.
- MVT::ValueType StoredVT;
-
- // SrcValue - Memory location for alias analysis.
- const Value *SrcValue;
-
- // SVOffset - Memory location offset.
- int SVOffset;
-
- // Alignment - Alignment of memory location in bytes.
- unsigned Alignment;
-
- // IsVolatile - True if the store is volatile.
- bool IsVolatile;
protected:
friend class SelectionDAG;
StoreSDNode(SDOperand *ChainValuePtrOff, SDVTList VTs,
ISD::MemIndexedMode AM, bool isTrunc, MVT::ValueType SVT,
const Value *SV, int O=0, unsigned Align=0, bool Vol=false)
- : SDNode(ISD::STORE, VTs),
- AddrMode(AM), IsTruncStore(isTrunc), StoredVT(SVT), SrcValue(SV),
- SVOffset(O), Alignment(Align), IsVolatile(Vol) {
- Ops[0] = ChainValuePtrOff[0]; // Chain
- Ops[1] = ChainValuePtrOff[1]; // Value
- Ops[2] = ChainValuePtrOff[2]; // Ptr
- Ops[3] = ChainValuePtrOff[3]; // Off
- InitOperands(Ops, 4);
- assert(Align != 0 && "Stores should have non-zero aligment");
- assert((getOffset().getOpcode() == ISD::UNDEF ||
- AddrMode != ISD::UNINDEXED) &&
- "Only indexed store has a non-undef offset operand");
- }
+ : LSBaseSDNode(ISD::STORE, ChainValuePtrOff, 4,
+ VTs, AM, SVT, SV, O, Align, Vol),
+ IsTruncStore(isTrunc) {}
public:
- const SDOperand getChain() const { return getOperand(0); }
- const SDOperand getValue() const { return getOperand(1); }
- const SDOperand getBasePtr() const { return getOperand(2); }
- const SDOperand getOffset() const { return getOperand(3); }
- ISD::MemIndexedMode getAddressingMode() const { return AddrMode; }
bool isTruncatingStore() const { return IsTruncStore; }
- MVT::ValueType getStoredVT() const { return StoredVT; }
- const Value *getSrcValue() const { return SrcValue; }
- int getSrcValueOffset() const { return SVOffset; }
- unsigned getAlignment() const { return Alignment; }
- bool isVolatile() const { return IsVolatile; }
-
+ const SDOperand &getValue() const { return getOperand(1); }
+ const SDOperand &getBasePtr() const { return getOperand(2); }
+ const SDOperand &getOffset() const { return getOperand(3); }
+
static bool classof(const StoreSDNode *) { return true; }
static bool classof(const SDNode *N) {
return N->getOpcode() == ISD::STORE;
};
namespace ISD {
+ /// isNormalLoad - Returns true if the specified node is a non-extending
+ /// and unindexed load.
+ inline bool isNormalLoad(const SDNode *N) {
+ if (N->getOpcode() != ISD::LOAD)
+ return false;
+ const LoadSDNode *Ld = cast<LoadSDNode>(N);
+ return Ld->getExtensionType() == ISD::NON_EXTLOAD &&
+ Ld->getAddressingMode() == ISD::UNINDEXED;
+ }
+
/// isNON_EXTLoad - Returns true if the specified node is a non-extending
/// load.
inline bool isNON_EXTLoad(const SDNode *N) {