1 //===-- Instruction.cpp - Implement the Instruction class -----------------===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 // This file implements the Instruction class for the IR library.
12 //===----------------------------------------------------------------------===//
14 #include "llvm/IR/Instruction.h"
15 #include "llvm/IR/CallSite.h"
16 #include "llvm/IR/Constants.h"
17 #include "llvm/IR/Instructions.h"
18 #include "llvm/IR/LeakDetector.h"
19 #include "llvm/IR/Module.h"
20 #include "llvm/IR/Operator.h"
21 #include "llvm/IR/Type.h"
24 Instruction::Instruction(Type *ty, unsigned it, Use *Ops, unsigned NumOps,
25 Instruction *InsertBefore)
26 : User(ty, Value::InstructionVal + it, Ops, NumOps), Parent(0) {
27 // Make sure that we get added to a basicblock
28 LeakDetector::addGarbageObject(this);
30 // If requested, insert this instruction into a basic block...
32 assert(InsertBefore->getParent() &&
33 "Instruction to insert before is not in a basic block!");
34 InsertBefore->getParent()->getInstList().insert(InsertBefore, this);
38 const DataLayout *Instruction::getDataLayout() const {
39 return getParent()->getDataLayout();
42 Instruction::Instruction(Type *ty, unsigned it, Use *Ops, unsigned NumOps,
43 BasicBlock *InsertAtEnd)
44 : User(ty, Value::InstructionVal + it, Ops, NumOps), Parent(0) {
45 // Make sure that we get added to a basicblock
46 LeakDetector::addGarbageObject(this);
48 // append this instruction into the basic block
49 assert(InsertAtEnd && "Basic block to append to may not be NULL!");
50 InsertAtEnd->getInstList().push_back(this);
54 // Out of line virtual method, so the vtable, etc has a home.
55 Instruction::~Instruction() {
56 assert(Parent == 0 && "Instruction still linked in the program!");
57 if (hasMetadataHashEntry())
58 clearMetadataHashEntries();
62 void Instruction::setParent(BasicBlock *P) {
64 if (!P) LeakDetector::addGarbageObject(this);
66 if (P) LeakDetector::removeGarbageObject(this);
72 void Instruction::removeFromParent() {
73 getParent()->getInstList().remove(this);
76 void Instruction::eraseFromParent() {
77 getParent()->getInstList().erase(this);
80 /// insertBefore - Insert an unlinked instructions into a basic block
81 /// immediately before the specified instruction.
82 void Instruction::insertBefore(Instruction *InsertPos) {
83 InsertPos->getParent()->getInstList().insert(InsertPos, this);
86 /// insertAfter - Insert an unlinked instructions into a basic block
87 /// immediately after the specified instruction.
88 void Instruction::insertAfter(Instruction *InsertPos) {
89 InsertPos->getParent()->getInstList().insertAfter(InsertPos, this);
92 /// moveBefore - Unlink this instruction from its current basic block and
93 /// insert it into the basic block that MovePos lives in, right before
95 void Instruction::moveBefore(Instruction *MovePos) {
96 MovePos->getParent()->getInstList().splice(MovePos,getParent()->getInstList(),
100 /// Set or clear the unsafe-algebra flag on this instruction, which must be an
101 /// operator which supports this flag. See LangRef.html for the meaning of this
103 void Instruction::setHasUnsafeAlgebra(bool B) {
104 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
105 cast<FPMathOperator>(this)->setHasUnsafeAlgebra(B);
108 /// Set or clear the NoNaNs flag on this instruction, which must be an operator
109 /// which supports this flag. See LangRef.html for the meaning of this flag.
110 void Instruction::setHasNoNaNs(bool B) {
111 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
112 cast<FPMathOperator>(this)->setHasNoNaNs(B);
115 /// Set or clear the no-infs flag on this instruction, which must be an operator
116 /// which supports this flag. See LangRef.html for the meaning of this flag.
117 void Instruction::setHasNoInfs(bool B) {
118 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
119 cast<FPMathOperator>(this)->setHasNoInfs(B);
122 /// Set or clear the no-signed-zeros flag on this instruction, which must be an
123 /// operator which supports this flag. See LangRef.html for the meaning of this
125 void Instruction::setHasNoSignedZeros(bool B) {
126 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
127 cast<FPMathOperator>(this)->setHasNoSignedZeros(B);
130 /// Set or clear the allow-reciprocal flag on this instruction, which must be an
131 /// operator which supports this flag. See LangRef.html for the meaning of this
133 void Instruction::setHasAllowReciprocal(bool B) {
134 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
135 cast<FPMathOperator>(this)->setHasAllowReciprocal(B);
138 /// Convenience function for setting all the fast-math flags on this
139 /// instruction, which must be an operator which supports these flags. See
140 /// LangRef.html for the meaning of these flats.
141 void Instruction::setFastMathFlags(FastMathFlags FMF) {
142 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
143 cast<FPMathOperator>(this)->setFastMathFlags(FMF);
146 /// Determine whether the unsafe-algebra flag is set.
147 bool Instruction::hasUnsafeAlgebra() const {
148 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
149 return cast<FPMathOperator>(this)->hasUnsafeAlgebra();
152 /// Determine whether the no-NaNs flag is set.
153 bool Instruction::hasNoNaNs() const {
154 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
155 return cast<FPMathOperator>(this)->hasNoNaNs();
158 /// Determine whether the no-infs flag is set.
159 bool Instruction::hasNoInfs() const {
160 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
161 return cast<FPMathOperator>(this)->hasNoInfs();
164 /// Determine whether the no-signed-zeros flag is set.
165 bool Instruction::hasNoSignedZeros() const {
166 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
167 return cast<FPMathOperator>(this)->hasNoSignedZeros();
170 /// Determine whether the allow-reciprocal flag is set.
171 bool Instruction::hasAllowReciprocal() const {
172 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
173 return cast<FPMathOperator>(this)->hasAllowReciprocal();
176 /// Convenience function for getting all the fast-math flags, which must be an
177 /// operator which supports these flags. See LangRef.html for the meaning of
179 FastMathFlags Instruction::getFastMathFlags() const {
180 assert(isa<FPMathOperator>(this) && "setting fast-math flag on invalid op");
181 return cast<FPMathOperator>(this)->getFastMathFlags();
184 /// Copy I's fast-math flags
185 void Instruction::copyFastMathFlags(const Instruction *I) {
186 setFastMathFlags(I->getFastMathFlags());
190 const char *Instruction::getOpcodeName(unsigned OpCode) {
193 case Ret: return "ret";
194 case Br: return "br";
195 case Switch: return "switch";
196 case IndirectBr: return "indirectbr";
197 case Invoke: return "invoke";
198 case Resume: return "resume";
199 case Unreachable: return "unreachable";
201 // Standard binary operators...
202 case Add: return "add";
203 case FAdd: return "fadd";
204 case Sub: return "sub";
205 case FSub: return "fsub";
206 case Mul: return "mul";
207 case FMul: return "fmul";
208 case UDiv: return "udiv";
209 case SDiv: return "sdiv";
210 case FDiv: return "fdiv";
211 case URem: return "urem";
212 case SRem: return "srem";
213 case FRem: return "frem";
215 // Logical operators...
216 case And: return "and";
217 case Or : return "or";
218 case Xor: return "xor";
220 // Memory instructions...
221 case Alloca: return "alloca";
222 case Load: return "load";
223 case Store: return "store";
224 case AtomicCmpXchg: return "cmpxchg";
225 case AtomicRMW: return "atomicrmw";
226 case Fence: return "fence";
227 case GetElementPtr: return "getelementptr";
229 // Convert instructions...
230 case Trunc: return "trunc";
231 case ZExt: return "zext";
232 case SExt: return "sext";
233 case FPTrunc: return "fptrunc";
234 case FPExt: return "fpext";
235 case FPToUI: return "fptoui";
236 case FPToSI: return "fptosi";
237 case UIToFP: return "uitofp";
238 case SIToFP: return "sitofp";
239 case IntToPtr: return "inttoptr";
240 case PtrToInt: return "ptrtoint";
241 case BitCast: return "bitcast";
242 case AddrSpaceCast: return "addrspacecast";
244 // Other instructions...
245 case ICmp: return "icmp";
246 case FCmp: return "fcmp";
247 case PHI: return "phi";
248 case Select: return "select";
249 case Call: return "call";
250 case Shl: return "shl";
251 case LShr: return "lshr";
252 case AShr: return "ashr";
253 case VAArg: return "va_arg";
254 case ExtractElement: return "extractelement";
255 case InsertElement: return "insertelement";
256 case ShuffleVector: return "shufflevector";
257 case ExtractValue: return "extractvalue";
258 case InsertValue: return "insertvalue";
259 case LandingPad: return "landingpad";
261 default: return "<Invalid operator> ";
265 /// isIdenticalTo - Return true if the specified instruction is exactly
266 /// identical to the current one. This means that all operands match and any
267 /// extra information (e.g. load is volatile) agree.
268 bool Instruction::isIdenticalTo(const Instruction *I) const {
269 return isIdenticalToWhenDefined(I) &&
270 SubclassOptionalData == I->SubclassOptionalData;
273 /// isIdenticalToWhenDefined - This is like isIdenticalTo, except that it
274 /// ignores the SubclassOptionalData flags, which specify conditions
275 /// under which the instruction's result is undefined.
276 bool Instruction::isIdenticalToWhenDefined(const Instruction *I) const {
277 if (getOpcode() != I->getOpcode() ||
278 getNumOperands() != I->getNumOperands() ||
279 getType() != I->getType())
282 // We have two instructions of identical opcode and #operands. Check to see
283 // if all operands are the same.
284 if (!std::equal(op_begin(), op_end(), I->op_begin()))
287 // Check special state that is a part of some instructions.
288 if (const LoadInst *LI = dyn_cast<LoadInst>(this))
289 return LI->isVolatile() == cast<LoadInst>(I)->isVolatile() &&
290 LI->getAlignment() == cast<LoadInst>(I)->getAlignment() &&
291 LI->getOrdering() == cast<LoadInst>(I)->getOrdering() &&
292 LI->getSynchScope() == cast<LoadInst>(I)->getSynchScope();
293 if (const StoreInst *SI = dyn_cast<StoreInst>(this))
294 return SI->isVolatile() == cast<StoreInst>(I)->isVolatile() &&
295 SI->getAlignment() == cast<StoreInst>(I)->getAlignment() &&
296 SI->getOrdering() == cast<StoreInst>(I)->getOrdering() &&
297 SI->getSynchScope() == cast<StoreInst>(I)->getSynchScope();
298 if (const CmpInst *CI = dyn_cast<CmpInst>(this))
299 return CI->getPredicate() == cast<CmpInst>(I)->getPredicate();
300 if (const CallInst *CI = dyn_cast<CallInst>(this))
301 return CI->isTailCall() == cast<CallInst>(I)->isTailCall() &&
302 CI->getCallingConv() == cast<CallInst>(I)->getCallingConv() &&
303 CI->getAttributes() == cast<CallInst>(I)->getAttributes();
304 if (const InvokeInst *CI = dyn_cast<InvokeInst>(this))
305 return CI->getCallingConv() == cast<InvokeInst>(I)->getCallingConv() &&
306 CI->getAttributes() == cast<InvokeInst>(I)->getAttributes();
307 if (const InsertValueInst *IVI = dyn_cast<InsertValueInst>(this))
308 return IVI->getIndices() == cast<InsertValueInst>(I)->getIndices();
309 if (const ExtractValueInst *EVI = dyn_cast<ExtractValueInst>(this))
310 return EVI->getIndices() == cast<ExtractValueInst>(I)->getIndices();
311 if (const FenceInst *FI = dyn_cast<FenceInst>(this))
312 return FI->getOrdering() == cast<FenceInst>(FI)->getOrdering() &&
313 FI->getSynchScope() == cast<FenceInst>(FI)->getSynchScope();
314 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(this))
315 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I)->isVolatile() &&
316 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I)->getOrdering() &&
317 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I)->getSynchScope();
318 if (const AtomicRMWInst *RMWI = dyn_cast<AtomicRMWInst>(this))
319 return RMWI->getOperation() == cast<AtomicRMWInst>(I)->getOperation() &&
320 RMWI->isVolatile() == cast<AtomicRMWInst>(I)->isVolatile() &&
321 RMWI->getOrdering() == cast<AtomicRMWInst>(I)->getOrdering() &&
322 RMWI->getSynchScope() == cast<AtomicRMWInst>(I)->getSynchScope();
323 if (const PHINode *thisPHI = dyn_cast<PHINode>(this)) {
324 const PHINode *otherPHI = cast<PHINode>(I);
325 return std::equal(thisPHI->block_begin(), thisPHI->block_end(),
326 otherPHI->block_begin());
332 // This should be kept in sync with isEquivalentOperation in
333 // lib/Transforms/IPO/MergeFunctions.cpp.
334 bool Instruction::isSameOperationAs(const Instruction *I,
335 unsigned flags) const {
336 bool IgnoreAlignment = flags & CompareIgnoringAlignment;
337 bool UseScalarTypes = flags & CompareUsingScalarTypes;
339 if (getOpcode() != I->getOpcode() ||
340 getNumOperands() != I->getNumOperands() ||
342 getType()->getScalarType() != I->getType()->getScalarType() :
343 getType() != I->getType()))
346 // We have two instructions of identical opcode and #operands. Check to see
347 // if all operands are the same type
348 for (unsigned i = 0, e = getNumOperands(); i != e; ++i)
350 getOperand(i)->getType()->getScalarType() !=
351 I->getOperand(i)->getType()->getScalarType() :
352 getOperand(i)->getType() != I->getOperand(i)->getType())
355 // Check special state that is a part of some instructions.
356 if (const LoadInst *LI = dyn_cast<LoadInst>(this))
357 return LI->isVolatile() == cast<LoadInst>(I)->isVolatile() &&
358 (LI->getAlignment() == cast<LoadInst>(I)->getAlignment() ||
360 LI->getOrdering() == cast<LoadInst>(I)->getOrdering() &&
361 LI->getSynchScope() == cast<LoadInst>(I)->getSynchScope();
362 if (const StoreInst *SI = dyn_cast<StoreInst>(this))
363 return SI->isVolatile() == cast<StoreInst>(I)->isVolatile() &&
364 (SI->getAlignment() == cast<StoreInst>(I)->getAlignment() ||
366 SI->getOrdering() == cast<StoreInst>(I)->getOrdering() &&
367 SI->getSynchScope() == cast<StoreInst>(I)->getSynchScope();
368 if (const CmpInst *CI = dyn_cast<CmpInst>(this))
369 return CI->getPredicate() == cast<CmpInst>(I)->getPredicate();
370 if (const CallInst *CI = dyn_cast<CallInst>(this))
371 return CI->isTailCall() == cast<CallInst>(I)->isTailCall() &&
372 CI->getCallingConv() == cast<CallInst>(I)->getCallingConv() &&
373 CI->getAttributes() == cast<CallInst>(I)->getAttributes();
374 if (const InvokeInst *CI = dyn_cast<InvokeInst>(this))
375 return CI->getCallingConv() == cast<InvokeInst>(I)->getCallingConv() &&
376 CI->getAttributes() ==
377 cast<InvokeInst>(I)->getAttributes();
378 if (const InsertValueInst *IVI = dyn_cast<InsertValueInst>(this))
379 return IVI->getIndices() == cast<InsertValueInst>(I)->getIndices();
380 if (const ExtractValueInst *EVI = dyn_cast<ExtractValueInst>(this))
381 return EVI->getIndices() == cast<ExtractValueInst>(I)->getIndices();
382 if (const FenceInst *FI = dyn_cast<FenceInst>(this))
383 return FI->getOrdering() == cast<FenceInst>(I)->getOrdering() &&
384 FI->getSynchScope() == cast<FenceInst>(I)->getSynchScope();
385 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(this))
386 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I)->isVolatile() &&
387 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I)->getOrdering() &&
388 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I)->getSynchScope();
389 if (const AtomicRMWInst *RMWI = dyn_cast<AtomicRMWInst>(this))
390 return RMWI->getOperation() == cast<AtomicRMWInst>(I)->getOperation() &&
391 RMWI->isVolatile() == cast<AtomicRMWInst>(I)->isVolatile() &&
392 RMWI->getOrdering() == cast<AtomicRMWInst>(I)->getOrdering() &&
393 RMWI->getSynchScope() == cast<AtomicRMWInst>(I)->getSynchScope();
398 /// isUsedOutsideOfBlock - Return true if there are any uses of I outside of the
399 /// specified block. Note that PHI nodes are considered to evaluate their
400 /// operands in the corresponding predecessor block.
401 bool Instruction::isUsedOutsideOfBlock(const BasicBlock *BB) const {
402 for (const Use &U : uses()) {
403 // PHI nodes uses values in the corresponding predecessor block. For other
404 // instructions, just check to see whether the parent of the use matches up.
405 const Instruction *I = cast<Instruction>(U.getUser());
406 const PHINode *PN = dyn_cast<PHINode>(I);
408 if (I->getParent() != BB)
413 if (PN->getIncomingBlock(U) != BB)
419 /// mayReadFromMemory - Return true if this instruction may read memory.
421 bool Instruction::mayReadFromMemory() const {
422 switch (getOpcode()) {
423 default: return false;
424 case Instruction::VAArg:
425 case Instruction::Load:
426 case Instruction::Fence: // FIXME: refine definition of mayReadFromMemory
427 case Instruction::AtomicCmpXchg:
428 case Instruction::AtomicRMW:
430 case Instruction::Call:
431 return !cast<CallInst>(this)->doesNotAccessMemory();
432 case Instruction::Invoke:
433 return !cast<InvokeInst>(this)->doesNotAccessMemory();
434 case Instruction::Store:
435 return !cast<StoreInst>(this)->isUnordered();
439 /// mayWriteToMemory - Return true if this instruction may modify memory.
441 bool Instruction::mayWriteToMemory() const {
442 switch (getOpcode()) {
443 default: return false;
444 case Instruction::Fence: // FIXME: refine definition of mayWriteToMemory
445 case Instruction::Store:
446 case Instruction::VAArg:
447 case Instruction::AtomicCmpXchg:
448 case Instruction::AtomicRMW:
450 case Instruction::Call:
451 return !cast<CallInst>(this)->onlyReadsMemory();
452 case Instruction::Invoke:
453 return !cast<InvokeInst>(this)->onlyReadsMemory();
454 case Instruction::Load:
455 return !cast<LoadInst>(this)->isUnordered();
459 bool Instruction::mayThrow() const {
460 if (const CallInst *CI = dyn_cast<CallInst>(this))
461 return !CI->doesNotThrow();
462 return isa<ResumeInst>(this);
465 bool Instruction::mayReturn() const {
466 if (const CallInst *CI = dyn_cast<CallInst>(this))
467 return !CI->doesNotReturn();
471 /// isAssociative - Return true if the instruction is associative:
473 /// Associative operators satisfy: x op (y op z) === (x op y) op z
475 /// In LLVM, the Add, Mul, And, Or, and Xor operators are associative.
477 bool Instruction::isAssociative(unsigned Opcode) {
478 return Opcode == And || Opcode == Or || Opcode == Xor ||
479 Opcode == Add || Opcode == Mul;
482 bool Instruction::isAssociative() const {
483 unsigned Opcode = getOpcode();
484 if (isAssociative(Opcode))
490 return cast<FPMathOperator>(this)->hasUnsafeAlgebra();
496 /// isCommutative - Return true if the instruction is commutative:
498 /// Commutative operators satisfy: (x op y) === (y op x)
500 /// In LLVM, these are the associative operators, plus SetEQ and SetNE, when
501 /// applied to any type.
503 bool Instruction::isCommutative(unsigned op) {
518 /// isIdempotent - Return true if the instruction is idempotent:
520 /// Idempotent operators satisfy: x op x === x
522 /// In LLVM, the And and Or operators are idempotent.
524 bool Instruction::isIdempotent(unsigned Opcode) {
525 return Opcode == And || Opcode == Or;
528 /// isNilpotent - Return true if the instruction is nilpotent:
530 /// Nilpotent operators satisfy: x op x === Id,
532 /// where Id is the identity for the operator, i.e. a constant such that
533 /// x op Id === x and Id op x === x for all x.
535 /// In LLVM, the Xor operator is nilpotent.
537 bool Instruction::isNilpotent(unsigned Opcode) {
538 return Opcode == Xor;
541 Instruction *Instruction::clone() const {
542 Instruction *New = clone_impl();
543 New->SubclassOptionalData = SubclassOptionalData;
547 // Otherwise, enumerate and copy over metadata from the old instruction to the
549 SmallVector<std::pair<unsigned, MDNode*>, 4> TheMDs;
550 getAllMetadataOtherThanDebugLoc(TheMDs);
551 for (const auto &MD : TheMDs)
552 New->setMetadata(MD.first, MD.second);
554 New->setDebugLoc(getDebugLoc());