X-Git-Url: http://demsky.eecs.uci.edu/git/?a=blobdiff_plain;f=CDSPass.cpp;h=a33738a96d26880395afc7beaf3b284e1bf1e705;hb=HEAD;hp=d550140dff5639f147009f15bf998738318bfbfc;hpb=8aff644e1374177f7f1423adf500f518eef64674;p=c11llvm.git diff --git a/CDSPass.cpp b/CDSPass.cpp index d550140..a33738a 100644 --- a/CDSPass.cpp +++ b/CDSPass.cpp @@ -25,10 +25,12 @@ #include "llvm/ADT/SmallString.h" #include "llvm/Analysis/ValueTracking.h" #include "llvm/Analysis/CaptureTracking.h" +#include "llvm/Analysis/LoopInfo.h" #include "llvm/IR/BasicBlock.h" #include "llvm/IR/Function.h" #include "llvm/IR/IRBuilder.h" #include "llvm/IR/Instructions.h" +#include "llvm/IR/IntrinsicInst.h" #include "llvm/IR/LLVMContext.h" #include "llvm/IR/LegacyPassManager.h" #include "llvm/IR/Module.h" @@ -39,18 +41,18 @@ #include "llvm/Support/AtomicOrdering.h" #include "llvm/Support/Debug.h" #include "llvm/Transforms/Scalar.h" -#include "llvm/Transforms/Utils/Local.h" #include "llvm/Transforms/Utils/BasicBlockUtils.h" -#include "llvm/Transforms/IPO/PassManagerBuilder.h" #include "llvm/Transforms/Utils/EscapeEnumerator.h" +#include "llvm/Transforms/IPO/PassManagerBuilder.h" #include using namespace llvm; +#define CDS_DEBUG #define DEBUG_TYPE "CDS" #include -Value *getPosition( Instruction * I, IRBuilder <> IRB, bool print = false) +static inline Value *getPosition( Instruction * I, IRBuilder <> IRB, bool print = false) { const DebugLoc & debug_location = I->getDebugLoc (); std::string position_string; @@ -66,21 +68,38 @@ Value *getPosition( Instruction * I, IRBuilder <> IRB, bool print = false) return IRB.CreateGlobalStringPtr (position_string); } +static inline bool checkSignature(Function * func, Value * args[]) { + FunctionType * FType = func->getFunctionType(); + for (unsigned i = 0 ; i < FType->getNumParams(); i++) { + if (FType->getParamType(i) != args[i]->getType()) { +#ifdef CDS_DEBUG + errs() << "expects: " << *FType->getParamType(i) + << "\tbut receives: " << *args[i]->getType() << "\n"; +#endif + return false; + } + } + + return true; +} + STATISTIC(NumInstrumentedReads, "Number of instrumented reads"); STATISTIC(NumInstrumentedWrites, "Number of instrumented writes"); +STATISTIC(NumOmittedReadsBeforeWrite, + "Number of reads ignored due to following writes"); STATISTIC(NumAccessesWithBadSize, "Number of accesses with bad size"); // STATISTIC(NumInstrumentedVtableWrites, "Number of vtable ptr writes"); // STATISTIC(NumInstrumentedVtableReads, "Number of vtable ptr reads"); - -STATISTIC(NumOmittedReadsBeforeWrite, - "Number of reads ignored due to following writes"); STATISTIC(NumOmittedReadsFromConstantGlobals, "Number of reads from constant globals"); STATISTIC(NumOmittedReadsFromVtable, "Number of vtable reads"); STATISTIC(NumOmittedNonCaptured, "Number of accesses ignored due to capturing"); -Type * OrdTy; +// static const char *const kCDSModuleCtorName = "cds.module_ctor"; +// static const char *const kCDSInitName = "cds_init"; +Type * OrdTy; +Type * IntPtrTy; Type * Int8PtrTy; Type * Int16PtrTy; Type * Int32PtrTy; @@ -89,14 +108,13 @@ Type * Int64PtrTy; Type * VoidTy; static const size_t kNumberOfAccessSizes = 4; -static const int volatile_order = 6; -int getAtomicOrderIndex(AtomicOrdering order){ +int getAtomicOrderIndex(AtomicOrdering order) { switch (order) { case AtomicOrdering::Monotonic: return (int)AtomicOrderingCABI::relaxed; - // case AtomicOrdering::Consume: // not specified yet - // return AtomicOrderingCABI::consume; + //case AtomicOrdering::Consume: // not specified yet + // return AtomicOrderingCABI::consume; case AtomicOrdering::Acquire: return (int)AtomicOrderingCABI::acquire; case AtomicOrdering::Release: @@ -111,56 +129,112 @@ int getAtomicOrderIndex(AtomicOrdering order){ } } +AtomicOrderingCABI indexToAtomicOrder(int index) { + switch (index) { + case 0: + return AtomicOrderingCABI::relaxed; + case 1: + return AtomicOrderingCABI::consume; + case 2: + return AtomicOrderingCABI::acquire; + case 3: + return AtomicOrderingCABI::release; + case 4: + return AtomicOrderingCABI::acq_rel; + case 5: + return AtomicOrderingCABI::seq_cst; + default: + errs() << "Bad Atomic index\n"; + return AtomicOrderingCABI::seq_cst; + } +} + +/* According to atomic_base.h: __cmpexch_failure_order */ +int AtomicCasFailureOrderIndex(int index) { + AtomicOrderingCABI succ_order = indexToAtomicOrder(index); + AtomicOrderingCABI fail_order; + if (succ_order == AtomicOrderingCABI::acq_rel) + fail_order = AtomicOrderingCABI::acquire; + else if (succ_order == AtomicOrderingCABI::release) + fail_order = AtomicOrderingCABI::relaxed; + else + fail_order = succ_order; + + return (int) fail_order; +} + +/* The original function checkSanitizerInterfaceFunction was defined + * in llvm/Transforms/Utils/ModuleUtils.h + */ +static Function * checkCDSPassInterfaceFunction(Constant *FuncOrBitcast) { + if (isa(FuncOrBitcast)) + return cast(FuncOrBitcast); + FuncOrBitcast->print(errs()); + errs() << "\n"; + std::string Err; + raw_string_ostream Stream(Err); + Stream << "CDSPass interface function redefined: " << *FuncOrBitcast; + report_fatal_error(Err); +} + namespace { struct CDSPass : public FunctionPass { - static char ID; CDSPass() : FunctionPass(ID) {} - bool runOnFunction(Function &F) override; + StringRef getPassName() const override; + bool runOnFunction(Function &F) override; + bool doInitialization(Module &M) override; + static char ID; private: void initializeCallbacks(Module &M); bool instrumentLoadOrStore(Instruction *I, const DataLayout &DL); bool instrumentVolatile(Instruction *I, const DataLayout &DL); + bool instrumentMemIntrinsic(Instruction *I); bool isAtomicCall(Instruction *I); bool instrumentAtomic(Instruction *I, const DataLayout &DL); bool instrumentAtomicCall(CallInst *CI, const DataLayout &DL); + bool shouldInstrumentBeforeAtomics(Instruction *I); void chooseInstructionsToInstrument(SmallVectorImpl &Local, SmallVectorImpl &All, const DataLayout &DL); bool addrPointsToConstantData(Value *Addr); int getMemoryAccessFuncIndex(Value *Addr, const DataLayout &DL); - - // Callbacks to run-time library are computed in doInitialization. - Constant * CDSFuncEntry; - Constant * CDSFuncExit; - - Constant * CDSLoad[kNumberOfAccessSizes]; - Constant * CDSStore[kNumberOfAccessSizes]; - Constant * CDSVolatileLoad[kNumberOfAccessSizes]; - Constant * CDSVolatileStore[kNumberOfAccessSizes]; - Constant * CDSAtomicInit[kNumberOfAccessSizes]; - Constant * CDSAtomicLoad[kNumberOfAccessSizes]; - Constant * CDSAtomicStore[kNumberOfAccessSizes]; - Constant * CDSAtomicRMW[AtomicRMWInst::LAST_BINOP + 1][kNumberOfAccessSizes]; - Constant * CDSAtomicCAS_V1[kNumberOfAccessSizes]; - Constant * CDSAtomicCAS_V2[kNumberOfAccessSizes]; - Constant * CDSAtomicThreadFence; + bool instrumentLoops(Function &F); + + Function * CDSFuncEntry; + Function * CDSFuncExit; + + Function * CDSLoad[kNumberOfAccessSizes]; + Function * CDSStore[kNumberOfAccessSizes]; + Function * CDSVolatileLoad[kNumberOfAccessSizes]; + Function * CDSVolatileStore[kNumberOfAccessSizes]; + Function * CDSAtomicInit[kNumberOfAccessSizes]; + Function * CDSAtomicLoad[kNumberOfAccessSizes]; + Function * CDSAtomicStore[kNumberOfAccessSizes]; + Function * CDSAtomicRMW[AtomicRMWInst::LAST_BINOP + 1][kNumberOfAccessSizes]; + Function * CDSAtomicCAS_V1[kNumberOfAccessSizes]; + Function * CDSAtomicCAS_V2[kNumberOfAccessSizes]; + Function * CDSAtomicThreadFence; + Function * MemmoveFn, * MemcpyFn, * MemsetFn; + // Function * CDSCtorFunction; std::vector AtomicFuncNames; std::vector PartialAtomicFuncNames; }; } -static bool isVtableAccess(Instruction *I) { - if (MDNode *Tag = I->getMetadata(LLVMContext::MD_tbaa)) - return Tag->isTBAAVtableAccess(); - return false; +StringRef CDSPass::getPassName() const { + return "CDSPass"; } void CDSPass::initializeCallbacks(Module &M) { LLVMContext &Ctx = M.getContext(); + AttributeList Attr; + Attr = Attr.addAttribute(Ctx, AttributeList::FunctionIndex, + Attribute::NoUnwind); Type * Int1Ty = Type::getInt1Ty(Ctx); + Type * Int32Ty = Type::getInt32Ty(Ctx); OrdTy = Type::getInt32Ty(Ctx); Int8PtrTy = Type::getInt8PtrTy(Ctx); @@ -170,10 +244,12 @@ void CDSPass::initializeCallbacks(Module &M) { VoidTy = Type::getVoidTy(Ctx); - CDSFuncEntry = M.getOrInsertFunction("cds_func_entry", - VoidTy, Int8PtrTy); - CDSFuncExit = M.getOrInsertFunction("cds_func_exit", - VoidTy, Int8PtrTy); + CDSFuncEntry = checkCDSPassInterfaceFunction( + M.getOrInsertFunction("cds_func_entry", + Attr, VoidTy, Int8PtrTy)); + CDSFuncExit = checkCDSPassInterfaceFunction( + M.getOrInsertFunction("cds_func_exit", + Attr, VoidTy, Int8PtrTy)); // Get the function to call from our untime library. for (unsigned i = 0; i < kNumberOfAccessSizes; i++) { @@ -196,18 +272,25 @@ void CDSPass::initializeCallbacks(Module &M) { SmallString<32> AtomicLoadName("cds_atomic_load" + BitSizeStr); SmallString<32> AtomicStoreName("cds_atomic_store" + BitSizeStr); - CDSLoad[i] = M.getOrInsertFunction(LoadName, VoidTy, PtrTy); - CDSStore[i] = M.getOrInsertFunction(StoreName, VoidTy, PtrTy); - CDSVolatileLoad[i] = M.getOrInsertFunction(VolatileLoadName, - Ty, PtrTy, OrdTy, Int8PtrTy); - CDSVolatileStore[i] = M.getOrInsertFunction(VolatileStoreName, - VoidTy, PtrTy, Ty, OrdTy, Int8PtrTy); - CDSAtomicInit[i] = M.getOrInsertFunction(AtomicInitName, - VoidTy, PtrTy, Ty, Int8PtrTy); - CDSAtomicLoad[i] = M.getOrInsertFunction(AtomicLoadName, - Ty, PtrTy, OrdTy, Int8PtrTy); - CDSAtomicStore[i] = M.getOrInsertFunction(AtomicStoreName, - VoidTy, PtrTy, Ty, OrdTy, Int8PtrTy); + CDSLoad[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(LoadName, Attr, VoidTy, Int8PtrTy)); + CDSStore[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(StoreName, Attr, VoidTy, Int8PtrTy)); + CDSVolatileLoad[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(VolatileLoadName, + Attr, Ty, PtrTy, Int8PtrTy)); + CDSVolatileStore[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(VolatileStoreName, + Attr, VoidTy, PtrTy, Ty, Int8PtrTy)); + CDSAtomicInit[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(AtomicInitName, + Attr, VoidTy, PtrTy, Ty, Int8PtrTy)); + CDSAtomicLoad[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(AtomicLoadName, + Attr, Ty, PtrTy, OrdTy, Int8PtrTy)); + CDSAtomicStore[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(AtomicStoreName, + Attr, VoidTy, PtrTy, Ty, OrdTy, Int8PtrTy)); for (int op = AtomicRMWInst::FIRST_BINOP; op <= AtomicRMWInst::LAST_BINOP; ++op) { @@ -230,23 +313,71 @@ void CDSPass::initializeCallbacks(Module &M) { continue; SmallString<32> AtomicRMWName("cds_atomic" + NamePart + BitSizeStr); - CDSAtomicRMW[op][i] = M.getOrInsertFunction(AtomicRMWName, - Ty, PtrTy, Ty, OrdTy, Int8PtrTy); + CDSAtomicRMW[op][i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(AtomicRMWName, + Attr, Ty, PtrTy, Ty, OrdTy, Int8PtrTy)); } // only supportes strong version SmallString<32> AtomicCASName_V1("cds_atomic_compare_exchange" + BitSizeStr + "_v1"); SmallString<32> AtomicCASName_V2("cds_atomic_compare_exchange" + BitSizeStr + "_v2"); - CDSAtomicCAS_V1[i] = M.getOrInsertFunction(AtomicCASName_V1, - Ty, PtrTy, Ty, Ty, OrdTy, OrdTy, Int8PtrTy); - CDSAtomicCAS_V2[i] = M.getOrInsertFunction(AtomicCASName_V2, - Int1Ty, PtrTy, PtrTy, Ty, OrdTy, OrdTy, Int8PtrTy); + CDSAtomicCAS_V1[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(AtomicCASName_V1, + Attr, Ty, PtrTy, Ty, Ty, OrdTy, OrdTy, Int8PtrTy)); + CDSAtomicCAS_V2[i] = checkCDSPassInterfaceFunction( + M.getOrInsertFunction(AtomicCASName_V2, + Attr, Int1Ty, PtrTy, PtrTy, Ty, OrdTy, OrdTy, Int8PtrTy)); } - CDSAtomicThreadFence = M.getOrInsertFunction("cds_atomic_thread_fence", - VoidTy, OrdTy, Int8PtrTy); + CDSAtomicThreadFence = checkCDSPassInterfaceFunction( + M.getOrInsertFunction("cds_atomic_thread_fence", Attr, VoidTy, OrdTy, Int8PtrTy)); + + MemmoveFn = checkCDSPassInterfaceFunction( + M.getOrInsertFunction("memmove", Attr, Int8PtrTy, Int8PtrTy, + Int8PtrTy, IntPtrTy)); + MemcpyFn = checkCDSPassInterfaceFunction( + M.getOrInsertFunction("memcpy", Attr, Int8PtrTy, Int8PtrTy, + Int8PtrTy, IntPtrTy)); + MemsetFn = checkCDSPassInterfaceFunction( + M.getOrInsertFunction("memset", Attr, Int8PtrTy, Int8PtrTy, + Int32Ty, IntPtrTy)); +} + +bool CDSPass::doInitialization(Module &M) { + const DataLayout &DL = M.getDataLayout(); + IntPtrTy = DL.getIntPtrType(M.getContext()); + + // createSanitizerCtorAndInitFunctions is defined in "llvm/Transforms/Utils/ModuleUtils.h" + // We do not support it yet + /* + std::tie(CDSCtorFunction, std::ignore) = createSanitizerCtorAndInitFunctions( + M, kCDSModuleCtorName, kCDSInitName, {}, {}); + + appendToGlobalCtors(M, CDSCtorFunction, 0); + */ + + AtomicFuncNames = + { + "atomic_init", "atomic_load", "atomic_store", + "atomic_fetch_", "atomic_exchange", "atomic_compare_exchange_" + }; + + PartialAtomicFuncNames = + { + "load", "store", "fetch", "exchange", "compare_exchange_" + }; + + return true; } +static bool isVtableAccess(Instruction *I) { + if (MDNode *Tag = I->getMetadata(LLVMContext::MD_tbaa)) + return Tag->isTBAAVtableAccess(); + return false; +} + +// Do not instrument known races/"benign races" that come from compiler +// instrumentatin. The user has no way of suppressing them. static bool shouldInstrumentReadWriteFromAddress(const Module *M, Value *Addr) { // Peel off GEPs and BitCasts. Addr = Addr->stripInBoundsOffsets(); @@ -299,92 +430,27 @@ bool CDSPass::addrPointsToConstantData(Value *Addr) { return false; } -bool CDSPass::runOnFunction(Function &F) { - if (F.getName() == "main") { - F.setName("user_main"); - errs() << "main replaced by user_main\n"; - } - - if (true) { - initializeCallbacks( *F.getParent() ); - - AtomicFuncNames = - { - "atomic_init", "atomic_load", "atomic_store", - "atomic_fetch_", "atomic_exchange", "atomic_compare_exchange_" - }; - - PartialAtomicFuncNames = - { - "load", "store", "fetch", "exchange", "compare_exchange_" - }; - - SmallVector AllLoadsAndStores; - SmallVector LocalLoadsAndStores; - SmallVector VolatileLoadsAndStores; - SmallVector AtomicAccesses; - - std::vector worklist; - - bool Res = false; - bool HasAtomic = false; - const DataLayout &DL = F.getParent()->getDataLayout(); - - // errs() << "--- " << F.getName() << "---\n"; - - for (auto &B : F) { - for (auto &I : B) { - if ( (&I)->isAtomic() || isAtomicCall(&I) ) { - AtomicAccesses.push_back(&I); - HasAtomic = true; - } else if (isa(I) || isa(I)) { - LoadInst *LI = dyn_cast(&I); - StoreInst *SI = dyn_cast(&I); - bool isVolatile = ( LI ? LI->isVolatile() : SI->isVolatile() ); - - if (isVolatile) - VolatileLoadsAndStores.push_back(&I); - else - LocalLoadsAndStores.push_back(&I); - } else if (isa(I) || isa(I)) { - // not implemented yet - } - } - - chooseInstructionsToInstrument(LocalLoadsAndStores, AllLoadsAndStores, DL); - } - - for (auto Inst : AllLoadsAndStores) { - Res |= instrumentLoadOrStore(Inst, DL); - } - - for (auto Inst : VolatileLoadsAndStores) { - Res |= instrumentVolatile(Inst, DL); - } - - for (auto Inst : AtomicAccesses) { - Res |= instrumentAtomic(Inst, DL); - } - - // only instrument functions that contain atomics - if (Res && HasAtomic) { - IRBuilder<> IRB(F.getEntryBlock().getFirstNonPHI()); - /* Unused for now - Value *ReturnAddress = IRB.CreateCall( - Intrinsic::getDeclaration(F.getParent(), Intrinsic::returnaddress), - IRB.getInt32(0)); - */ - - Value * FuncName = IRB.CreateGlobalStringPtr(F.getName()); - IRB.CreateCall(CDSFuncEntry, FuncName); - - EscapeEnumerator EE(F, "cds_cleanup", true); - while (IRBuilder<> *AtExit = EE.Next()) { - AtExit->CreateCall(CDSFuncExit, FuncName); - } - - Res = true; - } +bool CDSPass::shouldInstrumentBeforeAtomics(Instruction * Inst) { + if (LoadInst *LI = dyn_cast(Inst)) { + AtomicOrdering ordering = LI->getOrdering(); + if ( isAtLeastOrStrongerThan(ordering, AtomicOrdering::Acquire) ) + return true; + } else if (StoreInst *SI = dyn_cast(Inst)) { + AtomicOrdering ordering = SI->getOrdering(); + if ( isAtLeastOrStrongerThan(ordering, AtomicOrdering::Acquire) ) + return true; + } else if (AtomicRMWInst *RMWI = dyn_cast(Inst)) { + AtomicOrdering ordering = RMWI->getOrdering(); + if ( isAtLeastOrStrongerThan(ordering, AtomicOrdering::Acquire) ) + return true; + } else if (AtomicCmpXchgInst *CASI = dyn_cast(Inst)) { + AtomicOrdering ordering = CASI->getSuccessOrdering(); + if ( isAtLeastOrStrongerThan(ordering, AtomicOrdering::Acquire) ) + return true; + } else if (FenceInst *FI = dyn_cast(Inst)) { + AtomicOrdering ordering = FI->getOrdering(); + if ( isAtLeastOrStrongerThan(ordering, AtomicOrdering::Acquire) ) + return true; } return false; @@ -432,6 +498,110 @@ void CDSPass::chooseInstructionsToInstrument( Local.clear(); } +/* Not implemented +void CDSPass::InsertRuntimeIgnores(Function &F) { + IRBuilder<> IRB(F.getEntryBlock().getFirstNonPHI()); + IRB.CreateCall(CDSIgnoreBegin); + EscapeEnumerator EE(F, "cds_ignore_cleanup", ClHandleCxxExceptions); + while (IRBuilder<> *AtExit = EE.Next()) { + AtExit->CreateCall(CDSIgnoreEnd); + } +}*/ + +bool CDSPass::runOnFunction(Function &F) { + initializeCallbacks( *F.getParent() ); + SmallVector AllLoadsAndStores; + SmallVector LocalLoadsAndStores; + SmallVector VolatileLoadsAndStores; + SmallVector AtomicAccesses; + SmallVector MemIntrinCalls; + + bool Res = false; + bool HasAtomic = false; + bool HasVolatile = false; + const DataLayout &DL = F.getParent()->getDataLayout(); + + // instrumentLoops(F); + + for (auto &BB : F) { + for (auto &Inst : BB) { + if ( (&Inst)->isAtomic() ) { + AtomicAccesses.push_back(&Inst); + HasAtomic = true; + + if (shouldInstrumentBeforeAtomics(&Inst)) { + chooseInstructionsToInstrument(LocalLoadsAndStores, AllLoadsAndStores, + DL); + } + } else if (isAtomicCall(&Inst) ) { + AtomicAccesses.push_back(&Inst); + HasAtomic = true; + chooseInstructionsToInstrument(LocalLoadsAndStores, AllLoadsAndStores, + DL); + } else if (isa(Inst) || isa(Inst)) { + LoadInst *LI = dyn_cast(&Inst); + StoreInst *SI = dyn_cast(&Inst); + bool isVolatile = ( LI ? LI->isVolatile() : SI->isVolatile() ); + + if (isVolatile) { + VolatileLoadsAndStores.push_back(&Inst); + HasVolatile = true; + } else + LocalLoadsAndStores.push_back(&Inst); + } else if (isa(Inst) || isa(Inst)) { + if (isa(Inst)) + MemIntrinCalls.push_back(&Inst); + + /*if (CallInst *CI = dyn_cast(&Inst)) + maybeMarkSanitizerLibraryCallNoBuiltin(CI, TLI); + */ + + chooseInstructionsToInstrument(LocalLoadsAndStores, AllLoadsAndStores, + DL); + } + } + + chooseInstructionsToInstrument(LocalLoadsAndStores, AllLoadsAndStores, DL); + } + + for (auto Inst : AllLoadsAndStores) { + Res |= instrumentLoadOrStore(Inst, DL); + } + + for (auto Inst : VolatileLoadsAndStores) { + Res |= instrumentVolatile(Inst, DL); + } + + for (auto Inst : AtomicAccesses) { + Res |= instrumentAtomic(Inst, DL); + } + + for (auto Inst : MemIntrinCalls) { + Res |= instrumentMemIntrinsic(Inst); + } + + // Instrument function entry and exit for functions containing atomics or volatiles + if (Res && ( HasAtomic || HasVolatile) ) { + IRBuilder<> IRB(F.getEntryBlock().getFirstNonPHI()); + /* Unused for now + Value *ReturnAddress = IRB.CreateCall( + Intrinsic::getDeclaration(F.getParent(), Intrinsic::returnaddress), + IRB.getInt32(0)); + */ + + Value * FuncName = IRB.CreateGlobalStringPtr(F.getName()); + IRB.CreateCall(CDSFuncEntry, FuncName); + + EscapeEnumerator EE(F, "cds_cleanup", true); + while (IRBuilder<> *AtExit = EE.Next()) { + AtExit->CreateCall(CDSFuncExit, FuncName); + } + + Res = true; + } + + return false; +} bool CDSPass::instrumentLoadOrStore(Instruction *I, const DataLayout &DL) { @@ -445,51 +615,46 @@ bool CDSPass::instrumentLoadOrStore(Instruction *I, // As such they cannot have regular uses like an instrumentation function and // it makes no sense to track them as memory. if (Addr->isSwiftError()) - return false; + return false; int Idx = getMemoryAccessFuncIndex(Addr, DL); if (Idx < 0) return false; -// not supported by CDS yet -/* if (IsWrite && isVtableAccess(I)) { - LLVM_DEBUG(dbgs() << " VPTR : " << *I << "\n"); - Value *StoredValue = cast(I)->getValueOperand(); - // StoredValue may be a vector type if we are storing several vptrs at once. - // In this case, just take the first element of the vector since this is - // enough to find vptr races. - if (isa(StoredValue->getType())) - StoredValue = IRB.CreateExtractElement( - StoredValue, ConstantInt::get(IRB.getInt32Ty(), 0)); - if (StoredValue->getType()->isIntegerTy()) - StoredValue = IRB.CreateIntToPtr(StoredValue, IRB.getInt8PtrTy()); - // Call TsanVptrUpdate. - IRB.CreateCall(TsanVptrUpdate, - {IRB.CreatePointerCast(Addr, IRB.getInt8PtrTy()), - IRB.CreatePointerCast(StoredValue, IRB.getInt8PtrTy())}); - NumInstrumentedVtableWrites++; - return true; - } - - if (!IsWrite && isVtableAccess(I)) { - IRB.CreateCall(TsanVptrLoad, - IRB.CreatePointerCast(Addr, IRB.getInt8PtrTy())); - NumInstrumentedVtableReads++; - return true; - } -*/ + if (IsWrite && isVtableAccess(I)) { + /* TODO + LLVM_DEBUG(dbgs() << " VPTR : " << *I << "\n"); + Value *StoredValue = cast(I)->getValueOperand(); + // StoredValue may be a vector type if we are storing several vptrs at once. + // In this case, just take the first element of the vector since this is + // enough to find vptr races. + if (isa(StoredValue->getType())) + StoredValue = IRB.CreateExtractElement( + StoredValue, ConstantInt::get(IRB.getInt32Ty(), 0)); + if (StoredValue->getType()->isIntegerTy()) + StoredValue = IRB.CreateIntToPtr(StoredValue, IRB.getInt8PtrTy()); + // Call TsanVptrUpdate. + IRB.CreateCall(TsanVptrUpdate, + {IRB.CreatePointerCast(Addr, IRB.getInt8PtrTy()), + IRB.CreatePointerCast(StoredValue, IRB.getInt8PtrTy())}); + NumInstrumentedVtableWrites++; + */ + return true; + } + + if (!IsWrite && isVtableAccess(I)) { + /* TODO + IRB.CreateCall(TsanVptrLoad, + IRB.CreatePointerCast(Addr, IRB.getInt8PtrTy())); + NumInstrumentedVtableReads++; + */ + return true; + } + // TODO: unaligned reads and writes Value *OnAccessFunc = nullptr; OnAccessFunc = IsWrite ? CDSStore[Idx] : CDSLoad[Idx]; - - Type *ArgType = IRB.CreatePointerCast(Addr, Addr->getType())->getType(); - - if ( ArgType != Int8PtrTy && ArgType != Int16PtrTy && - ArgType != Int32PtrTy && ArgType != Int64PtrTy ) { - // if other types of load or stores are passed in - return false; - } - IRB.CreateCall(OnAccessFunc, IRB.CreatePointerCast(Addr, Addr->getType())); + IRB.CreateCall(OnAccessFunc, IRB.CreatePointerCast(Addr, IRB.getInt8PtrTy())); if (IsWrite) NumInstrumentedWrites++; else NumInstrumentedReads++; return true; @@ -500,28 +665,34 @@ bool CDSPass::instrumentVolatile(Instruction * I, const DataLayout &DL) { Value *position = getPosition(I, IRB); if (LoadInst *LI = dyn_cast(I)) { - assert( LI->isVolatile() ); Value *Addr = LI->getPointerOperand(); int Idx=getMemoryAccessFuncIndex(Addr, DL); if (Idx < 0) return false; + const unsigned ByteSize = 1U << Idx; + const unsigned BitSize = ByteSize * 8; + Type *Ty = Type::getIntNTy(IRB.getContext(), BitSize); + Type *PtrTy = Ty->getPointerTo(); + Value *Args[] = {IRB.CreatePointerCast(Addr, PtrTy), position}; - Value *order = ConstantInt::get(OrdTy, volatile_order); - Value *args[] = {Addr, order, position}; - Instruction* funcInst=CallInst::Create(CDSVolatileLoad[Idx], args); - ReplaceInstWithInst(LI, funcInst); + Type *OrigTy = cast(Addr->getType())->getElementType(); + Value *C = IRB.CreateCall(CDSVolatileLoad[Idx], Args); + Value *Cast = IRB.CreateBitOrPointerCast(C, OrigTy); + I->replaceAllUsesWith(Cast); } else if (StoreInst *SI = dyn_cast(I)) { - assert( SI->isVolatile() ); Value *Addr = SI->getPointerOperand(); int Idx=getMemoryAccessFuncIndex(Addr, DL); if (Idx < 0) return false; - - Value *val = SI->getValueOperand(); - Value *order = ConstantInt::get(OrdTy, volatile_order); - Value *args[] = {Addr, val, order, position}; - Instruction* funcInst=CallInst::Create(CDSVolatileStore[Idx], args); - ReplaceInstWithInst(SI, funcInst); + const unsigned ByteSize = 1U << Idx; + const unsigned BitSize = ByteSize * 8; + Type *Ty = Type::getIntNTy(IRB.getContext(), BitSize); + Type *PtrTy = Ty->getPointerTo(); + Value *Args[] = {IRB.CreatePointerCast(Addr, PtrTy), + IRB.CreateBitOrPointerCast(SI->getValueOperand(), Ty), + position}; + CallInst *C = CallInst::Create(CDSVolatileStore[Idx], Args); + ReplaceInstWithInst(I, C); } else { return false; } @@ -529,6 +700,26 @@ bool CDSPass::instrumentVolatile(Instruction * I, const DataLayout &DL) { return true; } +bool CDSPass::instrumentMemIntrinsic(Instruction *I) { + IRBuilder<> IRB(I); + if (MemSetInst *M = dyn_cast(I)) { + IRB.CreateCall( + MemsetFn, + {IRB.CreatePointerCast(M->getArgOperand(0), IRB.getInt8PtrTy()), + IRB.CreateIntCast(M->getArgOperand(1), IRB.getInt32Ty(), false), + IRB.CreateIntCast(M->getArgOperand(2), IntPtrTy, false)}); + I->eraseFromParent(); + } else if (MemTransferInst *M = dyn_cast(I)) { + IRB.CreateCall( + isa(M) ? MemcpyFn : MemmoveFn, + {IRB.CreatePointerCast(M->getArgOperand(0), IRB.getInt8PtrTy()), + IRB.CreatePointerCast(M->getArgOperand(1), IRB.getInt8PtrTy()), + IRB.CreateIntCast(M->getArgOperand(2), IntPtrTy, false)}); + I->eraseFromParent(); + } + return false; +} + bool CDSPass::instrumentAtomic(Instruction * I, const DataLayout &DL) { IRBuilder<> IRB(I); @@ -537,7 +728,6 @@ bool CDSPass::instrumentAtomic(Instruction * I, const DataLayout &DL) { } Value *position = getPosition(I, IRB); - if (LoadInst *LI = dyn_cast(I)) { Value *Addr = LI->getPointerOperand(); int Idx=getMemoryAccessFuncIndex(Addr, DL); @@ -546,8 +736,8 @@ bool CDSPass::instrumentAtomic(Instruction * I, const DataLayout &DL) { int atomic_order_index = getAtomicOrderIndex(LI->getOrdering()); Value *order = ConstantInt::get(OrdTy, atomic_order_index); - Value *args[] = {Addr, order, position}; - Instruction* funcInst=CallInst::Create(CDSAtomicLoad[Idx], args); + Value *Args[] = {Addr, order, position}; + Instruction* funcInst = CallInst::Create(CDSAtomicLoad[Idx], Args); ReplaceInstWithInst(LI, funcInst); } else if (StoreInst *SI = dyn_cast(I)) { Value *Addr = SI->getPointerOperand(); @@ -558,8 +748,8 @@ bool CDSPass::instrumentAtomic(Instruction * I, const DataLayout &DL) { int atomic_order_index = getAtomicOrderIndex(SI->getOrdering()); Value *val = SI->getValueOperand(); Value *order = ConstantInt::get(OrdTy, atomic_order_index); - Value *args[] = {Addr, val, order, position}; - Instruction* funcInst=CallInst::Create(CDSAtomicStore[Idx], args); + Value *Args[] = {Addr, val, order, position}; + Instruction* funcInst = CallInst::Create(CDSAtomicStore[Idx], Args); ReplaceInstWithInst(SI, funcInst); } else if (AtomicRMWInst *RMWI = dyn_cast(I)) { Value *Addr = RMWI->getPointerOperand(); @@ -570,8 +760,8 @@ bool CDSPass::instrumentAtomic(Instruction * I, const DataLayout &DL) { int atomic_order_index = getAtomicOrderIndex(RMWI->getOrdering()); Value *val = RMWI->getValOperand(); Value *order = ConstantInt::get(OrdTy, atomic_order_index); - Value *args[] = {Addr, val, order, position}; - Instruction* funcInst = CallInst::Create(CDSAtomicRMW[RMWI->getOperation()][Idx], args); + Value *Args[] = {Addr, val, order, position}; + Instruction* funcInst = CallInst::Create(CDSAtomicRMW[RMWI->getOperation()][Idx], Args); ReplaceInstWithInst(RMWI, funcInst); } else if (AtomicCmpXchgInst *CASI = dyn_cast(I)) { IRBuilder<> IRB(CASI); @@ -634,7 +824,7 @@ bool CDSPass::isAtomicCall(Instruction *I) { StringRef funName = fun->getName(); - // todo: come up with better rules for function name checking + // TODO: come up with better rules for function name checking for (StringRef name : AtomicFuncNames) { if ( funName.contains(name) ) return true; @@ -680,13 +870,22 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { // atomic_init; args = {obj, order} if (funName.contains("atomic_init")) { + Value *OrigVal = parameters[1]; + Value *ptr = IRB.CreatePointerCast(OrigPtr, PtrTy); - Value *val = IRB.CreateBitOrPointerCast(parameters[1], Ty); + Value *val; + if (OrigVal->getType()->isPtrOrPtrVectorTy()) + val = IRB.CreatePointerCast(OrigVal, Ty); + else + val = IRB.CreateIntCast(OrigVal, Ty, true); + Value *args[] = {ptr, val, position}; + if (!checkSignature(CDSAtomicInit[Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicInit[Idx], args); ReplaceInstWithInst(CI, funcInst); - return true; } @@ -702,7 +901,10 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { order = ConstantInt::get(OrdTy, (int) AtomicOrderingCABI::seq_cst); Value *args[] = {ptr, order, position}; - + + if (!checkSignature(CDSAtomicLoad[Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicLoad[Idx], args); ReplaceInstWithInst(CI, funcInst); @@ -714,10 +916,14 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { Value *order = IRB.CreateBitOrPointerCast(parameters[1], OrdTy); Value *args[] = {ptr, order, position}; + // Without this check, gdax does not compile :( if (!CI->getType()->isPointerTy()) { return false; } + if (!checkSignature(CDSAtomicLoad[Idx], args)) + return false; + CallInst *funcInst = IRB.CreateCall(CDSAtomicLoad[Idx], args); Value *RetVal = IRB.CreateIntToPtr(funcInst, CI->getType()); @@ -741,21 +947,35 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { order = ConstantInt::get(OrdTy, (int) AtomicOrderingCABI::seq_cst); Value *args[] = {ptr, val, order, position}; - + + if (!checkSignature(CDSAtomicStore[Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicStore[Idx], args); ReplaceInstWithInst(CI, funcInst); return true; } else if (funName.contains("atomic") && - funName.contains("EEEE5store") ) { - // does this version of call always have an atomic order as an argument? - Value *OrigVal = parameters[1]; + funName.contains("store") ) { + // Does this version of call always have an atomic order as an argument? + if (parameters.size() < 3) + return false; + Value *OrigVal = parameters[1]; Value *ptr = IRB.CreatePointerCast(OrigPtr, PtrTy); - Value *val = IRB.CreatePointerCast(OrigVal, Ty); + + Value *val; + if (OrigVal->getType()->isPtrOrPtrVectorTy()) + val = IRB.CreatePointerCast(OrigVal, Ty); + else + val = IRB.CreateIntCast(OrigVal, Ty, true); + Value *order = IRB.CreateBitOrPointerCast(parameters[2], OrdTy); Value *args[] = {ptr, val, order, position}; + if (!checkSignature(CDSAtomicStore[Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicStore[Idx], args); ReplaceInstWithInst(CI, funcInst); @@ -764,7 +984,8 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { // atomic_fetch_*; args = {obj, val, order} if (funName.contains("atomic_fetch_") || - funName.contains("atomic_exchange") ) { + funName.contains("atomic_exchange")) { + bool isExplicit = funName.contains("_explicit"); Value *OrigVal = parameters[1]; @@ -787,7 +1008,12 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { } Value *ptr = IRB.CreatePointerCast(OrigPtr, PtrTy); - Value *val = IRB.CreatePointerCast(OrigVal, Ty); + Value *val; + if (OrigVal->getType()->isPtrOrPtrVectorTy()) + val = IRB.CreatePointerCast(OrigVal, Ty); + else + val = IRB.CreateIntCast(OrigVal, Ty, true); + Value *order; if (isExplicit) order = IRB.CreateBitOrPointerCast(parameters[2], OrdTy); @@ -795,20 +1021,57 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { order = ConstantInt::get(OrdTy, (int) AtomicOrderingCABI::seq_cst); Value *args[] = {ptr, val, order, position}; - + + if (!checkSignature(CDSAtomicRMW[op][Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicRMW[op][Idx], args); ReplaceInstWithInst(CI, funcInst); return true; } else if (funName.contains("fetch")) { - errs() << "atomic exchange captured. Not implemented yet. "; + errs() << "atomic fetch captured. Not implemented yet. "; errs() << "See source file :"; getPosition(CI, IRB, true); + return false; } else if (funName.contains("exchange") && !funName.contains("compare_exchange") ) { - errs() << "atomic exchange captured. Not implemented yet. "; - errs() << "See source file :"; - getPosition(CI, IRB, true); + if (CI->getType()->isPointerTy()) { + /** + * TODO: instrument the following case + * mcs-lock.h + * std::atomic m_tail; + * struct T * me; + * struct T * pred = m_tail.exchange(me, memory_order_*); + */ + errs() << "atomic exchange captured. Not implemented yet. "; + errs() << "See source file :"; + getPosition(CI, IRB, true); + + return false; + } + + Value *OrigVal = parameters[1]; + + Value *ptr = IRB.CreatePointerCast(OrigPtr, PtrTy); + Value *val; + if (OrigVal->getType()->isPtrOrPtrVectorTy()) + val = IRB.CreatePointerCast(OrigVal, Ty); + else + val = IRB.CreateIntCast(OrigVal, Ty, true); + + Value *order = IRB.CreateBitOrPointerCast(parameters[2], OrdTy); + Value *args[] = {ptr, val, order, position}; + + int op = AtomicRMWInst::Xchg; + + if (!checkSignature(CDSAtomicRMW[op][Idx], args)) + return false; + + Instruction* funcInst = CallInst::Create(CDSAtomicRMW[op][Idx], args); + ReplaceInstWithInst(CI, funcInst); + + return true; } /* atomic_compare_exchange_*; @@ -824,7 +1087,18 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { Value *order_succ, *order_fail; if (isExplicit) { order_succ = IRB.CreateBitOrPointerCast(parameters[3], OrdTy); - order_fail = IRB.CreateBitOrPointerCast(parameters[4], OrdTy); + + if (parameters.size() > 4) { + order_fail = IRB.CreateBitOrPointerCast(parameters[4], OrdTy); + } else { + /* The failure order is not provided */ + order_fail = order_succ; + ConstantInt * order_succ_cast = dyn_cast(order_succ); + int index = order_succ_cast->getSExtValue(); + + order_fail = ConstantInt::get(OrdTy, + AtomicCasFailureOrderIndex(index)); + } } else { order_succ = ConstantInt::get(OrdTy, (int) AtomicOrderingCABI::seq_cst); @@ -834,7 +1108,10 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { Value *args[] = {Addr, CmpOperand, NewOperand, order_succ, order_fail, position}; - + + if (!checkSignature(CDSAtomicCAS_V2[Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicCAS_V2[Idx], args); ReplaceInstWithInst(CI, funcInst); @@ -847,10 +1124,25 @@ bool CDSPass::instrumentAtomicCall(CallInst *CI, const DataLayout &DL) { Value *order_succ, *order_fail; order_succ = IRB.CreateBitOrPointerCast(parameters[3], OrdTy); - order_fail = IRB.CreateBitOrPointerCast(parameters[4], OrdTy); + + if (parameters.size() > 4) { + order_fail = IRB.CreateBitOrPointerCast(parameters[4], OrdTy); + } else { + /* The failure order is not provided */ + order_fail = order_succ; + ConstantInt * order_succ_cast = dyn_cast(order_succ); + int index = order_succ_cast->getSExtValue(); + + order_fail = ConstantInt::get(OrdTy, + AtomicCasFailureOrderIndex(index)); + } Value *args[] = {Addr, CmpOperand, NewOperand, order_succ, order_fail, position}; + + if (!checkSignature(CDSAtomicCAS_V2[Idx], args)) + return false; + Instruction* funcInst = CallInst::Create(CDSAtomicCAS_V2[Idx], args); ReplaceInstWithInst(CI, funcInst); @@ -880,6 +1172,58 @@ int CDSPass::getMemoryAccessFuncIndex(Value *Addr, return Idx; } +bool CDSPass::instrumentLoops(Function &F) +{ + DominatorTree DT(F); + LoopInfo LI(DT); + + SmallVector Loops = LI.getLoopsInPreorder(); + bool instrumented = false; + + // Do a post-order traversal of the loops so that counter updates can be + // iteratively hoisted outside the loop nest. + for (auto *Loop : llvm::reverse(Loops)) { + bool instrument_loop = false; + + // Iterator over loop blocks and search for atomics and volatiles + Loop::block_iterator it; + for (it = Loop->block_begin(); it != Loop->block_end(); it++) { + BasicBlock * block = *it; + for (auto &Inst : *block) { + if ( (&Inst)->isAtomic() ) { + instrument_loop = true; + break; + } else if (isAtomicCall(&Inst)) { + instrument_loop = true; + break; + } else if (isa(Inst) || isa(Inst)) { + LoadInst *LI = dyn_cast(&Inst); + StoreInst *SI = dyn_cast(&Inst); + bool isVolatile = ( LI ? LI->isVolatile() : SI->isVolatile() ); + + if (isVolatile) { + instrument_loop = true; + break; + } + } + } + + if (instrument_loop) + break; + } + + if (instrument_loop) { + // TODO: what to instrument? + errs() << "Function: " << F.getName() << "\n"; + BasicBlock * header = Loop->getHeader(); + header->dump(); + + instrumented = true; + } + } + + return instrumented; +} char CDSPass::ID = 0;