/external/llvm/lib/Target/NVPTX/ |
D | NVPTXLowerAggrCopies.cpp | 243 SmallVector<MemIntrinsic *, 4> MemCalls; in runOnFunction() 264 } else if (MemIntrinsic *IntrCall = dyn_cast<MemIntrinsic>(II)) { in runOnFunction() 305 for (MemIntrinsic *MemCall : MemCalls) { in runOnFunction()
|
/external/llvm/include/llvm/Analysis/ |
D | MemoryLocation.h | 28 class MemIntrinsic; variable 89 static MemoryLocation getForDest(const MemIntrinsic *MI);
|
D | PtrUseVisitor.h | 263 void visitMemIntrinsic(MemIntrinsic &I) {} in visitMemIntrinsic()
|
D | AliasAnalysis.h | 57 class MemIntrinsic; variable
|
/external/llvm/lib/Transforms/Scalar/ |
D | DeadStoreElimination.cpp | 207 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(Inst)) { in getLocForWrite() 267 return !cast<MemIntrinsic>(II)->isVolatile(); in isRemovable() 304 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) in getStoredPointerOperand() 613 MemIntrinsic* DepIntrinsic = cast<MemIntrinsic>(DepWrite); in runOnBasicBlock()
|
D | AlignmentFromAssumptions.cpp | 344 } else if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(J)) { in processAssumption() 391 assert((!isa<MemIntrinsic>(MI) || isa<MemSetInst>(MI)) && in processAssumption()
|
D | GVN.cpp | 542 static AvailableValueInBlock getMI(BasicBlock *BB, MemIntrinsic *MI, in getMI() 586 MemIntrinsic *getMemIntrinValue() const { in getMemIntrinValue() 588 return cast<MemIntrinsic>(Val.getPointer()); in getMemIntrinValue() 1080 MemIntrinsic *MI, in AnalyzeLoadFromClobberingMemInst() 1230 static Value *GetMemInstValueForLoad(MemIntrinsic *SrcInst, unsigned Offset, in GetMemInstValueForLoad() 1434 if (MemIntrinsic *DepMI = dyn_cast<MemIntrinsic>(DepInfo.getInst())) { in AnalyzeLoadAvailability() 1938 if (MemIntrinsic *DepMI = dyn_cast<MemIntrinsic>(Dep.getInst())) { in processLoad()
|
D | ScalarReplAggregates.cpp | 180 void RewriteMemIntrinUserOfAlloca(MemIntrinsic *MI, Instruction *Inst, 1623 } else if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(User)) { in isSafeForScalarRepl() 1899 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(User)) { in RewriteForScalarRepl() 2177 SROA::RewriteMemIntrinUserOfAlloca(MemIntrinsic *MI, Instruction *Inst, in RewriteMemIntrinUserOfAlloca()
|
D | LoopRerollPass.cpp | 624 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) in isSimpleLoadStore()
|
D | SROA.cpp | 1752 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(U->getUser())) { in isVectorPromotionViableForSlice() 1964 } else if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(U->getUser())) { in isIntegerWideningViableForSlice()
|
/external/llvm/include/llvm/IR/ |
D | IntrinsicInst.h | 143 class MemIntrinsic : public IntrinsicInst { 220 class MemSetInst : public MemIntrinsic { 245 class MemTransferInst : public MemIntrinsic {
|
D | InstVisitor.h | 214 RetTy visitMemSetInst(MemSetInst &I) { DELEGATE(MemIntrinsic); } in visitMemSetInst() 217 RetTy visitMemTransferInst(MemTransferInst &I) { DELEGATE(MemIntrinsic); } in visitMemTransferInst() 218 RetTy visitMemIntrinsic(MemIntrinsic &I) { DELEGATE(IntrinsicInst); } in visitMemIntrinsic()
|
/external/llvm/lib/Transforms/Instrumentation/ |
D | SafeStack.cpp | 168 bool IsMemIntrinsicSafe(const MemIntrinsic *MI, const Use &U, 238 bool SafeStack::IsMemIntrinsicSafe(const MemIntrinsic *MI, const Use &U, in IsMemIntrinsicSafe() 305 if (const MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) { in IsSafeStackAlloca()
|
D | AddressSanitizer.cpp | 452 void instrumentMemIntrinsic(MemIntrinsic *MI); 833 void AddressSanitizer::instrumentMemIntrinsic(MemIntrinsic *MI) { in instrumentMemIntrinsic() 1629 } else if (isa<MemIntrinsic>(Inst)) { in runOnFunction() 1666 instrumentMemIntrinsic(cast<MemIntrinsic>(Inst)); in runOnFunction()
|
D | ThreadSanitizer.cpp | 358 if (isa<MemIntrinsic>(Inst)) in runOnFunction()
|
/external/llvm/lib/Analysis/ |
D | MemoryLocation.cpp | 80 MemoryLocation MemoryLocation::getForDest(const MemIntrinsic *MTI) { in getForDest()
|
D | LazyValueInfo.cpp | 643 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) { in InstructionDereferencesPointer()
|
/external/llvm/lib/Transforms/InstCombine/ |
D | InstCombineInternal.h | 40 class MemIntrinsic; variable 561 Instruction *SimplifyMemTransfer(MemIntrinsic *MI);
|
D | InstCombineCalls.cpp | 63 Instruction *InstCombiner::SimplifyMemTransfer(MemIntrinsic *MI) { in SimplifyMemTransfer() 702 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(II)) { in visitCallInst()
|
D | InstructionCombining.cpp | 1900 MemIntrinsic *MI = cast<MemIntrinsic>(II); in isAllocSiteRemovable()
|
/external/llvm/include/llvm-c/ |
D | Core.h | 1100 macro(MemIntrinsic) \
|
/external/llvm/lib/Transforms/IPO/ |
D | GlobalOpt.cpp | 341 } else if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(U)) { // memset/cpy/mv in CleanupConstantGlobalUsers()
|
/external/llvm/lib/Transforms/Vectorize/ |
D | SLPVectorizer.cpp | 351 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) in isSimple()
|
/external/llvm/lib/CodeGen/ |
D | CodeGenPrepare.cpp | 1752 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(CI)) { in optimizeCallInst()
|
/external/llvm/lib/Target/ARM/ |
D | ARMISelLowering.cpp | 1261 if (!isa<MemIntrinsic>(CI)) in shouldAlignPointerArgs()
|