/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Transforms/Utils/ |
D | SimplifyLibCalls.h | 24 class CallInst; variable 53 Value *optimizeCall(CallInst *CI); 56 Value *optimizeMemCpyChk(CallInst *CI, IRBuilder<> &B); 57 Value *optimizeMemMoveChk(CallInst *CI, IRBuilder<> &B); 58 Value *optimizeMemSetChk(CallInst *CI, IRBuilder<> &B); 61 Value *optimizeStrpCpyChk(CallInst *CI, IRBuilder<> &B, LibFunc Func); 62 Value *optimizeStrpNCpyChk(CallInst *CI, IRBuilder<> &B, LibFunc Func); 63 Value *optimizeMemCCpyChk(CallInst *CI, IRBuilder<> &B); 64 Value *optimizeSNPrintfChk(CallInst *CI, IRBuilder<> &B); 65 Value *optimizeSPrintfChk(CallInst *CI,IRBuilder<> &B); [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/ |
D | LibCallsShrinkWrap.cpp | 78 void visitCallInst(CallInst &CI) { checkCandidate(CI); } in visitCallInst() 93 bool perform(CallInst *CI); 94 void checkCandidate(CallInst &CI); 95 void shrinkWrapCI(CallInst *CI, Value *Cond); 96 bool performCallDomainErrorOnly(CallInst *CI, const LibFunc &Func); 97 bool performCallErrors(CallInst *CI, const LibFunc &Func); 98 bool performCallRangeErrorOnly(CallInst *CI, const LibFunc &Func); 99 Value *generateOneRangeCond(CallInst *CI, const LibFunc &Func); 100 Value *generateTwoRangeCond(CallInst *CI, const LibFunc &Func); 101 Value *generateCondForPow(CallInst *CI, const LibFunc &Func); [all …]
|
D | SimplifyLibCalls.cpp | 60 static bool isCallingConvCCompatible(CallInst *CI) { in isCallingConvCCompatible() 104 static bool callHasFloatingPointArgument(const CallInst *CI) { in callHasFloatingPointArgument() 110 static bool callHasFP128Argument(const CallInst *CI) { in callHasFP128Argument() 116 static Value *convertStrToNumber(CallInst *CI, StringRef &Str, int64_t Base) { in convertStrToNumber() 141 static bool isLocallyOpenedFile(Value *File, CallInst *CI, IRBuilder<> &B, in isLocallyOpenedFile() 143 CallInst *FOpen = dyn_cast<CallInst>(File); in isLocallyOpenedFile() 175 static bool canTransformToMemCmp(CallInst *CI, Value *Str, uint64_t Len, in canTransformToMemCmp() 190 static void annotateDereferenceableBytes(CallInst *CI, in annotateDereferenceableBytes() 217 static void annotateNonNullBasedOnAccess(CallInst *CI, in annotateNonNullBasedOnAccess() 235 static void annotateNonNullAndDereferenceable(CallInst *CI, ArrayRef<unsigned> ArgNos, in annotateNonNullAndDereferenceable() [all …]
|
D | EntryExitInstrumenter.cpp | 35 CallInst *Call = CallInst::Create(Fn, "", InsertionPt); in insertCall() 46 Instruction *RetAddr = CallInst::Create( in insertCall() 55 CallInst *Call = in insertCall() 56 CallInst::Create(Fn, ArrayRef<Value *>(Args), "", InsertionPt); in insertCall() 102 if (CallInst *CI = dyn_cast_or_null<CallInst>(Prev)) { in runOnFunction()
|
D | EscapeEnumerator.cpp | 59 if (CallInst *CI = dyn_cast<CallInst>(&II)) in Next() 88 CallInst *CI = cast<CallInst>(Calls[--I]); in Next()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/IR/ |
D | IRBuilder.cpp | 74 static CallInst *createCallHelper(Function *Callee, ArrayRef<Value *> Ops, in createCallHelper() 78 CallInst *CI = CallInst::Create(Callee, Ops, Name); in createCallHelper() 99 CallInst *IRBuilderBase::CreateMemSet(Value *Ptr, Value *Val, Value *Size, in CreateMemSet() 109 CallInst *CI = createCallHelper(TheFn, Ops, this); in CreateMemSet() 127 CallInst *IRBuilderBase::CreateElementUnorderedAtomicMemSet( in CreateElementUnorderedAtomicMemSet() 138 CallInst *CI = createCallHelper(TheFn, Ops, this); in CreateElementUnorderedAtomicMemSet() 155 CallInst *IRBuilderBase::CreateMemCpy(Value *Dst, unsigned DstAlign, Value *Src, in CreateMemCpy() 165 CallInst *IRBuilderBase::CreateMemCpy(Value *Dst, MaybeAlign DstAlign, in CreateMemCpy() 178 CallInst *CI = createCallHelper(TheFn, Ops, this); in CreateMemCpy() 203 CallInst *IRBuilderBase::CreateElementUnorderedAtomicMemCpy( in CreateElementUnorderedAtomicMemCpy() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Analysis/ |
D | MemoryBuiltins.h | 33 class CallInst; variable 119 const CallInst * 122 inline CallInst * 125 return const_cast<CallInst *>(extractMallocCall((const Value *)I, GetTLI)); in extractMallocCall() 133 PointerType *getMallocType(const CallInst *CI, const TargetLibraryInfo *TLI); 140 Type *getMallocAllocatedType(const CallInst *CI, const TargetLibraryInfo *TLI); 147 Value *getMallocArraySize(CallInst *CI, const DataLayout &DL, 157 const CallInst *extractCallocCall(const Value *I, const TargetLibraryInfo *TLI); 158 inline CallInst *extractCallocCall(Value *I, const TargetLibraryInfo *TLI) { in extractCallocCall() 159 return const_cast<CallInst*>(extractCallocCall((const Value*)I, TLI)); in extractCallocCall() [all …]
|
D | TypeMetadataUtils.h | 43 SmallVectorImpl<CallInst *> &Assumes, const CallInst *CI, 52 const CallInst *CI, DominatorTree &DT);
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
D | AMDGPULibCalls.cpp | 77 bool replaceWithNative(CallInst *CI, const FuncInfo &FInfo); 82 bool TDOFold(CallInst *CI, const FuncInfo &FInfo); 87 bool fold_recip(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 90 bool fold_divide(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 93 bool fold_pow(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 96 bool fold_rootn(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 99 bool fold_fma_mad(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 102 bool sincosUseNative(CallInst *aCI, const FuncInfo &FInfo); 107 bool evaluateCall(CallInst *aCI, FuncInfo &FInfo); 110 bool fold_exp(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/SelectionDAG/ |
D | SelectionDAGBuilder.h | 50 class CallInst; variable 636 void LowerDeoptimizeCall(const CallInst *CI); 743 void visitMaskedLoad(const CallInst &I, bool IsExpanding = false); 744 void visitMaskedStore(const CallInst &I, bool IsCompressing = false); 745 void visitMaskedGather(const CallInst &I); 746 void visitMaskedScatter(const CallInst &I); 751 void visitCall(const CallInst &I); 752 bool visitMemCmpCall(const CallInst &I); 753 bool visitMemPCpyCall(const CallInst &I); 754 bool visitMemChrCall(const CallInst &I); [all …]
|
D | StatepointLowering.h | 27 class CallInst; variable 68 void scheduleRelocCall(const CallInst &RelocCall) { in scheduleRelocCall() 77 void relocCallVisited(const CallInst &RelocCall) { in relocCallVisited() 123 SmallVector<const CallInst *, 10> PendingGCRelocateCalls;
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ |
D | PreISelIntrinsicLowering.cpp | 40 auto CI = dyn_cast<CallInst>(I->getUser()); in lowerLoadRelative() 63 static CallInst::TailCallKind getOverridingTailCallKind(const Function &F) { in getOverridingTailCallKind() 66 return CallInst::TCK_Tail; in getOverridingTailCallKind() 68 return CallInst::TCK_NoTail; in getOverridingTailCallKind() 69 return CallInst::TCK_None; in getOverridingTailCallKind() 91 CallInst::TailCallKind OverridingTCK = getOverridingTailCallKind(F); in lowerObjCCall() 94 auto *CI = cast<CallInst>(I->getUser()); in lowerObjCCall() 100 CallInst *NewCI = Builder.CreateCall(FCache, Args); in lowerObjCCall() 110 CallInst::TailCallKind TCK = CI->getTailCallKind(); in lowerObjCCall()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/BPF/ |
D | BPFAbstractMemberAccess.cpp | 124 typedef std::stack<std::pair<CallInst *, CallInfo>> CallInfoStack; 138 std::map<CallInst *, std::pair<CallInst *, CallInfo>> AIChain; 142 std::map<CallInst *, CallInfo> BaseAICalls; 146 void traceAICall(CallInst *Call, CallInfo &ParentInfo); 147 void traceBitCast(BitCastInst *BitCast, CallInst *Parent, 149 void traceGEP(GetElementPtrInst *GEP, CallInst *Parent, 153 bool IsPreserveDIAccessIndexCall(const CallInst *Call, CallInfo &Cinfo); 157 void replaceWithGEP(std::vector<CallInst *> &CallList, 166 Value *computeBaseAndAccessKey(CallInst *Call, CallInfo &CInfo, 169 bool transformGEPChain(Module &M, CallInst *Call, CallInfo &CInfo); [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/ObjCARC/ |
D | ObjCARCOpts.cpp | 102 cast<CallInst>(Arg)->getArgOperand(0)); in FindSingleUseIdentifiedObject() 644 cast<CallInst>(RetainRV)->setCalledFunction(NewDecl); in OptimizeRetainRVCall() 680 cast<CallInst>(AutoreleaseRV)->getArgOperand(0)); in OptimizeInlinedAutoreleaseRVCall() 684 Inst->replaceAllUsesWith(cast<CallInst>(Inst)->getArgOperand(0)); in OptimizeInlinedAutoreleaseRVCall() 692 Value *CallArg = cast<CallInst>(Inst)->getArgOperand(0); in OptimizeInlinedAutoreleaseRVCall() 693 CallInst *Release = CallInst::Create( in OptimizeInlinedAutoreleaseRVCall() 747 CallInst *AutoreleaseRVCI = cast<CallInst>(AutoreleaseRV); in OptimizeAutoreleaseRVCall() 758 CloneCallInstForBB(CallInst &CI, BasicBlock &BB, in CloneCallInstForBB() 777 return CallInst::Create(&CI, OpBundles); in CloneCallInstForBB() 923 CallInst *CI = cast<CallInst>(Inst); in OptimizeIndividualCallImpl() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/ |
D | AssumptionCache.cpp | 57 static void findAffectedValues(CallInst *CI, in findAffectedValues() 114 void AssumptionCache::updateAffectedValues(CallInst *CI) { in updateAffectedValues() 125 void AssumptionCache::unregisterAssumption(CallInst *CI) { in unregisterAssumption() 187 updateAffectedValues(cast<CallInst>(A)); in scanFunction() 190 void AssumptionCache::registerAssumption(CallInst *CI) { in registerAssumption() 217 assert(match(cast<CallInst>(VH), m_Intrinsic<Intrinsic::assume>()) && in registerAssumption() 236 OS << " " << *cast<CallInst>(VH)->getArgOperand(0) << "\n"; in run() 280 SmallPtrSet<const CallInst *, 4> AssumptionSet; in verifyAnalysis() 284 AssumptionSet.insert(cast<CallInst>(VH)); in verifyAnalysis() 289 !AssumptionSet.count(cast<CallInst>(&II))) in verifyAnalysis()
|
D | TypeMetadataUtils.cpp | 26 const CallInst *CI, DominatorTree &DT) { in findCallsAtConstantOffset() 40 } else if (auto CI = dyn_cast<CallInst>(User)) { in findCallsAtConstantOffset() 53 int64_t Offset, const CallInst *CI, DominatorTree &DT) { in findLoadCallsAtConstantOffset() 75 SmallVectorImpl<CallInst *> &Assumes, const CallInst *CI, in findDevirtualizableCallsForTypeTest() 83 if (auto *AssumeCI = dyn_cast<CallInst>(CIU.getUser())) { in findDevirtualizableCallsForTypeTest() 101 const CallInst *CI, DominatorTree &DT) { in findDevirtualizableCallsForTypeCheckedLoad()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/CodeGen/ |
D | IntrinsicLowering.h | 21 class CallInst; variable 41 void LowerIntrinsicCall(CallInst *CI); 45 static bool LowerToByteSwap(CallInst *CI);
|
D | FastISel.h | 38 class CallInst; variable 509 bool lowerCallTo(const CallInst *CI, MCSymbol *Symbol, unsigned NumArgs); 510 bool lowerCallTo(const CallInst *CI, const char *SymName, 526 bool lowerCall(const CallInst *I); 532 bool selectStackmap(const CallInst *I); 533 bool selectPatchpoint(const CallInst *I); 540 bool selectXRayCustomEvent(const CallInst *II); 541 bool selectXRayTypedEvent(const CallInst *II); 595 const CallInst *CI, unsigned StartIdx); 596 bool lowerCallOperands(const CallInst *CI, unsigned ArgIdx, unsigned NumArgs,
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/IR/ |
D | IRBuilder.h | 265 void setConstrainedFPCallAttr(CallInst *I) { in setConstrainedFPCallAttr() 452 CallInst *CreateMemSet(Value *Ptr, Value *Val, uint64_t Size, 460 CallInst *CreateMemSet(Value *Ptr, Value *Val, Value *Size, MaybeAlign Align, 474 CallInst *CreateElementUnorderedAtomicMemSet( 484 CallInst *CreateElementUnorderedAtomicMemSet(Value *Ptr, Value *Val, 498 CallInst *CreateElementUnorderedAtomicMemSet( 508 CallInst *CreateElementUnorderedAtomicMemSet(Value *Ptr, Value *Val, 523 CallInst *CreateMemCpy(Value *Dst, unsigned DstAlign, Value *Src, 535 CallInst *CreateMemCpy(Value *Dst, MaybeAlign DstAlign, Value *Src, 549 CallInst *CreateMemCpy(Value *Dst, unsigned DstAlign, Value *Src, [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/ |
D | LowerExpectIntrinsic.cpp | 59 CallInst *CI = dyn_cast<CallInst>(SI.getCondition()); in handleSwitchExpect() 100 static void handlePhiDef(CallInst *Expect) { in handlePhiDef() 250 CallInst *CI; in handleBrSelExpect() 256 CI = dyn_cast<CallInst>(BSI.getCondition()); in handleBrSelExpect() 266 CI = dyn_cast<CallInst>(CmpI->getOperand(0)); in handleBrSelExpect() 340 CallInst *CI = dyn_cast<CallInst>(Inst); in lowerExpectIntrinsic()
|
D | TailRecursionElimination.cpp | 231 SmallVector<CallInst *, 32> DeferredTails; in markTails() 240 CallInst *CI = dyn_cast<CallInst>(&I); in markTails() 310 for (CallInst *CI : DeferredTails) { in markTails() 329 static bool canMoveAboveCall(Instruction *I, CallInst *CI, AliasAnalysis *AA) { in canMoveAboveCall() 363 static bool isDynamicConstant(Value *V, CallInst *CI, ReturnInst *RI) { in isDynamicConstant() 397 static Value *getCommonReturnValue(ReturnInst *IgnoreRI, CallInst *CI) { in getCommonReturnValue() 423 static Value *canTransformAccumulatorRecursion(Instruction *I, CallInst *CI) { in canTransformAccumulatorRecursion() 449 static CallInst *findTRECandidate(Instruction *TI, in findTRECandidate() 460 CallInst *CI = nullptr; in findTRECandidate() 463 CI = dyn_cast<CallInst>(BBI); in findTRECandidate() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ |
D | PPCBoolRetToInt.cpp | 83 if (CurrUser && !isa<CallInst>(Curr)) in findAllDefs() 141 return isa<ReturnInst>(V) || isa<CallInst>(V) || isa<PHINode>(V) || in getPromotablePHINodes() 145 return isa<Constant>(V) || isa<Argument>(V) || isa<CallInst>(V) || in getPromotablePHINodes() 208 if (auto *CI = dyn_cast<CallInst>(&I)) in runOnFunction() 231 !isa<Argument>(V) && !isa<CallInst>(V)) in runOnUse() 241 if (isa<CallInst>(U.getUser())) in runOnUse() 257 if (First && !isa<CallInst>(First)) in runOnUse()
|
/third_party/mesa3d/src/gallium/drivers/swr/rasterizer/jitter/ |
D | builder_misc.h | 121 CallInst* CALL(Value* Callee, const std::initializer_list<Value*>& args, const llvm::Twine& name = … 122 CallInst* CALL(Value* Callee) in CALL() 133 CallInst* CALL(Value* Callee, Value* arg); 134 CallInst* CALL2(Value* Callee, Value* arg1, Value* arg2); 135 CallInst* CALL3(Value* Callee, Value* arg1, Value* arg2, Value* arg3); 191 CallInst* PRINT(const std::string& printStr); 192 CallInst* PRINT(const std::string& printStr, const std::initializer_list<Value*>& printArgs);
|
/third_party/mesa3d/src/gallium/drivers/swr/rasterizer/jitter/functionpasses/ |
D | lower_x86.cpp | 67 typedef std::function<Instruction*(LowerX86*, TargetArch, TargetWidth, CallInst*)> EmuFunc; 93 Instruction* NO_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 95 VPERM_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 97 VGATHER_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 99 VSCATTER_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 101 VROUND_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 103 VHSUB_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 105 VCONVERT_EMU(LowerX86* pThis, TargetArch arch, TargetWidth width, CallInst* pCallInst); 110 CallInst* pCallInst, 240 void GetRequestedWidthAndType(CallInst* pCallInst, in GetRequestedWidthAndType() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/NVPTX/ |
D | NVVMReflect.cpp | 116 CallInst *Call = dyn_cast<CallInst>(&I); in runOnFunction() 131 if (const CallInst *ConvCall = dyn_cast<CallInst>(Str)) { in runOnFunction()
|