| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
| D | AMDGPUUnifyDivergentExitNodes.cpp | 125 if (IntrinsicInst *Intrin = llvm::dyn_cast<IntrinsicInst>(&I)) { in removeDoneExport() local
|
| D | AMDGPUCodeGenPrepare.cpp | 498 FunctionCallee Intrin = Intrinsic::getDeclaration(Mod, IntrID); in replaceMulWithMul24() local
|
| /external/llvm/lib/Target/AMDGPU/ |
| D | AMDGPUAnnotateKernelFeatures.cpp | 141 void AMDGPUAnnotateKernelFeatures::addAttrToCallers(Function *Intrin, in addAttrToCallers()
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Coroutines/ |
| D | CoroEarly.cpp | 65 void Lowerer::lowerCoroPromise(CoroPromiseInst *Intrin) { in lowerCoroPromise()
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/IR/ |
| D | AutoUpgrade.cpp | 1123 Function *Intrin = Intrinsic::getDeclaration(CI.getModule(), IID, Ty); in UpgradeX86AddSubSatIntrinsics() local 1150 Function *Intrin = Intrinsic::getDeclaration(CI.getModule(), IID, Ty); in upgradeX86Rotate() local 1220 Function *Intrin = Intrinsic::getDeclaration(CI.getModule(), IID, Ty); in upgradeX86ConcatShift() local 1385 Function *Intrin = Intrinsic::getDeclaration(CI.getModule(), IID); in UpgradeX86MaskedShift() local
|
| /external/llvm/lib/Transforms/Utils/ |
| D | Local.cpp | 1921 Intrinsic::ID Intrin; in recognizeBSwapOrBitReverseIdiom() local
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/NVPTX/ |
| D | NVPTXISelLowering.cpp | 4913 SDValue Intrin = N->getOperand(1); in ReplaceINTRINSIC_W_CHAIN() local
|
| /external/llvm/lib/Target/NVPTX/ |
| D | NVPTXISelLowering.cpp | 4391 SDValue Intrin = N->getOperand(1); in ReplaceINTRINSIC_W_CHAIN() local
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/ |
| D | Local.cpp | 2872 Intrinsic::ID Intrin; in recognizeBSwapOrBitReverseIdiom() local
|
| /external/llvm/lib/Target/PowerPC/ |
| D | PPCISelLowering.cpp | 7656 static bool getVectorCompareInfo(SDValue Intrin, int &CompareOpc, in getVectorCompareInfo() 10457 MemIntrinsicSDNode *Intrin = cast<MemIntrinsicSDNode>(N); in expandVSXLoadForLE() local 10519 MemIntrinsicSDNode *Intrin = cast<MemIntrinsicSDNode>(N); in expandVSXStoreForLE() local
|
| /external/clang/lib/CodeGen/ |
| D | CGBuiltin.cpp | 6158 llvm::Function *Intrin = CGM.getIntrinsic(Int, Ty); in EmitAArch64BuiltinExpr() local 6164 llvm::Function *Intrin = CGM.getIntrinsic(Int, Ty); in EmitAArch64BuiltinExpr() local
|
| /external/llvm/lib/Transforms/InstCombine/ |
| D | InstCombineCompares.cpp | 1108 } else if (auto *Intrin = dyn_cast<IntrinsicInst>(V)) { in FoldAllocaCmp() local
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ |
| D | PPCISelLowering.cpp | 9516 static bool getVectorCompareInfo(SDValue Intrin, int &CompareOpc, in getVectorCompareInfo() 13459 MemIntrinsicSDNode *Intrin = cast<MemIntrinsicSDNode>(N); in expandVSXLoadForLE() local 13529 MemIntrinsicSDNode *Intrin = cast<MemIntrinsicSDNode>(N); in expandVSXStoreForLE() local
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/InstCombine/ |
| D | InstCombineCompares.cpp | 1074 } else if (const auto *Intrin = dyn_cast<IntrinsicInst>(V)) { in foldAllocaCmp() local
|
| /external/swiftshader/third_party/subzero/src/ |
| D | IceTargetLoweringX8664.cpp | 699 } else if (auto *Intrin = llvm::dyn_cast<InstIntrinsic>(CurInst)) { in doLoadOpt() local
|
| D | IceTargetLoweringX8632.cpp | 688 } else if (auto *Intrin = llvm::dyn_cast<InstIntrinsic>(CurInst)) { in doLoadOpt() local
|
| /external/llvm/lib/Target/AArch64/ |
| D | AArch64ISelLowering.cpp | 5928 unsigned Intrin = in tryLowerToSLI() local
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
| D | AArch64ISelLowering.cpp | 7734 unsigned Intrin = in tryLowerToSLI() local
|
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
| D | X86ISelLowering.cpp | 36454 SDValue Intrin = DAG.getConstant(IntrinOp, DL, MVT::i32); in createMMXBuildVector() local
|