/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/ |
D | VNCoercion.cpp | 15 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, in canCoerceMustAliasedValueToLoad() argument 18 if (StoredTy == LoadTy) in canCoerceMustAliasedValueToLoad() 23 if (LoadTy->isStructTy() || LoadTy->isArrayTy() || StoredTy->isStructTy() || in canCoerceMustAliasedValueToLoad() 34 if (StoreSize < DL.getTypeSizeInBits(LoadTy)) in canCoerceMustAliasedValueToLoad() 39 DL.isNonIntegralPointerType(LoadTy->getScalarType())) { in canCoerceMustAliasedValueToLoad() 163 static int analyzeLoadFromClobberingWrite(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingWrite() argument 169 if (LoadTy->isStructTy() || LoadTy->isArrayTy()) in analyzeLoadFromClobberingWrite() 187 uint64_t LoadSize = DL.getTypeSizeInBits(LoadTy); in analyzeLoadFromClobberingWrite() 218 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingStore() argument 229 DL.isNonIntegralPointerType(LoadTy->getScalarType())) { in analyzeLoadFromClobberingStore() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Transforms/Utils/ |
D | VNCoercion.h | 37 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, 54 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, 62 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, 70 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr, 76 Value *getStoreValueForLoad(Value *SrcVal, unsigned Offset, Type *LoadTy, 81 Type *LoadTy, const DataLayout &DL); 87 Value *getLoadValueForLoad(LoadInst *SrcVal, unsigned Offset, Type *LoadTy, 92 Type *LoadTy, const DataLayout &DL); 99 Type *LoadTy, Instruction *InsertPt, 104 Type *LoadTy, const DataLayout &DL);
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Vectorize/ |
D | LoadStoreVectorizer.cpp | 1096 Type *LoadTy = nullptr; in vectorizeLoadChain() local 1098 LoadTy = cast<LoadInst>(V)->getType(); in vectorizeLoadChain() 1099 if (LoadTy->isIntOrIntVectorTy()) in vectorizeLoadChain() 1102 if (LoadTy->isPtrOrPtrVectorTy()) { in vectorizeLoadChain() 1103 LoadTy = Type::getIntNTy(F.getParent()->getContext(), in vectorizeLoadChain() 1104 DL.getTypeSizeInBits(LoadTy)); in vectorizeLoadChain() 1108 assert(LoadTy && "Can't determine LoadInst type from chain"); in vectorizeLoadChain() 1110 unsigned Sz = DL.getTypeSizeInBits(LoadTy); in vectorizeLoadChain() 1143 VectorType *VecLoadTy = dyn_cast<VectorType>(LoadTy); in vectorizeLoadChain() 1145 VecTy = VectorType::get(LoadTy->getScalarType(), in vectorizeLoadChain() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/ |
D | ConstantFolding.cpp | 502 Constant *FoldReinterpretLoadFromConstPtr(Constant *C, Type *LoadTy, in FoldReinterpretLoadFromConstPtr() argument 505 auto *IntType = dyn_cast<IntegerType>(LoadTy); in FoldReinterpretLoadFromConstPtr() 516 if (LoadTy->isHalfTy()) in FoldReinterpretLoadFromConstPtr() 518 else if (LoadTy->isFloatTy()) in FoldReinterpretLoadFromConstPtr() 520 else if (LoadTy->isDoubleTy()) in FoldReinterpretLoadFromConstPtr() 522 else if (LoadTy->isVectorTy()) { in FoldReinterpretLoadFromConstPtr() 524 DL.getTypeSizeInBits(LoadTy)); in FoldReinterpretLoadFromConstPtr() 530 if (Res->isNullValue() && !LoadTy->isX86_MMXTy()) in FoldReinterpretLoadFromConstPtr() 532 return Constant::getNullValue(LoadTy); in FoldReinterpretLoadFromConstPtr() 533 Type *CastTy = LoadTy->isPtrOrPtrVectorTy() ? DL.getIntPtrType(LoadTy) : LoadTy; in FoldReinterpretLoadFromConstPtr() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/ARM/ |
D | ARMParallelDSP.cpp | 229 LoadInst* CreateWideLoad(MemInstList &Loads, IntegerType *LoadTy); 734 IntegerType *LoadTy) { in CreateWideLoad() argument 774 LoadTy->getPointerTo(AddrSpace)); in CreateWideLoad() 775 LoadInst *WideLoad = IRB.CreateAlignedLoad(LoadTy, VecPtr, in CreateWideLoad() 790 Value *ShiftVal = ConstantInt::get(LoadTy, OffsetTy->getBitWidth()); in CreateWideLoad()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/ |
D | GVN.cpp | 776 Type *LoadTy = LI->getType(); in MaterializeAdjustedValue() local 780 if (Res->getType() != LoadTy) { in MaterializeAdjustedValue() 781 Res = getStoreValueForLoad(Res, Offset, LoadTy, InsertPt, DL); in MaterializeAdjustedValue() 790 if (Load->getType() == LoadTy && Offset == 0) { in MaterializeAdjustedValue() 793 Res = getLoadValueForLoad(Load, Offset, LoadTy, InsertPt, DL); in MaterializeAdjustedValue() 806 Res = getMemInstValueForLoad(getMemIntrinValue(), Offset, LoadTy, in MaterializeAdjustedValue() 815 return UndefValue::get(LoadTy); in MaterializeAdjustedValue()
|
D | SROA.cpp | 1267 Type *LoadTy = SomeLoad->getType(); in speculatePHINodeLoads() local 1269 PHINode *NewPN = PHIBuilder.CreatePHI(LoadTy, PN.getNumIncomingValues(), in speculatePHINodeLoads() 1304 LoadTy, InVal, in speculatePHINodeLoads()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
D | AMDGPURegisterBankInfo.cpp | 520 LLT LoadTy = MRI.getType(MI.getOperand(0).getReg()); in getInstrAlternativeMappings() local 535 {AMDGPU::getValueMappingLoadSGPROnly(AMDGPU::VGPRRegBankID, LoadTy), in getInstrAlternativeMappings() 1117 const LLT LoadTy = MRI.getType(DstReg); in applyMappingWideLoad() local 1118 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingWideLoad() 1152 MaxNonSmrdLoadSize / LoadTy.getScalarType().getSizeInBits(); in applyMappingWideLoad() 1153 const LLT LoadSplitTy = LLT::vector(SplitElts, LoadTy.getScalarType()); in applyMappingWideLoad() 2355 LLT LoadTy = MRI.getType(MI.getOperand(0).getReg()); in getInstrMappingForLoad() local 2374 ValMapping = AMDGPU::getValueMappingLoadSGPROnly(AMDGPU::VGPRRegBankID, LoadTy); in getInstrMappingForLoad()
|
D | AMDGPUISelLowering.cpp | 673 bool AMDGPUTargetLowering::isLoadBitCastBeneficial(EVT LoadTy, EVT CastTy, in isLoadBitCastBeneficial() argument 677 assert(LoadTy.getSizeInBits() == CastTy.getSizeInBits()); in isLoadBitCastBeneficial() 679 if (LoadTy.getScalarType() == MVT::i32) in isLoadBitCastBeneficial() 682 unsigned LScalarSize = LoadTy.getScalarSizeInBits(); in isLoadBitCastBeneficial()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Hexagon/ |
D | HexagonISelLowering.cpp | 2717 MVT LoadTy = ty(Op); in LowerUnalignedLoad() local 2718 unsigned NeedAlign = Subtarget.getTypeAlignment(LoadTy); in LowerUnalignedLoad() 2757 assert(LoadTy.getSizeInBits() == 8*NeedAlign); in LowerUnalignedLoad() 2788 SDValue Load0 = DAG.getLoad(LoadTy, dl, Chain, Base0, WideMMO); in LowerUnalignedLoad() 2789 SDValue Load1 = DAG.getLoad(LoadTy, dl, Chain, Base1, WideMMO); in LowerUnalignedLoad() 2791 SDValue Aligned = DAG.getNode(HexagonISD::VALIGN, dl, LoadTy, in LowerUnalignedLoad()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/InstCombine/ |
D | InstCombineLoadStoreAlloca.cpp | 536 static bool isMinMaxWithLoads(Value *V, Type *&LoadTy) { in isMinMaxWithLoads() argument 550 LoadTy = L1->getType(); in isMinMaxWithLoads()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/SelectionDAG/ |
D | LegalizeDAG.cpp | 1460 MVT LoadTy = TLI.getRegisterType(*DAG.getContext(), MVT::i8); in getSignAsIntValue() local 1462 SDValue StackPtr = DAG.CreateStackTemporary(FloatVT, LoadTy); in getSignAsIntValue() 1486 State.IntValue = DAG.getExtLoad(ISD::EXTLOAD, DL, LoadTy, State.Chain, IntPtr, in getSignAsIntValue() 1488 State.SignMask = APInt::getOneBitSet(LoadTy.getSizeInBits(), 7); in getSignAsIntValue()
|
D | SelectionDAGBuilder.cpp | 7265 Type *LoadTy = in getMemCmpLoad() local 7268 LoadTy = VectorType::get(LoadTy, LoadVT.getVectorNumElements()); in getMemCmpLoad() 7271 PointerType::getUnqual(LoadTy)); in getMemCmpLoad() 7274 const_cast<Constant *>(LoadInput), LoadTy, *Builder.DL)) in getMemCmpLoad()
|