/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/ |
D | VNCoercion.cpp | 187 uint64_t LoadSize = DL.getTypeSizeInBits(LoadTy); in analyzeLoadFromClobberingWrite() local 189 if ((WriteSizeInBits & 7) | (LoadSize & 7)) in analyzeLoadFromClobberingWrite() 192 LoadSize /= 8; in analyzeLoadFromClobberingWrite() 198 isAAFailure = LoadOffset + int64_t(LoadSize) <= StoreOffset; in analyzeLoadFromClobberingWrite() 208 StoreOffset + StoreSize < LoadOffset + LoadSize) in analyzeLoadFromClobberingWrite() 268 unsigned LoadSize = DL.getTypeStoreSize(LoadTy); in analyzeLoadFromClobberingLoad() local 271 LoadBase, LoadOffs, LoadSize, DepLI); in analyzeLoadFromClobberingLoad() 359 uint64_t LoadSize = (DL.getTypeSizeInBits(LoadTy) + 7) / 8; in getStoreValueForLoadHelper() local 372 ShiftAmt = (StoreSize - LoadSize - Offset) * 8; in getStoreValueForLoadHelper() 377 if (LoadSize != StoreSize) in getStoreValueForLoadHelper() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
D | AMDGPULowerKernelAttributes.cpp | 112 unsigned LoadSize = DL.getTypeStoreSize(Load->getType()); in processUse() local 117 if (LoadSize == 2) in processUse() 121 if (LoadSize == 2) in processUse() 125 if (LoadSize == 2) in processUse() 129 if (LoadSize == 4) in processUse() 133 if (LoadSize == 4) in processUse() 137 if (LoadSize == 4) in processUse()
|
D | AMDGPUTargetTransformInfo.cpp | 236 unsigned GCNTTIImpl::getLoadVectorFactor(unsigned VF, unsigned LoadSize, in getLoadVectorFactor() argument 239 unsigned VecRegBitWidth = VF * LoadSize; in getLoadVectorFactor() 242 return 128 / LoadSize; in getLoadVectorFactor()
|
D | AMDGPUTargetTransformInfo.h | 152 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize,
|
D | AMDGPURegisterBankInfo.cpp | 1118 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingWideLoad() local 1121 if (LoadSize <= MaxNonSmrdLoadSize) in applyMappingWideLoad() 1136 assert(LoadSize % MaxNonSmrdLoadSize == 0); in applyMappingWideLoad()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ |
D | ExpandMemCmp.cpp | 81 LoadEntry(unsigned LoadSize, uint64_t Offset) in LoadEntry() 82 : LoadSize(LoadSize), Offset(Offset) { in LoadEntry() 86 unsigned LoadSize; member 135 const unsigned LoadSize = LoadSizes.front(); in computeGreedyLoadSequence() local 136 const uint64_t NumLoadsForThisSize = Size / LoadSize; in computeGreedyLoadSequence() 146 LoadSequence.push_back({LoadSize, Offset}); in computeGreedyLoadSequence() 147 Offset += LoadSize; in computeGreedyLoadSequence() 149 if (LoadSize > 1) in computeGreedyLoadSequence() 151 Size = Size % LoadSize; in computeGreedyLoadSequence() 345 IntegerType::get(CI->getContext(), CurLoadEntry.LoadSize * 8); in getCompareLoadPairs() [all …]
|
D | StackSlotColoring.cpp | 448 unsigned LoadSize = 0; in RemoveDeadStores() local 450 if (!(LoadReg = TII->isLoadFromStackSlot(*I, FirstSS, LoadSize))) in RemoveDeadStores() 461 LoadSize != StoreSize) in RemoveDeadStores()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 580 int LoadSize = TII->getMemScale(LoadInst); in isLdOffsetInRangeOfSt() local 587 : getLdStOffsetOp(LoadInst).getImm() * LoadSize; in isLdOffsetInRangeOfSt() 589 (UnscaledLdOffset + LoadSize <= (UnscaledStOffset + StoreSize)); in isLdOffsetInRangeOfSt() 1014 int LoadSize = TII->getMemScale(*LoadI); in promoteLoadFromStore() local 1026 if (LoadSize == StoreSize && (LoadSize == 4 || LoadSize == 8)) { in promoteLoadFromStore() 1029 if (StRt == LdRt && LoadSize == 8) { in promoteLoadFromStore() 1059 assert(LoadSize <= StoreSize && "Invalid load size"); in promoteLoadFromStore() 1062 : getLdStOffsetOp(*LoadI).getImm() * LoadSize; in promoteLoadFromStore() 1066 int Width = LoadSize * 8; in promoteLoadFromStore() 1073 (UnscaledLdOffset + LoadSize) <= UnscaledStOffset + StoreSize) && in promoteLoadFromStore()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/ |
D | Loads.cpp | 280 const uint64_t LoadSize = Size.getZExtValue(); in isSafeToLoadUnconditionally() local 331 LoadSize <= DL.getTypeStoreSize(AccessedTy)) in isSafeToLoadUnconditionally() 335 LoadSize <= DL.getTypeStoreSize(AccessedTy)) in isSafeToLoadUnconditionally()
|
D | TargetTransformInfo.cpp | 829 unsigned LoadSize, in getLoadVectorFactor() argument 832 return TTIImpl->getLoadVectorFactor(VF, LoadSize, ChainSizeInBytes, VecTy); in getLoadVectorFactor()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ |
D | PPCHazardRecognizers.h | 94 bool isLoadOfStoredAddress(uint64_t LoadSize, int64_t LoadOffset,
|
D | PPCHazardRecognizers.cpp | 298 isLoadOfStoredAddress(uint64_t LoadSize, int64_t LoadOffset, in isLoadOfStoredAddress() argument 313 if (int64_t(LoadOffset+LoadSize) > StoreOffset[i]) return true; in isLoadOfStoredAddress()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Analysis/ |
D | TargetTransformInfo.h | 1121 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, 1388 virtual unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, 1859 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, in getLoadVectorFactor() argument 1862 return Impl.getLoadVectorFactor(VF, LoadSize, ChainSizeInBytes, VecTy); in getLoadVectorFactor()
|
D | TargetTransformInfoImpl.h | 594 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, in getLoadVectorFactor() argument
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
D | X86AvoidStoreForwardingBlocks.cpp | 628 static bool isBlockingStore(int64_t LoadDispImm, unsigned LoadSize, in isBlockingStore() argument 631 (StoreDispImm <= LoadDispImm + (LoadSize - StoreSize))); in isBlockingStore()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/ |
D | SROA.cpp | 3841 uint64_t LoadSize = Ty->getBitWidth() / 8; in presplitLoadsAndStores() local 3842 assert(LoadSize > 0 && "Cannot have a zero-sized integer load!"); in presplitLoadsAndStores() 3845 assert(LoadSize == Offsets.S->endOffset() - Offsets.S->beginOffset() && in presplitLoadsAndStores() 3848 assert(BaseOffset + LoadSize > BaseOffset && in presplitLoadsAndStores() 3892 PartSize = (Idx < Size ? Offsets.Splits[Idx] : LoadSize) - PartOffset; in presplitLoadsAndStores()
|