/external/llvm-project/llvm/lib/Transforms/Scalar/ |
D | MergeICmps.cpp | 83 BCEAtom(GetElementPtrInst *GEP, LoadInst *LoadI, int BaseId, APInt Offset) in BCEAtom() 84 : GEP(GEP), LoadI(LoadI), BaseId(BaseId), Offset(Offset) {} in BCEAtom() 94 LoadI = that.LoadI; in operator =() 115 LoadInst *LoadI = nullptr; member 143 auto *const LoadI = dyn_cast<LoadInst>(Val); in visitICmpLoadOperand() local 144 if (!LoadI) in visitICmpLoadOperand() 147 if (LoadI->isUsedOutsideOfBlock(LoadI->getParent())) { in visitICmpLoadOperand() 152 if (!LoadI->isSimple()) { in visitICmpLoadOperand() 156 Value *const Addr = LoadI->getOperand(0); in visitICmpLoadOperand() 161 if (GEP->isUsedOutsideOfBlock(LoadI->getParent())) { in visitICmpLoadOperand() [all …]
|
D | JumpThreading.cpp | 1203 if (LoadInst *LoadI = dyn_cast<LoadInst>(SimplifyValue)) in processBlock() local 1204 if (simplifyPartiallyRedundantLoad(LoadI)) in processBlock() 1293 bool JumpThreadingPass::simplifyPartiallyRedundantLoad(LoadInst *LoadI) { in simplifyPartiallyRedundantLoad() argument 1295 if (!LoadI->isUnordered()) return false; in simplifyPartiallyRedundantLoad() 1299 BasicBlock *LoadBB = LoadI->getParent(); in simplifyPartiallyRedundantLoad() 1309 Value *LoadedPtr = LoadI->getOperand(0); in simplifyPartiallyRedundantLoad() 1318 BasicBlock::iterator BBIt(LoadI); in simplifyPartiallyRedundantLoad() 1321 LoadI, LoadBB, BBIt, DefMaxInstsToScan, AA, &IsLoadCSE)) { in simplifyPartiallyRedundantLoad() 1327 combineMetadataForCSE(NLoadI, LoadI, false); in simplifyPartiallyRedundantLoad() 1332 if (AvailableVal == LoadI) in simplifyPartiallyRedundantLoad() [all …]
|
D | DeadStoreElimination.cpp | 2448 if (auto *LoadI = dyn_cast<LoadInst>(Store->getOperand(0))) { in storeIsNoop() local 2449 if (LoadI->getPointerOperand() == Store->getOperand(1)) { in storeIsNoop() 2451 auto *LoadAccess = MSSA.getMemoryAccess(LoadI)->getDefiningAccess(); in storeIsNoop()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/ |
D | MergeICmps.cpp | 83 BCEAtom(GetElementPtrInst *GEP, LoadInst *LoadI, int BaseId, APInt Offset) in BCEAtom() 84 : GEP(GEP), LoadI(LoadI), BaseId(BaseId), Offset(Offset) {} in BCEAtom() 94 LoadI = that.LoadI; in operator =() 115 LoadInst *LoadI = nullptr; member 143 auto *const LoadI = dyn_cast<LoadInst>(Val); in visitICmpLoadOperand() local 144 if (!LoadI) in visitICmpLoadOperand() 147 if (LoadI->isUsedOutsideOfBlock(LoadI->getParent())) { in visitICmpLoadOperand() 152 if (!LoadI->isSimple()) { in visitICmpLoadOperand() 156 Value *const Addr = LoadI->getOperand(0); in visitICmpLoadOperand() 161 if (GEP->isUsedOutsideOfBlock(LoadI->getParent())) { in visitICmpLoadOperand() [all …]
|
D | JumpThreading.cpp | 1167 if (LoadInst *LoadI = dyn_cast<LoadInst>(SimplifyValue)) in ProcessBlock() local 1168 if (SimplifyPartiallyRedundantLoad(LoadI)) in ProcessBlock() 1252 bool JumpThreadingPass::SimplifyPartiallyRedundantLoad(LoadInst *LoadI) { in SimplifyPartiallyRedundantLoad() argument 1254 if (!LoadI->isUnordered()) return false; in SimplifyPartiallyRedundantLoad() 1258 BasicBlock *LoadBB = LoadI->getParent(); in SimplifyPartiallyRedundantLoad() 1268 Value *LoadedPtr = LoadI->getOperand(0); in SimplifyPartiallyRedundantLoad() 1277 BasicBlock::iterator BBIt(LoadI); in SimplifyPartiallyRedundantLoad() 1280 LoadI, LoadBB, BBIt, DefMaxInstsToScan, AA, &IsLoadCSE)) { in SimplifyPartiallyRedundantLoad() 1286 combineMetadataForCSE(NLoadI, LoadI, false); in SimplifyPartiallyRedundantLoad() 1291 if (AvailableVal == LoadI) in SimplifyPartiallyRedundantLoad() [all …]
|
/external/llvm-project/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 156 promoteLoadFromStore(MachineBasicBlock::iterator LoadI, 1009 AArch64LoadStoreOpt::promoteLoadFromStore(MachineBasicBlock::iterator LoadI, in promoteLoadFromStore() argument 1012 next_nodbg(LoadI, LoadI->getParent()->end()); in promoteLoadFromStore() 1014 int LoadSize = TII->getMemScale(*LoadI); in promoteLoadFromStore() 1016 Register LdRt = getLdStRegOp(*LoadI).getReg(); in promoteLoadFromStore() 1031 LoadI->getIterator())) { in promoteLoadFromStore() 1038 LLVM_DEBUG(LoadI->print(dbgs())); in promoteLoadFromStore() 1040 LoadI->eraseFromParent(); in promoteLoadFromStore() 1045 BuildMI(*LoadI->getParent(), LoadI, LoadI->getDebugLoc(), in promoteLoadFromStore() 1050 .setMIFlags(LoadI->getFlags()); in promoteLoadFromStore() [all …]
|
D | AArch64FalkorHWPFFix.cpp | 156 LoadInst *LoadI = dyn_cast<LoadInst>(&I); in runOnLoop() local 157 if (!LoadI) in runOnLoop() 160 Value *PtrValue = LoadI->getPointerOperand(); in runOnLoop() 169 LoadI->setMetadata(FALKOR_STRIDED_ACCESS_MD, in runOnLoop() 170 MDNode::get(LoadI->getContext(), {})); in runOnLoop()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 155 promoteLoadFromStore(MachineBasicBlock::iterator LoadI, 1009 AArch64LoadStoreOpt::promoteLoadFromStore(MachineBasicBlock::iterator LoadI, in promoteLoadFromStore() argument 1011 MachineBasicBlock::iterator NextI = LoadI; in promoteLoadFromStore() 1014 int LoadSize = TII->getMemScale(*LoadI); in promoteLoadFromStore() 1016 Register LdRt = getLdStRegOp(*LoadI).getReg(); in promoteLoadFromStore() 1031 LoadI->getIterator())) { in promoteLoadFromStore() 1038 LLVM_DEBUG(LoadI->print(dbgs())); in promoteLoadFromStore() 1040 LoadI->eraseFromParent(); in promoteLoadFromStore() 1045 BuildMI(*LoadI->getParent(), LoadI, LoadI->getDebugLoc(), in promoteLoadFromStore() 1050 .setMIFlags(LoadI->getFlags()); in promoteLoadFromStore() [all …]
|
D | AArch64FalkorHWPFFix.cpp | 156 LoadInst *LoadI = dyn_cast<LoadInst>(&I); in runOnLoop() local 157 if (!LoadI) in runOnLoop() 160 Value *PtrValue = LoadI->getPointerOperand(); in runOnLoop() 169 LoadI->setMetadata(FALKOR_STRIDED_ACCESS_MD, in runOnLoop() 170 MDNode::get(LoadI->getContext(), {})); in runOnLoop()
|
/external/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 128 promoteLoadFromStore(MachineBasicBlock::iterator LoadI, 928 AArch64LoadStoreOpt::promoteLoadFromStore(MachineBasicBlock::iterator LoadI, in promoteLoadFromStore() argument 930 MachineBasicBlock::iterator NextI = LoadI; in promoteLoadFromStore() 933 int LoadSize = getMemScale(*LoadI); in promoteLoadFromStore() 935 unsigned LdRt = getLdStRegOp(*LoadI).getReg(); in promoteLoadFromStore() 949 DEBUG(LoadI->print(dbgs())); in promoteLoadFromStore() 951 LoadI->eraseFromParent(); in promoteLoadFromStore() 956 BuildMI(*LoadI->getParent(), LoadI, LoadI->getDebugLoc(), in promoteLoadFromStore() 966 bool IsUnscaled = TII->isUnscaledLdSt(*LoadI); in promoteLoadFromStore() 971 ? getLdStOffsetOp(*LoadI).getImm() in promoteLoadFromStore() [all …]
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/InstCombine/ |
D | InstCombinePHI.cpp | 110 if (LoadInst *LoadI = dyn_cast<LoadInst>(U)) { in FoldIntegerTypedPHI() local 111 Ptr = LoadI->getPointerOperand(); in FoldIntegerTypedPHI() 165 auto *LoadI = dyn_cast<LoadInst>(Arg); in FoldIntegerTypedPHI() local 166 if (!LoadI) in FoldIntegerTypedPHI() 169 if (!LoadI->hasOneUse()) in FoldIntegerTypedPHI() 175 AvailablePtrVals.emplace_back(LoadI); in FoldIntegerTypedPHI() 254 LoadInst *LoadI = dyn_cast<LoadInst>(IncomingVal); in FoldIntegerTypedPHI() local 257 (LoadI && LoadI->hasOneUse())) && in FoldIntegerTypedPHI()
|
/external/llvm-project/llvm/lib/Transforms/InstCombine/ |
D | InstCombinePHI.cpp | 119 if (LoadInst *LoadI = dyn_cast<LoadInst>(U)) { in foldIntegerTypedPHI() local 120 Ptr = LoadI->getPointerOperand(); in foldIntegerTypedPHI() 174 auto *LoadI = dyn_cast<LoadInst>(Arg); in foldIntegerTypedPHI() local 175 if (!LoadI) in foldIntegerTypedPHI() 178 if (!LoadI->hasOneUse()) in foldIntegerTypedPHI() 184 AvailablePtrVals.emplace_back(LoadI); in foldIntegerTypedPHI() 263 LoadInst *LoadI = dyn_cast<LoadInst>(IncomingVal); in foldIntegerTypedPHI() local 266 (LoadI && LoadI->hasOneUse())) && in foldIntegerTypedPHI()
|
/external/llvm-project/polly/lib/Exchange/ |
D | JSONExporter.cpp | 449 if (LoadInst *LoadI = dyn_cast<LoadInst>(MA->getAccessInstruction())) { in importAccesses() local 451 LoadI->getAlignment() && in importAccesses() 452 DL.getABITypeAlignment(LoadI->getType()) != LoadI->getAlignment(); in importAccesses()
|
/external/llvm-project/llvm/lib/Target/Hexagon/ |
D | HexagonVLIWPacketizer.cpp | 1524 bool LoadI = I.mayLoad(), StoreI = I.mayStore(); in isLegalToPacketizeTogether() local 1532 (StoreJ && LoadI && !NVStoreJ)) && in isLegalToPacketizeTogether() 1540 if (StoreJ && LoadI && alias(J, I)) { in isLegalToPacketizeTogether() 1546 if (!LoadJ || (!LoadI && !StoreI)) { in isLegalToPacketizeTogether()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Hexagon/ |
D | HexagonVLIWPacketizer.cpp | 1517 bool LoadI = I.mayLoad(), StoreI = I.mayStore(); in isLegalToPacketizeTogether() local 1525 (StoreJ && LoadI && !NVStoreJ)) && in isLegalToPacketizeTogether() 1533 if (StoreJ && LoadI && alias(J, I)) { in isLegalToPacketizeTogether() 1539 if (!LoadJ || (!LoadI && !StoreI)) { in isLegalToPacketizeTogether()
|
/external/llvm/lib/Target/Hexagon/ |
D | HexagonVLIWPacketizer.cpp | 1359 bool LoadI = I->mayLoad(), StoreI = I->mayStore(); in isLegalToPacketizeTogether() local 1363 if (LoadI) { in isLegalToPacketizeTogether() 1367 } else if (!LoadJ || (!LoadI && !StoreI)) { in isLegalToPacketizeTogether()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/SystemZ/ |
D | SystemZISelLowering.cpp | 894 if (auto *LoadI = dyn_cast<LoadInst>(StoreI->getValueOperand())) in supportedAddressingMode() local 895 if (LoadI->hasOneUse() && LoadI->getParent() == I->getParent()) in supportedAddressingMode() 897 return getLoadStoreAddrMode(HasVector, LoadI->getType()); in supportedAddressingMode()
|
/external/llvm-project/llvm/lib/Target/SystemZ/ |
D | SystemZISelLowering.cpp | 917 if (auto *LoadI = dyn_cast<LoadInst>(StoreI->getValueOperand())) in supportedAddressingMode() local 918 if (LoadI->hasOneUse() && LoadI->getParent() == I->getParent()) in supportedAddressingMode() 920 return getLoadStoreAddrMode(HasVector, LoadI->getType()); in supportedAddressingMode()
|