Lines Matching refs:MU
316 const MemoryUseOrDef *MU, in instructionClobbersQuery() argument
322 return instructionClobbersQuery(MD, MemoryLocation(), MU->getMemoryInst(), in instructionClobbersQuery()
324 return instructionClobbersQuery(MD, UseMLOC.getLoc(), MU->getMemoryInst(), in instructionClobbersQuery()
329 bool MemorySSAUtil::defClobbersUseOrDef(MemoryDef *MD, const MemoryUseOrDef *MU, in defClobbersUseOrDef() argument
331 return instructionClobbersQuery(MD, MU, MemoryLocOrCall(MU), AA).IsClobber; in defClobbersUseOrDef()
463 if (const auto *MU = dyn_cast<MemoryUse>(MA)) { in checkClobberSanity() local
464 (void)MU; in checkClobberSanity()
465 assert (MU == Start && in checkClobberSanity()
969 if (auto *MU = dyn_cast<MemoryUse>(Start)) in findClobber() local
970 Current = MU->getDefiningAccess(); in findClobber()
1377 auto *MU = dyn_cast<MemoryUse>(&MA); in optimizeUsesInBlock() local
1378 if (!MU) { in optimizeUsesInBlock()
1384 if (isUseTriviallyOptimizableToLiveOnEntry(*AA, MU->getMemoryInst())) { in optimizeUsesInBlock()
1385 MU->setDefiningAccess(MSSA->getLiveOnEntryDef(), true, None); in optimizeUsesInBlock()
1389 MemoryLocOrCall UseMLOC(MU); in optimizeUsesInBlock()
1440 LLVM_DEBUG(dbgs() << "MemorySSA skipping optimization of " << *MU << " (" in optimizeUsesInBlock()
1441 << *(MU->getMemoryInst()) << ")" in optimizeUsesInBlock()
1456 Walker->getClobberingMemoryAccess(MU, UpwardWalkLimit); in optimizeUsesInBlock()
1476 ClobberAlias CA = instructionClobbersQuery(MD, MU, UseMLOC, *AA); in optimizeUsesInBlock()
1493 MU->setDefiningAccess(VersionStack[UpperBound], true, LocInfo.AR); in optimizeUsesInBlock()
1498 MU->setDefiningAccess(VersionStack[LocInfo.LastKill], true, LocInfo.AR); in optimizeUsesInBlock()