/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
D | AArch64StackTaggingPreRA.cpp | 176 MachineInstr *UseI = &*(UI++); in uncheckUsesOf() local 177 if (isUncheckedLoadOrStoreOpcode(UseI->getOpcode())) { in uncheckUsesOf() 179 unsigned OpIdx = TII->getLoadStoreImmIdx(UseI->getOpcode()) - 1; in uncheckUsesOf() 180 if (UseI->getOperand(OpIdx).isReg() && in uncheckUsesOf() 181 UseI->getOperand(OpIdx).getReg() == TaggedReg) { in uncheckUsesOf() 182 UseI->getOperand(OpIdx).ChangeToFrameIndex(FI); in uncheckUsesOf() 183 UseI->getOperand(OpIdx).setTargetFlags(AArch64II::MO_TAGGED); in uncheckUsesOf() 185 } else if (UseI->isCopy() && in uncheckUsesOf() 186 Register::isVirtualRegister(UseI->getOperand(0).getReg())) { in uncheckUsesOf() 187 uncheckUsesOf(UseI->getOperand(0).getReg(), FI); in uncheckUsesOf()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ |
D | MIRCanonicalizerPass.cpp | 243 MachineBasicBlock::iterator UseI = BBE; in rescheduleCanonically() local 247 if (DefI != BBE && UseI != BBE) in rescheduleCanonically() 256 UseI = BBI; in rescheduleCanonically() 261 if (DefI == BBE || UseI == BBE) in rescheduleCanonically() 268 UseI->dump(); in rescheduleCanonically() 273 MBB->splice(UseI, MBB, DefI); in rescheduleCanonically() 279 auto UseI = in rescheduleCanonically() local 283 if (UseI == MBB->instr_end()) in rescheduleCanonically() 290 [&]() -> MachineBasicBlock::iterator { return UseI; }); in rescheduleCanonically()
|
D | MachineCopyPropagation.cpp | 284 const MachineInstr &UseI, unsigned UseIdx); 286 const MachineInstr &UseI, 386 const MachineInstr &Copy, const MachineInstr &UseI, unsigned UseIdx) { in isBackwardPropagatableRegClassCopy() argument 390 UseI.getRegClassConstraint(UseIdx, TII, TRI)) in isBackwardPropagatableRegClassCopy() 402 const MachineInstr &UseI, in isForwardableRegClassCopy() argument 410 UseI.getRegClassConstraint(UseIdx, TII, TRI)) in isForwardableRegClassCopy() 413 if (!UseI.isCopy()) in isForwardableRegClassCopy() 433 TRI->getMinimalPhysRegClass(UseI.getOperand(0).getReg()); in isForwardableRegClassCopy()
|
D | SplitKit.cpp | 220 SmallVectorImpl<SlotIndex>::const_iterator UseI, UseE; in calcLiveBlockInfo() local 221 UseI = UseSlots.begin(); in calcLiveBlockInfo() 236 if (UseI == UseE || *UseI >= Stop) { in calcLiveBlockInfo() 245 BI.FirstInstr = *UseI; in calcLiveBlockInfo() 247 do ++UseI; in calcLiveBlockInfo() 248 while (UseI != UseE && *UseI < Stop); in calcLiveBlockInfo() 249 BI.LastInstr = UseI[-1]; in calcLiveBlockInfo()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
D | R600EmitClauseMarkers.cpp | 210 for (MachineBasicBlock::iterator UseI = Def; UseI != BBEnd; ++UseI) { in canClauseLocalKillFitInClause() local 211 AluInstCount += OccupiedDwords(*UseI); in canClauseLocalKillFitInClause() 213 if (!SubstituteKCacheBank(*UseI, KCacheBanks, false)) in canClauseLocalKillFitInClause() 228 if (UseI->readsRegister(MOI->getReg(), &TRI)) in canClauseLocalKillFitInClause() 232 if (UseI != Def && UseI->killsRegister(MOI->getReg(), &TRI)) in canClauseLocalKillFitInClause()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Hexagon/ |
D | HexagonSplitDouble.cpp | 253 MachineInstr *UseI = Op.getParent(); in partitionRegisters() local 254 if (isFixedInstr(UseI)) in partitionRegisters() 256 for (unsigned i = 0, n = UseI->getNumOperands(); i < n; ++i) { in partitionRegisters() 257 MachineOperand &MO = UseI->getOperand(i); in partitionRegisters() 438 MachineInstr *UseI = U->getParent(); in isProfitable() local 439 if (isFixedInstr(UseI)) { in isProfitable() 442 for (auto &Op : UseI->operands()) { in isProfitable() 453 if (UseI->isPHI()) { in isProfitable() 454 const MachineBasicBlock *PB = UseI->getParent(); in isProfitable() 460 int32_t P = profit(UseI); in isProfitable() [all …]
|
D | BitTracker.cpp | 985 for (MachineInstr &UseI : MRI.use_nodbg_instructions(Reg)) in visitUsesOf() 986 UseQ.push(&UseI); in visitUsesOf() 1104 MachineInstr &UseI = *UseQ.front(); in runUseQueue() local 1107 if (!InstrExec.count(&UseI)) in runUseQueue() 1109 if (UseI.isPHI()) in runUseQueue() 1110 visitPHI(UseI); in runUseQueue() 1111 else if (!UseI.isBranch()) in runUseQueue() 1112 visitNonBranch(UseI); in runUseQueue() 1114 visitBranchesFrom(UseI); in runUseQueue()
|
D | HexagonBitSimplify.cpp | 975 MachineInstr *UseI = I->getParent(); in isDead() local 976 if (UseI->isDebugValue()) in isDead() 978 if (UseI->isPHI()) { in isDead() 979 assert(!UseI->getOperand(0).getSubReg()); in isDead() 980 Register DR = UseI->getOperand(0).getReg(); in isDead() 1219 MachineInstr &UseI = *I->getParent(); in computeUsedBits() local 1220 if (UseI.isPHI() || UseI.isCopy()) { in computeUsedBits() 1221 Register DefR = UseI.getOperand(0).getReg(); in computeUsedBits() 1226 if (!computeUsedBits(UseI, I.getOperandNo(), Used, B)) in computeUsedBits() 3117 MachineInstr *UseI = UI->getParent(); in processLoop() local [all …]
|
D | HexagonLoopIdiomRecognition.cpp | 583 bool classifyInst(Instruction *UseI, ValueSeq &Early, ValueSeq &Late); 1183 bool PolynomialMultiplyRecognize::classifyInst(Instruction *UseI, in classifyInst() argument 1188 if (UseI->getOpcode() == Instruction::Select) { in classifyInst() 1189 Value *TV = UseI->getOperand(1), *FV = UseI->getOperand(2); in classifyInst() 1193 Early.insert(UseI); in classifyInst() 1197 Late.insert(UseI); in classifyInst() 1204 if (UseI->getNumOperands() == 0) in classifyInst() 1208 for (auto &I : UseI->operands()) { in classifyInst() 1229 Early.insert(UseI); in classifyInst() 1231 Late.insert(UseI); in classifyInst() [all …]
|
D | HexagonGenPredicate.cpp | 239 MachineInstr *UseI = I->getParent(); in processPredicateGPR() local 240 if (isConvertibleToPredForm(UseI)) in processPredicateGPR() 241 PUsers.insert(UseI); in processPredicateGPR()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/ |
D | LoopRotationUtils.cpp | 574 for (User *UseI : IVOpnd->users()) { in shouldSpeculateInstrs() 575 auto *UserInst = cast<Instruction>(UseI); in shouldSpeculateInstrs()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
D | X86CmovConversion.cpp | 323 [&](MachineInstr &UseI) { in collectCmovCandidates() argument 324 return UseI.getOpcode() == X86::SUBREG_TO_REG; in collectCmovCandidates()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/ |
D | LazyValueInfo.cpp | 2033 if (auto *UseI = dyn_cast<Instruction>(U)) in emitInstructionAnnot() local 2034 if (!isa<PHINode>(UseI) || DT.dominates(ParentBB, UseI->getParent())) in emitInstructionAnnot() 2035 printResult(UseI->getParent()); in emitInstructionAnnot()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/ |
D | IndVarSimplify.cpp | 1007 Instruction *UseI) { in getPostIncRangeInfo() argument 1008 DefUserPair Key(Def, UseI); in getPostIncRangeInfo() 1018 void updatePostIncRangeInfo(Value *Def, Instruction *UseI, ConstantRange R) { in updatePostIncRangeInfo() argument 1019 DefUserPair Key(Def, UseI); in updatePostIncRangeInfo()
|
D | LoopStrengthReduce.cpp | 3112 auto UseI = find(Inc.UserInst->operands(), Inc.IVOperand); in FinalizeChain() local 3113 assert(UseI != Inc.UserInst->op_end() && "cannot find IV operand"); in FinalizeChain() 3114 IVIncSet.insert(UseI); in FinalizeChain() 3248 User::op_iterator UseI = in CollectFixupsAndInitialFormulae() local 3250 assert(UseI != UserInst->op_end() && "cannot find IV operand"); in CollectFixupsAndInitialFormulae() 3251 if (IVIncSet.count(UseI)) { in CollectFixupsAndInitialFormulae() 3252 LLVM_DEBUG(dbgs() << "Use is in profitable chain: " << **UseI << '\n'); in CollectFixupsAndInitialFormulae()
|