| /external/llvm/lib/CodeGen/ | 
| D | ImplicitNullChecks.cpp | 485 ImplicitNullChecks::insertFaultingLoad(MachineInstr *LoadMI,  in insertFaultingLoad()
 | 
| D | TargetInstrInfo.cpp | 780                                                  MachineInstr &LoadMI,  in foldMemoryOperand()
 | 
| D | InlineSpiller.cpp | 728                   MachineInstr *LoadMI) {  in foldMemoryOperand()
 | 
| /external/llvm/lib/Target/X86/ | 
| D | X86MCInstLower.cpp | 908   MCInst LoadMI;  in LowerFAULTING_LOAD_OP()  local
 | 
| D | X86InstrInfo.cpp | 6165 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI,  in isNonFoldablePartialRegisterLoad()6220     MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,  in foldMemoryOperandImpl()
 
 | 
| D | X86FastISel.cpp | 778         MachineInstrBuilder LoadMI =  in handleConstantAddresses()  local
 | 
| /external/llvm/lib/Target/AArch64/ | 
| D | AArch64LoadStoreOptimizer.cpp | 1103   MachineInstr &LoadMI = *I;  in findMatchingStore()  local
 | 
| D | AArch64FastISel.cpp | 4371   const auto *LoadMI = MI;  in optimizeIntExtLoad()  local
 | 
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ | 
| D | PPCMIPeephole.cpp | 370             MachineInstr *LoadMI = MRI->getVRegDef(FeedReg1);  in simplifyCode()  local
 | 
| /external/llvm/lib/Target/SystemZ/ | 
| D | SystemZInstrInfo.cpp | 1029     MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,  in foldMemoryOperandImpl()
 | 
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ | 
| D | TargetInstrInfo.cpp | 622                                                  MachineInstr &LoadMI,  in foldMemoryOperand()
 | 
| D | InlineSpiller.cpp | 789                   MachineInstr *LoadMI) {  in foldMemoryOperand()
 | 
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ | 
| D | AArch64LoadStoreOptimizer.cpp | 1175   MachineInstr &LoadMI = *I;  in findMatchingStore()  local
 | 
| D | AArch64InstructionSelector.cpp | 2181       auto *LoadMI = getOpcodeDef(TargetOpcode::G_LOAD, SrcReg, MRI);  in select()  local3180   MachineInstr *LoadMI = nullptr;  in emitLoadFromConstantPool()  local
 
 | 
| D | AArch64InstrInfo.cpp | 3403         MachineInstr &LoadMI = *--InsertPt;  in foldMemoryOperandImpl()  local
 | 
| D | AArch64FastISel.cpp | 4549   const auto *LoadMI = MI;  in optimizeIntExtLoad()  local
 | 
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/SystemZ/ | 
| D | SystemZInstrInfo.cpp | 1220     MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,  in foldMemoryOperandImpl()
 | 
| /external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ | 
| D | X86InstrInfo.cpp | 5070 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI,  in isNonFoldablePartialRegisterLoad()5193     MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,  in foldMemoryOperandImpl()
 
 | 
| D | X86FastISel.cpp | 792         MachineInstrBuilder LoadMI =  in handleConstantAddresses()  local
 |