/external/llvm/lib/CodeGen/ |
D | ImplicitNullChecks.cpp | 485 ImplicitNullChecks::insertFaultingLoad(MachineInstr *LoadMI, in insertFaultingLoad()
|
D | TargetInstrInfo.cpp | 780 MachineInstr &LoadMI, in foldMemoryOperand()
|
D | InlineSpiller.cpp | 728 MachineInstr *LoadMI) { in foldMemoryOperand()
|
/external/llvm/lib/Target/X86/ |
D | X86MCInstLower.cpp | 908 MCInst LoadMI; in LowerFAULTING_LOAD_OP() local
|
D | X86InstrInfo.cpp | 6165 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI, in isNonFoldablePartialRegisterLoad() 6220 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl()
|
D | X86FastISel.cpp | 778 MachineInstrBuilder LoadMI = in handleConstantAddresses() local
|
/external/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 1103 MachineInstr &LoadMI = *I; in findMatchingStore() local
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ |
D | PPCMIPeephole.cpp | 370 MachineInstr *LoadMI = MRI->getVRegDef(FeedReg1); in simplifyCode() local
|
/external/llvm-project/llvm/lib/Target/PowerPC/ |
D | PPCMIPeephole.cpp | 531 MachineInstr *LoadMI = MRI->getVRegDef(FeedReg1); in simplifyCode() local
|
/external/llvm/lib/Target/SystemZ/ |
D | SystemZInstrInfo.cpp | 1029 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ |
D | InlineSpiller.cpp | 789 MachineInstr *LoadMI) { in foldMemoryOperand()
|
D | TargetInstrInfo.cpp | 622 MachineInstr &LoadMI, in foldMemoryOperand()
|
/external/llvm-project/llvm/lib/CodeGen/ |
D | InlineSpiller.cpp | 803 MachineInstr *LoadMI) { in foldMemoryOperand()
|
D | TargetInstrInfo.cpp | 657 MachineInstr &LoadMI, in foldMemoryOperand()
|
/external/llvm-project/llvm/lib/Target/AArch64/GISel/ |
D | AArch64InstructionSelector.cpp | 2352 auto *LoadMI = emitLoadFromConstantPool(FPImm, MIB); in select() local 2907 auto *LoadMI = getOpcodeDef(TargetOpcode::G_LOAD, SrcReg, MRI); in select() local 3944 MachineInstr *LoadMI = nullptr; in emitLoadFromConstantPool() local
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 1175 MachineInstr &LoadMI = *I; in findMatchingStore() local
|
D | AArch64InstructionSelector.cpp | 2181 auto *LoadMI = getOpcodeDef(TargetOpcode::G_LOAD, SrcReg, MRI); in select() local 3180 MachineInstr *LoadMI = nullptr; in emitLoadFromConstantPool() local
|
/external/llvm-project/llvm/lib/Target/AArch64/ |
D | AArch64LoadStoreOptimizer.cpp | 1162 MachineInstr &LoadMI = *I; in findMatchingStore() local
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/SystemZ/ |
D | SystemZInstrInfo.cpp | 1220 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl()
|
/external/llvm-project/llvm/lib/Target/SystemZ/ |
D | SystemZInstrInfo.cpp | 1325 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
D | X86InstrInfo.cpp | 5070 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI, in isNonFoldablePartialRegisterLoad() 5193 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl()
|
D | X86FastISel.cpp | 792 MachineInstrBuilder LoadMI = in handleConstantAddresses() local
|
/external/llvm-project/llvm/lib/Target/X86/ |
D | X86InstrInfo.cpp | 5714 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI, in isNonFoldablePartialRegisterLoad() 5945 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl()
|
D | X86FastISel.cpp | 791 MachineInstrBuilder LoadMI = in handleConstantAddresses() local
|
/external/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
D | CombinerHelper.cpp | 603 if (auto *LoadMI = getOpcodeDef(TargetOpcode::G_SEXTLOAD, LoadUser, MRI)) { in matchSextTruncSextLoad() local
|