/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
D | X86CallFrameOptimization.cpp | 291 MachineOperand ImmOp = MI->getOperand(X86::AddrNumOperands); in classifyInstruction() 297 MachineOperand ImmOp = MI->getOperand(X86::AddrNumOperands); in classifyInstruction() 509 MachineOperand PushOp = Store->getOperand(X86::AddrNumOperands); in adjustCallSequence() 563 for (unsigned i = NumOps - X86::AddrNumOperands; i != NumOps; ++i) in adjustCallSequence()
|
D | X86ExpandPseudo.cpp | 197 MachineOperand &StackAdjust = MBBI->getOperand(isMem ? X86::AddrNumOperands in ExpandMI() 263 for (unsigned i = 0; i != X86::AddrNumOperands; ++i) in ExpandMI()
|
D | X86MCInstLower.cpp | 479 assert(OutMI.getNumOperands() == 1 + X86::AddrNumOperands && in Lower() 768 assert(OutMI.getNumOperands() == X86::AddrNumOperands && in Lower() 2061 for (int i = 0; i < X86::AddrNumOperands; ++i) { in EmitInstruction() 2070 for (int i = 0; i < X86::AddrNumOperands; ++i) { in EmitInstruction() 2085 Register Reg = MI->getOperand(X86::AddrNumOperands).getReg(); in EmitInstruction() 2091 for (int i = 0; i < X86::AddrNumOperands; ++i) in EmitInstruction() 2098 for (int i = 0; i < X86::AddrNumOperands; ++i) { in EmitInstruction()
|
D | X86DomainReassignment.cpp | 538 MemOpEnd = MemOpStart + X86::AddrNumOperands; in usedAsAddr() 573 OpIdx += (X86::AddrNumOperands - 1); in buildClosure()
|
D | X86AvoidStoreForwardingBlocks.cpp | 430 MachineOperand &StoreSrcVReg = StoreInst->getOperand(X86::AddrNumOperands); in buildCopy() 432 NewStore->getOperand(X86::AddrNumOperands).setIsKill(StoreSrcVReg.isKill()); in buildCopy()
|
D | X86InstrInfo.h | 128 return Op + X86::AddrNumOperands <= MI.getNumOperands() && in isMem()
|
D | X86InstrInfo.cpp | 440 if (MI.getOperand(X86::AddrNumOperands).getSubReg() == 0 && in isStoreToStackSlot() 442 return MI.getOperand(X86::AddrNumOperands).getReg(); in isStoreToStackSlot() 4043 Register SrcReg = MIB->getOperand(X86::AddrNumOperands).getReg(); in expandNOVLXStore() 4053 MIB->getOperand(X86::AddrNumOperands).setReg(SrcReg); in expandNOVLXStore() 4857 if (MOs.size() == X86::AddrNumOperands && in foldMemoryOperandImpl() 5276 SmallVector<MachineOperand,X86::AddrNumOperands> MOs; in foldMemoryOperandImpl() 5355 MOs.append(LoadMI.operands_begin() + NumOps - X86::AddrNumOperands, in foldMemoryOperandImpl() 5480 SmallVector<MachineOperand, X86::AddrNumOperands> AddrOps; in unfoldMemoryOperand() 5486 if (i >= Index && i < Index + X86::AddrNumOperands) in unfoldMemoryOperand() 5518 for (unsigned i = 1; i != 1 + X86::AddrNumOperands; ++i) { in unfoldMemoryOperand() [all …]
|
D | X86FlagsCopyLowering.cpp | 1114 for (int i = 0; i < X86::AddrNumOperands; ++i) in rewriteSetCC()
|
D | X86FloatingPoint.cpp | 1122 assert((NumOps == X86::AddrNumOperands + 1 || NumOps == 1) && in handleOneArgFP()
|
D | X86ISelLowering.cpp | 30328 static_assert(X86::AddrNumOperands == 5, in EmitVAARG64WithCustomInserter() 31516 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitSetJmpShadowStackFix() 31624 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjSetJmp() 31777 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitLongJmpShadowStackFix() 31903 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjLongJmp() 31915 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjLongJmp() 31929 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjLongJmp() 32368 .addReg(MI.getOperand(X86::AddrNumOperands).getReg()); in EmitInstrWithCustomInserter()
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/MCTargetDesc/ |
D | X86MCCodeEmitter.cpp | 887 CurOp += X86::AddrNumOperands; in emitVEXOpcodePrefix() 947 VEX_4V = ~getX86RegEncoding(MI, CurOp + X86::AddrNumOperands) & 0xf; in emitVEXOpcodePrefix() 1212 CurOp += X86::AddrNumOperands; in determineREXPrefix() 1221 CurOp += X86::AddrNumOperands; in determineREXPrefix() 1490 unsigned SrcRegNum = CurOp + X86::AddrNumOperands; in encodeInstruction() 1573 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction() 1585 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction() 1602 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction() 1608 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction() 1648 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction() [all …]
|
D | X86BaseInfo.h | 41 AddrNumOperands = 5 enumerator
|
/external/llvm/lib/Target/X86/MCTargetDesc/ |
D | X86MCCodeEmitter.cpp | 727 CurOp += X86::AddrNumOperands; in EmitVEXOpcodePrefix() 779 VEX_4V = ~getX86RegEncoding(MI, CurOp + X86::AddrNumOperands) & 0xf; in EmitVEXOpcodePrefix() 998 CurOp += X86::AddrNumOperands; in DetermineREXPrefix() 1008 CurOp += X86::AddrNumOperands; in DetermineREXPrefix() 1292 unsigned SrcRegNum = CurOp + X86::AddrNumOperands; in encodeInstruction() 1346 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction() 1383 CurOp += X86::AddrNumOperands; in encodeInstruction()
|
D | X86BaseInfo.h | 42 AddrNumOperands = 5 enumerator
|
/external/llvm/lib/Target/X86/ |
D | X86CallFrameOptimization.cpp | 467 MachineOperand PushOp = MOV->getOperand(X86::AddrNumOperands); in adjustCallSequence() 516 for (unsigned i = NumOps - X86::AddrNumOperands; i != NumOps; ++i) in adjustCallSequence()
|
D | X86InstrInfo.h | 137 return Op + X86::AddrNumOperands <= MI.getNumOperands() && in isMem()
|
D | X86InstrInfo.cpp | 2275 if (MI.getOperand(X86::AddrNumOperands).getSubReg() == 0 && in isStoreToStackSlot() 2277 return MI.getOperand(X86::AddrNumOperands).getReg(); in isStoreToStackSlot() 6287 SmallVector<MachineOperand,X86::AddrNumOperands> MOs; in foldMemoryOperandImpl() 6355 MOs.append(LoadMI.operands_begin() + NumOps - X86::AddrNumOperands, in foldMemoryOperandImpl() 6390 SmallVector<MachineOperand, X86::AddrNumOperands> AddrOps; in unfoldMemoryOperand() 6396 if (i >= Index && i < Index + X86::AddrNumOperands) in unfoldMemoryOperand() 6413 for (unsigned i = 1; i != 1 + X86::AddrNumOperands; ++i) { in unfoldMemoryOperand() 6507 if (i >= Index-NumDefs && i < Index-NumDefs + X86::AddrNumOperands) in unfoldMemoryOperand()
|
D | X86FloatingPoint.cpp | 1061 assert((NumOps == X86::AddrNumOperands + 1 || NumOps == 1) && in handleOneArgFP()
|
D | X86MCInstLower.cpp | 404 assert(OutMI.getNumOperands() == 1+X86::AddrNumOperands && in Lower()
|
D | X86ISelLowering.cpp | 22717 for (int i = 0; i < X86::AddrNumOperands; ++i) in emitMonitor() 22720 unsigned ValOps = X86::AddrNumOperands; in emitMonitor() 22747 static_assert(X86::AddrNumOperands == 5, in EmitVAARG64WithCustomInserter() 23431 unsigned ValOpIdx = X86::AddrNumOperands; in EmitLoweredAtomicFP() 23437 for (int i = 0; i < X86::AddrNumOperands; ++i) { in EmitLoweredAtomicFP() 23447 for (int i = 0; i < X86::AddrNumOperands; ++i) in EmitLoweredAtomicFP() 23824 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjSetJmp() 23910 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) in emitEHSjLjLongJmp() 23915 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjLongJmp() 23924 for (unsigned i = 0; i < X86::AddrNumOperands; ++i) { in emitEHSjLjLongJmp() [all …]
|
/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/AsmParser/ |
D | X86AsmParser.cpp | 3078 X86::AddrNumOperands - 1).getReg(); in validateInstruction()
|