Home
last modified time | relevance | path

Searched refs:AddrNumOperands (Results 1 – 25 of 27) sorted by relevance

12

/external/swiftshader/third_party/LLVM/lib/Target/X86/
DX86CodeEmitter.cpp211 unsigned e = (isTwoAddr ? X86::AddrNumOperands+1 : X86::AddrNumOperands); in determineREX()
866 X86_MC::getX86RegNum(MI.getOperand(CurOp + X86::AddrNumOperands) in emitInstruction()
868 CurOp += X86::AddrNumOperands + 1; in emitInstruction()
886 int AddrOperands = X86::AddrNumOperands; in emitInstruction()
940 intptr_t PCAdj = (CurOp + X86::AddrNumOperands != NumOps) ? in emitInstruction()
941 (MI.getOperand(CurOp+X86::AddrNumOperands).isImm() ? in emitInstruction()
947 CurOp += X86::AddrNumOperands; in emitInstruction()
DX86InstrInfo.cpp1072 if (MI->getOperand(X86::AddrNumOperands).getSubReg() == 0 && in isStoreToStackSlot()
1074 return MI->getOperand(X86::AddrNumOperands).getReg(); in isStoreToStackSlot()
2755 SmallVector<MachineOperand,X86::AddrNumOperands> MOs; in foldMemoryOperandImpl()
2814 for (unsigned i = NumOps - X86::AddrNumOperands; i != NumOps; ++i) in foldMemoryOperandImpl()
2907 SmallVector<MachineOperand, X86::AddrNumOperands> AddrOps; in unfoldMemoryOperand()
2913 if (i >= Index && i < Index + X86::AddrNumOperands) in unfoldMemoryOperand()
2932 for (unsigned i = 1; i != 1 + X86::AddrNumOperands; ++i) { in unfoldMemoryOperand()
3029 if (i >= Index-NumDefs && i < Index-NumDefs + X86::AddrNumOperands) in unfoldMemoryOperand()
DX86MCInstLower.cpp353 assert(OutMI.getNumOperands() == 1+X86::AddrNumOperands && in Lower()
DX86ISelLowering.cpp10944 assert(bInstr->getNumOperands() < X86::AddrNumOperands + 4 && in EmitAtomicBitwiseWithCustomInserter()
10948 MachineOperand* argOpers[2 + X86::AddrNumOperands]; in EmitAtomicBitwiseWithCustomInserter()
10954 int lastAddrIndx = X86::AddrNumOperands - 1; // [0,3] in EmitAtomicBitwiseWithCustomInserter()
11057 assert(bInstr->getNumOperands() < X86::AddrNumOperands + 14 && in EmitAtomicBit6432WithCustomInserter()
11061 MachineOperand* argOpers[2 + X86::AddrNumOperands]; in EmitAtomicBit6432WithCustomInserter()
11062 for (int i=0; i < 2 + X86::AddrNumOperands; ++i) { in EmitAtomicBit6432WithCustomInserter()
11072 int lastAddrIndx = X86::AddrNumOperands - 1; // [0,3] in EmitAtomicBit6432WithCustomInserter()
11211 assert(mInstr->getNumOperands() < X86::AddrNumOperands + 4 && in EmitAtomicMinMaxWithCustomInserter()
11214 MachineOperand* argOpers[2 + X86::AddrNumOperands]; in EmitAtomicMinMaxWithCustomInserter()
11220 int lastAddrIndx = X86::AddrNumOperands - 1; // [0,3] in EmitAtomicMinMaxWithCustomInserter()
[all …]
DX86FloatingPoint.cpp985 assert((NumOps == X86::AddrNumOperands + 1 || NumOps == 1) && in handleOneArgFP()
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/X86/
DX86CallFrameOptimization.cpp289 MachineOperand ImmOp = MI->getOperand(X86::AddrNumOperands); in classifyInstruction()
295 MachineOperand ImmOp = MI->getOperand(X86::AddrNumOperands); in classifyInstruction()
507 MachineOperand PushOp = Store->getOperand(X86::AddrNumOperands); in adjustCallSequence()
561 for (unsigned i = NumOps - X86::AddrNumOperands; i != NumOps; ++i) in adjustCallSequence()
DX86DomainReassignment.cpp545 MemOpEnd = MemOpStart + X86::AddrNumOperands; in usedAsAddr()
579 OpIdx += (X86::AddrNumOperands - 1); in buildClosure()
DX86AvoidStoreForwardingBlocks.cpp428 MachineOperand &StoreSrcVReg = StoreInst->getOperand(X86::AddrNumOperands); in buildCopy()
430 NewStore->getOperand(X86::AddrNumOperands).setIsKill(StoreSrcVReg.isKill()); in buildCopy()
DX86InstrInfo.h163 return Op + X86::AddrNumOperands <= MI.getNumOperands() && in isMem()
DX86InstrInfo.cpp430 if (MI.getOperand(X86::AddrNumOperands).getSubReg() == 0 && in isStoreToStackSlot()
432 return MI.getOperand(X86::AddrNumOperands).getReg(); in isStoreToStackSlot()
4128 unsigned SrcReg = MIB->getOperand(X86::AddrNumOperands).getReg(); in expandNOVLXStore()
4138 MIB->getOperand(X86::AddrNumOperands).setReg(SrcReg); in expandNOVLXStore()
4864 if (MOs.size() == X86::AddrNumOperands && in foldMemoryOperandImpl()
5272 SmallVector<MachineOperand,X86::AddrNumOperands> MOs; in foldMemoryOperandImpl()
5347 MOs.append(LoadMI.operands_begin() + NumOps - X86::AddrNumOperands, in foldMemoryOperandImpl()
5382 SmallVector<MachineOperand, X86::AddrNumOperands> AddrOps; in unfoldMemoryOperand()
5388 if (i >= Index && i < Index + X86::AddrNumOperands) in unfoldMemoryOperand()
5405 for (unsigned i = 1; i != 1 + X86::AddrNumOperands; ++i) { in unfoldMemoryOperand()
[all …]
DX86FlagsCopyLowering.cpp1051 for (int i = 0; i < X86::AddrNumOperands; ++i) in rewriteSetCC()
DX86FloatingPoint.cpp1109 assert((NumOps == X86::AddrNumOperands + 1 || NumOps == 1) && in handleOneArgFP()
DX86MCInstLower.cpp433 assert(OutMI.getNumOperands() == 1 + X86::AddrNumOperands && in Lower()
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/X86/MCTargetDesc/
DX86MCCodeEmitter.cpp757 CurOp += X86::AddrNumOperands; in EmitVEXOpcodePrefix()
815 VEX_4V = ~getX86RegEncoding(MI, CurOp + X86::AddrNumOperands) & 0xf; in EmitVEXOpcodePrefix()
1070 CurOp += X86::AddrNumOperands; in DetermineREXPrefix()
1080 CurOp += X86::AddrNumOperands; in DetermineREXPrefix()
1380 unsigned SrcRegNum = CurOp + X86::AddrNumOperands; in encodeInstruction()
1452 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction()
1464 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction()
1481 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction()
1513 CurOp += X86::AddrNumOperands; in encodeInstruction()
DX86BaseInfo.h42 AddrNumOperands = 5 enumerator
/external/swiftshader/third_party/LLVM/lib/Target/X86/MCTargetDesc/
DX86MCCodeEmitter.cpp510 CurOp = X86::AddrNumOperands; in EmitVEXOpcodePrefix()
683 unsigned e = (isTwoAddr ? X86::AddrNumOperands+1 : X86::AddrNumOperands); in DetermineREXPrefix()
925 SrcRegNum = CurOp + X86::AddrNumOperands; in EncodeInstruction()
949 int AddrOperands = X86::AddrNumOperands; in EncodeInstruction()
982 CurOp += X86::AddrNumOperands; in EncodeInstruction()
DX86BaseInfo.h41 AddrNumOperands = 5 enumerator
/external/llvm/lib/Target/X86/MCTargetDesc/
DX86MCCodeEmitter.cpp727 CurOp += X86::AddrNumOperands; in EmitVEXOpcodePrefix()
779 VEX_4V = ~getX86RegEncoding(MI, CurOp + X86::AddrNumOperands) & 0xf; in EmitVEXOpcodePrefix()
998 CurOp += X86::AddrNumOperands; in DetermineREXPrefix()
1008 CurOp += X86::AddrNumOperands; in DetermineREXPrefix()
1292 unsigned SrcRegNum = CurOp + X86::AddrNumOperands; in encodeInstruction()
1346 CurOp = FirstMemOp + X86::AddrNumOperands; in encodeInstruction()
1383 CurOp += X86::AddrNumOperands; in encodeInstruction()
DX86BaseInfo.h42 AddrNumOperands = 5 enumerator
/external/llvm/lib/Target/X86/
DX86CallFrameOptimization.cpp467 MachineOperand PushOp = MOV->getOperand(X86::AddrNumOperands); in adjustCallSequence()
516 for (unsigned i = NumOps - X86::AddrNumOperands; i != NumOps; ++i) in adjustCallSequence()
DX86InstrInfo.h137 return Op + X86::AddrNumOperands <= MI.getNumOperands() && in isMem()
DX86InstrInfo.cpp2275 if (MI.getOperand(X86::AddrNumOperands).getSubReg() == 0 && in isStoreToStackSlot()
2277 return MI.getOperand(X86::AddrNumOperands).getReg(); in isStoreToStackSlot()
6287 SmallVector<MachineOperand,X86::AddrNumOperands> MOs; in foldMemoryOperandImpl()
6355 MOs.append(LoadMI.operands_begin() + NumOps - X86::AddrNumOperands, in foldMemoryOperandImpl()
6390 SmallVector<MachineOperand, X86::AddrNumOperands> AddrOps; in unfoldMemoryOperand()
6396 if (i >= Index && i < Index + X86::AddrNumOperands) in unfoldMemoryOperand()
6413 for (unsigned i = 1; i != 1 + X86::AddrNumOperands; ++i) { in unfoldMemoryOperand()
6507 if (i >= Index-NumDefs && i < Index-NumDefs + X86::AddrNumOperands) in unfoldMemoryOperand()
DX86FloatingPoint.cpp1061 assert((NumOps == X86::AddrNumOperands + 1 || NumOps == 1) && in handleOneArgFP()
DX86MCInstLower.cpp404 assert(OutMI.getNumOperands() == 1+X86::AddrNumOperands && in Lower()
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/X86/AsmParser/
DX86AsmParser.cpp2825 X86::AddrNumOperands - 1).getReg(); in validateInstruction()

12