/external/swiftshader/third_party/subzero/src/ |
D | IceTargetLoweringX86Base.h | 363 void lowerCaseCluster(const CaseCluster &Case, Operand *Src0, bool DoneCmp, 426 Operand *legalizeSrc0ForCmp(Operand *Src0, Operand *Src1); 523 void _adc(Variable *Dest, Operand *Src0) { in _adc() argument 524 AutoMemorySandboxer<> _(this, &Dest, &Src0); in _adc() 525 Context.insert<typename Traits::Insts::Adc>(Dest, Src0); in _adc() 531 void _add(Variable *Dest, Operand *Src0) { in _add() argument 532 AutoMemorySandboxer<> _(this, &Dest, &Src0); in _add() 533 Context.insert<typename Traits::Insts::Add>(Dest, Src0); in _add() 539 void _addps(Variable *Dest, Operand *Src0) { in _addps() argument 540 AutoMemorySandboxer<> _(this, &Dest, &Src0); in _addps() [all …]
|
D | IceTargetLoweringARM32.h | 213 Operand *Src0, Operand *Src1); 253 Operand *Src0, Operand *Src1); 254 CondWhenTrue lowerInt32IcmpCond(InstIcmp::ICond Condition, Operand *Src0, 256 CondWhenTrue lowerInt64IcmpCond(InstIcmp::ICond Condition, Operand *Src0, 258 CondWhenTrue lowerIcmpCond(InstIcmp::ICond Condition, Operand *Src0, 334 void _add(Variable *Dest, Variable *Src0, Operand *Src1, 336 Context.insert<InstARM32Add>(Dest, Src0, Src1, Pred); 338 void _adds(Variable *Dest, Variable *Src0, Operand *Src1, 341 Context.insert<InstARM32Add>(Dest, Src0, Src1, Pred, SetFlags); 346 void _adc(Variable *Dest, Variable *Src0, Operand *Src1, [all …]
|
D | IceTargetLoweringMIPS32.h | 166 void _add(Variable *Dest, Variable *Src0, Variable *Src1) { in _add() argument 167 Context.insert<InstMIPS32Add>(Dest, Src0, Src1); in _add() 170 void _addu(Variable *Dest, Variable *Src0, Variable *Src1) { in _addu() argument 171 Context.insert<InstMIPS32Addu>(Dest, Src0, Src1); in _addu() 174 void _and(Variable *Dest, Variable *Src0, Variable *Src1) { in _and() argument 175 Context.insert<InstMIPS32And>(Dest, Src0, Src1); in _and() 188 void _br(CfgNode *TargetTrue, CfgNode *TargetFalse, Operand *Src0, in _br() argument 190 Context.insert<InstMIPS32Br>(TargetTrue, TargetFalse, Src0, Src1, in _br() 194 void _br(CfgNode *TargetTrue, CfgNode *TargetFalse, Operand *Src0, in _br() argument 196 Context.insert<InstMIPS32Br>(TargetTrue, TargetFalse, Src0, Condition); in _br() [all …]
|
D | IceTargetLoweringX86BaseImpl.h | 817 Operand *&Src0, Operand *&Src1) { 818 if (Src0 == LoadDest && Src1 != LoadDest) { 819 Src0 = LoadSrc; 822 if (Src0 != LoadDest && Src1 == LoadDest) { 868 Operand *Src0 = Arith->getSrc(0); 870 if (canFoldLoadIntoBinaryInst(LoadSrc, LoadDest, Src0, Src1)) { 872 Arith->getDest(), Src0, Src1); 875 Operand *Src0 = Icmp->getSrc(0); 877 if (canFoldLoadIntoBinaryInst(LoadSrc, LoadDest, Src0, Src1)) { 879 Icmp->getDest(), Src0, Src1); [all …]
|
D | IceTargetLoweringARM32.cpp | 541 Operand *Src0 = Instr->getSrc(0); in genTargetHelperCallFor() local 547 Context.insert<InstCast>(CastKind, Src0_32, Src0); in genTargetHelperCallFor() 548 Src0 = Src0_32; in genTargetHelperCallFor() 574 assert(Src0->getType() == IceType_i32); in genTargetHelperCallFor() 575 Call->addArg(Src0); in genTargetHelperCallFor() 602 Operand *Src0 = Instr->getSrc(0); in genTargetHelperCallFor() local 604 const Type SrcTy = Src0->getType(); in genTargetHelperCallFor() 626 Call->addArg(Src0); in genTargetHelperCallFor() 645 Call->addArg(Src0); in genTargetHelperCallFor() 673 Context.insert<InstCast>(InstCast::Zext, Src0AsI32, Src0); in genTargetHelperCallFor() [all …]
|
D | IceInstARM32.h | 746 Variable *Src0, Operand *Src1, 750 InstARM32ThreeAddrGPR(Func, Dest, Src0, Src1, Predicate, SetFlags); 771 InstARM32ThreeAddrGPR(Cfg *Func, Variable *Dest, Variable *Src0, in InstARM32ThreeAddrGPR() argument 775 addSource(Src0); in InstARM32ThreeAddrGPR() 796 static InstARM32ThreeAddrFP *create(Cfg *Func, Variable *Dest, Variable *Src0, in create() argument 799 InstARM32ThreeAddrFP(Func, Dest, Src0, Src1); in create() 822 InstARM32ThreeAddrFP(Cfg *Func, Variable *Dest, Variable *Src0, Operand *Src1) in InstARM32ThreeAddrFP() argument 824 addSource(Src0); in InstARM32ThreeAddrFP() 847 Variable *Src0, Variable *Src1) { in create() argument 849 InstARM32ThreeAddrSignAwareFP(Func, Dest, Src0, Src1); in create() [all …]
|
D | IceInstMIPS32.h | 409 Variable *Src0) { in create() argument 411 InstMIPS32TwoAddrFPR(Func, Dest, Src0); in create() 436 InstMIPS32TwoAddrFPR(Cfg *Func, Variable *Dest, Variable *Src0) in InstMIPS32TwoAddrFPR() argument 438 addSource(Src0); in InstMIPS32TwoAddrFPR() 453 Variable *Src0) { in create() argument 455 InstMIPS32TwoAddrGPR(Func, Dest, Src0); in create() 480 InstMIPS32TwoAddrGPR(Cfg *Func, Variable *Dest, Variable *Src0) in InstMIPS32TwoAddrGPR() argument 482 addSource(Src0); in InstMIPS32TwoAddrGPR() 500 Variable *Src0, Variable *Src1) { in create() argument 502 InstMIPS32ThreeAddrFPR(Func, Dest, Src0, Src1); in create() [all …]
|
D | IceTargetLoweringMIPS32.cpp | 306 Operand *Src0 = Instr->getSrc(0); in genTargetHelperCallFor() local 316 Context.insert<InstExtractElement>(Op0, Src0, Index); in genTargetHelperCallFor() 333 Operand *Src0 = Instr->getSrc(0); in genTargetHelperCallFor() local 335 const Type SrcType = Src0->getType(); in genTargetHelperCallFor() 344 Context.insert<InstExtractElement>(Op0, Src0, Index); in genTargetHelperCallFor() 421 Operand *Src0 = Instr->getSrc(0); in genTargetHelperCallFor() local 422 const Type SrcTy = Src0->getType(); in genTargetHelperCallFor() 435 Context.insert<InstExtractElement>(Op, Src0, Index); in genTargetHelperCallFor() 475 Call->addArg(Src0); in genTargetHelperCallFor() 504 Call->addArg(Src0); in genTargetHelperCallFor() [all …]
|
D | IceInstARM32.cpp | 1168 const Operand *Src0 = getSrc(0); in emitIAS() local 1170 Type SrcTy = Src0->getType(); in emitIAS() 1184 Asm->vqmovn2(typeElementType(DestTy), Dest, Src0, Src1, Unsigned, in emitIAS() 1190 Asm->vqmovn2(typeElementType(DestTy), Dest, Src0, Src1, Unsigned, in emitIAS() 1196 Asm->vqmovn2(typeElementType(DestTy), Dest, Src0, Src1, Unsigned, in emitIAS() 1207 const Operand *Src0 = getSrc(0); in emitIAS() local 1208 Type SrcTy = Src0->getType(); in emitIAS() 1233 const Operand *Src0 = getSrc(0); in emitIAS() local 1235 Type SrcTy = Src0->getType(); in emitIAS() 1241 Asm->vmlap(typeElementType(SrcTy), Dest, Src0, Src1); in emitIAS() [all …]
|
/external/swiftshader/third_party/subzero/unittest/AssemblerX8664/ |
D | ControlFlow.cpp | 17 #define TestJ(C, Near, Dest, Src0, Value0, Src1, Value1) \ in TEST_F() argument 20 "(" #C ", " #Near ", " #Dest ", " #Src0 ", " #Value0 ", " #Src1 \ in TEST_F() 24 __ mov(IceType_i32, Encoded_GPR_##Src0(), Immediate(Value0)); \ in TEST_F() 27 __ cmp(IceType_i32, Encoded_GPR_##Src0(), Encoded_GPR_##Src1()); \ in TEST_F() 33 ASSERT_EQ(Value0, test.Src0()) << TestString; \ in TEST_F() 39 #define TestImpl(Dst, Src0, Src1) \ in TEST_F() argument 41 TestJ(o, Near, Dst, Src0, 0x80000000ul, Src1, 0x1ul); \ in TEST_F() 42 TestJ(o, Far, Dst, Src0, 0x80000000ul, Src1, 0x1ul); \ in TEST_F() 43 TestJ(no, Near, Dst, Src0, 0x1ul, Src1, 0x1ul); \ in TEST_F() 44 TestJ(no, Far, Dst, Src0, 0x1ul, Src1, 0x1ul); \ in TEST_F() [all …]
|
D | DataMov.cpp | 422 #define TestRegReg(C, Dest, IsTrue, Src0, Value0, Src1, Value1) \ in TEST_F() argument 425 "(" #C ", " #Dest ", " #IsTrue ", " #Src0 ", " #Value0 ", " #Src1 \ in TEST_F() 427 __ mov(IceType_i32, Encoded_GPR_##Src0(), Immediate(Value0)); \ in TEST_F() 430 __ cmp(IceType_i32, Encoded_GPR_##Src0(), Encoded_GPR_##Src1()); \ in TEST_F() 441 #define TestRegAddr(C, Dest, IsTrue, Src0, Value0, Value1) \ in TEST_F() argument 444 "(" #C ", " #Dest ", " #IsTrue ", " #Src0 ", " #Value0 \ in TEST_F() 448 __ mov(IceType_i32, Encoded_GPR_##Src0(), Immediate(Value0)); \ in TEST_F() 450 __ cmp(IceType_i32, Encoded_GPR_##Src0(), dwordAddress(T0)); \ in TEST_F() 462 #define TestValue(C, Dest, IsTrue, Src0, Value0, Src1, Value1) \ in TEST_F() argument 464 TestRegReg(C, Dest, IsTrue, Src0, Value0, Src1, Value1); \ in TEST_F() [all …]
|
D | GPRArith.cpp | 33 #define TestSetCC(C, Dest, IsTrue, Src0, Value0, Src1, Value1) \ in TEST_F() argument 36 "(" #C ", " #Dest ", " #IsTrue ", " #Src0 ", " #Value0 ", " #Src1 \ in TEST_F() 40 __ mov(IceType_i32, Encoded_GPR_##Src0(), Immediate(Value0)); \ in TEST_F() 42 __ cmp(IceType_i32, Encoded_GPR_##Src0(), Encoded_GPR_##Src1()); \ in TEST_F() 57 #define TestImpl(Dest, Src0, Src1) \ in TEST_F() argument 59 TestSetCC(o, Dest, 1u, Src0, 0x80000000u, Src1, 0x1u); \ in TEST_F() 60 TestSetCC(o, Dest, 0u, Src0, 0x1u, Src1, 0x10000000u); \ in TEST_F() 61 TestSetCC(no, Dest, 1u, Src0, 0x1u, Src1, 0x10000000u); \ in TEST_F() 62 TestSetCC(no, Dest, 0u, Src0, 0x80000000u, Src1, 0x1u); \ in TEST_F() 63 TestSetCC(b, Dest, 1u, Src0, 0x1, Src1, 0x80000000u); \ in TEST_F() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/unittests/CodeGen/GlobalISel/ |
D | PatternMatchTest.cpp | 164 unsigned Src0, Src1, Src2; in TEST() local 166 m_GAdd(m_Reg(Src0), m_Reg(Src1))); in TEST() 168 ASSERT_EQ(Src0, Copies[0]); in TEST() 176 m_GMul(m_Reg(Src0), m_Reg(Src1))); in TEST() 178 ASSERT_EQ(Src0, MIBAdd->getOperand(0).getReg()); in TEST() 183 m_GMul(m_GAdd(m_Reg(Src0), m_Reg(Src1)), m_Reg(Src2))); in TEST() 185 ASSERT_EQ(Src0, Copies[0]); in TEST() 195 m_GMul(m_ICst(Cst), m_Reg(Src0))); in TEST() 198 ASSERT_EQ(Src0, Copies[0]); in TEST() 203 m_GSub(m_ICst(Cst), m_Reg(Src0))); in TEST() [all …]
|
/external/llvm/lib/Target/AMDGPU/ |
D | SIShrinkInstructions.cpp | 139 MachineOperand &Src0 = MI.getOperand(Src0Idx); in foldImmediates() local 143 if (Src0.isImm() && in foldImmediates() 144 TII->isLiteralConstant(Src0, TII->getOpSize(MI, Src0Idx))) in foldImmediates() 150 if (Src0.isReg() && !isVGPR(&Src0, TRI, MRI)) in foldImmediates() 154 if (Src0.isReg() && MRI.hasOneUse(Src0.getReg())) { in foldImmediates() 155 unsigned Reg = Src0.getReg(); in foldImmediates() 162 Src0.ChangeToImmediate(MovSrc.getImm()); in foldImmediates() 276 const MachineOperand &Src0 = MI.getOperand(1); in runOnMachineFunction() local 283 Src0.isReg()) { in runOnMachineFunction() 284 MRI.setRegAllocationHint(Dest.getReg(), 0, Src0.getReg()); in runOnMachineFunction() [all …]
|
D | R600ExpandSpecialInstrs.cpp | 222 unsigned Src0 = BMI->getOperand( in runOnMachineFunction() local 228 (void) Src0; in runOnMachineFunction() 230 if ((TRI.getEncodingValue(Src0) & 0xff) < 127 && in runOnMachineFunction() 232 assert(TRI.getHWRegChan(Src0) == TRI.getHWRegChan(Src1)); in runOnMachineFunction() 274 unsigned Src0 = MI.getOperand( in runOnMachineFunction() local 287 Src0 = TRI.getSubReg(Src0, SubRegIndex); in runOnMachineFunction() 293 Src1 = TRI.getSubReg(Src0, SubRegIndex1); in runOnMachineFunction() 294 Src0 = TRI.getSubReg(Src0, SubRegIndex0); in runOnMachineFunction() 328 TII->buildDefaultInstruction(MBB, I, Opcode, DstReg, Src0, Src1); in runOnMachineFunction()
|
D | SIInstrInfo.cpp | 889 unsigned Src0 = MI.getOperand(1).getReg(); in expandPostRAPseudo() local 894 .addReg(RI.getSubReg(Src0, AMDGPU::sub0)) in expandPostRAPseudo() 899 .addReg(RI.getSubReg(Src0, AMDGPU::sub1)) in expandPostRAPseudo() 954 MachineOperand &Src0 = MI.getOperand(Src0Idx); in commuteInstructionImpl() local 955 if (!Src0.isReg()) in commuteInstructionImpl() 978 if (!isLegalRegOperand(MRI, InstrDesc.OpInfo[Src1Idx], Src0)) in commuteInstructionImpl() 1007 unsigned Reg = Src0.getReg(); in commuteInstructionImpl() 1008 unsigned SubReg = Src0.getSubReg(); in commuteInstructionImpl() 1010 Src0.ChangeToImmediate(Src1.getImm()); in commuteInstructionImpl() 1239 MachineOperand *Src0 = getNamedOperand(UseMI, AMDGPU::OpName::src0); in FoldImmediate() local [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/AMDGPU/ |
D | R600ExpandSpecialInstrs.cpp | 159 unsigned Src0 = BMI->getOperand( in runOnMachineFunction() local 165 (void) Src0; in runOnMachineFunction() 167 if ((TRI.getEncodingValue(Src0) & 0xff) < 127 && in runOnMachineFunction() 169 assert(TRI.getHWRegChan(Src0) == TRI.getHWRegChan(Src1)); in runOnMachineFunction() 211 unsigned Src0 = MI.getOperand( in runOnMachineFunction() local 224 Src0 = TRI.getSubReg(Src0, SubRegIndex); in runOnMachineFunction() 230 Src1 = TRI.getSubReg(Src0, SubRegIndex1); in runOnMachineFunction() 231 Src0 = TRI.getSubReg(Src0, SubRegIndex0); in runOnMachineFunction() 265 TII->buildDefaultInstruction(MBB, I, Opcode, DstReg, Src0, Src1); in runOnMachineFunction()
|
D | SIShrinkInstructions.cpp | 130 MachineOperand &Src0 = MI.getOperand(Src0Idx); in foldImmediates() local 131 if (Src0.isReg()) { in foldImmediates() 132 unsigned Reg = Src0.getReg(); in foldImmediates() 143 Src0.setSubReg(0); in foldImmediates() 144 Src0.ChangeToImmediate(MovSrc.getImm()); in foldImmediates() 147 Src0.setSubReg(0); in foldImmediates() 148 Src0.ChangeToFrameIndex(MovSrc.getIndex()); in foldImmediates() 357 MachineOperand *Src0 = &MI.getOperand(1); in runOnMachineFunction() local 360 if (!Src0->isReg() && Src1->isReg()) { in runOnMachineFunction() 362 std::swap(Src0, Src1); in runOnMachineFunction() [all …]
|
D | SIFoldOperands.cpp | 547 MachineOperand *Src0 = getImmOrMaterializedImm(MRI, MI->getOperand(Src0Idx)); in tryConstantFoldOp() local 550 if (!Src0->isImm() && !Src1->isImm()) in tryConstantFoldOp() 556 if (Src0->isImm() && Src1->isImm()) { in tryConstantFoldOp() 558 if (!evalBinaryInstruction(Opc, NewImm, Src0->getImm(), Src1->getImm())) in tryConstantFoldOp() 575 if (Src0->isImm() && !Src1->isImm()) { in tryConstantFoldOp() 576 std::swap(Src0, Src1); in tryConstantFoldOp() 638 const MachineOperand *Src0 = TII->getNamedOperand(*MI, AMDGPU::OpName::src0); in tryFoldInst() local 640 if (Src1->isIdenticalTo(*Src0)) { in tryFoldInst() 646 mutateCopyOp(*MI, TII->get(Src0->isReg() ? (unsigned)AMDGPU::COPY in tryFoldInst() 780 const MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); in isClamp() local [all …]
|
D | SIInstrInfo.cpp | 1327 MachineOperand &Src0, in swapSourceModifiers() argument 1386 MachineOperand &Src0 = MI.getOperand(Src0Idx); in commuteInstructionImpl() local 1390 if (Src0.isReg() && Src1.isReg()) { in commuteInstructionImpl() 1391 if (isOperandLegal(MI, Src1Idx, &Src0)) { in commuteInstructionImpl() 1397 } else if (Src0.isReg() && !Src1.isReg()) { in commuteInstructionImpl() 1400 CommutedMI = swapRegAndNonRegOperand(MI, Src0, Src1); in commuteInstructionImpl() 1401 } else if (!Src0.isReg() && Src1.isReg()) { in commuteInstructionImpl() 1402 if (isOperandLegal(MI, Src1Idx, &Src0)) in commuteInstructionImpl() 1403 CommutedMI = swapRegAndNonRegOperand(MI, Src1, Src0); in commuteInstructionImpl() 1410 swapSourceModifiers(MI, Src0, AMDGPU::OpName::src0_modifiers, in commuteInstructionImpl() [all …]
|
D | SIPeepholeSDWA.cpp | 562 MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); in matchSDWAOperand() local 563 auto Imm = foldToImm(*Src0); in matchSDWAOperand() 603 MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); in matchSDWAOperand() local 604 auto Imm = foldToImm(*Src0); in matchSDWAOperand() 672 MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); in matchSDWAOperand() local 675 if (TRI->isPhysicalRegister(Src0->getReg()) || in matchSDWAOperand() 680 Src0, Dst, SrcSel, false, false, Opcode != AMDGPU::V_BFE_U32); in matchSDWAOperand() 689 MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); in matchSDWAOperand() local 692 auto Imm = foldToImm(*Src0); in matchSDWAOperand() 696 ValSrc = Src0; in matchSDWAOperand() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/include/llvm/CodeGen/GlobalISel/ |
D | ConstantFoldingMIRBuilder.h | 83 unsigned Src0, unsigned Src1) { in buildBinaryOp() argument 84 validateBinaryOp(Dst, Src0, Src1); in buildBinaryOp() 85 auto MaybeCst = ConstantFoldBinOp(Opcode, Src0, Src1, getMF().getRegInfo()); in buildBinaryOp() 88 return buildInstr(Opcode).addDef(Dst).addUse(Src0).addUse(Src1); in buildBinaryOp() 101 MachineInstrBuilder buildInstr(unsigned Opc, unsigned Dst, unsigned Src0, in buildInstr() argument 119 return buildBinaryOp(Opc, Dst, Src0, Src1); in buildInstr() 122 return buildInstr(Opc).addDef(Dst).addUse(Src0).addUse(Src1); in buildInstr()
|
D | MachineIRBuilder.h | 979 MachineInstrBuilder buildAdd(unsigned Dst, unsigned Src0, unsigned Src1) { in buildAdd() argument 980 return base().buildBinaryOp(TargetOpcode::G_ADD, Dst, Src0, Src1); in buildAdd() 999 MachineInstrBuilder buildSub(unsigned Dst, unsigned Src0, unsigned Src1) { in buildSub() argument 1000 return base().buildBinaryOp(TargetOpcode::G_SUB, Dst, Src0, Src1); in buildSub() 1018 MachineInstrBuilder buildMul(unsigned Dst, unsigned Src0, unsigned Src1) { in buildMul() argument 1019 return base().buildBinaryOp(TargetOpcode::G_MUL, Dst, Src0, Src1); in buildMul() 1038 MachineInstrBuilder buildAnd(unsigned Dst, unsigned Src0, unsigned Src1) { in buildAnd() argument 1039 return base().buildBinaryOp(TargetOpcode::G_AND, Dst, Src0, Src1); in buildAnd() 1057 MachineInstrBuilder buildOr(unsigned Dst, unsigned Src0, unsigned Src1) { in buildOr() argument 1058 return base().buildBinaryOp(TargetOpcode::G_OR, Dst, Src0, Src1); in buildOr() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/AArch64/ |
D | AArch64AdvSIMDScalarPass.cpp | 299 unsigned Src0 = 0, SubReg0; in transformInstruction() local 310 Src0 = MOSrc0->getReg(); in transformInstruction() 342 if (!Src0) { in transformInstruction() 344 Src0 = MRI->createVirtualRegister(&AArch64::FPR64RegClass); in transformInstruction() 345 insertCopy(TII, MI, Src0, OrigSrc0, KillSrc0); in transformInstruction() 364 .addReg(Src0, getKillRegState(KillSrc0), SubReg0) in transformInstruction()
|
/external/llvm/lib/Target/AArch64/ |
D | AArch64AdvSIMDScalarPass.cpp | 306 unsigned Src0 = 0, SubReg0; in transformInstruction() local 317 Src0 = MOSrc0->getReg(); in transformInstruction() 349 if (!Src0) { in transformInstruction() 351 Src0 = MRI->createVirtualRegister(&AArch64::FPR64RegClass); in transformInstruction() 352 insertCopy(TII, MI, Src0, OrigSrc0, KillSrc0); in transformInstruction() 371 .addReg(Src0, getKillRegState(KillSrc0), SubReg0) in transformInstruction()
|