/external/swiftshader/third_party/LLVM/lib/Target/X86/ |
D | X86GenFastISel.inc | 53 if ((Subtarget->hasAVX())) { 63 if ((Subtarget->hasAVX())) { 91 if ((Subtarget->hasAVX())) { 101 if ((Subtarget->hasAVX())) { 475 if ((Subtarget->hasAVX())) { 485 if ((Subtarget->hasAVX())) { 503 if ((Subtarget->hasAVX())) { 513 if ((Subtarget->hasAVX())) { 533 if ((Subtarget->hasAVX())) { 545 if ((Subtarget->hasAVX())) { [all …]
|
D | X86GenDAGISel.inc | 28 /*20*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 64 /*101*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 99 /*179*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 119 /*223*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 137 /*263*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 3779 /*8369*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 3813 /*8444*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 3837 /*8502*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 3861 /*8560*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) 4178 /*9234*/ OPC_CheckPatternPredicate, 0, // (Subtarget->hasAVX()) [all …]
|
D | X86Subtarget.h | 183 bool hasAVX() const { return HasAVX; } in hasAVX() function 184 bool hasXMM() const { return hasSSE1() || hasAVX(); } in hasXMM() 185 bool hasXMMInt() const { return hasSSE2() || hasAVX(); } in hasXMMInt()
|
D | X86TargetMachine.cpp | 139 (Subtarget.hasSSE2() || Subtarget.hasAVX())) { in addPreEmitPass() 144 if (Subtarget.hasAVX() && UseVZeroUpper) { in addPreEmitPass()
|
D | X86FastISel.cpp | 63 X86ScalarSSEf64 = Subtarget->hasSSE2() || Subtarget->hasAVX(); in X86FastISel() 64 X86ScalarSSEf32 = Subtarget->hasSSE1() || Subtarget->hasAVX(); in X86FastISel() 203 Opc = Subtarget->hasAVX() ? X86::VMOVSSrm : X86::MOVSSrm; in X86FastEmitLoad() 212 Opc = Subtarget->hasAVX() ? X86::VMOVSDrm : X86::MOVSDrm; in X86FastEmitLoad() 255 (Subtarget->hasAVX() ? X86::VMOVSSmr : X86::MOVSSmr) : X86::ST_Fp32m; in X86FastEmitStore() 259 (Subtarget->hasAVX() ? X86::VMOVSDmr : X86::MOVSDmr) : X86::ST_Fp64m; in X86FastEmitStore() 820 bool HasAVX = Subtarget->hasAVX(); in X86ChooseCmpOpcode() 1998 Opc = Subtarget->hasAVX() ? X86::VMOVSSrm : X86::MOVSSrm; in TargetMaterializeConstant() 2007 Opc = Subtarget->hasAVX() ? X86::VMOVSDrm : X86::MOVSDrm; in TargetMaterializeConstant()
|
D | X86ISelLowering.cpp | 912 if (Subtarget->hasSSE41() || Subtarget->hasAVX()) { in X86TargetLowering() 971 if (Subtarget->hasSSE42() || Subtarget->hasAVX()) in X86TargetLowering() 974 if (!UseSoftFloat && Subtarget->hasAVX()) { in X86TargetLowering() 1257 if (Subtarget->hasAVX() && in getOptimalMemOpType() 3235 if (!Subtarget->hasAVX() || VT.getSizeInBits() != 256) in isVSHUFPSYMask() 3319 if (!Subtarget->hasAVX() || VT.getSizeInBits() != 256) in isVSHUFPDYMask() 3703 if (!Subtarget->hasAVX() || VT.getSizeInBits() != 256) in isVPERM2F128Mask() 3768 if (!Subtarget->hasAVX()) in isVPERMILPDMask() 3800 if (!Subtarget->hasAVX()) in isVPERMILPSMask() 3911 if (!Subtarget->hasSSE3() && !Subtarget->hasAVX()) in isMOVSHDUPMask() [all …]
|
D | X86CallingConv.td | 163 CCIfSubtarget<"hasAVX()", 278 CCIfSubtarget<"hasAVX()",
|
D | X86GenCallingConv.inc | 243 if (State.getTarget().getSubtarget<X86Subtarget>().hasAVX()) { 563 if (State.getTarget().getSubtarget<X86Subtarget>().hasAVX()) {
|
D | X86InstrInfo.cpp | 2180 bool HasAVX = TM.getSubtarget<X86Subtarget>().hasAVX(); in copyPhysReg() 2250 bool HasAVX = TM.getSubtarget<X86Subtarget>().hasAVX(); in getLoadStoreRegOpcode() 2420 bool HasAVX = TM.getSubtarget<X86Subtarget>().hasAVX(); in expandPostRAPseudo()
|
D | X86InstrInfo.td | 470 def HasAVX : Predicate<"Subtarget->hasAVX()">;
|
/external/llvm/lib/Target/X86/ |
D | X86TargetTransformInfo.cpp | 59 if (ST->hasAVX()) return 256; in getRegisterBitWidth() 82 if (ST->hasAVX()) in getMaxInterleaveFactor() 302 (ST->hasAVX() && !ST->hasAVX2())) in getArithmeticInstrCost() 388 if (ST->hasAVX() && !ST->hasAVX2()) { in getArithmeticInstrCost() 457 if (ST->hasAVX()) in getShuffleCost() 821 if (ST->hasSSE2() && !ST->hasAVX()) { in getCastInstrCost() 853 if (ST->hasAVX()) { in getCastInstrCost() 931 if (ST->hasAVX()) in getCmpSelInstrCost() 1020 if (ST->hasAVX()) in getIntrinsicInstrCost() 1245 if (ST->hasAVX()) in getReductionCost() [all …]
|
D | X86Subtarget.h | 388 bool hasAVX() const { return X86SSELevel >= AVX; } in hasAVX() function 391 bool hasFp256() const { return hasAVX(); } in hasFp256()
|
D | X86FastISel.cpp | 352 bool HasAVX = Subtarget->hasAVX(); in X86FastEmitLoad() 506 bool HasAVX = Subtarget->hasAVX(); in X86FastEmitStore() 1329 bool HasAVX = Subtarget->hasAVX(); in X86ChooseCmpOpcode() 2125 if (Subtarget->hasAVX()) { in X86FastEmitSSESelect() 2274 if (!Subtarget->hasAVX()) in X86SelectSIToFP() 2324 if (Subtarget->hasAVX()) in X86SelectFPExtOrFPTrunc() 2335 unsigned Opc = Subtarget->hasAVX() ? X86::VCVTSS2SDrr : X86::CVTSS2SDrr; in X86SelectFPExt() 2346 unsigned Opc = Subtarget->hasAVX() ? X86::VCVTSD2SSrr : X86::CVTSD2SSrr; in X86SelectFPTrunc() 2657 bool HasAVX = Subtarget->hasAVX(); in fastLowerIntrinsicCall() 2853 bool HasAVX = Subtarget->hasAVX(); in fastLowerIntrinsicCall() [all …]
|
D | X86RegisterInfo.cpp | 246 bool HasAVX = Subtarget.hasAVX(); in getCalleeSavedRegs() 345 bool HasAVX = Subtarget.hasAVX(); in getCallPreservedMask()
|
D | X86VZeroUpper.cpp | 260 if (!ST.hasAVX() || ST.hasAVX512() || ST.hasFastPartialYMMWrite()) in runOnMachineFunction()
|
D | X86InstrInfo.td | 779 def UseSSE1 : Predicate<"Subtarget->hasSSE1() && !Subtarget->hasAVX()">; 781 def UseSSE2 : Predicate<"Subtarget->hasSSE2() && !Subtarget->hasAVX()">; 783 def UseSSE3 : Predicate<"Subtarget->hasSSE3() && !Subtarget->hasAVX()">; 785 def UseSSSE3 : Predicate<"Subtarget->hasSSSE3() && !Subtarget->hasAVX()">; 788 def UseSSE41 : Predicate<"Subtarget->hasSSE41() && !Subtarget->hasAVX()">; 790 def UseSSE42 : Predicate<"Subtarget->hasSSE42() && !Subtarget->hasAVX()">; 792 def HasAVX : Predicate<"Subtarget->hasAVX()">; 794 def HasAVX1Only : Predicate<"Subtarget->hasAVX() && !Subtarget->hasAVX2()">; 797 def UseAVX : Predicate<"Subtarget->hasAVX() && !Subtarget->hasAVX512()">;
|
D | X86InstrInfo.cpp | 4340 bool HasAVX = Subtarget.hasAVX(); in CopyToFromAsymmetricReg() 4452 bool HasAVX = Subtarget.hasAVX(); in copyPhysReg() 4623 bool HasAVX = STI.hasAVX(); in getLoadStoreRegOpcode() 5515 bool HasAVX = Subtarget.hasAVX(); in expandPostRAPseudo() 6087 unsigned Opc = Subtarget.hasAVX() ? X86::VXORPSrr : X86::XORPSrr; in breakPartialRegDependency() 6768 if (!Subtarget.hasAVX()) in shouldScheduleAdjacent()
|
D | X86ISelLowering.cpp | 1815 if (Size >= 32 && Subtarget.hasAVX()) { in getOptimalMemOpType() 2815 else if (Subtarget.hasAVX()) in LowerFormalArguments() 5749 if (!Subtarget.hasAVX()) in LowerVectorBroadcast() 6271 (!Subtarget.hasAVX() || (VT != MVT::v8f32 && VT != MVT::v4f64))) in LowerToAddSub() 6409 if (!Subtarget.hasAVX()) in LowerToHorizontalOp() 6467 VT == MVT::v16i16) && Subtarget.hasAVX()) { in LowerToHorizontalOp() 6683 if (Subtarget.hasAVX()) { in LowerBUILD_VECTOR() 8542 (Subtarget.hasAVX() && VT.isFloatingPoint()) || in lowerVectorShuffleAsBroadcast() 8924 if (Subtarget.hasAVX()) { in lowerV2F64VectorShuffle() 9228 if (Subtarget.hasAVX()) { in lowerV4F32VectorShuffle() [all …]
|