Searched refs:X86ScalarSSEf64 (Results 1 – 9 of 9) sorted by relevance
56 bool X86ScalarSSEf64; member in __anon412e6d530111::X86FastISel63 X86ScalarSSEf64 = Subtarget->hasSSE2() || Subtarget->hasAVX(); in X86FastISel()134 return (VT == MVT::f64 && X86ScalarSSEf64) || // f64 is when SSE2 in isScalarFPTypeInSSEReg()157 if (VT == MVT::f64 && !X86ScalarSSEf64) in isTypeLegal()211 if (X86ScalarSSEf64) { in X86FastEmitLoad()258 Opc = X86ScalarSSEf64 ? in X86FastEmitStore()822 bool X86ScalarSSEf64 = HasAVX || Subtarget->hasSSE2(); in X86ChooseCmpOpcode() local833 return X86ScalarSSEf64 ? (HasAVX ? X86::VUCOMISDrr : X86::UCOMISDrr) : 0; in X86ChooseCmpOpcode()1236 if (X86ScalarSSEf64 && in X86SelectFPExt()1255 if (X86ScalarSSEf64) { in X86SelectFPTrunc()[all …]
693 return !X86ScalarSSEf64 || VT == MVT::f80; in ShouldShrinkFPConstant()703 return (VT == MVT::f64 && X86ScalarSSEf64) || // f64 is when SSE2 in isScalarFPTypeInSSEReg()739 bool X86ScalarSSEf64; variable
172 X86ScalarSSEf64 = Subtarget->hasXMMInt(); in X86TargetLowering()331 if (!X86ScalarSSEf64) { in X86TargetLowering()548 if (!UseSoftFloat && X86ScalarSSEf64) { in X86TargetLowering()7830 if (SrcVT == MVT::i64 && DstVT == MVT::f64 && X86ScalarSSEf64) in LowerUINT_TO_FP()7832 else if (SrcVT == MVT::i32 && X86ScalarSSEf64) in LowerUINT_TO_FP()
925 return !X86ScalarSSEf64 || VT == MVT::f80; in ShouldShrinkFPConstant()936 return (VT == MVT::f64 && X86ScalarSSEf64) || // f64 is when SSE2 in isScalarFPTypeInSSEReg()1016 bool X86ScalarSSEf64; variable
57 bool X86ScalarSSEf64; member in __anon350396270111::X86FastISel65 X86ScalarSSEf64 = Subtarget->hasSSE2(); in X86FastISel()157 return (VT == MVT::f64 && X86ScalarSSEf64) || // f64 is when SSE2 in isScalarFPTypeInSSEReg()329 if (VT == MVT::f64 && !X86ScalarSSEf64) in isTypeLegal()389 if (X86ScalarSSEf64) { in X86FastEmitLoad()1331 bool X86ScalarSSEf64 = Subtarget->hasSSE2(); in X86ChooseCmpOpcode() local1342 return X86ScalarSSEf64 ? (HasAVX ? X86::VUCOMISDrr : X86::UCOMISDrr) : 0; in X86ChooseCmpOpcode()2332 if (X86ScalarSSEf64 && I->getType()->isDoubleTy() && in X86SelectFPExt()2343 if (X86ScalarSSEf64 && I->getType()->isFloatTy() && in X86SelectFPTrunc()3559 if (X86ScalarSSEf64) { in X86MaterializeFP()[all …]
76 X86ScalarSSEf64 = Subtarget.hasSSE2(); in X86TargetLowering()261 if (!X86ScalarSSEf64) { in X86TargetLowering()484 if (!Subtarget.useSoftFloat() && X86ScalarSSEf64) { in X86TargetLowering()1851 return X86ScalarSSEf64; in isSafeMemOpType()13712 if (SrcVT == MVT::i64 && DstVT == MVT::f64 && X86ScalarSSEf64) in LowerUINT_TO_FP()13714 if (SrcVT == MVT::i32 && X86ScalarSSEf64) in LowerUINT_TO_FP()
57 bool X86ScalarSSEf64; member in __anon9f96237f0111::X86FastISel65 X86ScalarSSEf64 = Subtarget->hasSSE2(); in X86FastISel()161 return (VT == MVT::f64 && X86ScalarSSEf64) || // f64 is when SSE2 in isScalarFPTypeInSSEReg()301 if (VT == MVT::f64 && !X86ScalarSSEf64) in isTypeLegal()363 if (X86ScalarSSEf64) { in X86FastEmitLoad()1354 bool X86ScalarSSEf64 = Subtarget->hasSSE2(); in X86ChooseCmpOpcode() local1368 return X86ScalarSSEf64 in X86ChooseCmpOpcode()2515 if (X86ScalarSSEf64 && I->getType()->isDoubleTy() && in X86SelectFPExt()2530 if (X86ScalarSSEf64 && I->getType()->isFloatTy() && in X86SelectFPTrunc()3792 if (X86ScalarSSEf64) { in X86MaterializeFP()[all …]
1013 return !X86ScalarSSEf64 || VT == MVT::f80; in ShouldShrinkFPConstant()1024 return (VT == MVT::f64 && X86ScalarSSEf64) || // f64 is when SSE2 in isScalarFPTypeInSSEReg()1147 bool X86ScalarSSEf64; variable
104 X86ScalarSSEf64 = Subtarget.hasSSE2(); in X86TargetLowering()300 if (!X86ScalarSSEf64) { in X86TargetLowering()518 if (!Subtarget.useSoftFloat() && X86ScalarSSEf64) { in X86TargetLowering()1954 return X86ScalarSSEf64; in isSafeMemOpType()16636 if (SrcVT == MVT::i64 && DstVT == MVT::f64 && X86ScalarSSEf64) in LowerUINT_TO_FP()16638 if (SrcVT == MVT::i32 && X86ScalarSSEf64) in LowerUINT_TO_FP()