Lines Matching refs:is128BitVector
2625 else if (RegVT.is128BitVector()) in LowerFormalArguments()
3116 else if (RegVT.is128BitVector()) { in LowerCall()
4393 assert((VT.is128BitVector() || VT.is256BitVector() || VT.is512BitVector() || in getZeroVector()
4401 if (!Subtarget.hasSSE2() && VT.is128BitVector()) { in getZeroVector()
4499 assert(Vec.getValueType().is128BitVector() && "Unexpected vector size!"); in insert128BitVector()
4681 assert((VT.is128BitVector() || VT.is256BitVector() || VT.is512BitVector()) && in getOnesVector()
4699 assert(VT.is128BitVector() && "Expected a 128-bit vector type"); in getUnpackl()
4712 assert(VT.is128BitVector() && "Expected a 128-bit vector type"); in getUnpackh()
5389 if (!VT.is128BitVector()) in LowerBuildVectorv4x32()
5473 assert(VT.is128BitVector() && "Unknown type for VShift"); in getVShift()
5691 ((VT.is128BitVector() || VT.is256BitVector() || VT.is512BitVector()))) { in EltsFromConsecutiveLoads()
5722 ((VT.is128BitVector() || VT.is256BitVector() || VT.is512BitVector()))) { in EltsFromConsecutiveLoads()
5755 assert((VT.is128BitVector() || VT.is256BitVector() || VT.is512BitVector()) && in LowerVectorBroadcast()
6670 assert((VT.is128BitVector() || VT.is256BitVector()) && in LowerBUILD_VECTOR()
6694 assert(VT.is128BitVector() && "Expected an SSE value type!"); in LowerBUILD_VECTOR()
6748 if (VT.is128BitVector() || VT.is256BitVector() || VT.is512BitVector()) { in LowerBUILD_VECTOR()
6841 if (Values.size() > 1 && VT.is128BitVector()) { in LowerBUILD_VECTOR()
7313 assert((Subtarget.hasSSSE3() && VT.is128BitVector()) || in lowerVectorShuffleWithPSHUFB()
7578 assert((VT.is128BitVector() || Subtarget.hasAVX2()) && in lowerVectorShuffleAsBlend()
7806 assert(VT.is128BitVector() && in lowerVectorShuffleAsByteRotate()
8096 if (Offset && Scale == 2 && VT.is128BitVector()) in lowerVectorShuffleAsSpecificZeroOrAnyExtend()
8110 assert(VT.is128BitVector() && "Only 128-bit vectors can be extended."); in lowerVectorShuffleAsSpecificZeroOrAnyExtend()
8140 assert(VT.is128BitVector() && "Unexpected vector width!"); in lowerVectorShuffleAsSpecificZeroOrAnyExtend()
8635 Opcode = (BroadcastVT.is128BitVector() ? X86ISD::MOVDDUP : Opcode); in lowerVectorShuffleAsBroadcast()
8704 assert(V1.getSimpleValueType().is128BitVector() && "Bad operand type!"); in matchVectorShuffleAsInsertPS()
8705 assert(V2.getSimpleValueType().is128BitVector() && "Bad operand type!"); in matchVectorShuffleAsInsertPS()
8798 assert(VT.is128BitVector() && in lowerVectorShuffleAsPermuteAndUnpack()
12235 if (VT.is128BitVector()) in lowerVectorShuffle()
12324 if (!Op.getOperand(0).getSimpleValueType().is128BitVector()) in LowerEXTRACT_VECTOR_ELT_SSE4()
12472 assert(VecVT.is128BitVector() && "Unexpected vector length"); in LowerEXTRACT_VECTOR_ELT()
12618 assert(VT.is128BitVector() && "Only 128-bit vector types should be left!"); in LowerINSERT_VECTOR_ELT()
12695 if (!OpVT.is128BitVector()) { in LowerSCALAR_TO_VECTOR()
12712 assert(OpVT.is128BitVector() && "Expected an SSE type!"); in LowerSCALAR_TO_VECTOR()
12730 if (ResVT.is128BitVector() && in LowerEXTRACT_SUBVECTOR()
12769 OpVT.is256BitVector() && SubVecVT.is128BitVector()) { in LowerINSERT_SUBVECTOR()
12790 SubVecVT.is128BitVector()) in LowerINSERT_SUBVECTOR()
14077 assert(!VT.is256BitVector() || !SVT.is128BitVector() || in LowerZERO_EXTEND()
14106 assert((InVT.is256BitVector() || InVT.is128BitVector()) && in LowerTruncateVecI1()
14226 if (!VT.is128BitVector() || !InVT.is256BitVector()) in LowerTRUNCATE()
14532 if (!VT.is128BitVector() && !VT.is256BitVector()) in LowerVectorAllZeroTest()
14544 assert((VT.is128BitVector() || VT.is256BitVector()) && in LowerVectorAllZeroTest()
14556 MVT TestVT = VT.is128BitVector() ? MVT::v2i64 : MVT::v4i64; in LowerVectorAllZeroTest()
15323 if (VT.is128BitVector() && VTOp0.is256BitVector()) { in LowerVSETCC()
16066 if (!(VT.is128BitVector() && Subtarget.hasSSE2()) && in LowerSIGN_EXTEND_VECTOR_INREG()
18985 assert((VT.is256BitVector() || VT.is128BitVector()) && in LowerVectorCTLZ_AVX512()
19718 bool LShift = VT.is128BitVector() || in SupportedVectorShiftWithImm()
19749 bool LShift = VT.is128BitVector() || VT.is256BitVector(); in SupportedVectorVarShift()
20457 assert(VT.is128BitVector() && "Only rotate 128-bit vectors!"); in LowerRotate()
20956 assert(VT.is128BitVector() && in LowerVectorCTPOPBitmath()
21020 assert((VT.is512BitVector() || VT.is256BitVector() || VT.is128BitVector()) && in LowerVectorCTPOP()
21027 assert(VT.is128BitVector() && "Only 128-bit vectors supported in SSE!"); in LowerVectorCTPOP()
21095 assert(VT.is128BitVector() && in LowerBITREVERSE_XOP()
24757 if (!FloatDomain && SrcVT.is128BitVector() && in matchUnaryVectorShuffle()
24767 if (SrcVT.is128BitVector() && Subtarget.hasSSE3() && FloatDomain) { in matchUnaryVectorShuffle()
24970 if (SrcVT.is128BitVector()) { in matchBinaryVectorShuffle()
25149 ((Subtarget.hasSSE41() && VT.is128BitVector()) || in combineX86ShuffleChain()
25262 ((VT.is128BitVector() && Subtarget.hasSSSE3()) || in combineX86ShuffleChain()
25728 if (!VT.is128BitVector()) in combineTargetShuffle()
26600 if ((OpVT.is128BitVector() || OpVT.is256BitVector()) && in combineSelect()
26877 if (VT.is128BitVector() && !Subtarget.hasSSE41()) in combineSelect()
28011 if (!NarrowVT.is128BitVector()) in WidenMaskArithmetic()
29458 assert((VT.is128BitVector() || VT.is256BitVector()) && in isHorizontalBinOp()
30131 if (!Subtarget.hasSSE41() || VT.is128BitVector() || in combineToExtendVectorInReg()