/external/llvm/lib/Target/PowerPC/ |
D | PPCInstrQPX.td | 42 [(set v4f64:$FRT, (IntID v4f64:$FRA, v4f64:$FRB, v4f64:$FRC))]>; 47 [(set v4f64:$FRT, (IntID v4f64:$FRA, v4f64:$FRB, v4f64:$FRC))]>; 52 [(set v4f64:$FRT, (IntID v4f64:$FRA, v4f64:$FRB))]>; 57 [(set v4f64:$FRT, (IntID v4f64:$FRA, v4f64:$FRC))]>; 62 [(set v4f64:$FRT, (IntID v4f64:$FRB))]>; 67 [(set v4f64:$FRT, (IntID v4f64:$FRA, v4f64:$FRB))]>; 72 [(set v4f64:$FRT, (IntID v4f64:$FRB))]>; 115 [(set v4f64:$FRT, (fadd v4f64:$FRA, v4f64:$FRB))]>; 126 [(set v4f64:$FRT, (fsub v4f64:$FRA, v4f64:$FRB))]>; 137 [(set v4f64:$FRT, (PPCfre v4f64:$FRB))]>; [all …]
|
D | PPCISelLowering.cpp | 660 setOperationAction(ISD::FADD, MVT::v4f64, Legal); in PPCTargetLowering() 661 setOperationAction(ISD::FSUB, MVT::v4f64, Legal); in PPCTargetLowering() 662 setOperationAction(ISD::FMUL, MVT::v4f64, Legal); in PPCTargetLowering() 663 setOperationAction(ISD::FREM, MVT::v4f64, Expand); in PPCTargetLowering() 665 setOperationAction(ISD::FCOPYSIGN, MVT::v4f64, Legal); in PPCTargetLowering() 666 setOperationAction(ISD::FGETSIGN, MVT::v4f64, Expand); in PPCTargetLowering() 668 setOperationAction(ISD::LOAD , MVT::v4f64, Custom); in PPCTargetLowering() 669 setOperationAction(ISD::STORE , MVT::v4f64, Custom); in PPCTargetLowering() 671 setTruncStoreAction(MVT::v4f64, MVT::v4f32, Custom); in PPCTargetLowering() 672 setLoadExtAction(ISD::EXTLOAD, MVT::v4f64, MVT::v4f32, Custom); in PPCTargetLowering() [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vector-intrinsics.ll | 3 declare <4 x double> @llvm.sin.v4f64(<4 x double> %p) 4 declare <4 x double> @llvm.cos.v4f64(<4 x double> %p) 5 declare <4 x double> @llvm.pow.v4f64(<4 x double> %p, <4 x double> %q) 6 declare <4 x double> @llvm.powi.v4f64(<4 x double> %p, i32) 10 %t = call <4 x double> @llvm.sin.v4f64(<4 x double> %p) 15 %t = call <4 x double> @llvm.cos.v4f64(<4 x double> %p) 20 %t = call <4 x double> @llvm.pow.v4f64(<4 x double> %p, <4 x double> %q) 25 %t = call <4 x double> @llvm.powi.v4f64(<4 x double> %p, i32 %q)
|
D | vec_floor.ll | 26 %t = call <4 x double> @llvm.floor.v4f64(<4 x double> %p) 29 declare <4 x double> @llvm.floor.v4f64(<4 x double> %p) 62 %t = call <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 65 declare <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 98 %t = call <4 x double> @llvm.trunc.v4f64(<4 x double> %p) 101 declare <4 x double> @llvm.trunc.v4f64(<4 x double> %p) 134 %t = call <4 x double> @llvm.rint.v4f64(<4 x double> %p) 137 declare <4 x double> @llvm.rint.v4f64(<4 x double> %p) 170 %t = call <4 x double> @llvm.nearbyint.v4f64(<4 x double> %p) 173 declare <4 x double> @llvm.nearbyint.v4f64(<4 x double> %p)
|
D | fma-phi-213-to-231.ll | 28 %add = call <4 x double> @llvm.fma.v4f64(<4 x double> %x, <4 x double> %y, <4 x double> %acc.04) 37 declare <4 x double> @llvm.fma.v4f64(<4 x double>, <4 x double>, <4 x double>)
|
D | vec_fabs.ll | 26 %t = call <4 x double> @llvm.fabs.v4f64(<4 x double> %p) 29 declare <4 x double> @llvm.fabs.v4f64(<4 x double> %p)
|
/external/llvm/test/CodeGen/PowerPC/ |
D | qpx-rounding-ops.ll | 20 %call = tail call <4 x double> @llvm.floor.v4f64(<4 x double> %x) nounwind readnone 30 declare <4 x double> @llvm.floor.v4f64(<4 x double>) nounwind readnone 46 %call = tail call <4 x double> @llvm.nearbyint.v4f64(<4 x double> %x) nounwind readnone 56 declare <4 x double> @llvm.nearbyint.v4f64(<4 x double>) nounwind readnone 72 %call = tail call <4 x double> @llvm.ceil.v4f64(<4 x double> %x) nounwind readnone 82 declare <4 x double> @llvm.ceil.v4f64(<4 x double>) nounwind readnone 98 %call = tail call <4 x double> @llvm.trunc.v4f64(<4 x double> %x) nounwind readnone 108 declare <4 x double> @llvm.trunc.v4f64(<4 x double>) nounwind readnone
|
D | vec_rounding.ll | 19 declare <4 x double> @llvm.floor.v4f64(<4 x double> %p) 22 %t = call <4 x double> @llvm.floor.v4f64(<4 x double> %p) 41 declare <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 44 %t = call <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 63 declare <4 x double> @llvm.trunc.v4f64(<4 x double> %p) 66 %t = call <4 x double> @llvm.trunc.v4f64(<4 x double> %p) 85 declare <4 x double> @llvm.nearbyint.v4f64(<4 x double> %p) 88 %t = call <4 x double> @llvm.nearbyint.v4f64(<4 x double> %p)
|
D | qpx-recipest.ll | 6 declare <4 x double> @llvm.sqrt.v4f64(<4 x double>) 11 %x = call <4 x double> @llvm.sqrt.v4f64(<4 x double> %b) 61 %x = call <4 x double> @llvm.sqrt.v4f64(<4 x double> %b) 148 %r = call <4 x double> @llvm.sqrt.v4f64(<4 x double> %a)
|
D | vec_fmuladd.ll | 10 declare <4 x double> @llvm.fmuladd.v4f64(<4 x double> %val, <4 x double>, <4 x double>) 49 …%fmuladd = call <4 x double> @llvm.fmuladd.v4f64 (<4 x double> %x, <4 x double> %x, <4 x double> %…
|
D | vec_sqrt.ll | 13 declare <4 x double> @llvm.sqrt.v4f64(<4 x double> %val) 64 %sqrt = call <4 x double> @llvm.sqrt.v4f64 (<4 x double> %x)
|
/external/llvm/lib/Target/X86/ |
D | X86TargetTransformInfo.cpp | 443 {ISD::VECTOR_SHUFFLE, MVT::v4f64, 1}, // vblendpd in getShuffleCost() 536 { ISD::UINT_TO_FP, MVT::v4f64, MVT::v4i64, 1 }, in getCastInstrCost() 543 { ISD::FP_TO_UINT, MVT::v4i64, MVT::v4f64, 1 }, in getCastInstrCost() 594 { ISD::UINT_TO_FP, MVT::v4f64, MVT::v4i8, 2 }, in getCastInstrCost() 595 { ISD::UINT_TO_FP, MVT::v4f64, MVT::v4i16, 2 }, in getCastInstrCost() 596 { ISD::UINT_TO_FP, MVT::v4f64, MVT::v4i32, 1 }, in getCastInstrCost() 601 { ISD::UINT_TO_FP, MVT::v4f64, MVT::v4i64, 12 }, in getCastInstrCost() 675 { ISD::SINT_TO_FP, MVT::v4f64, MVT::v4i1, 3 }, in getCastInstrCost() 676 { ISD::SINT_TO_FP, MVT::v4f64, MVT::v4i8, 3 }, in getCastInstrCost() 677 { ISD::SINT_TO_FP, MVT::v4f64, MVT::v4i16, 3 }, in getCastInstrCost() [all …]
|
D | X86InstrFMA.td | 107 loadv4f64, X86Fmadd, v2f64, v4f64>, VEX_W; 109 loadv4f64, X86Fmsub, v2f64, v4f64>, VEX_W; 112 v2f64, v4f64>, VEX_W; 115 v2f64, v4f64>, VEX_W; 127 loadv4f64, X86Fnmadd, v2f64, v4f64>, VEX_W; 130 v4f64>, VEX_W; 428 defm VFMADDPD4 : fma4p<0x69, "vfmaddpd", X86Fmadd, v2f64, v4f64, 430 defm VFMSUBPD4 : fma4p<0x6D, "vfmsubpd", X86Fmsub, v2f64, v4f64, 432 defm VFNMADDPD4 : fma4p<0x79, "vfnmaddpd", X86Fnmadd, v2f64, v4f64, 434 defm VFNMSUBPD4 : fma4p<0x7D, "vfnmsubpd", X86Fnmsub, v2f64, v4f64, [all …]
|
D | X86CallingConv.td | 62 CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 118 CCIfType<[v8f32, v4f64, v8i32, v4i64], 145 CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 306 CCIfNotVarArg<CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 328 CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 369 CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], CCPassIndirect<i64>>, 411 CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 486 CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 502 CCIfNotVarArg<CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], 521 CCIfNotVarArg<CCIfType<[v32i8, v16i16, v8i32, v4i64, v8f32, v4f64], [all …]
|
/external/llvm/include/llvm/CodeGen/ |
D | MachineValueType.h | 117 v4f64 = 62, // 4 x f64 enumerator 250 return (SimpleTy == MVT::v8f32 || SimpleTy == MVT::v4f64 || in is256BitVector() 359 case v4f64: in getVectorElementType() 403 case v4f64: return 4; in getVectorNumElements() 488 case v4f64: return 256; in getSizeInBits() 651 if (NumElements == 4) return MVT::v4f64; in getVectorVT()
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | llvm.rint.f64.ll | 38 %0 = call <4 x double> @llvm.rint.v4f64(<4 x double> %in) 46 declare <4 x double> @llvm.rint.v4f64(<4 x double>) #0
|
D | fnearbyint.ll | 13 declare <4 x double> @llvm.nearbyint.v4f64(<4 x double>) #0 52 %0 = call <4 x double> @llvm.nearbyint.v4f64(<4 x double> %in)
|
D | fcopysign.f64.ll | 6 declare <4 x double> @llvm.copysign.v4f64(<4 x double>, <4 x double>) nounwind readnone 37 %result = call <4 x double> @llvm.copysign.v4f64(<4 x double> %mag, <4 x double> %sign)
|
D | llvm.round.f64.ll | 50 %result = call <4 x double> @llvm.round.v4f64(<4 x double> %in) #1 67 declare <4 x double> @llvm.round.v4f64(<4 x double>) #1
|
D | fma.f64.ll | 6 declare <4 x double> @llvm.fma.v4f64(<4 x double>, <4 x double>, <4 x double>) nounwind readnone 44 … %r3 = tail call <4 x double> @llvm.fma.v4f64(<4 x double> %r0, <4 x double> %r1, <4 x double> %r2)
|
D | fminnum.f64.ll | 6 declare <4 x double> @llvm.minnum.v4f64(<4 x double>, <4 x double>) #0 33 %val = call <4 x double> @llvm.minnum.v4f64(<4 x double> %a, <4 x double> %b) #0
|
D | fmaxnum.f64.ll | 6 declare <4 x double> @llvm.maxnum.v4f64(<4 x double>, <4 x double>) #0 33 %val = call <4 x double> @llvm.maxnum.v4f64(<4 x double> %a, <4 x double> %b) #0
|
D | fabs.f64.ll | 8 declare <4 x double> @llvm.fabs.v4f64(<4 x double>) readnone 50 %fabs = call <4 x double> @llvm.fabs.v4f64(<4 x double> %in)
|
D | fceil64.ll | 8 declare <4 x double> @llvm.ceil.v4f64(<4 x double>) nounwind readnone 64 %y = call <4 x double> @llvm.ceil.v4f64(<4 x double> %x) nounwind readnone
|
/external/llvm/test/CodeGen/AArch64/ |
D | vector-fcopysign.ll | 147 ;============ v4f64 160 %r = call <4 x double> @llvm.copysign.v4f64(<4 x double> %a, <4 x double> %tmp0) 172 %r = call <4 x double> @llvm.copysign.v4f64(<4 x double> %a, <4 x double> %b) 176 declare <4 x double> @llvm.copysign.v4f64(<4 x double> %a, <4 x double> %b) #0
|