Lines Matching refs:v8f32
342 def : Pat<(v4f32 (extract_subvector (v8f32 VR256:$src), (iPTR 0))),
343 (v4f32 (EXTRACT_SUBREG (v8f32 VR256:$src), sub_xmm))>;
365 (INSERT_SUBREG (v8f32 (IMPLICIT_DEF)), VR128:$src, sub_xmm)>;
375 def : Pat<(v8f32 (scalar_to_vector FR32:$src)),
423 def : Pat<(v4f64 (bitconvert (v8f32 VR256:$src))), (v4f64 VR256:$src)>;
428 def : Pat<(v8f32 (bitconvert (v8i32 VR256:$src))), (v8f32 VR256:$src)>;
429 def : Pat<(v8f32 (bitconvert (v4i64 VR256:$src))), (v8f32 VR256:$src)>;
430 def : Pat<(v8f32 (bitconvert (v4f64 VR256:$src))), (v8f32 VR256:$src)>;
431 def : Pat<(v8f32 (bitconvert (v32i8 VR256:$src))), (v8f32 VR256:$src)>;
432 def : Pat<(v8f32 (bitconvert (v16i16 VR256:$src))), (v8f32 VR256:$src)>;
433 def : Pat<(v4i64 (bitconvert (v8f32 VR256:$src))), (v4i64 VR256:$src)>;
440 def : Pat<(v32i8 (bitconvert (v8f32 VR256:$src))), (v32i8 VR256:$src)>;
445 def : Pat<(v8i32 (bitconvert (v8f32 VR256:$src))), (v8i32 VR256:$src)>;
448 def : Pat<(v16i16 (bitconvert (v8f32 VR256:$src))), (v16i16 VR256:$src)>;
494 [(set VR256:$dst, (v8f32 immAllZerosV))]>;
511 def : Pat<(bc_v32i8 (v8f32 immAllZerosV)),
515 def : Pat<(bc_v16i16 (v8f32 immAllZerosV)),
519 def : Pat<(bc_v8i32 (v8f32 immAllZerosV)),
523 def : Pat<(bc_v4i64 (v8f32 immAllZerosV)),
646 def : Pat<(v8f32 (X86vzmovl (insert_subvector undef,
676 def : Pat<(v8f32 (X86Movss VR256:$src1, VR256:$src2)),
678 (VMOVSSrr (EXTRACT_SUBREG (v8f32 VR256:$src1), sub_xmm),
679 (EXTRACT_SUBREG (v8f32 VR256:$src2), sub_xmm)),
887 [(alignedstore256 (v8f32 VR256:$src), addr:$dst)],
895 [(store (v8f32 VR256:$src), addr:$dst)],
1048 (v8f32 VR256:$src), (iPTR 0))), addr:$dst),
1067 (v8f32 VR256:$src), (iPTR 0))), addr:$dst),
2084 def : Pat<(v8f32 (sint_to_fp (v8i32 VR256:$src))),
2086 def : Pat<(v8f32 (sint_to_fp (bc_v8i32 (loadv4i64 addr:$src)))),
2089 def : Pat<(v8i32 (fp_to_sint (v8f32 VR256:$src))),
2559 def : Pat<(v8i32 (X86cmpp (v8f32 VR256:$src1), VR256:$src2, imm:$cc)),
2560 (VCMPPSYrri (v8f32 VR256:$src1), (v8f32 VR256:$src2), imm:$cc)>;
2561 def : Pat<(v8i32 (X86cmpp (v8f32 VR256:$src1), (loadv8f32 addr:$src2), imm:$cc)),
2562 (VCMPPSYrmi (v8f32 VR256:$src1), addr:$src2, imm:$cc)>;
2607 defm VSHUFPSY : sse12_shuffle<VR256, f256mem, v8f32,
2707 defm VUNPCKHPSY: sse12_unpack_interleave<0x15, X86Unpckh, v8f32, loadv8f32,
2713 defm VUNPCKLPSY: sse12_unpack_interleave<0x14, X86Unpckl, v8f32, loadv8f32,
2921 VR256, v8f32, f256mem, loadv8f32, SSEPackedSingle, itins, 0>,
2961 [(set VR256:$dst, (OpNode (bc_v4i64 (v8f32 VR256:$src1)),
3053 OpNode, VR256, v8f32, f256mem, loadv8f32,
3464 [(set VR256:$dst, (v8f32 (OpNode VR256:$src)))],
3628 [(alignednontemporalstore (v8f32 VR256:$src),
5114 v8f32, VR256, loadv8f32, f256mem>, VEX, VEX_L;
5116 v8f32, VR256, loadv8f32, f256mem>, VEX, VEX_L;
5308 def : Pat<(v8f32 (X86Addsub (v8f32 VR256:$lhs), (v8f32 VR256:$rhs))),
5310 def : Pat<(v8f32 (X86Addsub (v8f32 VR256:$lhs), (loadv8f32 addr:$rhs))),
5375 defm VHADDPSY : S3D_Int<0x7C, "vhaddps", v8f32, VR256, f256mem,
5377 defm VHSUBPSY : S3D_Int<0x7D, "vhsubps", v8f32, VR256, f256mem,
5782 def : Pat<(v8f32 (X86PAlignr VR256:$src1, VR256:$src2, (i8 imm:$imm))),
6575 def : Pat<(v8f32 (ffloor VR256:$src)),
6577 def : Pat<(v8f32 (fnearbyint VR256:$src)),
6579 def : Pat<(v8f32 (fceil VR256:$src)),
6581 def : Pat<(v8f32 (frint VR256:$src)),
6583 def : Pat<(v8f32 (ftrunc VR256:$src)),
6703 defm VTESTPSY : avx_bittest<0x0E, "vtestps", VR256, f256mem, loadv8f32, v8f32>,
7002 defm VBLENDPSY : SS41I_binop_rmi<0x0C, "vblendps", X86Blendi, v8f32,
7141 def : Pat<(v8f32 (vselect (v8i32 VR256:$mask), (v8f32 VR256:$src1),
7142 (v8f32 VR256:$src2))),
7177 def : Pat<(v8f32 (X86vzmovl (v8f32 VR256:$src))),
7178 (VBLENDPSYrri (v8f32 (AVX_SET0)), VR256:$src, (i8 1))>;
7185 def : Pat<(v8f32 (X86vzmovl (insert_subvector undef,
7825 f32mem, v8f32, loadf32,
7836 v8f32, v4f32, WriteFShuffle256>, VEX_L;
7876 def : Pat<(vinsert128_insert:$ins (v8f32 VR256:$src1), (v4f32 VR128:$src2),
7885 def : Pat<(vinsert128_insert:$ins (v8f32 VR256:$src1), (loadv4f32 addr:$src2),
7953 (v8f32 VR256:$src1),
7960 def : Pat<(store (v4f32 (vextract128_extract:$ext (v8f32 VR256:$src1),
8083 loadv4i64, int_x86_avx_vpermilvar_ps_256, v8f32>, VEX_L;
8093 def : Pat<(v8f32 (X86VPermilpv VR256:$src1, (v8i32 VR256:$src2))),
8095 def : Pat<(v8f32 (X86VPermilpv VR256:$src1, (bc_v8i32 (loadv4i64 addr:$src2)))),
8134 [(set VR256:$dst, (v8f32 (X86VPerm2x128 VR256:$src1, VR256:$src2,
8351 def : Pat<(v8f32 (X86VBroadcast (v8f32 VR256:$src))),
8352 (VBROADCASTSSYrr (v4f32 (EXTRACT_SUBREG (v8f32 VR256:$src),
8363 def : Pat<(v8f32 (X86VBroadcast FR32:$src)),
8419 def : Pat<(v8f32 (X86VBroadcast FR32:$src)),
8420 (VINSERTF128rr (INSERT_SUBREG (v8f32 (IMPLICIT_DEF)),
8471 defm VPERMPS : avx2_perm<0x16, "vpermps", loadv8f32, v8f32, WriteFShuffle256>;
8673 def: Pat<(X86mstore addr:$ptr, (v8i32 VR256:$mask), (v8f32 VR256:$src)),
8685 def: Pat<(v8f32 (masked_load addr:$ptr, (v8i32 VR256:$mask), undef)),
8688 def: Pat<(v8f32 (masked_load addr:$ptr, (v8i32 VR256:$mask),
8692 def: Pat<(v8f32 (masked_load addr:$ptr, (v8i32 VR256:$mask), (v8f32 VR256:$src0))),