/third_party/ffmpeg/libavcodec/mips/ |
D | hevc_lpf_sao_msa.c | 50 v16i8 zero = { 0 }; in hevc_loopfilter_luma_hor_msa() 189 dst2 = (v16u8) __msa_pckev_b((v16i8) dst5, (v16i8) dst4); in hevc_loopfilter_luma_hor_msa() 193 dst5 = (v16u8) __msa_pckev_b((v16i8) q2_src, (v16i8) q1_src); in hevc_loopfilter_luma_hor_msa() 347 dst2 = (v16u8) __msa_pckev_b((v16i8) dst5, (v16i8) dst4); in hevc_loopfilter_luma_hor_msa() 425 dst5 = (v16u8) __msa_pckev_b((v16i8) q2_src, (v16i8) delta2); in hevc_loopfilter_luma_hor_msa() 434 dst5 = (v16u8) __msa_pckev_b((v16i8) q2_src, (v16i8) q1_src); in hevc_loopfilter_luma_hor_msa() 473 v16i8 zero = { 0 }; in hevc_loopfilter_luma_ver_msa() 927 v16i8 zero = { 0 }; in hevc_loopfilter_chroma_hor_msa() 974 temp0 = (v8i16) __msa_pckev_b((v16i8) temp1, (v16i8) temp0); in hevc_loopfilter_chroma_hor_msa() 987 v16i8 zero = { 0 }; in hevc_loopfilter_chroma_ver_msa() [all …]
|
D | h263dsp_msa.c | 42 temp0 = (v8i16) __msa_ilvr_b((v16i8) in0, (v16i8) in1); in h263_h_loop_filter_msa() 44 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_h_loop_filter_msa() 80 in0 = (v16u8) ((v16i8) in0 - (v16i8) diff6); in h263_h_loop_filter_msa() 81 in1 = (v16u8) ((v16i8) in1 + (v16i8) diff6); in h263_h_loop_filter_msa() 83 in3 = (v16u8) __msa_adds_s_b((v16i8) in3, (v16i8) d0); in h263_h_loop_filter_msa() 85 in2 = __msa_subsus_u_b(in2, (v16i8) d0); in h263_h_loop_filter_msa() 102 temp0 = (v8i16) __msa_ilvr_b((v16i8) in0, (v16i8) in1); in h263_v_loop_filter_msa() 104 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_v_loop_filter_msa() 140 in0 = (v16u8) ((v16i8) in0 - (v16i8) diff6); in h263_v_loop_filter_msa() 141 in1 = (v16u8) ((v16i8) in1 + (v16i8) diff6); in h263_v_loop_filter_msa() [all …]
|
D | h264chroma_msa.c | 36 v16i8 src0, src1; in avc_chroma_hz_2x2_msa() 39 v16i8 mask; in avc_chroma_hz_2x2_msa() 40 v16i8 coeff_vec0 = __msa_fill_b(coeff0); in avc_chroma_hz_2x2_msa() 41 v16i8 coeff_vec1 = __msa_fill_b(coeff1); in avc_chroma_hz_2x2_msa() 53 res = (v8i16) __msa_pckev_b((v16i8) res_r, (v16i8) res_r); in avc_chroma_hz_2x2_msa() 69 v16i8 mask; in avc_chroma_hz_2x4_msa() 70 v16i8 coeff_vec0 = __msa_fill_b(coeff0); in avc_chroma_hz_2x4_msa() 71 v16i8 coeff_vec1 = __msa_fill_b(coeff1); in avc_chroma_hz_2x4_msa() 86 res = (v8i16) __msa_pckev_b((v16i8) res_r, (v16i8) res_r); in avc_chroma_hz_2x4_msa() 105 v16i8 src0, src1; in avc_chroma_hz_4x2_msa() [all …]
|
D | vp8_mc_msa.c | 57 v16i8 vec0_m, vec1_m, vec2_m; \ 76 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m; \ 91 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \ 111 tmp0 = __msa_dotp_s_h((v16i8) vec0, (v16i8) filt0); \ 112 tmp0 = __msa_dpadd_s_h(tmp0, (v16i8) vec1, (v16i8) filt1); \ 119 v16i8 vec0_m, vec1_m; \ 135 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \ 147 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \ 163 v16i8 src0, src1, src2, src3, filt0, filt1, filt2; in common_hz_6t_4x4_msa() 191 v16i8 src0, src1, src2, src3, filt0, filt1, filt2; in common_hz_6t_4x8_msa() [all …]
|
D | hevc_macros_msa.h | 29 out_m = __msa_dotp_s_h((v16i8) in0, (v16i8) filt0); \ 30 out_m = __msa_dpadd_s_h(out_m, (v16i8) in1, (v16i8) filt1); \ 50 out_m = __msa_dotp_s_h((v16i8) in0, (v16i8) filt0); \ 51 out_m = __msa_dpadd_s_h(out_m, (v16i8) in1, (v16i8) filt1); \
|
D | vp9_lpf_msa.c | 28 v16i8 p1_m, p0_m, q0_m, q1_m, q0_sub_p0, filt, filt1, filt2; \ 29 const v16i8 cnst4b = __msa_ldi_b(4); \ 30 const v16i8 cnst3b = __msa_ldi_b(3); \ 32 p1_m = (v16i8) __msa_xori_b(p1_in, 0x80); \ 33 p0_m = (v16i8) __msa_xori_b(p0_in, 0x80); \ 34 q0_m = (v16i8) __msa_xori_b(q0_in, 0x80); \ 35 q1_m = (v16i8) __msa_xori_b(q1_in, 0x80); \ 39 filt = filt & (v16i8) hev_in; \ 45 filt = filt & (v16i8) mask_in; \ 60 filt = filt & (v16i8) hev_in; \ [all …]
|
D | vp8_lpf_msa.c | 31 p1_a_sub_q1 = (v16u8) __msa_srli_b((v16i8) p1_a_sub_q1, 1); \ 40 v16i8 p1_m, p0_m, q0_m, q1_m, q0_sub_p0, filt_sign; \ 41 v16i8 filt, filt1, filt2, cnst4b, cnst3b; \ 44 p1_m = (v16i8) __msa_xori_b(p1_in_out, 0x80); \ 45 p0_m = (v16i8) __msa_xori_b(p0_in_out, 0x80); \ 46 q0_m = (v16i8) __msa_xori_b(q0_in_out, 0x80); \ 47 q1_m = (v16i8) __msa_xori_b(q1_in_out, 0x80); \ 51 filt = filt & (v16i8) hev_in; \ 58 q0_sub_p0_r = __msa_dotp_s_h((v16i8) q0_sub_p0_r, (v16i8) cnst3h); \ 64 q0_sub_p0_l = __msa_dotp_s_h((v16i8) q0_sub_p0_l, (v16i8) cnst3h); \ [all …]
|
D | h264dsp_msa.c | 41 src0_r = (v8i16) __msa_ilvr_b((v16i8) zero, (v16i8) src0); in avc_wgt_4x2_msa() 47 src0 = (v16u8) __msa_pckev_b((v16i8) tmp0, (v16i8) tmp0); in avc_wgt_4x2_msa() 74 src0 = (v16u8) __msa_pckev_b((v16i8) tmp1, (v16i8) tmp0); in avc_wgt_4x4_msa() 235 v16i8 src_wgt, dst_wgt, wgt, vec0; in avc_biwgt_4x2_msa() 254 vec0 = (v16i8) __msa_ilvr_b((v16i8) dst0, (v16i8) src0); in avc_biwgt_4x2_msa() 259 dst0 = (v16u8) __msa_pckev_b((v16i8) tmp0, (v16i8) tmp0); in avc_biwgt_4x2_msa() 268 v16i8 src_wgt, dst_wgt, wgt, vec0, vec1; in avc_biwgt_4x4_msa() 293 dst0 = (v16u8) __msa_pckev_b((v16i8) tmp1, (v16i8) tmp0); in avc_biwgt_4x4_msa() 302 v16i8 src_wgt, dst_wgt, wgt, vec0, vec1, vec2, vec3; in avc_biwgt_4x8_msa() 342 v16i8 src_wgt, dst_wgt, wgt, vec0, vec1, vec2, vec3; in avc_biwgt_8x4_msa() [all …]
|
D | h264qpel_msa.c | 39 v16i8 tmp0_m, tmp1_m; \ 40 v16i8 minus5b_m = __msa_ldi_b(-5); \ 41 v16i8 plus20b_m = __msa_ldi_b(20); \ 54 v16i8 tmp0_m; \ 55 v16i8 minus5b = __msa_ldi_b(-5); \ 56 v16i8 plus20b = __msa_ldi_b(20); \ 58 tmp0_m = __msa_vshf_b((v16i8) mask0, in1, in0); \ 61 tmp0_m = __msa_vshf_b((v16i8) mask1, in1, in0); \ 64 tmp0_m = __msa_vshf_b((v16i8) mask2, in1, in0); \ 74 out0_m = __msa_dotp_s_h((v16i8) in0, (v16i8) coeff0); \ [all …]
|
D | hevc_mc_bi_msa.c | 71 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_bi_copy_4w_msa() 72 v16i8 zero = { 0 }; in hevc_bi_copy_4w_msa() 88 dst0 = (v8i16) __msa_pckev_b((v16i8) dst0, (v16i8) dst0); in hevc_bi_copy_4w_msa() 99 dst0 = (v8i16) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_bi_copy_4w_msa() 140 v16i8 zero = { 0 }; in hevc_bi_copy_6w_msa() 141 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }, src3 = { 0 }; in hevc_bi_copy_6w_msa() 191 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }, src3 = { 0 }; in hevc_bi_copy_8w_msa() 192 v16i8 zero = { 0 }; in hevc_bi_copy_8w_msa() 203 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_bi_copy_8w_msa() 278 v16i8 zero = { 0 }; in hevc_bi_copy_12w_msa() [all …]
|
D | hevc_mc_biw_msa.c | 99 v16i8 zero = { 0 }; in hevc_biwgt_copy_4w_msa() 100 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_biwgt_copy_4w_msa() 128 out0 = (v16u8) __msa_pckev_b((v16i8) dst0, (v16i8) dst0); in hevc_biwgt_copy_4w_msa() 140 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_biwgt_copy_4w_msa() 188 v16i8 zero = { 0 }; in hevc_biwgt_copy_6w_msa() 189 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_biwgt_copy_6w_msa() 241 v16i8 zero = { 0 }; in hevc_biwgt_copy_8w_msa() 242 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }; in hevc_biwgt_copy_8w_msa() 266 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_biwgt_copy_8w_msa() 328 v16i8 zero = { 0 }; in hevc_biwgt_copy_12w_msa() [all …]
|
D | hevcdsp_msa.c | 36 v16i8 zero = { 0 }; in hevc_copy_4w_msa() 39 v16i8 src0, src1; in hevc_copy_4w_msa() 44 src0 = (v16i8) __msa_ilvr_w((v4i32) src1, (v4i32) src0); in hevc_copy_4w_msa() 49 v16i8 src0, src1, src2, src3; in hevc_copy_4w_msa() 60 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_4w_msa() 85 v16i8 zero = { 0 }; in hevc_copy_6w_msa() 86 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_6w_msa() 108 v16i8 zero = { 0 }; in hevc_copy_8w_msa() 111 v16i8 src0, src1; in hevc_copy_8w_msa() 121 v16i8 src0, src1, src2, src3; in hevc_copy_8w_msa() [all …]
|
D | hevc_mc_uniw_msa.c | 67 v16i8 zero = { 0 }; in hevc_uniwgt_copy_4w_msa() 69 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_uniwgt_copy_4w_msa() 92 out0 = (v16u8) __msa_pckev_b((v16i8) dst0, (v16i8) dst0); in hevc_uniwgt_copy_4w_msa() 101 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_uniwgt_copy_4w_msa() 135 v16i8 zero = { 0 }; in hevc_uniwgt_copy_6w_msa() 137 v16i8 src0, src1, src2, src3; in hevc_uniwgt_copy_6w_msa() 197 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }, src3 = { 0 }; in hevc_uniwgt_copy_8w_msa() 198 v16i8 zero = { 0 }; in hevc_uniwgt_copy_8w_msa() 215 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_uniwgt_copy_8w_msa() 292 v16i8 src0, src1, src2, src3; in hevc_uniwgt_copy_12w_msa() [all …]
|
D | me_cmp_msa.c | 236 v16i8 mask = { 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7, 8 }; in sad_hv_bilinear_filter_8width_msa() 253 comp0 = (v8u16) __msa_pckev_b((v16i8) comp0, (v16i8) comp0); in sad_hv_bilinear_filter_8width_msa() 255 temp0 = (v16u8) __msa_vshf_b(mask, (v16i8) ref1, (v16i8) ref1); in sad_hv_bilinear_filter_8width_msa() 259 comp1 = (v8u16) __msa_pckev_b((v16i8) comp1, (v16i8) comp1); in sad_hv_bilinear_filter_8width_msa() 264 temp1 = (v16u8) __msa_vshf_b(mask, (v16i8) ref2, (v16i8) ref2); in sad_hv_bilinear_filter_8width_msa() 268 comp2 = (v8u16) __msa_pckev_b((v16i8) comp2, (v16i8) comp2); in sad_hv_bilinear_filter_8width_msa() 270 temp0 = (v16u8) __msa_vshf_b(mask, (v16i8) ref3, (v16i8) ref3); in sad_hv_bilinear_filter_8width_msa() 274 comp3 = (v8u16) __msa_pckev_b((v16i8) comp3, (v16i8) comp3); in sad_hv_bilinear_filter_8width_msa() 312 comp = (v16u8) __msa_pckev_b((v16i8) comp1, (v16i8) comp0); in sad_hv_bilinear_filter_16width_msa() 322 comp = (v16u8) __msa_pckev_b((v16i8) comp3, (v16i8) comp2); in sad_hv_bilinear_filter_16width_msa() [all …]
|
D | vp9_mc_msa.c | 57 tmp0 = __msa_dotp_s_h((v16i8) vec0, (v16i8) filt0); \ 58 tmp0 = __msa_dpadd_s_h(tmp0, (v16i8) vec1, (v16i8) filt1); \ 59 tmp1 = __msa_dotp_s_h((v16i8) vec2, (v16i8) filt2); \ 60 tmp1 = __msa_dpadd_s_h(tmp1, (v16i8) vec3, (v16i8) filt3); \ 69 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \ 88 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \ 107 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \ 143 tmp_m = (v16u8) __msa_pckev_b((v16i8) in0, (v16i8) in1); \ 164 v16i8 src0, src1, src2, src3, filt0, filt1, filt2, filt3; in common_hz_8t_4x4_msa() 192 v16i8 filt0, filt1, filt2, filt3; in common_hz_8t_4x8_msa() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ |
D | PPCInstrAltivec.td | 33 // Since we canonicalize buildvectors to v16i8, all vnots "-1" operands will be 36 (xor node:$in, (bitconvert (v16i8 immAllOnesV)))>; 79 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 83 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 87 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 91 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 95 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 99 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 105 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ 133 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{ [all …]
|
D | PPCInstrVSX.td | 1036 def : Pat<(v2f64 (bitconvert v16i8:$A)), 1045 def : Pat<(v16i8 (bitconvert v2f64:$A)), 1054 def : Pat<(v2i64 (bitconvert v16i8:$A)), 1063 def : Pat<(v16i8 (bitconvert v2i64:$A)), 1082 def : Pat<(v16i8 (bitconvert f128:$A)), 1205 def : Pat<(v16i8 (vselect v16i8:$vA, v16i8:$vB, v16i8:$vC)), 1373 [(set v4i32:$XT, (bitconvert (v16i8 immAllOnesV)))]>; 1833 dag LE_VBYTE_PERM_VEC = (v16i8 (LVSL ZERO8, (ANDC8 (LI8 8), $Idx))); 1837 dag LE_VBYTE_PERMUTE = (v16i8 (VPERM $S, $S, LE_VBYTE_PERM_VEC)); 1866 (v16i8 (LVSL ZERO8, (RLDICR (ANDC8 (LI8 4), $Idx), 1, 62))); [all …]
|
/third_party/openh264/codec/common/mips/ |
D | deblock_msa.c | 48 v16i8 iTc, negiTc, negTc, flags, f; in DeblockLumaLt4V_msa() 56 v16i8 const_1_b = __msa_ldi_b(1); in DeblockLumaLt4V_msa() 60 v16i8 zero = { 0 }; in DeblockLumaLt4V_msa() 61 v16i8 tc = { pTc[0 >> 2], pTc[1 >> 2], pTc[2 >> 2], pTc[3 >> 2], in DeblockLumaLt4V_msa() 98 f = (v16i8)bDetaP0Q0 & (v16i8)bDetaP1P0 & (v16i8)bDetaQ1Q0; in DeblockLumaLt4V_msa() 99 flags = f & (v16i8)bDetaP2P0; in DeblockLumaLt4V_msa() 102 flags = f & (v16i8)bDetaQ2Q0; in DeblockLumaLt4V_msa() 133 flags = (v16i8)__msa_ceq_h(t2, (v8i16)zero); in DeblockLumaLt4V_msa() 139 flags = (v16i8)__msa_ceq_h(t2, (v8i16)zero); in DeblockLumaLt4V_msa() 163 flags = (v16i8)__msa_ceq_h(t2, (v8i16)zero); in DeblockLumaLt4V_msa() [all …]
|
/third_party/ffmpeg/libavutil/mips/ |
D | generic_macros_msa.h | 37 #define LD_SB(...) LD_V(v16i8, __VA_ARGS__) 45 #define ST_SB(...) ST_V(v16i8, __VA_ARGS__) 282 #define LD_SB2(...) LD_V2(v16i8, __VA_ARGS__) 293 #define LD_SB3(...) LD_V3(v16i8, __VA_ARGS__) 301 #define LD_SB4(...) LD_V4(v16i8, __VA_ARGS__) 312 #define LD_SB5(...) LD_V5(v16i8, __VA_ARGS__) 320 #define LD_SB6(...) LD_V6(v16i8, __VA_ARGS__) 331 #define LD_SB7(...) LD_V7(v16i8, __VA_ARGS__) 340 #define LD_SB8(...) LD_V8(v16i8, __VA_ARGS__) 368 #define ST_SB2(...) ST_V2(v16i8, __VA_ARGS__) [all …]
|
/third_party/skia/third_party/externals/libpng/mips/ |
D | filter_msa_intrinsics.c | 43 #define MSA_SRLI_B(a, b) __msa_srli_b((v16i8) a, b) 301 out0 = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \ 302 out1 = (RTYPE) __msa_ilvr_b((v16i8) in2, (v16i8) in3); \ 315 v16i8 zero_m = { 0 }; \ 316 out0 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in0, slide_val); \ 317 out1 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in1, slide_val); \ 323 v16i8 zero_m = { 0 }; \ 325 out2 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in2, slide_val); \ 348 out0 = (RTYPE) __msa_vshf_b((v16i8) mask0, (v16i8) in1, (v16i8) in0); \ 349 out1 = (RTYPE) __msa_vshf_b((v16i8) mask1, (v16i8) in3, (v16i8) in2); \ [all …]
|
/third_party/skia/third_party/externals/libwebp/src/dsp/ |
D | msa_macro.h | 28 #define SRAI_B(a, b) __msa_srai_b((v16i8)a, b) 50 #define LD_SB(...) LD_B(v16i8, __VA_ARGS__) 62 #define ST_SB(...) ST_B(v16i8, __VA_ARGS__) 225 #define LD_SB2(...) LD_B2(v16i8, __VA_ARGS__) 232 #define LD_SB3(...) LD_B3(v16i8, __VA_ARGS__) 239 #define LD_SB4(...) LD_B4(v16i8, __VA_ARGS__) 247 #define LD_SB8(...) LD_B8(v16i8, __VA_ARGS__) 301 #define ST_SB2(...) ST_B2(v16i8, __VA_ARGS__) 308 #define ST_SB4(...) ST_B4(v16i8, __VA_ARGS__) 417 (RTYPE)__msa_sldi_b((v16i8)in0, (v16i8)in1, slide_val) \ [all …]
|
/third_party/openh264/codec/common/inc/ |
D | msa_macros.h | 376 out = (RTYPE) __msa_vshf_b((v16i8) mask, (v16i8) in0, (v16i8) in1); \ 453 out = (RTYPE) __msa_ilvev_b((v16i8) in0, (v16i8) in1); \ 553 out = (RTYPE) __msa_ilvod_b((v16i8) in0, (v16i8) in1); \ 653 out = (RTYPE) __msa_ilvl_b((v16i8) in0, (v16i8) in1); \ 753 out = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \ 943 out = (RTYPE) __msa_splati_b((v16i8) in, idx); \ 1023 out = (RTYPE) __msa_pckev_b((v16i8) in0, (v16i8) in1); \ 1127 out = (RTYPE) __msa_pckod_b((v16i8) in0, (v16i8) in1); \ 1264 out = (RTYPE) __msa_dotp_s_h((v16i8) mult, (v16i8) cnst); \ 1385 (v16i8) mult, (v16i8) cnst); \ [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/configs/common/lib/Target/ARM/ |
D | ARMGenDAGISel.inc | 1236 /* 2625*/ /*SwitchType*/ 48, MVT::v16i8,// ->2675 1247 …4i32] } QPR:{ *:[v4i32] }:$Vd, (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]… 1255 …2i64] } QPR:{ *:[v2i64] }:$Vd, (bitconvert:{ *:[v2i64] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]… 2005 /* 4099*/ OPC_CheckType, MVT::v16i8, 2017 …i32] }:$Vm, (xor:{ *:[v4i32] } (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]… 2032 /* 4152*/ OPC_CheckType, MVT::v16i8, 2044 …4i32] } QPR:{ *:[v4i32] }:$Vd, (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]… 2055 /* 4196*/ OPC_CheckType, MVT::v16i8, 2068 …*:[v4i32] } (xor:{ *:[v4i32] } (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]… 2085 /* 4252*/ OPC_CheckType, MVT::v16i8, [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/configs/common/lib/Target/Mips/ |
D | MipsGenDAGISel.inc | 458 /* 730*/ OPC_CheckChild1Type, MVT::v16i8, 467 …// Src: (st MSA128BOpnd:{ *:[v16i8] }:$wd, addrimm10:{ *:[iPTR] }:$addr)<<P:Predicate_unindexedsto… 468 // Dst: (ST_B MSA128BOpnd:{ *:[v16i8] }:$wd, addrimm10:{ *:[iPTR] }:$addr) 1295 /* 2297*/ /*SwitchType*/ 14, MVT::v16i8,// ->2313 1300 MVT::v16i8, 2/*#Ops*/, 2, 3, 1301 …// Src: (ld:{ *:[v16i8] } addrimm10:{ *:[iPTR] }:$addr)<<P:Predicate_unindexedload>><<P:Predicate_… 1302 // Dst: (LD_B:{ *:[v16i8] } addrimm10:{ *:[iPTR] }:$addr) 8369 MVT::v16i8, 2/*#Ops*/, 0, 1, 8370 …// Src: (intrinsic_wo_chain:{ *:[v16i8] } 4175:{ *:[iPTR] }, MSA128BOpnd:{ *:[v16i8] }:$ws, (timm:… 8371 // Dst: (SAT_S_B:{ *:[v16i8] } MSA128BOpnd:{ *:[v16i8] }:$ws, (timm:{ *:[i32] }):$m) [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
D | X86TargetTransformInfo.cpp | 200 { ISD::MUL, MVT::v16i8, 14 }, // extend/pmullw/trunc sequence. in getArithmeticInstrCost() 332 { ISD::SHL, MVT::v16i8, 2 }, // psllw + pand. in getArithmeticInstrCost() 333 { ISD::SRL, MVT::v16i8, 2 }, // psrlw + pand. in getArithmeticInstrCost() 334 { ISD::SRA, MVT::v16i8, 4 }, // psrlw, pand, pxor, psubb. in getArithmeticInstrCost() 408 { ISD::SDIV, MVT::v16i8, 14 }, // 2*ext+2*pmulhw sequence in getArithmeticInstrCost() 409 { ISD::SREM, MVT::v16i8, 16 }, // 2*ext+2*pmulhw+mul+sub sequence in getArithmeticInstrCost() 412 { ISD::UDIV, MVT::v16i8, 14 }, // 2*ext+2*pmulhw sequence in getArithmeticInstrCost() 413 { ISD::UREM, MVT::v16i8, 16 }, // 2*ext+2*pmulhw+mul+sub sequence in getArithmeticInstrCost() 516 { ISD::MUL, MVT::v16i8, 4 }, // extend/pmullw/trunc sequence. in getArithmeticInstrCost() 537 { ISD::MUL, MVT::v16i8, 5 }, // extend/pmullw/trunc sequence. in getArithmeticInstrCost() [all …]
|