Home
last modified time | relevance | path

Searched refs:v16i8 (Results 1 – 25 of 139) sorted by relevance

123456

/third_party/ffmpeg/libavcodec/mips/
Dhevc_lpf_sao_msa.c50 v16i8 zero = { 0 }; in hevc_loopfilter_luma_hor_msa()
189 dst2 = (v16u8) __msa_pckev_b((v16i8) dst5, (v16i8) dst4); in hevc_loopfilter_luma_hor_msa()
193 dst5 = (v16u8) __msa_pckev_b((v16i8) q2_src, (v16i8) q1_src); in hevc_loopfilter_luma_hor_msa()
347 dst2 = (v16u8) __msa_pckev_b((v16i8) dst5, (v16i8) dst4); in hevc_loopfilter_luma_hor_msa()
425 dst5 = (v16u8) __msa_pckev_b((v16i8) q2_src, (v16i8) delta2); in hevc_loopfilter_luma_hor_msa()
434 dst5 = (v16u8) __msa_pckev_b((v16i8) q2_src, (v16i8) q1_src); in hevc_loopfilter_luma_hor_msa()
473 v16i8 zero = { 0 }; in hevc_loopfilter_luma_ver_msa()
927 v16i8 zero = { 0 }; in hevc_loopfilter_chroma_hor_msa()
974 temp0 = (v8i16) __msa_pckev_b((v16i8) temp1, (v16i8) temp0); in hevc_loopfilter_chroma_hor_msa()
987 v16i8 zero = { 0 }; in hevc_loopfilter_chroma_ver_msa()
[all …]
Dh263dsp_msa.c42 temp0 = (v8i16) __msa_ilvr_b((v16i8) in0, (v16i8) in1); in h263_h_loop_filter_msa()
44 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_h_loop_filter_msa()
80 in0 = (v16u8) ((v16i8) in0 - (v16i8) diff6); in h263_h_loop_filter_msa()
81 in1 = (v16u8) ((v16i8) in1 + (v16i8) diff6); in h263_h_loop_filter_msa()
83 in3 = (v16u8) __msa_adds_s_b((v16i8) in3, (v16i8) d0); in h263_h_loop_filter_msa()
85 in2 = __msa_subsus_u_b(in2, (v16i8) d0); in h263_h_loop_filter_msa()
102 temp0 = (v8i16) __msa_ilvr_b((v16i8) in0, (v16i8) in1); in h263_v_loop_filter_msa()
104 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_v_loop_filter_msa()
140 in0 = (v16u8) ((v16i8) in0 - (v16i8) diff6); in h263_v_loop_filter_msa()
141 in1 = (v16u8) ((v16i8) in1 + (v16i8) diff6); in h263_v_loop_filter_msa()
[all …]
Dh264chroma_msa.c36 v16i8 src0, src1; in avc_chroma_hz_2x2_msa()
39 v16i8 mask; in avc_chroma_hz_2x2_msa()
40 v16i8 coeff_vec0 = __msa_fill_b(coeff0); in avc_chroma_hz_2x2_msa()
41 v16i8 coeff_vec1 = __msa_fill_b(coeff1); in avc_chroma_hz_2x2_msa()
53 res = (v8i16) __msa_pckev_b((v16i8) res_r, (v16i8) res_r); in avc_chroma_hz_2x2_msa()
69 v16i8 mask; in avc_chroma_hz_2x4_msa()
70 v16i8 coeff_vec0 = __msa_fill_b(coeff0); in avc_chroma_hz_2x4_msa()
71 v16i8 coeff_vec1 = __msa_fill_b(coeff1); in avc_chroma_hz_2x4_msa()
86 res = (v8i16) __msa_pckev_b((v16i8) res_r, (v16i8) res_r); in avc_chroma_hz_2x4_msa()
105 v16i8 src0, src1; in avc_chroma_hz_4x2_msa()
[all …]
Dvp8_mc_msa.c57 v16i8 vec0_m, vec1_m, vec2_m; \
76 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m; \
91 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \
111 tmp0 = __msa_dotp_s_h((v16i8) vec0, (v16i8) filt0); \
112 tmp0 = __msa_dpadd_s_h(tmp0, (v16i8) vec1, (v16i8) filt1); \
119 v16i8 vec0_m, vec1_m; \
135 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \
147 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \
163 v16i8 src0, src1, src2, src3, filt0, filt1, filt2; in common_hz_6t_4x4_msa()
191 v16i8 src0, src1, src2, src3, filt0, filt1, filt2; in common_hz_6t_4x8_msa()
[all …]
Dhevc_macros_msa.h29 out_m = __msa_dotp_s_h((v16i8) in0, (v16i8) filt0); \
30 out_m = __msa_dpadd_s_h(out_m, (v16i8) in1, (v16i8) filt1); \
50 out_m = __msa_dotp_s_h((v16i8) in0, (v16i8) filt0); \
51 out_m = __msa_dpadd_s_h(out_m, (v16i8) in1, (v16i8) filt1); \
Dvp9_lpf_msa.c28 v16i8 p1_m, p0_m, q0_m, q1_m, q0_sub_p0, filt, filt1, filt2; \
29 const v16i8 cnst4b = __msa_ldi_b(4); \
30 const v16i8 cnst3b = __msa_ldi_b(3); \
32 p1_m = (v16i8) __msa_xori_b(p1_in, 0x80); \
33 p0_m = (v16i8) __msa_xori_b(p0_in, 0x80); \
34 q0_m = (v16i8) __msa_xori_b(q0_in, 0x80); \
35 q1_m = (v16i8) __msa_xori_b(q1_in, 0x80); \
39 filt = filt & (v16i8) hev_in; \
45 filt = filt & (v16i8) mask_in; \
60 filt = filt & (v16i8) hev_in; \
[all …]
Dvp8_lpf_msa.c31 p1_a_sub_q1 = (v16u8) __msa_srli_b((v16i8) p1_a_sub_q1, 1); \
40 v16i8 p1_m, p0_m, q0_m, q1_m, q0_sub_p0, filt_sign; \
41 v16i8 filt, filt1, filt2, cnst4b, cnst3b; \
44 p1_m = (v16i8) __msa_xori_b(p1_in_out, 0x80); \
45 p0_m = (v16i8) __msa_xori_b(p0_in_out, 0x80); \
46 q0_m = (v16i8) __msa_xori_b(q0_in_out, 0x80); \
47 q1_m = (v16i8) __msa_xori_b(q1_in_out, 0x80); \
51 filt = filt & (v16i8) hev_in; \
58 q0_sub_p0_r = __msa_dotp_s_h((v16i8) q0_sub_p0_r, (v16i8) cnst3h); \
64 q0_sub_p0_l = __msa_dotp_s_h((v16i8) q0_sub_p0_l, (v16i8) cnst3h); \
[all …]
Dh264dsp_msa.c41 src0_r = (v8i16) __msa_ilvr_b((v16i8) zero, (v16i8) src0); in avc_wgt_4x2_msa()
47 src0 = (v16u8) __msa_pckev_b((v16i8) tmp0, (v16i8) tmp0); in avc_wgt_4x2_msa()
74 src0 = (v16u8) __msa_pckev_b((v16i8) tmp1, (v16i8) tmp0); in avc_wgt_4x4_msa()
235 v16i8 src_wgt, dst_wgt, wgt, vec0; in avc_biwgt_4x2_msa()
254 vec0 = (v16i8) __msa_ilvr_b((v16i8) dst0, (v16i8) src0); in avc_biwgt_4x2_msa()
259 dst0 = (v16u8) __msa_pckev_b((v16i8) tmp0, (v16i8) tmp0); in avc_biwgt_4x2_msa()
268 v16i8 src_wgt, dst_wgt, wgt, vec0, vec1; in avc_biwgt_4x4_msa()
293 dst0 = (v16u8) __msa_pckev_b((v16i8) tmp1, (v16i8) tmp0); in avc_biwgt_4x4_msa()
302 v16i8 src_wgt, dst_wgt, wgt, vec0, vec1, vec2, vec3; in avc_biwgt_4x8_msa()
342 v16i8 src_wgt, dst_wgt, wgt, vec0, vec1, vec2, vec3; in avc_biwgt_8x4_msa()
[all …]
Dh264qpel_msa.c39 v16i8 tmp0_m, tmp1_m; \
40 v16i8 minus5b_m = __msa_ldi_b(-5); \
41 v16i8 plus20b_m = __msa_ldi_b(20); \
54 v16i8 tmp0_m; \
55 v16i8 minus5b = __msa_ldi_b(-5); \
56 v16i8 plus20b = __msa_ldi_b(20); \
58 tmp0_m = __msa_vshf_b((v16i8) mask0, in1, in0); \
61 tmp0_m = __msa_vshf_b((v16i8) mask1, in1, in0); \
64 tmp0_m = __msa_vshf_b((v16i8) mask2, in1, in0); \
74 out0_m = __msa_dotp_s_h((v16i8) in0, (v16i8) coeff0); \
[all …]
Dhevc_mc_bi_msa.c71 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_bi_copy_4w_msa()
72 v16i8 zero = { 0 }; in hevc_bi_copy_4w_msa()
88 dst0 = (v8i16) __msa_pckev_b((v16i8) dst0, (v16i8) dst0); in hevc_bi_copy_4w_msa()
99 dst0 = (v8i16) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_bi_copy_4w_msa()
140 v16i8 zero = { 0 }; in hevc_bi_copy_6w_msa()
141 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }, src3 = { 0 }; in hevc_bi_copy_6w_msa()
191 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }, src3 = { 0 }; in hevc_bi_copy_8w_msa()
192 v16i8 zero = { 0 }; in hevc_bi_copy_8w_msa()
203 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_bi_copy_8w_msa()
278 v16i8 zero = { 0 }; in hevc_bi_copy_12w_msa()
[all …]
Dhevc_mc_biw_msa.c99 v16i8 zero = { 0 }; in hevc_biwgt_copy_4w_msa()
100 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_biwgt_copy_4w_msa()
128 out0 = (v16u8) __msa_pckev_b((v16i8) dst0, (v16i8) dst0); in hevc_biwgt_copy_4w_msa()
140 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_biwgt_copy_4w_msa()
188 v16i8 zero = { 0 }; in hevc_biwgt_copy_6w_msa()
189 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_biwgt_copy_6w_msa()
241 v16i8 zero = { 0 }; in hevc_biwgt_copy_8w_msa()
242 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }; in hevc_biwgt_copy_8w_msa()
266 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_biwgt_copy_8w_msa()
328 v16i8 zero = { 0 }; in hevc_biwgt_copy_12w_msa()
[all …]
Dhevcdsp_msa.c36 v16i8 zero = { 0 }; in hevc_copy_4w_msa()
39 v16i8 src0, src1; in hevc_copy_4w_msa()
44 src0 = (v16i8) __msa_ilvr_w((v4i32) src1, (v4i32) src0); in hevc_copy_4w_msa()
49 v16i8 src0, src1, src2, src3; in hevc_copy_4w_msa()
60 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_4w_msa()
85 v16i8 zero = { 0 }; in hevc_copy_6w_msa()
86 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_6w_msa()
108 v16i8 zero = { 0 }; in hevc_copy_8w_msa()
111 v16i8 src0, src1; in hevc_copy_8w_msa()
121 v16i8 src0, src1, src2, src3; in hevc_copy_8w_msa()
[all …]
Dhevc_mc_uniw_msa.c67 v16i8 zero = { 0 }; in hevc_uniwgt_copy_4w_msa()
69 v16i8 src0 = { 0 }, src1 = { 0 }; in hevc_uniwgt_copy_4w_msa()
92 out0 = (v16u8) __msa_pckev_b((v16i8) dst0, (v16i8) dst0); in hevc_uniwgt_copy_4w_msa()
101 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_uniwgt_copy_4w_msa()
135 v16i8 zero = { 0 }; in hevc_uniwgt_copy_6w_msa()
137 v16i8 src0, src1, src2, src3; in hevc_uniwgt_copy_6w_msa()
197 v16i8 src0 = { 0 }, src1 = { 0 }, src2 = { 0 }, src3 = { 0 }; in hevc_uniwgt_copy_8w_msa()
198 v16i8 zero = { 0 }; in hevc_uniwgt_copy_8w_msa()
215 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_uniwgt_copy_8w_msa()
292 v16i8 src0, src1, src2, src3; in hevc_uniwgt_copy_12w_msa()
[all …]
Dme_cmp_msa.c236 v16i8 mask = { 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7, 8 }; in sad_hv_bilinear_filter_8width_msa()
253 comp0 = (v8u16) __msa_pckev_b((v16i8) comp0, (v16i8) comp0); in sad_hv_bilinear_filter_8width_msa()
255 temp0 = (v16u8) __msa_vshf_b(mask, (v16i8) ref1, (v16i8) ref1); in sad_hv_bilinear_filter_8width_msa()
259 comp1 = (v8u16) __msa_pckev_b((v16i8) comp1, (v16i8) comp1); in sad_hv_bilinear_filter_8width_msa()
264 temp1 = (v16u8) __msa_vshf_b(mask, (v16i8) ref2, (v16i8) ref2); in sad_hv_bilinear_filter_8width_msa()
268 comp2 = (v8u16) __msa_pckev_b((v16i8) comp2, (v16i8) comp2); in sad_hv_bilinear_filter_8width_msa()
270 temp0 = (v16u8) __msa_vshf_b(mask, (v16i8) ref3, (v16i8) ref3); in sad_hv_bilinear_filter_8width_msa()
274 comp3 = (v8u16) __msa_pckev_b((v16i8) comp3, (v16i8) comp3); in sad_hv_bilinear_filter_8width_msa()
312 comp = (v16u8) __msa_pckev_b((v16i8) comp1, (v16i8) comp0); in sad_hv_bilinear_filter_16width_msa()
322 comp = (v16u8) __msa_pckev_b((v16i8) comp3, (v16i8) comp2); in sad_hv_bilinear_filter_16width_msa()
[all …]
Dvp9_mc_msa.c57 tmp0 = __msa_dotp_s_h((v16i8) vec0, (v16i8) filt0); \
58 tmp0 = __msa_dpadd_s_h(tmp0, (v16i8) vec1, (v16i8) filt1); \
59 tmp1 = __msa_dotp_s_h((v16i8) vec2, (v16i8) filt2); \
60 tmp1 = __msa_dpadd_s_h(tmp1, (v16i8) vec3, (v16i8) filt3); \
69 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \
88 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \
107 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \
143 tmp_m = (v16u8) __msa_pckev_b((v16i8) in0, (v16i8) in1); \
164 v16i8 src0, src1, src2, src3, filt0, filt1, filt2, filt3; in common_hz_8t_4x4_msa()
192 v16i8 filt0, filt1, filt2, filt3; in common_hz_8t_4x8_msa()
[all …]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/
DPPCInstrAltivec.td33 // Since we canonicalize buildvectors to v16i8, all vnots "-1" operands will be
36 (xor node:$in, (bitconvert (v16i8 immAllOnesV)))>;
79 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
83 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
87 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
91 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
95 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
99 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
105 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
133 (vector_shuffle (v16i8 node:$lhs), node:$rhs), [{
[all …]
DPPCInstrVSX.td1036 def : Pat<(v2f64 (bitconvert v16i8:$A)),
1045 def : Pat<(v16i8 (bitconvert v2f64:$A)),
1054 def : Pat<(v2i64 (bitconvert v16i8:$A)),
1063 def : Pat<(v16i8 (bitconvert v2i64:$A)),
1082 def : Pat<(v16i8 (bitconvert f128:$A)),
1205 def : Pat<(v16i8 (vselect v16i8:$vA, v16i8:$vB, v16i8:$vC)),
1373 [(set v4i32:$XT, (bitconvert (v16i8 immAllOnesV)))]>;
1833 dag LE_VBYTE_PERM_VEC = (v16i8 (LVSL ZERO8, (ANDC8 (LI8 8), $Idx)));
1837 dag LE_VBYTE_PERMUTE = (v16i8 (VPERM $S, $S, LE_VBYTE_PERM_VEC));
1866 (v16i8 (LVSL ZERO8, (RLDICR (ANDC8 (LI8 4), $Idx), 1, 62)));
[all …]
/third_party/openh264/codec/common/mips/
Ddeblock_msa.c48 v16i8 iTc, negiTc, negTc, flags, f; in DeblockLumaLt4V_msa()
56 v16i8 const_1_b = __msa_ldi_b(1); in DeblockLumaLt4V_msa()
60 v16i8 zero = { 0 }; in DeblockLumaLt4V_msa()
61 v16i8 tc = { pTc[0 >> 2], pTc[1 >> 2], pTc[2 >> 2], pTc[3 >> 2], in DeblockLumaLt4V_msa()
98 f = (v16i8)bDetaP0Q0 & (v16i8)bDetaP1P0 & (v16i8)bDetaQ1Q0; in DeblockLumaLt4V_msa()
99 flags = f & (v16i8)bDetaP2P0; in DeblockLumaLt4V_msa()
102 flags = f & (v16i8)bDetaQ2Q0; in DeblockLumaLt4V_msa()
133 flags = (v16i8)__msa_ceq_h(t2, (v8i16)zero); in DeblockLumaLt4V_msa()
139 flags = (v16i8)__msa_ceq_h(t2, (v8i16)zero); in DeblockLumaLt4V_msa()
163 flags = (v16i8)__msa_ceq_h(t2, (v8i16)zero); in DeblockLumaLt4V_msa()
[all …]
/third_party/ffmpeg/libavutil/mips/
Dgeneric_macros_msa.h37 #define LD_SB(...) LD_V(v16i8, __VA_ARGS__)
45 #define ST_SB(...) ST_V(v16i8, __VA_ARGS__)
282 #define LD_SB2(...) LD_V2(v16i8, __VA_ARGS__)
293 #define LD_SB3(...) LD_V3(v16i8, __VA_ARGS__)
301 #define LD_SB4(...) LD_V4(v16i8, __VA_ARGS__)
312 #define LD_SB5(...) LD_V5(v16i8, __VA_ARGS__)
320 #define LD_SB6(...) LD_V6(v16i8, __VA_ARGS__)
331 #define LD_SB7(...) LD_V7(v16i8, __VA_ARGS__)
340 #define LD_SB8(...) LD_V8(v16i8, __VA_ARGS__)
368 #define ST_SB2(...) ST_V2(v16i8, __VA_ARGS__)
[all …]
/third_party/skia/third_party/externals/libpng/mips/
Dfilter_msa_intrinsics.c43 #define MSA_SRLI_B(a, b) __msa_srli_b((v16i8) a, b)
301 out0 = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \
302 out1 = (RTYPE) __msa_ilvr_b((v16i8) in2, (v16i8) in3); \
315 v16i8 zero_m = { 0 }; \
316 out0 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in0, slide_val); \
317 out1 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in1, slide_val); \
323 v16i8 zero_m = { 0 }; \
325 out2 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in2, slide_val); \
348 out0 = (RTYPE) __msa_vshf_b((v16i8) mask0, (v16i8) in1, (v16i8) in0); \
349 out1 = (RTYPE) __msa_vshf_b((v16i8) mask1, (v16i8) in3, (v16i8) in2); \
[all …]
/third_party/skia/third_party/externals/libwebp/src/dsp/
Dmsa_macro.h28 #define SRAI_B(a, b) __msa_srai_b((v16i8)a, b)
50 #define LD_SB(...) LD_B(v16i8, __VA_ARGS__)
62 #define ST_SB(...) ST_B(v16i8, __VA_ARGS__)
225 #define LD_SB2(...) LD_B2(v16i8, __VA_ARGS__)
232 #define LD_SB3(...) LD_B3(v16i8, __VA_ARGS__)
239 #define LD_SB4(...) LD_B4(v16i8, __VA_ARGS__)
247 #define LD_SB8(...) LD_B8(v16i8, __VA_ARGS__)
301 #define ST_SB2(...) ST_B2(v16i8, __VA_ARGS__)
308 #define ST_SB4(...) ST_B4(v16i8, __VA_ARGS__)
417 (RTYPE)__msa_sldi_b((v16i8)in0, (v16i8)in1, slide_val) \
[all …]
/third_party/openh264/codec/common/inc/
Dmsa_macros.h376 out = (RTYPE) __msa_vshf_b((v16i8) mask, (v16i8) in0, (v16i8) in1); \
453 out = (RTYPE) __msa_ilvev_b((v16i8) in0, (v16i8) in1); \
553 out = (RTYPE) __msa_ilvod_b((v16i8) in0, (v16i8) in1); \
653 out = (RTYPE) __msa_ilvl_b((v16i8) in0, (v16i8) in1); \
753 out = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \
943 out = (RTYPE) __msa_splati_b((v16i8) in, idx); \
1023 out = (RTYPE) __msa_pckev_b((v16i8) in0, (v16i8) in1); \
1127 out = (RTYPE) __msa_pckod_b((v16i8) in0, (v16i8) in1); \
1264 out = (RTYPE) __msa_dotp_s_h((v16i8) mult, (v16i8) cnst); \
1385 (v16i8) mult, (v16i8) cnst); \
[all …]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/configs/common/lib/Target/ARM/
DARMGenDAGISel.inc1236 /* 2625*/ /*SwitchType*/ 48, MVT::v16i8,// ->2675
1247 …4i32] } QPR:{ *:[v4i32] }:$Vd, (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]…
1255 …2i64] } QPR:{ *:[v2i64] }:$Vd, (bitconvert:{ *:[v2i64] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]…
2005 /* 4099*/ OPC_CheckType, MVT::v16i8,
2017 …i32] }:$Vm, (xor:{ *:[v4i32] } (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]…
2032 /* 4152*/ OPC_CheckType, MVT::v16i8,
2044 …4i32] } QPR:{ *:[v4i32] }:$Vd, (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]…
2055 /* 4196*/ OPC_CheckType, MVT::v16i8,
2068 …*:[v4i32] } (xor:{ *:[v4i32] } (bitconvert:{ *:[v4i32] } (ARMvmovImm:{ *:[v16i8] } (timm:{ *:[i32]…
2085 /* 4252*/ OPC_CheckType, MVT::v16i8,
[all …]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/configs/common/lib/Target/Mips/
DMipsGenDAGISel.inc458 /* 730*/ OPC_CheckChild1Type, MVT::v16i8,
467 …// Src: (st MSA128BOpnd:{ *:[v16i8] }:$wd, addrimm10:{ *:[iPTR] }:$addr)<<P:Predicate_unindexedsto…
468 // Dst: (ST_B MSA128BOpnd:{ *:[v16i8] }:$wd, addrimm10:{ *:[iPTR] }:$addr)
1295 /* 2297*/ /*SwitchType*/ 14, MVT::v16i8,// ->2313
1300 MVT::v16i8, 2/*#Ops*/, 2, 3,
1301 …// Src: (ld:{ *:[v16i8] } addrimm10:{ *:[iPTR] }:$addr)<<P:Predicate_unindexedload>><<P:Predicate_…
1302 // Dst: (LD_B:{ *:[v16i8] } addrimm10:{ *:[iPTR] }:$addr)
8369 MVT::v16i8, 2/*#Ops*/, 0, 1,
8370 …// Src: (intrinsic_wo_chain:{ *:[v16i8] } 4175:{ *:[iPTR] }, MSA128BOpnd:{ *:[v16i8] }:$ws, (timm:…
8371 // Dst: (SAT_S_B:{ *:[v16i8] } MSA128BOpnd:{ *:[v16i8] }:$ws, (timm:{ *:[i32] }):$m)
[all …]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/
DX86TargetTransformInfo.cpp200 { ISD::MUL, MVT::v16i8, 14 }, // extend/pmullw/trunc sequence. in getArithmeticInstrCost()
332 { ISD::SHL, MVT::v16i8, 2 }, // psllw + pand. in getArithmeticInstrCost()
333 { ISD::SRL, MVT::v16i8, 2 }, // psrlw + pand. in getArithmeticInstrCost()
334 { ISD::SRA, MVT::v16i8, 4 }, // psrlw, pand, pxor, psubb. in getArithmeticInstrCost()
408 { ISD::SDIV, MVT::v16i8, 14 }, // 2*ext+2*pmulhw sequence in getArithmeticInstrCost()
409 { ISD::SREM, MVT::v16i8, 16 }, // 2*ext+2*pmulhw+mul+sub sequence in getArithmeticInstrCost()
412 { ISD::UDIV, MVT::v16i8, 14 }, // 2*ext+2*pmulhw sequence in getArithmeticInstrCost()
413 { ISD::UREM, MVT::v16i8, 16 }, // 2*ext+2*pmulhw+mul+sub sequence in getArithmeticInstrCost()
516 { ISD::MUL, MVT::v16i8, 4 }, // extend/pmullw/trunc sequence. in getArithmeticInstrCost()
537 { ISD::MUL, MVT::v16i8, 5 }, // extend/pmullw/trunc sequence. in getArithmeticInstrCost()
[all …]

123456