Home
last modified time | relevance | path

Searched refs:v8u16 (Results 1 – 25 of 40) sorted by relevance

12

/external/libyuv/files/source/
Drow_msa.cc140 v8u16 reg0_m, reg1_m; \
152 reg0_m = (v8u16)__msa_srai_h((v8i16)reg0_m, shift); \
153 reg1_m = (v8u16)__msa_srai_h((v8i16)reg1_m, shift); \
163 v8u16 reg0_m, reg1_m, reg2_m, reg3_m, reg4_m, reg5_m, reg6_m, reg7_m; \
164 v8u16 reg8_m, reg9_m; \
190 reg8_m = (v8u16)__msa_pckev_d((v2i64)reg4_m, (v2i64)reg0_m); \
191 reg9_m = (v8u16)__msa_pckev_d((v2i64)reg5_m, (v2i64)reg1_m); \
192 reg8_m += (v8u16)__msa_pckod_d((v2i64)reg4_m, (v2i64)reg0_m); \
193 reg9_m += (v8u16)__msa_pckod_d((v2i64)reg5_m, (v2i64)reg1_m); \
194 reg0_m = (v8u16)__msa_pckev_d((v2i64)reg6_m, (v2i64)reg2_m); \
[all …]
Dscale_msa.cc70 v8u16 reg0, reg1, reg2, reg3; in ScaleARGBRowDown2Box_MSA()
88 reg0 = (v8u16)__msa_srari_h((v8i16)reg0, 2); in ScaleARGBRowDown2Box_MSA()
89 reg1 = (v8u16)__msa_srari_h((v8i16)reg1, 2); in ScaleARGBRowDown2Box_MSA()
133 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in ScaleARGBRowDownEvenBox_MSA()
161 reg4 = (v8u16)__msa_pckev_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA()
162 reg5 = (v8u16)__msa_pckev_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA()
163 reg6 = (v8u16)__msa_pckod_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA()
164 reg7 = (v8u16)__msa_pckod_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA()
167 reg4 = (v8u16)__msa_srari_h((v8i16)reg4, 2); in ScaleARGBRowDownEvenBox_MSA()
168 reg5 = (v8u16)__msa_srari_h((v8i16)reg5, 2); in ScaleARGBRowDownEvenBox_MSA()
[all …]
/external/libvpx/libvpx/third_party/libyuv/source/
Drow_msa.cc140 v8u16 reg0_m, reg1_m; \
152 reg0_m = (v8u16)__msa_srai_h((v8i16)reg0_m, shift); \
153 reg1_m = (v8u16)__msa_srai_h((v8i16)reg1_m, shift); \
163 v8u16 reg0_m, reg1_m, reg2_m, reg3_m, reg4_m, reg5_m, reg6_m, reg7_m; \
164 v8u16 reg8_m, reg9_m; \
190 reg8_m = (v8u16)__msa_pckev_d((v2i64)reg4_m, (v2i64)reg0_m); \
191 reg9_m = (v8u16)__msa_pckev_d((v2i64)reg5_m, (v2i64)reg1_m); \
192 reg8_m += (v8u16)__msa_pckod_d((v2i64)reg4_m, (v2i64)reg0_m); \
193 reg9_m += (v8u16)__msa_pckod_d((v2i64)reg5_m, (v2i64)reg1_m); \
194 reg0_m = (v8u16)__msa_pckev_d((v2i64)reg6_m, (v2i64)reg2_m); \
[all …]
Dscale_msa.cc78 v8u16 reg0, reg1, reg2, reg3; in ScaleARGBRowDown2Box_MSA()
96 reg0 = (v8u16)__msa_srari_h((v8i16)reg0, 2); in ScaleARGBRowDown2Box_MSA()
97 reg1 = (v8u16)__msa_srari_h((v8i16)reg1, 2); in ScaleARGBRowDown2Box_MSA()
141 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in ScaleARGBRowDownEvenBox_MSA()
169 reg4 = (v8u16)__msa_pckev_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA()
170 reg5 = (v8u16)__msa_pckev_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA()
171 reg6 = (v8u16)__msa_pckod_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA()
172 reg7 = (v8u16)__msa_pckod_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA()
175 reg4 = (v8u16)__msa_srari_h((v8i16)reg4, 2); in ScaleARGBRowDownEvenBox_MSA()
176 reg5 = (v8u16)__msa_srari_h((v8i16)reg5, 2); in ScaleARGBRowDownEvenBox_MSA()
[all …]
/external/libvpx/libvpx/vpx_dsp/mips/
Dsad_msa.c31 v8u16 sad = { 0 }; in sad_4width_msa()
54 v8u16 sad = { 0 }; in sad_8width_msa()
75 v8u16 sad = { 0 }; in sad_16width_msa()
99 v8u16 sad = { 0 }; in sad_32width_msa()
137 v8u16 sad0 = { 0 }; in sad_64width_msa()
138 v8u16 sad1 = { 0 }; in sad_64width_msa()
170 v8u16 sad0 = { 0 }; in sad_4width_x3_msa()
171 v8u16 sad1 = { 0 }; in sad_4width_x3_msa()
172 v8u16 sad2 = { 0 }; in sad_4width_x3_msa()
209 v8u16 sad0 = { 0 }; in sad_8width_x3_msa()
[all …]
Davg_msa.c18 v8u16 sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7; in vpx_avg_8x8_msa()
29 sum0 = (v8u16)__msa_pckev_h((v8i16)sum, (v8i16)sum); in vpx_avg_8x8_msa()
41 v8u16 sum0; in vpx_avg_4x4_msa()
50 sum0 = (v8u16)__msa_pckev_h((v8i16)sum1, (v8i16)sum1); in vpx_avg_4x4_msa()
262 v8u16 tmp0_h, tmp1_h, tmp2_h, tmp3_h, tmp4_h, tmp5_h, tmp6_h, tmp7_h; in vpx_satd_msa()
267 tmp0_h = (v8u16)__msa_asub_s_h(src0, zero); in vpx_satd_msa()
268 tmp1_h = (v8u16)__msa_asub_s_h(src1, zero); in vpx_satd_msa()
275 tmp0_h = (v8u16)__msa_asub_s_h(src0, zero); in vpx_satd_msa()
276 tmp1_h = (v8u16)__msa_asub_s_h(src1, zero); in vpx_satd_msa()
277 tmp2_h = (v8u16)__msa_asub_s_h(src2, zero); in vpx_satd_msa()
[all …]
Dloopfilter_16_msa.c23 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in hz_lpf_t4_and_t8_16w()
24 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in hz_lpf_t4_and_t8_16w()
86 v8u16 p7_r_in, p6_r_in, p5_r_in, p4_r_in, p3_r_in, p2_r_in, p1_r_in, p0_r_in; in hz_lpf_t16_16w()
87 v8u16 q7_r_in, q6_r_in, q5_r_in, q4_r_in, q3_r_in, q2_r_in, q1_r_in, q0_r_in; in hz_lpf_t16_16w()
88 v8u16 p7_l_in, p6_l_in, p5_l_in, p4_l_in, p3_l_in, p2_l_in, p1_l_in, p0_l_in; in hz_lpf_t16_16w()
89 v8u16 q7_l_in, q6_l_in, q5_l_in, q4_l_in, q3_l_in, q2_l_in, q1_l_in, q0_l_in; in hz_lpf_t16_16w()
90 v8u16 tmp0_r, tmp1_r, tmp0_l, tmp1_l; in hz_lpf_t16_16w()
114 q0_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q0); in hz_lpf_t16_16w()
133 q0_l_in = (v8u16)__msa_ilvl_b(zero, (v16i8)q0); in hz_lpf_t16_16w()
154 q1_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q1); in hz_lpf_t16_16w()
[all …]
Dintrapred_msa.c158 v8u16 sum_h; in intra_predict_dc_4x4_msa()
179 v8u16 sum_h; in intra_predict_dc_tl_4x4_msa()
208 v8u16 sum_h; in intra_predict_dc_8x8_msa()
234 v8u16 sum_h; in intra_predict_dc_tl_8x8_msa()
267 v8u16 sum_h, sum_top, sum_left; in intra_predict_dc_16x16_msa()
290 v8u16 sum_h; in intra_predict_dc_tl_16x16_msa()
321 v8u16 sum_h, sum_top0, sum_top1, sum_left0, sum_left1; in intra_predict_dc_32x32_msa()
350 v8u16 sum_h, sum_data0, sum_data1; in intra_predict_dc_tl_32x32_msa()
391 v8u16 src_top_left, vec0, vec1, vec2, vec3; in intra_predict_tm_4x4_msa()
393 src_top_left = (v8u16)__msa_fill_h(top_left); in intra_predict_tm_4x4_msa()
[all …]
Dmacros_msa.h22 #define LD_UH(...) LD_V(v8u16, __VA_ARGS__)
541 #define SLDI_B3_UH(...) SLDI_B3(v8u16, __VA_ARGS__)
557 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__)
584 #define DOTP_UB2_UH(...) DOTP_UB2(v8u16, __VA_ARGS__)
592 #define DOTP_UB4_UH(...) DOTP_UB4(v8u16, __VA_ARGS__)
730 in0 = (RTYPE)__msa_min_u_h((v8u16)in0, min_vec); \
731 in1 = (RTYPE)__msa_min_u_h((v8u16)in1, min_vec); \
733 #define MIN_UH2_UH(...) MIN_UH2(v8u16, __VA_ARGS__)
740 #define MIN_UH4_UH(...) MIN_UH4(v8u16, __VA_ARGS__)
818 res_m = __msa_hadd_u_w((v8u16)in, (v8u16)in); \
[all …]
Dsub_pixel_variance_msa.c391 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_4width_h_msa()
430 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_h_msa()
471 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_h_msa()
472 v8u16 out0, out1, out2, out3, out4, out5, out6, out7; in sub_pixel_sse_diff_16width_h_msa()
559 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_v_msa()
597 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_v_msa()
598 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_8width_v_msa()
642 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_16width_v_msa()
739 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4; in sub_pixel_sse_diff_4width_hv_msa()
740 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_hv_msa()
[all …]
Dvpx_convolve8_avg_msa.c235 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4, tmp0, tmp1, filt; in common_hv_2ht_2vt_and_aver_dst_4x4_msa()
251 hz_out1 = (v8u16)__msa_sldi_b((v16i8)hz_out2, (v16i8)hz_out0, 8); in common_hv_2ht_2vt_and_aver_dst_4x4_msa()
252 hz_out3 = (v8u16)__msa_pckod_d((v2i64)hz_out4, (v2i64)hz_out2); in common_hv_2ht_2vt_and_aver_dst_4x4_msa()
271 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4, hz_out5, hz_out6; in common_hv_2ht_2vt_and_aver_dst_4x8_msa()
272 v8u16 hz_out7, hz_out8, tmp0, tmp1, tmp2, tmp3; in common_hv_2ht_2vt_and_aver_dst_4x8_msa()
295 hz_out7 = (v8u16)__msa_pckod_d((v2i64)hz_out8, (v2i64)hz_out6); in common_hv_2ht_2vt_and_aver_dst_4x8_msa()
329 v8u16 hz_out0, hz_out1, tmp0, tmp1, tmp2, tmp3; in common_hv_2ht_2vt_and_aver_dst_8x4_msa()
375 v8u16 hz_out0, hz_out1, tmp0, tmp1, tmp2, tmp3; in common_hv_2ht_2vt_and_aver_dst_8x8mult_msa()
441 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, tmp0, tmp1; in common_hv_2ht_2vt_and_aver_dst_16w_msa()
Dloopfilter_8_msa.c23 v8u16 p3_r, p2_r, p1_r, p0_r, q3_r, q2_r, q1_r, q0_r; in vpx_lpf_horizontal_8_msa()
89 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in vpx_lpf_horizontal_8_dual_msa()
90 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in vpx_lpf_horizontal_8_dual_msa()
160 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in vpx_lpf_vertical_8_msa()
238 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in vpx_lpf_vertical_8_dual_msa()
239 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in vpx_lpf_vertical_8_dual_msa()
Dvpx_convolve8_avg_vert_msa.c256 v8u16 tmp0, tmp1; in common_vt_2t_and_aver_dst_4x4_msa()
290 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_and_aver_dst_4x8_msa()
336 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_and_aver_dst_8x4_msa()
363 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_and_aver_dst_8x8mult_msa()
422 v8u16 tmp0, tmp1, tmp2, tmp3, filt; in common_vt_2t_and_aver_dst_16w_msa()
472 v8u16 tmp0, tmp1, tmp2, tmp3, filt; in common_vt_2t_and_aver_dst_32w_msa()
544 v8u16 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; in common_vt_2t_and_aver_dst_64w_msa()
545 v8u16 filt; in common_vt_2t_and_aver_dst_64w_msa()
/external/libaom/libaom/aom_dsp/mips/
Dsad_msa.c33 v8u16 sad = { 0 }; in sad_4width_msa()
56 v8u16 sad = { 0 }; in sad_8width_msa()
77 v8u16 sad = { 0 }; in sad_16width_msa()
101 v8u16 sad = { 0 }; in sad_32width_msa()
139 v8u16 sad0 = { 0 }; in sad_64width_msa()
140 v8u16 sad1 = { 0 }; in sad_64width_msa()
175 v8u16 sad0 = { 0 }; in sad_4width_x4d_msa()
176 v8u16 sad1 = { 0 }; in sad_4width_x4d_msa()
177 v8u16 sad2 = { 0 }; in sad_4width_x4d_msa()
178 v8u16 sad3 = { 0 }; in sad_4width_x4d_msa()
[all …]
Dloopfilter_16_msa.c22 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_hz_lpf_t4_and_t8_16w()
23 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in aom_hz_lpf_t4_and_t8_16w()
85 v8u16 p7_r_in, p6_r_in, p5_r_in, p4_r_in, p3_r_in, p2_r_in, p1_r_in, p0_r_in; in aom_hz_lpf_t16_16w()
86 v8u16 q7_r_in, q6_r_in, q5_r_in, q4_r_in, q3_r_in, q2_r_in, q1_r_in, q0_r_in; in aom_hz_lpf_t16_16w()
87 v8u16 p7_l_in, p6_l_in, p5_l_in, p4_l_in, p3_l_in, p2_l_in, p1_l_in, p0_l_in; in aom_hz_lpf_t16_16w()
88 v8u16 q7_l_in, q6_l_in, q5_l_in, q4_l_in, q3_l_in, q2_l_in, q1_l_in, q0_l_in; in aom_hz_lpf_t16_16w()
89 v8u16 tmp0_r, tmp1_r, tmp0_l, tmp1_l; in aom_hz_lpf_t16_16w()
113 q0_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q0); in aom_hz_lpf_t16_16w()
132 q0_l_in = (v8u16)__msa_ilvl_b(zero, (v16i8)q0); in aom_hz_lpf_t16_16w()
153 q1_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q1); in aom_hz_lpf_t16_16w()
[all …]
Dmacros_msa.h26 #define LD_UH(...) LD_H(v8u16, __VA_ARGS__)
669 #define SLDI_B3_UH(...) SLDI_B3(v8u16, __VA_ARGS__)
685 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__)
711 #define DOTP_UB2_UH(...) DOTP_UB2(v8u16, __VA_ARGS__)
719 #define DOTP_UB4_UH(...) DOTP_UB4(v8u16, __VA_ARGS__)
857 in0 = (RTYPE)__msa_min_u_h((v8u16)in0, min_vec); \
858 in1 = (RTYPE)__msa_min_u_h((v8u16)in1, min_vec); \
860 #define MIN_UH2_UH(...) MIN_UH2(v8u16, __VA_ARGS__)
867 #define MIN_UH4_UH(...) MIN_UH4(v8u16, __VA_ARGS__)
927 res_m = __msa_hadd_u_w((v8u16)in, (v8u16)in); \
[all …]
Dsub_pixel_variance_msa.c389 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_4width_h_msa()
428 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_h_msa()
469 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_h_msa()
470 v8u16 out0, out1, out2, out3, out4, out5, out6, out7; in sub_pixel_sse_diff_16width_h_msa()
557 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_v_msa()
595 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_v_msa()
596 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_8width_v_msa()
640 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_16width_v_msa()
737 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4; in sub_pixel_sse_diff_4width_hv_msa()
738 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_hv_msa()
[all …]
Dintrapred_msa.c160 v8u16 sum_h; in intra_predict_dc_4x4_msa()
181 v8u16 sum_h; in intra_predict_dc_tl_4x4_msa()
210 v8u16 sum_h; in intra_predict_dc_8x8_msa()
236 v8u16 sum_h; in intra_predict_dc_tl_8x8_msa()
269 v8u16 sum_h, sum_top, sum_left; in intra_predict_dc_16x16_msa()
292 v8u16 sum_h; in intra_predict_dc_tl_16x16_msa()
323 v8u16 sum_h, sum_top0, sum_top1, sum_left0, sum_left1; in intra_predict_dc_32x32_msa()
352 v8u16 sum_h, sum_data0, sum_data1; in intra_predict_dc_tl_32x32_msa()
Dloopfilter_8_msa.c23 v8u16 p3_r, p2_r, p1_r, p0_r, q3_r, q2_r, q1_r, q0_r; in aom_lpf_horizontal_8_msa()
89 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_lpf_horizontal_8_dual_msa()
90 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in aom_lpf_horizontal_8_dual_msa()
160 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_lpf_vertical_8_msa()
238 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_lpf_vertical_8_dual_msa()
239 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in aom_lpf_vertical_8_dual_msa()
/external/webp/src/dsp/
Dmsa_macro.h53 #define LD_UH(...) LD_H(v8u16, __VA_ARGS__)
65 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__)
259 #define LD_UH2(...) LD_H2(v8u16, __VA_ARGS__)
354 #define ST_UH2(...) ST_H2(v8u16, __VA_ARGS__)
435 #define VSHF_UH(...) VSHF_B(v8u16, __VA_ARGS__)
444 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__)
458 #define VSHF_H2_UH(...) VSHF_H2(v8u16, __VA_ARGS__)
553 const v8u16 max_m = (v8u16)__msa_ldi_h(255); \
554 in = __msa_maxi_u_h((v8u16) in, 0); \
555 in = __msa_min_u_h((v8u16) max_m, (v8u16) in); \
[all …]
Dupsampling_msa.c49 out0 = (v8u16)__msa_pckod_h((v8i16)temp1, (v8i16)temp0); \
96 const v8u16 const_a = (v8u16)__msa_fill_h(17685); \
97 const v8u16 a0 = __msa_adds_u_h((v8u16)y0, u0); \
98 const v8u16 a1 = __msa_adds_u_h((v8u16)y1, u1); \
99 v8u16 b0 = __msa_subs_u_h(a0, const_a); \
100 v8u16 b1 = __msa_subs_u_h(a1, const_a); \
107 const v8u16 const_a = (v8u16)__msa_fill_h(17685); \
108 const v8u16 a0 = __msa_adds_u_h((v8u16)y0, u0); \
109 v8u16 b0 = __msa_subs_u_h(a0, const_a); \
117 v8u16 y0, y1, u0, u1, v0, v1; \
[all …]
/external/libvpx/libvpx/vp8/common/mips/msa/
Dbilinear_filter_msa.c35 v8u16 vec2, vec3, filt; in common_hz_2t_4x4_msa()
56 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa()
91 v8u16 vec0, vec1, vec2, vec3, filt; in common_hz_2t_8x4_msa()
113 v8u16 vec0, vec1, vec2, vec3, filt; in common_hz_2t_8x8mult_msa()
186 v8u16 out0, out1, out2, out3, out4, out5, out6, out7, filt; in common_hz_2t_16w_msa()
251 v8u16 tmp0, tmp1; in common_vt_2t_4x4_msa()
274 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_4x8_msa()
316 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_8x4_msa()
339 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_8x8mult_msa()
390 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_16w_msa()
[all …]
Dvp8_macros_msa.h24 #define LD_UH(...) LD_H(v8u16, __VA_ARGS__)
36 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__)
540 #define SLDI_B3_UH(...) SLDI_B3(v8u16, __VA_ARGS__)
556 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__)
595 #define DOTP_UB2_UH(...) DOTP_UB2(v8u16, __VA_ARGS__)
603 #define DOTP_UB4_UH(...) DOTP_UB4(v8u16, __VA_ARGS__)
812 res_m = __msa_hadd_u_w((v8u16)in, (v8u16)in); \
833 #define HADD_UB2_UH(...) HADD_UB2(v8u16, __VA_ARGS__)
1016 #define ILVR_B4_UH(...) ILVR_B4(v8u16, __VA_ARGS__)
1090 #define ILVRL_B2_UH(...) ILVRL_B2(v8u16, __VA_ARGS__)
[all …]
/external/libvpx/libvpx/third_party/libyuv/include/libyuv/
Dmacros_msa.h147 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__)
200 #define ST_UH2(...) ST_H2(v8u16, __VA_ARGS__)
/external/libyuv/files/include/libyuv/
Dmacros_msa.h147 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__)
200 #define ST_UH2(...) ST_H2(v8u16, __VA_ARGS__)

12