/external/libvpx/libvpx/third_party/libyuv/source/ |
D | row_msa.cc | 140 v8u16 reg0_m, reg1_m; \ 152 reg0_m = (v8u16)__msa_srai_h((v8i16)reg0_m, shift); \ 153 reg1_m = (v8u16)__msa_srai_h((v8i16)reg1_m, shift); \ 163 v8u16 reg0_m, reg1_m, reg2_m, reg3_m, reg4_m, reg5_m, reg6_m, reg7_m; \ 164 v8u16 reg8_m, reg9_m; \ 190 reg8_m = (v8u16)__msa_pckev_d((v2i64)reg4_m, (v2i64)reg0_m); \ 191 reg9_m = (v8u16)__msa_pckev_d((v2i64)reg5_m, (v2i64)reg1_m); \ 192 reg8_m += (v8u16)__msa_pckod_d((v2i64)reg4_m, (v2i64)reg0_m); \ 193 reg9_m += (v8u16)__msa_pckod_d((v2i64)reg5_m, (v2i64)reg1_m); \ 194 reg0_m = (v8u16)__msa_pckev_d((v2i64)reg6_m, (v2i64)reg2_m); \ [all …]
|
D | scale_msa.cc | 78 v8u16 reg0, reg1, reg2, reg3; in ScaleARGBRowDown2Box_MSA() 96 reg0 = (v8u16)__msa_srari_h((v8i16)reg0, 2); in ScaleARGBRowDown2Box_MSA() 97 reg1 = (v8u16)__msa_srari_h((v8i16)reg1, 2); in ScaleARGBRowDown2Box_MSA() 141 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in ScaleARGBRowDownEvenBox_MSA() 169 reg4 = (v8u16)__msa_pckev_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA() 170 reg5 = (v8u16)__msa_pckev_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 171 reg6 = (v8u16)__msa_pckod_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA() 172 reg7 = (v8u16)__msa_pckod_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 175 reg4 = (v8u16)__msa_srari_h((v8i16)reg4, 2); in ScaleARGBRowDownEvenBox_MSA() 176 reg5 = (v8u16)__msa_srari_h((v8i16)reg5, 2); in ScaleARGBRowDownEvenBox_MSA() [all …]
|
/external/libyuv/files/source/ |
D | row_msa.cc | 140 v8u16 reg0_m, reg1_m; \ 152 reg0_m = (v8u16)__msa_srai_h((v8i16)reg0_m, shift); \ 153 reg1_m = (v8u16)__msa_srai_h((v8i16)reg1_m, shift); \ 163 v8u16 reg0_m, reg1_m, reg2_m, reg3_m, reg4_m, reg5_m, reg6_m, reg7_m; \ 164 v8u16 reg8_m, reg9_m; \ 190 reg8_m = (v8u16)__msa_pckev_d((v2i64)reg4_m, (v2i64)reg0_m); \ 191 reg9_m = (v8u16)__msa_pckev_d((v2i64)reg5_m, (v2i64)reg1_m); \ 192 reg8_m += (v8u16)__msa_pckod_d((v2i64)reg4_m, (v2i64)reg0_m); \ 193 reg9_m += (v8u16)__msa_pckod_d((v2i64)reg5_m, (v2i64)reg1_m); \ 194 reg0_m = (v8u16)__msa_pckev_d((v2i64)reg6_m, (v2i64)reg2_m); \ [all …]
|
D | scale_msa.cc | 78 v8u16 reg0, reg1, reg2, reg3; in ScaleARGBRowDown2Box_MSA() 96 reg0 = (v8u16)__msa_srari_h((v8i16)reg0, 2); in ScaleARGBRowDown2Box_MSA() 97 reg1 = (v8u16)__msa_srari_h((v8i16)reg1, 2); in ScaleARGBRowDown2Box_MSA() 141 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in ScaleARGBRowDownEvenBox_MSA() 169 reg4 = (v8u16)__msa_pckev_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA() 170 reg5 = (v8u16)__msa_pckev_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 171 reg6 = (v8u16)__msa_pckod_d((v2i64)reg2, (v2i64)reg0); in ScaleARGBRowDownEvenBox_MSA() 172 reg7 = (v8u16)__msa_pckod_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 175 reg4 = (v8u16)__msa_srari_h((v8i16)reg4, 2); in ScaleARGBRowDownEvenBox_MSA() 176 reg5 = (v8u16)__msa_srari_h((v8i16)reg5, 2); in ScaleARGBRowDownEvenBox_MSA() [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | sad_msa.c | 31 v8u16 sad = { 0 }; in sad_4width_msa() 54 v8u16 sad = { 0 }; in sad_8width_msa() 75 v8u16 sad = { 0 }; in sad_16width_msa() 99 v8u16 sad = { 0 }; in sad_32width_msa() 137 v8u16 sad0 = { 0 }; in sad_64width_msa() 138 v8u16 sad1 = { 0 }; in sad_64width_msa() 170 v8u16 sad0 = { 0 }; in sad_4width_x3_msa() 171 v8u16 sad1 = { 0 }; in sad_4width_x3_msa() 172 v8u16 sad2 = { 0 }; in sad_4width_x3_msa() 209 v8u16 sad0 = { 0 }; in sad_8width_x3_msa() [all …]
|
D | avg_msa.c | 19 v8u16 sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7; in vpx_avg_8x8_msa() 30 sum0 = (v8u16)__msa_pckev_h((v8i16)sum, (v8i16)sum); in vpx_avg_8x8_msa() 42 v8u16 sum0; in vpx_avg_4x4_msa() 51 sum0 = (v8u16)__msa_pckev_h((v8i16)sum1, (v8i16)sum1); in vpx_avg_4x4_msa() 264 v8u16 tmp0_h, tmp1_h, tmp2_h, tmp3_h, tmp4_h, tmp5_h, tmp6_h, tmp7_h; in vpx_satd_msa() 269 tmp0_h = (v8u16)__msa_asub_s_h(src0, zero); in vpx_satd_msa() 270 tmp1_h = (v8u16)__msa_asub_s_h(src1, zero); in vpx_satd_msa() 277 tmp0_h = (v8u16)__msa_asub_s_h(src0, zero); in vpx_satd_msa() 278 tmp1_h = (v8u16)__msa_asub_s_h(src1, zero); in vpx_satd_msa() 279 tmp2_h = (v8u16)__msa_asub_s_h(src2, zero); in vpx_satd_msa() [all …]
|
D | loopfilter_16_msa.c | 23 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in hz_lpf_t4_and_t8_16w() 24 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in hz_lpf_t4_and_t8_16w() 86 v8u16 p7_r_in, p6_r_in, p5_r_in, p4_r_in, p3_r_in, p2_r_in, p1_r_in, p0_r_in; in hz_lpf_t16_16w() 87 v8u16 q7_r_in, q6_r_in, q5_r_in, q4_r_in, q3_r_in, q2_r_in, q1_r_in, q0_r_in; in hz_lpf_t16_16w() 88 v8u16 p7_l_in, p6_l_in, p5_l_in, p4_l_in, p3_l_in, p2_l_in, p1_l_in, p0_l_in; in hz_lpf_t16_16w() 89 v8u16 q7_l_in, q6_l_in, q5_l_in, q4_l_in, q3_l_in, q2_l_in, q1_l_in, q0_l_in; in hz_lpf_t16_16w() 90 v8u16 tmp0_r, tmp1_r, tmp0_l, tmp1_l; in hz_lpf_t16_16w() 114 q0_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q0); in hz_lpf_t16_16w() 133 q0_l_in = (v8u16)__msa_ilvl_b(zero, (v16i8)q0); in hz_lpf_t16_16w() 154 q1_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q1); in hz_lpf_t16_16w() [all …]
|
D | intrapred_msa.c | 158 v8u16 sum_h; in intra_predict_dc_4x4_msa() 179 v8u16 sum_h; in intra_predict_dc_tl_4x4_msa() 208 v8u16 sum_h; in intra_predict_dc_8x8_msa() 234 v8u16 sum_h; in intra_predict_dc_tl_8x8_msa() 267 v8u16 sum_h, sum_top, sum_left; in intra_predict_dc_16x16_msa() 290 v8u16 sum_h; in intra_predict_dc_tl_16x16_msa() 321 v8u16 sum_h, sum_top0, sum_top1, sum_left0, sum_left1; in intra_predict_dc_32x32_msa() 350 v8u16 sum_h, sum_data0, sum_data1; in intra_predict_dc_tl_32x32_msa() 391 v8u16 src_top_left, vec0, vec1, vec2, vec3; in intra_predict_tm_4x4_msa() 393 src_top_left = (v8u16)__msa_fill_h(top_left); in intra_predict_tm_4x4_msa() [all …]
|
D | macros_msa.h | 22 #define LD_UH(...) LD_V(v8u16, __VA_ARGS__) 541 #define SLDI_B3_UH(...) SLDI_B3(v8u16, __VA_ARGS__) 557 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__) 584 #define DOTP_UB2_UH(...) DOTP_UB2(v8u16, __VA_ARGS__) 592 #define DOTP_UB4_UH(...) DOTP_UB4(v8u16, __VA_ARGS__) 730 in0 = (RTYPE)__msa_min_u_h((v8u16)in0, min_vec); \ 731 in1 = (RTYPE)__msa_min_u_h((v8u16)in1, min_vec); \ 733 #define MIN_UH2_UH(...) MIN_UH2(v8u16, __VA_ARGS__) 740 #define MIN_UH4_UH(...) MIN_UH4(v8u16, __VA_ARGS__) 818 res_m = __msa_hadd_u_w((v8u16)in, (v8u16)in); \ [all …]
|
D | sub_pixel_variance_msa.c | 392 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_4width_h_msa() 431 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_h_msa() 472 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_h_msa() 473 v8u16 out0, out1, out2, out3, out4, out5, out6, out7; in sub_pixel_sse_diff_16width_h_msa() 560 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_v_msa() 598 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_v_msa() 599 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_8width_v_msa() 643 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_16width_v_msa() 740 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4; in sub_pixel_sse_diff_4width_hv_msa() 741 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_hv_msa() [all …]
|
D | loopfilter_8_msa.c | 23 v8u16 p3_r, p2_r, p1_r, p0_r, q3_r, q2_r, q1_r, q0_r; in vpx_lpf_horizontal_8_msa() 89 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in vpx_lpf_horizontal_8_dual_msa() 90 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in vpx_lpf_horizontal_8_dual_msa() 160 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in vpx_lpf_vertical_8_msa() 238 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in vpx_lpf_vertical_8_dual_msa() 239 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in vpx_lpf_vertical_8_dual_msa()
|
D | vpx_convolve8_avg_msa.c | 235 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4, tmp0, tmp1, filt; in common_hv_2ht_2vt_and_aver_dst_4x4_msa() 251 hz_out1 = (v8u16)__msa_sldi_b((v16i8)hz_out2, (v16i8)hz_out0, 8); in common_hv_2ht_2vt_and_aver_dst_4x4_msa() 252 hz_out3 = (v8u16)__msa_pckod_d((v2i64)hz_out4, (v2i64)hz_out2); in common_hv_2ht_2vt_and_aver_dst_4x4_msa() 271 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4, hz_out5, hz_out6; in common_hv_2ht_2vt_and_aver_dst_4x8_msa() 272 v8u16 hz_out7, hz_out8, tmp0, tmp1, tmp2, tmp3; in common_hv_2ht_2vt_and_aver_dst_4x8_msa() 295 hz_out7 = (v8u16)__msa_pckod_d((v2i64)hz_out8, (v2i64)hz_out6); in common_hv_2ht_2vt_and_aver_dst_4x8_msa() 329 v8u16 hz_out0, hz_out1, tmp0, tmp1, tmp2, tmp3; in common_hv_2ht_2vt_and_aver_dst_8x4_msa() 375 v8u16 hz_out0, hz_out1, tmp0, tmp1, tmp2, tmp3; in common_hv_2ht_2vt_and_aver_dst_8x8mult_msa() 441 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, tmp0, tmp1; in common_hv_2ht_2vt_and_aver_dst_16w_msa()
|
D | vpx_convolve8_avg_vert_msa.c | 256 v8u16 tmp0, tmp1; in common_vt_2t_and_aver_dst_4x4_msa() 290 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_and_aver_dst_4x8_msa() 336 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_and_aver_dst_8x4_msa() 363 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_and_aver_dst_8x8mult_msa() 422 v8u16 tmp0, tmp1, tmp2, tmp3, filt; in common_vt_2t_and_aver_dst_16w_msa() 472 v8u16 tmp0, tmp1, tmp2, tmp3, filt; in common_vt_2t_and_aver_dst_32w_msa() 544 v8u16 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; in common_vt_2t_and_aver_dst_64w_msa() 545 v8u16 filt; in common_vt_2t_and_aver_dst_64w_msa()
|
/external/libaom/libaom/aom_dsp/mips/ |
D | sad_msa.c | 33 v8u16 sad = { 0 }; in sad_4width_msa() 56 v8u16 sad = { 0 }; in sad_8width_msa() 77 v8u16 sad = { 0 }; in sad_16width_msa() 101 v8u16 sad = { 0 }; in sad_32width_msa() 139 v8u16 sad0 = { 0 }; in sad_64width_msa() 140 v8u16 sad1 = { 0 }; in sad_64width_msa() 175 v8u16 sad0 = { 0 }; in sad_4width_x4d_msa() 176 v8u16 sad1 = { 0 }; in sad_4width_x4d_msa() 177 v8u16 sad2 = { 0 }; in sad_4width_x4d_msa() 178 v8u16 sad3 = { 0 }; in sad_4width_x4d_msa() [all …]
|
D | loopfilter_16_msa.c | 22 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_hz_lpf_t4_and_t8_16w() 23 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in aom_hz_lpf_t4_and_t8_16w() 85 v8u16 p7_r_in, p6_r_in, p5_r_in, p4_r_in, p3_r_in, p2_r_in, p1_r_in, p0_r_in; in aom_hz_lpf_t16_16w() 86 v8u16 q7_r_in, q6_r_in, q5_r_in, q4_r_in, q3_r_in, q2_r_in, q1_r_in, q0_r_in; in aom_hz_lpf_t16_16w() 87 v8u16 p7_l_in, p6_l_in, p5_l_in, p4_l_in, p3_l_in, p2_l_in, p1_l_in, p0_l_in; in aom_hz_lpf_t16_16w() 88 v8u16 q7_l_in, q6_l_in, q5_l_in, q4_l_in, q3_l_in, q2_l_in, q1_l_in, q0_l_in; in aom_hz_lpf_t16_16w() 89 v8u16 tmp0_r, tmp1_r, tmp0_l, tmp1_l; in aom_hz_lpf_t16_16w() 113 q0_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q0); in aom_hz_lpf_t16_16w() 132 q0_l_in = (v8u16)__msa_ilvl_b(zero, (v16i8)q0); in aom_hz_lpf_t16_16w() 153 q1_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q1); in aom_hz_lpf_t16_16w() [all …]
|
D | macros_msa.h | 26 #define LD_UH(...) LD_H(v8u16, __VA_ARGS__) 669 #define SLDI_B3_UH(...) SLDI_B3(v8u16, __VA_ARGS__) 685 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__) 711 #define DOTP_UB2_UH(...) DOTP_UB2(v8u16, __VA_ARGS__) 719 #define DOTP_UB4_UH(...) DOTP_UB4(v8u16, __VA_ARGS__) 857 in0 = (RTYPE)__msa_min_u_h((v8u16)in0, min_vec); \ 858 in1 = (RTYPE)__msa_min_u_h((v8u16)in1, min_vec); \ 860 #define MIN_UH2_UH(...) MIN_UH2(v8u16, __VA_ARGS__) 867 #define MIN_UH4_UH(...) MIN_UH4(v8u16, __VA_ARGS__) 927 res_m = __msa_hadd_u_w((v8u16)in, (v8u16)in); \ [all …]
|
D | intrapred_msa.c | 160 v8u16 sum_h; in intra_predict_dc_4x4_msa() 181 v8u16 sum_h; in intra_predict_dc_tl_4x4_msa() 210 v8u16 sum_h; in intra_predict_dc_8x8_msa() 236 v8u16 sum_h; in intra_predict_dc_tl_8x8_msa() 269 v8u16 sum_h, sum_top, sum_left; in intra_predict_dc_16x16_msa() 292 v8u16 sum_h; in intra_predict_dc_tl_16x16_msa() 323 v8u16 sum_h, sum_top0, sum_top1, sum_left0, sum_left1; in intra_predict_dc_32x32_msa() 352 v8u16 sum_h, sum_data0, sum_data1; in intra_predict_dc_tl_32x32_msa()
|
D | sub_pixel_variance_msa.c | 389 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_4width_h_msa() 428 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_h_msa() 469 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_h_msa() 470 v8u16 out0, out1, out2, out3, out4, out5, out6, out7; in sub_pixel_sse_diff_16width_h_msa() 557 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_v_msa() 595 v8u16 vec0, vec1, vec2, vec3; in sub_pixel_sse_diff_8width_v_msa() 596 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_8width_v_msa() 640 v8u16 tmp0, tmp1, tmp2, tmp3; in sub_pixel_sse_diff_16width_v_msa() 737 v8u16 hz_out0, hz_out1, hz_out2, hz_out3, hz_out4; in sub_pixel_sse_diff_4width_hv_msa() 738 v8u16 tmp0, tmp1; in sub_pixel_sse_diff_4width_hv_msa() [all …]
|
D | loopfilter_8_msa.c | 23 v8u16 p3_r, p2_r, p1_r, p0_r, q3_r, q2_r, q1_r, q0_r; in aom_lpf_horizontal_8_msa() 89 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_lpf_horizontal_8_dual_msa() 90 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in aom_lpf_horizontal_8_dual_msa() 160 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_lpf_vertical_8_msa() 238 v8u16 p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r; in aom_lpf_vertical_8_dual_msa() 239 v8u16 p3_l, p2_l, p1_l, p0_l, q0_l, q1_l, q2_l, q3_l; in aom_lpf_vertical_8_dual_msa()
|
/external/webp/src/dsp/ |
D | msa_macro.h | 53 #define LD_UH(...) LD_H(v8u16, __VA_ARGS__) 65 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__) 259 #define LD_UH2(...) LD_H2(v8u16, __VA_ARGS__) 354 #define ST_UH2(...) ST_H2(v8u16, __VA_ARGS__) 435 #define VSHF_UH(...) VSHF_B(v8u16, __VA_ARGS__) 444 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__) 458 #define VSHF_H2_UH(...) VSHF_H2(v8u16, __VA_ARGS__) 553 const v8u16 max_m = (v8u16)__msa_ldi_h(255); \ 554 in = __msa_maxi_u_h((v8u16) in, 0); \ 555 in = __msa_min_u_h((v8u16) max_m, (v8u16) in); \ [all …]
|
D | upsampling_msa.c | 49 out0 = (v8u16)__msa_pckod_h((v8i16)temp1, (v8i16)temp0); \ 96 const v8u16 const_a = (v8u16)__msa_fill_h(17685); \ 97 const v8u16 a0 = __msa_adds_u_h((v8u16)y0, u0); \ 98 const v8u16 a1 = __msa_adds_u_h((v8u16)y1, u1); \ 99 v8u16 b0 = __msa_subs_u_h(a0, const_a); \ 100 v8u16 b1 = __msa_subs_u_h(a1, const_a); \ 107 const v8u16 const_a = (v8u16)__msa_fill_h(17685); \ 108 const v8u16 a0 = __msa_adds_u_h((v8u16)y0, u0); \ 109 v8u16 b0 = __msa_subs_u_h(a0, const_a); \ 117 v8u16 y0, y1, u0, u1, v0, v1; \ [all …]
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | bilinear_filter_msa.c | 35 v8u16 vec2, vec3, filt; in common_hz_2t_4x4_msa() 56 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() 91 v8u16 vec0, vec1, vec2, vec3, filt; in common_hz_2t_8x4_msa() 113 v8u16 vec0, vec1, vec2, vec3, filt; in common_hz_2t_8x8mult_msa() 186 v8u16 out0, out1, out2, out3, out4, out5, out6, out7, filt; in common_hz_2t_16w_msa() 251 v8u16 tmp0, tmp1; in common_vt_2t_4x4_msa() 274 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_4x8_msa() 316 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_8x4_msa() 339 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_8x8mult_msa() 390 v8u16 tmp0, tmp1, tmp2, tmp3; in common_vt_2t_16w_msa() [all …]
|
D | vp8_macros_msa.h | 24 #define LD_UH(...) LD_H(v8u16, __VA_ARGS__) 36 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__) 540 #define SLDI_B3_UH(...) SLDI_B3(v8u16, __VA_ARGS__) 556 #define VSHF_B2_UH(...) VSHF_B2(v8u16, __VA_ARGS__) 595 #define DOTP_UB2_UH(...) DOTP_UB2(v8u16, __VA_ARGS__) 603 #define DOTP_UB4_UH(...) DOTP_UB4(v8u16, __VA_ARGS__) 812 res_m = __msa_hadd_u_w((v8u16)in, (v8u16)in); \ 833 #define HADD_UB2_UH(...) HADD_UB2(v8u16, __VA_ARGS__) 1016 #define ILVR_B4_UH(...) ILVR_B4(v8u16, __VA_ARGS__) 1090 #define ILVRL_B2_UH(...) ILVRL_B2(v8u16, __VA_ARGS__) [all …]
|
/external/libyuv/files/include/libyuv/ |
D | macros_msa.h | 147 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__) 200 #define ST_UH2(...) ST_H2(v8u16, __VA_ARGS__)
|
/external/libvpx/libvpx/third_party/libyuv/include/libyuv/ |
D | macros_msa.h | 147 #define ST_UH(...) ST_H(v8u16, __VA_ARGS__) 200 #define ST_UH2(...) ST_H2(v8u16, __VA_ARGS__)
|