/external/libyuv/files/source/ |
D | row_msa.cc | 48 out_y = (v16u8)__msa_insert_d((v2i64)zero_m, 0, (int64)y_m); \ 49 out_u = (v16u8)__msa_insert_w(zero_m, 0, (int32)u_m); \ 50 out_v = (v16u8)__msa_insert_w(zero_m, 0, (int32)v_m); \ 127 v16u8 dst0_m, dst1_m; \ 130 dst0_m = (v16u8)__msa_ilvr_h(vec1_m, vec0_m); \ 131 dst1_m = (v16u8)__msa_ilvl_h(vec1_m, vec0_m); \ 139 v16u8 vec0_m, vec1_m, vec2_m, vec3_m; \ 142 vec0_m = (v16u8)__msa_pckev_h((v8i16)argb1, (v8i16)argb0); \ 143 vec1_m = (v16u8)__msa_pckev_h((v8i16)argb3, (v8i16)argb2); \ 144 vec2_m = (v16u8)__msa_pckod_h((v8i16)argb1, (v8i16)argb0); \ [all …]
|
D | scale_msa.cc | 29 v16u8 src0, src1, dst0; in ScaleARGBRowDown2_MSA() 33 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb, 0); in ScaleARGBRowDown2_MSA() 34 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb, 16); in ScaleARGBRowDown2_MSA() 35 dst0 = (v16u8)__msa_pckod_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2_MSA() 47 v16u8 src0, src1, vec0, vec1, dst0; in ScaleARGBRowDown2Linear_MSA() 51 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb, 0); in ScaleARGBRowDown2Linear_MSA() 52 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb, 16); in ScaleARGBRowDown2Linear_MSA() 53 vec0 = (v16u8)__msa_pckev_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2Linear_MSA() 54 vec1 = (v16u8)__msa_pckod_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2Linear_MSA() 55 dst0 = (v16u8)__msa_aver_u_b((v16u8)vec0, (v16u8)vec1); in ScaleARGBRowDown2Linear_MSA() [all …]
|
D | rotate_msa.cc | 24 out0 = (v16u8)__msa_ilvr_b((v16i8)in1, (v16i8)in0); \ 25 out1 = (v16u8)__msa_ilvl_b((v16i8)in1, (v16i8)in0); \ 26 out2 = (v16u8)__msa_ilvr_b((v16i8)in3, (v16i8)in2); \ 27 out3 = (v16u8)__msa_ilvl_b((v16i8)in3, (v16i8)in2); \ 32 out0 = (v16u8)__msa_ilvr_h((v8i16)in1, (v8i16)in0); \ 33 out1 = (v16u8)__msa_ilvl_h((v8i16)in1, (v8i16)in0); \ 34 out2 = (v16u8)__msa_ilvr_h((v8i16)in3, (v8i16)in2); \ 35 out3 = (v16u8)__msa_ilvl_h((v8i16)in3, (v8i16)in2); \ 40 out0 = (v16u8)__msa_ilvr_w((v4i32)in1, (v4i32)in0); \ 41 out1 = (v16u8)__msa_ilvl_w((v4i32)in1, (v4i32)in0); \ [all …]
|
/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
D | denoising_msa.c | 29 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in vp8_denoiser_filter_msa() 30 v16u8 src8, src9, src10, src11, src12, src13, src14, src15; in vp8_denoiser_filter_msa() 31 v16u8 mc_running_avg_y0, running_avg_y, sig0; in vp8_denoiser_filter_msa() 32 v16u8 mc_running_avg_y1, running_avg_y1, sig1; in vp8_denoiser_filter_msa() 33 v16u8 coeff0, coeff1; in vp8_denoiser_filter_msa() 109 adjust0 = (v8i16)__msa_bmnz_v((v16u8)adjust0, (v16u8)diff0, (v16u8)temp2_h); in vp8_denoiser_filter_msa() 110 adjust1 = (v8i16)__msa_bmnz_v((v16u8)adjust1, (v16u8)diff1, (v16u8)temp3_h); in vp8_denoiser_filter_msa() 117 running_avg_y = (v16u8)__msa_pckev_b((v16i8)temp1_h, (v16i8)temp0_h); in vp8_denoiser_filter_msa() 119 __msa_bmnz_v(running_avg_y, mc_running_avg_y0, (v16u8)temp2_h); in vp8_denoiser_filter_msa() 157 adjust0 = (v8i16)__msa_bmnz_v((v16u8)adjust0, (v16u8)diff0, (v16u8)temp2_h); in vp8_denoiser_filter_msa() [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | loopfilter_4_msa.c | 18 v16u8 mask, hev, flat, thresh, b_limit, limit; in vpx_lpf_horizontal_4_msa() 19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0, p1_out, p0_out, q0_out, q1_out; in vpx_lpf_horizontal_4_msa() 24 thresh = (v16u8)__msa_fill_b(*thresh_ptr); in vpx_lpf_horizontal_4_msa() 25 b_limit = (v16u8)__msa_fill_b(*b_limit_ptr); in vpx_lpf_horizontal_4_msa() 26 limit = (v16u8)__msa_fill_b(*limit_ptr); in vpx_lpf_horizontal_4_msa() 46 v16u8 mask, hev, flat, thresh0, b_limit0, limit0, thresh1, b_limit1, limit1; in vpx_lpf_horizontal_4_dual_msa() 47 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_4_dual_msa() 52 thresh0 = (v16u8)__msa_fill_b(*thresh0_ptr); in vpx_lpf_horizontal_4_dual_msa() 53 thresh1 = (v16u8)__msa_fill_b(*thresh1_ptr); in vpx_lpf_horizontal_4_dual_msa() 54 thresh0 = (v16u8)__msa_ilvr_d((v2i64)thresh1, (v2i64)thresh0); in vpx_lpf_horizontal_4_dual_msa() [all …]
|
D | loopfilter_8_msa.c | 18 v16u8 mask, hev, flat, thresh, b_limit, limit; in vpx_lpf_horizontal_8_msa() 19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_8_msa() 20 v16u8 p2_out, p1_out, p0_out, q0_out, q1_out, q2_out; in vpx_lpf_horizontal_8_msa() 28 thresh = (v16u8)__msa_fill_b(*thresh_ptr); in vpx_lpf_horizontal_8_msa() 29 b_limit = (v16u8)__msa_fill_b(*b_limit_ptr); in vpx_lpf_horizontal_8_msa() 30 limit = (v16u8)__msa_fill_b(*limit_ptr); in vpx_lpf_horizontal_8_msa() 37 flat = (v16u8)__msa_ilvr_d((v2i64)zero, (v2i64)flat); in vpx_lpf_horizontal_8_msa() 57 p2_out = __msa_bmnz_v(p2, (v16u8)p2_filter8, flat); in vpx_lpf_horizontal_8_msa() 58 p1_out = __msa_bmnz_v(p1_out, (v16u8)p1_filter8, flat); in vpx_lpf_horizontal_8_msa() 59 p0_out = __msa_bmnz_v(p0_out, (v16u8)p0_filter8, flat); in vpx_lpf_horizontal_8_msa() [all …]
|
D | loopfilter_16_msa.c | 18 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_hz_lpf_t4_and_t8_16w() 19 v16u8 p2_out, p1_out, p0_out, q0_out, q1_out, q2_out; in vpx_hz_lpf_t4_and_t8_16w() 20 v16u8 flat, mask, hev, thresh, b_limit, limit; in vpx_hz_lpf_t4_and_t8_16w() 25 v16u8 zero = { 0 }; in vpx_hz_lpf_t4_and_t8_16w() 30 thresh = (v16u8)__msa_fill_b(*thresh_ptr); in vpx_hz_lpf_t4_and_t8_16w() 31 b_limit = (v16u8)__msa_fill_b(*b_limit_ptr); in vpx_hz_lpf_t4_and_t8_16w() 32 limit = (v16u8)__msa_fill_b(*limit_ptr); in vpx_hz_lpf_t4_and_t8_16w() 63 p2_out = __msa_bmnz_v(p2, (v16u8)p2_filt8_r, flat); in vpx_hz_lpf_t4_and_t8_16w() 64 p1_out = __msa_bmnz_v(p1_out, (v16u8)p1_filt8_r, flat); in vpx_hz_lpf_t4_and_t8_16w() 65 p0_out = __msa_bmnz_v(p0_out, (v16u8)p0_filt8_r, flat); in vpx_hz_lpf_t4_and_t8_16w() [all …]
|
D | deblock_msa.c | 37 out0 = (v16u8)temp6; \ 38 out2 = (v16u8)temp7; \ 39 out4 = (v16u8)temp8; \ 40 out6 = (v16u8)temp9; \ 41 out9 = (v16u8)__msa_ilvl_d((v2i64)out8, (v2i64)out8); \ 42 out11 = (v16u8)__msa_ilvl_d((v2i64)out10, (v2i64)out10); \ 43 out13 = (v16u8)__msa_ilvl_d((v2i64)out12, (v2i64)out12); \ 44 out15 = (v16u8)__msa_ilvl_d((v2i64)out14, (v2i64)out14); \ 45 out1 = (v16u8)__msa_ilvl_d((v2i64)out0, (v2i64)out0); \ 46 out3 = (v16u8)__msa_ilvl_d((v2i64)out2, (v2i64)out2); \ [all …]
|
D | sub_pixel_variance_msa.c | 23 v16u8 src_l0_m, src_l1_m; \ 47 v16u8 pred, src = { 0 }; in avg_sse_diff_4width_msa() 48 v16u8 ref = { 0 }; in avg_sse_diff_4width_msa() 80 v16u8 src0, src1, src2, src3; in avg_sse_diff_8width_msa() 81 v16u8 ref0, ref1, ref2, ref3; in avg_sse_diff_8width_msa() 82 v16u8 pred0, pred1; in avg_sse_diff_8width_msa() 114 v16u8 src, ref, pred; in avg_sse_diff_16width_msa() 169 v16u8 src0, src1, ref0, ref1, pred0, pred1; in avg_sse_diff_32width_msa() 227 v16u8 src0, src1, ref0, ref1, pred0, pred1; in avg_sse_diff_32x64_msa() 287 v16u8 src0, src1, src2, src3; in avg_sse_diff_64x32_msa() [all …]
|
D | loopfilter_msa.h | 40 q0_out = __msa_xori_b((v16u8)q0_m, 0x80); \ 42 p0_out = __msa_xori_b((v16u8)p0_m, 0x80); \ 47 q1_out = __msa_xori_b((v16u8)q1_m, 0x80); \ 49 p1_out = __msa_xori_b((v16u8)p1_m, 0x80); \ 54 v16u8 tmp_flat4, p2_a_sub_p0, q2_a_sub_q0, p3_a_sub_p0, q3_a_sub_q0; \ 55 v16u8 zero_in = { 0 }; \ 68 flat_out = (tmp_flat4 < (v16u8)flat_out); \ 76 v16u8 tmp_flat5, zero_in = { 0 }; \ 77 v16u8 p4_a_sub_p0, q4_a_sub_q0, p5_a_sub_p0, q5_a_sub_q0; \ 78 v16u8 p6_a_sub_p0, q6_a_sub_q0, p7_a_sub_p0, q7_a_sub_q0; \ [all …]
|
D | macros_msa.h | 20 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__) 31 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__) 298 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__) 306 #define LD_UB3(...) LD_B3(v16u8, __VA_ARGS__) 313 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) 321 #define LD_UB5(...) LD_B5(v16u8, __VA_ARGS__) 337 #define LD_UB8(...) LD_B8(v16u8, __VA_ARGS__) 412 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__) 419 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__) 426 #define ST_UB8(...) ST_B8(v16u8, __VA_ARGS__) [all …]
|
D | intrapred_msa.c | 47 v16u8 src0; in intra_predict_vert_16x16_msa() 60 v16u8 src1, src2; in intra_predict_vert_32x32_msa() 105 v16u8 src0, src1, src2, src3; in intra_predict_horiz_16x16_msa() 114 src0 = (v16u8)__msa_fill_b(inp0); in intra_predict_horiz_16x16_msa() 115 src1 = (v16u8)__msa_fill_b(inp1); in intra_predict_horiz_16x16_msa() 116 src2 = (v16u8)__msa_fill_b(inp2); in intra_predict_horiz_16x16_msa() 117 src3 = (v16u8)__msa_fill_b(inp3); in intra_predict_horiz_16x16_msa() 128 v16u8 src0, src1, src2, src3; in intra_predict_horiz_32x32_msa() 137 src0 = (v16u8)__msa_fill_b(inp0); in intra_predict_horiz_32x32_msa() 138 src1 = (v16u8)__msa_fill_b(inp1); in intra_predict_horiz_32x32_msa() [all …]
|
D | sad_msa.c | 21 #define SAD_INSVE_W4_UB(...) SAD_INSVE_W4(v16u8, __VA_ARGS__) 28 v16u8 src = { 0 }; in sad_4width_msa() 29 v16u8 ref = { 0 }; in sad_4width_msa() 30 v16u8 diff; in sad_4width_msa() 53 v16u8 src0, src1, src2, src3, ref0, ref1, ref2, ref3; in sad_8width_msa() 74 v16u8 src0, src1, ref0, ref1; in sad_16width_msa() 98 v16u8 src0, src1, ref0, ref1; in sad_32width_msa() 135 v16u8 src0, src1, src2, src3; in sad_64width_msa() 136 v16u8 ref0, ref1, ref2, ref3; in sad_64width_msa() 167 v16u8 src = { 0 }; in sad_4width_x3_msa() [all …]
|
D | vpx_convolve_avg_msa.c | 17 v16u8 src0, src1, src2, src3; in avg_width4_msa() 18 v16u8 dst0, dst1, dst2, dst3; in avg_width4_msa() 60 v16u8 src0, src1, src2, src3; in avg_width8_msa() 61 v16u8 dst0, dst1, dst2, dst3; in avg_width8_msa() 83 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in avg_width16_msa() 84 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in avg_width16_msa() 104 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in avg_width32_msa() 105 v16u8 src8, src9, src10, src11, src12, src13, src14, src15; in avg_width32_msa() 106 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in avg_width32_msa() 107 v16u8 dst8, dst9, dst10, dst11, dst12, dst13, dst14, dst15; in avg_width32_msa() [all …]
|
/external/webp/src/dsp/ |
D | dec_msa.c | 138 v16u8 dest0, dest1, dest2, dest3; in TransformAC3() 186 q = __msa_xori_b((v16u8)q_m, 0x80); \ 187 p = __msa_xori_b((v16u8)p_m, 0x80); \ 205 q0 = __msa_xori_b((v16u8)q0_m, 0x80); \ 207 p0 = __msa_xori_b((v16u8)p0_m, 0x80); \ 212 q1 = __msa_xori_b((v16u8)q1_m, 0x80); \ 214 p1 = __msa_xori_b((v16u8)p1_m, 0x80); \ 262 v16u8 p3_asub_p2_m, p2_asub_p1_m, p1_asub_p0_m, q1_asub_q0_m; \ 263 v16u8 p1_asub_q1_m, p0_asub_q0_m, q3_asub_q2_m, q2_asub_q1_m; \ 264 v16u8 flat_out; \ [all …]
|
D | enc_msa.c | 85 v16u8 srcl0, srcl1, src0, src1; in FTransform() 121 tmp4 = (v4i32)__msa_nor_v((v16u8)tmp5, (v16u8)tmp5); in FTransform() 123 tmp5 = (v4i32)__msa_and_v((v16u8)tmp5, (v16u8)tmp4); in FTransform() 263 const v16u8 A = (v16u8)__msa_insert_d((v2i64)A, 0, val_m); in VE4() 264 const v16u8 B = SLDI_UB(A, A, 1); in VE4() 265 const v16u8 C = SLDI_UB(A, A, 2); in VE4() 266 const v16u8 AC = __msa_ave_u_b(A, C); in VE4() 267 const v16u8 B2 = __msa_ave_u_b(B, B); in VE4() 268 const v16u8 R = __msa_aver_u_b(AC, B2); in VE4() 296 const v16u8 A1 = (v16u8)__msa_insert_d((v2i64)A1, 0, val_m); in RD4() [all …]
|
D | upsampling_msa.c | 31 v16u8 t0, t1; \ 60 dst = (v16u8)__msa_pckev_b((v16i8)b1, (v16i8)b0); \ 69 dst = (v16u8)__msa_pckev_b((v16i8)b0, (v16i8)b0); \ 82 dst = (v16u8)__msa_pckev_b((v16i8)a1, (v16i8)a0); \ 92 dst = (v16u8)__msa_pckev_b((v16i8)a0, (v16i8)a0); \ 103 dst = (v16u8)__msa_pckev_b((v16i8)b1, (v16i8)b0); \ 112 dst = (v16u8)__msa_pckev_b((v16i8)b0, (v16i8)b0); \ 116 const v16u8 zero = { 0 }; \ 119 const v16u8 in_y = LD_UB(y); \ 120 const v16u8 in_u = LD_UB(u); \ [all …]
|
D | lossless_msa.c | 25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \ 35 v16u8 src0, src1, src2, dst0, dst1, dst2; \ 46 v16u8 src0, src1, src2, dst0, dst1; \ 55 const v16u8 src0 = LD_UB(psrc); \ 56 const v16u8 dst0 = VSHF_UB(src0, src0, m); \ 117 v16u8 src0, dst0; in ConvertBGRAToRGBA() 118 const v16u8 mask = { 2, 1, 0, 3, 6, 5, 4, 7, 10, 9, 8, 11, 14, 13, 12, 15 }; in ConvertBGRAToRGBA() 121 v16u8 src1, dst1; in ConvertBGRAToRGBA() 157 const v16u8 mask0 = { 0, 1, 2, 4, 5, 6, 8, 9, 10, 12, 13, 14, in ConvertBGRAToBGR() 159 const v16u8 mask1 = { 5, 6, 8, 9, 10, 12, 13, 14, 16, 17, 18, 20, in ConvertBGRAToBGR() [all …]
|
D | filters_msa.c | 25 v16u8 src0, pred0, dst0; in PredictLineInverse0() 28 v16u8 src1, pred1, dst1; in PredictLineInverse0() 105 v16u8 pred0, dst0; in PredictLineGradient() 107 const v16u8 tmp0 = LD_UB(ppred - 1); in PredictLineGradient() 108 const v16u8 tmp1 = LD_UB(ppred - stride); in PredictLineGradient() 109 const v16u8 tmp2 = LD_UB(ppred - stride - 1); in PredictLineGradient() 110 const v16u8 src0 = LD_UB(pinput); in PredictLineGradient() 117 pred0 = (v16u8)__msa_pckev_b((v16i8)a1, (v16i8)a0); in PredictLineGradient()
|
D | msa_macro.h | 32 #define ANDI_B(a, b) __msa_andi_b((v16u8)a, b) 33 #define ORI_B(a, b) __msa_ori_b((v16u8)a, b) 47 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__) 59 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__) 222 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__) 229 #define LD_UB3(...) LD_B3(v16u8, __VA_ARGS__) 236 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) 244 #define LD_UB8(...) LD_B8(v16u8, __VA_ARGS__) 298 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__) 305 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__) [all …]
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | loopfilter_filters_msa.c | 17 v16u8 p1_a_sub_q1, p0_a_sub_q0; \ 21 p1_a_sub_q1 = (v16u8)__msa_srli_b((v16i8)p1_a_sub_q1, 1); \ 24 mask = ((v16u8)mask <= b_limit); \ 50 q0 = __msa_xori_b((v16u8)q0_m, 0x80); \ 52 p0 = __msa_xori_b((v16u8)p0_m, 0x80); \ 57 q1 = __msa_xori_b((v16u8)q1_m, 0x80); \ 59 p1 = __msa_xori_b((v16u8)p1_m, 0x80); \ 86 q0_in = __msa_xori_b((v16u8)q0_m, 0x80); \ 87 p0_in = __msa_xori_b((v16u8)p0_m, 0x80); \ 138 q2 = __msa_xori_b((v16u8)q2_m, 0x80); \ [all …]
|
D | vp8_macros_msa.h | 20 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__) 32 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__) 273 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__) 281 #define LD_UB3(...) LD_B3(v16u8, __VA_ARGS__) 289 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) 297 #define LD_UB5(...) LD_B5(v16u8, __VA_ARGS__) 306 #define LD_UB8(...) LD_B8(v16u8, __VA_ARGS__) 350 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__) 357 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__) 365 #define ST_UB8(...) ST_B8(v16u8, __VA_ARGS__) [all …]
|
D | bilinear_filter_msa.c | 34 v16u8 filt0, vec0, vec1, res0, res1; in common_hz_2t_4x4_msa() 40 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_4x4_msa() 53 v16u8 vec0, vec1, vec2, vec3, filt0; in common_hz_2t_4x8_msa() 61 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_4x8_msa() 89 v16u8 filt0; in common_hz_2t_8x4_msa() 96 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_8x4_msa() 111 v16u8 filt0; in common_hz_2t_8x8mult_msa() 118 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_8x8mult_msa() 185 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() 193 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_16w_msa() [all …]
|
/external/libpng/mips/ |
D | filter_msa_intrinsics.c | 251 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__) 257 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__) 263 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) 266 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__) 272 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__) 278 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__) 307 out0 = (RTYPE) __msa_hsub_u_h((v16u8) in0, (v16u8) in0); \ 308 out1 = (RTYPE) __msa_hsub_u_h((v16u8) in1, (v16u8) in1); \ 318 #define SLDI_B2_0_UB(...) SLDI_B2_0(v16u8, __VA_ARGS__) 326 #define SLDI_B3_0_UB(...) SLDI_B3_0(v16u8, __VA_ARGS__) [all …]
|
/external/libyuv/files/include/libyuv/ |
D | macros_msa.h | 141 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__) 144 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__) 161 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__) 168 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) 181 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__) 188 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__) 215 #define VSHF_B2_UB(...) VSHF_B2(v16u8, __VA_ARGS__) 229 #define ILVRL_B2_UB(...) ILVRL_B2(v16u8, __VA_ARGS__)
|