Home
last modified time | relevance | path

Searched refs:v16u8 (Results 1 – 25 of 50) sorted by relevance

12

/external/libyuv/files/source/
Drow_msa.cc48 out_y = (v16u8)__msa_insert_d((v2i64)zero_m, 0, (int64)y_m); \
49 out_u = (v16u8)__msa_insert_w(zero_m, 0, (int32)u_m); \
50 out_v = (v16u8)__msa_insert_w(zero_m, 0, (int32)v_m); \
127 v16u8 dst0_m, dst1_m; \
130 dst0_m = (v16u8)__msa_ilvr_h(vec1_m, vec0_m); \
131 dst1_m = (v16u8)__msa_ilvl_h(vec1_m, vec0_m); \
139 v16u8 vec0_m, vec1_m, vec2_m, vec3_m; \
142 vec0_m = (v16u8)__msa_pckev_h((v8i16)argb1, (v8i16)argb0); \
143 vec1_m = (v16u8)__msa_pckev_h((v8i16)argb3, (v8i16)argb2); \
144 vec2_m = (v16u8)__msa_pckod_h((v8i16)argb1, (v8i16)argb0); \
[all …]
Dscale_msa.cc29 v16u8 src0, src1, dst0; in ScaleARGBRowDown2_MSA()
33 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb, 0); in ScaleARGBRowDown2_MSA()
34 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb, 16); in ScaleARGBRowDown2_MSA()
35 dst0 = (v16u8)__msa_pckod_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2_MSA()
47 v16u8 src0, src1, vec0, vec1, dst0; in ScaleARGBRowDown2Linear_MSA()
51 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb, 0); in ScaleARGBRowDown2Linear_MSA()
52 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb, 16); in ScaleARGBRowDown2Linear_MSA()
53 vec0 = (v16u8)__msa_pckev_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2Linear_MSA()
54 vec1 = (v16u8)__msa_pckod_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2Linear_MSA()
55 dst0 = (v16u8)__msa_aver_u_b((v16u8)vec0, (v16u8)vec1); in ScaleARGBRowDown2Linear_MSA()
[all …]
Drotate_msa.cc24 out0 = (v16u8)__msa_ilvr_b((v16i8)in1, (v16i8)in0); \
25 out1 = (v16u8)__msa_ilvl_b((v16i8)in1, (v16i8)in0); \
26 out2 = (v16u8)__msa_ilvr_b((v16i8)in3, (v16i8)in2); \
27 out3 = (v16u8)__msa_ilvl_b((v16i8)in3, (v16i8)in2); \
32 out0 = (v16u8)__msa_ilvr_h((v8i16)in1, (v8i16)in0); \
33 out1 = (v16u8)__msa_ilvl_h((v8i16)in1, (v8i16)in0); \
34 out2 = (v16u8)__msa_ilvr_h((v8i16)in3, (v8i16)in2); \
35 out3 = (v16u8)__msa_ilvl_h((v8i16)in3, (v8i16)in2); \
40 out0 = (v16u8)__msa_ilvr_w((v4i32)in1, (v4i32)in0); \
41 out1 = (v16u8)__msa_ilvl_w((v4i32)in1, (v4i32)in0); \
[all …]
/external/libvpx/libvpx/vp8/encoder/mips/msa/
Ddenoising_msa.c29 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in vp8_denoiser_filter_msa()
30 v16u8 src8, src9, src10, src11, src12, src13, src14, src15; in vp8_denoiser_filter_msa()
31 v16u8 mc_running_avg_y0, running_avg_y, sig0; in vp8_denoiser_filter_msa()
32 v16u8 mc_running_avg_y1, running_avg_y1, sig1; in vp8_denoiser_filter_msa()
33 v16u8 coeff0, coeff1; in vp8_denoiser_filter_msa()
109 adjust0 = (v8i16)__msa_bmnz_v((v16u8)adjust0, (v16u8)diff0, (v16u8)temp2_h); in vp8_denoiser_filter_msa()
110 adjust1 = (v8i16)__msa_bmnz_v((v16u8)adjust1, (v16u8)diff1, (v16u8)temp3_h); in vp8_denoiser_filter_msa()
117 running_avg_y = (v16u8)__msa_pckev_b((v16i8)temp1_h, (v16i8)temp0_h); in vp8_denoiser_filter_msa()
119 __msa_bmnz_v(running_avg_y, mc_running_avg_y0, (v16u8)temp2_h); in vp8_denoiser_filter_msa()
157 adjust0 = (v8i16)__msa_bmnz_v((v16u8)adjust0, (v16u8)diff0, (v16u8)temp2_h); in vp8_denoiser_filter_msa()
[all …]
/external/libvpx/libvpx/vpx_dsp/mips/
Dloopfilter_4_msa.c18 v16u8 mask, hev, flat, thresh, b_limit, limit; in vpx_lpf_horizontal_4_msa()
19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0, p1_out, p0_out, q0_out, q1_out; in vpx_lpf_horizontal_4_msa()
24 thresh = (v16u8)__msa_fill_b(*thresh_ptr); in vpx_lpf_horizontal_4_msa()
25 b_limit = (v16u8)__msa_fill_b(*b_limit_ptr); in vpx_lpf_horizontal_4_msa()
26 limit = (v16u8)__msa_fill_b(*limit_ptr); in vpx_lpf_horizontal_4_msa()
46 v16u8 mask, hev, flat, thresh0, b_limit0, limit0, thresh1, b_limit1, limit1; in vpx_lpf_horizontal_4_dual_msa()
47 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_4_dual_msa()
52 thresh0 = (v16u8)__msa_fill_b(*thresh0_ptr); in vpx_lpf_horizontal_4_dual_msa()
53 thresh1 = (v16u8)__msa_fill_b(*thresh1_ptr); in vpx_lpf_horizontal_4_dual_msa()
54 thresh0 = (v16u8)__msa_ilvr_d((v2i64)thresh1, (v2i64)thresh0); in vpx_lpf_horizontal_4_dual_msa()
[all …]
Dloopfilter_8_msa.c18 v16u8 mask, hev, flat, thresh, b_limit, limit; in vpx_lpf_horizontal_8_msa()
19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_8_msa()
20 v16u8 p2_out, p1_out, p0_out, q0_out, q1_out, q2_out; in vpx_lpf_horizontal_8_msa()
28 thresh = (v16u8)__msa_fill_b(*thresh_ptr); in vpx_lpf_horizontal_8_msa()
29 b_limit = (v16u8)__msa_fill_b(*b_limit_ptr); in vpx_lpf_horizontal_8_msa()
30 limit = (v16u8)__msa_fill_b(*limit_ptr); in vpx_lpf_horizontal_8_msa()
37 flat = (v16u8)__msa_ilvr_d((v2i64)zero, (v2i64)flat); in vpx_lpf_horizontal_8_msa()
57 p2_out = __msa_bmnz_v(p2, (v16u8)p2_filter8, flat); in vpx_lpf_horizontal_8_msa()
58 p1_out = __msa_bmnz_v(p1_out, (v16u8)p1_filter8, flat); in vpx_lpf_horizontal_8_msa()
59 p0_out = __msa_bmnz_v(p0_out, (v16u8)p0_filter8, flat); in vpx_lpf_horizontal_8_msa()
[all …]
Dloopfilter_16_msa.c18 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_hz_lpf_t4_and_t8_16w()
19 v16u8 p2_out, p1_out, p0_out, q0_out, q1_out, q2_out; in vpx_hz_lpf_t4_and_t8_16w()
20 v16u8 flat, mask, hev, thresh, b_limit, limit; in vpx_hz_lpf_t4_and_t8_16w()
25 v16u8 zero = { 0 }; in vpx_hz_lpf_t4_and_t8_16w()
30 thresh = (v16u8)__msa_fill_b(*thresh_ptr); in vpx_hz_lpf_t4_and_t8_16w()
31 b_limit = (v16u8)__msa_fill_b(*b_limit_ptr); in vpx_hz_lpf_t4_and_t8_16w()
32 limit = (v16u8)__msa_fill_b(*limit_ptr); in vpx_hz_lpf_t4_and_t8_16w()
63 p2_out = __msa_bmnz_v(p2, (v16u8)p2_filt8_r, flat); in vpx_hz_lpf_t4_and_t8_16w()
64 p1_out = __msa_bmnz_v(p1_out, (v16u8)p1_filt8_r, flat); in vpx_hz_lpf_t4_and_t8_16w()
65 p0_out = __msa_bmnz_v(p0_out, (v16u8)p0_filt8_r, flat); in vpx_hz_lpf_t4_and_t8_16w()
[all …]
Ddeblock_msa.c37 out0 = (v16u8)temp6; \
38 out2 = (v16u8)temp7; \
39 out4 = (v16u8)temp8; \
40 out6 = (v16u8)temp9; \
41 out9 = (v16u8)__msa_ilvl_d((v2i64)out8, (v2i64)out8); \
42 out11 = (v16u8)__msa_ilvl_d((v2i64)out10, (v2i64)out10); \
43 out13 = (v16u8)__msa_ilvl_d((v2i64)out12, (v2i64)out12); \
44 out15 = (v16u8)__msa_ilvl_d((v2i64)out14, (v2i64)out14); \
45 out1 = (v16u8)__msa_ilvl_d((v2i64)out0, (v2i64)out0); \
46 out3 = (v16u8)__msa_ilvl_d((v2i64)out2, (v2i64)out2); \
[all …]
Dsub_pixel_variance_msa.c23 v16u8 src_l0_m, src_l1_m; \
47 v16u8 pred, src = { 0 }; in avg_sse_diff_4width_msa()
48 v16u8 ref = { 0 }; in avg_sse_diff_4width_msa()
80 v16u8 src0, src1, src2, src3; in avg_sse_diff_8width_msa()
81 v16u8 ref0, ref1, ref2, ref3; in avg_sse_diff_8width_msa()
82 v16u8 pred0, pred1; in avg_sse_diff_8width_msa()
114 v16u8 src, ref, pred; in avg_sse_diff_16width_msa()
169 v16u8 src0, src1, ref0, ref1, pred0, pred1; in avg_sse_diff_32width_msa()
227 v16u8 src0, src1, ref0, ref1, pred0, pred1; in avg_sse_diff_32x64_msa()
287 v16u8 src0, src1, src2, src3; in avg_sse_diff_64x32_msa()
[all …]
Dloopfilter_msa.h40 q0_out = __msa_xori_b((v16u8)q0_m, 0x80); \
42 p0_out = __msa_xori_b((v16u8)p0_m, 0x80); \
47 q1_out = __msa_xori_b((v16u8)q1_m, 0x80); \
49 p1_out = __msa_xori_b((v16u8)p1_m, 0x80); \
54 v16u8 tmp_flat4, p2_a_sub_p0, q2_a_sub_q0, p3_a_sub_p0, q3_a_sub_q0; \
55 v16u8 zero_in = { 0 }; \
68 flat_out = (tmp_flat4 < (v16u8)flat_out); \
76 v16u8 tmp_flat5, zero_in = { 0 }; \
77 v16u8 p4_a_sub_p0, q4_a_sub_q0, p5_a_sub_p0, q5_a_sub_q0; \
78 v16u8 p6_a_sub_p0, q6_a_sub_q0, p7_a_sub_p0, q7_a_sub_q0; \
[all …]
Dmacros_msa.h20 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__)
31 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__)
298 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__)
306 #define LD_UB3(...) LD_B3(v16u8, __VA_ARGS__)
313 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__)
321 #define LD_UB5(...) LD_B5(v16u8, __VA_ARGS__)
337 #define LD_UB8(...) LD_B8(v16u8, __VA_ARGS__)
412 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__)
419 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__)
426 #define ST_UB8(...) ST_B8(v16u8, __VA_ARGS__)
[all …]
Dintrapred_msa.c47 v16u8 src0; in intra_predict_vert_16x16_msa()
60 v16u8 src1, src2; in intra_predict_vert_32x32_msa()
105 v16u8 src0, src1, src2, src3; in intra_predict_horiz_16x16_msa()
114 src0 = (v16u8)__msa_fill_b(inp0); in intra_predict_horiz_16x16_msa()
115 src1 = (v16u8)__msa_fill_b(inp1); in intra_predict_horiz_16x16_msa()
116 src2 = (v16u8)__msa_fill_b(inp2); in intra_predict_horiz_16x16_msa()
117 src3 = (v16u8)__msa_fill_b(inp3); in intra_predict_horiz_16x16_msa()
128 v16u8 src0, src1, src2, src3; in intra_predict_horiz_32x32_msa()
137 src0 = (v16u8)__msa_fill_b(inp0); in intra_predict_horiz_32x32_msa()
138 src1 = (v16u8)__msa_fill_b(inp1); in intra_predict_horiz_32x32_msa()
[all …]
Dsad_msa.c21 #define SAD_INSVE_W4_UB(...) SAD_INSVE_W4(v16u8, __VA_ARGS__)
28 v16u8 src = { 0 }; in sad_4width_msa()
29 v16u8 ref = { 0 }; in sad_4width_msa()
30 v16u8 diff; in sad_4width_msa()
53 v16u8 src0, src1, src2, src3, ref0, ref1, ref2, ref3; in sad_8width_msa()
74 v16u8 src0, src1, ref0, ref1; in sad_16width_msa()
98 v16u8 src0, src1, ref0, ref1; in sad_32width_msa()
135 v16u8 src0, src1, src2, src3; in sad_64width_msa()
136 v16u8 ref0, ref1, ref2, ref3; in sad_64width_msa()
167 v16u8 src = { 0 }; in sad_4width_x3_msa()
[all …]
Dvpx_convolve_avg_msa.c17 v16u8 src0, src1, src2, src3; in avg_width4_msa()
18 v16u8 dst0, dst1, dst2, dst3; in avg_width4_msa()
60 v16u8 src0, src1, src2, src3; in avg_width8_msa()
61 v16u8 dst0, dst1, dst2, dst3; in avg_width8_msa()
83 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in avg_width16_msa()
84 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in avg_width16_msa()
104 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in avg_width32_msa()
105 v16u8 src8, src9, src10, src11, src12, src13, src14, src15; in avg_width32_msa()
106 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in avg_width32_msa()
107 v16u8 dst8, dst9, dst10, dst11, dst12, dst13, dst14, dst15; in avg_width32_msa()
[all …]
/external/webp/src/dsp/
Ddec_msa.c138 v16u8 dest0, dest1, dest2, dest3; in TransformAC3()
186 q = __msa_xori_b((v16u8)q_m, 0x80); \
187 p = __msa_xori_b((v16u8)p_m, 0x80); \
205 q0 = __msa_xori_b((v16u8)q0_m, 0x80); \
207 p0 = __msa_xori_b((v16u8)p0_m, 0x80); \
212 q1 = __msa_xori_b((v16u8)q1_m, 0x80); \
214 p1 = __msa_xori_b((v16u8)p1_m, 0x80); \
262 v16u8 p3_asub_p2_m, p2_asub_p1_m, p1_asub_p0_m, q1_asub_q0_m; \
263 v16u8 p1_asub_q1_m, p0_asub_q0_m, q3_asub_q2_m, q2_asub_q1_m; \
264 v16u8 flat_out; \
[all …]
Denc_msa.c85 v16u8 srcl0, srcl1, src0, src1; in FTransform()
121 tmp4 = (v4i32)__msa_nor_v((v16u8)tmp5, (v16u8)tmp5); in FTransform()
123 tmp5 = (v4i32)__msa_and_v((v16u8)tmp5, (v16u8)tmp4); in FTransform()
263 const v16u8 A = (v16u8)__msa_insert_d((v2i64)A, 0, val_m); in VE4()
264 const v16u8 B = SLDI_UB(A, A, 1); in VE4()
265 const v16u8 C = SLDI_UB(A, A, 2); in VE4()
266 const v16u8 AC = __msa_ave_u_b(A, C); in VE4()
267 const v16u8 B2 = __msa_ave_u_b(B, B); in VE4()
268 const v16u8 R = __msa_aver_u_b(AC, B2); in VE4()
296 const v16u8 A1 = (v16u8)__msa_insert_d((v2i64)A1, 0, val_m); in RD4()
[all …]
Dupsampling_msa.c31 v16u8 t0, t1; \
60 dst = (v16u8)__msa_pckev_b((v16i8)b1, (v16i8)b0); \
69 dst = (v16u8)__msa_pckev_b((v16i8)b0, (v16i8)b0); \
82 dst = (v16u8)__msa_pckev_b((v16i8)a1, (v16i8)a0); \
92 dst = (v16u8)__msa_pckev_b((v16i8)a0, (v16i8)a0); \
103 dst = (v16u8)__msa_pckev_b((v16i8)b1, (v16i8)b0); \
112 dst = (v16u8)__msa_pckev_b((v16i8)b0, (v16i8)b0); \
116 const v16u8 zero = { 0 }; \
119 const v16u8 in_y = LD_UB(y); \
120 const v16u8 in_u = LD_UB(u); \
[all …]
Dlossless_msa.c25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \
35 v16u8 src0, src1, src2, dst0, dst1, dst2; \
46 v16u8 src0, src1, src2, dst0, dst1; \
55 const v16u8 src0 = LD_UB(psrc); \
56 const v16u8 dst0 = VSHF_UB(src0, src0, m); \
117 v16u8 src0, dst0; in ConvertBGRAToRGBA()
118 const v16u8 mask = { 2, 1, 0, 3, 6, 5, 4, 7, 10, 9, 8, 11, 14, 13, 12, 15 }; in ConvertBGRAToRGBA()
121 v16u8 src1, dst1; in ConvertBGRAToRGBA()
157 const v16u8 mask0 = { 0, 1, 2, 4, 5, 6, 8, 9, 10, 12, 13, 14, in ConvertBGRAToBGR()
159 const v16u8 mask1 = { 5, 6, 8, 9, 10, 12, 13, 14, 16, 17, 18, 20, in ConvertBGRAToBGR()
[all …]
Dfilters_msa.c25 v16u8 src0, pred0, dst0; in PredictLineInverse0()
28 v16u8 src1, pred1, dst1; in PredictLineInverse0()
105 v16u8 pred0, dst0; in PredictLineGradient()
107 const v16u8 tmp0 = LD_UB(ppred - 1); in PredictLineGradient()
108 const v16u8 tmp1 = LD_UB(ppred - stride); in PredictLineGradient()
109 const v16u8 tmp2 = LD_UB(ppred - stride - 1); in PredictLineGradient()
110 const v16u8 src0 = LD_UB(pinput); in PredictLineGradient()
117 pred0 = (v16u8)__msa_pckev_b((v16i8)a1, (v16i8)a0); in PredictLineGradient()
Dmsa_macro.h32 #define ANDI_B(a, b) __msa_andi_b((v16u8)a, b)
33 #define ORI_B(a, b) __msa_ori_b((v16u8)a, b)
47 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__)
59 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__)
222 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__)
229 #define LD_UB3(...) LD_B3(v16u8, __VA_ARGS__)
236 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__)
244 #define LD_UB8(...) LD_B8(v16u8, __VA_ARGS__)
298 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__)
305 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__)
[all …]
/external/libvpx/libvpx/vp8/common/mips/msa/
Dloopfilter_filters_msa.c17 v16u8 p1_a_sub_q1, p0_a_sub_q0; \
21 p1_a_sub_q1 = (v16u8)__msa_srli_b((v16i8)p1_a_sub_q1, 1); \
24 mask = ((v16u8)mask <= b_limit); \
50 q0 = __msa_xori_b((v16u8)q0_m, 0x80); \
52 p0 = __msa_xori_b((v16u8)p0_m, 0x80); \
57 q1 = __msa_xori_b((v16u8)q1_m, 0x80); \
59 p1 = __msa_xori_b((v16u8)p1_m, 0x80); \
86 q0_in = __msa_xori_b((v16u8)q0_m, 0x80); \
87 p0_in = __msa_xori_b((v16u8)p0_m, 0x80); \
138 q2 = __msa_xori_b((v16u8)q2_m, 0x80); \
[all …]
Dvp8_macros_msa.h20 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__)
32 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__)
273 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__)
281 #define LD_UB3(...) LD_B3(v16u8, __VA_ARGS__)
289 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__)
297 #define LD_UB5(...) LD_B5(v16u8, __VA_ARGS__)
306 #define LD_UB8(...) LD_B8(v16u8, __VA_ARGS__)
350 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__)
357 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__)
365 #define ST_UB8(...) ST_B8(v16u8, __VA_ARGS__)
[all …]
Dbilinear_filter_msa.c34 v16u8 filt0, vec0, vec1, res0, res1; in common_hz_2t_4x4_msa()
40 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_4x4_msa()
53 v16u8 vec0, vec1, vec2, vec3, filt0; in common_hz_2t_4x8_msa()
61 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_4x8_msa()
89 v16u8 filt0; in common_hz_2t_8x4_msa()
96 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_8x4_msa()
111 v16u8 filt0; in common_hz_2t_8x8mult_msa()
118 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_8x8mult_msa()
185 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa()
193 filt0 = (v16u8)__msa_splati_h((v8i16)filt, 0); in common_hz_2t_16w_msa()
[all …]
/external/libpng/mips/
Dfilter_msa_intrinsics.c251 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__)
257 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__)
263 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__)
266 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__)
272 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__)
278 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__)
307 out0 = (RTYPE) __msa_hsub_u_h((v16u8) in0, (v16u8) in0); \
308 out1 = (RTYPE) __msa_hsub_u_h((v16u8) in1, (v16u8) in1); \
318 #define SLDI_B2_0_UB(...) SLDI_B2_0(v16u8, __VA_ARGS__)
326 #define SLDI_B3_0_UB(...) SLDI_B3_0(v16u8, __VA_ARGS__)
[all …]
/external/libyuv/files/include/libyuv/
Dmacros_msa.h141 #define LD_UB(...) LD_B(v16u8, __VA_ARGS__)
144 #define ST_UB(...) ST_B(v16u8, __VA_ARGS__)
161 #define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__)
168 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__)
181 #define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__)
188 #define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__)
215 #define VSHF_B2_UB(...) VSHF_B2(v16u8, __VA_ARGS__)
229 #define ILVRL_B2_UB(...) ILVRL_B2(v16u8, __VA_ARGS__)

12