Home
last modified time | relevance | path

Searched refs:in3 (Results 1 – 25 of 82) sorted by relevance

1234

/external/libvpx/libvpx/vp9/encoder/mips/msa/
Dvp9_fdct4x4_msa.c18 v8i16 in0, in1, in2, in3, in4; in vp9_fwht4x4_msa() local
20 LD_SH4(input, src_stride, in0, in1, in2, in3); in vp9_fwht4x4_msa()
23 in3 -= in2; in vp9_fwht4x4_msa()
24 in4 = (in0 - in3) >> 1; in vp9_fwht4x4_msa()
27 in3 += in1; in vp9_fwht4x4_msa()
29 TRANSPOSE4x4_SH_SH(in0, in2, in3, in1, in0, in2, in3, in1); in vp9_fwht4x4_msa()
32 in1 -= in3; in vp9_fwht4x4_msa()
34 SUB2(in4, in2, in4, in3, in2, in3); in vp9_fwht4x4_msa()
35 in0 -= in3; in vp9_fwht4x4_msa()
38 SLLI_4V(in0, in1, in2, in3, 2); in vp9_fwht4x4_msa()
[all …]
Dvp9_fdct8x8_msa.c18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_fht8x8_msa() local
20 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
21 SLLI_4V(in0, in1, in2, in3, 2); in vp9_fht8x8_msa()
26 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
28 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa()
29 in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
30 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
34 VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
36 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa()
37 in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
[all …]
Dvp9_fdct_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
36 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
38 cnst2_m, cnst3_m, in7, in0, in4, in3); \
62 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
82 #define VP9_FADST4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
90 UNPCK_R_SH_SW(in3, in3_r_m); \
/external/libvpx/libvpx/vp9/common/mips/msa/
Dvp9_idct4x4_msa.c18 v8i16 in0, in1, in2, in3; in vp9_iht4x4_16_add_msa() local
21 LD4x4_SH(input, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
22 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
27 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
29 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
30 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
34 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
36 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
37 VP9_IADST4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
41 VP9_IADST4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa()
[all …]
Dvp9_idct8x8_msa.c18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_iht8x8_64_add_msa() local
21 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
23 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
29 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
32 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_iht8x8_64_add_msa()
33 in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
34 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
39 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
42 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_iht8x8_64_add_msa()
43 in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
[all …]
/external/libvpx/libvpx/vp8/encoder/mips/msa/
Ddct_msa.c14 #define TRANSPOSE4x4_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
18 ILVR_H2_SH(in2, in0, in3, in1, s0_m, s1_m); \
20 ILVL_H2_SH(in2, in0, in3, in1, s0_m, s1_m); \
70 v8i16 in0, in1, in2, in3; in vp8_short_fdct4x4_msa() local
77 LD_SH4(input, pitch / 2, in0, in1, in2, in3); in vp8_short_fdct4x4_msa()
78 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp8_short_fdct4x4_msa()
80 BUTTERFLY_4(in0, in1, in2, in3, temp0, temp1, in1, in3); in vp8_short_fdct4x4_msa()
81 SLLI_4V(temp0, temp1, in1, in3, 3); in vp8_short_fdct4x4_msa()
85 temp0 = __msa_ilvr_h(in3, in1); in vp8_short_fdct4x4_msa()
93 PCKEV_H2_SH(out0, out0, out1, out1, in1, in3); in vp8_short_fdct4x4_msa()
[all …]
/external/libvpx/libvpx/vpx_dsp/mips/
Didct4x4_msa.c15 v8i16 in0, in1, in2, in3; in vpx_iwht4x4_16_add_msa() local
19 LD4x4_SH(input, in0, in2, in3, in1); in vpx_iwht4x4_16_add_msa()
20 TRANSPOSE4x4_SH_SH(in0, in2, in3, in1, in0, in2, in3, in1); in vpx_iwht4x4_16_add_msa()
23 UNPCK_R_SH_SW(in3, in3_r); in vpx_iwht4x4_16_add_msa()
46 in2, in3); in vpx_iwht4x4_16_add_msa()
47 ADDBLK_ST4x4_UB(in0, in3, in1, in2, dst, dst_stride); in vpx_iwht4x4_16_add_msa()
72 v8i16 in0, in1, in2, in3; in vpx_idct4x4_16_add_msa() local
75 LD4x4_SH(input, in0, in1, in2, in3); in vpx_idct4x4_16_add_msa()
77 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vpx_idct4x4_16_add_msa()
78 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vpx_idct4x4_16_add_msa()
[all …]
Didct8x8_msa.c15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_msa() local
18 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
21 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
24 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
27 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
30 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
33 SRARI_H4_SH(in0, in1, in2, in3, 5); in vpx_idct8x8_64_add_msa()
36 VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3); in vpx_idct8x8_64_add_msa()
43 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_12_add_msa() local
49 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_12_add_msa()
[all …]
Dfwd_txfm_msa.c16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x16_1d_column() local
29 LD_SH16(input, src_stride, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in fdct8x16_1d_column()
31 SLLI_4V(in0, in1, in2, in3, 2); in fdct8x16_1d_column()
35 ADD4(in0, in15, in1, in14, in2, in13, in3, in12, tmp0, tmp1, tmp2, tmp3); in fdct8x16_1d_column()
40 SUB4(in0, in15, in1, in14, in2, in13, in3, in12, in15, in14, in13, in12); in fdct8x16_1d_column()
135 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct16x8_1d_row() local
138 LD_SH8(input, 16, in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row()
140 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in fdct16x8_1d_row()
144 ADD4(in0, 1, in1, 1, in2, 1, in3, 1, in0, in1, in2, in3); in fdct16x8_1d_row()
148 SRA_4V(in0, in1, in2, in3, 2); in fdct16x8_1d_row()
[all …]
Dmacros_msa.h263 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument
268 SW(in3, (pdst) + 3 * stride); \
278 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument
283 SD(in3, (pdst) + 3 * stride); \
414 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
417 ST_B2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \
421 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
423 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \
440 #define ST_H4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
443 ST_H2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \
[all …]
Dfwd_dct32x32_msa.c16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_load_butterfly() local
22 LD_SH4(input, src_stride, in0, in1, in2, in3); in fdct8x32_1d_column_load_butterfly()
26 SLLI_4V(in0, in1, in2, in3, 2); in fdct8x32_1d_column_load_butterfly()
30 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly()
40 LD_SH4(input + (8 * src_stride), src_stride, in0, in1, in2, in3); in fdct8x32_1d_column_load_butterfly()
44 SLLI_4V(in0, in1, in2, in3, 2); in fdct8x32_1d_column_load_butterfly()
48 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly()
59 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_even_store() local
65 LD_SH4(input, 8, in0, in1, in2, in3); in fdct8x32_1d_column_even_store()
67 BUTTERFLY_8(in0, in1, in2, in3, in12, in13, in14, in15, vec0, vec1, vec2, in fdct8x32_1d_column_even_store()
[all …]
Dinv_txfm_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
36 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
38 cnst2_m, cnst3_m, in7, in0, in4, in3); \
62 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
93 #define VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3) \ argument
104 ADD4(res0_m, in0, res1_m, in1, res2_m, in2, res3_m, in3, res0_m, res1_m, \
111 #define VP9_IDCT4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
124 step1_m = __msa_ilvr_h(in3, in1); \
134 #define VP9_IADST4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
146 ILVR_H2_SH(in0, in2, in1, in3, res0_m, res1_m); \
[all …]
/external/webp/src/dsp/
Dmsa_macro.h166 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument
174 SW(in3, ptmp); \
200 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument
208 SD(in3, ptmp); \
301 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
303 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
308 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
310 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \
336 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
338 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
[all …]
Dcommon_sse2.h70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument
78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
132 __m128i* const in2, __m128i* const in3, in VP8PlanarTo24b() argument
165 __m128i* const in3) { in VP8L32bToPlanar() argument
169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar()
170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar()
185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar()
/external/libvpx/libvpx/vp8/common/mips/msa/
Dvp8_macros_msa.h238 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument
243 SW(in3, (pdst) + 3 * stride); \
253 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument
258 SD(in3, (pdst) + 3 * stride); \
352 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
355 ST_B2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \
360 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
362 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \
549 #define VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument
552 out1 = (RTYPE)__msa_vshf_b((v16i8)mask1, (v16i8)in3, (v16i8)in2); \
[all …]
Didct_msa.c18 #define TRANSPOSE_TWO_4x4_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
22 TRANSPOSE8X4_SH_SH(in0, in1, in2, in3, s4_m, s5_m, s6_m, s7_m); \
45 #define VP8_IDCT_1D_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
55 c_tmp2_m = __msa_mul_q_h(in3, const_cospi8sqrt2minus1_m); \
57 c_tmp2_m = in3 + c_tmp2_m; \
62 d_tmp2_m = EXPAND_TO_H_MULTIPLY_SINPI8SQRT2_PCK_TO_W(in3); \
67 #define VP8_IDCT_1D_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
78 c_tmp2_m = in3 + ((in3 * const_cospi8sqrt2minus1_m) >> 16); \
81 d_tmp2_m = (in3 * sinpi8_sqrt2_m) >> 16; \
90 v4i32 in0, in1, in2, in3, hz0, hz1, hz2, hz3, vt0, vt1, vt2, vt3; in idct4x4_addblk_msa() local
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_txfm_sse2.c51 __m128i in3 = _mm_load_si128((const __m128i *)(input + 3 * stride)); in vpx_fdct8x8_1_sse2() local
55 u1 = _mm_add_epi16(in2, in3); in vpx_fdct8x8_1_sse2()
60 in3 = _mm_load_si128((const __m128i *)(input + 7 * stride)); in vpx_fdct8x8_1_sse2()
65 in2 = _mm_add_epi16(in2, in3); in vpx_fdct8x8_1_sse2()
89 __m128i in0, in1, in2, in3; in vpx_fdct16x16_1_sse2() local
98 in3 = _mm_load_si128((const __m128i *)(input + 1 * stride + 8)); in vpx_fdct16x16_1_sse2()
101 u1 = _mm_add_epi16(in2, in3); in vpx_fdct16x16_1_sse2()
107 in3 = _mm_load_si128((const __m128i *)(input + 3 * stride + 8)); in vpx_fdct16x16_1_sse2()
111 u1 = _mm_add_epi16(in2, in3); in vpx_fdct16x16_1_sse2()
117 in3 = _mm_load_si128((const __m128i *)(input + 5 * stride + 8)); in vpx_fdct16x16_1_sse2()
[all …]
/external/python/cpython2/Modules/_ctypes/libffi/testsuite/libffi.call/
Dreturn_dbl1.c10 static double return_dbl(double dbl1, float fl2, unsigned int in3, double dbl4) in return_dbl() argument
12 return dbl1 + fl2 + in3 + dbl4; in return_dbl()
21 unsigned int in3; in main() local
28 values[2] = &in3; in main()
36 in3 = 255; in main()
40 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, fl2, in3, dbl4)); in main()
41 CHECK(rdbl == dbl1 + fl2 + in3 + dbl4); in main()
Dreturn_dbl2.c10 static double return_dbl(double dbl1, double dbl2, unsigned int in3, double dbl4) in return_dbl() argument
12 return dbl1 + dbl2 + in3 + dbl4; in return_dbl()
20 unsigned int in3; in main() local
27 values[2] = &in3; in main()
35 in3 = 255; in main()
39 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, dbl2, in3, dbl4)); in main()
40 CHECK(rdbl == dbl1 + dbl2 + in3 + dbl4); in main()
Dreturn_fl3.c10 static float return_fl(float fl1, float fl2, unsigned int in3, float fl4) in return_fl() argument
12 return fl1 + fl2 + in3 + fl4; in return_fl()
20 unsigned int in3; in main() local
27 values[2] = &in3; in main()
35 in3 = 255; in main()
39 printf ("%f vs %f\n", rfl, return_fl(fl1, fl2, in3, fl4)); in main()
40 CHECK(rfl == fl1 + fl2 + in3 + fl4); in main()
/external/python/cpython2/Modules/_ctypes/libffi/src/ia64/
Dunix.S102 zxt1 r16 = in3
113 shr in3 = in3, 8
179 cmp.lt p6, p0 = 8, in3
180 cmp.lt p7, p0 = 16, in3
181 cmp.lt p8, p0 = 24, in3
193 mov out2 = in3
203 cmp.lt p6, p0 = 4, in3
207 cmp.lt p7, p0 = 8, in3
208 cmp.lt p8, p0 = 12, in3
212 cmp.lt p9, p0 = 16, in3
[all …]
/external/llvm/test/CodeGen/AMDGPU/
Dfma.f64.ll12 double addrspace(1)* %in2, double addrspace(1)* %in3) {
15 %r2 = load double, double addrspace(1)* %in3
25 <2 x double> addrspace(1)* %in2, <2 x double> addrspace(1)* %in3) {
28 %r2 = load <2 x double>, <2 x double> addrspace(1)* %in3
40 <4 x double> addrspace(1)* %in2, <4 x double> addrspace(1)* %in3) {
43 %r2 = load <4 x double>, <4 x double> addrspace(1)* %in3
/external/eigen/unsupported/test/
Dcxx11_tensor_sycl.cpp38 Tensor<float, 3> in3(tensorRange); in test_sycl_cpu() local
42 in3 = in3.random(); in test_sycl_cpu()
46 …float * gpu_in3_data = static_cast<float*>(sycl_device.allocate(in3.dimensions().TotalSize()*size… in test_sycl_cpu()
137 …sycl_device.memcpyHostToDevice(gpu_in3_data, in3.data(),(in3.dimensions().TotalSize())*sizeof(floa… in test_sycl_cpu()
145 : in3(i, j, k)); in test_sycl_cpu()
/external/boringssl/src/crypto/fipsmodule/aes/asm/
Daesp8-ppc.pl672 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
676 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment
778 lvx_u $in3,$x30,$inp
784 le?vperm $in3,$in3,$in3,$inpperm
792 vxor $out3,$in3,$rndkey0
899 vxor $in3,$in3,v31
915 vncipherlast $out4,$out4,$in3
917 lvx_u $in3,$x30,$inp
922 le?vperm $in3,$in3,$in3,$inpperm
946 vxor $out3,$in3,$rndkey0
[all …]
/external/libyuv/files/source/
Drotate_msa.cc22 #define ILVRL_B(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
26 out2 = (v16u8)__msa_ilvr_b((v16i8)in3, (v16i8)in2); \
27 out3 = (v16u8)__msa_ilvl_b((v16i8)in3, (v16i8)in2); \
30 #define ILVRL_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
34 out2 = (v16u8)__msa_ilvr_h((v8i16)in3, (v8i16)in2); \
35 out3 = (v16u8)__msa_ilvl_h((v8i16)in3, (v8i16)in2); \
38 #define ILVRL_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
42 out2 = (v16u8)__msa_ilvr_w((v4i32)in3, (v4i32)in2); \
43 out3 = (v16u8)__msa_ilvl_w((v4i32)in3, (v4i32)in2); \
46 #define ILVRL_D(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
[all …]

1234