/external/libvpx/libvpx/vp9/encoder/mips/msa/ |
D | vp9_fdct4x4_msa.c | 18 v8i16 in0, in1, in2, in3, in4; in vp9_fwht4x4_msa() local 20 LD_SH4(input, src_stride, in0, in1, in2, in3); in vp9_fwht4x4_msa() 23 in3 -= in2; in vp9_fwht4x4_msa() 24 in4 = (in0 - in3) >> 1; in vp9_fwht4x4_msa() 27 in3 += in1; in vp9_fwht4x4_msa() 29 TRANSPOSE4x4_SH_SH(in0, in2, in3, in1, in0, in2, in3, in1); in vp9_fwht4x4_msa() 32 in1 -= in3; in vp9_fwht4x4_msa() 34 SUB2(in4, in2, in4, in3, in2, in3); in vp9_fwht4x4_msa() 35 in0 -= in3; in vp9_fwht4x4_msa() 38 SLLI_4V(in0, in1, in2, in3, 2); in vp9_fwht4x4_msa() [all …]
|
D | vp9_fdct8x8_msa.c | 18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_fht8x8_msa() local 20 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa() 21 SLLI_4V(in0, in1, in2, in3, 2); in vp9_fht8x8_msa() 26 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa() 28 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa() 29 in3, in4, in5, in6, in7); in vp9_fht8x8_msa() 30 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa() 34 VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa() 36 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa() 37 in3, in4, in5, in6, in7); in vp9_fht8x8_msa() [all …]
|
D | vp9_fdct_msa.h | 18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument 36 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \ 38 cnst2_m, cnst3_m, in7, in0, in4, in3); \ 62 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \ 82 #define VP9_FADST4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 90 UNPCK_R_SH_SW(in3, in3_r_m); \
|
/external/libvpx/libvpx/vp9/common/mips/msa/ |
D | vp9_idct4x4_msa.c | 18 v8i16 in0, in1, in2, in3; in vp9_iht4x4_16_add_msa() local 21 LD4x4_SH(input, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 22 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 27 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 29 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 30 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 34 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 36 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 37 VP9_IADST4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() 41 VP9_IADST4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vp9_iht4x4_16_add_msa() [all …]
|
D | vp9_idct8x8_msa.c | 18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_iht8x8_64_add_msa() local 21 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() 23 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa() 29 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa() 32 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_iht8x8_64_add_msa() 33 in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() 34 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa() 39 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa() 42 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_iht8x8_64_add_msa() 43 in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() [all …]
|
/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
D | dct_msa.c | 14 #define TRANSPOSE4x4_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 18 ILVR_H2_SH(in2, in0, in3, in1, s0_m, s1_m); \ 20 ILVL_H2_SH(in2, in0, in3, in1, s0_m, s1_m); \ 70 v8i16 in0, in1, in2, in3; in vp8_short_fdct4x4_msa() local 77 LD_SH4(input, pitch / 2, in0, in1, in2, in3); in vp8_short_fdct4x4_msa() 78 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vp8_short_fdct4x4_msa() 80 BUTTERFLY_4(in0, in1, in2, in3, temp0, temp1, in1, in3); in vp8_short_fdct4x4_msa() 81 SLLI_4V(temp0, temp1, in1, in3, 3); in vp8_short_fdct4x4_msa() 85 temp0 = __msa_ilvr_h(in3, in1); in vp8_short_fdct4x4_msa() 93 PCKEV_H2_SH(out0, out0, out1, out1, in1, in3); in vp8_short_fdct4x4_msa() [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | idct4x4_msa.c | 15 v8i16 in0, in1, in2, in3; in vpx_iwht4x4_16_add_msa() local 19 LD4x4_SH(input, in0, in2, in3, in1); in vpx_iwht4x4_16_add_msa() 20 TRANSPOSE4x4_SH_SH(in0, in2, in3, in1, in0, in2, in3, in1); in vpx_iwht4x4_16_add_msa() 23 UNPCK_R_SH_SW(in3, in3_r); in vpx_iwht4x4_16_add_msa() 46 in2, in3); in vpx_iwht4x4_16_add_msa() 47 ADDBLK_ST4x4_UB(in0, in3, in1, in2, dst, dst_stride); in vpx_iwht4x4_16_add_msa() 72 v8i16 in0, in1, in2, in3; in vpx_idct4x4_16_add_msa() local 75 LD4x4_SH(input, in0, in1, in2, in3); in vpx_idct4x4_16_add_msa() 77 TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3); in vpx_idct4x4_16_add_msa() 78 VP9_IDCT4x4(in0, in1, in2, in3, in0, in1, in2, in3); in vpx_idct4x4_16_add_msa() [all …]
|
D | idct8x8_msa.c | 15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_msa() local 18 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa() 21 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa() 24 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa() 27 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa() 30 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa() 33 SRARI_H4_SH(in0, in1, in2, in3, 5); in vpx_idct8x8_64_add_msa() 36 VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3); in vpx_idct8x8_64_add_msa() 43 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_12_add_msa() local 49 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_12_add_msa() [all …]
|
D | fwd_txfm_msa.c | 16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x16_1d_column() local 29 LD_SH16(input, src_stride, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in fdct8x16_1d_column() 31 SLLI_4V(in0, in1, in2, in3, 2); in fdct8x16_1d_column() 35 ADD4(in0, in15, in1, in14, in2, in13, in3, in12, tmp0, tmp1, tmp2, tmp3); in fdct8x16_1d_column() 40 SUB4(in0, in15, in1, in14, in2, in13, in3, in12, in15, in14, in13, in12); in fdct8x16_1d_column() 135 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct16x8_1d_row() local 138 LD_SH8(input, 16, in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row() 140 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in fdct16x8_1d_row() 144 ADD4(in0, 1, in1, 1, in2, 1, in3, 1, in0, in1, in2, in3); in fdct16x8_1d_row() 148 SRA_4V(in0, in1, in2, in3, 2); in fdct16x8_1d_row() [all …]
|
D | macros_msa.h | 263 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument 268 SW(in3, (pdst) + 3 * stride); \ 278 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument 283 SD(in3, (pdst) + 3 * stride); \ 414 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 417 ST_B2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \ 421 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument 423 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \ 440 #define ST_H4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 443 ST_H2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \ [all …]
|
D | fwd_dct32x32_msa.c | 16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_load_butterfly() local 22 LD_SH4(input, src_stride, in0, in1, in2, in3); in fdct8x32_1d_column_load_butterfly() 26 SLLI_4V(in0, in1, in2, in3, 2); in fdct8x32_1d_column_load_butterfly() 30 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly() 40 LD_SH4(input + (8 * src_stride), src_stride, in0, in1, in2, in3); in fdct8x32_1d_column_load_butterfly() 44 SLLI_4V(in0, in1, in2, in3, 2); in fdct8x32_1d_column_load_butterfly() 48 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly() 59 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_even_store() local 65 LD_SH4(input, 8, in0, in1, in2, in3); in fdct8x32_1d_column_even_store() 67 BUTTERFLY_8(in0, in1, in2, in3, in12, in13, in14, in15, vec0, vec1, vec2, in fdct8x32_1d_column_even_store() [all …]
|
D | inv_txfm_msa.h | 18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument 36 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \ 38 cnst2_m, cnst3_m, in7, in0, in4, in3); \ 62 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \ 93 #define VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3) \ argument 104 ADD4(res0_m, in0, res1_m, in1, res2_m, in2, res3_m, in3, res0_m, res1_m, \ 111 #define VP9_IDCT4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 124 step1_m = __msa_ilvr_h(in3, in1); \ 134 #define VP9_IADST4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 146 ILVR_H2_SH(in0, in2, in1, in3, res0_m, res1_m); \ [all …]
|
/external/webp/src/dsp/ |
D | msa_macro.h | 166 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument 174 SW(in3, ptmp); \ 200 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument 208 SD(in3, ptmp); \ 301 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 303 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ 308 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 310 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \ 336 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 338 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ [all …]
|
D | common_sse2.h | 70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument 78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 132 __m128i* const in2, __m128i* const in3, in VP8PlanarTo24b() argument 165 __m128i* const in3) { in VP8L32bToPlanar() argument 169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar() 170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar() 185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar()
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | vp8_macros_msa.h | 238 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument 243 SW(in3, (pdst) + 3 * stride); \ 253 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument 258 SD(in3, (pdst) + 3 * stride); \ 352 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 355 ST_B2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \ 360 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument 362 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \ 549 #define VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument 552 out1 = (RTYPE)__msa_vshf_b((v16i8)mask1, (v16i8)in3, (v16i8)in2); \ [all …]
|
D | idct_msa.c | 18 #define TRANSPOSE_TWO_4x4_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 22 TRANSPOSE8X4_SH_SH(in0, in1, in2, in3, s4_m, s5_m, s6_m, s7_m); \ 45 #define VP8_IDCT_1D_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 55 c_tmp2_m = __msa_mul_q_h(in3, const_cospi8sqrt2minus1_m); \ 57 c_tmp2_m = in3 + c_tmp2_m; \ 62 d_tmp2_m = EXPAND_TO_H_MULTIPLY_SINPI8SQRT2_PCK_TO_W(in3); \ 67 #define VP8_IDCT_1D_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 78 c_tmp2_m = in3 + ((in3 * const_cospi8sqrt2minus1_m) >> 16); \ 81 d_tmp2_m = (in3 * sinpi8_sqrt2_m) >> 16; \ 90 v4i32 in0, in1, in2, in3, hz0, hz1, hz2, hz3, vt0, vt1, vt2, vt3; in idct4x4_addblk_msa() local [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_txfm_sse2.c | 51 __m128i in3 = _mm_load_si128((const __m128i *)(input + 3 * stride)); in vpx_fdct8x8_1_sse2() local 55 u1 = _mm_add_epi16(in2, in3); in vpx_fdct8x8_1_sse2() 60 in3 = _mm_load_si128((const __m128i *)(input + 7 * stride)); in vpx_fdct8x8_1_sse2() 65 in2 = _mm_add_epi16(in2, in3); in vpx_fdct8x8_1_sse2() 89 __m128i in0, in1, in2, in3; in vpx_fdct16x16_1_sse2() local 98 in3 = _mm_load_si128((const __m128i *)(input + 1 * stride + 8)); in vpx_fdct16x16_1_sse2() 101 u1 = _mm_add_epi16(in2, in3); in vpx_fdct16x16_1_sse2() 107 in3 = _mm_load_si128((const __m128i *)(input + 3 * stride + 8)); in vpx_fdct16x16_1_sse2() 111 u1 = _mm_add_epi16(in2, in3); in vpx_fdct16x16_1_sse2() 117 in3 = _mm_load_si128((const __m128i *)(input + 5 * stride + 8)); in vpx_fdct16x16_1_sse2() [all …]
|
/external/python/cpython2/Modules/_ctypes/libffi/testsuite/libffi.call/ |
D | return_dbl1.c | 10 static double return_dbl(double dbl1, float fl2, unsigned int in3, double dbl4) in return_dbl() argument 12 return dbl1 + fl2 + in3 + dbl4; in return_dbl() 21 unsigned int in3; in main() local 28 values[2] = &in3; in main() 36 in3 = 255; in main() 40 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, fl2, in3, dbl4)); in main() 41 CHECK(rdbl == dbl1 + fl2 + in3 + dbl4); in main()
|
D | return_dbl2.c | 10 static double return_dbl(double dbl1, double dbl2, unsigned int in3, double dbl4) in return_dbl() argument 12 return dbl1 + dbl2 + in3 + dbl4; in return_dbl() 20 unsigned int in3; in main() local 27 values[2] = &in3; in main() 35 in3 = 255; in main() 39 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, dbl2, in3, dbl4)); in main() 40 CHECK(rdbl == dbl1 + dbl2 + in3 + dbl4); in main()
|
D | return_fl3.c | 10 static float return_fl(float fl1, float fl2, unsigned int in3, float fl4) in return_fl() argument 12 return fl1 + fl2 + in3 + fl4; in return_fl() 20 unsigned int in3; in main() local 27 values[2] = &in3; in main() 35 in3 = 255; in main() 39 printf ("%f vs %f\n", rfl, return_fl(fl1, fl2, in3, fl4)); in main() 40 CHECK(rfl == fl1 + fl2 + in3 + fl4); in main()
|
/external/python/cpython2/Modules/_ctypes/libffi/src/ia64/ |
D | unix.S | 102 zxt1 r16 = in3 113 shr in3 = in3, 8 179 cmp.lt p6, p0 = 8, in3 180 cmp.lt p7, p0 = 16, in3 181 cmp.lt p8, p0 = 24, in3 193 mov out2 = in3 203 cmp.lt p6, p0 = 4, in3 207 cmp.lt p7, p0 = 8, in3 208 cmp.lt p8, p0 = 12, in3 212 cmp.lt p9, p0 = 16, in3 [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | fma.f64.ll | 12 double addrspace(1)* %in2, double addrspace(1)* %in3) { 15 %r2 = load double, double addrspace(1)* %in3 25 <2 x double> addrspace(1)* %in2, <2 x double> addrspace(1)* %in3) { 28 %r2 = load <2 x double>, <2 x double> addrspace(1)* %in3 40 <4 x double> addrspace(1)* %in2, <4 x double> addrspace(1)* %in3) { 43 %r2 = load <4 x double>, <4 x double> addrspace(1)* %in3
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_sycl.cpp | 38 Tensor<float, 3> in3(tensorRange); in test_sycl_cpu() local 42 in3 = in3.random(); in test_sycl_cpu() 46 …float * gpu_in3_data = static_cast<float*>(sycl_device.allocate(in3.dimensions().TotalSize()*size… in test_sycl_cpu() 137 …sycl_device.memcpyHostToDevice(gpu_in3_data, in3.data(),(in3.dimensions().TotalSize())*sizeof(floa… in test_sycl_cpu() 145 : in3(i, j, k)); in test_sycl_cpu()
|
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
D | aesp8-ppc.pl | 672 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13)); 676 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment 778 lvx_u $in3,$x30,$inp 784 le?vperm $in3,$in3,$in3,$inpperm 792 vxor $out3,$in3,$rndkey0 899 vxor $in3,$in3,v31 915 vncipherlast $out4,$out4,$in3 917 lvx_u $in3,$x30,$inp 922 le?vperm $in3,$in3,$in3,$inpperm 946 vxor $out3,$in3,$rndkey0 [all …]
|
/external/libyuv/files/source/ |
D | rotate_msa.cc | 22 #define ILVRL_B(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 26 out2 = (v16u8)__msa_ilvr_b((v16i8)in3, (v16i8)in2); \ 27 out3 = (v16u8)__msa_ilvl_b((v16i8)in3, (v16i8)in2); \ 30 #define ILVRL_H(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 34 out2 = (v16u8)__msa_ilvr_h((v8i16)in3, (v8i16)in2); \ 35 out3 = (v16u8)__msa_ilvl_h((v8i16)in3, (v8i16)in2); \ 38 #define ILVRL_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 42 out2 = (v16u8)__msa_ilvr_w((v4i32)in3, (v4i32)in2); \ 43 out3 = (v16u8)__msa_ilvl_w((v4i32)in3, (v4i32)in2); \ 46 #define ILVRL_D(in0, in1, in2, in3, out0, out1, out2, out3) \ argument [all …]
|