/third_party/openh264/codec/common/inc/ |
D | msa_macros.h | 231 #define MSA_ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 234 MSA_ST_V2(RTYPE, in2, in3, (pdst) + 2 * (stride), stride); \ 237 #define MSA_ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument 239 MSA_ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \ 359 #define MSA_ST_D8(in0, in1, in2, in3, idx0, idx1, idx2, idx3, \ argument 363 MSA_ST_D4(in2, in3, idx4, idx5, idx6, idx7, pdst + 4 * stride, stride) \ 379 #define MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument 382 MSA_VSHF_B(RTYPE, in2, in3, mask1, out1) \ 385 #define MSA_VSHF_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 388 MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1); \ [all …]
|
/third_party/ffmpeg/libavutil/mips/ |
D | generic_macros_msa.h | 245 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument 250 SW(in3, (pdst) + 3 * stride); \ 260 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument 265 SD(in3, (pdst) + 3 * stride); \ 373 #define ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 376 ST_V2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \ 383 #define ST_V6(RTYPE, in0, in1, in2, in3, in4, in5, pdst, stride) \ argument 385 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \ 390 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument 392 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \ [all …]
|
/third_party/ffmpeg/libavcodec/mips/ |
D | idctdsp_msa.c | 28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_pixels_clamped_msa() local 30 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa() 31 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa() 32 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_pixels_clamped_msa() 38 in3_d = __msa_copy_u_d((v2i64) in3, 0); in put_pixels_clamped_msa() 52 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_signed_pixels_clamped_msa() local 54 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa() 59 in3 += 128; in put_signed_pixels_clamped_msa() 65 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa() 66 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_signed_pixels_clamped_msa() [all …]
|
D | h263dsp_msa.c | 32 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in h263_h_loop_filter_msa() local 38 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in h263_h_loop_filter_msa() 39 TRANSPOSE8x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in h263_h_loop_filter_msa() 40 in0, in3, in2, in1); in h263_h_loop_filter_msa() 44 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_h_loop_filter_msa() 82 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa() 83 in3 = (v16u8) __msa_adds_s_b((v16i8) in3, (v16i8) d0); in h263_h_loop_filter_msa() 84 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa() 86 ILVR_B2_SH(in3, in0, in1, in2, temp0, temp1); in h263_h_loop_filter_msa() 88 in3 = (v16u8) __msa_ilvl_h(temp1, temp0); in h263_h_loop_filter_msa() [all …]
|
D | hevc_idct_msa.c | 94 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument 104 ILVR_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \ 106 ILVL_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \ 146 PCKEV_H2_SH(sum1_l, sum1_r, sum2_l, sum2_r, in3, in4); \ 334 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_idct_8x8_msa() local 336 LD_SH8(coeffs, 8, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa() 337 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in hevc_idct_8x8_msa() 338 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in hevc_idct_8x8_msa() 339 in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa() 340 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 12); in hevc_idct_8x8_msa() [all …]
|
D | vp9_idct_msa.c | 86 #define VP9_DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 94 DOTP_SH4_SW(in2, in3, in2, in3, in6, in6, in7, in7, \ 116 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 134 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \ 137 in4, in3); \ 163 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \ 229 #define VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3) \ argument 240 ADD4(res0_m, in0, res1_m, in1, res2_m, in2, res3_m, in3, \ 247 #define VP9_IDCT4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 261 step1_m = __msa_ilvr_h(in3, in1); \ [all …]
|
D | vc1dsp_msa.c | 30 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x8_msa() local 44 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_msa() 48 UNPCK_SH_SW(in3, in_r3, in_l3); in ff_vc1_inv_trans_8x8_msa() 134 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_msa() 137 ST_SH8(in0, in1, in2, in3, in4, in5, in6, in7, block, 8); in ff_vc1_inv_trans_8x8_msa() 142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_4x8_msa() local 159 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_4x8_msa() 163 UNPCK_R_SH_SW(in3, in_r3); in ff_vc1_inv_trans_4x8_msa() 191 in0, in1, in2, in3); in ff_vc1_inv_trans_4x8_msa() 192 ST_D8(in0, in1, in2, in3, 0, 1, 0, 1, 0, 1, 0, 1, block, 8); in ff_vc1_inv_trans_4x8_msa() [all …]
|
D | simple_idct_msa.c | 28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in simple_idct_msa() local 41 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa() 46 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa() 47 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa() 48 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7; in simple_idct_msa() 79 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa() 111 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp, in simple_idct_msa() 121 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa() 122 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa() 152 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa() [all …]
|
D | vp8_idct_msa.c | 29 #define VP8_IDCT_1D_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 40 c_tmp2_m = in3 + (((in3) * const_cospi8sqrt2minus1_m) >> 16); \ 43 d_tmp2_m = ((in3) * sinpi8_sqrt2_m) >> 16; \ 51 v4i32 in0, in1, in2, in3, hz0, hz1, hz2, hz3, vt0, vt1, vt2, vt3; in ff_vp8_idct_add_msa() local 60 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_idct_add_msa() 61 VP8_IDCT_1D_W(in0, in1, in2, in3, hz0, hz1, hz2, hz3); in ff_vp8_idct_add_msa() 108 v4i32 in0, in1, in2, in3, a1, b1, c1, d1; in ff_vp8_luma_dc_wht_msa() local 114 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_luma_dc_wht_msa() 115 BUTTERFLY_4(in0, in1, in2, in3, a1, b1, c1, d1); in ff_vp8_luma_dc_wht_msa()
|
D | mpegvideoencdsp_msa.c | 27 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in sum_u8src_16width_msa() local 30 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in sum_u8src_16width_msa() 34 HADD_UB4_UB(in0, in1, in2, in3, in0, in1, in2, in3); in sum_u8src_16width_msa() 42 sum += HADD_UH_U32(in3); in sum_u8src_16width_msa()
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/spirv_assembly/ |
D | vktSpvAsmTrinaryMinMaxTests.cpp | 196 static inline void runOpFunc (F f, void* out, const void* in1, const void* in2, const void* in3) in runOpFunc() argument 198 …interpret_cast<const T*>(in1), *reinterpret_cast<const T*>(in2), *reinterpret_cast<const T*>(in3)); in runOpFunc() 202 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (min3<deInt8>, out, in1,… in minInt8() argument 203 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (max3<deInt8>, out, in1,… in maxInt8() argument 204 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (mid3<deInt8>, out, in1,… in midInt8() argument 205 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (min3<deUint8>, out, in1… in minUint8() argument 206 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (max3<deUint8>, out, in1… in maxUint8() argument 207 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (mid3<deUint8>, out, in1… in midUint8() argument 208 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (min3<deInt16>, out, in1… in minInt16() argument 209 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (max3<deInt16>, out, in1… in maxInt16() argument [all …]
|
/third_party/skia/third_party/externals/libwebp/src/dsp/ |
D | msa_macro.h | 168 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument 176 SW(in3, ptmp); \ 202 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument 210 SD(in3, ptmp); \ 303 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 305 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ 310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 312 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \ 338 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 340 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ [all …]
|
D | common_sse2.h | 70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument 78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 133 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE2() argument 165 __m128i* const in3) { in VP8L32bToPlanar_SSE2() argument 169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2() 170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2() 185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar_SSE2()
|
/third_party/flutter/skia/third_party/externals/libwebp/src/dsp/ |
D | msa_macro.h | 168 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument 176 SW(in3, ptmp); \ 202 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument 210 SD(in3, ptmp); \ 303 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 305 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ 310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 312 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \ 338 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 340 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ [all …]
|
D | common_sse2.h | 70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument 78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 133 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE2() argument 165 __m128i* const in3) { in VP8L32bToPlanar_SSE2() argument 169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2() 170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2() 185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar_SSE2()
|
D | common_sse41.h | 40 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE41() argument 66 WEBP_SSE41_SHUFF(G, in2, in3) in VP8PlanarTo24b_SSE41() 91 *in3 = _mm_or_si128(RG3, B3); in VP8PlanarTo24b_SSE41() 104 __m128i* const in3) { in VP8L32bToPlanar_SSE41() argument 111 const __m128i A3 = _mm_shuffle_epi8(*in3, shuff0); in VP8L32bToPlanar_SSE41() 120 *in3 = _mm_unpacklo_epi64(B0, B2); in VP8L32bToPlanar_SSE41()
|
/third_party/cmsis/CMSIS/DSP/Source/ComplexMathFunctions/ |
D | arm_cmplx_conj_q15.c | 110 q31_t in2, in3, in4; /* Temporary input variables */ in arm_cmplx_conj_q15() local 128 in3 = read_q15x2_ia ((q15_t **) &pSrc); in arm_cmplx_conj_q15() 134 in3 = __QASX(0, in3); in arm_cmplx_conj_q15() 139 in3 = __QSAX(0, in3); in arm_cmplx_conj_q15() 145 in3 = ((uint32_t) in3 >> 16) | ((uint32_t) in3 << 16); in arm_cmplx_conj_q15() 150 write_q15x2_ia (&pDst, in3); in arm_cmplx_conj_q15()
|
/third_party/libffi/testsuite/libffi.call/ |
D | return_fl3.c | 10 static float return_fl(float fl1, float fl2, unsigned int in3, float fl4) in return_fl() argument 12 return fl1 + fl2 + in3 + fl4; in return_fl() 20 unsigned int in3; in main() local 27 values[2] = &in3; in main() 35 in3 = 255; in main() 39 printf ("%f vs %f\n", rfl, return_fl(fl1, fl2, in3, fl4)); in main() 40 CHECK(rfl == fl1 + fl2 + in3 + fl4); in main()
|
D | return_dbl1.c | 10 static double return_dbl(double dbl1, float fl2, unsigned int in3, double dbl4) in return_dbl() argument 12 return dbl1 + fl2 + in3 + dbl4; in return_dbl() 21 unsigned int in3; in main() local 28 values[2] = &in3; in main() 36 in3 = 255; in main() 40 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, fl2, in3, dbl4)); in main() 41 CHECK(rdbl == dbl1 + fl2 + in3 + dbl4); in main()
|
D | return_dbl2.c | 10 static double return_dbl(double dbl1, double dbl2, unsigned int in3, double dbl4) in return_dbl() argument 12 return dbl1 + dbl2 + in3 + dbl4; in return_dbl() 20 unsigned int in3; in main() local 27 values[2] = &in3; in main() 35 in3 = 255; in main() 39 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, dbl2, in3, dbl4)); in main() 40 CHECK(rdbl == dbl1 + dbl2 + in3 + dbl4); in main()
|
/third_party/libffi/src/ia64/ |
D | unix.S | 102 zxt1 r16 = in3 113 shr in3 = in3, 8 178 cmp.lt p6, p0 = 8, in3 179 cmp.lt p7, p0 = 16, in3 180 cmp.lt p8, p0 = 24, in3 192 mov out2 = in3 208 cmp.lt p6, p0 = 4, in3 212 cmp.lt p7, p0 = 8, in3 213 cmp.lt p8, p0 = 12, in3 217 cmp.lt p9, p0 = 16, in3 [all …]
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/fp32/ |
D | instance_norm_fp32.h | 26 #define MS_ADDQ_F32_VEC(in1, in2, in3, in4, v1, v2, v3, v4) \ argument 29 in3 = MS_ADDQ_F32(in3, v3); \ 32 #define MS_DIVQ_F32_VEC(in1, in2, in3, in4, v) \ argument 35 in3 = MS_DIVQ_F32(in3, v); \
|
/third_party/libffi/testsuite/libffi.complex/ |
D | return_complex1.inc | 5 static _Complex T_C_TYPE return_c(_Complex T_C_TYPE c1, float fl2, unsigned int in3, _Complex T_C_T… 7 return c1 + fl2 + in3 + c4; 16 unsigned int in3; 23 values[2] = &in3; 31 in3 = 255; 35 rc2 = return_c(c1, fl2, in3, c4);
|
D | return_complex2.inc | 7 unsigned int in3, _Complex T_C_TYPE c4) 9 volatile _Complex T_C_TYPE r = c1 + c2 + in3 + c4; 19 unsigned int in3; 26 values[2] = &in3; 34 in3 = 255; 38 rc2 = return_c(c1, c2, in3, c4);
|
/third_party/openssl/crypto/des/asm/ |
D | des_enc.m4 | 79 #define in3 %i3 125 ! parameter 7 1 for move in1 to in3 126 ! parameter 8 1 for move in3 to in4, 2 for move in4 to in3 127 ! parameter 9 1 for load ks3 and ks2 to in4 and in3 138 ifelse($7,1,{mov in1, in3},{nop}) 142 ifelse($8,1,{mov in3, in4},{}) 143 ifelse($8,2,{mov in4, in3},{}) 157 ifelse($9,1,{LDPTR KS2, in3},{}) 620 ! parameter 5 1: load ks1/ks2 to in3/in4, add 120 to in4 621 ! 2: mov in4 to in3 [all …]
|