Home
last modified time | relevance | path

Searched refs:in3 (Results 1 – 25 of 121) sorted by relevance

12345

/third_party/openh264/codec/common/inc/
Dmsa_macros.h231 #define MSA_ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
234 MSA_ST_V2(RTYPE, in2, in3, (pdst) + 2 * (stride), stride); \
237 #define MSA_ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
239 MSA_ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
359 #define MSA_ST_D8(in0, in1, in2, in3, idx0, idx1, idx2, idx3, \ argument
363 MSA_ST_D4(in2, in3, idx4, idx5, idx6, idx7, pdst + 4 * stride, stride) \
379 #define MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument
382 MSA_VSHF_B(RTYPE, in2, in3, mask1, out1) \
385 #define MSA_VSHF_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
388 MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1); \
[all …]
/third_party/ffmpeg/libavutil/mips/
Dgeneric_macros_msa.h245 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument
250 SW(in3, (pdst) + 3 * stride); \
260 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument
265 SD(in3, (pdst) + 3 * stride); \
373 #define ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
376 ST_V2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \
383 #define ST_V6(RTYPE, in0, in1, in2, in3, in4, in5, pdst, stride) \ argument
385 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
390 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
392 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
[all …]
/third_party/ffmpeg/libavcodec/mips/
Didctdsp_msa.c28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_pixels_clamped_msa() local
30 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa()
31 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa()
32 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_pixels_clamped_msa()
38 in3_d = __msa_copy_u_d((v2i64) in3, 0); in put_pixels_clamped_msa()
52 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_signed_pixels_clamped_msa() local
54 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa()
59 in3 += 128; in put_signed_pixels_clamped_msa()
65 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa()
66 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_signed_pixels_clamped_msa()
[all …]
Dh263dsp_msa.c32 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in h263_h_loop_filter_msa() local
38 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in h263_h_loop_filter_msa()
39 TRANSPOSE8x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in h263_h_loop_filter_msa()
40 in0, in3, in2, in1); in h263_h_loop_filter_msa()
44 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_h_loop_filter_msa()
82 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa()
83 in3 = (v16u8) __msa_adds_s_b((v16i8) in3, (v16i8) d0); in h263_h_loop_filter_msa()
84 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa()
86 ILVR_B2_SH(in3, in0, in1, in2, temp0, temp1); in h263_h_loop_filter_msa()
88 in3 = (v16u8) __msa_ilvl_h(temp1, temp0); in h263_h_loop_filter_msa()
[all …]
Dhevc_idct_msa.c94 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument
104 ILVR_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \
106 ILVL_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \
146 PCKEV_H2_SH(sum1_l, sum1_r, sum2_l, sum2_r, in3, in4); \
334 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_idct_8x8_msa() local
336 LD_SH8(coeffs, 8, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa()
337 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in hevc_idct_8x8_msa()
338 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in hevc_idct_8x8_msa()
339 in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa()
340 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 12); in hevc_idct_8x8_msa()
[all …]
Dvp9_idct_msa.c86 #define VP9_DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
94 DOTP_SH4_SW(in2, in3, in2, in3, in6, in6, in7, in7, \
116 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
134 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
137 in4, in3); \
163 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
229 #define VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3) \ argument
240 ADD4(res0_m, in0, res1_m, in1, res2_m, in2, res3_m, in3, \
247 #define VP9_IDCT4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
261 step1_m = __msa_ilvr_h(in3, in1); \
[all …]
Dvc1dsp_msa.c30 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x8_msa() local
44 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_msa()
48 UNPCK_SH_SW(in3, in_r3, in_l3); in ff_vc1_inv_trans_8x8_msa()
134 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_msa()
137 ST_SH8(in0, in1, in2, in3, in4, in5, in6, in7, block, 8); in ff_vc1_inv_trans_8x8_msa()
142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_4x8_msa() local
159 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_4x8_msa()
163 UNPCK_R_SH_SW(in3, in_r3); in ff_vc1_inv_trans_4x8_msa()
191 in0, in1, in2, in3); in ff_vc1_inv_trans_4x8_msa()
192 ST_D8(in0, in1, in2, in3, 0, 1, 0, 1, 0, 1, 0, 1, block, 8); in ff_vc1_inv_trans_4x8_msa()
[all …]
Dsimple_idct_msa.c28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in simple_idct_msa() local
41 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
46 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa()
47 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
48 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7; in simple_idct_msa()
79 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa()
111 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp, in simple_idct_msa()
121 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa()
122 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
152 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa()
[all …]
Dvp8_idct_msa.c29 #define VP8_IDCT_1D_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
40 c_tmp2_m = in3 + (((in3) * const_cospi8sqrt2minus1_m) >> 16); \
43 d_tmp2_m = ((in3) * sinpi8_sqrt2_m) >> 16; \
51 v4i32 in0, in1, in2, in3, hz0, hz1, hz2, hz3, vt0, vt1, vt2, vt3; in ff_vp8_idct_add_msa() local
60 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_idct_add_msa()
61 VP8_IDCT_1D_W(in0, in1, in2, in3, hz0, hz1, hz2, hz3); in ff_vp8_idct_add_msa()
108 v4i32 in0, in1, in2, in3, a1, b1, c1, d1; in ff_vp8_luma_dc_wht_msa() local
114 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_luma_dc_wht_msa()
115 BUTTERFLY_4(in0, in1, in2, in3, a1, b1, c1, d1); in ff_vp8_luma_dc_wht_msa()
Dmpegvideoencdsp_msa.c27 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in sum_u8src_16width_msa() local
30 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in sum_u8src_16width_msa()
34 HADD_UB4_UB(in0, in1, in2, in3, in0, in1, in2, in3); in sum_u8src_16width_msa()
42 sum += HADD_UH_U32(in3); in sum_u8src_16width_msa()
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/spirv_assembly/
DvktSpvAsmTrinaryMinMaxTests.cpp196 static inline void runOpFunc (F f, void* out, const void* in1, const void* in2, const void* in3) in runOpFunc() argument
198 …interpret_cast<const T*>(in1), *reinterpret_cast<const T*>(in2), *reinterpret_cast<const T*>(in3)); in runOpFunc()
202 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (min3<deInt8>, out, in1,… in minInt8() argument
203 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (max3<deInt8>, out, in1,… in maxInt8() argument
204 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (mid3<deInt8>, out, in1,… in midInt8() argument
205 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (min3<deUint8>, out, in1… in minUint8() argument
206 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (max3<deUint8>, out, in1… in maxUint8() argument
207 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (mid3<deUint8>, out, in1… in midUint8() argument
208 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (min3<deInt16>, out, in1… in minInt16() argument
209 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (max3<deInt16>, out, in1… in maxInt16() argument
[all …]
/third_party/skia/third_party/externals/libwebp/src/dsp/
Dmsa_macro.h168 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument
176 SW(in3, ptmp); \
202 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument
210 SD(in3, ptmp); \
303 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
305 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
312 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \
338 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
340 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
[all …]
Dcommon_sse2.h70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument
78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
133 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE2() argument
165 __m128i* const in3) { in VP8L32bToPlanar_SSE2() argument
169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2()
170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2()
185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar_SSE2()
/third_party/flutter/skia/third_party/externals/libwebp/src/dsp/
Dmsa_macro.h168 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument
176 SW(in3, ptmp); \
202 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument
210 SD(in3, ptmp); \
303 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
305 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
312 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \
338 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
340 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
[all …]
Dcommon_sse2.h70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument
78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
133 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE2() argument
165 __m128i* const in3) { in VP8L32bToPlanar_SSE2() argument
169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2()
170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2()
185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar_SSE2()
Dcommon_sse41.h40 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE41() argument
66 WEBP_SSE41_SHUFF(G, in2, in3) in VP8PlanarTo24b_SSE41()
91 *in3 = _mm_or_si128(RG3, B3); in VP8PlanarTo24b_SSE41()
104 __m128i* const in3) { in VP8L32bToPlanar_SSE41() argument
111 const __m128i A3 = _mm_shuffle_epi8(*in3, shuff0); in VP8L32bToPlanar_SSE41()
120 *in3 = _mm_unpacklo_epi64(B0, B2); in VP8L32bToPlanar_SSE41()
/third_party/cmsis/CMSIS/DSP/Source/ComplexMathFunctions/
Darm_cmplx_conj_q15.c110 q31_t in2, in3, in4; /* Temporary input variables */ in arm_cmplx_conj_q15() local
128 in3 = read_q15x2_ia ((q15_t **) &pSrc); in arm_cmplx_conj_q15()
134 in3 = __QASX(0, in3); in arm_cmplx_conj_q15()
139 in3 = __QSAX(0, in3); in arm_cmplx_conj_q15()
145 in3 = ((uint32_t) in3 >> 16) | ((uint32_t) in3 << 16); in arm_cmplx_conj_q15()
150 write_q15x2_ia (&pDst, in3); in arm_cmplx_conj_q15()
/third_party/libffi/testsuite/libffi.call/
Dreturn_fl3.c10 static float return_fl(float fl1, float fl2, unsigned int in3, float fl4) in return_fl() argument
12 return fl1 + fl2 + in3 + fl4; in return_fl()
20 unsigned int in3; in main() local
27 values[2] = &in3; in main()
35 in3 = 255; in main()
39 printf ("%f vs %f\n", rfl, return_fl(fl1, fl2, in3, fl4)); in main()
40 CHECK(rfl == fl1 + fl2 + in3 + fl4); in main()
Dreturn_dbl1.c10 static double return_dbl(double dbl1, float fl2, unsigned int in3, double dbl4) in return_dbl() argument
12 return dbl1 + fl2 + in3 + dbl4; in return_dbl()
21 unsigned int in3; in main() local
28 values[2] = &in3; in main()
36 in3 = 255; in main()
40 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, fl2, in3, dbl4)); in main()
41 CHECK(rdbl == dbl1 + fl2 + in3 + dbl4); in main()
Dreturn_dbl2.c10 static double return_dbl(double dbl1, double dbl2, unsigned int in3, double dbl4) in return_dbl() argument
12 return dbl1 + dbl2 + in3 + dbl4; in return_dbl()
20 unsigned int in3; in main() local
27 values[2] = &in3; in main()
35 in3 = 255; in main()
39 printf ("%f vs %f\n", rdbl, return_dbl(dbl1, dbl2, in3, dbl4)); in main()
40 CHECK(rdbl == dbl1 + dbl2 + in3 + dbl4); in main()
/third_party/libffi/src/ia64/
Dunix.S102 zxt1 r16 = in3
113 shr in3 = in3, 8
178 cmp.lt p6, p0 = 8, in3
179 cmp.lt p7, p0 = 16, in3
180 cmp.lt p8, p0 = 24, in3
192 mov out2 = in3
208 cmp.lt p6, p0 = 4, in3
212 cmp.lt p7, p0 = 8, in3
213 cmp.lt p8, p0 = 12, in3
217 cmp.lt p9, p0 = 16, in3
[all …]
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/fp32/
Dinstance_norm_fp32.h26 #define MS_ADDQ_F32_VEC(in1, in2, in3, in4, v1, v2, v3, v4) \ argument
29 in3 = MS_ADDQ_F32(in3, v3); \
32 #define MS_DIVQ_F32_VEC(in1, in2, in3, in4, v) \ argument
35 in3 = MS_DIVQ_F32(in3, v); \
/third_party/libffi/testsuite/libffi.complex/
Dreturn_complex1.inc5 static _Complex T_C_TYPE return_c(_Complex T_C_TYPE c1, float fl2, unsigned int in3, _Complex T_C_T…
7 return c1 + fl2 + in3 + c4;
16 unsigned int in3;
23 values[2] = &in3;
31 in3 = 255;
35 rc2 = return_c(c1, fl2, in3, c4);
Dreturn_complex2.inc7 unsigned int in3, _Complex T_C_TYPE c4)
9 volatile _Complex T_C_TYPE r = c1 + c2 + in3 + c4;
19 unsigned int in3;
26 values[2] = &in3;
34 in3 = 255;
38 rc2 = return_c(c1, c2, in3, c4);
/third_party/openssl/crypto/des/asm/
Ddes_enc.m479 #define in3 %i3
125 ! parameter 7 1 for move in1 to in3
126 ! parameter 8 1 for move in3 to in4, 2 for move in4 to in3
127 ! parameter 9 1 for load ks3 and ks2 to in4 and in3
138 ifelse($7,1,{mov in1, in3},{nop})
142 ifelse($8,1,{mov in3, in4},{})
143 ifelse($8,2,{mov in4, in3},{})
157 ifelse($9,1,{LDPTR KS2, in3},{})
620 ! parameter 5 1: load ks1/ks2 to in3/in4, add 120 to in4
621 ! 2: mov in4 to in3
[all …]

12345