Home
last modified time | relevance | path

Searched refs:in4 (Results 1 – 25 of 30) sorted by relevance

12

/external/libvpx/libvpx/vp9/encoder/mips/msa/
Dvp9_fdct8x8_msa.c18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_fht8x8_msa() local
20 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
22 SLLI_4V(in4, in5, in6, in7, 2); in vp9_fht8x8_msa()
26 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa()
27 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
28 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa()
29 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
30 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa()
31 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
34 VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa()
[all …]
Dvp9_fdct4x4_msa.c18 v8i16 in0, in1, in2, in3, in4; in vp9_fwht4x4_msa() local
24 in4 = (in0 - in3) >> 1; in vp9_fwht4x4_msa()
25 SUB2(in4, in1, in4, in2, in1, in2); in vp9_fwht4x4_msa()
33 in4 = (in0 - in1) >> 1; in vp9_fwht4x4_msa()
34 SUB2(in4, in2, in4, in3, in2, in3); in vp9_fwht4x4_msa()
Dvp9_fdct_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
35 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
38 in4, in3); \
63 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
Dvp9_fdct16x16_msa.c368 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11; in postproc_fdct16x8_1d_row() local
371 LD_SH8(temp, 16, in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row()
374 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in postproc_fdct16x8_1d_row()
375 in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row()
380 FDCT_POSTPROC_2V_NEG_H(in4, in5); in postproc_fdct16x8_1d_row()
386 BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, in postproc_fdct16x8_1d_row()
397 in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row()
401 TRANSPOSE8x8_SH_SH(tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7, in postproc_fdct16x8_1d_row()
402 tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7); in postproc_fdct16x8_1d_row()
404 ST_SH8(tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7, out, 16); in postproc_fdct16x8_1d_row()
/external/libvpx/libvpx/vp9/common/mips/msa/
Dvp9_idct8x8_msa.c18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_iht8x8_64_add_msa() local
21 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
23 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa()
24 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
29 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa()
30 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
32 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa()
33 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
34 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa()
35 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
[all …]
/external/libvpx/libvpx/vpx_dsp/mips/
Didct8x8_msa.c15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_msa() local
18 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
21 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa()
22 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
24 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa()
25 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
27 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa()
28 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
30 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa()
31 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
[all …]
Dfwd_txfm_msa.c16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x16_1d_column() local
29 in0, in1, in2, in3, in4, in5, in6, in7, in fdct8x16_1d_column()
32 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x16_1d_column()
36 ADD4(in4, in11, in5, in10, in6, in9, in7, in8, tmp4, tmp5, tmp6, tmp7); in fdct8x16_1d_column()
41 SUB4(in4, in11, in5, in10, in6, in9, in7, in8, in11, in10, in9, in8); in fdct8x16_1d_column()
135 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct16x8_1d_row() local
138 LD_SH8(input, 16, in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row()
140 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in fdct16x8_1d_row()
141 in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row()
145 ADD4(in4, 1, in5, 1, in6, 1, in7, 1, in4, in5, in6, in7); in fdct16x8_1d_row()
[all …]
Dfwd_dct32x32_msa.c16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_load_butterfly() local
23 LD_SH4(input + (28 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
27 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly()
30 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, in fdct8x32_1d_column_load_butterfly()
31 step0, step1, step2, step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
35 ST_SH4(in4, in5, in6, in7, temp_buff + (28 * 8), 8); in fdct8x32_1d_column_load_butterfly()
41 LD_SH4(input + (20 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
45 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly()
48 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, in fdct8x32_1d_column_load_butterfly()
49 step0, step1, step2, step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
[all …]
Dmacros_msa.h412 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
415 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
436 #define ST_H8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) { \ argument
438 ST_H4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
593 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
596 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \
1057 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1060 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1108 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1111 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
[all …]
Dinv_txfm_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
35 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
38 in4, in3); \
64 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
213 #define VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
239 VP9_MADD(in0, in4, in2, in6, k1_m, k0_m, k2_m, k3_m, \
240 in0, in4, in2, in6); \
241 BUTTERFLY_4(in0, in4, in2, in6, tp0_m, tp1_m, tp2_m, tp3_m); \
246 #define VP9_IADST8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
265 ILVRL_H2_SH(in5, in4, in_s1, in_s0); \
[all …]
Dfwd_txfm_msa.h59 #define SRLI_AVE_S_4V_H(in0, in1, in2, in3, in4, in5, in6, in7) { \ argument
63 SRLI_H4_SH(in4, in5, in6, in7, vec4_m, vec5_m, vec6_m, vec7_m, 15); \
66 AVE_SH4_SH(vec4_m, in4, vec5_m, in5, vec6_m, in6, vec7_m, in7, \
67 in4, in5, in6, in7); \
70 #define VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
79 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, \
130 #define FDCT8x16_EVEN(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
138 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, \
Dtxfm_macros_msa.h34 #define DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
39 DOTP_SH4_SW(in0, in1, in0, in1, in4, in4, in5, in5, \
Dvpx_convolve_msa.h109 #define PCKEV_AVG_ST8x4_UB(in1, dst0, in2, dst1, in3, dst2, in4, dst3, \ argument
114 PCKEV_B2_UB(in2, in1, in4, in3, tmp0_m, tmp1_m); \
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_dct_ssse3.c54 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in vp9_fdct8x8_quant_ssse3() local
71 in4 = _mm_slli_epi16(in4, 2); in vp9_fdct8x8_quant_ssse3()
80 in[4] = &in4; in vp9_fdct8x8_quant_ssse3()
96 const __m128i q3 = _mm_add_epi16(in3, in4); in vp9_fdct8x8_quant_ssse3()
97 const __m128i q4 = _mm_sub_epi16(in3, in4); in vp9_fdct8x8_quant_ssse3()
246 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in vp9_fdct8x8_quant_ssse3()
269 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in vp9_fdct8x8_quant_ssse3()
277 in4 = _mm_sub_epi16(in4, sign_in4); in vp9_fdct8x8_quant_ssse3()
285 in4 = _mm_srai_epi16(in4, 1); in vp9_fdct8x8_quant_ssse3()
Dvp9_dct_sse2.c215 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in vp9_fdct8x8_quant_sse2() local
232 in4 = _mm_slli_epi16(in4, 2); in vp9_fdct8x8_quant_sse2()
241 in[4] = &in4; in vp9_fdct8x8_quant_sse2()
257 const __m128i q3 = _mm_add_epi16(in3, in4); in vp9_fdct8x8_quant_sse2()
258 const __m128i q4 = _mm_sub_epi16(in3, in4); in vp9_fdct8x8_quant_sse2()
413 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in vp9_fdct8x8_quant_sse2()
436 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in vp9_fdct8x8_quant_sse2()
444 in4 = _mm_sub_epi16(in4, sign_in4); in vp9_fdct8x8_quant_sse2()
452 in4 = _mm_srai_epi16(in4, 1); in vp9_fdct8x8_quant_sse2()
930 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in fadst8_sse2() local
[all …]
/external/libvpx/libvpx/vp8/common/mips/msa/
Dvp8_macros_msa.h378 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
382 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
577 #define VSHF_B3(RTYPE, in0, in1, in2, in3, in4, in5, mask0, mask1, mask2, \ argument
581 out2 = (RTYPE)__msa_vshf_b((v16i8)mask2, (v16i8)in5, (v16i8)in4); \
974 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
978 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1029 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1033 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1056 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1060 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
[all …]
Dpostproc_msa.c63 #define VP8_TRANSPOSE8x16_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
72 ILVR_B4_SH(in1, in0, in3, in2, in5, in4, in7, in6, \
78 ILVL_B4_SH(in1, in0, in3, in2, in5, in4, in7, in6, \
121 #define TRANSPOSE12x16_B(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
129 ILVR_B2_SH(in5, in4, in7, in6, temp0, temp1); \
144 ILVL_B2_SH(in5, in4, in7, in6, temp0, temp1); \
145 ILVR_D2_UB(temp6, temp2, temp7, temp3, in4, in6); \
160 #define VP8_TRANSPOSE12x8_UB_UB(in0, in1, in2, in3, in4, in5, \ argument
168 ILVR_B2_SH(in5, in4, in7, in6, temp0, temp1); \
174 ILVL_B2_SH(in5, in4, in7, in6, temp6, temp7); \
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_txfm_impl_sse2.h290 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in FDCT8x8_2D() local
299 in4 = _mm_slli_epi16(in4, 2); in FDCT8x8_2D()
315 const __m128i q3 = ADD_EPI16(in3, in4); in FDCT8x8_2D()
316 const __m128i q4 = SUB_EPI16(in3, in4); in FDCT8x8_2D()
522 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in FDCT8x8_2D()
545 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D()
553 in4 = _mm_sub_epi16(in4, sign_in4); in FDCT8x8_2D()
561 in4 = _mm_srai_epi16(in4, 1); in FDCT8x8_2D()
570 store_output(&in4, (output + 4 * 8)); in FDCT8x8_2D()
Dinv_txfm_sse2.c264 #define TRANSPOSE_8X8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
271 const __m128i tr0_4 = _mm_unpacklo_epi16(in4, in5); \
273 const __m128i tr0_6 = _mm_unpackhi_epi16(in4, in5); \
380 #define IDCT8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
397 const __m128i lo_04 = _mm_unpacklo_epi16(in0, in4); \
398 const __m128i hi_04 = _mm_unpackhi_epi16(in0, in4); \
465 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_sse2() local
476 in4 = _mm_load_si128((const __m128i *)(input + 8 * 4)); in vpx_idct8x8_64_add_sse2()
484 TRANSPOSE_8X8(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_sse2()
485 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_sse2()
[all …]
/external/openssh/
Daudit-bsm.c141 struct sockaddr_in *in4; in aug_get_machine() local
153 in4 = (struct sockaddr_in *)ai->ai_addr; in aug_get_machine()
155 memcpy(addr, &in4->sin_addr, sizeof(struct in_addr)); in aug_get_machine()
Daddrmatch.c84 struct sockaddr_in *in4 = (struct sockaddr_in *)sa; in addr_sa_to_xaddr() local
91 if (slen < (socklen_t)sizeof(*in4)) in addr_sa_to_xaddr()
94 memcpy(&xa->v4, &in4->sin_addr, sizeof(xa->v4)); in addr_sa_to_xaddr()
/external/llvm/test/Transforms/LoopUnswitch/
D2007-08-01-LCSSA.ll34 %iftmp.37.0.in4 = icmp eq i32 %tmp138, 0 ; <i1> [#uses=1]
35 br i1 %iftmp.37.0.in4, label %bb250, label %bb166
/external/boringssl/src/crypto/aes/asm/
Daesni-x86_64.pl1160 my ($in0,$in1,$in2,$in3,$in4,$in5)=map("%xmm$_",(10..15));
1481 movdqu 0x40($inp),$in4
1484 pxor $rndkey0,$in4
1507 aesenclast $in4,$inout4
1509 movdqa 0x30(%rsp),$in4
1523 movdqa $in4,$inout3
2719 my ($iv,$in0,$in1,$in2,$in3,$in4)=map("%xmm$_",(10..15));
2838 movdqa $inout4,$in4
2926 pxor $rndkey0,$in4
2944 aesdeclast $in4,$inout5
[all …]
/external/llvm/test/CodeGen/X86/
Dtailcall-stackalign.ll10 define fastcc i32 @tailcaller(i32 %in1, i32 %in2, i32 %in3, i32 %in4) {
/external/compiler-rt/test/dfsan/
Dcustom.cc535 struct in_addr in4; in test_inet_pton() local
536 int ret4 = inet_pton(AF_INET, addr4, &in4); in test_inet_pton()
538 ASSERT_READ_LABEL(&in4, sizeof(in4), i_label); in test_inet_pton()
539 assert(in4.s_addr == 0x0100007f); in test_inet_pton()

12