Home
last modified time | relevance | path

Searched refs:in7 (Results 1 – 25 of 36) sorted by relevance

12

/external/libvpx/libvpx/vp9/encoder/mips/msa/
Dvp9_fdct8x8_msa.c18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_fht8x8_msa() local
20 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
22 SLLI_4V(in4, in5, in6, in7, 2); in vp9_fht8x8_msa()
26 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
27 in5, in6, in7); in vp9_fht8x8_msa()
28 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa()
29 in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
30 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
31 in5, in6, in7); in vp9_fht8x8_msa()
34 VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
[all …]
Dvp9_fdct_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
35 ILVRL_H2_SH(in0, in7, vec1_m, vec0_m); \
38 cnst2_m, cnst3_m, in7, in0, in4, in3); \
52 BUTTERFLY_4(in7, in0, in2, in5, s1_m, s0_m, in2, in5); \
Dvp9_fdct16x16_msa.c368 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11; in postproc_fdct16x8_1d_row() local
371 LD_SH8(temp, 16, in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row()
374 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in postproc_fdct16x8_1d_row()
375 in4, in5, in6, in7); in postproc_fdct16x8_1d_row()
381 FDCT_POSTPROC_2V_NEG_H(in6, in7); in postproc_fdct16x8_1d_row()
386 BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, in postproc_fdct16x8_1d_row()
396 in4, in5, in6, in7); in postproc_fdct16x8_1d_row()
400 TRANSPOSE8x8_SH_SH(tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7, tmp4, in4, in postproc_fdct16x8_1d_row()
401 tmp5, in5, tmp6, in6, tmp7, in7); in postproc_fdct16x8_1d_row()
403 ST_SH8(tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7, out, 16); in postproc_fdct16x8_1d_row()
/external/libvpx/libvpx/vp9/common/mips/msa/
Dvp9_idct8x8_msa.c18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_iht8x8_64_add_msa() local
21 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
23 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
24 in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
29 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
30 in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
32 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_iht8x8_64_add_msa()
33 in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
34 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
35 in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
[all …]
/external/libvpx/libvpx/vpx_dsp/mips/
Didct8x8_msa.c15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_msa() local
18 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
21 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
22 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
24 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
25 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
27 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
28 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
30 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
31 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
[all …]
Dfwd_txfm_msa.c15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_fdct8x8_1_msa() local
18 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_fdct8x8_1_msa()
19 ADD4(in0, in1, in2, in3, in4, in5, in6, in7, in0, in2, in4, in6); in vpx_fdct8x8_1_msa()
31 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x16_1d_column() local
44 LD_SH16(input, src_stride, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in fdct8x16_1d_column()
47 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x16_1d_column()
51 ADD4(in4, in11, in5, in10, in6, in9, in7, in8, tmp4, tmp5, tmp6, tmp7); in fdct8x16_1d_column()
56 SUB4(in4, in11, in5, in10, in6, in9, in7, in8, in11, in10, in9, in8); in fdct8x16_1d_column()
150 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct16x8_1d_row() local
153 LD_SH8(input, 16, in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row()
[all …]
Dmacros_msa.h331 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
334 ST_V4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
488 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
492 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \
1005 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1009 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1062 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1066 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1073 #define ILVR_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, \ argument
1077 ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \
[all …]
Dfwd_dct32x32_msa.c16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_load_butterfly() local
23 LD_SH4(input + (28 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
27 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly()
30 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly()
31 step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
35 ST_SH4(in4, in5, in6, in7, temp_buff + (28 * 8), 8); in fdct8x32_1d_column_load_butterfly()
41 LD_SH4(input + (20 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
45 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly()
48 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly()
49 step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
[all …]
Dinv_txfm_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
35 ILVRL_H2_SH(in0, in7, vec1_m, vec0_m); \
38 cnst2_m, cnst3_m, in7, in0, in4, in3); \
52 BUTTERFLY_4(in7, in0, in2, in5, s1_m, s0_m, in2, in5); \
214 #define VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
227 VP9_MADD(in1, in7, in3, in5, k0_m, k1_m, k2_m, k3_m, in1, in7, in3, in5); \
228 SUB2(in1, in3, in7, in5, res0_m, res1_m); \
238 tp7_m = in7 + in5; \
247 #define VP9_IADST8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
286 ILVRL_H2_SH(in7, in6, in_s1, in_s0); \
[all …]
Dfwd_txfm_msa.h46 #define SRLI_AVE_S_4V_H(in0, in1, in2, in3, in4, in5, in6, in7) \ argument
51 SRLI_H4_SH(in4, in5, in6, in7, vec4_m, vec5_m, vec6_m, vec7_m, 15); \
54 AVE_SH4_SH(vec4_m, in4, vec5_m, in5, vec6_m, in6, vec7_m, in7, in4, in5, \
55 in6, in7); \
58 #define VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
67 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, s0_m, s1_m, s2_m, \
118 #define FDCT8x16_EVEN(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
127 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, s0_m, s1_m, s2_m, \
Dtxfm_macros_msa.h40 #define DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, dst0, \ argument
48 DOTP_SH4_SW(in2, in3, in2, in3, in6, in6, in7, in7, tp5_m, tp6_m, tp7_m, \
Ddeblock_msa.c17 #define VPX_TRANSPOSE8x16_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, out0, \ argument
25 ILVR_B4_SH(in1, in0, in3, in2, in5, in4, in7, in6, temp0, temp1, temp2, \
31 ILVL_B4_SH(in1, in0, in3, in2, in5, in4, in7, in6, temp0, temp1, temp2, \
74 #define TRANSPOSE12x16_B(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, \ argument
82 ILVR_B2_SH(in5, in4, in7, in6, temp0, temp1); \
97 ILVL_B2_SH(in5, in4, in7, in6, temp0, temp1); \
100 in7 = (v16u8)__msa_ilvl_d((v2i64)temp7, (v2i64)temp3); \
113 #define VPX_TRANSPOSE12x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, \ argument
121 ILVR_B2_SH(in5, in4, in7, in6, temp0, temp1); \
127 ILVL_B2_SH(in5, in4, in7, in6, temp6, temp7); \
[all …]
/external/tensorflow/tensorflow/core/kernels/
Daggregate_ops_cpu.h94 typename TTypes<T>::ConstFlat in7) {
96 in7);
107 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) {
109 in7, in8);
120 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) {
122 in7, in8);
133 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8,
136 in7, in8, in9);
202 typename TTypes<T>::ConstFlat in7) {
204 in7);
[all …]
Daggregate_ops_gpu.cu.cc96 typename TTypes<T>::ConstFlat in7) { in operator ()()
98 in7); in operator ()()
109 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in operator ()()
111 in7, in8); in operator ()()
122 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in operator ()()
124 in7, in8); in operator ()()
135 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8, in operator ()()
138 in7, in8, in9); in operator ()()
Daggregate_ops.h136 typename TTypes<T>::ConstFlat in7);
148 typename TTypes<T>::ConstFlat in7) { in Compute()
149 out.device(d) = in1 + in2 + in3 + in4 + in5 + in6 + in7; in Compute()
160 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8);
170 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in Compute()
171 out.device(d) = in1 + in2 + in3 + in4 + in5 + in6 + in7 + in8; in Compute()
184 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8);
194 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in Compute()
195 out.device(d) += in1 + in2 + in3 + in4 + in5 + in6 + in7 + in8; in Compute()
206 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8,
[all …]
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_dct_ssse3.c51 __m128i in7 = _mm_load_si128((const __m128i *)(input + 7 * stride)); in vp9_fdct8x8_quant_ssse3() local
66 in7 = _mm_slli_epi16(in7, 2); in vp9_fdct8x8_quant_ssse3()
75 in[7] = &in7; in vp9_fdct8x8_quant_ssse3()
85 const __m128i q0 = _mm_add_epi16(in0, in7); in vp9_fdct8x8_quant_ssse3()
92 const __m128i q7 = _mm_sub_epi16(in0, in7); in vp9_fdct8x8_quant_ssse3()
241 in7 = _mm_unpackhi_epi64(tr1_3, tr1_7); in vp9_fdct8x8_quant_ssse3()
264 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in vp9_fdct8x8_quant_ssse3()
272 in7 = _mm_sub_epi16(in7, sign_in7); in vp9_fdct8x8_quant_ssse3()
280 in7 = _mm_srai_epi16(in7, 1); in vp9_fdct8x8_quant_ssse3()
Dvp9_dct_intrin_sse2.c116 __m128i in7 = _mm_add_epi16(in[0], in[1]); in fadst4_sse2() local
120 u[2] = _mm_unpacklo_epi16(in7, kZero); in fadst4_sse2()
214 __m128i in7 = _mm_load_si128((const __m128i *)(input + 7 * stride)); in vp9_fdct8x8_quant_sse2() local
229 in7 = _mm_slli_epi16(in7, 2); in vp9_fdct8x8_quant_sse2()
238 in[7] = &in7; in vp9_fdct8x8_quant_sse2()
248 const __m128i q0 = _mm_add_epi16(in0, in7); in vp9_fdct8x8_quant_sse2()
255 const __m128i q7 = _mm_sub_epi16(in0, in7); in vp9_fdct8x8_quant_sse2()
410 in7 = _mm_unpackhi_epi64(tr1_3, tr1_7); in vp9_fdct8x8_quant_sse2()
433 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in vp9_fdct8x8_quant_sse2()
441 in7 = _mm_sub_epi16(in7, sign_in7); in vp9_fdct8x8_quant_sse2()
[all …]
/external/libvpx/libvpx/vp8/common/mips/msa/
Dvp8_macros_msa.h360 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
363 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
953 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
957 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1008 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1012 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1035 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1039 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
1067 #define ILVR_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1071 ILVR_D2(RTYPE, in4, in5, in6, in7, out2, out3); \
[all …]
/external/boringssl/src/crypto/fipsmodule/aes/asm/
Daesp8-ppc.pl672 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
789 lvx_u $in7,$x70,$inp
795 le?vperm $in7,$in7,$in7,$inpperm
798 vxor $out7,$in7,$rndkey0
927 vmr $ivec,$in7
929 lvx_u $in7,$x70,$inp
939 le?vperm $in7,$in7,$in7,$inpperm
958 vxor $out7,$in7,$rndkey0
1073 vmr $ivec,$in7
1100 vmr $ivec,$in7
[all …]
/external/webp/src/dsp/
Dmsa_macro.h310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
313 ST_B4(RTYPE, in4, in5, in6, in7, pdst + 4 * stride, stride); \
873 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
876 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
899 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
902 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
923 #define ILVR_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
926 ILVR_D2(RTYPE, in4, in5, in6, in7, out2, out3); \
984 #define PCKEV_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
987 PCKEV_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
[all …]
/external/libvpx/libvpx/vpx_dsp/ppc/
Dinv_txfm_vsx.c159 #define TRANSPOSE8x8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
167 out6 = vec_mergeh(in6, in7); \
168 out7 = vec_mergel(in6, in7); \
176 in7 = (int16x8_t)vec_mergel((int32x4_t)out5, (int32x4_t)out7); \
183 out6 = vec_perm(in3, in7, tr8_mask0); \
184 out7 = vec_perm(in3, in7, tr8_mask1);
218 #define IDCT8(in0, in1, in2, in3, in4, in5, in6, in7) \ argument
225 STEP8_0(in1, in7, step4, step7, cospi28_v, cospi4_v); \
234 in7 = vec_add(step6, step7); \
243 step7 = in7; \
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_txfm_impl_sse2.h284 __m128i in7 = _mm_load_si128((const __m128i *)(input + 7 * stride)); in FDCT8x8_2D() local
293 in7 = _mm_slli_epi16(in7, 2); in FDCT8x8_2D()
303 const __m128i q0 = ADD_EPI16(in0, in7); in FDCT8x8_2D()
310 const __m128i q7 = SUB_EPI16(in0, in7); in FDCT8x8_2D()
516 in7 = _mm_unpackhi_epi64(tr1_3, tr1_7); in FDCT8x8_2D()
539 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in FDCT8x8_2D()
547 in7 = _mm_sub_epi16(in7, sign_in7); in FDCT8x8_2D()
555 in7 = _mm_srai_epi16(in7, 1); in FDCT8x8_2D()
564 store_output(&in7, (output + 7 * 8)); in FDCT8x8_2D()
Dinv_txfm_sse2.c110 __m128i u[8], v[8], in7; in iadst4_sse2() local
113 in7 = _mm_srli_si128(in[1], 8); in iadst4_sse2()
114 in7 = _mm_add_epi16(in7, in[0]); in iadst4_sse2()
115 in7 = _mm_sub_epi16(in7, in[1]); in iadst4_sse2()
119 u[2] = _mm_unpacklo_epi16(in7, kZero); in iadst4_sse2()
258 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in iadst8_sse2() local
271 in7 = in[6]; in iadst8_sse2()
282 s6 = _mm_unpacklo_epi16(in6, in7); in iadst8_sse2()
283 s7 = _mm_unpackhi_epi16(in6, in7); in iadst8_sse2()
/external/libjpeg-turbo/simd/
Djquanti-altivec.c57 __vector unsigned char in0, in1, in2, in3, in4, in5, in6, in7; in jsimd_convsamp_altivec() local
80 out7 = (__vector short)VEC_UNPACKHU(in7); in jsimd_convsamp_altivec()
/external/libvpx/libvpx/vpx_dsp/arm/
Dhighbd_idct32x32_135_add_neon.c22 int32x4x2_t *const in5, int32x4x2_t *const in6, int32x4x2_t *const in7) { in load_8x8_s32_dual() argument
44 in7->val[0] = vld1q_s32(input); in load_8x8_s32_dual()
45 in7->val[1] = vld1q_s32(input + 4); in load_8x8_s32_dual()
53 int32x4_t *const in7) { in load_4x8_s32_dual() argument
68 *in7 = vld1q_s32(input); in load_4x8_s32_dual()

12