Home
last modified time | relevance | path

Searched refs:in4 (Results 1 – 25 of 95) sorted by relevance

1234

/external/libvpx/libvpx/vp9/encoder/mips/msa/
Dvp9_fdct8x8_msa.c19 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_fht8x8_msa() local
21 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
23 SLLI_4V(in4, in5, in6, in7, 2); in vp9_fht8x8_msa()
27 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
29 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa()
30 in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
31 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
35 VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in4, in vp9_fht8x8_msa()
37 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_fht8x8_msa()
38 in3, in4, in5, in6, in7); in vp9_fht8x8_msa()
[all …]
Dvp9_fdct4x4_msa.c19 v8i16 in0, in1, in2, in3, in4; in vp9_fwht4x4_msa() local
25 in4 = (in0 - in3) >> 1; in vp9_fwht4x4_msa()
26 SUB2(in4, in1, in4, in2, in1, in2); in vp9_fwht4x4_msa()
34 in4 = (in0 - in1) >> 1; in vp9_fwht4x4_msa()
35 SUB2(in4, in2, in4, in3, in2, in3); in vp9_fwht4x4_msa()
Dvp9_fdct_msa.h18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ argument
36 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
38 cnst2_m, cnst3_m, in7, in0, in4, in3); \
62 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
/external/libvpx/libvpx/vp9/common/mips/msa/
Dvp9_idct8x8_msa.c19 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_iht8x8_64_add_msa() local
22 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
24 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
25 in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
30 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
31 in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
33 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in vp9_iht8x8_64_add_msa()
34 in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
35 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vp9_iht8x8_64_add_msa()
36 in4, in5, in6, in7); in vp9_iht8x8_64_add_msa()
[all …]
/external/libvpx/libvpx/vpx_dsp/mips/
Dfwd_txfm_msa.c15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_fdct8x8_1_msa() local
18 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_fdct8x8_1_msa()
19 ADD4(in0, in1, in2, in3, in4, in5, in6, in7, in0, in2, in4, in6); in vpx_fdct8x8_1_msa()
20 ADD2(in0, in2, in4, in6, in0, in4); in vpx_fdct8x8_1_msa()
22 vec_w += __msa_hadd_s_w(in4, in4); in vpx_fdct8x8_1_msa()
31 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x16_1d_column() local
44 LD_SH16(input, src_stride, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in fdct8x16_1d_column()
47 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x16_1d_column()
51 ADD4(in4, in11, in5, in10, in6, in9, in7, in8, tmp4, tmp5, tmp6, tmp7); in fdct8x16_1d_column()
56 SUB4(in4, in11, in5, in10, in6, in9, in7, in8, in11, in10, in9, in8); in fdct8x16_1d_column()
[all …]
Didct8x8_msa.c16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_msa() local
19 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
22 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
23 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
25 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
26 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
28 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
29 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
31 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in0, in1, in2, in3, in vpx_idct8x8_64_add_msa()
32 in4, in5, in6, in7); in vpx_idct8x8_64_add_msa()
[all …]
Dfwd_dct32x32_msa.c17 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_load_butterfly() local
24 LD_SH4(input + (28 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
28 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly()
31 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly()
32 step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
36 ST_SH4(in4, in5, in6, in7, temp_buff + (28 * 8), 8); in fdct8x32_1d_column_load_butterfly()
42 LD_SH4(input + (20 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
46 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly()
49 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, step0, step1, step2, in fdct8x32_1d_column_load_butterfly()
50 step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly()
[all …]
Dmacros_msa.h331 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
334 ST_V4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
488 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
492 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \
1005 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1009 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1062 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1066 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1073 #define ILVR_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, \ argument
1077 ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \
[all …]
/external/libaom/libaom/av1/encoder/mips/msa/
Dfdct4x4_msa.c18 v8i16 in0, in1, in2, in3, in4; in av1_fwht4x4_msa() local
24 in4 = (in0 - in3) >> 1; in av1_fwht4x4_msa()
25 SUB2(in4, in1, in4, in2, in1, in2); in av1_fwht4x4_msa()
33 in4 = (in0 - in1) >> 1; in av1_fwht4x4_msa()
34 SUB2(in4, in2, in4, in3, in2, in3); in av1_fwht4x4_msa()
/external/deqp-deps/glslang/Test/
Dspv.swizzleInversion.frag3 in vec4 in4;
8 vec3 v43 = interpolateAtCentroid(in4.wzx);
9 vec2 v42 = interpolateAtSample(in4.zx, 1);
10 vec4 v44 = interpolateAtOffset(in4.zyxw, vec2(2.0));
11 float v41 = interpolateAtOffset(in4.y, vec2(2.0));
15 float v31 = interpolateAtOffset(in4.y, vec2(2.0));
D450.frag6 in vec4 in4;
12 vec4 v4 = fwidth(in4);
13 v4 = dFdyFine(in4);
16 v4 = fwidthCoarse(in4) + fwidthFine(in4);
D400.frag69 in vec4 in4;
75 vec4 v4 = fwidthCoarse(in4) + fwidthFine(in4); // ERROR
84 vec4 v4 = fwidthCoarse(in4) + fwidthFine(in4);
/external/angle/third_party/vulkan-deps/glslang/src/Test/
Dspv.swizzleInversion.frag3 in vec4 in4;
8 vec3 v43 = interpolateAtCentroid(in4.wzx);
9 vec2 v42 = interpolateAtSample(in4.zx, 1);
10 vec4 v44 = interpolateAtOffset(in4.zyxw, vec2(2.0));
11 float v41 = interpolateAtOffset(in4.y, vec2(2.0));
15 float v31 = interpolateAtOffset(in4.y, vec2(2.0));
D450.frag6 in vec4 in4;
12 vec4 v4 = fwidth(in4);
13 v4 = dFdyFine(in4);
16 v4 = fwidthCoarse(in4) + fwidthFine(in4);
D400.frag69 in vec4 in4;
75 vec4 v4 = fwidthCoarse(in4) + fwidthFine(in4); // ERROR
84 vec4 v4 = fwidthCoarse(in4) + fwidthFine(in4);
/external/tensorflow/tensorflow/core/kernels/
Daggregate_ops_gpu.cu.cc58 typename TTypes<T>::ConstFlat in4) { in operator ()()
59 Add4EigenImpl<GPUDevice, T>::Compute(d, out, in1, in2, in3, in4); in operator ()()
69 typename TTypes<T>::ConstFlat in4, in operator ()()
71 Add5EigenImpl<GPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5); in operator ()()
81 typename TTypes<T>::ConstFlat in4, in operator ()()
84 Add6EigenImpl<GPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5, in6); in operator ()()
94 typename TTypes<T>::ConstFlat in4, in operator ()()
98 Add7EigenImpl<GPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5, in6, in operator ()()
108 typename TTypes<T>::ConstFlat in3, typename TTypes<T>::ConstFlat in4, in operator ()()
111 Add8EigenImpl<GPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5, in6, in operator ()()
[all …]
Daggregate_ops_cpu.h55 typename TTypes<T>::ConstFlat in4) {
56 Add4EigenImpl<CPUDevice, T>::Compute(d, out, in1, in2, in3, in4);
65 typename TTypes<T>::ConstFlat in4,
67 Add5EigenImpl<CPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5);
76 typename TTypes<T>::ConstFlat in4,
79 Add6EigenImpl<CPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5, in6);
88 typename TTypes<T>::ConstFlat in4,
92 Add7EigenImpl<CPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5, in6,
102 typename TTypes<T>::ConstFlat in3, typename TTypes<T>::ConstFlat in4,
105 Add8EigenImpl<CPUDevice, T>::Compute(d, out, in1, in2, in3, in4, in5, in6,
[all …]
Daggregate_ops.h68 typename TTypes<T>::ConstFlat in4);
77 typename TTypes<T>::ConstFlat in4) { in Compute()
78 out.device(d) = in1 + in2 + in3 + in4; in Compute()
88 typename TTypes<T>::ConstFlat in4,
98 typename TTypes<T>::ConstFlat in4, in Compute()
100 out.device(d) = in1 + in2 + in3 + in4 + in5; in Compute()
110 typename TTypes<T>::ConstFlat in4,
121 typename TTypes<T>::ConstFlat in4, in Compute()
124 out.device(d) = in1 + in2 + in3 + in4 + in5 + in6; in Compute()
134 typename TTypes<T>::ConstFlat in4,
[all …]
/external/deqp-deps/glslang/Test/baseResults/
Dspv.swizzleInversion.frag.out15 Name 12 "in4"
23 Decorate 12(in4) Location 0
32 12(in4): 11(ptr) Variable Input
55 13: 10(fvec4) ExtInst 1(GLSL.std.450) 76(InterpolateAtCentroid) 12(in4)
58 20: 10(fvec4) ExtInst 1(GLSL.std.450) 77(InterpolateAtSample) 12(in4) 19
61 26: 10(fvec4) ExtInst 1(GLSL.std.450) 78(InterpolateAtOffset) 12(in4) 25
64 33: 32(ptr) AccessChain 12(in4) 31
73 44: 32(ptr) AccessChain 12(in4) 31
/external/angle/third_party/vulkan-deps/glslang/src/Test/baseResults/
Dspv.swizzleInversion.frag.out15 Name 12 "in4"
23 Decorate 12(in4) Location 0
32 12(in4): 11(ptr) Variable Input
55 13: 10(fvec4) ExtInst 1(GLSL.std.450) 76(InterpolateAtCentroid) 12(in4)
58 20: 10(fvec4) ExtInst 1(GLSL.std.450) 77(InterpolateAtSample) 12(in4) 19
61 26: 10(fvec4) ExtInst 1(GLSL.std.450) 78(InterpolateAtOffset) 12(in4) 25
64 33: 32(ptr) AccessChain 12(in4) 31
73 44: 32(ptr) AccessChain 12(in4) 31
/external/boringssl/src/crypto/fipsmodule/aes/asm/
Daesp8-ppc.pl674 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
678 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment
782 lvx_u $in4,$x40,$inp
789 le?vperm $in4,$in4,$in4,$inpperm
796 vxor $out4,$in4,$rndkey0
903 vxor $in4,$in4,v31
920 vncipherlast $out5,$out5,$in4
922 lvx_u $in4,$x40,$inp
927 le?vperm $in4,$in4,$in4,$inpperm
951 vxor $out4,$in4,$rndkey0
[all …]
/external/libaom/libaom/aom_dsp/mips/
Dmacros_msa.h423 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
426 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
449 #define ST_H8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
452 ST_H4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
616 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
620 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \
1116 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1120 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1173 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1177 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
[all …]
/external/libvpx/libvpx/vp8/common/mips/msa/
Dvp8_macros_msa.h360 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
363 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
558 #define VSHF_B3(RTYPE, in0, in1, in2, in3, in4, in5, mask0, mask1, mask2, \ argument
562 out2 = (RTYPE)__msa_vshf_b((v16i8)mask2, (v16i8)in5, (v16i8)in4); \
953 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
957 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1008 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1012 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1035 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \ argument
1039 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
[all …]
/external/libaom/libaom/aom_dsp/x86/
Dfwd_txfm_impl_sse2.h256 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in FDCT8x8_2D() local
265 in4 = _mm_slli_epi16(in4, 2); in FDCT8x8_2D()
281 const __m128i q3 = ADD_EPI16(in3, in4); in FDCT8x8_2D()
282 const __m128i q4 = SUB_EPI16(in3, in4); in FDCT8x8_2D()
488 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in FDCT8x8_2D()
511 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D()
519 in4 = _mm_sub_epi16(in4, sign_in4); in FDCT8x8_2D()
527 in4 = _mm_srai_epi16(in4, 1); in FDCT8x8_2D()
536 store_output(&in4, (output + 4 * 8)); in FDCT8x8_2D()
/external/llvm-project/llvm/test/Transforms/Reassociate/
Dlong-chains.ll4 define i8 @longchain(i8 %in1, i8 %in2, i8 %in3, i8 %in4, i8 %in5, i8 %in6, i8 %in7, i8 %in8, i8 %in…
7 %tmp3 = add i8 %tmp2, %in4
9 %tmp5 = add i8 %tmp4, %in4

1234