Home
last modified time | relevance | path

Searched refs:sum1 (Results 1 – 25 of 130) sorted by relevance

123456

/external/libvpx/libvpx/vpx_dsp/x86/
Dconvolve_avx2.h61 __m256i sum1, sum2; in convolve8_16_avx2() local
66 sum1 = _mm256_add_epi16(x0, x2); in convolve8_16_avx2()
69 sum1 = _mm256_add_epi16(sum1, k_64); in convolve8_16_avx2()
70 sum1 = _mm256_adds_epi16(sum1, sum2); in convolve8_16_avx2()
72 sum1 = _mm256_srai_epi16(sum1, 7); in convolve8_16_avx2()
73 return sum1; in convolve8_16_avx2()
88 __m128i sum1, sum2; in convolve8_8_avx2() local
93 sum1 = _mm_add_epi16(x0, x2); in convolve8_8_avx2()
96 sum1 = _mm_add_epi16(sum1, k_64); in convolve8_8_avx2()
97 sum1 = _mm_adds_epi16(sum1, sum2); in convolve8_8_avx2()
[all …]
Dconvolve_ssse3.h51 __m128i sum1, sum2; in convolve8_8_ssse3() local
56 sum1 = _mm_add_epi16(x0, x2); in convolve8_8_ssse3()
59 sum1 = _mm_add_epi16(sum1, k_64); in convolve8_8_ssse3()
60 sum1 = _mm_adds_epi16(sum1, sum2); in convolve8_8_ssse3()
62 sum1 = _mm_srai_epi16(sum1, 7); in convolve8_8_ssse3()
63 return sum1; in convolve8_8_ssse3()
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopUnrollAndJam/
Dunprofitable.ll30 %sum1 = phi i32 [ 0, %for.outer ], [ %add, %for.inner2 ]
33 %add = add i32 %0, %sum1
77 %sum1 = phi i32 [ 0, %for.outer ], [ %add, %for.inner ]
80 %add = add i32 %0, %sum1
121 %sum1 = phi i32 [ 0, %for.outer ], [ %add, %for.inner ]
124 %add = add i32 %0, %sum1
125 %add0 = add i32 %0, %sum1
126 %add1 = add i32 %0, %sum1
127 %add2 = add i32 %0, %sum1
128 %add3 = add i32 %0, %sum1
[all …]
Ddisable.ll33 %sum1.025 = phi i32 [ %0, %for.outer ], [ %add, %for.inner ]
36 %add = add i32 %1, %sum1.025
78 %sum1.029 = phi i32 [ %sum1.1, %for.inner ], [ 0, %for.outer ]
82 %sub = add i32 %sum1.029, 10
84 %sum1.1 = select i1 %tobool7, i32 %sum1.029, i32 %add
90 %sum1.1.lcssa = phi i32 [ 0, %for.outer ], [ %sum1.1, %for.inner ]
92 store i32 %sum1.1.lcssa, i32* %arrayidx11, align 4
178 %sum1.020 = phi i32 [ 0, %for.outer ], [ %add, %for.inner ]
181 %add = add i32 %0, %sum1.020
387 %sum1 = phi i32 [ 0, %for.outer ], [ %add, %for.inner ]
[all …]
/external/libopus/celt/x86/
Dpitch_sse4_1.c117 __m128i sum0, sum1, sum2, sum3, vecSum; in xcorr_kernel_sse4_1() local
123 sum1 = _mm_setzero_si128(); in xcorr_kernel_sse4_1()
136 sum1 = _mm_add_epi32(sum1, _mm_madd_epi16(vecX, vecY1)); in xcorr_kernel_sse4_1()
144 sum1 = _mm_add_epi32(sum1, _mm_unpackhi_epi64( sum1, sum1)); in xcorr_kernel_sse4_1()
145 sum1 = _mm_add_epi32(sum1, _mm_shufflelo_epi16( sum1, 0x0E)); in xcorr_kernel_sse4_1()
153 vecSum = _mm_unpacklo_epi64(_mm_unpacklo_epi32(sum0, sum1), in xcorr_kernel_sse4_1()
170 sum1 = _mm_mullo_epi32(vecX1, vecY1); in xcorr_kernel_sse4_1()
174 sum0 = _mm_add_epi32(sum0, sum1); in xcorr_kernel_sse4_1()
/external/libaom/libaom/av1/common/arm/
Dconvolve_neon.h24 int32x4_t sum0, sum1; in wiener_convolve8_vert_4x8() local
43 sum1 = vmull_n_s16(vget_high_s16(ss0), filter_y[0]); in wiener_convolve8_vert_4x8()
44 sum1 = vmlal_n_s16(sum1, vget_high_s16(ss1), filter_y[1]); in wiener_convolve8_vert_4x8()
45 sum1 = vmlal_n_s16(sum1, vget_high_s16(ss2), filter_y[2]); in wiener_convolve8_vert_4x8()
46 sum1 = vmlal_n_s16(sum1, vget_high_s16(s3), filter_y[3]); in wiener_convolve8_vert_4x8()
49 sum1 = vsubq_s32(sum1, round_vec); in wiener_convolve8_vert_4x8()
53 sum1 = vrshlq_s32(sum1, round_bits); in wiener_convolve8_vert_4x8()
56 sum1 = vmaxq_s32(sum1, zero); in wiener_convolve8_vert_4x8()
60 tmp1 = vqmovn_u32(vreinterpretq_u32_s32(sum1)); in wiener_convolve8_vert_4x8()
/external/flac/libFLAC/
Dlpc_intrin_sse.c97 __m128 sum0, sum1; in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_8_new() local
104 sum1 = _mm_setzero_ps(); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_8_new()
112 sum1 = _mm_add_ps(sum1, _mm_mul_ps(d1, d)); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_8_new()
127 sum1 = _mm_add_ps(sum1, _mm_mul_ps(d, d1)); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_8_new()
133 _mm_storeu_ps(autoc+4, sum1); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_8_new()
141 __m128 sum0, sum1, sum2; in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_12_new() local
148 sum1 = _mm_setzero_ps(); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_12_new()
158 sum1 = _mm_add_ps(sum1, _mm_mul_ps(d1, d)); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_12_new()
178 sum1 = _mm_add_ps(sum1, _mm_mul_ps(d, d1)); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_12_new()
184 _mm_storeu_ps(autoc+4, sum1); in FLAC__lpc_compute_autocorrelation_intrin_sse_lag_12_new()
[all …]
/external/python/cpython2/Modules/zlib/
Dadler32.c148 unsigned long sum1; local
159 sum1 = adler1 & 0xffff;
160 sum2 = rem * sum1;
162 sum1 += (adler2 & 0xffff) + BASE - 1;
164 if (sum1 >= BASE) sum1 -= BASE;
165 if (sum1 >= BASE) sum1 -= BASE;
168 return sum1 | (sum2 << 16);
/external/zlib/src/
Dadler32.c148 unsigned long sum1; local
159 sum1 = adler1 & 0xffff;
160 sum2 = rem * sum1;
162 sum1 += (adler2 & 0xffff) + BASE - 1;
164 if (sum1 >= BASE) sum1 -= BASE;
165 if (sum1 >= BASE) sum1 -= BASE;
168 return sum1 | (sum2 << 16);
/external/libaom/libaom/aom_dsp/x86/
Dsad4d_avx2.c176 uint32_t sum1[4]; in aom_sad32x64x4d_avx2() local
188 aom_sad32x32x4d_avx2(src, src_stride, rf, ref_stride, sum1); in aom_sad32x64x4d_avx2()
189 res[0] = sum0[0] + sum1[0]; in aom_sad32x64x4d_avx2()
190 res[1] = sum0[1] + sum1[1]; in aom_sad32x64x4d_avx2()
191 res[2] = sum0[2] + sum1[2]; in aom_sad32x64x4d_avx2()
192 res[3] = sum0[3] + sum1[3]; in aom_sad32x64x4d_avx2()
200 uint32_t sum1[4]; in aom_sad64x32x4d_avx2() local
213 aom_sad32x32x4d_avx2(src, src_stride, rf, ref_stride, sum1); in aom_sad64x32x4d_avx2()
214 res[0] = sum0[0] + sum1[0]; in aom_sad64x32x4d_avx2()
215 res[1] = sum0[1] + sum1[1]; in aom_sad64x32x4d_avx2()
[all …]
Dsad_impl_avx2.c98 __m128i sum0, sum1; in aom_sad64x128x4d_avx2() local
111 sad64x64x4d(src, src_stride, rf, ref_stride, &sum1); in aom_sad64x128x4d_avx2()
112 sum0 = _mm_add_epi32(sum0, sum1); in aom_sad64x128x4d_avx2()
119 __m128i sum0, sum1; in aom_sad128x64x4d_avx2() local
133 sad64x64x4d(src, src_stride, rf, ref_stride, &sum1); in aom_sad128x64x4d_avx2()
134 sum0 = _mm_add_epi32(sum0, sum1); in aom_sad128x64x4d_avx2()
143 uint32_t sum1[4]; in aom_sad128x128x4d_avx2() local
155 aom_sad128x64x4d_avx2(src, src_stride, rf, ref_stride, sum1); in aom_sad128x128x4d_avx2()
156 res[0] = sum0[0] + sum1[0]; in aom_sad128x128x4d_avx2()
157 res[1] = sum0[1] + sum1[1]; in aom_sad128x128x4d_avx2()
[all …]
/external/libvpx/libvpx/vpx_dsp/arm/
Dvariance_neon.c212 int sum1, sum2; in vpx_variance32x64_neon() local
215 &sum1); in vpx_variance32x64_neon()
220 sum1 += sum2; in vpx_variance32x64_neon()
221 return *sse - (unsigned int)(((int64_t)sum1 * sum1) >> 11); in vpx_variance32x64_neon()
227 int sum1, sum2; in vpx_variance64x32_neon() local
230 &sum1); in vpx_variance64x32_neon()
235 sum1 += sum2; in vpx_variance64x32_neon()
236 return *sse - (unsigned int)(((int64_t)sum1 * sum1) >> 11); in vpx_variance64x32_neon()
242 int sum1, sum2; in vpx_variance64x64_neon() local
246 &sum1); in vpx_variance64x64_neon()
[all …]
Dloopfilter_neon.c186 uint16x8_t *sum1) { in filter_update_16() argument
188 *sum1 = vsubw_u8(*sum1, vget_high_u8(sub0)); in filter_update_16()
190 *sum1 = vsubw_u8(*sum1, vget_high_u8(sub1)); in filter_update_16()
192 *sum1 = vaddw_u8(*sum1, vget_high_u8(add0)); in filter_update_16()
194 *sum1 = vaddw_u8(*sum1, vget_high_u8(add1)); in filter_update_16()
208 const uint8x16_t add1, uint16x8_t *sum0, uint16x8_t *sum1) { in calc_7_tap_filter_16_kernel() argument
209 filter_update_16(sub0, sub1, add0, add1, sum0, sum1); in calc_7_tap_filter_16_kernel()
210 return vcombine_u8(vrshrn_n_u16(*sum0, 3), vrshrn_n_u16(*sum1, 3)); in calc_7_tap_filter_16_kernel()
224 uint16x8_t *sum0, uint16x8_t *sum1) { in apply_15_tap_filter_16_kernel() argument
226 filter_update_16(sub0, sub1, add0, add1, sum0, sum1); in apply_15_tap_filter_16_kernel()
[all …]
/external/speex/libspeexdsp/
Dresample_sse.h104 __m128d sum1 = _mm_setzero_pd(); in interpolate_product_double() local
113 sum1 = _mm_add_pd(sum1, _mm_cvtps_pd(t)); in interpolate_product_double()
117 sum1 = _mm_add_pd(sum1, _mm_cvtps_pd(t)); in interpolate_product_double()
120 sum1 = _mm_mul_pd(f1, sum1); in interpolate_product_double()
122 sum = _mm_add_pd(sum1, sum2); in interpolate_product_double()
/external/libaom/libaom/aom_dsp/arm/
Dvariance_neon.c101 int sum1, sum2; in aom_variance32x64_neon() local
103 variance_neon_w8(a, a_stride, b, b_stride, 32, 32, &sse1, &sum1); in aom_variance32x64_neon()
107 sum1 += sum2; in aom_variance32x64_neon()
108 return *sse - (unsigned int)(((int64_t)sum1 * sum1) >> 11); in aom_variance32x64_neon()
114 int sum1, sum2; in aom_variance64x32_neon() local
116 variance_neon_w8(a, a_stride, b, b_stride, 64, 16, &sse1, &sum1); in aom_variance64x32_neon()
120 sum1 += sum2; in aom_variance64x32_neon()
121 return *sse - (unsigned int)(((int64_t)sum1 * sum1) >> 11); in aom_variance64x32_neon()
127 int sum1, sum2; in aom_variance64x64_neon() local
130 variance_neon_w8(a, a_stride, b, b_stride, 64, 16, &sse1, &sum1); in aom_variance64x64_neon()
[all …]
/external/webp/src/dsp/
Dssim_sse2.c37 const __m128i sum1 = _mm_madd_epi16(C0, C0); in SubtractAndSquare_SSE2() local
39 *sum = _mm_add_epi32(sum1, sum2); in SubtractAndSquare_SSE2()
52 __m128i sum1; in AccumulateSSE_SSE2() local
62 SubtractAndSquare_SSE2(a0, b0, &sum1); in AccumulateSSE_SSE2()
63 sum = _mm_add_epi32(sum, sum1); in AccumulateSSE_SSE2()
70 SubtractAndSquare_SSE2(a0, b0, &sum1); in AccumulateSSE_SSE2()
71 sum = _mm_add_epi32(sum, sum1); in AccumulateSSE_SSE2()
/external/libavc/common/x86/
Dih264_resi_trans_quant_sse42.c114 __m128i sum0, sum1, sum2, cmp0, cmp1; in ih264_resi_trans_quant_4x4_sse42() local
303 sum1 = _mm_hadd_epi16(sum0, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42()
304 sum2 = _mm_hadd_epi16(sum1, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42()
316 sum1 = _mm_hadd_epi16(sum0, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42()
317 sum2 = _mm_hadd_epi16(sum1, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42()
388 __m128i cmp0, cmp1, sum0, sum1, sum2; in ih264_resi_trans_quant_chroma_4x4_sse42() local
588 sum1 = _mm_hadd_epi16(sum0, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42()
589 sum2 = _mm_hadd_epi16(sum1, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42()
601 sum1 = _mm_hadd_epi16(sum0, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42()
602 sum2 = _mm_hadd_epi16(sum1, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42()
[all …]
/external/google-breakpad/src/common/mac/
Dmacho_id.cc89 #define DO1(buf,i) {sum1 += (buf)[i]; sum2 += sum1;} in UpdateCRC()
95 uint32_t sum1 = crc_ & 0xFFFF; in UpdateCRC() local
106 sum1 %= MOD_ADLER; in UpdateCRC()
118 sum1 += *bytes++; in UpdateCRC()
119 sum2 += sum1; in UpdateCRC()
121 sum1 %= MOD_ADLER; in UpdateCRC()
123 crc_ = (sum2 << 16) | sum1; in UpdateCRC()
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Hexagon/
Dswp-multi-loops.ll28 %sum1.026 = phi i32 [ %add8, %for.inc12 ], [ 0, %for.body3.lr.ph.preheader ]
49 %sum1.121 = phi i32 [ %sum1.026, %for.end ], [ %add8, %for.body6 ]
53 %add8 = add nsw i32 %1, %sum1.121
68 %sum1.0.lcssa = phi i32 [ 0, %entry ], [ %add8, %for.end14.loopexit ]
70 %add15 = add nsw i32 %sum1.0.lcssa, %sum.0.lcssa
/external/llvm/test/CodeGen/Hexagon/
Dswp-multi-loops.ll28 %sum1.026 = phi i32 [ %add8, %for.inc12 ], [ 0, %for.body3.lr.ph.preheader ]
49 %sum1.121 = phi i32 [ %sum1.026, %for.end ], [ %add8, %for.body6 ]
53 %add8 = add nsw i32 %1, %sum1.121
68 %sum1.0.lcssa = phi i32 [ 0, %entry ], [ %add8, %for.end14.loopexit ]
70 %add15 = add nsw i32 %sum1.0.lcssa, %sum.0.lcssa
/external/libopus/celt/arm/
Dcelt_pitch_xcorr_arm_gnu.s361 @ r7 = opus_val32 sum1
429 @ maxcorr = max(maxcorr, sum0, sum1, sum2, sum3)
447 @ {r10, r11} = {sum0, sum1} = {0, 0}
457 SMLABT r11, r6, r8, r11 @ sum1 = MAC16_16(sum1, x_0, y_1)
461 SMLATB r11, r6, r9, r11 @ sum1 = MAC16_16(sum1, x_1, y_2)
464 SMLABT r11, r7, r9, r11 @ sum1 = MAC16_16(sum1, x_2, y_3)
467 SMLATB r11, r7, r8, r11 @ sum1 = MAC16_16(sum1, x_3, y_4)
476 SMLABT r11, r6, r8, r11 @ sum1 = MAC16_16(sum1, x_0, y_1)
480 SMLATB r11, r6, r9, r11 @ sum1 = MAC16_16(sum1, x_1, y_2)
487 SMLABT r11, r6, r8, r11 @ sum1 = MAC16_16(sum1, x_0, y_1)
[all …]
/external/webrtc/webrtc/common_audio/signal_processing/
Dcross_correlation_neon.c24 int64x2_t sum1 = vdupq_n_s64(0); in DotProductWithScaleNeon() local
40 sum1 = vpadalq_s32(sum1, tmp1); in DotProductWithScaleNeon()
53 sum0 = vaddq_s64(sum0, sum1); in DotProductWithScaleNeon()
Dfilter_ar_fast_q12_armv7.S56 mov r7, #0 @ sum1
66 smlatb r7, r10, r5, r7 @ sum1 += coefficients[j] * data_out[i - j];
76 smlabb r7, r10, r5, r7 @ sum1 += coefficients[2] * data_out[i - 2];
82 smlatb r7, r10, r5, r7 @ sum1 += coefficients[1] * data_out[i - 1];
86 sub r6, r7 @ output1 -= sum1;
97 sub r6, r8 @ output1 -= sum1;
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/NVPTX/
Dldg-invariant.ll36 %sum1 = fadd half %v1, %v2
38 %sum = fadd half %sum1, %sum2
52 %sum1 = fadd half %v1, %v2
54 %sum = fadd half %sum1, %sum2
76 %sum1 = add i32 %v1, %v2
78 %sum3 = add i32 %sum1, %sum2
/external/libaom/libaom/av1/common/x86/
Dselfguided_sse4.c147 static __m128i compute_p(__m128i sum1, __m128i sum2, int bit_depth, int n) { in compute_p() argument
155 const __m128i b = _mm_srl_epi32(_mm_add_epi32(sum1, rounding_b), shift_b); in compute_p()
161 bb = _mm_madd_epi16(sum1, sum1); in compute_p()
196 __m128i sum1 = boxsum_from_ii(Dij, buf_stride, r); in calc_ab() local
206 sum1 = _mm_and_si128(mask[idx], sum1); in calc_ab()
210 const __m128i p = compute_p(sum1, sum2, bit_depth, n); in calc_ab()
233 const __m128i b_int = _mm_mullo_epi32(a_comp_over_n, sum1); in calc_ab()
336 __m128i sum1 = boxsum_from_ii(Dij, buf_stride, r); in calc_ab_fast() local
346 sum1 = _mm_and_si128(mask[idx], sum1); in calc_ab_fast()
350 const __m128i p = compute_p(sum1, sum2, bit_depth, n); in calc_ab_fast()
[all …]

123456