Home
last modified time | relevance | path

Searched refs:_mm256_add_epi64 (Results 1 – 14 of 14) sorted by relevance

/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_error_avx2.c82 sse_256 = _mm256_add_epi64(sse_256, exp_dqcoeff_lo); in vp9_block_error_avx2()
83 ssz_256 = _mm256_add_epi64(ssz_256, exp_coeff_lo); in vp9_block_error_avx2()
84 sse_256 = _mm256_add_epi64(sse_256, exp_dqcoeff_hi); in vp9_block_error_avx2()
85 ssz_256 = _mm256_add_epi64(ssz_256, exp_coeff_hi); in vp9_block_error_avx2()
92 sse_256 = _mm256_add_epi64(sse_256, sse_hi); in vp9_block_error_avx2()
93 ssz_256 = _mm256_add_epi64(ssz_256, ssz_hi); in vp9_block_error_avx2()
143 sse_256 = _mm256_add_epi64(sse_256, exp_error_lo); in vp9_block_error_fp_avx2()
144 sse_256 = _mm256_add_epi64(sse_256, exp_error_hi); in vp9_block_error_fp_avx2()
152 sse_256 = _mm256_add_epi64(sse_256, sse_hi); in vp9_block_error_fp_avx2()
/external/libaom/libaom/av1/encoder/x86/
Dhighbd_block_error_intrin_avx2.c40 __m256i res_diff = _mm256_add_epi64(_mm256_add_epi64(res, res1), in av1_highbd_block_error_avx2()
41 _mm256_add_epi64(res2, res3)); in av1_highbd_block_error_avx2()
48 __m256i res_sqcoeff = _mm256_add_epi64(_mm256_add_epi64(res, res1), in av1_highbd_block_error_avx2()
49 _mm256_add_epi64(res2, res3)); in av1_highbd_block_error_avx2()
Derror_intrin_avx2.c63 sse_reg = _mm256_add_epi64(sse_reg, exp_dqcoeff_lo); in av1_block_error_avx2()
64 ssz_reg = _mm256_add_epi64(ssz_reg, exp_coeff_lo); in av1_block_error_avx2()
65 sse_reg = _mm256_add_epi64(sse_reg, exp_dqcoeff_hi); in av1_block_error_avx2()
66 ssz_reg = _mm256_add_epi64(ssz_reg, exp_coeff_hi); in av1_block_error_avx2()
72 sse_reg = _mm256_add_epi64(sse_reg, sse_reg_64hi); in av1_block_error_avx2()
73 ssz_reg = _mm256_add_epi64(ssz_reg, ssz_reg_64hi); in av1_block_error_avx2()
Dwedge_utils_avx2.c63 const __m256i v_sum0_q = _mm256_add_epi64( in av1_wedge_sse_from_residuals_avx2()
66 v_acc0_q = _mm256_add_epi64(v_acc0_q, v_sum0_q); in av1_wedge_sse_from_residuals_avx2()
71 v_acc0_q = _mm256_add_epi64(v_acc0_q, _mm256_srli_si256(v_acc0_q, 8)); in av1_wedge_sse_from_residuals_avx2()
135 v_acc0_d = _mm256_add_epi64(_mm256_unpacklo_epi32(v_acc0_d, v_sign_d), in av1_wedge_sign_from_residuals_avx2()
138 __m256i v_acc_q = _mm256_add_epi64(v_acc0_d, _mm256_srli_si256(v_acc0_d, 8)); in av1_wedge_sign_from_residuals_avx2()
Dpickrst_avx2.c149 yy_storeu_256(dst, _mm256_add_epi64(d0l, dst0)); in acc_stat_highbd_avx2()
155 yy_storeu_256(dst + 4, _mm256_add_epi64(d0h, dst1)); in acc_stat_highbd_avx2()
549 sum64 = _mm256_add_epi64(sum64, sum64_0); in av1_lowbd_pixel_proj_error_avx2()
550 sum64 = _mm256_add_epi64(sum64, sum64_1); in av1_lowbd_pixel_proj_error_avx2()
593 sum64 = _mm256_add_epi64(sum64, sum64_0); in av1_lowbd_pixel_proj_error_avx2()
594 sum64 = _mm256_add_epi64(sum64, sum64_1); in av1_lowbd_pixel_proj_error_avx2()
617 sum64 = _mm256_add_epi64(sum64_0, sum64_1); in av1_lowbd_pixel_proj_error_avx2()
704 sum64 = _mm256_add_epi64(sum64, sum32l); in av1_highbd_pixel_proj_error_avx2()
707 sum64 = _mm256_add_epi64(sum64, sum32h); in av1_highbd_pixel_proj_error_avx2()
781 sum64 = _mm256_add_epi64(sum64, sum32l); in av1_highbd_pixel_proj_error_avx2()
[all …]
/external/flac/libFLAC/
Dlpc_intrin_avx2.c795 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-11)))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
796 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-10)))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
797 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-9 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
798 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-8 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
799 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-7 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
800 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-6 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
801 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-5 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
802 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-4 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
803 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-3 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
804 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-2 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
[all …]
Dstream_encoder_intrin_avx2.c104 sum256 = _mm256_add_epi64(sum256, res256); in FLAC__precompute_partition_info_sums_intrin_avx2()
/external/libaom/libaom/aom_dsp/x86/
Dsum_squares_avx2.c50 _mm256_add_epi64(v_acc_q, _mm256_and_si256(v_acc_d, v_zext_mask_q)); in aom_sum_squares_2d_i16_nxn_avx2()
51 v_acc_q = _mm256_add_epi64(v_acc_q, _mm256_srli_epi64(v_acc_d, 32)); in aom_sum_squares_2d_i16_nxn_avx2()
Dsse_avx2.c40 const __m256i sum_4x64 = _mm256_add_epi64(sum0_4x64, sum1_4x64); in summary_all_avx2()
53 const __m256i sum_4x64 = _mm256_add_epi64(sum0_4x64, sum1_4x64); in summary_32_avx2()
54 *sum = _mm256_add_epi64(*sum, sum_4x64); in summary_32_avx2()
Dintrapred_avx2.c23 y0 = _mm256_add_epi64(y0, y1); in dc_sum_64()
25 y0 = _mm256_add_epi64(u0, y0); in dc_sum_64()
35 y = _mm256_add_epi64(u, y); in dc_sum_32()
/external/libaom/libaom/aom_dsp/simd/
Dv256_intrinsics_x86.h103 SIMD_INLINE v256 v256_add_64(v256 a, v256 b) { return _mm256_add_epi64(a, b); } in v256_add_64()
393 r = _mm256_add_epi64(_mm256_cvtepi32_epi64(v256_high_v128(r)), in v256_dotp_s16()
395 t = v256_low_v128(_mm256_add_epi64( in v256_dotp_s16()
416 r = _mm256_add_epi64(_mm256_cvtepi32_epi64(v256_high_v128(r)), in v256_dotp_s32()
418 t = v256_low_v128(_mm256_add_epi64( in v256_dotp_s32()
453 return _mm256_add_epi64(s, _mm256_sad_epu8(a, b)); in v256_sad_u8()
481 return _mm256_add_epi64( in v256_ssd_u8()
/external/clang/test/CodeGen/
Davx2-builtins.c50 return _mm256_add_epi64(a, b); in test_mm256_add_epi64()
/external/clang/lib/Headers/
Davx2intrin.h100 _mm256_add_epi64(__m256i __a, __m256i __b) in _mm256_add_epi64() function
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_dct32x32_impl_avx2.h33 return _mm256_add_epi64(buf0, buf1); in k_madd_epi32_avx2()