/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_error_avx2.c | 82 sse_256 = _mm256_add_epi64(sse_256, exp_dqcoeff_lo); in vp9_block_error_avx2() 83 ssz_256 = _mm256_add_epi64(ssz_256, exp_coeff_lo); in vp9_block_error_avx2() 84 sse_256 = _mm256_add_epi64(sse_256, exp_dqcoeff_hi); in vp9_block_error_avx2() 85 ssz_256 = _mm256_add_epi64(ssz_256, exp_coeff_hi); in vp9_block_error_avx2() 92 sse_256 = _mm256_add_epi64(sse_256, sse_hi); in vp9_block_error_avx2() 93 ssz_256 = _mm256_add_epi64(ssz_256, ssz_hi); in vp9_block_error_avx2() 143 sse_256 = _mm256_add_epi64(sse_256, exp_error_lo); in vp9_block_error_fp_avx2() 144 sse_256 = _mm256_add_epi64(sse_256, exp_error_hi); in vp9_block_error_fp_avx2() 152 sse_256 = _mm256_add_epi64(sse_256, sse_hi); in vp9_block_error_fp_avx2()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | highbd_block_error_intrin_avx2.c | 40 __m256i res_diff = _mm256_add_epi64(_mm256_add_epi64(res, res1), in av1_highbd_block_error_avx2() 41 _mm256_add_epi64(res2, res3)); in av1_highbd_block_error_avx2() 48 __m256i res_sqcoeff = _mm256_add_epi64(_mm256_add_epi64(res, res1), in av1_highbd_block_error_avx2() 49 _mm256_add_epi64(res2, res3)); in av1_highbd_block_error_avx2()
|
D | error_intrin_avx2.c | 63 sse_reg = _mm256_add_epi64(sse_reg, exp_dqcoeff_lo); in av1_block_error_avx2() 64 ssz_reg = _mm256_add_epi64(ssz_reg, exp_coeff_lo); in av1_block_error_avx2() 65 sse_reg = _mm256_add_epi64(sse_reg, exp_dqcoeff_hi); in av1_block_error_avx2() 66 ssz_reg = _mm256_add_epi64(ssz_reg, exp_coeff_hi); in av1_block_error_avx2() 72 sse_reg = _mm256_add_epi64(sse_reg, sse_reg_64hi); in av1_block_error_avx2() 73 ssz_reg = _mm256_add_epi64(ssz_reg, ssz_reg_64hi); in av1_block_error_avx2()
|
D | wedge_utils_avx2.c | 63 const __m256i v_sum0_q = _mm256_add_epi64( in av1_wedge_sse_from_residuals_avx2() 66 v_acc0_q = _mm256_add_epi64(v_acc0_q, v_sum0_q); in av1_wedge_sse_from_residuals_avx2() 71 v_acc0_q = _mm256_add_epi64(v_acc0_q, _mm256_srli_si256(v_acc0_q, 8)); in av1_wedge_sse_from_residuals_avx2() 135 v_acc0_d = _mm256_add_epi64(_mm256_unpacklo_epi32(v_acc0_d, v_sign_d), in av1_wedge_sign_from_residuals_avx2() 138 __m256i v_acc_q = _mm256_add_epi64(v_acc0_d, _mm256_srli_si256(v_acc0_d, 8)); in av1_wedge_sign_from_residuals_avx2()
|
D | pickrst_avx2.c | 149 yy_storeu_256(dst, _mm256_add_epi64(d0l, dst0)); in acc_stat_highbd_avx2() 155 yy_storeu_256(dst + 4, _mm256_add_epi64(d0h, dst1)); in acc_stat_highbd_avx2() 549 sum64 = _mm256_add_epi64(sum64, sum64_0); in av1_lowbd_pixel_proj_error_avx2() 550 sum64 = _mm256_add_epi64(sum64, sum64_1); in av1_lowbd_pixel_proj_error_avx2() 593 sum64 = _mm256_add_epi64(sum64, sum64_0); in av1_lowbd_pixel_proj_error_avx2() 594 sum64 = _mm256_add_epi64(sum64, sum64_1); in av1_lowbd_pixel_proj_error_avx2() 617 sum64 = _mm256_add_epi64(sum64_0, sum64_1); in av1_lowbd_pixel_proj_error_avx2() 704 sum64 = _mm256_add_epi64(sum64, sum32l); in av1_highbd_pixel_proj_error_avx2() 707 sum64 = _mm256_add_epi64(sum64, sum32h); in av1_highbd_pixel_proj_error_avx2() 781 sum64 = _mm256_add_epi64(sum64, sum32l); in av1_highbd_pixel_proj_error_avx2() [all …]
|
/external/flac/libFLAC/ |
D | lpc_intrin_avx2.c | 795 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-11)))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 796 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-10)))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 797 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-9 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 798 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-8 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 799 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-7 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 800 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-6 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 801 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-5 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 802 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-4 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 803 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-3 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 804 …cvtepu32_epi64(_mm_loadu_si128((const __m128i*)(data+i-2 )))); summ = _mm256_add_epi64(summ, mull); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() [all …]
|
D | stream_encoder_intrin_avx2.c | 104 sum256 = _mm256_add_epi64(sum256, res256); in FLAC__precompute_partition_info_sums_intrin_avx2()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | sum_squares_avx2.c | 50 _mm256_add_epi64(v_acc_q, _mm256_and_si256(v_acc_d, v_zext_mask_q)); in aom_sum_squares_2d_i16_nxn_avx2() 51 v_acc_q = _mm256_add_epi64(v_acc_q, _mm256_srli_epi64(v_acc_d, 32)); in aom_sum_squares_2d_i16_nxn_avx2()
|
D | sse_avx2.c | 40 const __m256i sum_4x64 = _mm256_add_epi64(sum0_4x64, sum1_4x64); in summary_all_avx2() 53 const __m256i sum_4x64 = _mm256_add_epi64(sum0_4x64, sum1_4x64); in summary_32_avx2() 54 *sum = _mm256_add_epi64(*sum, sum_4x64); in summary_32_avx2()
|
D | intrapred_avx2.c | 23 y0 = _mm256_add_epi64(y0, y1); in dc_sum_64() 25 y0 = _mm256_add_epi64(u0, y0); in dc_sum_64() 35 y = _mm256_add_epi64(u, y); in dc_sum_32()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v256_intrinsics_x86.h | 103 SIMD_INLINE v256 v256_add_64(v256 a, v256 b) { return _mm256_add_epi64(a, b); } in v256_add_64() 393 r = _mm256_add_epi64(_mm256_cvtepi32_epi64(v256_high_v128(r)), in v256_dotp_s16() 395 t = v256_low_v128(_mm256_add_epi64( in v256_dotp_s16() 416 r = _mm256_add_epi64(_mm256_cvtepi32_epi64(v256_high_v128(r)), in v256_dotp_s32() 418 t = v256_low_v128(_mm256_add_epi64( in v256_dotp_s32() 453 return _mm256_add_epi64(s, _mm256_sad_epu8(a, b)); in v256_sad_u8() 481 return _mm256_add_epi64( in v256_ssd_u8()
|
/external/clang/test/CodeGen/ |
D | avx2-builtins.c | 50 return _mm256_add_epi64(a, b); in test_mm256_add_epi64()
|
/external/clang/lib/Headers/ |
D | avx2intrin.h | 100 _mm256_add_epi64(__m256i __a, __m256i __b) in _mm256_add_epi64() function
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_dct32x32_impl_avx2.h | 33 return _mm256_add_epi64(buf0, buf1); in k_madd_epi32_avx2()
|