Home
last modified time | relevance | path

Searched refs:offset_const (Results 1 – 12 of 12) sorted by relevance

/external/libaom/libaom/av1/common/x86/
Dhighbd_jnt_convolve_sse4.c50 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_y_sse4_1() local
111 __m128i res_unsigned_lo_0 = _mm_add_epi32(res_a_round0, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1()
112 __m128i res_unsigned_lo_1 = _mm_add_epi32(res_a_round1, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1()
132 highbd_convolve_rounding_sse2(&comp_avg_res_0, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1()
135 highbd_convolve_rounding_sse2(&comp_avg_res_1, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1()
175 __m128i res_unsigned_hi_0 = _mm_add_epi32(res_b_round0, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1()
176 __m128i res_unsigned_hi_1 = _mm_add_epi32(res_b_round1, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1()
203 highbd_convolve_rounding_sse2(&comp_avg_res_lo_0, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1()
206 highbd_convolve_rounding_sse2(&comp_avg_res_lo_1, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1()
209 highbd_convolve_rounding_sse2(&comp_avg_res_hi_0, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1()
[all …]
Dhighbd_jnt_convolve_avx2.c52 const __m256i offset_const = _mm256_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() local
79 _mm256_add_epi32(res_32b_lo, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
87 _mm256_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
94 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
96 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
138 _mm256_add_epi32(res_32b, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
145 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
183 _mm256_add_epi32(res_32b_lo, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
191 _mm256_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
198 highbd_convolve_rounding(&comp_avg_res_lo, &offset_const, in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
[all …]
Dhighbd_convolve_2d_sse4.c51 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() local
76 _mm_add_epi32(res_32b_lo, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
84 _mm_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
91 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
93 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
130 const __m128i res_unsigned_lo = _mm_add_epi32(res_32b, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
134 _mm_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
142 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
144 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
197 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_sse4_1() local
[all …]
Djnt_convolve_avx2.c55 const __m256i offset_const = _mm256_set1_epi16(offset); in av1_dist_wtd_convolve_x_avx2() local
96 const __m256i res_unsigned = _mm256_add_epi16(res, offset_const); in av1_dist_wtd_convolve_x_avx2()
106 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_avx2()
151 const __m256i res_unsigned = _mm256_add_epi16(res, offset_const); in av1_dist_wtd_convolve_x_avx2()
161 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_avx2()
211 const __m256i offset_const = _mm256_set1_epi16(offset); in av1_dist_wtd_convolve_y_avx2() local
303 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_avx2()
368 convolve_rounding(&comp_avg_res_lo, &offset_const, in av1_dist_wtd_convolve_y_avx2()
372 convolve_rounding(&comp_avg_res_hi, &offset_const, in av1_dist_wtd_convolve_y_avx2()
482 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_avx2()
[all …]
Djnt_convolve_sse2.c44 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_x_sse2() local
73 const __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_x_sse2()
83 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_sse2()
129 const __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_x_sse2()
140 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_sse2()
174 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_y_sse2() local
221 __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_y_sse2()
231 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_sse2()
250 res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_y_sse2()
260 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_sse2()
[all …]
Dconvolve_2d_sse2.c387 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_2d_copy_sse2() local
403 const __m128i res_unsigned_lo = _mm_add_epi16(res_lo, offset_const); in av1_dist_wtd_convolve_2d_copy_sse2()
406 const __m128i res_unsigned_hi = _mm_add_epi16(res_hi, offset_const); in av1_dist_wtd_convolve_2d_copy_sse2()
417 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_copy_sse2()
423 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_copy_sse2()
445 const __m128i res_unsigned = _mm_add_epi16(res, offset_const); in av1_dist_wtd_convolve_2d_copy_sse2()
454 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_copy_sse2()
Djnt_convolve_ssse3.c50 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_2d_ssse3() local
205 const __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_2d_ssse3()
216 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_ssse3()
/external/libaom/libaom/aom_dsp/x86/
Dconvolve_sse2.h102 const __m128i *const offset_const, in convolve_rounding() argument
105 const __m128i res_signed = _mm_sub_epi16(*res_unsigned, *offset_const); in convolve_rounding()
112 const __m128i *const res_unsigned, const __m128i *const offset_const, in highbd_convolve_rounding_sse2() argument
114 const __m128i res_signed = _mm_sub_epi32(*res_unsigned, *offset_const); in highbd_convolve_rounding_sse2()
Dconvolve_avx2.h204 const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \
213 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); \
232 const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \
242 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); \
426 const __m256i *const offset_const, in convolve_rounding() argument
429 const __m256i res_signed = _mm256_sub_epi16(*res_unsigned, *offset_const); in convolve_rounding()
454 const __m256i *const res_unsigned, const __m256i *const offset_const, in highbd_convolve_rounding() argument
456 const __m256i res_signed = _mm256_sub_epi32(*res_unsigned, *offset_const); in highbd_convolve_rounding()
/external/libaom/libaom/av1/common/arm/
Dconvolve_neon.c125 const int32x4_t round_shift_vec, const int32x4_t offset_const, in convolve8_vert_4x4_s32() argument
140 sum0 = vaddq_s32(sum0, offset_const); in convolve8_vert_4x4_s32()
154 const int32x4_t round_shift_vec, const int32x4_t offset_const, in convolve8_vert_8x4_s32() argument
178 sum0 = vaddq_s32(sum0, offset_const); in convolve8_vert_8x4_s32()
179 sum1 = vaddq_s32(sum1, offset_const); in convolve8_vert_8x4_s32()
1293 const int32x4_t offset_const = vdupq_n_s32(1 << offset_bits); in av1_convolve_2d_sr_neon() local
1342 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon()
1345 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon()
1348 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon()
1351 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon()
[all …]
Dconvolve_neon.h206 const int32x4_t round_shift_vec, const int32x4_t offset_const) { in convolve8_4x4_s32() argument
220 sum0 = vaddq_s32(sum0, offset_const); in convolve8_4x4_s32()
Djnt_convolve_neon.c581 const int32x4_t offset_const = vdupq_n_s32(1 << offset); in dist_wtd_convolve_2d_vert_neon() local
638 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon()
640 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon()
642 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon()
644 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon()
683 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon()