Home
last modified time | relevance | path

Searched refs:round_const (Results 1 – 25 of 29) sorted by relevance

12

/external/libaom/libaom/av1/common/x86/
Dconvolve_sse2.c86 const __m128i round_const = _mm_set1_epi32((1 << FILTER_BITS) >> 1); in av1_convolve_y_sr_sse2() local
130 res_round = _mm_sra_epi32(_mm_add_epi32(res, round_const), round_shift); in av1_convolve_y_sr_sse2()
143 res_round = _mm_sra_epi32(_mm_add_epi32(res, round_const), round_shift); in av1_convolve_y_sr_sse2()
203 _mm_sra_epi32(_mm_add_epi32(res_lo, round_const), round_shift); in av1_convolve_y_sr_sse2()
205 _mm_sra_epi32(_mm_add_epi32(res_hi, round_const), round_shift); in av1_convolve_y_sr_sse2()
217 _mm_sra_epi32(_mm_add_epi32(res_lo, round_const), round_shift); in av1_convolve_y_sr_sse2()
219 _mm_sra_epi32(_mm_add_epi32(res_hi, round_const), round_shift); in av1_convolve_y_sr_sse2()
250 const __m128i round_const = _mm_set1_epi32((1 << bits) >> 1); in av1_convolve_x_sr_sse2() local
280 _mm_sra_epi32(_mm_add_epi32(res_lo_round, round_const), round_shift); in av1_convolve_x_sr_sse2()
323 res_lo_round = _mm_sra_epi32(_mm_add_epi32(res_lo_round, round_const), in av1_convolve_x_sr_sse2()
[all …]
Djnt_convolve_sse2.c32 const __m128i round_const = _mm_set1_epi32((1 << conv_params->round_0) >> 1); in av1_dist_wtd_convolve_x_sse2() local
69 _mm_sra_epi32(_mm_add_epi32(res_lo, round_const), round_shift); in av1_dist_wtd_convolve_x_sse2()
122 _mm_sra_epi32(_mm_add_epi32(res_lo, round_const), round_shift); in av1_dist_wtd_convolve_x_sse2()
124 _mm_sra_epi32(_mm_add_epi32(res_hi, round_const), round_shift); in av1_dist_wtd_convolve_x_sse2()
178 const __m128i round_const = _mm_set1_epi32((1 << conv_params->round_1) >> 1); in av1_dist_wtd_convolve_y_sse2() local
218 _mm_sra_epi32(_mm_add_epi32(res_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_sse2()
247 _mm_sra_epi32(_mm_add_epi32(res_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_sse2()
320 res_lo_shift = _mm_sra_epi32(_mm_add_epi32(res_lo_shift, round_const), in av1_dist_wtd_convolve_y_sse2()
322 res_hi_shift = _mm_sra_epi32(_mm_add_epi32(res_hi_shift, round_const), in av1_dist_wtd_convolve_y_sse2()
350 res_lo_shift = _mm_sra_epi32(_mm_add_epi32(res_lo_shift, round_const), in av1_dist_wtd_convolve_y_sse2()
[all …]
Dhighbd_wiener_convolve_ssse3.c65 const __m128i round_const = _mm_set1_epi32( in av1_highbd_wiener_convolve_add_src_ssse3() local
86 res_even = _mm_srai_epi32(_mm_add_epi32(res_even, round_const), in av1_highbd_wiener_convolve_add_src_ssse3()
101 res_odd = _mm_srai_epi32(_mm_add_epi32(res_odd, round_const), in av1_highbd_wiener_convolve_add_src_ssse3()
133 const __m128i round_const = in av1_highbd_wiener_convolve_add_src_ssse3() local
189 _mm_add_epi32(res_lo, round_const), conv_params->round_1); in av1_highbd_wiener_convolve_add_src_ssse3()
191 _mm_add_epi32(res_hi, round_const), conv_params->round_1); in av1_highbd_wiener_convolve_add_src_ssse3()
Dwiener_convolve_sse2.c64 const __m128i round_const = _mm_set1_epi32( in av1_wiener_convolve_add_src_sse2() local
84 res_even = _mm_srai_epi32(_mm_add_epi32(res_even, round_const), in av1_wiener_convolve_add_src_sse2()
99 res_odd = _mm_srai_epi32(_mm_add_epi32(res_odd, round_const), in av1_wiener_convolve_add_src_sse2()
131 const __m128i round_const = in av1_wiener_convolve_add_src_sse2() local
187 _mm_add_epi32(res_lo, round_const), conv_params->round_1); in av1_wiener_convolve_add_src_sse2()
189 _mm_add_epi32(res_hi, round_const), conv_params->round_1); in av1_wiener_convolve_add_src_sse2()
Dhighbd_wiener_convolve_avx2.c87 const __m256i round_const = _mm256_set1_epi32( in av1_highbd_wiener_convolve_add_src_avx2() local
119 _mm256_add_epi32(res_even_sum, round_const), conv_params->round_0); in av1_highbd_wiener_convolve_add_src_avx2()
124 _mm256_add_epi32(res_odd_sum, round_const), conv_params->round_0); in av1_highbd_wiener_convolve_add_src_avx2()
170 const __m256i round_const = in av1_highbd_wiener_convolve_add_src_avx2() local
229 _mm256_add_epi32(res_lo, round_const), conv_params->round_1); in av1_highbd_wiener_convolve_add_src_avx2()
231 _mm256_add_epi32(res_hi, round_const), conv_params->round_1); in av1_highbd_wiener_convolve_add_src_avx2()
Djnt_convolve_ssse3.c75 const __m128i round_const = _mm_set1_epi32( in av1_dist_wtd_convolve_2d_ssse3() local
99 _mm_sra_epi32(_mm_add_epi32(res_even, round_const), round_shift); in av1_dist_wtd_convolve_2d_ssse3()
114 _mm_sra_epi32(_mm_add_epi32(res_odd, round_const), round_shift); in av1_dist_wtd_convolve_2d_ssse3()
143 const __m128i round_const = _mm_set1_epi32( in av1_dist_wtd_convolve_2d_ssse3() local
200 _mm_sra_epi32(_mm_add_epi32(res_lo, round_const), round_shift); in av1_dist_wtd_convolve_2d_ssse3()
202 _mm_sra_epi32(_mm_add_epi32(res_hi, round_const), round_shift); in av1_dist_wtd_convolve_2d_ssse3()
Dwarp_plane_avx2.c80 const __m256i *round_const, in filter_src_pixels_avx2() argument
95 _mm256_add_epi16(_mm256_add_epi16(res_even, res_odd), *round_const); in filter_src_pixels_avx2()
257 const __m256i *round_const, in horizontal_filter_avx2() argument
261 filter_src_pixels_avx2(src, horz_out, coeff, shuffle_src, round_const, shift, in horizontal_filter_avx2()
302 const __m256i *round_const, const __m128i *shift, in warp_horizontal_filter_avx2() argument
319 round_const, shift); in warp_horizontal_filter_avx2()
328 filter_src_pixels_avx2(src_01, horz_out, coeff, shuffle_src, round_const, in warp_horizontal_filter_avx2()
335 const __m256i *round_const, const __m128i *shift, in warp_horizontal_filter_alpha0_avx2() argument
353 filter_src_pixels_avx2(src_01, horz_out, coeff, shuffle_src, round_const, in warp_horizontal_filter_alpha0_avx2()
363 filter_src_pixels_avx2(src_01, horz_out, coeff, shuffle_src, round_const, in warp_horizontal_filter_alpha0_avx2()
[all …]
Djnt_convolve_avx2.c63 const __m256i round_const = in av1_dist_wtd_convolve_x_avx2() local
93 res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const), round_shift); in av1_dist_wtd_convolve_x_avx2()
147 res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const), round_shift); in av1_dist_wtd_convolve_x_avx2()
202 const __m256i round_const = in av1_dist_wtd_convolve_y_avx2() local
280 _mm256_add_epi32(res_lo_0_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_avx2()
286 _mm256_add_epi32(res_lo_1_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_avx2()
338 _mm256_add_epi32(res_hi_0_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_avx2()
344 _mm256_add_epi32(res_hi_1_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_avx2()
459 _mm256_add_epi32(res_lo_0_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_avx2()
465 _mm256_add_epi32(res_lo_1_shift, round_const), round_shift); in av1_dist_wtd_convolve_y_avx2()
[all …]
Dhighbd_convolve_2d_sse4.c228 const __m128i round_const = _mm_set1_epi32( in av1_highbd_dist_wtd_convolve_2d_sse4_1() local
251 _mm_sra_epi32(_mm_add_epi32(res_even, round_const), round_shift); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
266 _mm_sra_epi32(_mm_add_epi32(res_odd, round_const), round_shift); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
295 const __m128i round_const = _mm_set1_epi32( in av1_highbd_dist_wtd_convolve_2d_sse4_1() local
352 _mm_sra_epi32(_mm_add_epi32(res_lo, round_const), round_shift); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
383 _mm_sra_epi32(_mm_add_epi32(res_hi, round_const), round_shift); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
Dconvolve_2d_sse2.c65 const __m128i round_const = _mm_set1_epi32( in av1_convolve_2d_sr_sse2() local
87 _mm_sra_epi32(_mm_add_epi32(res_even, round_const), round_shift); in av1_convolve_2d_sr_sse2()
102 _mm_sra_epi32(_mm_add_epi32(res_odd, round_const), round_shift); in av1_convolve_2d_sr_sse2()
135 const __m128i round_const = _mm_set1_epi32( in av1_convolve_2d_sr_sse2() local
196 res_lo_round = _mm_sra_epi32(_mm_add_epi32(res_lo_round, round_const), in av1_convolve_2d_sr_sse2()
198 res_hi_round = _mm_sra_epi32(_mm_add_epi32(res_hi_round, round_const), in av1_convolve_2d_sr_sse2()
Dconvolve_avx2.c274 const __m256i round_const = _mm256_set1_epi16((1 << bits) >> 1); in av1_convolve_x_sr_avx2() local
313 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2()
351 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2()
386 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2()
423 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2()
Dreconinter_sse4.c105 const __m128i round_const = _mm_set1_epi16((1 << round) >> 1); in av1_build_compound_diffwtd_mask_d16_sse4_1() local
130 _mm_srli_epi16(_mm_adds_epu16(diff, round_const), round); in av1_build_compound_diffwtd_mask_d16_sse4_1()
Dhighbd_warp_plane_sse4.c121 const __m128i round_const = _mm_set1_epi32((1 << offset_bits_horiz) + in highbd_filter_src_pixels() local
134 res_even = _mm_sra_epi32(_mm_add_epi32(res_even, round_const), in highbd_filter_src_pixels()
148 res_odd = _mm_sra_epi32(_mm_add_epi32(res_odd, round_const), in highbd_filter_src_pixels()
602 const __m128i round_const = in av1_highbd_warp_affine_sse4_1() local
607 _mm_add_epi32(res_lo, round_const), reduce_bits_vert); in av1_highbd_warp_affine_sse4_1()
609 _mm_add_epi32(res_hi, round_const), reduce_bits_vert); in av1_highbd_warp_affine_sse4_1()
Dreconinter_avx2.c140 const __m256i *round_const, in calc_mask_d16_avx2() argument
147 _mm256_srli_epi16(_mm256_adds_epu16(diff, *round_const), round); in calc_mask_d16_avx2()
156 const __m256i *round_const, in calc_mask_d16_inv_avx2() argument
164 _mm256_srli_epi16(_mm256_adds_epu16(diff, *round_const), round); in calc_mask_d16_inv_avx2()
Dwarp_plane_sse4.c265 const __m128i round_const = _mm_set1_epi16((1 << offset_bits_horiz) + in filter_src_pixels() local
279 _mm_add_epi16(_mm_add_epi16(res_even, res_odd), round_const); in filter_src_pixels()
/external/libaom/libaom/aom_dsp/x86/
Dconvolve_sse2.h103 const __m128i *const round_const, in convolve_rounding() argument
107 _mm_srai_epi16(_mm_add_epi16(res_signed, *round_const), round_shift); in convolve_rounding()
113 const __m128i *const round_const, const int round_shift) { in highbd_convolve_rounding_sse2() argument
116 _mm_srai_epi32(_mm_add_epi32(res_signed, *round_const), round_shift); in highbd_convolve_rounding_sse2()
Dmasked_sad_intrin_ssse3.c296 const __m128i round_const = in highbd_masked_sad_ssse3() local
313 pred_l = _mm_srai_epi32(_mm_add_epi32(pred_l, round_const), in highbd_masked_sad_ssse3()
319 pred_r = _mm_srai_epi32(_mm_add_epi32(pred_r, round_const), in highbd_masked_sad_ssse3()
355 const __m128i round_const = in aom_highbd_masked_sad4xh_ssse3() local
380 pred_l = _mm_srai_epi32(_mm_add_epi32(pred_l, round_const), in aom_highbd_masked_sad4xh_ssse3()
386 pred_r = _mm_srai_epi32(_mm_add_epi32(pred_r, round_const), in aom_highbd_masked_sad4xh_ssse3()
Dmasked_sad_intrin_avx2.c212 const __m256i round_const = in highbd_masked_sad8xh_avx2() local
229 pred_l = _mm256_srai_epi32(_mm256_add_epi32(pred_l, round_const), in highbd_masked_sad8xh_avx2()
235 pred_r = _mm256_srai_epi32(_mm256_add_epi32(pred_r, round_const), in highbd_masked_sad8xh_avx2()
269 const __m256i round_const = in highbd_masked_sad16xh_avx2() local
286 pred_l = _mm256_srai_epi32(_mm256_add_epi32(pred_l, round_const), in highbd_masked_sad16xh_avx2()
292 pred_r = _mm256_srai_epi32(_mm256_add_epi32(pred_r, round_const), in highbd_masked_sad16xh_avx2()
Dconvolve_avx2.h427 const __m256i *const round_const, in convolve_rounding() argument
431 _mm256_add_epi16(res_signed, *round_const), round_shift); in convolve_rounding()
455 const __m256i *const round_const, const int round_shift) { in highbd_convolve_rounding() argument
458 _mm256_add_epi32(res_signed, *round_const), round_shift); in highbd_convolve_rounding()
Dvariance_avx2.c424 const __m256i round_const = in highbd_comp_mask_pred_line_avx2() local
432 _mm256_add_epi32(pred_lo, round_const), AOM_BLEND_A64_ROUND_BITS); in highbd_comp_mask_pred_line_avx2()
438 _mm256_add_epi32(pred_hi, round_const), AOM_BLEND_A64_ROUND_BITS); in highbd_comp_mask_pred_line_avx2()
Dmasked_variance_intrin_ssse3.c906 const __m128i round_const = in highbd_masked_variance() local
923 pred_l = _mm_srai_epi32(_mm_add_epi32(pred_l, round_const), in highbd_masked_variance()
929 pred_r = _mm_srai_epi32(_mm_add_epi32(pred_r, round_const), in highbd_masked_variance()
978 const __m128i round_const = in highbd_masked_variance4xh() local
998 pred_l = _mm_srai_epi32(_mm_add_epi32(pred_l, round_const), in highbd_masked_variance4xh()
1004 pred_r = _mm_srai_epi32(_mm_add_epi32(pred_r, round_const), in highbd_masked_variance4xh()
Dvariance_sse2.c655 const __m128i round_const = in highbd_comp_mask_pred_line_sse2() local
662 const __m128i pred_l = _mm_srai_epi32(_mm_add_epi32(pred_lo, round_const), in highbd_comp_mask_pred_line_sse2()
668 const __m128i pred_h = _mm_srai_epi32(_mm_add_epi32(pred_hi, round_const), in highbd_comp_mask_pred_line_sse2()
/external/libaom/libaom/av1/common/arm/
Dconvolve_neon.h29 const int32_t round_const = (1 << (bd + round1_bits - 1)); in wiener_convolve8_vert_4x8() local
32 const int32x4_t round_vec = vdupq_n_s32(round_const); in wiener_convolve8_vert_4x8()
/external/vixl/src/aarch64/
Dlogic-aarch64.cc3952 int round_const = round ? (1 << (esize - 2)) : 0; in sqrdmulh() local
3958 product += round_const; in sqrdmulh()
4030 int round_const = round ? (1 << (esize - 2)) : 0; in sqrdmlash() local
4041 accum += round_const; in sqrdmlash()
/external/arm-neon-tests/
Dref-rvct-neon-nofp16.txt2113 VRSHL/VRSHLQ (checking round_const overflow: shift by -1) output:
2137 VRSHL/VRSHLQ (checking round_const overflow: shift by -3) output:

12