Home
last modified time | relevance | path

Searched refs:round_0 (Results 1 – 25 of 40) sorted by relevance

12

/external/libaom/libaom/av1/common/
Dconvolve.h25 int round_0; member
74 conv_params.round_0 = ROUND0_BITS; in get_conv_params_no_round()
76 : 2 * FILTER_BITS - conv_params.round_0; in get_conv_params_no_round()
77 const int intbufrange = bd + FILTER_BITS - conv_params.round_0 + 2; in get_conv_params_no_round()
80 conv_params.round_0 += intbufrange - 16; in get_conv_params_no_round()
105 conv_params.round_0 = WIENER_ROUND0_BITS; in get_conv_params_wiener()
106 conv_params.round_1 = 2 * FILTER_BITS - conv_params.round_0; in get_conv_params_wiener()
107 const int intbufrange = bd + FILTER_BITS - conv_params.round_0 + 2; in get_conv_params_wiener()
110 conv_params.round_0 += intbufrange - 16; in get_conv_params_wiener()
Dconvolve.c129 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_convolve_2d_sr_c()
143 (int16_t)ROUND_POWER_OF_TWO(sum, conv_params->round_0); in av1_convolve_2d_sr_c()
151 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_convolve_2d_sr_c()
178 assert(conv_params->round_0 <= FILTER_BITS); in av1_convolve_y_sr_c()
179 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_convolve_y_sr_c()
180 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_convolve_y_sr_c()
204 const int bits = FILTER_BITS - conv_params->round_0; in av1_convolve_x_sr_c()
211 ((conv_params->round_0 + conv_params->round_1) == 2 * FILTER_BITS)); in av1_convolve_x_sr_c()
223 res = ROUND_POWER_OF_TWO(res, conv_params->round_0); in av1_convolve_x_sr_c()
261 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_c()
[all …]
Dwarped_motion.c331 conv_params->round_0 + in av1_highbd_warp_affine_c()
332 AOMMAX(bd + FILTER_BITS - conv_params->round_0 - 14, 0); in av1_highbd_warp_affine_c()
340 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_warp_affine_c()
341 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_highbd_warp_affine_c()
602 const int reduce_bits_horiz = conv_params->round_0; in av1_warp_affine_c()
610 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_warp_affine_c()
611 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_warp_affine_c()
/external/libaom/libaom/av1/common/arm/
Dwiener_convolve_neon.c64 assert((conv_params->round_0) >= 1); in av1_wiener_convolve_add_src_neon()
123 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
130 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
137 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
144 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
151 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
158 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
165 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
172 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
227 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon()
[all …]
Djnt_convolve_neon.c320 int16_t *x_filter_tmp, const int im_h, int w, const int round_0) { in dist_wtd_convolve_2d_horiz_neon() argument
338 const int16x4_t shift_round_0 = vdup_n_s16(-(round_0)); in dist_wtd_convolve_2d_horiz_neon()
438 const int16x8_t shift_round_0 = vdupq_n_s16(-(round_0)); in dist_wtd_convolve_2d_horiz_neon()
573 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in dist_wtd_convolve_2d_vert_neon()
578 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in dist_wtd_convolve_2d_vert_neon()
579 const int offset = bd + 2 * FILTER_BITS - conv_params->round_0; in dist_wtd_convolve_2d_vert_neon()
732 const int round_0 = conv_params->round_0 - 1; in av1_dist_wtd_convolve_2d_neon() local
748 x_filter_tmp, im_h, w, round_0); in av1_dist_wtd_convolve_2d_neon()
769 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_dist_wtd_convolve_2d_copy_neon()
771 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_dist_wtd_convolve_2d_copy_neon()
[all …]
/external/libaom/libaom/av1/common/x86/
Dwiener_convolve_avx2.c65 assert(conv_params->round_0 > 0); in av1_wiener_convolve_add_src_avx2()
91 _mm256_set1_epi16((1 << (conv_params->round_0 - 1))); in av1_wiener_convolve_add_src_avx2()
93 _mm256_set1_epi16((1 << (bd + FILTER_BITS - conv_params->round_0 - 1))); in av1_wiener_convolve_add_src_avx2()
96 _mm256_set1_epi16(WIENER_CLAMP_LIMIT(conv_params->round_0, bd) - 1); in av1_wiener_convolve_add_src_avx2()
97 const __m128i round_shift_h = _mm_cvtsi32_si128(conv_params->round_0); in av1_wiener_convolve_add_src_avx2()
142 data_0 = _mm256_slli_epi16(data_0, FILTER_BITS - conv_params->round_0); in av1_wiener_convolve_add_src_avx2()
Djnt_convolve_avx2.c53 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_avx2()
57 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_avx2()
61 assert(conv_params->round_0 > 0); in av1_dist_wtd_convolve_x_avx2()
64 _mm256_set1_epi16((1 << (conv_params->round_0 - 1)) >> 1); in av1_dist_wtd_convolve_x_avx2()
65 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0 - 1); in av1_dist_wtd_convolve_x_avx2()
201 const int left_shift = FILTER_BITS - conv_params->round_0 + 1; in av1_dist_wtd_convolve_y_avx2()
209 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_avx2()
216 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_avx2()
221 assert((FILTER_BITS - conv_params->round_0) >= 0); in av1_dist_wtd_convolve_y_avx2()
613 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_avx2()
[all …]
Dhighbd_convolve_2d_sse4.c37 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
49 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
54 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
195 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_sse4_1()
199 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_sse4_1()
206 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
229 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
230 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
297 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
Dhighbd_wiener_convolve_ssse3.c28 assert(bd + FILTER_BITS - conv_params->round_0 + 2 <= 16); in av1_highbd_wiener_convolve_add_src_ssse3()
66 (1 << (conv_params->round_0 - 1)) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_wiener_convolve_add_src_ssse3()
87 conv_params->round_0); in av1_highbd_wiener_convolve_add_src_ssse3()
102 conv_params->round_0); in av1_highbd_wiener_convolve_add_src_ssse3()
106 _mm_set1_epi16((WIENER_CLAMP_LIMIT(conv_params->round_0, bd)) - 1); in av1_highbd_wiener_convolve_add_src_ssse3()
Dhighbd_jnt_convolve_avx2.c38 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
50 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
55 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
248 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_dist_wtd_convolve_2d_avx2()
261 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_dist_wtd_convolve_2d_avx2()
262 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_dist_wtd_convolve_2d_avx2()
266 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_dist_wtd_convolve_2d_avx2()
270 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_avx2()
274 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_avx2()
492 _mm256_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_dist_wtd_convolve_x_avx2()
[all …]
Dconvolve_sse2.c94 assert(conv_params->round_0 <= FILTER_BITS); in av1_convolve_y_sr_sse2()
95 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_convolve_y_sr_sse2()
96 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_convolve_y_sr_sse2()
247 const int bits = FILTER_BITS - conv_params->round_0; in av1_convolve_x_sr_sse2()
249 _mm_set1_epi32((1 << conv_params->round_0) >> 1); in av1_convolve_x_sr_sse2()
251 const __m128i round_0_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_convolve_x_sr_sse2()
260 ((conv_params->round_0 + conv_params->round_1) == 2 * FILTER_BITS)); in av1_convolve_x_sr_sse2()
Djnt_convolve_sse2.c32 const __m128i round_const = _mm_set1_epi32((1 << conv_params->round_0) >> 1); in av1_dist_wtd_convolve_x_sse2()
33 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_dist_wtd_convolve_x_sse2()
42 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_sse2()
46 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_sse2()
164 const int bits = FILTER_BITS - conv_params->round_0; in av1_dist_wtd_convolve_y_sse2()
172 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_sse2()
176 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_sse2()
417 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_sse2()
421 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_sse2()
445 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_dist_wtd_convolve_2d_sse2()
[all …]
Dconvolve_avx2.c34 assert(conv_params->round_0 <= FILTER_BITS); in av1_convolve_y_sr_avx2()
35 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_convolve_y_sr_avx2()
36 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_convolve_y_sr_avx2()
269 const int bits = FILTER_BITS - conv_params->round_0; in av1_convolve_x_sr_avx2()
272 _mm256_set1_epi16((1 << (conv_params->round_0 - 1)) >> 1); in av1_convolve_x_sr_avx2()
273 const __m128i round_0_shift = _mm_cvtsi32_si128(conv_params->round_0 - 1); in av1_convolve_x_sr_avx2()
282 ((conv_params->round_0 + conv_params->round_1) == 2 * FILTER_BITS)); in av1_convolve_x_sr_avx2()
283 assert(conv_params->round_0 > 0); in av1_convolve_x_sr_avx2()
Dhighbd_wiener_convolve_avx2.c35 assert(bd + FILTER_BITS - conv_params->round_0 + 2 <= 16); in av1_highbd_wiener_convolve_add_src_avx2()
59 _mm256_set1_epi16(WIENER_CLAMP_LIMIT(conv_params->round_0, bd) - 1); in av1_highbd_wiener_convolve_add_src_avx2()
88 (1 << (conv_params->round_0 - 1)) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_wiener_convolve_add_src_avx2()
119 _mm256_add_epi32(res_even_sum, round_const), conv_params->round_0); in av1_highbd_wiener_convolve_add_src_avx2()
124 _mm256_add_epi32(res_odd_sum, round_const), conv_params->round_0); in av1_highbd_wiener_convolve_add_src_avx2()
Dhighbd_convolve_2d_ssse3.c37 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_convolve_2d_sr_ssse3()
41 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_convolve_2d_sr_ssse3()
42 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_convolve_2d_sr_ssse3()
46 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_convolve_2d_sr_ssse3()
50 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_highbd_convolve_2d_sr_ssse3()
Djnt_convolve_ssse3.c48 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_ssse3()
52 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_ssse3()
76 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_dist_wtd_convolve_2d_ssse3()
77 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_dist_wtd_convolve_2d_ssse3()
145 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_dist_wtd_convolve_2d_ssse3()
Dconvolve_2d_sse2.c40 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_convolve_2d_sr_sse2()
41 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_convolve_2d_sr_sse2()
43 assert(conv_params->round_0 > 0); in av1_convolve_2d_sr_sse2()
66 (1 << (bd + FILTER_BITS - 1)) + ((1 << conv_params->round_0) >> 1)); in av1_convolve_2d_sr_sse2()
67 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_convolve_2d_sr_sse2()
371 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_dist_wtd_convolve_2d_copy_sse2()
385 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_copy_sse2()
389 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_copy_sse2()
Dhighbd_jnt_convolve_sse4.c29 const int bits = FILTER_BITS - conv_params->round_0; in av1_highbd_dist_wtd_convolve_y_sse4_1()
48 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_y_sse4_1()
52 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_y_sse4_1()
287 _mm_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_dist_wtd_convolve_x_sse4_1()
288 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_dist_wtd_convolve_x_sse4_1()
292 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_x_sse4_1()
296 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_x_sse4_1()
Dwiener_convolve_sse2.c65 (1 << (conv_params->round_0 - 1)) + (1 << (bd + FILTER_BITS - 1))); in av1_wiener_convolve_add_src_sse2()
85 conv_params->round_0); in av1_wiener_convolve_add_src_sse2()
100 conv_params->round_0); in av1_wiener_convolve_add_src_sse2()
106 _mm_set1_epi16(WIENER_CLAMP_LIMIT(conv_params->round_0, bd) - 1)); in av1_wiener_convolve_add_src_sse2()
Dconvolve_2d_avx2.c34 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_convolve_2d_sr_avx2()
35 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_convolve_2d_sr_avx2()
37 assert(conv_params->round_0 > 0); in av1_convolve_2d_sr_avx2()
40 ((1 << (conv_params->round_0 - 1)) >> 1) + (1 << (bd + FILTER_BITS - 2))); in av1_convolve_2d_sr_avx2()
41 const __m128i round_shift_h = _mm_cvtsi32_si128(conv_params->round_0 - 1); in av1_convolve_2d_sr_avx2()
Dhighbd_convolve_2d_avx2.c40 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_convolve_2d_sr_avx2()
45 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_convolve_2d_sr_avx2()
46 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_convolve_2d_sr_avx2()
50 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_convolve_2d_sr_avx2()
54 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_highbd_convolve_2d_sr_avx2()
Dav1_convolve_scale_sse4.c111 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in vfilter8()
123 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in vfilter8()
251 x_step_qn, filter_params_x, conv_params->round_0); in av1_convolve_2d_scale_sse4_1()
337 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in highbd_vfilter8()
351 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in highbd_vfilter8()
358 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in highbd_vfilter8()
492 subpel_x_qn, x_step_qn, filter_params_x, conv_params->round_0, in av1_highbd_convolve_2d_scale_sse4_1()
/external/libaom/libaom/aom_dsp/x86/
Dhighbd_convolve_ssse3.c33 assert(conv_params->round_0 <= FILTER_BITS); in av1_highbd_convolve_y_sr_ssse3()
34 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_highbd_convolve_y_sr_ssse3()
35 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_highbd_convolve_y_sr_ssse3()
182 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_convolve_x_sr_ssse3()
187 _mm_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_convolve_x_sr_ssse3()
188 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_convolve_x_sr_ssse3()
190 const int bits = FILTER_BITS - conv_params->round_0; in av1_highbd_convolve_x_sr_ssse3()
/external/libaom/libaom/aom_dsp/
Dblend_a64_mask.c43 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in aom_lowbd_blend_a64_d16_mask_c()
47 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in aom_lowbd_blend_a64_d16_mask_c()
129 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in aom_highbd_blend_a64_d16_mask_c()
133 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in aom_highbd_blend_a64_d16_mask_c()
/external/libaom/libaom/test/
Dreconinter_test.cc106 bd + 2 * FILTER_BITS - conv_params.round_0 - conv_params.round_1 + 2; in RunCheckOutput()
146 bd + 2 * FILTER_BITS - conv_params.round_0 - conv_params.round_1 + 2; in RunSpeedTest()

12