/external/libaom/libaom/av1/common/ |
D | convolve.h | 25 int round_0; member 74 conv_params.round_0 = ROUND0_BITS; in get_conv_params_no_round() 76 : 2 * FILTER_BITS - conv_params.round_0; in get_conv_params_no_round() 77 const int intbufrange = bd + FILTER_BITS - conv_params.round_0 + 2; in get_conv_params_no_round() 80 conv_params.round_0 += intbufrange - 16; in get_conv_params_no_round() 105 conv_params.round_0 = WIENER_ROUND0_BITS; in get_conv_params_wiener() 106 conv_params.round_1 = 2 * FILTER_BITS - conv_params.round_0; in get_conv_params_wiener() 107 const int intbufrange = bd + FILTER_BITS - conv_params.round_0 + 2; in get_conv_params_wiener() 110 conv_params.round_0 += intbufrange - 16; in get_conv_params_wiener()
|
D | convolve.c | 129 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_convolve_2d_sr_c() 143 (int16_t)ROUND_POWER_OF_TWO(sum, conv_params->round_0); in av1_convolve_2d_sr_c() 151 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_convolve_2d_sr_c() 178 assert(conv_params->round_0 <= FILTER_BITS); in av1_convolve_y_sr_c() 179 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_convolve_y_sr_c() 180 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_convolve_y_sr_c() 204 const int bits = FILTER_BITS - conv_params->round_0; in av1_convolve_x_sr_c() 211 ((conv_params->round_0 + conv_params->round_1) == 2 * FILTER_BITS)); in av1_convolve_x_sr_c() 223 res = ROUND_POWER_OF_TWO(res, conv_params->round_0); in av1_convolve_x_sr_c() 261 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_c() [all …]
|
D | warped_motion.c | 331 conv_params->round_0 + in av1_highbd_warp_affine_c() 332 AOMMAX(bd + FILTER_BITS - conv_params->round_0 - 14, 0); in av1_highbd_warp_affine_c() 340 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_warp_affine_c() 341 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_highbd_warp_affine_c() 602 const int reduce_bits_horiz = conv_params->round_0; in av1_warp_affine_c() 610 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_warp_affine_c() 611 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_warp_affine_c()
|
/external/libaom/libaom/av1/common/arm/ |
D | wiener_convolve_neon.c | 64 assert((conv_params->round_0) >= 1); in av1_wiener_convolve_add_src_neon() 123 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 130 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 137 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 144 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 151 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 158 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 165 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 172 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() 227 bd, conv_params->round_0); in av1_wiener_convolve_add_src_neon() [all …]
|
D | jnt_convolve_neon.c | 320 int16_t *x_filter_tmp, const int im_h, int w, const int round_0) { in dist_wtd_convolve_2d_horiz_neon() argument 338 const int16x4_t shift_round_0 = vdup_n_s16(-(round_0)); in dist_wtd_convolve_2d_horiz_neon() 438 const int16x8_t shift_round_0 = vdupq_n_s16(-(round_0)); in dist_wtd_convolve_2d_horiz_neon() 573 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in dist_wtd_convolve_2d_vert_neon() 578 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in dist_wtd_convolve_2d_vert_neon() 579 const int offset = bd + 2 * FILTER_BITS - conv_params->round_0; in dist_wtd_convolve_2d_vert_neon() 732 const int round_0 = conv_params->round_0 - 1; in av1_dist_wtd_convolve_2d_neon() local 748 x_filter_tmp, im_h, w, round_0); in av1_dist_wtd_convolve_2d_neon() 769 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_dist_wtd_convolve_2d_copy_neon() 771 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_dist_wtd_convolve_2d_copy_neon() [all …]
|
/external/libaom/libaom/av1/common/x86/ |
D | wiener_convolve_avx2.c | 65 assert(conv_params->round_0 > 0); in av1_wiener_convolve_add_src_avx2() 91 _mm256_set1_epi16((1 << (conv_params->round_0 - 1))); in av1_wiener_convolve_add_src_avx2() 93 _mm256_set1_epi16((1 << (bd + FILTER_BITS - conv_params->round_0 - 1))); in av1_wiener_convolve_add_src_avx2() 96 _mm256_set1_epi16(WIENER_CLAMP_LIMIT(conv_params->round_0, bd) - 1); in av1_wiener_convolve_add_src_avx2() 97 const __m128i round_shift_h = _mm_cvtsi32_si128(conv_params->round_0); in av1_wiener_convolve_add_src_avx2() 142 data_0 = _mm256_slli_epi16(data_0, FILTER_BITS - conv_params->round_0); in av1_wiener_convolve_add_src_avx2()
|
D | jnt_convolve_avx2.c | 53 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_avx2() 57 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_avx2() 61 assert(conv_params->round_0 > 0); in av1_dist_wtd_convolve_x_avx2() 64 _mm256_set1_epi16((1 << (conv_params->round_0 - 1)) >> 1); in av1_dist_wtd_convolve_x_avx2() 65 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0 - 1); in av1_dist_wtd_convolve_x_avx2() 201 const int left_shift = FILTER_BITS - conv_params->round_0 + 1; in av1_dist_wtd_convolve_y_avx2() 209 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_avx2() 216 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_avx2() 221 assert((FILTER_BITS - conv_params->round_0) >= 0); in av1_dist_wtd_convolve_y_avx2() 613 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_avx2() [all …]
|
D | highbd_convolve_2d_sse4.c | 37 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 49 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 54 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 195 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_sse4_1() 199 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_sse4_1() 206 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 229 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 230 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 297 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
|
D | highbd_wiener_convolve_ssse3.c | 28 assert(bd + FILTER_BITS - conv_params->round_0 + 2 <= 16); in av1_highbd_wiener_convolve_add_src_ssse3() 66 (1 << (conv_params->round_0 - 1)) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_wiener_convolve_add_src_ssse3() 87 conv_params->round_0); in av1_highbd_wiener_convolve_add_src_ssse3() 102 conv_params->round_0); in av1_highbd_wiener_convolve_add_src_ssse3() 106 _mm_set1_epi16((WIENER_CLAMP_LIMIT(conv_params->round_0, bd)) - 1); in av1_highbd_wiener_convolve_add_src_ssse3()
|
D | highbd_jnt_convolve_avx2.c | 38 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 50 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 55 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 248 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_dist_wtd_convolve_2d_avx2() 261 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_dist_wtd_convolve_2d_avx2() 262 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_dist_wtd_convolve_2d_avx2() 266 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_dist_wtd_convolve_2d_avx2() 270 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_avx2() 274 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_2d_avx2() 492 _mm256_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_dist_wtd_convolve_x_avx2() [all …]
|
D | convolve_sse2.c | 94 assert(conv_params->round_0 <= FILTER_BITS); in av1_convolve_y_sr_sse2() 95 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_convolve_y_sr_sse2() 96 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_convolve_y_sr_sse2() 247 const int bits = FILTER_BITS - conv_params->round_0; in av1_convolve_x_sr_sse2() 249 _mm_set1_epi32((1 << conv_params->round_0) >> 1); in av1_convolve_x_sr_sse2() 251 const __m128i round_0_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_convolve_x_sr_sse2() 260 ((conv_params->round_0 + conv_params->round_1) == 2 * FILTER_BITS)); in av1_convolve_x_sr_sse2()
|
D | jnt_convolve_sse2.c | 32 const __m128i round_const = _mm_set1_epi32((1 << conv_params->round_0) >> 1); in av1_dist_wtd_convolve_x_sse2() 33 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_dist_wtd_convolve_x_sse2() 42 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_sse2() 46 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_x_sse2() 164 const int bits = FILTER_BITS - conv_params->round_0; in av1_dist_wtd_convolve_y_sse2() 172 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_sse2() 176 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_y_sse2() 417 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_sse2() 421 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_sse2() 445 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_dist_wtd_convolve_2d_sse2() [all …]
|
D | convolve_avx2.c | 34 assert(conv_params->round_0 <= FILTER_BITS); in av1_convolve_y_sr_avx2() 35 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_convolve_y_sr_avx2() 36 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_convolve_y_sr_avx2() 269 const int bits = FILTER_BITS - conv_params->round_0; in av1_convolve_x_sr_avx2() 272 _mm256_set1_epi16((1 << (conv_params->round_0 - 1)) >> 1); in av1_convolve_x_sr_avx2() 273 const __m128i round_0_shift = _mm_cvtsi32_si128(conv_params->round_0 - 1); in av1_convolve_x_sr_avx2() 282 ((conv_params->round_0 + conv_params->round_1) == 2 * FILTER_BITS)); in av1_convolve_x_sr_avx2() 283 assert(conv_params->round_0 > 0); in av1_convolve_x_sr_avx2()
|
D | highbd_wiener_convolve_avx2.c | 35 assert(bd + FILTER_BITS - conv_params->round_0 + 2 <= 16); in av1_highbd_wiener_convolve_add_src_avx2() 59 _mm256_set1_epi16(WIENER_CLAMP_LIMIT(conv_params->round_0, bd) - 1); in av1_highbd_wiener_convolve_add_src_avx2() 88 (1 << (conv_params->round_0 - 1)) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_wiener_convolve_add_src_avx2() 119 _mm256_add_epi32(res_even_sum, round_const), conv_params->round_0); in av1_highbd_wiener_convolve_add_src_avx2() 124 _mm256_add_epi32(res_odd_sum, round_const), conv_params->round_0); in av1_highbd_wiener_convolve_add_src_avx2()
|
D | highbd_convolve_2d_ssse3.c | 37 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_convolve_2d_sr_ssse3() 41 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_convolve_2d_sr_ssse3() 42 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_convolve_2d_sr_ssse3() 46 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_convolve_2d_sr_ssse3() 50 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_highbd_convolve_2d_sr_ssse3()
|
D | jnt_convolve_ssse3.c | 48 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_ssse3() 52 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_ssse3() 76 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_dist_wtd_convolve_2d_ssse3() 77 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_dist_wtd_convolve_2d_ssse3() 145 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_dist_wtd_convolve_2d_ssse3()
|
D | convolve_2d_sse2.c | 40 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_convolve_2d_sr_sse2() 41 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_convolve_2d_sr_sse2() 43 assert(conv_params->round_0 > 0); in av1_convolve_2d_sr_sse2() 66 (1 << (bd + FILTER_BITS - 1)) + ((1 << conv_params->round_0) >> 1)); in av1_convolve_2d_sr_sse2() 67 const __m128i round_shift = _mm_cvtsi32_si128(conv_params->round_0); in av1_convolve_2d_sr_sse2() 371 FILTER_BITS * 2 - conv_params->round_1 - conv_params->round_0; in av1_dist_wtd_convolve_2d_copy_sse2() 385 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_copy_sse2() 389 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_dist_wtd_convolve_2d_copy_sse2()
|
D | highbd_jnt_convolve_sse4.c | 29 const int bits = FILTER_BITS - conv_params->round_0; in av1_highbd_dist_wtd_convolve_y_sse4_1() 48 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_y_sse4_1() 52 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_y_sse4_1() 287 _mm_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_dist_wtd_convolve_x_sse4_1() 288 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_dist_wtd_convolve_x_sse4_1() 292 bd + 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_x_sse4_1() 296 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in av1_highbd_dist_wtd_convolve_x_sse4_1()
|
D | wiener_convolve_sse2.c | 65 (1 << (conv_params->round_0 - 1)) + (1 << (bd + FILTER_BITS - 1))); in av1_wiener_convolve_add_src_sse2() 85 conv_params->round_0); in av1_wiener_convolve_add_src_sse2() 100 conv_params->round_0); in av1_wiener_convolve_add_src_sse2() 106 _mm_set1_epi16(WIENER_CLAMP_LIMIT(conv_params->round_0, bd) - 1)); in av1_wiener_convolve_add_src_sse2()
|
D | convolve_2d_avx2.c | 34 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_convolve_2d_sr_avx2() 35 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in av1_convolve_2d_sr_avx2() 37 assert(conv_params->round_0 > 0); in av1_convolve_2d_sr_avx2() 40 ((1 << (conv_params->round_0 - 1)) >> 1) + (1 << (bd + FILTER_BITS - 2))); in av1_convolve_2d_sr_avx2() 41 const __m128i round_shift_h = _mm_cvtsi32_si128(conv_params->round_0 - 1); in av1_convolve_2d_sr_avx2()
|
D | highbd_convolve_2d_avx2.c | 40 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_convolve_2d_sr_avx2() 45 ((1 << conv_params->round_0) >> 1) + (1 << (bd + FILTER_BITS - 1))); in av1_highbd_convolve_2d_sr_avx2() 46 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_convolve_2d_sr_avx2() 50 (1 << (bd + 2 * FILTER_BITS - conv_params->round_0 - 1))); in av1_highbd_convolve_2d_sr_avx2() 54 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in av1_highbd_convolve_2d_sr_avx2()
|
D | av1_convolve_scale_sse4.c | 111 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in vfilter8() 123 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in vfilter8() 251 x_step_qn, filter_params_x, conv_params->round_0); in av1_convolve_2d_scale_sse4_1() 337 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in highbd_vfilter8() 351 FILTER_BITS * 2 - conv_params->round_0 - conv_params->round_1; in highbd_vfilter8() 358 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in highbd_vfilter8() 492 subpel_x_qn, x_step_qn, filter_params_x, conv_params->round_0, in av1_highbd_convolve_2d_scale_sse4_1()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | highbd_convolve_ssse3.c | 33 assert(conv_params->round_0 <= FILTER_BITS); in av1_highbd_convolve_y_sr_ssse3() 34 assert(((conv_params->round_0 + conv_params->round_1) <= (FILTER_BITS + 1)) || in av1_highbd_convolve_y_sr_ssse3() 35 ((conv_params->round_0 + conv_params->round_1) == (2 * FILTER_BITS))); in av1_highbd_convolve_y_sr_ssse3() 182 assert(bd + FILTER_BITS + 2 - conv_params->round_0 <= 16); in av1_highbd_convolve_x_sr_ssse3() 187 _mm_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_convolve_x_sr_ssse3() 188 const __m128i round_shift_x = _mm_cvtsi32_si128(conv_params->round_0); in av1_highbd_convolve_x_sr_ssse3() 190 const int bits = FILTER_BITS - conv_params->round_0; in av1_highbd_convolve_x_sr_ssse3()
|
/external/libaom/libaom/aom_dsp/ |
D | blend_a64_mask.c | 43 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in aom_lowbd_blend_a64_d16_mask_c() 47 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in aom_lowbd_blend_a64_d16_mask_c() 129 const int offset_bits = bd + 2 * FILTER_BITS - conv_params->round_0; in aom_highbd_blend_a64_d16_mask_c() 133 2 * FILTER_BITS - conv_params->round_0 - conv_params->round_1; in aom_highbd_blend_a64_d16_mask_c()
|
/external/libaom/libaom/test/ |
D | reconinter_test.cc | 106 bd + 2 * FILTER_BITS - conv_params.round_0 - conv_params.round_1 + 2; in RunCheckOutput() 146 bd + 2 * FILTER_BITS - conv_params.round_0 - conv_params.round_1 + 2; in RunSpeedTest()
|