Home
last modified time | relevance | path

Searched refs:xx_roundn_epu16 (Results 1 – 7 of 7) sorted by relevance

/external/libaom/libaom/aom_dsp/x86/
Dblend_sse4.h36 const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); in blend_4()
53 const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); in blend_8()
116 const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); in blend_4_b10()
131 const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); in blend_8_b10()
Dmasked_sad4d_ssse3.c29 pred_l = xx_roundn_epu16(pred_l, AOM_BLEND_A64_ROUND_BITS); \
34 pred_r = xx_roundn_epu16(pred_r, AOM_BLEND_A64_ROUND_BITS); \
96 pred_l = xx_roundn_epu16(pred_l, AOM_BLEND_A64_ROUND_BITS); \
101 pred_r = xx_roundn_epu16(pred_r, AOM_BLEND_A64_ROUND_BITS); \
162 pred = xx_roundn_epu16(pred, AOM_BLEND_A64_ROUND_BITS); \
Dmasked_sad_intrin_ssse3.c118 pred_l = xx_roundn_epu16(pred_l, AOM_BLEND_A64_ROUND_BITS); in masked_sad_ssse3()
123 pred_r = xx_roundn_epu16(pred_r, AOM_BLEND_A64_ROUND_BITS); in masked_sad_ssse3()
165 pred_l = xx_roundn_epu16(pred_l, AOM_BLEND_A64_ROUND_BITS); in aom_masked_sad8xh_ssse3()
170 pred_r = xx_roundn_epu16(pred_r, AOM_BLEND_A64_ROUND_BITS); in aom_masked_sad8xh_ssse3()
214 pred_16bit = xx_roundn_epu16(pred_16bit, AOM_BLEND_A64_ROUND_BITS); in aom_masked_sad4xh_ssse3()
Dmasked_variance_intrin_ssse3.c140 v0 = xx_roundn_epu16(v0, FILTER_BITS); in filter_block()
144 v1 = xx_roundn_epu16(v1, FILTER_BITS); in filter_block()
226 v0 = xx_roundn_epu16(v0, FILTER_BITS); in filter_block_2rows()
230 v1 = xx_roundn_epu16(v1, FILTER_BITS); in filter_block_2rows()
277 v0 = xx_roundn_epu16(v0, FILTER_BITS); in bilinear_filter8xh()
358 v0 = xx_roundn_epu16(v0, FILTER_BITS); in bilinear_filter4xh()
409 pred_l = xx_roundn_epu16(pred_l, AOM_BLEND_A64_ROUND_BITS); in accumulate_block()
414 pred_r = xx_roundn_epu16(pred_r, AOM_BLEND_A64_ROUND_BITS); in accumulate_block()
Dsynonyms.h92 static INLINE __m128i xx_roundn_epu16(__m128i v_val_w, int bits) { in xx_roundn_epu16() function
Dblend_a64_mask_sse4.c290 const __m128i v_m0_w = xx_roundn_epu16(v_rs_w, 2); in blend_a64_mask_sx_sy_w4_sse4_1()
323 const __m128i v_m0_w = xx_roundn_epu16(v_rs_w, 2); in blend_a64_mask_sx_sy_w8_sse4_1()
364 const __m128i v_m0l_w = xx_roundn_epu16(v_rsl_w, 2); in blend_a64_mask_sx_sy_w16n_sse4_1()
365 const __m128i v_m0h_w = xx_roundn_epu16(v_rsh_w, 2); in blend_a64_mask_sx_sy_w16n_sse4_1()
727 const __m128i v_m0_w = xx_roundn_epu16(v_rs_w, 2); in blend_a64_mask_bn_sx_sy_w4_sse4_1()
781 const __m128i v_m0_w = xx_roundn_epu16(v_rs_w, 2); in blend_a64_mask_bn_sx_sy_w8n_sse4_1()
Dblend_a64_mask_avx2.c532 const __m128i v_m0_w = xx_roundn_epu16(v_rs_w, 2); in blend_a64_mask_sx_sy_avx2()
555 const __m128i v_m0_w = xx_roundn_epu16(v_rs_w, 2); in blend_a64_mask_sx_sy_avx2()