/external/libaom/libaom/av1/encoder/x86/ |
D | encodetxb_sse4.c | 45 xx_storeu_128(ls, lsAB); in av1_txb_init_levels_sse4_1() 57 xx_storeu_128(ls, absAB8); in av1_txb_init_levels_sse4_1() 75 xx_storeu_128(ls + j, absABCD); in av1_txb_init_levels_sse4_1()
|
D | rdopt_sse4.c | 112 xx_storeu_128(xy_tmp, xy_sum_32); in av1_get_horver_correlation_full_sse4_1() 113 xx_storeu_128(xz_tmp, xz_sum_32); in av1_get_horver_correlation_full_sse4_1() 114 xx_storeu_128(x_tmp, x_sum_32); in av1_get_horver_correlation_full_sse4_1() 115 xx_storeu_128(x2_tmp, x2_sum_32); in av1_get_horver_correlation_full_sse4_1()
|
D | encodetxb_avx2.c | 94 xx_storeu_128(ls, _mm256_castsi256_si128(res)); in av1_txb_init_levels_avx2() 95 xx_storeu_128(ls + stride, _mm256_extracti128_si256(res, 1)); in av1_txb_init_levels_avx2()
|
D | pickrst_sse4.c | 30 xx_storeu_128(dst, r0); in acc_stat_sse41() 31 xx_storeu_128(dst + 4, r1); in acc_stat_sse41() 159 xx_storeu_128(dst, rll); in acc_stat_highbd_sse41() 161 xx_storeu_128(dst + 2, rlh); in acc_stat_highbd_sse41() 163 xx_storeu_128(dst + 4, rhl); in acc_stat_highbd_sse41() 165 xx_storeu_128(dst + 6, rhh); in acc_stat_highbd_sse41() 620 xx_storeu_128(sum, sum64); in av1_lowbd_pixel_proj_error_sse4_1() 826 xx_storeu_128(sum, sum64); in av1_highbd_pixel_proj_error_sse4_1()
|
/external/libaom/libaom/av1/common/x86/ |
D | selfguided_sse4.c | 224 xx_storeu_128(A + i * buf_stride + j, a_res); in calc_ab() 237 xx_storeu_128(B + i * buf_stride + j, b_res); in calc_ab() 301 xx_storeu_128(dst + i * dst_stride + j, w); in final_filter() 364 xx_storeu_128(A + i * buf_stride + j, a_res); in calc_ab_fast() 377 xx_storeu_128(B + i * buf_stride + j, b_res); in calc_ab_fast() 481 xx_storeu_128(dst + i * dst_stride + j, w); in final_filter_fast() 496 xx_storeu_128(dst + i * dst_stride + j, w); in final_filter_fast() 651 xx_storeu_128(CONVERT_TO_SHORTPTR(dst8 + m), res); in apply_selfguided_restoration_sse4_1()
|
D | reconinter_avx2.c | 60 xx_storeu_128(mask, x_m8); in av1_build_compound_diffwtd_mask_avx2() 196 xx_storeu_128(mask, in build_compound_diffwtd_mask_d16_avx2() 358 xx_storeu_128(mask, in build_compound_diffwtd_mask_d16_inv_avx2()
|
D | wiener_convolve_avx2.c | 256 xx_storeu_128(dst + i * dst_stride + j, in av1_wiener_convolve_add_src_avx2()
|
D | selfguided_avx2.c | 720 xx_storeu_128(dst8 + m, res2); in apply_selfguided_restoration_avx2()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | blend_a64_mask_sse4.c | 89 xx_storeu_128(dst + c, v_res_b); in blend_a64_mask_w16n_sse4_1() 180 xx_storeu_128(dst + c, v_res_b); in blend_a64_mask_sx_w16n_sse4_1() 260 xx_storeu_128(dst + c, v_res_b); in blend_a64_mask_sy_w16n_sse4_1() 372 xx_storeu_128(dst + c, v_res_b); in blend_a64_mask_sx_sy_w16n_sse4_1() 486 xx_storeu_128(dst + c, v_res_w); in blend_a64_mask_bn_w8n_sse4_1() 583 xx_storeu_128(dst + c, v_res_w); in blend_a64_mask_bn_sx_w8n_sse4_1() 678 xx_storeu_128(dst + c, v_res_w); in blend_a64_mask_bn_sy_w8n_sse4_1() 785 xx_storeu_128(dst + c, v_res_w); in blend_a64_mask_bn_sx_sy_w8n_sse4_1() 1303 xx_storeu_128(dst, clipa); in highbd_blend_a64_d16_mask_w8_sse4_1() 1304 xx_storeu_128(dst + dst_stride, clipb); in highbd_blend_a64_d16_mask_w8_sse4_1() [all …]
|
D | blend_a64_vmask_sse4.c | 99 xx_storeu_128(dst + c, v_res_b); in blend_a64_vmask_w16n_sse4_1() 206 xx_storeu_128(dst + c, v_res_w); in blend_a64_vmask_bn_w8n_sse4_1()
|
D | synonyms.h | 62 static INLINE void xx_storeu_128(void *const a, const __m128i v) { in xx_storeu_128() function
|
D | variance_impl_ssse3.c | 53 xx_storeu_128(b, res); in aom_var_filter_block2d_bil_first_pass_ssse3()
|
D | variance_sse2.c | 587 xx_storeu_128(comp_pred, xx_loadu_128(ref)); in aom_upsampled_pred_sse2() 601 xx_storeu_128(comp_pred, _mm_unpacklo_epi64(s0, s1)); in aom_upsampled_pred_sse2() 617 xx_storeu_128(comp_pred, reg); in aom_upsampled_pred_sse2() 669 xx_storeu_128(comp_pred, _mm_avg_epu8(s0, p0)); in aom_comp_avg_upsampled_pred_sse2()
|
D | jnt_variance_ssse3.c | 45 xx_storeu_128(result, _mm_packus_epi16(shift_lo, shift_hi)); in compute_dist_wtd_avg()
|
D | blend_a64_mask_avx2.c | 465 xx_storeu_128(dst, _mm256_castsi256_si128(y_res_b)); in blend_a64_mask_sx_sy_w16_avx2() 598 xx_storeu_128(dst, _mm256_castsi256_si128(v_res_b)); in blend_a64_mask_sx_w16_avx2() 711 xx_storeu_128(dst, v_res_b); in blend_a64_mask_sy_w16_avx2() 856 xx_storeu_128(dst, v_res_b); in blend_a64_mask_avx2()
|
D | highbd_variance_sse2.c | 803 xx_storeu_128(result, shift); in highbd_compute_dist_wtd_comp_avg()
|