Home
last modified time | relevance | path

Searched refs:_mm256_srai_epi32 (Results 1 – 25 of 27) sorted by relevance

12

/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_dct32x32_impl_avx2.h414 const __m256i s2_20_6 = _mm256_srai_epi32(s2_20_4, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
415 const __m256i s2_20_7 = _mm256_srai_epi32(s2_20_5, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
416 const __m256i s2_21_6 = _mm256_srai_epi32(s2_21_4, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
417 const __m256i s2_21_7 = _mm256_srai_epi32(s2_21_5, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
418 const __m256i s2_22_6 = _mm256_srai_epi32(s2_22_4, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
419 const __m256i s2_22_7 = _mm256_srai_epi32(s2_22_5, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
420 const __m256i s2_23_6 = _mm256_srai_epi32(s2_23_4, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
421 const __m256i s2_23_7 = _mm256_srai_epi32(s2_23_5, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
422 const __m256i s2_24_6 = _mm256_srai_epi32(s2_24_4, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
423 const __m256i s2_24_7 = _mm256_srai_epi32(s2_24_5, DCT_CONST_BITS); in FDCT32x32_2D_AVX2()
[all …]
Davg_intrin_avx2.c159 b0 = _mm256_srai_epi32(b0, 1); in vpx_highbd_hadamard_16x16_avx2()
160 b1 = _mm256_srai_epi32(b1, 1); in vpx_highbd_hadamard_16x16_avx2()
161 b2 = _mm256_srai_epi32(b2, 1); in vpx_highbd_hadamard_16x16_avx2()
162 b3 = _mm256_srai_epi32(b3, 1); in vpx_highbd_hadamard_16x16_avx2()
200 b0 = _mm256_srai_epi32(b0, 2); in vpx_highbd_hadamard_32x32_avx2()
201 b1 = _mm256_srai_epi32(b1, 2); in vpx_highbd_hadamard_32x32_avx2()
202 b2 = _mm256_srai_epi32(b2, 2); in vpx_highbd_hadamard_32x32_avx2()
203 b3 = _mm256_srai_epi32(b3, 2); in vpx_highbd_hadamard_32x32_avx2()
Dconvolve_avx2.h141 return _mm256_srai_epi32(nearest_src, depth); in mm256_round_epi32()
Dhighbd_convolve_avx2.c310 *y = _mm256_srai_epi32(a, CONV8_ROUNDING_BITS); in filter_8x1_pixels()
447 *y0 = _mm256_srai_epi32(x0, CONV8_ROUNDING_BITS); in filter_16_2t_pixels()
448 *y1 = _mm256_srai_epi32(x1, CONV8_ROUNDING_BITS); in filter_16_2t_pixels()
456 *y0 = _mm256_srai_epi32(x0, CONV8_ROUNDING_BITS); in filter_8x1_2t_pixels()
/external/libaom/libaom/av1/encoder/x86/
Dav1_fwd_txfm_avx2.h29 *in0 = _mm256_srai_epi32(temp0, cos_bit); in btf_32_avx2_type0()
34 *in1 = _mm256_srai_epi32(temp1, cos_bit); in btf_32_avx2_type0()
48 *in0 = _mm256_srai_epi32(temp0, cos_bit); in btf_32_avx2_type1()
53 *in1 = _mm256_srai_epi32(temp1, cos_bit); in btf_32_avx2_type1()
68 *in0 = _mm256_srai_epi32(temp0, cos_bit); in btf_32_avx2_type0_new()
73 *in1 = _mm256_srai_epi32(temp1, cos_bit); in btf_32_avx2_type0_new()
88 *in0 = _mm256_srai_epi32(temp0, cos_bit); in btf_32_avx2_type1_new()
93 *in1 = _mm256_srai_epi32(temp1, cos_bit); in btf_32_avx2_type1_new()
Dhighbd_fwd_txfm_avx2.c88 in[0] = _mm256_srai_epi32(in[0], shift); in col_txfm_8x8_rounding()
89 in[1] = _mm256_srai_epi32(in[1], shift); in col_txfm_8x8_rounding()
90 in[2] = _mm256_srai_epi32(in[2], shift); in col_txfm_8x8_rounding()
91 in[3] = _mm256_srai_epi32(in[3], shift); in col_txfm_8x8_rounding()
92 in[4] = _mm256_srai_epi32(in[4], shift); in col_txfm_8x8_rounding()
93 in[5] = _mm256_srai_epi32(in[5], shift); in col_txfm_8x8_rounding()
94 in[6] = _mm256_srai_epi32(in[6], shift); in col_txfm_8x8_rounding()
95 in[7] = _mm256_srai_epi32(in[7], shift); in col_txfm_8x8_rounding()
190 in[stride * i] = _mm256_srai_epi32(in[stride * i], bit); in av1_round_shift_32_8xn_avx2()
222 x = _mm256_srai_epi32(x, bit); in av1_half_btf_avx2()
[all …]
Dpickrst_avx2.c527 _mm256_srai_epi32(_mm256_add_epi32(v0, rounding), shift); in av1_lowbd_pixel_proj_error_avx2()
529 _mm256_srai_epi32(_mm256_add_epi32(v1, rounding), shift); in av1_lowbd_pixel_proj_error_avx2()
572 _mm256_srai_epi32(_mm256_add_epi32(v0, rounding), shift); in av1_lowbd_pixel_proj_error_avx2()
574 _mm256_srai_epi32(_mm256_add_epi32(v1, rounding), shift); in av1_lowbd_pixel_proj_error_avx2()
681 _mm256_srai_epi32(_mm256_add_epi32(vl, rounding), shift); in av1_highbd_pixel_proj_error_avx2()
683 _mm256_srai_epi32(_mm256_add_epi32(vh, rounding), shift); in av1_highbd_pixel_proj_error_avx2()
758 _mm256_srai_epi32(_mm256_add_epi32(vh, rounding), shift); in av1_highbd_pixel_proj_error_avx2()
760 _mm256_srai_epi32(_mm256_add_epi32(vl, rounding), shift); in av1_highbd_pixel_proj_error_avx2()
Dav1_highbd_quantize_avx2.c68 dq = _mm256_srai_epi32(dq, log_scale); in quantize()
Dwedge_utils_avx2.c134 __m256i v_sign_d = _mm256_srai_epi32(v_acc0_d, 31); in av1_wedge_sign_from_residuals_avx2()
Dav1_fwd_txfm2d_avx2.c2010 __m256i c0 = _mm256_srai_epi32(a0, *cos_bit); in btf_16_avx2()
2011 __m256i c1 = _mm256_srai_epi32(a1, *cos_bit); in btf_16_avx2()
2012 __m256i d0 = _mm256_srai_epi32(b0, *cos_bit); in btf_16_avx2()
2013 __m256i d1 = _mm256_srai_epi32(b1, *cos_bit); in btf_16_avx2()
/external/libaom/libaom/av1/common/x86/
Dhighbd_inv_txfm_avx2.c206 x = _mm256_srai_epi32(x, bit); in half_btf_0_avx2()
219 x = _mm256_srai_epi32(x, bit); in half_btf_avx2()
480 x = _mm256_srai_epi32(x, bit); in idct32_low1_avx2()
1229 in[0] = _mm256_srai_epi32(in[0], bit); in idct16_low1_avx2()
1334 u[0] = _mm256_srai_epi32(u[0], bit); in idct16_low8_avx2()
1358 u[5] = _mm256_srai_epi32(u[5], bit); in idct16_low8_avx2()
1362 u[6] = _mm256_srai_epi32(u[6], bit); in idct16_low8_avx2()
1379 u[10] = _mm256_srai_epi32(u[10], bit); in idct16_low8_avx2()
1383 u[13] = _mm256_srai_epi32(u[13], bit); in idct16_low8_avx2()
1389 u[11] = _mm256_srai_epi32(u[11], bit); in idct16_low8_avx2()
[all …]
Dhighbd_wiener_convolve_avx2.c118 const __m256i res_even = _mm256_srai_epi32( in av1_highbd_wiener_convolve_add_src_avx2()
123 const __m256i res_odd = _mm256_srai_epi32( in av1_highbd_wiener_convolve_add_src_avx2()
228 const __m256i res_lo_round = _mm256_srai_epi32( in av1_highbd_wiener_convolve_add_src_avx2()
230 const __m256i res_hi_round = _mm256_srai_epi32( in av1_highbd_wiener_convolve_add_src_avx2()
Dwiener_convolve_avx2.c129 const __m256i res_even = _mm256_srai_epi32( in av1_wiener_convolve_add_src_avx2()
131 const __m256i res_odd = _mm256_srai_epi32( in av1_wiener_convolve_add_src_avx2()
234 const __m256i res_lo_round = _mm256_srai_epi32( in av1_wiener_convolve_add_src_avx2()
236 const __m256i res_hi_round = _mm256_srai_epi32( in av1_wiener_convolve_add_src_avx2()
Dselfguided_avx2.c344 __m256i w = _mm256_srai_epi32(_mm256_add_epi32(v, rounding), in final_filter()
523 _mm256_srai_epi32(_mm256_add_epi32(v, rounding0), in final_filter_fast()
540 _mm256_srai_epi32(_mm256_add_epi32(v, rounding1), in final_filter_fast()
696 const __m256i w_0 = _mm256_srai_epi32( in apply_selfguided_restoration_avx2()
698 const __m256i w_1 = _mm256_srai_epi32( in apply_selfguided_restoration_avx2()
Dav1_inv_txfm_avx2.c1699 lo = _mm256_srai_epi32(lo, NewSqrt2Bits - shift); in iidentity_row_16xn_avx2()
1700 hi = _mm256_srai_epi32(hi, NewSqrt2Bits - shift); in iidentity_row_16xn_avx2()
1714 lo = _mm256_srai_epi32(lo, NewSqrt2Bits - shift); in iidentity_row_16xn_avx2()
1715 hi = _mm256_srai_epi32(hi, NewSqrt2Bits - shift); in iidentity_row_16xn_avx2()
1734 lo = _mm256_srai_epi32(lo, NewSqrt2Bits); in iidentity_col_16xn_avx2()
1735 hi = _mm256_srai_epi32(hi, NewSqrt2Bits); in iidentity_col_16xn_avx2()
1738 lo = _mm256_srai_epi32(lo, -shift); in iidentity_col_16xn_avx2()
1739 hi = _mm256_srai_epi32(hi, -shift); in iidentity_col_16xn_avx2()
/external/libaom/libaom/aom_dsp/x86/
Dobmc_variance_avx2.c56 const __m256i v_sign_d = _mm256_srai_epi32(v_diff0_d, 31); in obmc_variance_w8n()
59 const __m256i v_rdiff0_d = _mm256_srai_epi32(v_tmp_d, 12); in obmc_variance_w8n()
119 const __m256i v_sign0_d = _mm256_srai_epi32(v_diff0_d, 31); in obmc_variance_w16n()
120 const __m256i v_sign1_d = _mm256_srai_epi32(v_diff1_d, 31); in obmc_variance_w16n()
127 const __m256i v_rdiff0_d = _mm256_srai_epi32(v_tmp0_d, 12); in obmc_variance_w16n()
128 const __m256i v_rdiff2_d = _mm256_srai_epi32(v_tmp1_d, 12); in obmc_variance_w16n()
Dtxfm_common_avx2.h43 __m256i c0 = _mm256_srai_epi32(a0, cos_bit); in btf_16_w16_avx2()
44 __m256i c1 = _mm256_srai_epi32(a1, cos_bit); in btf_16_w16_avx2()
45 __m256i d0 = _mm256_srai_epi32(b0, cos_bit); in btf_16_w16_avx2()
46 __m256i d1 = _mm256_srai_epi32(b1, cos_bit); in btf_16_w16_avx2()
226 return _mm256_srai_epi32(tmp, bit); in av1_round_shift_32_avx2()
273 return _mm256_srai_epi32(b, NewSqrt2Bits); in scale_round_avx2()
Dconvolve_avx2.h370 d = _mm256_srai_epi32(d, 1); in add_store_aligned_256()
389 const __m256i res_lo = _mm256_srai_epi32(wt_res_lo, DIST_PRECISION_BITS); in comp_avg()
390 const __m256i res_hi = _mm256_srai_epi32(wt_res_hi, DIST_PRECISION_BITS); in comp_avg()
420 res = _mm256_srai_epi32(wt_res, DIST_PRECISION_BITS); in highbd_comp_avg()
423 res = _mm256_srai_epi32(wt_res, 1); in highbd_comp_avg()
432 const __m256i res_round = _mm256_srai_epi32( in highbd_convolve_rounding()
Dblend_a64_mask_avx2.c40 _mm256_srai_epi32(_mm256_sub_epi32(res0_lo, *v_round_offset), shift); in blend_a64_d16_mask_w16_avx2()
42 _mm256_srai_epi32(_mm256_sub_epi32(res0_hi, *v_round_offset), shift); in blend_a64_d16_mask_w16_avx2()
68 _mm256_srai_epi32(_mm256_sub_epi32(res0_lo, *v_round_offset), shift); in blend_a64_d16_mask_w32_avx2()
70 _mm256_srai_epi32(_mm256_sub_epi32(res0_hi, *v_round_offset), shift); in blend_a64_d16_mask_w32_avx2()
72 _mm256_srai_epi32(_mm256_sub_epi32(res1_lo, *v_round_offset), shift); in blend_a64_d16_mask_w32_avx2()
74 _mm256_srai_epi32(_mm256_sub_epi32(res1_hi, *v_round_offset), shift); in blend_a64_d16_mask_w32_avx2()
941 _mm256_srai_epi32(_mm256_sub_epi32(sumh, *round_offset), shift); in highbd_blend_a64_d16_mask_w4_avx2()
943 _mm256_srai_epi32(_mm256_sub_epi32(suml, *round_offset), shift); in highbd_blend_a64_d16_mask_w4_avx2()
1073 _mm256_srai_epi32(_mm256_sub_epi32(sumah, *round_offset), shift); in highbd_blend_a64_d16_mask_w8_avx2()
1075 _mm256_srai_epi32(_mm256_sub_epi32(sumal, *round_offset), shift); in highbd_blend_a64_d16_mask_w8_avx2()
[all …]
Dmasked_sad_intrin_avx2.c229 pred_l = _mm256_srai_epi32(_mm256_add_epi32(pred_l, round_const), in highbd_masked_sad8xh_avx2()
235 pred_r = _mm256_srai_epi32(_mm256_add_epi32(pred_r, round_const), in highbd_masked_sad8xh_avx2()
286 pred_l = _mm256_srai_epi32(_mm256_add_epi32(pred_l, round_const), in highbd_masked_sad16xh_avx2()
292 pred_r = _mm256_srai_epi32(_mm256_add_epi32(pred_r, round_const), in highbd_masked_sad16xh_avx2()
Dhighbd_convolve_avx2.c489 *y = _mm256_srai_epi32(a, CONV8_ROUNDING_BITS); in filter_8x1_pixels()
600 _mm256_srai_epi32(_mm256_add_epi32(res, rounding), CONV8_ROUNDING_BITS); in aom_highbd_filter_block1d4_h4_avx2()
626 _mm256_srai_epi32(_mm256_add_epi32(res, rounding), CONV8_ROUNDING_BITS); in aom_highbd_filter_block1d4_h4_avx2()
676 res_even = _mm256_srai_epi32(_mm256_add_epi32(res_even, rounding), in aom_highbd_filter_block1d8_h4_avx2()
684 res_odd = _mm256_srai_epi32(_mm256_add_epi32(res_odd, rounding), in aom_highbd_filter_block1d8_h4_avx2()
714 _mm256_srai_epi32(_mm256_add_epi32(res, rounding), CONV8_ROUNDING_BITS); in aom_highbd_filter_block1d8_h4_avx2()
796 *y0 = _mm256_srai_epi32(x0, CONV8_ROUNDING_BITS); in filter_16_2t_pixels()
797 *y1 = _mm256_srai_epi32(x1, CONV8_ROUNDING_BITS); in filter_16_2t_pixels()
805 *y0 = _mm256_srai_epi32(x0, CONV8_ROUNDING_BITS); in filter_8x1_2t_pixels()
Dvariance_avx2.c422 const __m256i pred_l = _mm256_srai_epi32( in highbd_comp_mask_pred_line_avx2()
428 const __m256i pred_h = _mm256_srai_epi32( in highbd_comp_mask_pred_line_avx2()
/external/libaom/libaom/aom_dsp/simd/
Dv256_intrinsics_x86.h221 _mm256_packs_epi32(_mm256_srai_epi32(b, 16), _mm256_srai_epi32(a, 16)), in v256_unziphi_16()
328 return _mm256_srai_epi32( in v256_unpacklo_s16_s32()
341 return _mm256_srai_epi32( in v256_unpackhi_s16_s32()
694 #define v256_shr_n_s32(a, c) _mm256_srai_epi32(a, c)
/external/clang/test/CodeGen/
Davx2-builtins.c1039 return _mm256_srai_epi32(a, 3); in test_mm256_srai_epi32()
/external/clang/lib/Headers/
Davx2intrin.h644 _mm256_srai_epi32(__m256i __a, int __count) in _mm256_srai_epi32() function

12