Home
last modified time | relevance | path

Searched refs:_mm_srai_epi32 (Results 1 – 25 of 113) sorted by relevance

12345

/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_dct_intrin_sse2.c98 u[0] = _mm_srai_epi32(v[0], DCT_CONST_BITS); in fdct4_sse2()
99 u[1] = _mm_srai_epi32(v[1], DCT_CONST_BITS); in fdct4_sse2()
100 u[2] = _mm_srai_epi32(v[2], DCT_CONST_BITS); in fdct4_sse2()
101 u[3] = _mm_srai_epi32(v[3], DCT_CONST_BITS); in fdct4_sse2()
146 u[0] = _mm_srai_epi32(v[0], DCT_CONST_BITS); in fadst4_sse2()
147 u[1] = _mm_srai_epi32(v[1], DCT_CONST_BITS); in fadst4_sse2()
148 u[2] = _mm_srai_epi32(v[2], DCT_CONST_BITS); in fadst4_sse2()
149 u[3] = _mm_srai_epi32(v[3], DCT_CONST_BITS); in fadst4_sse2()
326 u0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in fdct8_sse2()
327 u1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in fdct8_sse2()
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_dct32x32_impl_sse2.h438 const __m128i s2_20_6 = _mm_srai_epi32(s2_20_4, DCT_CONST_BITS); in FDCT32x32_2D()
439 const __m128i s2_20_7 = _mm_srai_epi32(s2_20_5, DCT_CONST_BITS); in FDCT32x32_2D()
440 const __m128i s2_21_6 = _mm_srai_epi32(s2_21_4, DCT_CONST_BITS); in FDCT32x32_2D()
441 const __m128i s2_21_7 = _mm_srai_epi32(s2_21_5, DCT_CONST_BITS); in FDCT32x32_2D()
442 const __m128i s2_22_6 = _mm_srai_epi32(s2_22_4, DCT_CONST_BITS); in FDCT32x32_2D()
443 const __m128i s2_22_7 = _mm_srai_epi32(s2_22_5, DCT_CONST_BITS); in FDCT32x32_2D()
444 const __m128i s2_23_6 = _mm_srai_epi32(s2_23_4, DCT_CONST_BITS); in FDCT32x32_2D()
445 const __m128i s2_23_7 = _mm_srai_epi32(s2_23_5, DCT_CONST_BITS); in FDCT32x32_2D()
446 const __m128i s2_24_6 = _mm_srai_epi32(s2_24_4, DCT_CONST_BITS); in FDCT32x32_2D()
447 const __m128i s2_24_7 = _mm_srai_epi32(s2_24_5, DCT_CONST_BITS); in FDCT32x32_2D()
[all …]
Dfwd_txfm_sse2.c32 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct4x4_1_sse2()
33 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct4x4_1_sse2()
73 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct8x8_1_sse2()
74 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct8x8_1_sse2()
141 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct16x16_1_sse2()
142 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct16x16_1_sse2()
152 in1 = _mm_srai_epi32(in1, 1); in vpx_fdct16x16_1_sse2()
213 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct32x32_1_sse2()
214 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct32x32_1_sse2()
224 in1 = _mm_srai_epi32(in1, 3); in vpx_fdct32x32_1_sse2()
Dfwd_txfm_impl_sse2.h159 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT4x4_2D()
160 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT4x4_2D()
161 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS); in FDCT4x4_2D()
162 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS); in FDCT4x4_2D()
219 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS2); in FDCT4x4_2D()
220 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS2); in FDCT4x4_2D()
221 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS2); in FDCT4x4_2D()
222 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS2); in FDCT4x4_2D()
358 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT8x8_2D()
359 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT8x8_2D()
[all …]
/external/libavc/common/x86/
Dih264_iquant_itrans_recon_ssse3.c165 resq_r0 = _mm_srai_epi32(temp4, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3()
166 resq_r1 = _mm_srai_epi32(temp5, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3()
167 resq_r2 = _mm_srai_epi32(temp6, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3()
168 resq_r3 = _mm_srai_epi32(temp7, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3()
205 temp2 = _mm_srai_epi32(resq_r1, 1); //(w1>>1) in ih264_iquant_itrans_recon_4x4_ssse3()
208 temp3 = _mm_srai_epi32(resq_r3, 1); //(w3>>1) + w1 in ih264_iquant_itrans_recon_4x4_ssse3()
264 temp2 = _mm_srai_epi32(resq_r1, 1); //(y1j>>1) in ih264_iquant_itrans_recon_4x4_ssse3()
267 temp3 = _mm_srai_epi32(resq_r3, 1); //(y3j>>1) in ih264_iquant_itrans_recon_4x4_ssse3()
273 temp4 = _mm_srai_epi32(temp4, 6); in ih264_iquant_itrans_recon_4x4_ssse3()
278 temp5 = _mm_srai_epi32(temp5, 6); in ih264_iquant_itrans_recon_4x4_ssse3()
[all …]
Dih264_iquant_itrans_recon_sse42.c164 resq_r0 = _mm_srai_epi32(temp4, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42()
165 resq_r1 = _mm_srai_epi32(temp5, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42()
166 resq_r2 = _mm_srai_epi32(temp6, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42()
167 resq_r3 = _mm_srai_epi32(temp7, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42()
198 temp2 = _mm_srai_epi32(resq_r1, 1); //(w1>>1) in ih264_iquant_itrans_recon_4x4_sse42()
201 temp3 = _mm_srai_epi32(resq_r3, 1); //(w3>>1) + w1 in ih264_iquant_itrans_recon_4x4_sse42()
250 temp2 = _mm_srai_epi32(resq_r1, 1); //(y1j>>1) in ih264_iquant_itrans_recon_4x4_sse42()
253 temp3 = _mm_srai_epi32(resq_r3, 1); //(y3j>>1) in ih264_iquant_itrans_recon_4x4_sse42()
259 temp4 = _mm_srai_epi32(temp4, 6); in ih264_iquant_itrans_recon_4x4_sse42()
264 temp5 = _mm_srai_epi32(temp5, 6); in ih264_iquant_itrans_recon_4x4_sse42()
[all …]
Dih264_ihadamard_scaling_sse42.c192 src_r0 = _mm_srai_epi32(temp0, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42()
193 src_r1 = _mm_srai_epi32(temp1, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42()
194 src_r2 = _mm_srai_epi32(temp2, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42()
195 src_r3 = _mm_srai_epi32(temp3, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42()
240 temp0 = _mm_srai_epi32(temp0, 5); in ih264_ihadamard_scaling_2x2_uv_sse42()
241 temp1 = _mm_srai_epi32(temp1, 5); in ih264_ihadamard_scaling_2x2_uv_sse42()
Dih264_ihadamard_scaling_ssse3.c199 src_r0 = _mm_srai_epi32(temp0, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3()
200 src_r1 = _mm_srai_epi32(temp1, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3()
201 src_r2 = _mm_srai_epi32(temp2, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3()
202 src_r3 = _mm_srai_epi32(temp3, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3()
/external/libmpeg2/common/x86/
Dimpeg2_idct_recon_sse42_intr.c300 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42()
302 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42()
330 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42()
331 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42()
360 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42()
361 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42()
386 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42()
387 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42()
512 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42()
514 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42()
[all …]
/external/libhevc/common/x86/
Dihevc_itrans_recon_sse42_intr.c962 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42()
964 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42()
992 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42()
993 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42()
1022 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42()
1023 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42()
1048 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42()
1049 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42()
1174 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42()
1176 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42()
[all …]
Dihevc_weighted_pred_sse42_intr.c212 src_temp0_4x32b = _mm_srai_epi32(src_temp0_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
213 src_temp1_4x32b = _mm_srai_epi32(src_temp1_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
214 src_temp2_4x32b = _mm_srai_epi32(src_temp2_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
215 src_temp3_4x32b = _mm_srai_epi32(src_temp3_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
224 src_temp4_4x32b = _mm_srai_epi32(src_temp4_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
225 src_temp5_4x32b = _mm_srai_epi32(src_temp5_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
226 src_temp6_4x32b = _mm_srai_epi32(src_temp6_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
227 src_temp7_4x32b = _mm_srai_epi32(src_temp7_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
310 src_temp0_4x32b = _mm_srai_epi32(src_temp0_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
311 src_temp1_4x32b = _mm_srai_epi32(src_temp1_4x32b, shift); in ihevc_weighted_pred_uni_sse42()
[all …]
Dihevc_itrans_recon_ssse3_intr.c1141 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1143 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1171 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1172 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1201 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1202 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1227 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1228 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1365 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
1367 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3()
[all …]
Dihevc_weighted_pred_ssse3_intr.c207 res_temp0_4x32b = _mm_srai_epi32(res_temp0_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
208 res_temp1_4x32b = _mm_srai_epi32(res_temp1_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
209 res_temp2_4x32b = _mm_srai_epi32(res_temp2_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
210 res_temp3_4x32b = _mm_srai_epi32(res_temp3_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
213 res_temp4_4x32b = _mm_srai_epi32(res_temp4_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
214 res_temp5_4x32b = _mm_srai_epi32(res_temp5_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
215 res_temp6_4x32b = _mm_srai_epi32(res_temp6_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
216 res_temp7_4x32b = _mm_srai_epi32(res_temp7_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
300 res_temp0_4x32b = _mm_srai_epi32(res_temp0_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
301 res_temp1_4x32b = _mm_srai_epi32(res_temp1_4x32b, shift); in ihevc_weighted_pred_uni_ssse3()
[all …]
/external/libaom/libaom/aom_dsp/x86/
Dfwd_txfm_impl_sse2.h134 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT4x4_2D_HELPER()
135 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT4x4_2D_HELPER()
136 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS); in FDCT4x4_2D_HELPER()
137 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS); in FDCT4x4_2D_HELPER()
179 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER()
180 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER()
181 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER()
182 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER()
333 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT8x8_2D()
334 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT8x8_2D()
[all …]
Dconvolve_sse4_1.h27 d = _mm_srai_epi32(d, DIST_PRECISION_BITS); in mult_add_store()
45 res = _mm_srai_epi32(wt_res, DIST_PRECISION_BITS); in highbd_comp_avg_sse4_1()
48 res = _mm_srai_epi32(wt_res, 1); in highbd_comp_avg_sse4_1()
Dhighbd_convolve_sse2.c76 resReg23_45_lo = _mm_srai_epi32(resReg23_45_lo, 7); in aom_highbd_filter_block1d4_v4_sse2()
77 resReg34_56_lo = _mm_srai_epi32(resReg34_56_lo, 7); in aom_highbd_filter_block1d4_v4_sse2()
142 srcRegFilt32b1_1 = _mm_srai_epi32(srcRegFilt32b1_1, 7); in aom_highbd_filter_block1d4_h4_sse2()
233 resReg23_45_lo = _mm_srai_epi32(resReg23_45_lo, 7); in aom_highbd_filter_block1d8_v4_sse2()
234 resReg34_56_lo = _mm_srai_epi32(resReg34_56_lo, 7); in aom_highbd_filter_block1d8_v4_sse2()
235 resReg23_45_hi = _mm_srai_epi32(resReg23_45_hi, 7); in aom_highbd_filter_block1d8_v4_sse2()
236 resReg34_56_hi = _mm_srai_epi32(resReg34_56_hi, 7); in aom_highbd_filter_block1d8_v4_sse2()
317 res_lo_1 = _mm_srai_epi32(res_lo_1, 7); in aom_highbd_filter_block1d8_h4_sse2()
318 res_hi_1 = _mm_srai_epi32(res_hi_1, 7); in aom_highbd_filter_block1d8_h4_sse2()
Dhighbd_adaptive_quantize_sse2.c26 __m128i sign = _mm_srai_epi32(*y, 31); in highbd_mul_shift_sse2()
88 __m128i coeff_sign = _mm_srai_epi32(qcoeff, 31); in highbd_calculate_dqcoeff()
154 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_adaptive_sse2()
155 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2()
210 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_adaptive_sse2()
211 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2()
371 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2()
372 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2()
427 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2()
428 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2()
[all …]
Dobmc_intrinsic_ssse3.h48 const __m128i v_sign_d = _mm_srai_epi32(v_val_d, 31); in xx_roundn_epi32()
51 return _mm_srai_epi32(v_tmp_d, bits); in xx_roundn_epi32()
Dconvolve_sse2.h90 const __m128i res_lo = _mm_srai_epi32(wt_res_lo, DIST_PRECISION_BITS); in comp_avg()
91 const __m128i res_hi = _mm_srai_epi32(wt_res_hi, DIST_PRECISION_BITS); in comp_avg()
116 _mm_srai_epi32(_mm_add_epi32(res_signed, *round_const), round_shift); in highbd_convolve_rounding_sse2()
/external/libaom/libaom/av1/encoder/x86/
Dhighbd_fwd_txfm_sse4.c89 u0 = _mm_srai_epi32(u3, bit); in fdct4x4_sse4_1()
90 u2 = _mm_srai_epi32(v1, bit); in fdct4x4_sse4_1()
98 u1 = _mm_srai_epi32(v3, bit); in fdct4x4_sse4_1()
105 u3 = _mm_srai_epi32(v3, bit); in fdct4x4_sse4_1()
170 u0 = _mm_srai_epi32(u0, bit); in fadst4x4_sse4_1()
173 u1 = _mm_srai_epi32(u1, bit); in fadst4x4_sse4_1()
176 u2 = _mm_srai_epi32(u2, bit); in fadst4x4_sse4_1()
179 u3 = _mm_srai_epi32(u3, bit); in fadst4x4_sse4_1()
201 out[i] = _mm_srai_epi32(a_low, NewSqrt2Bits); in idtx4x4_sse4_1()
430 in[0] = _mm_srai_epi32(in[0], shift); in col_txfm_8x8_rounding()
[all …]
/external/libaom/libaom/av1/common/x86/
Dav1_txfm_sse2.h37 const __m128i c0 = _mm_srai_epi32(a0, cos_bit); in btf_16_w4_sse2()
38 const __m128i d0 = _mm_srai_epi32(b0, cos_bit); in btf_16_w4_sse2()
53 __m128i c0 = _mm_srai_epi32(a0, cos_bit); \
54 __m128i d0 = _mm_srai_epi32(b0, cos_bit); \
74 __m128i c0 = _mm_srai_epi32(a0, cos_bit); \
75 __m128i c1 = _mm_srai_epi32(a1, cos_bit); \
76 __m128i d0 = _mm_srai_epi32(b0, cos_bit); \
77 __m128i d1 = _mm_srai_epi32(b1, cos_bit); \
100 const __m128i a_1 = _mm_srai_epi32(a_lo, 16); in store_16bit_to_32bit_w4()
108 const __m128i a_1 = _mm_srai_epi32(a_lo, 16); in store_16bit_to_32bit()
[all …]
Dhighbd_inv_txfm_sse4.c48 in[0] = _mm_srai_epi32(in[0], shift); in round_shift_4x4()
49 in[1] = _mm_srai_epi32(in[1], shift); in round_shift_4x4()
50 in[2] = _mm_srai_epi32(in[2], shift); in round_shift_4x4()
51 in[3] = _mm_srai_epi32(in[3], shift); in round_shift_4x4()
417 v0 = _mm_srai_epi32(v0, bit); in idct4x4_sse4_1()
421 v1 = _mm_srai_epi32(v1, bit); in idct4x4_sse4_1()
427 v2 = _mm_srai_epi32(v2, bit); in idct4x4_sse4_1()
433 v3 = _mm_srai_epi32(v3, bit); in idct4x4_sse4_1()
851 u4 = _mm_srai_epi32(u4, bit); in idct8x8_sse4_1()
857 u7 = _mm_srai_epi32(u7, bit); in idct8x8_sse4_1()
[all …]
Dhighbd_wiener_convolve_ssse3.c86 res_even = _mm_srai_epi32(_mm_add_epi32(res_even, round_const), in av1_highbd_wiener_convolve_add_src_ssse3()
101 res_odd = _mm_srai_epi32(_mm_add_epi32(res_odd, round_const), in av1_highbd_wiener_convolve_add_src_ssse3()
188 const __m128i res_lo_round = _mm_srai_epi32( in av1_highbd_wiener_convolve_add_src_ssse3()
190 const __m128i res_hi_round = _mm_srai_epi32( in av1_highbd_wiener_convolve_add_src_ssse3()
Dwiener_convolve_sse2.c84 res_even = _mm_srai_epi32(_mm_add_epi32(res_even, round_const), in av1_wiener_convolve_add_src_sse2()
99 res_odd = _mm_srai_epi32(_mm_add_epi32(res_odd, round_const), in av1_wiener_convolve_add_src_sse2()
186 const __m128i res_lo_round = _mm_srai_epi32( in av1_wiener_convolve_add_src_sse2()
188 const __m128i res_hi_round = _mm_srai_epi32( in av1_wiener_convolve_add_src_sse2()
/external/flac/src/libFLAC/
Dfixed_intrin_sse2.c102 tmp = _mm_srai_epi32(err0, 31); in FLAC__fixed_compute_best_predictor_intrin_sse2()
105 tmp = _mm_srai_epi32(err1, 31); in FLAC__fixed_compute_best_predictor_intrin_sse2()
199 tmp = _mm_srai_epi32(err0, 31); in FLAC__fixed_compute_best_predictor_wide_intrin_sse2()
202 tmp = _mm_srai_epi32(err1, 31); in FLAC__fixed_compute_best_predictor_wide_intrin_sse2()

12345