/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_dct_intrin_sse2.c | 98 u[0] = _mm_srai_epi32(v[0], DCT_CONST_BITS); in fdct4_sse2() 99 u[1] = _mm_srai_epi32(v[1], DCT_CONST_BITS); in fdct4_sse2() 100 u[2] = _mm_srai_epi32(v[2], DCT_CONST_BITS); in fdct4_sse2() 101 u[3] = _mm_srai_epi32(v[3], DCT_CONST_BITS); in fdct4_sse2() 146 u[0] = _mm_srai_epi32(v[0], DCT_CONST_BITS); in fadst4_sse2() 147 u[1] = _mm_srai_epi32(v[1], DCT_CONST_BITS); in fadst4_sse2() 148 u[2] = _mm_srai_epi32(v[2], DCT_CONST_BITS); in fadst4_sse2() 149 u[3] = _mm_srai_epi32(v[3], DCT_CONST_BITS); in fadst4_sse2() 326 u0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in fdct8_sse2() 327 u1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in fdct8_sse2() [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_dct32x32_impl_sse2.h | 438 const __m128i s2_20_6 = _mm_srai_epi32(s2_20_4, DCT_CONST_BITS); in FDCT32x32_2D() 439 const __m128i s2_20_7 = _mm_srai_epi32(s2_20_5, DCT_CONST_BITS); in FDCT32x32_2D() 440 const __m128i s2_21_6 = _mm_srai_epi32(s2_21_4, DCT_CONST_BITS); in FDCT32x32_2D() 441 const __m128i s2_21_7 = _mm_srai_epi32(s2_21_5, DCT_CONST_BITS); in FDCT32x32_2D() 442 const __m128i s2_22_6 = _mm_srai_epi32(s2_22_4, DCT_CONST_BITS); in FDCT32x32_2D() 443 const __m128i s2_22_7 = _mm_srai_epi32(s2_22_5, DCT_CONST_BITS); in FDCT32x32_2D() 444 const __m128i s2_23_6 = _mm_srai_epi32(s2_23_4, DCT_CONST_BITS); in FDCT32x32_2D() 445 const __m128i s2_23_7 = _mm_srai_epi32(s2_23_5, DCT_CONST_BITS); in FDCT32x32_2D() 446 const __m128i s2_24_6 = _mm_srai_epi32(s2_24_4, DCT_CONST_BITS); in FDCT32x32_2D() 447 const __m128i s2_24_7 = _mm_srai_epi32(s2_24_5, DCT_CONST_BITS); in FDCT32x32_2D() [all …]
|
D | fwd_txfm_sse2.c | 32 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct4x4_1_sse2() 33 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct4x4_1_sse2() 73 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct8x8_1_sse2() 74 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct8x8_1_sse2() 141 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct16x16_1_sse2() 142 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct16x16_1_sse2() 152 in1 = _mm_srai_epi32(in1, 1); in vpx_fdct16x16_1_sse2() 213 in0 = _mm_srai_epi32(in0, 16); in vpx_fdct32x32_1_sse2() 214 in1 = _mm_srai_epi32(in1, 16); in vpx_fdct32x32_1_sse2() 224 in1 = _mm_srai_epi32(in1, 3); in vpx_fdct32x32_1_sse2()
|
D | fwd_txfm_impl_sse2.h | 159 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT4x4_2D() 160 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT4x4_2D() 161 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS); in FDCT4x4_2D() 162 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS); in FDCT4x4_2D() 219 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS2); in FDCT4x4_2D() 220 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS2); in FDCT4x4_2D() 221 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS2); in FDCT4x4_2D() 222 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS2); in FDCT4x4_2D() 358 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT8x8_2D() 359 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT8x8_2D() [all …]
|
/external/libavc/common/x86/ |
D | ih264_iquant_itrans_recon_ssse3.c | 165 resq_r0 = _mm_srai_epi32(temp4, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3() 166 resq_r1 = _mm_srai_epi32(temp5, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3() 167 resq_r2 = _mm_srai_epi32(temp6, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3() 168 resq_r3 = _mm_srai_epi32(temp7, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_ssse3() 205 temp2 = _mm_srai_epi32(resq_r1, 1); //(w1>>1) in ih264_iquant_itrans_recon_4x4_ssse3() 208 temp3 = _mm_srai_epi32(resq_r3, 1); //(w3>>1) + w1 in ih264_iquant_itrans_recon_4x4_ssse3() 264 temp2 = _mm_srai_epi32(resq_r1, 1); //(y1j>>1) in ih264_iquant_itrans_recon_4x4_ssse3() 267 temp3 = _mm_srai_epi32(resq_r3, 1); //(y3j>>1) in ih264_iquant_itrans_recon_4x4_ssse3() 273 temp4 = _mm_srai_epi32(temp4, 6); in ih264_iquant_itrans_recon_4x4_ssse3() 278 temp5 = _mm_srai_epi32(temp5, 6); in ih264_iquant_itrans_recon_4x4_ssse3() [all …]
|
D | ih264_iquant_itrans_recon_sse42.c | 164 resq_r0 = _mm_srai_epi32(temp4, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42() 165 resq_r1 = _mm_srai_epi32(temp5, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42() 166 resq_r2 = _mm_srai_epi32(temp6, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42() 167 resq_r3 = _mm_srai_epi32(temp7, 4 - u4_qp_div_6); in ih264_iquant_itrans_recon_4x4_sse42() 198 temp2 = _mm_srai_epi32(resq_r1, 1); //(w1>>1) in ih264_iquant_itrans_recon_4x4_sse42() 201 temp3 = _mm_srai_epi32(resq_r3, 1); //(w3>>1) + w1 in ih264_iquant_itrans_recon_4x4_sse42() 250 temp2 = _mm_srai_epi32(resq_r1, 1); //(y1j>>1) in ih264_iquant_itrans_recon_4x4_sse42() 253 temp3 = _mm_srai_epi32(resq_r3, 1); //(y3j>>1) in ih264_iquant_itrans_recon_4x4_sse42() 259 temp4 = _mm_srai_epi32(temp4, 6); in ih264_iquant_itrans_recon_4x4_sse42() 264 temp5 = _mm_srai_epi32(temp5, 6); in ih264_iquant_itrans_recon_4x4_sse42() [all …]
|
D | ih264_ihadamard_scaling_sse42.c | 192 src_r0 = _mm_srai_epi32(temp0, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42() 193 src_r1 = _mm_srai_epi32(temp1, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42() 194 src_r2 = _mm_srai_epi32(temp2, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42() 195 src_r3 = _mm_srai_epi32(temp3, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_sse42() 240 temp0 = _mm_srai_epi32(temp0, 5); in ih264_ihadamard_scaling_2x2_uv_sse42() 241 temp1 = _mm_srai_epi32(temp1, 5); in ih264_ihadamard_scaling_2x2_uv_sse42()
|
D | ih264_ihadamard_scaling_ssse3.c | 199 src_r0 = _mm_srai_epi32(temp0, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3() 200 src_r1 = _mm_srai_epi32(temp1, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3() 201 src_r2 = _mm_srai_epi32(temp2, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3() 202 src_r3 = _mm_srai_epi32(temp3, 6 - u4_qp_div_6); in ih264_ihadamard_scaling_4x4_ssse3()
|
/external/libmpeg2/common/x86/ |
D | impeg2_idct_recon_sse42_intr.c | 300 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42() 302 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42() 330 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42() 331 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42() 360 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42() 361 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42() 386 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42() 387 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42() 512 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in impeg2_idct_recon_sse42() 514 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in impeg2_idct_recon_sse42() [all …]
|
/external/libhevc/common/x86/ |
D | ihevc_itrans_recon_sse42_intr.c | 962 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42() 964 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42() 992 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42() 993 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42() 1022 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42() 1023 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42() 1048 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42() 1049 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42() 1174 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_sse42() 1176 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_sse42() [all …]
|
D | ihevc_weighted_pred_sse42_intr.c | 212 src_temp0_4x32b = _mm_srai_epi32(src_temp0_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 213 src_temp1_4x32b = _mm_srai_epi32(src_temp1_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 214 src_temp2_4x32b = _mm_srai_epi32(src_temp2_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 215 src_temp3_4x32b = _mm_srai_epi32(src_temp3_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 224 src_temp4_4x32b = _mm_srai_epi32(src_temp4_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 225 src_temp5_4x32b = _mm_srai_epi32(src_temp5_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 226 src_temp6_4x32b = _mm_srai_epi32(src_temp6_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 227 src_temp7_4x32b = _mm_srai_epi32(src_temp7_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 310 src_temp0_4x32b = _mm_srai_epi32(src_temp0_4x32b, shift); in ihevc_weighted_pred_uni_sse42() 311 src_temp1_4x32b = _mm_srai_epi32(src_temp1_4x32b, shift); in ihevc_weighted_pred_uni_sse42() [all …]
|
D | ihevc_itrans_recon_ssse3_intr.c | 1141 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1143 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1171 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1172 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1201 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1202 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1227 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1228 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1365 m_temp_reg_62 = _mm_srai_epi32(m_temp_reg_62, i4_shift); in ihevc_itrans_recon_8x8_ssse3() 1367 m_temp_reg_66 = _mm_srai_epi32(m_temp_reg_66, i4_shift); in ihevc_itrans_recon_8x8_ssse3() [all …]
|
D | ihevc_weighted_pred_ssse3_intr.c | 207 res_temp0_4x32b = _mm_srai_epi32(res_temp0_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 208 res_temp1_4x32b = _mm_srai_epi32(res_temp1_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 209 res_temp2_4x32b = _mm_srai_epi32(res_temp2_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 210 res_temp3_4x32b = _mm_srai_epi32(res_temp3_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 213 res_temp4_4x32b = _mm_srai_epi32(res_temp4_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 214 res_temp5_4x32b = _mm_srai_epi32(res_temp5_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 215 res_temp6_4x32b = _mm_srai_epi32(res_temp6_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 216 res_temp7_4x32b = _mm_srai_epi32(res_temp7_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 300 res_temp0_4x32b = _mm_srai_epi32(res_temp0_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() 301 res_temp1_4x32b = _mm_srai_epi32(res_temp1_4x32b, shift); in ihevc_weighted_pred_uni_ssse3() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | fwd_txfm_impl_sse2.h | 134 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT4x4_2D_HELPER() 135 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT4x4_2D_HELPER() 136 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS); in FDCT4x4_2D_HELPER() 137 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS); in FDCT4x4_2D_HELPER() 179 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER() 180 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER() 181 const __m128i w2 = _mm_srai_epi32(v2, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER() 182 const __m128i w3 = _mm_srai_epi32(v3, DCT_CONST_BITS2); in FDCT4x4_2D_HELPER() 333 const __m128i w0 = _mm_srai_epi32(v0, DCT_CONST_BITS); in FDCT8x8_2D() 334 const __m128i w1 = _mm_srai_epi32(v1, DCT_CONST_BITS); in FDCT8x8_2D() [all …]
|
D | convolve_sse4_1.h | 27 d = _mm_srai_epi32(d, DIST_PRECISION_BITS); in mult_add_store() 45 res = _mm_srai_epi32(wt_res, DIST_PRECISION_BITS); in highbd_comp_avg_sse4_1() 48 res = _mm_srai_epi32(wt_res, 1); in highbd_comp_avg_sse4_1()
|
D | highbd_convolve_sse2.c | 76 resReg23_45_lo = _mm_srai_epi32(resReg23_45_lo, 7); in aom_highbd_filter_block1d4_v4_sse2() 77 resReg34_56_lo = _mm_srai_epi32(resReg34_56_lo, 7); in aom_highbd_filter_block1d4_v4_sse2() 142 srcRegFilt32b1_1 = _mm_srai_epi32(srcRegFilt32b1_1, 7); in aom_highbd_filter_block1d4_h4_sse2() 233 resReg23_45_lo = _mm_srai_epi32(resReg23_45_lo, 7); in aom_highbd_filter_block1d8_v4_sse2() 234 resReg34_56_lo = _mm_srai_epi32(resReg34_56_lo, 7); in aom_highbd_filter_block1d8_v4_sse2() 235 resReg23_45_hi = _mm_srai_epi32(resReg23_45_hi, 7); in aom_highbd_filter_block1d8_v4_sse2() 236 resReg34_56_hi = _mm_srai_epi32(resReg34_56_hi, 7); in aom_highbd_filter_block1d8_v4_sse2() 317 res_lo_1 = _mm_srai_epi32(res_lo_1, 7); in aom_highbd_filter_block1d8_h4_sse2() 318 res_hi_1 = _mm_srai_epi32(res_hi_1, 7); in aom_highbd_filter_block1d8_h4_sse2()
|
D | highbd_adaptive_quantize_sse2.c | 26 __m128i sign = _mm_srai_epi32(*y, 31); in highbd_mul_shift_sse2() 88 __m128i coeff_sign = _mm_srai_epi32(qcoeff, 31); in highbd_calculate_dqcoeff() 154 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_adaptive_sse2() 155 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2() 210 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_adaptive_sse2() 211 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2() 371 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2() 372 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2() 427 coeff0_sign = _mm_srai_epi32(coeff0, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2() 428 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_32x32_adaptive_sse2() [all …]
|
D | obmc_intrinsic_ssse3.h | 48 const __m128i v_sign_d = _mm_srai_epi32(v_val_d, 31); in xx_roundn_epi32() 51 return _mm_srai_epi32(v_tmp_d, bits); in xx_roundn_epi32()
|
D | convolve_sse2.h | 90 const __m128i res_lo = _mm_srai_epi32(wt_res_lo, DIST_PRECISION_BITS); in comp_avg() 91 const __m128i res_hi = _mm_srai_epi32(wt_res_hi, DIST_PRECISION_BITS); in comp_avg() 116 _mm_srai_epi32(_mm_add_epi32(res_signed, *round_const), round_shift); in highbd_convolve_rounding_sse2()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | highbd_fwd_txfm_sse4.c | 89 u0 = _mm_srai_epi32(u3, bit); in fdct4x4_sse4_1() 90 u2 = _mm_srai_epi32(v1, bit); in fdct4x4_sse4_1() 98 u1 = _mm_srai_epi32(v3, bit); in fdct4x4_sse4_1() 105 u3 = _mm_srai_epi32(v3, bit); in fdct4x4_sse4_1() 170 u0 = _mm_srai_epi32(u0, bit); in fadst4x4_sse4_1() 173 u1 = _mm_srai_epi32(u1, bit); in fadst4x4_sse4_1() 176 u2 = _mm_srai_epi32(u2, bit); in fadst4x4_sse4_1() 179 u3 = _mm_srai_epi32(u3, bit); in fadst4x4_sse4_1() 201 out[i] = _mm_srai_epi32(a_low, NewSqrt2Bits); in idtx4x4_sse4_1() 430 in[0] = _mm_srai_epi32(in[0], shift); in col_txfm_8x8_rounding() [all …]
|
/external/libaom/libaom/av1/common/x86/ |
D | av1_txfm_sse2.h | 37 const __m128i c0 = _mm_srai_epi32(a0, cos_bit); in btf_16_w4_sse2() 38 const __m128i d0 = _mm_srai_epi32(b0, cos_bit); in btf_16_w4_sse2() 53 __m128i c0 = _mm_srai_epi32(a0, cos_bit); \ 54 __m128i d0 = _mm_srai_epi32(b0, cos_bit); \ 74 __m128i c0 = _mm_srai_epi32(a0, cos_bit); \ 75 __m128i c1 = _mm_srai_epi32(a1, cos_bit); \ 76 __m128i d0 = _mm_srai_epi32(b0, cos_bit); \ 77 __m128i d1 = _mm_srai_epi32(b1, cos_bit); \ 100 const __m128i a_1 = _mm_srai_epi32(a_lo, 16); in store_16bit_to_32bit_w4() 108 const __m128i a_1 = _mm_srai_epi32(a_lo, 16); in store_16bit_to_32bit() [all …]
|
D | highbd_inv_txfm_sse4.c | 48 in[0] = _mm_srai_epi32(in[0], shift); in round_shift_4x4() 49 in[1] = _mm_srai_epi32(in[1], shift); in round_shift_4x4() 50 in[2] = _mm_srai_epi32(in[2], shift); in round_shift_4x4() 51 in[3] = _mm_srai_epi32(in[3], shift); in round_shift_4x4() 417 v0 = _mm_srai_epi32(v0, bit); in idct4x4_sse4_1() 421 v1 = _mm_srai_epi32(v1, bit); in idct4x4_sse4_1() 427 v2 = _mm_srai_epi32(v2, bit); in idct4x4_sse4_1() 433 v3 = _mm_srai_epi32(v3, bit); in idct4x4_sse4_1() 851 u4 = _mm_srai_epi32(u4, bit); in idct8x8_sse4_1() 857 u7 = _mm_srai_epi32(u7, bit); in idct8x8_sse4_1() [all …]
|
D | highbd_wiener_convolve_ssse3.c | 86 res_even = _mm_srai_epi32(_mm_add_epi32(res_even, round_const), in av1_highbd_wiener_convolve_add_src_ssse3() 101 res_odd = _mm_srai_epi32(_mm_add_epi32(res_odd, round_const), in av1_highbd_wiener_convolve_add_src_ssse3() 188 const __m128i res_lo_round = _mm_srai_epi32( in av1_highbd_wiener_convolve_add_src_ssse3() 190 const __m128i res_hi_round = _mm_srai_epi32( in av1_highbd_wiener_convolve_add_src_ssse3()
|
D | wiener_convolve_sse2.c | 84 res_even = _mm_srai_epi32(_mm_add_epi32(res_even, round_const), in av1_wiener_convolve_add_src_sse2() 99 res_odd = _mm_srai_epi32(_mm_add_epi32(res_odd, round_const), in av1_wiener_convolve_add_src_sse2() 186 const __m128i res_lo_round = _mm_srai_epi32( in av1_wiener_convolve_add_src_sse2() 188 const __m128i res_hi_round = _mm_srai_epi32( in av1_wiener_convolve_add_src_sse2()
|
/external/flac/src/libFLAC/ |
D | fixed_intrin_sse2.c | 102 tmp = _mm_srai_epi32(err0, 31); in FLAC__fixed_compute_best_predictor_intrin_sse2() 105 tmp = _mm_srai_epi32(err1, 31); in FLAC__fixed_compute_best_predictor_intrin_sse2() 199 tmp = _mm_srai_epi32(err0, 31); in FLAC__fixed_compute_best_predictor_wide_intrin_sse2() 202 tmp = _mm_srai_epi32(err1, 31); in FLAC__fixed_compute_best_predictor_wide_intrin_sse2()
|