/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_idct_intrin_sse2.c | 46 in[0] = _mm_srai_epi16(in[0], 4); in vp9_iht4x4_16_add_sse2() 47 in[1] = _mm_srai_epi16(in[1], 4); in vp9_iht4x4_16_add_sse2() 97 in[0] = _mm_srai_epi16(in[0], 5); in vp9_iht8x8_64_add_sse2() 98 in[1] = _mm_srai_epi16(in[1], 5); in vp9_iht8x8_64_add_sse2() 99 in[2] = _mm_srai_epi16(in[2], 5); in vp9_iht8x8_64_add_sse2() 100 in[3] = _mm_srai_epi16(in[3], 5); in vp9_iht8x8_64_add_sse2() 101 in[4] = _mm_srai_epi16(in[4], 5); in vp9_iht8x8_64_add_sse2() 102 in[5] = _mm_srai_epi16(in[5], 5); in vp9_iht8x8_64_add_sse2() 103 in[6] = _mm_srai_epi16(in[6], 5); in vp9_iht8x8_64_add_sse2() 104 in[7] = _mm_srai_epi16(in[7], 5); in vp9_iht8x8_64_add_sse2() [all …]
|
/external/libavc/common/x86/ |
D | ih264_chroma_intra_pred_filters_ssse3.c | 345 res1_sh_l_8x16b = _mm_srai_epi16(res1_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 346 res1_sh_h_8x16b = _mm_srai_epi16(res1_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 347 res2_sh_l_8x16b = _mm_srai_epi16(res2_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 348 res2_sh_h_8x16b = _mm_srai_epi16(res2_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 365 res1_sh_l_8x16b = _mm_srai_epi16(res1_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 366 res1_sh_h_8x16b = _mm_srai_epi16(res1_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 367 res2_sh_l_8x16b = _mm_srai_epi16(res2_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 368 res2_sh_h_8x16b = _mm_srai_epi16(res2_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 384 res1_sh_l_8x16b = _mm_srai_epi16(res1_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() 385 res1_sh_h_8x16b = _mm_srai_epi16(res1_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3() [all …]
|
D | ih264_weighted_pred_sse42.c | 425 y_0_8x16b = _mm_srai_epi16(y_0_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 426 y_2_8x16b = _mm_srai_epi16(y_2_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 473 y_0_8x16b = _mm_srai_epi16(y_0_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 474 y_1_8x16b = _mm_srai_epi16(y_1_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 475 y_2_8x16b = _mm_srai_epi16(y_2_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 476 y_3_8x16b = _mm_srai_epi16(y_3_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 541 y_0L_8x16b = _mm_srai_epi16(y_0L_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 542 y_0H_8x16b = _mm_srai_epi16(y_0H_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 543 y_1L_8x16b = _mm_srai_epi16(y_1L_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() 544 y_1H_8x16b = _mm_srai_epi16(y_1H_8x16b, log_wd); in ih264_weighted_pred_luma_sse42() [all …]
|
D | ih264_luma_intra_pred_filters_ssse3.c | 351 res1_8x16b = _mm_srai_epi16(res1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_diag_dl_ssse3() 437 res1_8x16b = _mm_srai_epi16(res1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_diag_dr_ssse3() 526 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_vert_r_ssse3() 620 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_horz_d_ssse3() 714 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_vert_l_ssse3() 809 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_horz_u_ssse3() 1113 res1_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dl_ssse3() 1126 res2_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dl_ssse3() 1214 res1_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dr_ssse3() 1228 res2_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dr_ssse3() [all …]
|
D | ih264_deblk_luma_ssse3.c | 218 Alpha_8x16 = _mm_srai_epi16(Alpha_8x16, 2); in ih264_deblk_luma_vert_bs4_ssse3() 274 p0_16x8_1 = _mm_srai_epi16(temp1, 2); in ih264_deblk_luma_vert_bs4_ssse3() 275 q0_16x8_1 = _mm_srai_epi16(temp2, 2); in ih264_deblk_luma_vert_bs4_ssse3() 282 p1_16x8_2 = _mm_srai_epi16(temp1, 2); in ih264_deblk_luma_vert_bs4_ssse3() 283 q1_16x8_2 = _mm_srai_epi16(temp2, 2); in ih264_deblk_luma_vert_bs4_ssse3() 296 p0_16x8_2 = _mm_srai_epi16(temp1, 3); in ih264_deblk_luma_vert_bs4_ssse3() 297 q0_16x8_2 = _mm_srai_epi16(temp2, 3); in ih264_deblk_luma_vert_bs4_ssse3() 312 p2_16x8_2 = _mm_srai_epi16(temp1, 3); in ih264_deblk_luma_vert_bs4_ssse3() 313 q2_16x8_2 = _mm_srai_epi16(temp2, 3); in ih264_deblk_luma_vert_bs4_ssse3() 334 temp1 = _mm_srai_epi16(temp1, 2); in ih264_deblk_luma_vert_bs4_ssse3() [all …]
|
/external/libhevc/common/x86/ |
D | ihevc_intra_pred_filters_sse42_intr.c | 252 src_temp1 = _mm_srai_epi16(src_temp1, 2); in ihevc_intra_pred_ref_filtering_sse42() 268 src_temp4 = _mm_srai_epi16(src_temp4, 2); in ihevc_intra_pred_ref_filtering_sse42() 1038 src_temp3 = _mm_srai_epi16(src_temp3, 1); in ihevc_intra_pred_luma_horz_sse42() 1096 src_temp10 = _mm_srai_epi16(src_temp10, 1); in ihevc_intra_pred_luma_horz_sse42() 1509 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42() 1510 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42() 1511 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42() 1512 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42() 1590 ref_main_idx_4x32b = _mm_sub_epi16(two_nt_4x32b, _mm_srai_epi16(res_temp5_4x32b, 5)); in ihevc_intra_pred_luma_mode_3_to_9_sse42() 1654 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42() [all …]
|
D | ihevc_intra_pred_filters_ssse3_intr.c | 537 src_temp1 = _mm_srai_epi16(src_temp1, 2); in ihevc_intra_pred_ref_filtering_ssse3() 553 src_temp4 = _mm_srai_epi16(src_temp4, 2); in ihevc_intra_pred_ref_filtering_ssse3() 1547 src_temp3 = _mm_srai_epi16(src_temp3, 1); in ihevc_intra_pred_luma_horz_ssse3() 1606 src_temp10 = _mm_srai_epi16(src_temp10, 1); in ihevc_intra_pred_luma_horz_ssse3() 2427 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3() 2428 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3() 2429 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3() 2430 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3() 2508 ref_main_idx_4x32b = _mm_sub_epi16(two_nt_4x32b, _mm_srai_epi16(res_temp5_4x32b, 5)); in ihevc_intra_pred_luma_mode_3_to_9_ssse3() 2572 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3() [all …]
|
D | ihevc_chroma_intra_pred_filters_ssse3_intr.c | 1210 ref_main_idx_4x32b = _mm_srai_epi16(res_temp5_4x32b, 5); in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1277 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1278 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1279 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1280 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1338 ref_main_idx_4x32b = _mm_srai_epi16(res_temp5_4x32b, 5); in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1451 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1452 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1453 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() 1454 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3() [all …]
|
D | ihevc_weighted_pred_ssse3_intr.c | 1385 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1386 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1387 src_temp5_8x16b = _mm_srai_epi16(src_temp5_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1388 src_temp7_8x16b = _mm_srai_epi16(src_temp7_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1397 src_temp9_8x16b = _mm_srai_epi16(src_temp9_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1398 src_temp11_8x16b = _mm_srai_epi16(src_temp11_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1399 src_temp13_8x16b = _mm_srai_epi16(src_temp13_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1400 src_temp15_8x16b = _mm_srai_epi16(src_temp15_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1460 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() 1461 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | aom_subpixel_8t_intrin_ssse3.c | 77 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d4_h4_ssse3() 99 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d4_h4_ssse3() 130 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d4_v4_ssse3() 173 resReglo = _mm_srai_epi16(resReglo, 6); in aom_filter_block1d4_v4_ssse3() 174 resReghi = _mm_srai_epi16(resReghi, 6); in aom_filter_block1d4_v4_ssse3() 250 srcRegFilt1 = _mm_srai_epi16(srcRegFilt1, 7); in aom_filter_block1d4_h8_intrin_ssse3() 275 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d8_h4_ssse3() 305 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d8_h4_ssse3() 334 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d8_v4_ssse3() 379 resReg23_45 = _mm_srai_epi16(resReg23_45, 6); in aom_filter_block1d8_v4_ssse3() [all …]
|
D | aom_subpixel_8t_intrin_sse2.c | 33 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d16_h4_sse2() 93 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d16_h4_sse2() 94 srcRegFilt32b2_1 = _mm_srai_epi16(srcRegFilt32b2_1, 6); in aom_filter_block1d16_h4_sse2() 127 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d16_v4_sse2() 226 resReg23_45_lo = _mm_srai_epi16(resReg23_45_lo, 6); in aom_filter_block1d16_v4_sse2() 227 resReg34_56_lo = _mm_srai_epi16(resReg34_56_lo, 6); in aom_filter_block1d16_v4_sse2() 228 resReg23_45_hi = _mm_srai_epi16(resReg23_45_hi, 6); in aom_filter_block1d16_v4_sse2() 229 resReg34_56_hi = _mm_srai_epi16(resReg34_56_hi, 6); in aom_filter_block1d16_v4_sse2() 271 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d8_h4_sse2() 306 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d8_h4_sse2() [all …]
|
D | fwd_txfm_impl_sse2.h | 507 const __m128i sign_in0 = _mm_srai_epi16(in0, 15); in FDCT8x8_2D() 508 const __m128i sign_in1 = _mm_srai_epi16(in1, 15); in FDCT8x8_2D() 509 const __m128i sign_in2 = _mm_srai_epi16(in2, 15); in FDCT8x8_2D() 510 const __m128i sign_in3 = _mm_srai_epi16(in3, 15); in FDCT8x8_2D() 511 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D() 512 const __m128i sign_in5 = _mm_srai_epi16(in5, 15); in FDCT8x8_2D() 513 const __m128i sign_in6 = _mm_srai_epi16(in6, 15); in FDCT8x8_2D() 514 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in FDCT8x8_2D() 523 in0 = _mm_srai_epi16(in0, 1); in FDCT8x8_2D() 524 in1 = _mm_srai_epi16(in1, 1); in FDCT8x8_2D() [all …]
|
D | avg_intrin_sse2.c | 336 b0 = _mm_srai_epi16(b0, 1); in hadamard_16x16_sse2() 337 b1 = _mm_srai_epi16(b1, 1); in hadamard_16x16_sse2() 338 b2 = _mm_srai_epi16(b2, 1); in hadamard_16x16_sse2() 339 b3 = _mm_srai_epi16(b3, 1); in hadamard_16x16_sse2() 396 b0 = _mm_srai_epi16(b0, 2); in aom_hadamard_32x32_sse2() 397 b1 = _mm_srai_epi16(b1, 2); in aom_hadamard_32x32_sse2() 398 b2 = _mm_srai_epi16(b2, 2); in aom_hadamard_32x32_sse2() 399 b3 = _mm_srai_epi16(b3, 2); in aom_hadamard_32x32_sse2() 475 s0 = _mm_srai_epi16(s0, 6); in aom_int_pro_row_sse2() 476 s1 = _mm_srai_epi16(s1, 6); in aom_int_pro_row_sse2() [all …]
|
D | quantize_sse2.c | 48 coeff0_sign = _mm_srai_epi16(coeff0, 15); in aom_quantize_b_sse2() 49 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2() 91 coeff0_sign = _mm_srai_epi16(coeff0, 15); in aom_quantize_b_sse2() 92 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2()
|
D | intrapred_sse2.c | 318 sum_above = _mm_srai_epi16(sum_above, 2); in aom_dc_top_predictor_4x8_sse2() 332 sum_above = _mm_srai_epi16(sum_above, 2); in aom_dc_top_predictor_4x16_sse2() 346 sum_above = _mm_srai_epi16(sum_above, 3); in aom_dc_top_predictor_8x4_sse2() 358 sum_above = _mm_srai_epi16(sum_above, 3); in aom_dc_top_predictor_8x16_sse2() 370 sum_above = _mm_srai_epi16(sum_above, 3); in aom_dc_top_predictor_8x32_sse2() 382 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x4_sse2() 395 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x8_sse2() 409 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x32_sse2() 423 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x64_sse2() 436 sum_above = _mm_srai_epi16(sum_above, 5); in aom_dc_top_predictor_32x8_sse2() [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_txfm_impl_sse2.h | 532 const __m128i sign_in0 = _mm_srai_epi16(in0, 15); in FDCT8x8_2D() 533 const __m128i sign_in1 = _mm_srai_epi16(in1, 15); in FDCT8x8_2D() 534 const __m128i sign_in2 = _mm_srai_epi16(in2, 15); in FDCT8x8_2D() 535 const __m128i sign_in3 = _mm_srai_epi16(in3, 15); in FDCT8x8_2D() 536 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D() 537 const __m128i sign_in5 = _mm_srai_epi16(in5, 15); in FDCT8x8_2D() 538 const __m128i sign_in6 = _mm_srai_epi16(in6, 15); in FDCT8x8_2D() 539 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in FDCT8x8_2D() 548 in0 = _mm_srai_epi16(in0, 1); in FDCT8x8_2D() 549 in1 = _mm_srai_epi16(in1, 1); in FDCT8x8_2D() [all …]
|
D | quantize_sse2.c | 49 coeff0_sign = _mm_srai_epi16(coeff0, 15); in vpx_quantize_b_sse2() 50 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2() 88 coeff0_sign = _mm_srai_epi16(coeff0, 15); in vpx_quantize_b_sse2() 89 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2()
|
D | avg_intrin_sse2.c | 357 b0 = _mm_srai_epi16(b0, 1); in hadamard_16x16_sse2() 358 b1 = _mm_srai_epi16(b1, 1); in hadamard_16x16_sse2() 359 b2 = _mm_srai_epi16(b2, 1); in hadamard_16x16_sse2() 360 b3 = _mm_srai_epi16(b3, 1); in hadamard_16x16_sse2() 421 b0 = _mm_srai_epi16(b0, 2); in vpx_hadamard_32x32_sse2() 422 b1 = _mm_srai_epi16(b1, 2); in vpx_hadamard_32x32_sse2() 423 b2 = _mm_srai_epi16(b2, 2); in vpx_hadamard_32x32_sse2() 424 b3 = _mm_srai_epi16(b3, 2); in vpx_hadamard_32x32_sse2() 501 s0 = _mm_srai_epi16(s0, 5); in vpx_int_pro_row_sse2() 502 s1 = _mm_srai_epi16(s1, 5); in vpx_int_pro_row_sse2() [all …]
|
/external/XNNPACK/src/qs8-vadd/gen/ |
D | minmax-sse41-mul16-ld64-x32.c | 75 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 76 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 77 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 78 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 79 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 80 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 81 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 82 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 158 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 159 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
|
D | minmax-sse2-mul16-ld64-x32.c | 83 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 84 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 85 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 86 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 87 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 88 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 89 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 90 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 168 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 169 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
|
D | minmax-sse41-mul16-ld64-x24.c | 67 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 68 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 69 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 70 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 71 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 72 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 137 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 138 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
|
D | minmax-sse2-mul16-ld64-x24.c | 73 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 74 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 75 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 76 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 77 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 78 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 145 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 146 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
|
D | minmax-sse2-mul16-ld64-x16.c | 63 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 64 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 65 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 66 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 120 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 121 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
|
D | minmax-sse41-mul16-ld64-x16.c | 59 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 60 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 61 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 62 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 114 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 115 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | bilinear_filter_sse2.c | 63 _mm_srai_epi16(compensated_lo, VP8_FILTER_SHIFT); in horizontal_16x16() 65 _mm_srai_epi16(compensated_hi, VP8_FILTER_SHIFT); in horizontal_16x16() 117 _mm_srai_epi16(compensated_lo, VP8_FILTER_SHIFT); in vertical_16x16() 119 _mm_srai_epi16(compensated_hi, VP8_FILTER_SHIFT); in vertical_16x16() 176 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in horizontal_8xN() 212 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in vertical_8xN() 276 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in horizontal_4x4() 314 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in vertical_4x4()
|