Home
last modified time | relevance | path

Searched refs:_mm_srai_epi16 (Results 1 – 25 of 125) sorted by relevance

12345

/external/libvpx/libvpx/vp9/common/x86/
Dvp9_idct_intrin_sse2.c46 in[0] = _mm_srai_epi16(in[0], 4); in vp9_iht4x4_16_add_sse2()
47 in[1] = _mm_srai_epi16(in[1], 4); in vp9_iht4x4_16_add_sse2()
97 in[0] = _mm_srai_epi16(in[0], 5); in vp9_iht8x8_64_add_sse2()
98 in[1] = _mm_srai_epi16(in[1], 5); in vp9_iht8x8_64_add_sse2()
99 in[2] = _mm_srai_epi16(in[2], 5); in vp9_iht8x8_64_add_sse2()
100 in[3] = _mm_srai_epi16(in[3], 5); in vp9_iht8x8_64_add_sse2()
101 in[4] = _mm_srai_epi16(in[4], 5); in vp9_iht8x8_64_add_sse2()
102 in[5] = _mm_srai_epi16(in[5], 5); in vp9_iht8x8_64_add_sse2()
103 in[6] = _mm_srai_epi16(in[6], 5); in vp9_iht8x8_64_add_sse2()
104 in[7] = _mm_srai_epi16(in[7], 5); in vp9_iht8x8_64_add_sse2()
[all …]
/external/libavc/common/x86/
Dih264_chroma_intra_pred_filters_ssse3.c345 res1_sh_l_8x16b = _mm_srai_epi16(res1_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
346 res1_sh_h_8x16b = _mm_srai_epi16(res1_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
347 res2_sh_l_8x16b = _mm_srai_epi16(res2_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
348 res2_sh_h_8x16b = _mm_srai_epi16(res2_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
365 res1_sh_l_8x16b = _mm_srai_epi16(res1_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
366 res1_sh_h_8x16b = _mm_srai_epi16(res1_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
367 res2_sh_l_8x16b = _mm_srai_epi16(res2_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
368 res2_sh_h_8x16b = _mm_srai_epi16(res2_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
384 res1_sh_l_8x16b = _mm_srai_epi16(res1_l_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
385 res1_sh_h_8x16b = _mm_srai_epi16(res1_h_8x16b, 5); in ih264_intra_pred_chroma_8x8_mode_plane_ssse3()
[all …]
Dih264_weighted_pred_sse42.c425 y_0_8x16b = _mm_srai_epi16(y_0_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
426 y_2_8x16b = _mm_srai_epi16(y_2_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
473 y_0_8x16b = _mm_srai_epi16(y_0_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
474 y_1_8x16b = _mm_srai_epi16(y_1_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
475 y_2_8x16b = _mm_srai_epi16(y_2_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
476 y_3_8x16b = _mm_srai_epi16(y_3_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
541 y_0L_8x16b = _mm_srai_epi16(y_0L_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
542 y_0H_8x16b = _mm_srai_epi16(y_0H_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
543 y_1L_8x16b = _mm_srai_epi16(y_1L_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
544 y_1H_8x16b = _mm_srai_epi16(y_1H_8x16b, log_wd); in ih264_weighted_pred_luma_sse42()
[all …]
Dih264_luma_intra_pred_filters_ssse3.c351 res1_8x16b = _mm_srai_epi16(res1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_diag_dl_ssse3()
437 res1_8x16b = _mm_srai_epi16(res1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_diag_dr_ssse3()
526 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_vert_r_ssse3()
620 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_horz_d_ssse3()
714 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_vert_l_ssse3()
809 w121_a1_8x16b = _mm_srai_epi16(w121_a1_8x16b, 2); in ih264_intra_pred_luma_4x4_mode_horz_u_ssse3()
1113 res1_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dl_ssse3()
1126 res2_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dl_ssse3()
1214 res1_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dr_ssse3()
1228 res2_8x16 = _mm_srai_epi16(a0_8x16, 2); in ih264_intra_pred_luma_8x8_mode_diag_dr_ssse3()
[all …]
Dih264_deblk_luma_ssse3.c218 Alpha_8x16 = _mm_srai_epi16(Alpha_8x16, 2); in ih264_deblk_luma_vert_bs4_ssse3()
274 p0_16x8_1 = _mm_srai_epi16(temp1, 2); in ih264_deblk_luma_vert_bs4_ssse3()
275 q0_16x8_1 = _mm_srai_epi16(temp2, 2); in ih264_deblk_luma_vert_bs4_ssse3()
282 p1_16x8_2 = _mm_srai_epi16(temp1, 2); in ih264_deblk_luma_vert_bs4_ssse3()
283 q1_16x8_2 = _mm_srai_epi16(temp2, 2); in ih264_deblk_luma_vert_bs4_ssse3()
296 p0_16x8_2 = _mm_srai_epi16(temp1, 3); in ih264_deblk_luma_vert_bs4_ssse3()
297 q0_16x8_2 = _mm_srai_epi16(temp2, 3); in ih264_deblk_luma_vert_bs4_ssse3()
312 p2_16x8_2 = _mm_srai_epi16(temp1, 3); in ih264_deblk_luma_vert_bs4_ssse3()
313 q2_16x8_2 = _mm_srai_epi16(temp2, 3); in ih264_deblk_luma_vert_bs4_ssse3()
334 temp1 = _mm_srai_epi16(temp1, 2); in ih264_deblk_luma_vert_bs4_ssse3()
[all …]
/external/libhevc/common/x86/
Dihevc_intra_pred_filters_sse42_intr.c252 src_temp1 = _mm_srai_epi16(src_temp1, 2); in ihevc_intra_pred_ref_filtering_sse42()
268 src_temp4 = _mm_srai_epi16(src_temp4, 2); in ihevc_intra_pred_ref_filtering_sse42()
1038 src_temp3 = _mm_srai_epi16(src_temp3, 1); in ihevc_intra_pred_luma_horz_sse42()
1096 src_temp10 = _mm_srai_epi16(src_temp10, 1); in ihevc_intra_pred_luma_horz_sse42()
1509 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42()
1510 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42()
1511 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42()
1512 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42()
1590 ref_main_idx_4x32b = _mm_sub_epi16(two_nt_4x32b, _mm_srai_epi16(res_temp5_4x32b, 5)); in ihevc_intra_pred_luma_mode_3_to_9_sse42()
1654 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_sse42()
[all …]
Dihevc_intra_pred_filters_ssse3_intr.c537 src_temp1 = _mm_srai_epi16(src_temp1, 2); in ihevc_intra_pred_ref_filtering_ssse3()
553 src_temp4 = _mm_srai_epi16(src_temp4, 2); in ihevc_intra_pred_ref_filtering_ssse3()
1547 src_temp3 = _mm_srai_epi16(src_temp3, 1); in ihevc_intra_pred_luma_horz_ssse3()
1606 src_temp10 = _mm_srai_epi16(src_temp10, 1); in ihevc_intra_pred_luma_horz_ssse3()
2427 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3()
2428 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3()
2429 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3()
2430 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3()
2508 ref_main_idx_4x32b = _mm_sub_epi16(two_nt_4x32b, _mm_srai_epi16(res_temp5_4x32b, 5)); in ihevc_intra_pred_luma_mode_3_to_9_ssse3()
2572 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_luma_mode_3_to_9_ssse3()
[all …]
Dihevc_chroma_intra_pred_filters_ssse3_intr.c1210 ref_main_idx_4x32b = _mm_srai_epi16(res_temp5_4x32b, 5); in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1277 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1278 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1279 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1280 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1338 ref_main_idx_4x32b = _mm_srai_epi16(res_temp5_4x32b, 5); in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1451 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, 5); /* col=0*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1452 src_temp2_8x16b = _mm_srai_epi16(src_temp2_8x16b, 5); /* col=1*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1453 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, 5); /* col=2*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
1454 src_temp4_8x16b = _mm_srai_epi16(src_temp4_8x16b, 5); /* col=3*/ in ihevc_intra_pred_chroma_mode_3_to_9_ssse3()
[all …]
Dihevc_weighted_pred_ssse3_intr.c1385 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1386 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1387 src_temp5_8x16b = _mm_srai_epi16(src_temp5_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1388 src_temp7_8x16b = _mm_srai_epi16(src_temp7_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1397 src_temp9_8x16b = _mm_srai_epi16(src_temp9_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1398 src_temp11_8x16b = _mm_srai_epi16(src_temp11_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1399 src_temp13_8x16b = _mm_srai_epi16(src_temp13_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1400 src_temp15_8x16b = _mm_srai_epi16(src_temp15_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1460 src_temp1_8x16b = _mm_srai_epi16(src_temp1_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
1461 src_temp3_8x16b = _mm_srai_epi16(src_temp3_8x16b, shift); in ihevc_weighted_pred_bi_default_ssse3()
[all …]
/external/libaom/libaom/aom_dsp/x86/
Daom_subpixel_8t_intrin_ssse3.c77 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d4_h4_ssse3()
99 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d4_h4_ssse3()
130 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d4_v4_ssse3()
173 resReglo = _mm_srai_epi16(resReglo, 6); in aom_filter_block1d4_v4_ssse3()
174 resReghi = _mm_srai_epi16(resReghi, 6); in aom_filter_block1d4_v4_ssse3()
250 srcRegFilt1 = _mm_srai_epi16(srcRegFilt1, 7); in aom_filter_block1d4_h8_intrin_ssse3()
275 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d8_h4_ssse3()
305 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d8_h4_ssse3()
334 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d8_v4_ssse3()
379 resReg23_45 = _mm_srai_epi16(resReg23_45, 6); in aom_filter_block1d8_v4_ssse3()
[all …]
Daom_subpixel_8t_intrin_sse2.c33 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d16_h4_sse2()
93 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d16_h4_sse2()
94 srcRegFilt32b2_1 = _mm_srai_epi16(srcRegFilt32b2_1, 6); in aom_filter_block1d16_h4_sse2()
127 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d16_v4_sse2()
226 resReg23_45_lo = _mm_srai_epi16(resReg23_45_lo, 6); in aom_filter_block1d16_v4_sse2()
227 resReg34_56_lo = _mm_srai_epi16(resReg34_56_lo, 6); in aom_filter_block1d16_v4_sse2()
228 resReg23_45_hi = _mm_srai_epi16(resReg23_45_hi, 6); in aom_filter_block1d16_v4_sse2()
229 resReg34_56_hi = _mm_srai_epi16(resReg34_56_hi, 6); in aom_filter_block1d16_v4_sse2()
271 filtersReg = _mm_srai_epi16(filtersReg, 1); in aom_filter_block1d8_h4_sse2()
306 srcRegFilt32b1_1 = _mm_srai_epi16(srcRegFilt32b1_1, 6); in aom_filter_block1d8_h4_sse2()
[all …]
Dfwd_txfm_impl_sse2.h507 const __m128i sign_in0 = _mm_srai_epi16(in0, 15); in FDCT8x8_2D()
508 const __m128i sign_in1 = _mm_srai_epi16(in1, 15); in FDCT8x8_2D()
509 const __m128i sign_in2 = _mm_srai_epi16(in2, 15); in FDCT8x8_2D()
510 const __m128i sign_in3 = _mm_srai_epi16(in3, 15); in FDCT8x8_2D()
511 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D()
512 const __m128i sign_in5 = _mm_srai_epi16(in5, 15); in FDCT8x8_2D()
513 const __m128i sign_in6 = _mm_srai_epi16(in6, 15); in FDCT8x8_2D()
514 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in FDCT8x8_2D()
523 in0 = _mm_srai_epi16(in0, 1); in FDCT8x8_2D()
524 in1 = _mm_srai_epi16(in1, 1); in FDCT8x8_2D()
[all …]
Davg_intrin_sse2.c336 b0 = _mm_srai_epi16(b0, 1); in hadamard_16x16_sse2()
337 b1 = _mm_srai_epi16(b1, 1); in hadamard_16x16_sse2()
338 b2 = _mm_srai_epi16(b2, 1); in hadamard_16x16_sse2()
339 b3 = _mm_srai_epi16(b3, 1); in hadamard_16x16_sse2()
396 b0 = _mm_srai_epi16(b0, 2); in aom_hadamard_32x32_sse2()
397 b1 = _mm_srai_epi16(b1, 2); in aom_hadamard_32x32_sse2()
398 b2 = _mm_srai_epi16(b2, 2); in aom_hadamard_32x32_sse2()
399 b3 = _mm_srai_epi16(b3, 2); in aom_hadamard_32x32_sse2()
475 s0 = _mm_srai_epi16(s0, 6); in aom_int_pro_row_sse2()
476 s1 = _mm_srai_epi16(s1, 6); in aom_int_pro_row_sse2()
[all …]
Dquantize_sse2.c48 coeff0_sign = _mm_srai_epi16(coeff0, 15); in aom_quantize_b_sse2()
49 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2()
91 coeff0_sign = _mm_srai_epi16(coeff0, 15); in aom_quantize_b_sse2()
92 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2()
Dintrapred_sse2.c318 sum_above = _mm_srai_epi16(sum_above, 2); in aom_dc_top_predictor_4x8_sse2()
332 sum_above = _mm_srai_epi16(sum_above, 2); in aom_dc_top_predictor_4x16_sse2()
346 sum_above = _mm_srai_epi16(sum_above, 3); in aom_dc_top_predictor_8x4_sse2()
358 sum_above = _mm_srai_epi16(sum_above, 3); in aom_dc_top_predictor_8x16_sse2()
370 sum_above = _mm_srai_epi16(sum_above, 3); in aom_dc_top_predictor_8x32_sse2()
382 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x4_sse2()
395 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x8_sse2()
409 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x32_sse2()
423 sum_above = _mm_srai_epi16(sum_above, 4); in aom_dc_top_predictor_16x64_sse2()
436 sum_above = _mm_srai_epi16(sum_above, 5); in aom_dc_top_predictor_32x8_sse2()
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_txfm_impl_sse2.h532 const __m128i sign_in0 = _mm_srai_epi16(in0, 15); in FDCT8x8_2D()
533 const __m128i sign_in1 = _mm_srai_epi16(in1, 15); in FDCT8x8_2D()
534 const __m128i sign_in2 = _mm_srai_epi16(in2, 15); in FDCT8x8_2D()
535 const __m128i sign_in3 = _mm_srai_epi16(in3, 15); in FDCT8x8_2D()
536 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D()
537 const __m128i sign_in5 = _mm_srai_epi16(in5, 15); in FDCT8x8_2D()
538 const __m128i sign_in6 = _mm_srai_epi16(in6, 15); in FDCT8x8_2D()
539 const __m128i sign_in7 = _mm_srai_epi16(in7, 15); in FDCT8x8_2D()
548 in0 = _mm_srai_epi16(in0, 1); in FDCT8x8_2D()
549 in1 = _mm_srai_epi16(in1, 1); in FDCT8x8_2D()
[all …]
Dquantize_sse2.c49 coeff0_sign = _mm_srai_epi16(coeff0, 15); in vpx_quantize_b_sse2()
50 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2()
88 coeff0_sign = _mm_srai_epi16(coeff0, 15); in vpx_quantize_b_sse2()
89 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2()
Davg_intrin_sse2.c357 b0 = _mm_srai_epi16(b0, 1); in hadamard_16x16_sse2()
358 b1 = _mm_srai_epi16(b1, 1); in hadamard_16x16_sse2()
359 b2 = _mm_srai_epi16(b2, 1); in hadamard_16x16_sse2()
360 b3 = _mm_srai_epi16(b3, 1); in hadamard_16x16_sse2()
421 b0 = _mm_srai_epi16(b0, 2); in vpx_hadamard_32x32_sse2()
422 b1 = _mm_srai_epi16(b1, 2); in vpx_hadamard_32x32_sse2()
423 b2 = _mm_srai_epi16(b2, 2); in vpx_hadamard_32x32_sse2()
424 b3 = _mm_srai_epi16(b3, 2); in vpx_hadamard_32x32_sse2()
501 s0 = _mm_srai_epi16(s0, 5); in vpx_int_pro_row_sse2()
502 s1 = _mm_srai_epi16(s1, 5); in vpx_int_pro_row_sse2()
[all …]
/external/XNNPACK/src/qs8-vadd/gen/
Dminmax-sse41-mul16-ld64-x32.c75 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
76 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
77 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
78 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
79 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
80 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
81 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
82 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
158 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
159 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
Dminmax-sse2-mul16-ld64-x32.c83 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
84 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
85 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
86 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
87 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
88 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
89 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
90 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
168 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
169 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
Dminmax-sse41-mul16-ld64-x24.c67 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
68 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
69 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
70 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
71 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
72 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
137 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
138 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
Dminmax-sse2-mul16-ld64-x24.c73 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
74 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
75 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
76 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
77 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
78 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
145 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
146 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
Dminmax-sse2-mul16-ld64-x16.c63 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
64 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
65 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
66 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
120 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
121 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
Dminmax-sse41-mul16-ld64-x16.c59 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
60 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
61 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
62 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
114 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
115 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
/external/libvpx/libvpx/vp8/common/x86/
Dbilinear_filter_sse2.c63 _mm_srai_epi16(compensated_lo, VP8_FILTER_SHIFT); in horizontal_16x16()
65 _mm_srai_epi16(compensated_hi, VP8_FILTER_SHIFT); in horizontal_16x16()
117 _mm_srai_epi16(compensated_lo, VP8_FILTER_SHIFT); in vertical_16x16()
119 _mm_srai_epi16(compensated_hi, VP8_FILTER_SHIFT); in vertical_16x16()
176 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in horizontal_8xN()
212 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in vertical_8xN()
276 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in horizontal_4x4()
314 const __m128i shifted = _mm_srai_epi16(compensated, VP8_FILTER_SHIFT); in vertical_4x4()

12345