/external/libaom/libaom/av1/encoder/x86/ |
D | av1_fwd_txfm_sse2.c | 66 x1[0] = _mm_adds_epi16(input[0], input[3]); in fdct8x4_new_sse2() 68 x1[1] = _mm_adds_epi16(input[1], input[2]); in fdct8x4_new_sse2() 100 x1[0] = _mm_adds_epi16(input[0], input[7]); in fdct4x8_new_sse2() 102 x1[1] = _mm_adds_epi16(input[1], input[6]); in fdct4x8_new_sse2() 104 x1[2] = _mm_adds_epi16(input[2], input[5]); in fdct4x8_new_sse2() 106 x1[3] = _mm_adds_epi16(input[3], input[4]); in fdct4x8_new_sse2() 111 x2[0] = _mm_adds_epi16(x1[0], x1[3]); in fdct4x8_new_sse2() 113 x2[1] = _mm_adds_epi16(x1[1], x1[2]); in fdct4x8_new_sse2() 126 x3[4] = _mm_adds_epi16(x2[4], x2[5]); in fdct4x8_new_sse2() 129 x3[7] = _mm_adds_epi16(x2[7], x2[6]); in fdct4x8_new_sse2() [all …]
|
D | av1_fwd_txfm_sse2.h | 64 output[0] = _mm_adds_epi16(input[0], input[0]); in fidentity8x8_new_sse2() 65 output[1] = _mm_adds_epi16(input[1], input[1]); in fidentity8x8_new_sse2() 66 output[2] = _mm_adds_epi16(input[2], input[2]); in fidentity8x8_new_sse2() 67 output[3] = _mm_adds_epi16(input[3], input[3]); in fidentity8x8_new_sse2() 68 output[4] = _mm_adds_epi16(input[4], input[4]); in fidentity8x8_new_sse2() 69 output[5] = _mm_adds_epi16(input[5], input[5]); in fidentity8x8_new_sse2() 70 output[6] = _mm_adds_epi16(input[6], input[6]); in fidentity8x8_new_sse2() 71 output[7] = _mm_adds_epi16(input[7], input[7]); in fidentity8x8_new_sse2()
|
/external/libavc/common/x86/ |
D | ih264_weighted_pred_sse42.c | 422 y_0_8x16b = _mm_adds_epi16(round_8x16b, y_0_8x16b); in ih264_weighted_pred_luma_sse42() 423 y_2_8x16b = _mm_adds_epi16(round_8x16b, y_2_8x16b); in ih264_weighted_pred_luma_sse42() 428 y_0_8x16b = _mm_adds_epi16(ofst_8x16b, y_0_8x16b); in ih264_weighted_pred_luma_sse42() 429 y_2_8x16b = _mm_adds_epi16(ofst_8x16b, y_2_8x16b); in ih264_weighted_pred_luma_sse42() 468 y_0_8x16b = _mm_adds_epi16(round_8x16b, y_0_8x16b); in ih264_weighted_pred_luma_sse42() 469 y_1_8x16b = _mm_adds_epi16(round_8x16b, y_1_8x16b); in ih264_weighted_pred_luma_sse42() 470 y_2_8x16b = _mm_adds_epi16(round_8x16b, y_2_8x16b); in ih264_weighted_pred_luma_sse42() 471 y_3_8x16b = _mm_adds_epi16(round_8x16b, y_3_8x16b); in ih264_weighted_pred_luma_sse42() 478 y_0_8x16b = _mm_adds_epi16(ofst_8x16b, y_0_8x16b); in ih264_weighted_pred_luma_sse42() 479 y_1_8x16b = _mm_adds_epi16(ofst_8x16b, y_1_8x16b); in ih264_weighted_pred_luma_sse42() [all …]
|
/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_idct_intrin_sse2.c | 88 in[0] = _mm_adds_epi16(in[0], final_rounding); in vp9_iht8x8_64_add_sse2() 89 in[1] = _mm_adds_epi16(in[1], final_rounding); in vp9_iht8x8_64_add_sse2() 90 in[2] = _mm_adds_epi16(in[2], final_rounding); in vp9_iht8x8_64_add_sse2() 91 in[3] = _mm_adds_epi16(in[3], final_rounding); in vp9_iht8x8_64_add_sse2() 92 in[4] = _mm_adds_epi16(in[4], final_rounding); in vp9_iht8x8_64_add_sse2() 93 in[5] = _mm_adds_epi16(in[5], final_rounding); in vp9_iht8x8_64_add_sse2() 94 in[6] = _mm_adds_epi16(in[6], final_rounding); in vp9_iht8x8_64_add_sse2() 95 in[7] = _mm_adds_epi16(in[7], final_rounding); in vp9_iht8x8_64_add_sse2() 141 in[0] = _mm_adds_epi16(in[0], final_rounding); in write_buffer_8x16() 142 in[1] = _mm_adds_epi16(in[1], final_rounding); in write_buffer_8x16() [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | convolve_ssse3.h | 60 sum1 = _mm_adds_epi16(sum1, sum2); in convolve8_8_ssse3() 77 __m128i temp = _mm_adds_epi16(x0, x3); in convolve8_8_even_offset_ssse3() 78 temp = _mm_adds_epi16(temp, x1); in convolve8_8_even_offset_ssse3() 79 temp = _mm_adds_epi16(temp, x2); in convolve8_8_even_offset_ssse3() 80 temp = _mm_adds_epi16(temp, x4); in convolve8_8_even_offset_ssse3() 82 temp = _mm_adds_epi16(temp, k_64); in convolve8_8_even_offset_ssse3() 101 temp = _mm_adds_epi16(x0, x1); in convolve8_8_odd_offset_ssse3() 102 temp = _mm_adds_epi16(temp, x2); in convolve8_8_odd_offset_ssse3() 103 temp = _mm_adds_epi16(temp, x3); in convolve8_8_odd_offset_ssse3() 104 temp = _mm_adds_epi16(temp, x4); in convolve8_8_odd_offset_ssse3() [all …]
|
D | highbd_loopfilter_sse2.c | 159 filt = _mm_adds_epi16(filt, work_a); in vpx_highbd_lpf_horizontal_16_sse2() 160 filt = _mm_adds_epi16(filt, work_a); in vpx_highbd_lpf_horizontal_16_sse2() 161 filt = signed_char_clamp_bd_sse2(_mm_adds_epi16(filt, work_a), bd); in vpx_highbd_lpf_horizontal_16_sse2() 163 filter1 = signed_char_clamp_bd_sse2(_mm_adds_epi16(filt, t4), bd); in vpx_highbd_lpf_horizontal_16_sse2() 164 filter2 = signed_char_clamp_bd_sse2(_mm_adds_epi16(filt, t3), bd); in vpx_highbd_lpf_horizontal_16_sse2() 170 qs0 = _mm_adds_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 172 ps0 = _mm_adds_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 173 signed_char_clamp_bd_sse2(_mm_adds_epi16(ps0, filter2), bd), t80); in vpx_highbd_lpf_horizontal_16_sse2() 174 filt = _mm_adds_epi16(filter1, t1); in vpx_highbd_lpf_horizontal_16_sse2() 177 qs1 = _mm_adds_epi16(signed_char_clamp_bd_sse2(_mm_subs_epi16(qs1, filt), bd), in vpx_highbd_lpf_horizontal_16_sse2() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | aom_subpixel_8t_intrin_ssse3.c | 98 srcRegFilt32b1_1 = _mm_adds_epi16(srcRegFilt32b1_1, addFilterReg32); in aom_filter_block1d4_h4_ssse3() 171 resReglo = _mm_adds_epi16(resReglo, addFilterReg32); in aom_filter_block1d4_v4_ssse3() 172 resReghi = _mm_adds_epi16(resReghi, addFilterReg32); in aom_filter_block1d4_v4_ssse3() 243 srcRegFilt1 = _mm_adds_epi16(srcRegFilt1, srcRegFilt4); in aom_filter_block1d4_h8_intrin_ssse3() 245 srcRegFilt1 = _mm_adds_epi16(srcRegFilt1, minReg); in aom_filter_block1d4_h8_intrin_ssse3() 246 srcRegFilt1 = _mm_adds_epi16(srcRegFilt1, srcRegFilt3); in aom_filter_block1d4_h8_intrin_ssse3() 247 srcRegFilt1 = _mm_adds_epi16(srcRegFilt1, addFilterReg64); in aom_filter_block1d4_h8_intrin_ssse3() 301 srcRegFilt32b1_1 = _mm_adds_epi16(srcRegFilt32b3, srcRegFilt32b2); in aom_filter_block1d8_h4_ssse3() 304 srcRegFilt32b1_1 = _mm_adds_epi16(srcRegFilt32b1_1, addFilterReg32); in aom_filter_block1d8_h4_ssse3() 373 resReg23_45 = _mm_adds_epi16(resReg23, resReg45); in aom_filter_block1d8_v4_ssse3() [all …]
|
D | aom_subpixel_8t_intrin_sse2.c | 91 srcRegFilt32b1_1 = _mm_adds_epi16(srcRegFilt32b1_1, addFilterReg32); in aom_filter_block1d16_h4_sse2() 92 srcRegFilt32b2_1 = _mm_adds_epi16(srcRegFilt32b2_1, addFilterReg32); in aom_filter_block1d16_h4_sse2() 192 resReg23_45_lo = _mm_adds_epi16(resReg23_lo, resReg45_lo); in aom_filter_block1d16_v4_sse2() 193 resReg34_56_lo = _mm_adds_epi16(resReg34_lo, resReg56_lo); in aom_filter_block1d16_v4_sse2() 218 resReg23_45_hi = _mm_adds_epi16(resReg23_hi, resReg45_hi); in aom_filter_block1d16_v4_sse2() 219 resReg34_56_hi = _mm_adds_epi16(resReg34_hi, resReg56_hi); in aom_filter_block1d16_v4_sse2() 222 resReg23_45_lo = _mm_adds_epi16(resReg23_45_lo, addFilterReg32); in aom_filter_block1d16_v4_sse2() 223 resReg34_56_lo = _mm_adds_epi16(resReg34_56_lo, addFilterReg32); in aom_filter_block1d16_v4_sse2() 224 resReg23_45_hi = _mm_adds_epi16(resReg23_45_hi, addFilterReg32); in aom_filter_block1d16_v4_sse2() 225 resReg34_56_hi = _mm_adds_epi16(resReg34_56_hi, addFilterReg32); in aom_filter_block1d16_v4_sse2() [all …]
|
D | aom_subpixel_8t_intrin_avx2.c | 146 _mm_adds_epi16(srcRegFilt1_1, _mm256_castsi256_si128(addFilterReg32)); in aom_filter_block1d4_h4_avx2() 247 srcRegFilt1_1 = _mm_adds_epi16(srcRegFilt1_1, srcRegFilt2); in aom_filter_block1d4_h8_avx2() 251 _mm_adds_epi16(srcRegFilt1_1, _mm256_castsi256_si128(addFilterReg32)); in aom_filter_block1d4_h8_avx2() 342 srcRegFilt1_1 = _mm_adds_epi16(srcRegFilt2, srcRegFilt3); in aom_filter_block1d8_h4_avx2() 346 _mm_adds_epi16(srcRegFilt1_1, _mm256_castsi256_si128(addFilterReg32)); in aom_filter_block1d8_h4_avx2() 458 srcRegFilt1_1 = _mm_adds_epi16(srcRegFilt1_1, srcRegFilt2); in aom_filter_block1d8_h8_avx2() 472 _mm_adds_epi16(srcRegFilt1_1, _mm_adds_epi16(srcRegFilt3, srcRegFilt2)); in aom_filter_block1d8_h8_avx2() 476 _mm_adds_epi16(srcRegFilt1_1, _mm256_castsi256_si128(addFilterReg32)); in aom_filter_block1d8_h8_avx2() 734 srcRegFilt1_1 = _mm_adds_epi16(srcRegFilt1_1, srcRegFilt2); in aom_filter_block1d16_h8_avx2() 748 _mm_adds_epi16(srcRegFilt1_1, _mm_adds_epi16(srcRegFilt3, srcRegFilt2)); in aom_filter_block1d16_h8_avx2() [all …]
|
/external/libhevc/common/x86/ |
D | ihevc_weighted_pred_ssse3_intr.c | 1324 lvl_shift1_8x16b = _mm_adds_epi16(lvl_shift1_8x16b, lvl_shift2_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1325 lvl_shift1_8x16b = _mm_adds_epi16(lvl_shift1_8x16b, const_temp_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1362 src_temp1_8x16b = _mm_adds_epi16(src_temp1_8x16b, src_temp2_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1363 src_temp3_8x16b = _mm_adds_epi16(src_temp3_8x16b, src_temp4_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1364 src_temp5_8x16b = _mm_adds_epi16(src_temp5_8x16b, src_temp6_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1365 src_temp7_8x16b = _mm_adds_epi16(src_temp7_8x16b, src_temp8_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1373 src_temp1_8x16b = _mm_adds_epi16(src_temp1_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1374 src_temp3_8x16b = _mm_adds_epi16(src_temp3_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1375 src_temp5_8x16b = _mm_adds_epi16(src_temp5_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_ssse3() 1376 src_temp7_8x16b = _mm_adds_epi16(src_temp7_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_ssse3() [all …]
|
D | ihevc_weighted_pred_sse42_intr.c | 1810 lvl_shift1_8x16b = _mm_adds_epi16(lvl_shift1_8x16b, lvl_shift2_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1811 lvl_shift1_8x16b = _mm_adds_epi16(lvl_shift1_8x16b, const_temp_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1848 src_temp1_8x16b = _mm_adds_epi16(src_temp1_8x16b, src_temp2_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1849 src_temp3_8x16b = _mm_adds_epi16(src_temp3_8x16b, src_temp4_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1850 src_temp5_8x16b = _mm_adds_epi16(src_temp5_8x16b, src_temp6_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1851 src_temp7_8x16b = _mm_adds_epi16(src_temp7_8x16b, src_temp8_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1859 src_temp1_8x16b = _mm_adds_epi16(src_temp1_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1860 src_temp3_8x16b = _mm_adds_epi16(src_temp3_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1861 src_temp5_8x16b = _mm_adds_epi16(src_temp5_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_sse42() 1862 src_temp7_8x16b = _mm_adds_epi16(src_temp7_8x16b, lvl_shift1_8x16b); in ihevc_weighted_pred_bi_default_sse42() [all …]
|
/external/libgav1/libgav1/src/dsp/x86/ |
D | inverse_transform_sse4.cc | 189 y = _mm_adds_epi16(*b, *a); in HadamardRotation() 192 x = _mm_adds_epi16(*a, *b); in HadamardRotation() 1608 const __m128i b = _mm_adds_epi16(a, v_src); in Identity4_SSE4_1() 1655 const __m128i v_dst_i = _mm_adds_epi16(v_src_mult, v_src); in Identity4ColumnStoreToFrame() 1656 const __m128i a = _mm_adds_epi16(v_dst_i, v_eight); in Identity4ColumnStoreToFrame() 1659 const __m128i d = _mm_adds_epi16(c, b); in Identity4ColumnStoreToFrame() 1673 const __m128i v_dst_i = _mm_adds_epi16(v_src_mult, v_src); in Identity4ColumnStoreToFrame() 1674 const __m128i a = _mm_adds_epi16(v_dst_i, v_eight); in Identity4ColumnStoreToFrame() 1677 const __m128i d = _mm_adds_epi16(c, b); in Identity4ColumnStoreToFrame() 1704 const __m128i v_dst_row = _mm_adds_epi16(v_src_mult, v_src); in Identity4RowColumnStoreToFrame() [all …]
|
/external/XNNPACK/src/qs8-vaddc/gen/ |
D | minmax-sse41-mul32-ld32-x32.c | 75 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 76 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 77 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 78 __m128i voutOPQRSTUV = _mm_adds_epi16(_mm_packs_epi32(vaccOPQR, vaccSTUV), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 112 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x32.c | 80 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 81 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 82 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 83 __m128i voutOPQRSTUV = _mm_adds_epi16(_mm_packs_epi32(vaccOPQR, vaccSTUV), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 117 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x24.c | 72 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24() 73 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24() 74 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24() 106 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24()
|
D | minmax-sse2-mul16-ld64-x32.c | 94 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 95 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 96 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 97 __m128i voutOPQRSTUV = _mm_adds_epi16(_mm_packs_epi32(vaccOPQR, vaccSTUV), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 139 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
|
D | minmax-sse41-mul16-ld64-x32.c | 90 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 91 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 92 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 93 __m128i voutOPQRSTUV = _mm_adds_epi16(_mm_packs_epi32(vaccOPQR, vaccSTUV), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 134 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
|
D | minmax-sse41-mul32-ld32-x24.c | 67 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24() 68 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24() 69 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24() 101 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24()
|
/external/fec/ |
D | viterbi39_sse2.c | 147 m0 = _mm_adds_epi16(vp->old_metrics->v[i],metric); in update_viterbi39_blk_sse2() 148 m3 = _mm_adds_epi16(vp->old_metrics->v[16+i],metric); in update_viterbi39_blk_sse2() 149 m1 = _mm_adds_epi16(vp->old_metrics->v[16+i],m_metric); in update_viterbi39_blk_sse2() 150 m2 = _mm_adds_epi16(vp->old_metrics->v[i],m_metric); in update_viterbi39_blk_sse2()
|
D | viterbi615_sse2.c | 150 m0 = _mm_adds_epi16(vp->old_metrics->v[i],metric); in update_viterbi615_blk_sse2() 151 m3 = _mm_adds_epi16(vp->old_metrics->v[1024+i],metric); in update_viterbi615_blk_sse2() 152 m1 = _mm_adds_epi16(vp->old_metrics->v[1024+i],m_metric); in update_viterbi615_blk_sse2() 153 m2 = _mm_adds_epi16(vp->old_metrics->v[i],m_metric); in update_viterbi615_blk_sse2()
|
/external/XNNPACK/src/qs8-vadd/gen/ |
D | minmax-sse41-mul32-ld32-x32.c | 91 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 92 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 93 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 94 __m128i voutOPQRSTUV = _mm_adds_epi16(_mm_packs_epi32(vaccOPQR, vaccSTUV), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 134 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x32.c | 96 __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 97 __m128i vout89ABCDEF = _mm_adds_epi16(_mm_packs_epi32(vacc89AB, vaccCDEF), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 98 __m128i voutGHIJKLMN = _mm_adds_epi16(_mm_packs_epi32(vaccGHIJ, vaccKLMN), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 99 __m128i voutOPQRSTUV = _mm_adds_epi16(_mm_packs_epi32(vaccOPQR, vaccSTUV), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 139 … __m128i vout01234567 = _mm_adds_epi16(_mm_packs_epi32(vacc0123, vacc4567), voutput_zero_point); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_quantize_sse2.c | 69 qcoeff0 = _mm_adds_epi16(qcoeff0, round); in vp9_quantize_fp_sse2() 71 qcoeff1 = _mm_adds_epi16(qcoeff1, round); in vp9_quantize_fp_sse2() 140 qcoeff0 = _mm_adds_epi16(qcoeff0, round); in vp9_quantize_fp_sse2() 141 qcoeff1 = _mm_adds_epi16(qcoeff1, round); in vp9_quantize_fp_sse2()
|
/external/libaom/libaom/av1/common/x86/ |
D | av1_inv_txfm_ssse3.h | 41 in0 = _mm_adds_epi16(_in0, _in1); \ 50 in0 = _mm_adds_epi16(_in0, _in1); \ 57 out0 = _mm_adds_epi16(_in0, _in1); \
|
/external/XNNPACK/src/qu8-requantization/ |
D | fp32-sse2.c | 74 const __m128i xy_packed = _mm_adds_epi16(_mm_packs_epi32(x_rounded, y_rounded), vzero_point); in xnn_qu8_requantize_fp32__sse2() 75 const __m128i zw_packed = _mm_adds_epi16(_mm_packs_epi32(z_rounded, w_rounded), vzero_point); in xnn_qu8_requantize_fp32__sse2()
|