/external/libgav1/libgav1/src/dsp/x86/ |
D | intrapred_smooth_sse4.cc | 149 const __m128i top = _mm_cvtepu8_epi16(Load4(above)); in LoadSmoothPixels4() 163 weight_h[0] = _mm_cvtepu8_epi16(x_weights); in LoadSmoothWeights4() 169 weight_h[0] = _mm_cvtepu8_epi16(y_weights); in LoadSmoothWeights4() 174 weight_h[0] = _mm_cvtepu8_epi16(y_weights); in LoadSmoothWeights4() 289 __m128i top_row = _mm_cvtepu8_epi16(LoadLo8(above)); in LoadSmoothPixels8() 324 weight_h[0] = _mm_cvtepu8_epi16(loaded_weights); in LoadSmoothWeights8() 330 __m128i weights_x = _mm_cvtepu8_epi16(loaded_weights); in LoadSmoothWeights8() 342 weight_h[0] = _mm_cvtepu8_epi16(loaded_weights); in LoadSmoothWeights8() 349 weight_h[0] = _mm_cvtepu8_epi16(weight_lo); in LoadSmoothWeights8() 354 weight_h[4] = _mm_cvtepu8_epi16(weight_hi); in LoadSmoothWeights8() [all …]
|
D | intra_edge_sse4.cc | 48 const __m128i outers_lo = _mm_cvtepu8_epi16(edge_lo); in ComputeKernel1Store12() 49 const __m128i outers_hi = _mm_cvtepu8_epi16(edge_hi); in ComputeKernel1Store12() 83 const __m128i outers_lo = _mm_cvtepu8_epi16(edge_lo); in ComputeKernel2Store12() 85 const __m128i outers_hi = _mm_cvtepu8_epi16(edge_hi); in ComputeKernel2Store12() 123 const __m128i source2_lo = _mm_slli_epi16(_mm_cvtepu8_epi16(edge_lo), 1); in ComputeKernel3Store8() 136 const __m128i source2_hi = _mm_slli_epi16(_mm_cvtepu8_epi16(edge_hi), 1); in ComputeKernel3Store8() 218 const __m128i src_lo = _mm_cvtepu8_epi16(data); in IntraEdgeUpsampler_SSE4_1() 230 const __m128i src_hi_extra = _mm_cvtepu8_epi16(LoadLo8(temp + 16)); in IntraEdgeUpsampler_SSE4_1()
|
D | mask_blend_sse4.cc | 41 const __m128i mask_val_0 = _mm_cvtepu8_epi16(LoadLo8(mask)); in GetMask4x2() 43 _mm_cvtepu8_epi16(LoadLo8(mask + (mask_stride << subsampling_y))); in GetMask4x2() 47 _mm_cvtepu8_epi16(LoadLo8(mask + mask_stride)); in GetMask4x2() 49 _mm_cvtepu8_epi16(LoadLo8(mask + mask_stride * 3)); in GetMask4x2() 57 return _mm_cvtepu8_epi16( in GetMask4x2() 69 const __m128i mask_val_0 = _mm_cvtepu8_epi16(row_vals); in GetMask8() 70 const __m128i mask_val_1 = _mm_cvtepu8_epi16(_mm_srli_si128(row_vals, 8)); in GetMask8() 75 const __m128i next_mask_val_0 = _mm_cvtepu8_epi16(next_row_vals); in GetMask8() 77 _mm_cvtepu8_epi16(_mm_srli_si128(next_row_vals, 8)); in GetMask8() 85 return _mm_cvtepu8_epi16(mask_val); in GetMask8() [all …]
|
D | intrapred_cfl_sse4.cc | 147 samples = _mm_slli_epi16(_mm_cvtepu8_epi16(samples), 3); in CflSubsampler444_4xH_SSE4_1() 240 samples0 = _mm_slli_epi16(_mm_cvtepu8_epi16(samples0), 3); in CflSubsampler444_8xH_SSE4_1() 253 samples1 = _mm_slli_epi16(_mm_cvtepu8_epi16(samples1), 3); in CflSubsampler444_8xH_SSE4_1() 355 samples0 = _mm_slli_epi16(_mm_cvtepu8_epi16(samples01), 3); in CflSubsampler444_SSE4_1() 371 samples2 = _mm_slli_epi16(_mm_cvtepu8_epi16(samples23), 3); in CflSubsampler444_SSE4_1() 455 const __m128i samples_row0 = _mm_cvtepu8_epi16(LoadLo8(src)); in CflSubsampler420_4xH_SSE4_1() 457 const __m128i samples_row1 = _mm_cvtepu8_epi16(LoadLo8(src)); in CflSubsampler420_4xH_SSE4_1() 461 const __m128i samples_row2 = _mm_cvtepu8_epi16(LoadLo8(src)); in CflSubsampler420_4xH_SSE4_1() 463 const __m128i samples_row3 = _mm_cvtepu8_epi16(LoadLo8(src)); in CflSubsampler420_4xH_SSE4_1() 469 const __m128i samples_row4 = _mm_cvtepu8_epi16(LoadLo8(src)); in CflSubsampler420_4xH_SSE4_1() [all …]
|
D | film_grain_sse4.cc | 48 return _mm_cvtepu8_epi16(LoadLo8(src)); in LoadSource() 52 return _mm_cvtepu8_epi16(LoadLo8Msan(src, 8 - valid_range)); in LoadSourceMsan() 79 _mm_hadd_epi16(_mm_cvtepu8_epi16(src), in GetAverageLuma() 83 return _mm_cvtepu8_epi16(LoadLo8(luma)); in GetAverageLuma() 92 _mm_hadd_epi16(_mm_cvtepu8_epi16(src), in GetAverageLumaMsan() 96 return _mm_cvtepu8_epi16(LoadLo8Msan(luma, 8 - valid_range)); in GetAverageLumaMsan()
|
D | intrapred_sse4.cc | 671 const __m128i top_dists_y16 = _mm_cvtepu8_epi16(top_dists_y8); in WritePaethLine16() 673 const __m128i lefts_y16 = _mm_cvtepu8_epi16(lefts_y8); in WritePaethLine16() 867 const __m128i left = _mm_cvtepu8_epi16(Load4(left_column)); in Paeth8x4_SSE4_1() 868 const __m128i top = _mm_cvtepu8_epi16(LoadLo8(top_row)); in Paeth8x4_SSE4_1() 896 const __m128i left = _mm_cvtepu8_epi16(LoadLo8(left_column)); in Paeth8x8_SSE4_1() 897 const __m128i top = _mm_cvtepu8_epi16(LoadLo8(top_row)); in Paeth8x8_SSE4_1() 939 const __m128i left_lo = _mm_cvtepu8_epi16(left); in Paeth8x16_SSE4_1() 940 const __m128i left_hi = _mm_cvtepu8_epi16(_mm_srli_si128(left, 8)); in Paeth8x16_SSE4_1() 941 const __m128i top = _mm_cvtepu8_epi16(LoadLo8(top_row)); in Paeth8x16_SSE4_1() 1018 const __m128i top_lo = _mm_cvtepu8_epi16(top); in Paeth16x4_SSE4_1() [all …]
|
D | convolve_sse4.cc | 499 const __m128i v_src_ext_lo = _mm_cvtepu8_epi16(v_src); in ConvolveCompoundCopy_SSE4() 501 _mm_cvtepu8_epi16(_mm_srli_si128(v_src, 8)); in ConvolveCompoundCopy_SSE4() 518 const __m128i v_src_ext = _mm_cvtepu8_epi16(v_src); in ConvolveCompoundCopy_SSE4() 530 const __m128i v_src_ext = _mm_cvtepu8_epi16(v_src); in ConvolveCompoundCopy_SSE4() 1626 const __m128i a = _mm_cvtepu8_epi16(LoadLo8(src)); in LoadU8AndAddLong() 1627 const __m128i b = _mm_cvtepu8_epi16(LoadLo8(src1)); in LoadU8AndAddLong() 1818 __m128i left = _mm_cvtepu8_epi16(Load4(src)); in ConvolveIntraBlockCopy2D_SSE4_1() 1819 __m128i right = _mm_cvtepu8_epi16(Load4(src + 1)); in ConvolveIntraBlockCopy2D_SSE4_1() 1834 _mm_add_epi16(_mm_cvtepu8_epi16(left), _mm_cvtepu8_epi16(right)); in ConvolveIntraBlockCopy2D_SSE4_1() 1852 _mm_add_epi16(_mm_cvtepu8_epi16(left), _mm_cvtepu8_epi16(right)); in ConvolveIntraBlockCopy2D_SSE4_1() [all …]
|
/external/libhevc/common/x86/ |
D | ihevc_inter_pred_filters_sse42_intr.c | 135 src0_16x8b = _mm_cvtepu8_epi16(src0_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 136 src1_16x8b = _mm_cvtepu8_epi16(src1_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 137 src2_16x8b = _mm_cvtepu8_epi16(src2_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 138 src3_16x8b = _mm_cvtepu8_epi16(src3_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 171 src0_16x8b = _mm_cvtepu8_epi16(src0_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 172 src1_16x8b = _mm_cvtepu8_epi16(src1_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 173 src2_16x8b = _mm_cvtepu8_epi16(src2_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 174 src3_16x8b = _mm_cvtepu8_epi16(src3_16x8b); in ihevc_inter_pred_luma_copy_w16out_sse42() 487 src0_16x8b = _mm_cvtepu8_epi16(src0_16x8b); in ihevc_inter_pred_chroma_copy_w16out_sse42() 488 src1_16x8b = _mm_cvtepu8_epi16(src1_16x8b); in ihevc_inter_pred_chroma_copy_w16out_sse42() [all …]
|
D | ihevc_chroma_intra_pred_filters_sse42_intr.c | 199 src_temp_8x16b = _mm_cvtepu8_epi16(src_temp_8x16b); /* row=0*/ in ihevc_intra_pred_chroma_planar_sse42() 315 src_temp5 = _mm_cvtepu8_epi16(src_temp3); in ihevc_intra_pred_chroma_dc_sse42() 316 src_temp6 = _mm_cvtepu8_epi16(src_temp4); in ihevc_intra_pred_chroma_dc_sse42() 317 src_temp9 = _mm_cvtepu8_epi16(src_temp7); in ihevc_intra_pred_chroma_dc_sse42() 318 src_temp10 = _mm_cvtepu8_epi16(src_temp8); in ihevc_intra_pred_chroma_dc_sse42() 325 src_temp3 = _mm_cvtepu8_epi16(src_temp3); in ihevc_intra_pred_chroma_dc_sse42() 326 src_temp4 = _mm_cvtepu8_epi16(src_temp4); in ihevc_intra_pred_chroma_dc_sse42() 327 src_temp7 = _mm_cvtepu8_epi16(src_temp7); in ihevc_intra_pred_chroma_dc_sse42() 328 src_temp8 = _mm_cvtepu8_epi16(src_temp8); in ihevc_intra_pred_chroma_dc_sse42() 355 src_temp5 = _mm_cvtepu8_epi16(src_temp3); in ihevc_intra_pred_chroma_dc_sse42() [all …]
|
/external/libavc/common/x86/ |
D | ih264_weighted_pred_sse42.c | 416 y_0_8x16b = _mm_cvtepu8_epi16(y_0_16x8b); in ih264_weighted_pred_luma_sse42() 417 y_2_8x16b = _mm_cvtepu8_epi16(y_2_16x8b); in ih264_weighted_pred_luma_sse42() 458 y_0_8x16b = _mm_cvtepu8_epi16(y_0_16x8b); in ih264_weighted_pred_luma_sse42() 459 y_1_8x16b = _mm_cvtepu8_epi16(y_1_16x8b); in ih264_weighted_pred_luma_sse42() 460 y_2_8x16b = _mm_cvtepu8_epi16(y_2_16x8b); in ih264_weighted_pred_luma_sse42() 461 y_3_8x16b = _mm_cvtepu8_epi16(y_3_16x8b); in ih264_weighted_pred_luma_sse42() 514 y_0L_8x16b = _mm_cvtepu8_epi16(y_0_16x8b); in ih264_weighted_pred_luma_sse42() 516 y_1L_8x16b = _mm_cvtepu8_epi16(y_1_16x8b); in ih264_weighted_pred_luma_sse42() 518 y_2L_8x16b = _mm_cvtepu8_epi16(y_2_16x8b); in ih264_weighted_pred_luma_sse42() 520 y_3L_8x16b = _mm_cvtepu8_epi16(y_3_16x8b); in ih264_weighted_pred_luma_sse42() [all …]
|
D | ih264_resi_trans_quant_sse42.c | 134 src_r0 = _mm_cvtepu8_epi16(src_r0); in ih264_resi_trans_quant_4x4_sse42() 135 src_r1 = _mm_cvtepu8_epi16(src_r1); in ih264_resi_trans_quant_4x4_sse42() 136 src_r2 = _mm_cvtepu8_epi16(src_r2); in ih264_resi_trans_quant_4x4_sse42() 137 src_r3 = _mm_cvtepu8_epi16(src_r3); in ih264_resi_trans_quant_4x4_sse42() 144 pred_r0 = _mm_cvtepu8_epi16(pred_r0); //p00 p01 p02 p03 -- all 16 bits in ih264_resi_trans_quant_4x4_sse42() 145 pred_r1 = _mm_cvtepu8_epi16(pred_r1); //p10 p11 p12 p13 -- all 16 bits in ih264_resi_trans_quant_4x4_sse42() 146 pred_r2 = _mm_cvtepu8_epi16(pred_r2); //p20 p21 p22 p23 -- all 16 bits in ih264_resi_trans_quant_4x4_sse42() 147 pred_r3 = _mm_cvtepu8_epi16(pred_r3); //p30 p31 p32 p33 -- all 16 bits in ih264_resi_trans_quant_4x4_sse42()
|
/external/libaom/libaom/av1/common/x86/ |
D | reconinter_sse4.c | 39 const __m128i s0 = _mm_cvtepu8_epi16(s0AB); in av1_build_compound_diffwtd_mask_sse4_1() 44 const __m128i s1 = _mm_cvtepu8_epi16(s1AB); in av1_build_compound_diffwtd_mask_sse4_1() 60 s0 = _mm_cvtepu8_epi16(s0); in av1_build_compound_diffwtd_mask_sse4_1() 61 s1 = _mm_cvtepu8_epi16(s1); in av1_build_compound_diffwtd_mask_sse4_1() 77 const __m128i s0L = _mm_cvtepu8_epi16(s0); in av1_build_compound_diffwtd_mask_sse4_1() 78 const __m128i s1L = _mm_cvtepu8_epi16(s1); in av1_build_compound_diffwtd_mask_sse4_1()
|
D | av1_convolve_horiz_rs_sse4.c | 89 const __m128i src0_16 = _mm_cvtepu8_epi16(src0_8); in av1_convolve_horiz_rs_sse4_1() 90 const __m128i src1_16 = _mm_cvtepu8_epi16(src1_8); in av1_convolve_horiz_rs_sse4_1() 91 const __m128i src2_16 = _mm_cvtepu8_epi16(src2_8); in av1_convolve_horiz_rs_sse4_1() 92 const __m128i src3_16 = _mm_cvtepu8_epi16(src3_8); in av1_convolve_horiz_rs_sse4_1()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | corner_match_sse4.c | 66 const __m128i v1_l = _mm_cvtepu8_epi16(v1); in av1_compute_cross_correlation_sse4_1() 67 const __m128i v1_r = _mm_cvtepu8_epi16(_mm_srli_si128(v1, 8)); in av1_compute_cross_correlation_sse4_1() 68 const __m128i v2_l = _mm_cvtepu8_epi16(v2); in av1_compute_cross_correlation_sse4_1() 69 const __m128i v2_r = _mm_cvtepu8_epi16(_mm_srli_si128(v2, 8)); in av1_compute_cross_correlation_sse4_1()
|
D | pickrst_sse4.c | 23 const __m128i d0 = _mm_madd_epi16(*kl, _mm_cvtepu8_epi16(s)); in acc_stat_sse41() 25 _mm_madd_epi16(*kl, _mm_cvtepu8_epi16(_mm_srli_si128(s, 8))); in acc_stat_sse41() 56 _mm_cvtepu8_epi16(_mm_set1_epi16(*((uint16_t *)(dgd_ijk + l)))); in acc_stat_win7_one_line_sse4_1() 415 _mm_cvtepu8_epi16(_mm_set1_epi16(*((uint16_t *)(dgd_ijk + l)))); in acc_stat_win5_one_line_sse4_1() 520 const __m128i d0 = _mm_cvtepu8_epi16(xx_loadl_64(dat + j)); in av1_lowbd_pixel_proj_error_sse4_1() 521 const __m128i s0 = _mm_cvtepu8_epi16(xx_loadl_64(src + j)); in av1_lowbd_pixel_proj_error_sse4_1() 564 const __m128i d0 = _mm_cvtepu8_epi16(xx_loadl_64(dat + j)); in av1_lowbd_pixel_proj_error_sse4_1() 565 const __m128i s0 = _mm_cvtepu8_epi16(xx_loadl_64(src + j)); in av1_lowbd_pixel_proj_error_sse4_1() 599 const __m128i d0 = _mm_cvtepu8_epi16(d); in av1_lowbd_pixel_proj_error_sse4_1() 600 const __m128i d1 = _mm_cvtepu8_epi16(_mm_srli_si128(d, 8)); in av1_lowbd_pixel_proj_error_sse4_1() [all …]
|
/external/libmpeg2/common/x86/ |
D | impeg2_inter_pred_sse42_intr.c | 493 src_r0 = _mm_cvtepu8_epi16(src_r0); in impeg2_mc_halfx_halfy_8x8_sse42() 494 src_r0_1 = _mm_cvtepu8_epi16(src_r0_1); in impeg2_mc_halfx_halfy_8x8_sse42() 495 src_r1 = _mm_cvtepu8_epi16(src_r1); in impeg2_mc_halfx_halfy_8x8_sse42() 496 src_r1_1 = _mm_cvtepu8_epi16(src_r1_1); in impeg2_mc_halfx_halfy_8x8_sse42() 515 src_r0 = _mm_cvtepu8_epi16(src_r0); in impeg2_mc_halfx_halfy_8x8_sse42() 516 src_r0_1 = _mm_cvtepu8_epi16(src_r0_1); in impeg2_mc_halfx_halfy_8x8_sse42() 534 src_r0 = _mm_cvtepu8_epi16(src_r0); in impeg2_mc_halfx_halfy_8x8_sse42() 535 src_r0_1 = _mm_cvtepu8_epi16(src_r0_1); in impeg2_mc_halfx_halfy_8x8_sse42() 554 src_r0 = _mm_cvtepu8_epi16(src_r0); in impeg2_mc_halfx_halfy_8x8_sse42() 555 src_r0_1 = _mm_cvtepu8_epi16(src_r0_1); in impeg2_mc_halfx_halfy_8x8_sse42() [all …]
|
D | impeg2_idct_recon_sse42_intr.c | 1870 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 1896 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 1922 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 1948 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 1974 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 2000 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 2026 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 2052 pred_r = _mm_cvtepu8_epi16(pred_r); in impeg2_idct_recon_dc_mismatch_sse42() 2104 pred_r0 = _mm_cvtepu8_epi16(pred_r0); in impeg2_idct_recon_dc_sse42() 2105 pred_r1 = _mm_cvtepu8_epi16(pred_r1); in impeg2_idct_recon_dc_sse42() [all …]
|
/external/gemmlowp/internal/ |
D | pack_sse.h | 97 xmm1 = _mm_cvtepu8_epi16(xmm9); in Pack() 103 xmm1 = _mm_cvtepu8_epi16(xmm10); in Pack() 107 xmm1 = _mm_cvtepu8_epi16(xmm11); in Pack() 111 xmm1 = _mm_cvtepu8_epi16(xmm12); in Pack()
|
D | pack_avx.h | 251 xmm1 = _mm_cvtepu8_epi16(xmm9); in Pack() 257 xmm1 = _mm_cvtepu8_epi16(xmm10); in Pack() 261 xmm1 = _mm_cvtepu8_epi16(xmm11); in Pack() 265 xmm1 = _mm_cvtepu8_epi16(xmm12); in Pack()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | sse_sse4.c | 44 const __m128i v_a00_w = _mm_cvtepu8_epi16(v_a0); in sse_w16_sse4_1() 45 const __m128i v_a01_w = _mm_cvtepu8_epi16(_mm_srli_si128(v_a0, 8)); in sse_w16_sse4_1() 46 const __m128i v_b00_w = _mm_cvtepu8_epi16(v_b0); in sse_w16_sse4_1() 47 const __m128i v_b01_w = _mm_cvtepu8_epi16(_mm_srli_si128(v_b0, 8)); in sse_w16_sse4_1() 60 const __m128i v_a_w = _mm_cvtepu8_epi16(_mm_unpacklo_epi32(v_a0, v_a1)); in sse4x2_sse4_1() 61 const __m128i v_b_w = _mm_cvtepu8_epi16(_mm_unpacklo_epi32(v_b0, v_b1)); in sse4x2_sse4_1() 69 const __m128i v_a_w = _mm_cvtepu8_epi16(v_a0); in sse8_sse4_1() 70 const __m128i v_b_w = _mm_cvtepu8_epi16(v_b0); in sse8_sse4_1()
|
D | blend_sse4.h | 30 const __m128i v_s0_w = _mm_cvtepu8_epi16(v_s0_b); in blend_4() 31 const __m128i v_s1_w = _mm_cvtepu8_epi16(v_s1_b); in blend_4() 45 const __m128i v_s0_w = _mm_cvtepu8_epi16(v_s0_b); in blend_8() 46 const __m128i v_s1_w = _mm_cvtepu8_epi16(v_s1_b); in blend_8()
|
D | blend_mask_sse4.h | 72 const __m128i m = _mm_cvtepu8_epi16(m0); in aom_lowbd_blend_a64_d16_mask_subw0_subh0_w4_sse4_1() 91 const __m128i m = _mm_cvtepu8_epi16(m0); in aom_lowbd_blend_a64_d16_mask_subw0_subh0_w8_sse4_1() 205 const __m128i m = _mm_cvtepu8_epi16(_mm_avg_epu8(m_ac, zeros)); in aom_lowbd_blend_a64_d16_mask_subw0_subh1_w4_sse4_1() 227 const __m128i m = _mm_cvtepu8_epi16(_mm_avg_epu8(m_ac, zeros)); in aom_lowbd_blend_a64_d16_mask_subw0_subh1_w8_sse4_1()
|
D | blend_a64_mask_sse4.c | 287 const __m128i v_r0_s_w = _mm_cvtepu8_epi16(v_r_s_b); in blend_a64_mask_sx_sy_w4_sse4_1() 288 const __m128i v_r1_s_w = _mm_cvtepu8_epi16(_mm_srli_si128(v_r_s_b, 8)); in blend_a64_mask_sx_sy_w4_sse4_1() 320 const __m128i v_r0_s_w = _mm_cvtepu8_epi16(v_r_s_b); in blend_a64_mask_sx_sy_w8_sse4_1() 321 const __m128i v_r1_s_w = _mm_cvtepu8_epi16(_mm_srli_si128(v_r_s_b, 8)); in blend_a64_mask_sx_sy_w8_sse4_1() 439 const __m128i v_m0_w = _mm_cvtepu8_epi16(v_m0_b); in blend_a64_mask_bn_w4_sse4_1() 482 const __m128i v_m0_w = _mm_cvtepu8_epi16(v_m0_b); in blend_a64_mask_bn_w8n_sse4_1() 626 const __m128i v_m0_w = _mm_cvtepu8_epi16(v_a_b); in blend_a64_mask_bn_sy_w4_sse4_1() 674 const __m128i v_m0_w = _mm_cvtepu8_epi16(v_a_b); in blend_a64_mask_bn_sy_w8n_sse4_1() 913 const __m128i m0 = _mm_cvtepu8_epi16(m); in lowbd_blend_a64_d16_mask_subw0_subh0_w16_sse4_1() 914 const __m128i m1 = _mm_cvtepu8_epi16(_mm_srli_si128(m, 8)); in lowbd_blend_a64_d16_mask_subw0_subh0_w16_sse4_1() [all …]
|
/external/webp/src/dsp/ |
D | enc_sse41.c | 98 tmp_0 = _mm_cvtepu8_epi16(inAB_0); in TTransform_SSE41() 99 tmp_1 = _mm_cvtepu8_epi16(inAB_1); in TTransform_SSE41() 100 tmp_2 = _mm_cvtepu8_epi16(inAB_2); in TTransform_SSE41() 101 tmp_3 = _mm_cvtepu8_epi16(inAB_3); in TTransform_SSE41()
|
/external/libavc/encoder/x86/ |
D | ime_distortion_metrics_sse42.c | 1748 src_r0 = _mm_cvtepu8_epi16(src_r0); in ime_compute_satqd_16x16_lumainter_sse42() 1749 src_r1 = _mm_cvtepu8_epi16(src_r1); in ime_compute_satqd_16x16_lumainter_sse42() 1750 src_r2 = _mm_cvtepu8_epi16(src_r2); in ime_compute_satqd_16x16_lumainter_sse42() 1751 src_r3 = _mm_cvtepu8_epi16(src_r3); in ime_compute_satqd_16x16_lumainter_sse42() 1758 est_r0 = _mm_cvtepu8_epi16(est_r0); in ime_compute_satqd_16x16_lumainter_sse42() 1759 est_r1 = _mm_cvtepu8_epi16(est_r1); in ime_compute_satqd_16x16_lumainter_sse42() 1760 est_r2 = _mm_cvtepu8_epi16(est_r2); in ime_compute_satqd_16x16_lumainter_sse42() 1761 est_r3 = _mm_cvtepu8_epi16(est_r3); in ime_compute_satqd_16x16_lumainter_sse42() 1847 src_r0 = _mm_cvtepu8_epi16(src_r0); in ime_compute_satqd_16x16_lumainter_sse42() 1848 src_r1 = _mm_cvtepu8_epi16(src_r1); in ime_compute_satqd_16x16_lumainter_sse42() [all …]
|