/external/libaom/libaom/av1/common/arm/ |
D | warp_plane_neon.c | 202 uint8x8_t src_1, int16x4_t *res) { in convolve() argument 214 pix_1 = vreinterpretq_s16_u16(vmovl_u8(src_1)); in convolve() 220 static INLINE void horizontal_filter_neon(uint8x16_t src_1, uint8x16_t src_2, in horizontal_filter_neon() argument 237 uint8x16_t tmp_0 = vandq_u8(src_1, mask); in horizontal_filter_neon() 245 src_1 = vaddq_u8(tmp_0, tmp_2); in horizontal_filter_neon() 248 src_1_low = vget_low_u8(src_1); in horizontal_filter_neon() 250 src_3_low = vget_low_u8(vextq_u8(src_1, src_1, 4)); in horizontal_filter_neon() 252 src_5_low = vget_low_u8(vextq_u8(src_1, src_1, 2)); in horizontal_filter_neon() 253 src_6_low = vget_low_u8(vextq_u8(src_1, src_1, 6)); in horizontal_filter_neon() 312 int16x4_t src_0, src_1, fltr_0, fltr_1; in vertical_filter_neon() local [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | convolve_sse2.h | 35 static INLINE __m128i mm_madd_add_epi8_sse2(const __m128i *const src_1, in mm_madd_add_epi8_sse2() argument 39 const __m128i src_1_half = _mm_unpacklo_epi8(*src_1, _mm_setzero_si128()); in mm_madd_add_epi8_sse2() 49 static INLINE __m128i mm_madd_add_epi16_sse2(const __m128i *const src_1, in mm_madd_add_epi16_sse2() argument 53 const __m128i madd_1 = _mm_madd_epi16(*src_1, *ker_1); in mm_madd_add_epi16_sse2() 59 const __m128i *const src_1, in mm_madd_packs_epi16_sse2() argument 62 const __m128i madd_2 = _mm_madd_epi16(*src_1, *ker); in mm_madd_packs_epi16_sse2() 67 static INLINE __m128i mm_zip_epi32_sse2(const __m128i *const src_1, in mm_zip_epi32_sse2() argument 69 const __m128i tmp_1 = _mm_unpacklo_epi32(*src_1, *src_2); in mm_zip_epi32_sse2() 70 const __m128i tmp_2 = _mm_unpackhi_epi32(*src_1, *src_2); in mm_zip_epi32_sse2()
|
D | variance_avx2.c | 265 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + sstep)); in spv32_half_zero() local 266 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_half_zero() 316 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + 1)); in spv32_x4_y4() local 317 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_x4_y4() 354 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + sstep)); in spv32_bilin_zero() local 355 exp_src_lo = _mm256_unpacklo_epi8(src_0, src_1); in spv32_bilin_zero() 356 exp_src_hi = _mm256_unpackhi_epi8(src_0, src_1); in spv32_bilin_zero() 409 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + 1)); in spv32_x4_yb() local 410 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_x4_yb() 454 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + 1)); in spv32_xb_y4() local [all …]
|
D | convolve_avx2.h | 152 const __m256i *const src_1, in mm256_madd_add_epi32() argument 156 const __m256i tmp_1 = _mm256_madd_epi16(*src_1, *ker_1); in mm256_madd_add_epi32()
|
/external/libhevc/encoder/arm/ |
D | ihevce_copy_neon.c | 143 uint8x8_t src_0, src_1; in copy_2d_neon() local 149 src_1 = vld1_u8(pu1_src + src_strd); in copy_2d_neon() 150 vst1_u8(pu1_dst + dst_strd, src_1); in copy_2d_neon() 160 uint8x16_t src_0, src_1; in copy_2d_neon() local 166 src_1 = vld1q_u8(pu1_src + src_strd); in copy_2d_neon() 167 vst1q_u8(pu1_dst + dst_strd, src_1); in copy_2d_neon() 176 uint8x16_t src_0, src_1; in copy_2d_neon() local 181 src_1 = vld1q_u8(pu1_src + 16); in copy_2d_neon() 182 vst1q_u8(pu1_dst + 16, src_1); in copy_2d_neon() 198 uint8x16_t src_0, src_1, src_2, src_3; in copy_2d_neon() local [all …]
|
D | ihevce_ssd_calculator_neon.c | 136 uint8x16_t src_0, pred_0, src_1, pred_1, abs_0, abs_1; in ihevce_1x32_ssd_computer_neon() local 143 src_1 = vld1q_u8(pu1_src + 16); in ihevce_1x32_ssd_computer_neon() 150 src_1 = vld2q_u8(pu1_src + 32).val[0]; in ihevce_1x32_ssd_computer_neon() 154 abs_1 = vabdq_u8(src_1, pred_1); in ihevce_1x32_ssd_computer_neon() 174 uint8x16_t src_0, src_1, src_2, src_3; in ihevce_1x64_ssd_computer_neon() local 184 src_1 = vld1q_u8(pu1_src + 16); in ihevce_1x64_ssd_computer_neon() 195 src_1 = vld2q_u8(pu1_src + 32).val[0]; in ihevce_1x64_ssd_computer_neon() 203 abs_1 = vabdq_u8(src_1, pred_1); in ihevce_1x64_ssd_computer_neon()
|
D | ihevce_sad_compute_neon.c | 145 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_32xn_sad_computer_neon() local 150 abs_1 = vabal_u8(abs_1, vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_32xn_sad_computer_neon() 151 abs_1 = vabal_u8(abs_1, vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_32xn_sad_computer_neon() 179 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_64xn_sad_computer_neon() local 188 abs_0 = vabal_u8(abs_0, vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_64xn_sad_computer_neon() 189 abs_0 = vabal_u8(abs_0, vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_64xn_sad_computer_neon()
|
D | ihevce_ssd_and_sad_calculator_neon.c | 187 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_ssd_and_sad_calculator_neon() local 201 abs_l = vabd_u8(vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon() 202 abs_h = vabd_u8(vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon() 246 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_ssd_and_sad_calculator_neon() local 264 abs_l = vabd_u8(vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon() 265 abs_h = vabd_u8(vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | subpel_variance_neon.c | 37 const uint8x8_t src_1 = in var_filter_block2d_bil_w4() local 40 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w4() 60 const uint8x8_t src_1 = vld1_u8(&src_ptr[pixel_step]); in var_filter_block2d_bil_w8() local 62 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8() 84 const uint8x16_t src_1 = vld1q_u8(&src_ptr[j + pixel_step]); in var_filter_block2d_bil_w16() local 86 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16() 89 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | subpel_variance_neon.c | 35 const uint8x8_t src_1 = vld1_u8(&src_ptr[pixel_step]); in var_filter_block2d_bil_w8() local 37 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8() 59 const uint8x16_t src_1 = vld1q_u8(&src_ptr[j + pixel_step]); in var_filter_block2d_bil_w16() local 61 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16() 64 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
|
D | blend_a64_mask_neon.c | 22 static INLINE void blend8x1(int16x8_t mask, int16x8_t src_0, int16x8_t src_1, in blend8x1() argument 29 vmlal_s16(im_res_low, vget_low_s16(max_minus_mask), vget_low_s16(src_1)); in blend8x1() 33 vget_high_s16(src_1)); in blend8x1()
|
/external/libaom/libaom/av1/common/x86/ |
D | highbd_warp_plane_sse4.c | 110 const __m128i src_1 = *src; in highbd_filter_src_pixels() local 116 const __m128i res_0 = _mm_madd_epi16(src_1, coeff[0]); in highbd_filter_src_pixels() 118 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 4), coeff[2]); in highbd_filter_src_pixels() 120 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 8), coeff[4]); in highbd_filter_src_pixels() 122 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 12), coeff[6]); in highbd_filter_src_pixels() 130 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 2), coeff[1]); in highbd_filter_src_pixels() 132 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 6), coeff[3]); in highbd_filter_src_pixels() 134 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 10), coeff[5]); in highbd_filter_src_pixels() 136 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 14), coeff[7]); in highbd_filter_src_pixels() 488 const __m128i src_1 = _mm_unpackhi_epi16(src[0], src[1]); in av1_highbd_warp_affine_sse4_1() local [all …]
|
D | wiener_convolve_sse2.c | 88 const __m128i src_1 = _mm_unpacklo_epi8(_mm_srli_si128(data, 1), zero); in av1_wiener_convolve_add_src_sse2() local 89 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_wiener_convolve_add_src_sse2() 161 const __m128i src_1 = in av1_wiener_convolve_add_src_sse2() local 174 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_wiener_convolve_add_src_sse2()
|
D | highbd_wiener_convolve_avx2.c | 96 const __m256i src_1 = yy_loadu_256(src_ij + 1); in av1_highbd_wiener_convolve_add_src_avx2() local 106 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_highbd_wiener_convolve_add_src_avx2() 205 const __m256i src_1 = _mm256_unpackhi_epi16(data_0, data_1); in av1_highbd_wiener_convolve_add_src_avx2() local 210 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_highbd_wiener_convolve_add_src_avx2()
|
D | wiener_convolve_avx2.c | 104 const __m256i src_1 = _mm256_cvtepu8_epi16(data_1); in av1_wiener_convolve_add_src_avx2() local 114 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_wiener_convolve_add_src_avx2() 211 const __m256i src_1 = _mm256_unpackhi_epi16(data_0, data_1); in av1_wiener_convolve_add_src_avx2() local 216 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_wiener_convolve_add_src_avx2()
|
D | jnt_convolve_ssse3.c | 102 const __m128i src_1 = _mm_alignr_epi8(src_hi, src_lo, 2); in av1_dist_wtd_convolve_2d_ssse3() local 103 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_ssse3() 174 const __m128i src_1 = in av1_dist_wtd_convolve_2d_ssse3() local 187 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_ssse3()
|
D | convolve_2d_avx2.c | 120 __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); in av1_convolve_2d_sr_avx2() local 124 s[0] = _mm256_unpacklo_epi16(src_0, src_1); in av1_convolve_2d_sr_avx2() 126 s[3] = _mm256_unpackhi_epi16(src_0, src_1); in av1_convolve_2d_sr_avx2()
|
D | convolve_2d_sse2.c | 90 const __m128i src_1 = _mm_unpacklo_epi8(_mm_srli_si128(data, 1), zero); in av1_convolve_2d_sr_sse2() local 91 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_convolve_2d_sr_sse2() 166 const __m128i src_1 = in av1_convolve_2d_sr_sse2() local 179 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_convolve_2d_sr_sse2()
|
D | highbd_wiener_convolve_ssse3.c | 163 const __m128i src_1 = in av1_highbd_wiener_convolve_add_src_ssse3() local 176 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_highbd_wiener_convolve_add_src_ssse3()
|
D | jnt_convolve_sse2.c | 480 const __m128i src_1 = _mm_or_si128(temp_hi, temp_lo); in av1_dist_wtd_convolve_2d_sse2() local 481 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_sse2() 558 const __m128i src_1 = in av1_dist_wtd_convolve_2d_sse2() local 571 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_sse2()
|
D | selfguided_avx2.c | 660 __m128i src_0, src_1; in apply_selfguided_restoration_avx2() local 663 src_1 = xx_loadu_128(CONVERT_TO_SHORTPTR(dat8ij + 8)); in apply_selfguided_restoration_avx2() 665 ep_1 = _mm256_cvtepu16_epi32(src_1); in apply_selfguided_restoration_avx2()
|
D | highbd_convolve_2d_sse4.c | 326 const __m128i src_1 = in av1_highbd_dist_wtd_convolve_2d_sse4_1() local 339 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
|
/external/tensorflow/tensorflow/core/util/ |
D | tensor_slice_set_test.cc | 52 const float src_1[] = {0, 1, 2, 3, 4, 5, 6, 7, 8, 9}; in TEST() local 54 TF_CHECK_OK(tss.Register(slice_1, "", src_1)); in TEST()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | convolve_avx2.h | 64 __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \ 71 s[0] = _mm256_unpacklo_epi16(src_0, src_1); \ 75 s[4] = _mm256_unpackhi_epi16(src_0, src_1); \
|
/external/webp/src/dsp/ |
D | enc_sse2.c | 309 const __m128i src_1 = _mm_unpacklo_epi16(src2, src3); in FTransform_SSE2() local 323 const __m128i src_1_16b = _mm_unpacklo_epi8(src_1, zero); in FTransform_SSE2() 349 const __m128i src_1 = _mm_unpacklo_epi8(src1, zero); in FTransform2_SSE2() local 363 const __m128i diff1 = _mm_sub_epi16(src_1, ref_1); in FTransform2_SSE2()
|