Home
last modified time | relevance | path

Searched refs:src_1 (Results 1 – 25 of 26) sorted by relevance

12

/external/libaom/libaom/av1/common/arm/
Dwarp_plane_neon.c202 uint8x8_t src_1, int16x4_t *res) { in convolve() argument
214 pix_1 = vreinterpretq_s16_u16(vmovl_u8(src_1)); in convolve()
220 static INLINE void horizontal_filter_neon(uint8x16_t src_1, uint8x16_t src_2, in horizontal_filter_neon() argument
237 uint8x16_t tmp_0 = vandq_u8(src_1, mask); in horizontal_filter_neon()
245 src_1 = vaddq_u8(tmp_0, tmp_2); in horizontal_filter_neon()
248 src_1_low = vget_low_u8(src_1); in horizontal_filter_neon()
250 src_3_low = vget_low_u8(vextq_u8(src_1, src_1, 4)); in horizontal_filter_neon()
252 src_5_low = vget_low_u8(vextq_u8(src_1, src_1, 2)); in horizontal_filter_neon()
253 src_6_low = vget_low_u8(vextq_u8(src_1, src_1, 6)); in horizontal_filter_neon()
312 int16x4_t src_0, src_1, fltr_0, fltr_1; in vertical_filter_neon() local
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dconvolve_sse2.h35 static INLINE __m128i mm_madd_add_epi8_sse2(const __m128i *const src_1, in mm_madd_add_epi8_sse2() argument
39 const __m128i src_1_half = _mm_unpacklo_epi8(*src_1, _mm_setzero_si128()); in mm_madd_add_epi8_sse2()
49 static INLINE __m128i mm_madd_add_epi16_sse2(const __m128i *const src_1, in mm_madd_add_epi16_sse2() argument
53 const __m128i madd_1 = _mm_madd_epi16(*src_1, *ker_1); in mm_madd_add_epi16_sse2()
59 const __m128i *const src_1, in mm_madd_packs_epi16_sse2() argument
62 const __m128i madd_2 = _mm_madd_epi16(*src_1, *ker); in mm_madd_packs_epi16_sse2()
67 static INLINE __m128i mm_zip_epi32_sse2(const __m128i *const src_1, in mm_zip_epi32_sse2() argument
69 const __m128i tmp_1 = _mm_unpacklo_epi32(*src_1, *src_2); in mm_zip_epi32_sse2()
70 const __m128i tmp_2 = _mm_unpackhi_epi32(*src_1, *src_2); in mm_zip_epi32_sse2()
Dvariance_avx2.c265 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + sstep)); in spv32_half_zero() local
266 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_half_zero()
316 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + 1)); in spv32_x4_y4() local
317 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_x4_y4()
354 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + sstep)); in spv32_bilin_zero() local
355 exp_src_lo = _mm256_unpacklo_epi8(src_0, src_1); in spv32_bilin_zero()
356 exp_src_hi = _mm256_unpackhi_epi8(src_0, src_1); in spv32_bilin_zero()
409 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + 1)); in spv32_x4_yb() local
410 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_x4_yb()
454 const __m256i src_1 = _mm256_loadu_si256((__m256i const *)(src + 1)); in spv32_xb_y4() local
[all …]
Dconvolve_avx2.h152 const __m256i *const src_1, in mm256_madd_add_epi32() argument
156 const __m256i tmp_1 = _mm256_madd_epi16(*src_1, *ker_1); in mm256_madd_add_epi32()
/external/libhevc/encoder/arm/
Dihevce_copy_neon.c143 uint8x8_t src_0, src_1; in copy_2d_neon() local
149 src_1 = vld1_u8(pu1_src + src_strd); in copy_2d_neon()
150 vst1_u8(pu1_dst + dst_strd, src_1); in copy_2d_neon()
160 uint8x16_t src_0, src_1; in copy_2d_neon() local
166 src_1 = vld1q_u8(pu1_src + src_strd); in copy_2d_neon()
167 vst1q_u8(pu1_dst + dst_strd, src_1); in copy_2d_neon()
176 uint8x16_t src_0, src_1; in copy_2d_neon() local
181 src_1 = vld1q_u8(pu1_src + 16); in copy_2d_neon()
182 vst1q_u8(pu1_dst + 16, src_1); in copy_2d_neon()
198 uint8x16_t src_0, src_1, src_2, src_3; in copy_2d_neon() local
[all …]
Dihevce_ssd_calculator_neon.c136 uint8x16_t src_0, pred_0, src_1, pred_1, abs_0, abs_1; in ihevce_1x32_ssd_computer_neon() local
143 src_1 = vld1q_u8(pu1_src + 16); in ihevce_1x32_ssd_computer_neon()
150 src_1 = vld2q_u8(pu1_src + 32).val[0]; in ihevce_1x32_ssd_computer_neon()
154 abs_1 = vabdq_u8(src_1, pred_1); in ihevce_1x32_ssd_computer_neon()
174 uint8x16_t src_0, src_1, src_2, src_3; in ihevce_1x64_ssd_computer_neon() local
184 src_1 = vld1q_u8(pu1_src + 16); in ihevce_1x64_ssd_computer_neon()
195 src_1 = vld2q_u8(pu1_src + 32).val[0]; in ihevce_1x64_ssd_computer_neon()
203 abs_1 = vabdq_u8(src_1, pred_1); in ihevce_1x64_ssd_computer_neon()
Dihevce_sad_compute_neon.c145 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_32xn_sad_computer_neon() local
150 abs_1 = vabal_u8(abs_1, vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_32xn_sad_computer_neon()
151 abs_1 = vabal_u8(abs_1, vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_32xn_sad_computer_neon()
179 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_64xn_sad_computer_neon() local
188 abs_0 = vabal_u8(abs_0, vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_64xn_sad_computer_neon()
189 abs_0 = vabal_u8(abs_0, vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_64xn_sad_computer_neon()
Dihevce_ssd_and_sad_calculator_neon.c187 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_ssd_and_sad_calculator_neon() local
201 abs_l = vabd_u8(vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon()
202 abs_h = vabd_u8(vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon()
246 const uint8x16_t src_1 = vld1q_u8(pu1_src + 16); in ihevce_ssd_and_sad_calculator_neon() local
264 abs_l = vabd_u8(vget_low_u8(src_1), vget_low_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon()
265 abs_h = vabd_u8(vget_high_u8(src_1), vget_high_u8(pred_1)); in ihevce_ssd_and_sad_calculator_neon()
/external/libvpx/libvpx/vpx_dsp/arm/
Dsubpel_variance_neon.c37 const uint8x8_t src_1 = in var_filter_block2d_bil_w4() local
40 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w4()
60 const uint8x8_t src_1 = vld1_u8(&src_ptr[pixel_step]); in var_filter_block2d_bil_w8() local
62 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8()
84 const uint8x16_t src_1 = vld1q_u8(&src_ptr[j + pixel_step]); in var_filter_block2d_bil_w16() local
86 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16()
89 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
/external/libaom/libaom/aom_dsp/arm/
Dsubpel_variance_neon.c35 const uint8x8_t src_1 = vld1_u8(&src_ptr[pixel_step]); in var_filter_block2d_bil_w8() local
37 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8()
59 const uint8x16_t src_1 = vld1q_u8(&src_ptr[j + pixel_step]); in var_filter_block2d_bil_w16() local
61 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16()
64 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
Dblend_a64_mask_neon.c22 static INLINE void blend8x1(int16x8_t mask, int16x8_t src_0, int16x8_t src_1, in blend8x1() argument
29 vmlal_s16(im_res_low, vget_low_s16(max_minus_mask), vget_low_s16(src_1)); in blend8x1()
33 vget_high_s16(src_1)); in blend8x1()
/external/libaom/libaom/av1/common/x86/
Dhighbd_warp_plane_sse4.c110 const __m128i src_1 = *src; in highbd_filter_src_pixels() local
116 const __m128i res_0 = _mm_madd_epi16(src_1, coeff[0]); in highbd_filter_src_pixels()
118 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 4), coeff[2]); in highbd_filter_src_pixels()
120 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 8), coeff[4]); in highbd_filter_src_pixels()
122 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 12), coeff[6]); in highbd_filter_src_pixels()
130 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 2), coeff[1]); in highbd_filter_src_pixels()
132 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 6), coeff[3]); in highbd_filter_src_pixels()
134 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 10), coeff[5]); in highbd_filter_src_pixels()
136 _mm_madd_epi16(_mm_alignr_epi8(src2_1, src_1, 14), coeff[7]); in highbd_filter_src_pixels()
488 const __m128i src_1 = _mm_unpackhi_epi16(src[0], src[1]); in av1_highbd_warp_affine_sse4_1() local
[all …]
Dwiener_convolve_sse2.c88 const __m128i src_1 = _mm_unpacklo_epi8(_mm_srli_si128(data, 1), zero); in av1_wiener_convolve_add_src_sse2() local
89 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_wiener_convolve_add_src_sse2()
161 const __m128i src_1 = in av1_wiener_convolve_add_src_sse2() local
174 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_wiener_convolve_add_src_sse2()
Dhighbd_wiener_convolve_avx2.c96 const __m256i src_1 = yy_loadu_256(src_ij + 1); in av1_highbd_wiener_convolve_add_src_avx2() local
106 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_highbd_wiener_convolve_add_src_avx2()
205 const __m256i src_1 = _mm256_unpackhi_epi16(data_0, data_1); in av1_highbd_wiener_convolve_add_src_avx2() local
210 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_highbd_wiener_convolve_add_src_avx2()
Dwiener_convolve_avx2.c104 const __m256i src_1 = _mm256_cvtepu8_epi16(data_1); in av1_wiener_convolve_add_src_avx2() local
114 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_wiener_convolve_add_src_avx2()
211 const __m256i src_1 = _mm256_unpackhi_epi16(data_0, data_1); in av1_wiener_convolve_add_src_avx2() local
216 const __m256i res_1 = _mm256_madd_epi16(src_1, coeffs_01); in av1_wiener_convolve_add_src_avx2()
Djnt_convolve_ssse3.c102 const __m128i src_1 = _mm_alignr_epi8(src_hi, src_lo, 2); in av1_dist_wtd_convolve_2d_ssse3() local
103 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_ssse3()
174 const __m128i src_1 = in av1_dist_wtd_convolve_2d_ssse3() local
187 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_ssse3()
Dconvolve_2d_avx2.c120 __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); in av1_convolve_2d_sr_avx2() local
124 s[0] = _mm256_unpacklo_epi16(src_0, src_1); in av1_convolve_2d_sr_avx2()
126 s[3] = _mm256_unpackhi_epi16(src_0, src_1); in av1_convolve_2d_sr_avx2()
Dconvolve_2d_sse2.c90 const __m128i src_1 = _mm_unpacklo_epi8(_mm_srli_si128(data, 1), zero); in av1_convolve_2d_sr_sse2() local
91 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_convolve_2d_sr_sse2()
166 const __m128i src_1 = in av1_convolve_2d_sr_sse2() local
179 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_convolve_2d_sr_sse2()
Dhighbd_wiener_convolve_ssse3.c163 const __m128i src_1 = in av1_highbd_wiener_convolve_add_src_ssse3() local
176 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_highbd_wiener_convolve_add_src_ssse3()
Djnt_convolve_sse2.c480 const __m128i src_1 = _mm_or_si128(temp_hi, temp_lo); in av1_dist_wtd_convolve_2d_sse2() local
481 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_sse2()
558 const __m128i src_1 = in av1_dist_wtd_convolve_2d_sse2() local
571 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_dist_wtd_convolve_2d_sse2()
Dselfguided_avx2.c660 __m128i src_0, src_1; in apply_selfguided_restoration_avx2() local
663 src_1 = xx_loadu_128(CONVERT_TO_SHORTPTR(dat8ij + 8)); in apply_selfguided_restoration_avx2()
665 ep_1 = _mm256_cvtepu16_epi32(src_1); in apply_selfguided_restoration_avx2()
Dhighbd_convolve_2d_sse4.c326 const __m128i src_1 = in av1_highbd_dist_wtd_convolve_2d_sse4_1() local
339 const __m128i res_1 = _mm_madd_epi16(src_1, coeff_01); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
/external/tensorflow/tensorflow/core/util/
Dtensor_slice_set_test.cc52 const float src_1[] = {0, 1, 2, 3, 4, 5, 6, 7, 8, 9}; in TEST() local
54 TF_CHECK_OK(tss.Register(slice_1, "", src_1)); in TEST()
/external/libaom/libaom/aom_dsp/x86/
Dconvolve_avx2.h64 __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \
71 s[0] = _mm256_unpacklo_epi16(src_0, src_1); \
75 s[4] = _mm256_unpackhi_epi16(src_0, src_1); \
/external/webp/src/dsp/
Denc_sse2.c309 const __m128i src_1 = _mm_unpacklo_epi16(src2, src3); in FTransform_SSE2() local
323 const __m128i src_1_16b = _mm_unpacklo_epi8(src_1, zero); in FTransform_SSE2()
349 const __m128i src_1 = _mm_unpacklo_epi8(src1, zero); in FTransform2_SSE2() local
363 const __m128i diff1 = _mm_sub_epi16(src_1, ref_1); in FTransform2_SSE2()

12