Home
last modified time | relevance | path

Searched refs:uint16x8_t (Results 1 – 25 of 138) sorted by relevance

123456

/external/libvpx/libvpx/vpx_dsp/arm/
Dhighbd_loopfilter_neon.c17 const uint8_t *thresh, uint16x8_t *blimit_vec, in load_thresh()
18 uint16x8_t *limit_vec, uint16x8_t *thresh_vec, in load_thresh()
35 static INLINE uint32_t calc_flat_status(const uint16x8_t flat) { in calc_flat_status()
42 static INLINE uint16x8_t
43 filter_hev_mask4(const uint16x8_t limit, const uint16x8_t blimit, in filter_hev_mask4()
44 const uint16x8_t thresh, const uint16x8_t p3, in filter_hev_mask4()
45 const uint16x8_t p2, const uint16x8_t p1, const uint16x8_t p0, in filter_hev_mask4()
46 const uint16x8_t q0, const uint16x8_t q1, const uint16x8_t q2, in filter_hev_mask4()
47 const uint16x8_t q3, uint16x8_t *hev, uint16x8_t *mask) { in filter_hev_mask4()
48 uint16x8_t max, t0, t1; in filter_hev_mask4()
[all …]
Dhighbd_intrapred_neon.c83 const uint16x8_t ref_u16 = vld1q_u16(ref); in dc_sum_8()
91 const uint16x8_t dc_dup = vdupq_lane_u16(dc, 0); in dc_store_8x8()
101 const uint16x8_t above_u16 = vld1q_u16(above); in vpx_highbd_dc_predictor_8x8_neon()
102 const uint16x8_t left_u16 = vld1q_u16(left); in vpx_highbd_dc_predictor_8x8_neon()
103 const uint16x8_t p0 = vaddq_u16(above_u16, left_u16); in vpx_highbd_dc_predictor_8x8_neon()
147 const uint16x8_t p0 = vaddq_u16(ref_u16.val[0], ref_u16.val[1]); in dc_sum_16()
168 const uint16x8_t pa = vaddq_u16(a.val[0], a.val[1]); in vpx_highbd_dc_predictor_16x16_neon()
169 const uint16x8_t pl = vaddq_u16(l.val[0], l.val[1]); in vpx_highbd_dc_predictor_16x16_neon()
170 const uint16x8_t pal0 = vaddq_u16(pa, pl); in vpx_highbd_dc_predictor_16x16_neon()
215 const uint16x8_t p0 = vaddq_u16(r.val[0], r.val[1]); in dc_sum_32()
[all …]
Dhighbd_idct32x32_add_neon.c21 const uint16x8_t a0 = vld1q_u16(*dest); in highbd_idct32x32_1_add_pos_kernel()
22 const uint16x8_t a1 = vld1q_u16(*dest + 8); in highbd_idct32x32_1_add_pos_kernel()
23 const uint16x8_t a2 = vld1q_u16(*dest + 16); in highbd_idct32x32_1_add_pos_kernel()
24 const uint16x8_t a3 = vld1q_u16(*dest + 24); in highbd_idct32x32_1_add_pos_kernel()
43 const uint16x8_t a0 = vld1q_u16(*dest); in highbd_idct32x32_1_add_neg_kernel()
44 const uint16x8_t a1 = vld1q_u16(*dest + 8); in highbd_idct32x32_1_add_neg_kernel()
45 const uint16x8_t a2 = vld1q_u16(*dest + 16); in highbd_idct32x32_1_add_neg_kernel()
46 const uint16x8_t a3 = vld1q_u16(*dest + 24); in highbd_idct32x32_1_add_neg_kernel()
51 const uint16x8_t c0 = vqshluq_n_s16(b0, 0); in highbd_idct32x32_1_add_neg_kernel()
52 const uint16x8_t c1 = vqshluq_n_s16(b1, 0); in highbd_idct32x32_1_add_neg_kernel()
[all …]
Dsad_neon.c24 uint16x8_t abs = vabdl_u8(vget_low_u8(src_u8), vget_low_u8(ref_u8)); in vpx_sad4x4_neon()
36 uint16x8_t abs = vabdl_u8(vget_low_u8(src_u8), vget_low_u8(avg)); in vpx_sad4x4_avg_neon()
44 uint16x8_t abs = vdupq_n_u16(0); in vpx_sad4x8_neon()
61 uint16x8_t abs = vdupq_n_u16(0); in vpx_sad4x8_avg_neon()
77 static INLINE uint16x8_t sad8x(const uint8_t *src_ptr, int src_stride, in sad8x()
81 uint16x8_t abs = vdupq_n_u16(0); in sad8x()
93 static INLINE uint16x8_t sad8x_avg(const uint8_t *src_ptr, int src_stride, in sad8x_avg()
98 uint16x8_t abs = vdupq_n_u16(0); in sad8x_avg()
116 const uint16x8_t abs = sad8x(src_ptr, src_stride, ref_ptr, ref_stride, n); \
123 const uint16x8_t abs = \
[all …]
Dsubtract_neon.c32 const uint16x8_t d0 = vsubl_u8(vget_low_u8(s0), vget_low_u8(p0)); in vpx_subtract_block_neon()
33 const uint16x8_t d1 = vsubl_u8(vget_high_u8(s0), vget_high_u8(p0)); in vpx_subtract_block_neon()
34 const uint16x8_t d2 = vsubl_u8(vget_low_u8(s1), vget_low_u8(p1)); in vpx_subtract_block_neon()
35 const uint16x8_t d3 = vsubl_u8(vget_high_u8(s1), vget_high_u8(p1)); in vpx_subtract_block_neon()
49 const uint16x8_t d0 = vsubl_u8(vget_low_u8(s), vget_low_u8(p)); in vpx_subtract_block_neon()
50 const uint16x8_t d1 = vsubl_u8(vget_high_u8(s), vget_high_u8(p)); in vpx_subtract_block_neon()
61 const uint16x8_t v_diff = vsubl_u8(s, p); in vpx_subtract_block_neon()
72 const uint16x8_t d = vsubl_u8(s, p); in vpx_subtract_block_neon()
Dsubpel_variance_neon.c39 const uint16x8_t a = vmull_u8(src_0, f0); in var_filter_block2d_bil_w4()
40 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w4()
61 const uint16x8_t a = vmull_u8(src_0, f0); in var_filter_block2d_bil_w8()
62 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8()
85 const uint16x8_t a = vmull_u8(vget_low_u8(src_0), f0); in var_filter_block2d_bil_w16()
86 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16()
88 const uint16x8_t c = vmull_u8(vget_high_u8(src_0), f0); in var_filter_block2d_bil_w16()
89 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
Dsad4d_neon.c36 uint16x8_t abs[2] = { vdupq_n_u16(0), vdupq_n_u16(0) }; in sad4x_4d()
75 static INLINE void sad_512_pel_final_neon(const uint16x8_t *sum /*[4]*/, in sad_512_pel_final_neon()
88 static INLINE void sad_1024_pel_final_neon(const uint16x8_t *sum /*[4]*/, in sad_1024_pel_final_neon()
102 static INLINE void sad_2048_pel_final_neon(const uint16x8_t *sum /*[4]*/, in sad_2048_pel_final_neon()
118 static INLINE void sad_4096_pel_final_neon(const uint16x8_t *sum /*[8]*/, in sad_4096_pel_final_neon()
147 uint16x8_t sum[4] = { vdupq_n_u16(0), vdupq_n_u16(0), vdupq_n_u16(0), in sad8x_4d()
184 uint16x8_t *const sum) { in sad16_neon()
196 uint16x8_t sum[4] = { vdupq_n_u16(0), vdupq_n_u16(0), vdupq_n_u16(0), in sad16x_4d()
233 const int height, uint16x8_t *const sum) { in sad32x_4d()
266 uint16x8_t sum[4]; in vpx_sad32x16x4d_neon()
[all …]
/external/libaom/libaom/aom_dsp/arm/
Dsad4d_neon.c19 static INLINE unsigned int horizontal_long_add_16x8(const uint16x8_t vec_lo, in horizontal_long_add_16x8()
20 const uint16x8_t vec_hi) { in horizontal_long_add_16x8()
39 uint16x8_t *vec_sum_ref_lo, in sad_neon_64()
40 uint16x8_t *vec_sum_ref_hi) { in sad_neon_64()
68 uint16x8_t *vec_sum_ref_lo, in sad_neon_32()
69 uint16x8_t *vec_sum_ref_hi) { in sad_neon_32()
87 uint16x8_t vec_sum_ref0_lo = vdupq_n_u16(0); in aom_sad64x64x4d_neon()
88 uint16x8_t vec_sum_ref0_hi = vdupq_n_u16(0); in aom_sad64x64x4d_neon()
89 uint16x8_t vec_sum_ref1_lo = vdupq_n_u16(0); in aom_sad64x64x4d_neon()
90 uint16x8_t vec_sum_ref1_hi = vdupq_n_u16(0); in aom_sad64x64x4d_neon()
[all …]
Dsad_neon.c21 uint16x8_t q12; in aom_sad8x16_neon()
52 uint16x8_t q12; in aom_sad4x4_neon()
80 uint16x8_t q12, q13; in aom_sad16x8_neon()
111 static INLINE unsigned int horizontal_long_add_16x8(const uint16x8_t vec_lo, in horizontal_long_add_16x8()
112 const uint16x8_t vec_hi) { in horizontal_long_add_16x8()
123 static INLINE unsigned int horizontal_add_16x8(const uint16x8_t vec_16x8) { in horizontal_add_16x8()
134 uint16x8_t vec_accum_lo = vdupq_n_u16(0); in aom_sad64x64_neon()
135 uint16x8_t vec_accum_hi = vdupq_n_u16(0); in aom_sad64x64_neon()
170 uint16x8_t vec_accum_lo = vdupq_n_u16(0); in aom_sad32x32_neon()
171 uint16x8_t vec_accum_hi = vdupq_n_u16(0); in aom_sad32x32_neon()
[all …]
Dsubtract_neon.c31 const uint16x8_t v_diff_lo_00 = in aom_subtract_block_neon()
33 const uint16x8_t v_diff_hi_00 = in aom_subtract_block_neon()
35 const uint16x8_t v_diff_lo_16 = in aom_subtract_block_neon()
37 const uint16x8_t v_diff_hi_16 = in aom_subtract_block_neon()
52 const uint16x8_t v_diff_lo = in aom_subtract_block_neon()
54 const uint16x8_t v_diff_hi = in aom_subtract_block_neon()
66 const uint16x8_t v_diff = vsubl_u8(v_src, v_pred); in aom_subtract_block_neon()
Dintrapred_neon.c25 uint16x8_t sum_top; in dc_4x4()
26 uint16x8_t sum_left; in dc_4x4()
44 const uint16x8_t sum = vaddq_u16(sum_left, sum_top); in dc_4x4()
93 uint16x8_t sum_top; in dc_8x8()
94 uint16x8_t sum_left; in dc_8x8()
114 const uint16x8_t sum = vaddq_u16(sum_left, sum_top); in dc_8x8()
164 uint16x8_t sum_top; in dc_16x16()
165 uint16x8_t sum_left; in dc_16x16()
170 const uint16x8_t p0 = vpaddlq_u8(A); // cascading summation of the top in dc_16x16()
179 const uint16x8_t p0 = vpaddlq_u8(L); // cascading summation of the left in dc_16x16()
[all …]
Dsubpel_variance_neon.c36 const uint16x8_t a = vmull_u8(src_0, f0); in var_filter_block2d_bil_w8()
37 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8()
60 const uint16x8_t a = vmull_u8(vget_low_u8(src_0), f0); in var_filter_block2d_bil_w16()
61 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16()
63 const uint16x8_t c = vmull_u8(vget_high_u8(src_0), f0); in var_filter_block2d_bil_w16()
64 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
/external/libaom/libaom/av1/common/arm/
Dcfl_neon.c23 static INLINE uint16x8_t vldaddq_u16(const uint16_t *buf, size_t offset) { in vldaddq_u16()
58 const uint16x8_t top = vpaddlq_u8(vld1q_u8(input)); in cfl_luma_subsampling_420_lbd_neon()
59 const uint16x8_t sum = vpadalq_u8(top, vld1q_u8(input + input_stride)); in cfl_luma_subsampling_420_lbd_neon()
65 const uint16x8_t top_0 = vaddl_u8(top.val[0], top.val[1]); in cfl_luma_subsampling_420_lbd_neon()
67 const uint16x8_t bot_0 = vaddl_u8(bot.val[0], bot.val[1]); in cfl_luma_subsampling_420_lbd_neon()
69 const uint16x8_t top_1 = vaddl_u8(top.val[2], top.val[3]); in cfl_luma_subsampling_420_lbd_neon()
71 const uint16x8_t bot_1 = vaddl_u8(bot.val[2], bot.val[3]); in cfl_luma_subsampling_420_lbd_neon()
94 const uint16x8_t top = vpaddlq_u8(vld1q_u8(input)); in cfl_luma_subsampling_422_lbd_neon()
115 const uint16x8_t top = vshll_n_u8(vldh_dup_u8(input), 3); in cfl_luma_subsampling_444_lbd_neon()
118 const uint16x8_t top = vshll_n_u8(vld1_u8(input), 3); in cfl_luma_subsampling_444_lbd_neon()
[all …]
Dmem_neon.h93 uint16x8_t *const s0, uint16x8_t *const s1, in load_u16_8x4()
94 uint16x8_t *const s2, uint16x8_t *const s3) { in load_u16_8x4()
193 const uint16x8_t s0, const uint16x8_t s1, in store_u16_8x8()
194 const uint16x8_t s2, const uint16x8_t s3, in store_u16_8x8()
195 const uint16x8_t s4, const uint16x8_t s5, in store_u16_8x8()
196 const uint16x8_t s6, const uint16x8_t s7) { in store_u16_8x8()
227 const uint16x8_t s0, const uint16x8_t s1, in store_u16_8x4()
228 const uint16x8_t s2, const uint16x8_t s3) { in store_u16_8x4()
/external/libhevc/encoder/arm/
Dihevce_coarse_layer_sad_neon.c190 uint16x8_t abs_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_speed_neon()
191 uint16x8_t abs_23 = vdupq_n_u16(0); in hme_store_4x4_sads_high_speed_neon()
211 uint16x8_t abs_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_speed_neon()
231 uint16x8_t abs = vabdl_u8(vget_low_u8(src), vget_low_u8(ref)); in hme_store_4x4_sads_high_speed_neon()
341 uint16x8_t abs_a_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon()
342 uint16x8_t abs_a_23 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon()
343 uint16x8_t abs_b_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon()
344 uint16x8_t abs_b_23 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon()
372 uint16x8_t abs_a_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon()
373 uint16x8_t abs_b_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon()
[all …]
Dihevce_ssd_and_sad_calculator_neon.c71 const uint16x8_t sq_abs_l = vmull_u8(abs_l, abs_l); in ihevce_ssd_and_sad_calculator_neon()
72 const uint16x8_t sq_abs_h = vmull_u8(abs_h, abs_h); in ihevce_ssd_and_sad_calculator_neon()
73 uint16x8_t abs_sum; in ihevce_ssd_and_sad_calculator_neon()
93 uint16x8_t abs_sum = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon()
95 uint16x8_t abs, sqabs; in ihevce_ssd_and_sad_calculator_neon()
125 uint16x8_t abs_sum_l = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon()
126 uint16x8_t abs_sum_h = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon()
129 uint16x8_t abs_l, abs_h; in ihevce_ssd_and_sad_calculator_neon()
130 uint16x8_t sqabs_l, sqabs_h; in ihevce_ssd_and_sad_calculator_neon()
172 uint16x8_t abs_sum = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon()
[all …]
/external/clang/test/CodeGen/
Daarch64-neon-3v.c87 uint16x8_t test_vandq_u16(uint16x8_t a, uint16x8_t b) { in test_vandq_u16()
199 uint16x8_t test_vorrq_u16(uint16x8_t a, uint16x8_t b) { in test_vorrq_u16()
311 uint16x8_t test_veorq_u16(uint16x8_t a, uint16x8_t b) { in test_veorq_u16()
435 uint16x8_t test_vbicq_u16(uint16x8_t a, uint16x8_t b) { in test_vbicq_u16()
563 uint16x8_t test_vornq_u16(uint16x8_t a, uint16x8_t b) { in test_vornq_u16()
Darm_neon_intrinsics.c135 uint16x8_t test_vabaq_u16(uint16x8_t a, uint16x8_t b, uint16x8_t c) { in test_vabaq_u16()
202 uint16x8_t test_vabal_u8(uint16x8_t a, uint8x8_t b, uint8x8_t c) { in test_vabal_u8()
369 uint16x8_t test_vabdq_u16(uint16x8_t a, uint16x8_t b) { in test_vabdq_u16()
444 uint16x8_t test_vabdl_u8(uint8x8_t a, uint8x8_t b) { in test_vabdl_u8()
658 uint16x8_t test_vaddq_u16(uint16x8_t a, uint16x8_t b) { in test_vaddq_u16()
725 uint8x8_t test_vaddhn_u16(uint16x8_t a, uint16x8_t b) { in test_vaddhn_u16()
796 uint16x8_t test_vaddl_u8(uint8x8_t a, uint8x8_t b) { in test_vaddl_u8()
859 uint16x8_t test_vaddw_u8(uint16x8_t a, uint8x8_t b) { in test_vaddw_u8()
978 uint16x8_t test_vandq_u16(uint16x8_t a, uint16x8_t b) { in test_vandq_u16()
1105 uint16x8_t test_vbicq_u16(uint16x8_t a, uint16x8_t b) { in test_vbicq_u16()
[all …]
Daarch64-neon-2velem.c195 uint16x8_t test_vmulq_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmulq_lane_u16()
251 uint16x4_t test_vmul_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmul_laneq_u16()
259 uint16x8_t test_vmulq_laneq_u16(uint16x8_t a, uint16x8_t v) { in test_vmulq_laneq_u16()
1011 uint32x4_t test_vmull_high_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmull_high_lane_u16()
1060 uint32x4_t test_vmull_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmull_laneq_u16()
1111 uint32x4_t test_vmull_high_laneq_u16(uint16x8_t a, uint16x8_t v) { in test_vmull_high_laneq_u16()
1817 uint16x8_t test_vmulq_lane_u16_0(uint16x8_t a, uint16x4_t v) { in test_vmulq_lane_u16_0()
1873 uint16x4_t test_vmul_laneq_u16_0(uint16x4_t a, uint16x8_t v) { in test_vmul_laneq_u16_0()
1881 uint16x8_t test_vmulq_laneq_u16_0(uint16x8_t a, uint16x8_t v) { in test_vmulq_laneq_u16_0()
2561 uint32x4_t test_vmull_high_lane_u16_0(uint16x8_t a, uint16x4_t v) { in test_vmull_high_lane_u16_0()
[all …]
Daarch64-neon-intrinsics.c125 uint16x8_t test_vaddq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vaddq_u16()
256 uint16x8_t test_vsubq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vsubq_u16()
355 uint16x8_t test_vmulq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vmulq_u16()
500 uint16x8_t test_vmlaq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vmlaq_u16()
619 uint16x8_t test_vmlsq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vmlsq_u16()
851 uint16x8_t test_vabaq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vabaq_u16()
979 uint16x8_t test_vabdq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vabdq_u16()
1217 int16x8_t test_vbslq_s16(uint16x8_t v1, int16x8_t v2, int16x8_t v3) { in test_vbslq_s16()
1275 uint16x8_t test_vbslq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vbslq_u16()
1350 poly16x8_t test_vbslq_p16(uint16x8_t v1, poly16x8_t v2, poly16x8_t v3) { in test_vbslq_p16()
[all …]
Daarch64-neon-perm.c80 uint16x8_t test_vuzp1q_u16(uint16x8_t a, uint16x8_t b) { in test_vuzp1q_u16()
227 uint16x8_t test_vuzp2q_u16(uint16x8_t a, uint16x8_t b) { in test_vuzp2q_u16()
374 uint16x8_t test_vzip1q_u16(uint16x8_t a, uint16x8_t b) { in test_vzip1q_u16()
521 uint16x8_t test_vzip2q_u16(uint16x8_t a, uint16x8_t b) { in test_vzip2q_u16()
668 uint16x8_t test_vtrn1q_u16(uint16x8_t a, uint16x8_t b) { in test_vtrn1q_u16()
815 uint16x8_t test_vtrn2q_u16(uint16x8_t a, uint16x8_t b) { in test_vtrn2q_u16()
1245 uint16x8x2_t test_vuzpq_u16(uint16x8_t a, uint16x8_t b) { in test_vuzpq_u16()
1709 uint16x8x2_t test_vzipq_u16(uint16x8_t a, uint16x8_t b) { in test_vzipq_u16()
2173 uint16x8x2_t test_vtrnq_u16(uint16x8_t a, uint16x8_t b) { in test_vtrnq_u16()
Daarch64-neon-misc.c81 uint16x8_t test_vceqzq_s16(int16x8_t a) { in test_vceqzq_s16()
147 uint16x8_t test_vceqzq_u16(uint16x8_t a) { in test_vceqzq_u16()
233 uint16x8_t test_vceqzq_p16(poly16x8_t a) { in test_vceqzq_p16()
309 uint16x8_t test_vcgezq_s16(int16x8_t a) { in test_vcgezq_s16()
425 uint16x8_t test_vclezq_s16(int16x8_t a) { in test_vclezq_s16()
541 uint16x8_t test_vcgtzq_s16(int16x8_t a) { in test_vcgtzq_s16()
657 uint16x8_t test_vcltzq_s16(int16x8_t a) { in test_vcltzq_s16()
829 uint16x8_t test_vrev32q_u16(uint16x8_t a) { in test_vrev32q_u16()
941 uint16x8_t test_vrev64q_u16(uint16x8_t a) { in test_vrev64q_u16()
1051 uint16x8_t test_vpaddlq_u8(uint8x16_t a) { in test_vpaddlq_u8()
[all …]
Dstruct-matching-constraint.c4 typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t; typedef
6 void b(uint16x8_t sat, uint16x8_t luma) in b()
/external/libvpx/libvpx/vpx_dsp/ppc/
Dtypes_vsx.h19 typedef vector unsigned short uint16x8_t; typedef
47 (uint16x8_t) vec_mergeh(vec_splat_u8(0), (uint8x16_t)v)
49 (uint16x8_t) vec_mergel(vec_splat_u8(0), (uint8x16_t)v)
59 (uint16x8_t) vec_mergeh((uint8x16_t)v, vec_splat_u8(0))
61 (uint16x8_t) vec_mergel((uint8x16_t)v, vec_splat_u8(0))
87 static const uint16x8_t vec_ones_u16 = { 1, 1, 1, 1, 1, 1, 1, 1 };
91 static const uint16x8_t vec_shift_sign_s16 = { 15, 15, 15, 15, 15, 15, 15, 15 };
/external/webp/src/dsp/
Dalpha_processing_neon.c27 const uint16x8_t r1 = vmull_u8((V).val[1], alpha); \
28 const uint16x8_t g1 = vmull_u8((V).val[2], alpha); \
29 const uint16x8_t b1 = vmull_u8((V).val[(OTHER)], alpha); \
31 const uint16x8_t r2 = vsraq_n_u16(r1, r1, 8); \
32 const uint16x8_t g2 = vsraq_n_u16(g1, g1, 8); \
33 const uint16x8_t b2 = vsraq_n_u16(b1, b1, 8); \
34 const uint16x8_t r3 = vaddq_u16(r2, kOne); \
35 const uint16x8_t g3 = vaddq_u16(g2, kOne); \
36 const uint16x8_t b3 = vaddq_u16(b2, kOne); \
44 const uint16x8_t kOne = vdupq_n_u16(1u); in ApplyAlphaMultiply_NEON()

123456