/external/libvpx/libvpx/vpx_dsp/arm/ |
D | highbd_loopfilter_neon.c | 17 const uint8_t *thresh, uint16x8_t *blimit_vec, in load_thresh() 18 uint16x8_t *limit_vec, uint16x8_t *thresh_vec, in load_thresh() 35 static INLINE uint32_t calc_flat_status(const uint16x8_t flat) { in calc_flat_status() 42 static INLINE uint16x8_t 43 filter_hev_mask4(const uint16x8_t limit, const uint16x8_t blimit, in filter_hev_mask4() 44 const uint16x8_t thresh, const uint16x8_t p3, in filter_hev_mask4() 45 const uint16x8_t p2, const uint16x8_t p1, const uint16x8_t p0, in filter_hev_mask4() 46 const uint16x8_t q0, const uint16x8_t q1, const uint16x8_t q2, in filter_hev_mask4() 47 const uint16x8_t q3, uint16x8_t *hev, uint16x8_t *mask) { in filter_hev_mask4() 48 uint16x8_t max, t0, t1; in filter_hev_mask4() [all …]
|
D | highbd_intrapred_neon.c | 83 const uint16x8_t ref_u16 = vld1q_u16(ref); in dc_sum_8() 91 const uint16x8_t dc_dup = vdupq_lane_u16(dc, 0); in dc_store_8x8() 101 const uint16x8_t above_u16 = vld1q_u16(above); in vpx_highbd_dc_predictor_8x8_neon() 102 const uint16x8_t left_u16 = vld1q_u16(left); in vpx_highbd_dc_predictor_8x8_neon() 103 const uint16x8_t p0 = vaddq_u16(above_u16, left_u16); in vpx_highbd_dc_predictor_8x8_neon() 147 const uint16x8_t p0 = vaddq_u16(ref_u16.val[0], ref_u16.val[1]); in dc_sum_16() 168 const uint16x8_t pa = vaddq_u16(a.val[0], a.val[1]); in vpx_highbd_dc_predictor_16x16_neon() 169 const uint16x8_t pl = vaddq_u16(l.val[0], l.val[1]); in vpx_highbd_dc_predictor_16x16_neon() 170 const uint16x8_t pal0 = vaddq_u16(pa, pl); in vpx_highbd_dc_predictor_16x16_neon() 215 const uint16x8_t p0 = vaddq_u16(r.val[0], r.val[1]); in dc_sum_32() [all …]
|
D | highbd_idct32x32_add_neon.c | 21 const uint16x8_t a0 = vld1q_u16(*dest); in highbd_idct32x32_1_add_pos_kernel() 22 const uint16x8_t a1 = vld1q_u16(*dest + 8); in highbd_idct32x32_1_add_pos_kernel() 23 const uint16x8_t a2 = vld1q_u16(*dest + 16); in highbd_idct32x32_1_add_pos_kernel() 24 const uint16x8_t a3 = vld1q_u16(*dest + 24); in highbd_idct32x32_1_add_pos_kernel() 43 const uint16x8_t a0 = vld1q_u16(*dest); in highbd_idct32x32_1_add_neg_kernel() 44 const uint16x8_t a1 = vld1q_u16(*dest + 8); in highbd_idct32x32_1_add_neg_kernel() 45 const uint16x8_t a2 = vld1q_u16(*dest + 16); in highbd_idct32x32_1_add_neg_kernel() 46 const uint16x8_t a3 = vld1q_u16(*dest + 24); in highbd_idct32x32_1_add_neg_kernel() 51 const uint16x8_t c0 = vqshluq_n_s16(b0, 0); in highbd_idct32x32_1_add_neg_kernel() 52 const uint16x8_t c1 = vqshluq_n_s16(b1, 0); in highbd_idct32x32_1_add_neg_kernel() [all …]
|
D | sad_neon.c | 24 uint16x8_t abs = vabdl_u8(vget_low_u8(src_u8), vget_low_u8(ref_u8)); in vpx_sad4x4_neon() 36 uint16x8_t abs = vabdl_u8(vget_low_u8(src_u8), vget_low_u8(avg)); in vpx_sad4x4_avg_neon() 44 uint16x8_t abs = vdupq_n_u16(0); in vpx_sad4x8_neon() 61 uint16x8_t abs = vdupq_n_u16(0); in vpx_sad4x8_avg_neon() 77 static INLINE uint16x8_t sad8x(const uint8_t *src_ptr, int src_stride, in sad8x() 81 uint16x8_t abs = vdupq_n_u16(0); in sad8x() 93 static INLINE uint16x8_t sad8x_avg(const uint8_t *src_ptr, int src_stride, in sad8x_avg() 98 uint16x8_t abs = vdupq_n_u16(0); in sad8x_avg() 116 const uint16x8_t abs = sad8x(src_ptr, src_stride, ref_ptr, ref_stride, n); \ 123 const uint16x8_t abs = \ [all …]
|
D | subtract_neon.c | 32 const uint16x8_t d0 = vsubl_u8(vget_low_u8(s0), vget_low_u8(p0)); in vpx_subtract_block_neon() 33 const uint16x8_t d1 = vsubl_u8(vget_high_u8(s0), vget_high_u8(p0)); in vpx_subtract_block_neon() 34 const uint16x8_t d2 = vsubl_u8(vget_low_u8(s1), vget_low_u8(p1)); in vpx_subtract_block_neon() 35 const uint16x8_t d3 = vsubl_u8(vget_high_u8(s1), vget_high_u8(p1)); in vpx_subtract_block_neon() 49 const uint16x8_t d0 = vsubl_u8(vget_low_u8(s), vget_low_u8(p)); in vpx_subtract_block_neon() 50 const uint16x8_t d1 = vsubl_u8(vget_high_u8(s), vget_high_u8(p)); in vpx_subtract_block_neon() 61 const uint16x8_t v_diff = vsubl_u8(s, p); in vpx_subtract_block_neon() 72 const uint16x8_t d = vsubl_u8(s, p); in vpx_subtract_block_neon()
|
D | subpel_variance_neon.c | 39 const uint16x8_t a = vmull_u8(src_0, f0); in var_filter_block2d_bil_w4() 40 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w4() 61 const uint16x8_t a = vmull_u8(src_0, f0); in var_filter_block2d_bil_w8() 62 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8() 85 const uint16x8_t a = vmull_u8(vget_low_u8(src_0), f0); in var_filter_block2d_bil_w16() 86 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16() 88 const uint16x8_t c = vmull_u8(vget_high_u8(src_0), f0); in var_filter_block2d_bil_w16() 89 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
|
D | sad4d_neon.c | 36 uint16x8_t abs[2] = { vdupq_n_u16(0), vdupq_n_u16(0) }; in sad4x_4d() 75 static INLINE void sad_512_pel_final_neon(const uint16x8_t *sum /*[4]*/, in sad_512_pel_final_neon() 88 static INLINE void sad_1024_pel_final_neon(const uint16x8_t *sum /*[4]*/, in sad_1024_pel_final_neon() 102 static INLINE void sad_2048_pel_final_neon(const uint16x8_t *sum /*[4]*/, in sad_2048_pel_final_neon() 118 static INLINE void sad_4096_pel_final_neon(const uint16x8_t *sum /*[8]*/, in sad_4096_pel_final_neon() 147 uint16x8_t sum[4] = { vdupq_n_u16(0), vdupq_n_u16(0), vdupq_n_u16(0), in sad8x_4d() 184 uint16x8_t *const sum) { in sad16_neon() 196 uint16x8_t sum[4] = { vdupq_n_u16(0), vdupq_n_u16(0), vdupq_n_u16(0), in sad16x_4d() 233 const int height, uint16x8_t *const sum) { in sad32x_4d() 266 uint16x8_t sum[4]; in vpx_sad32x16x4d_neon() [all …]
|
/external/libaom/libaom/aom_dsp/arm/ |
D | sad4d_neon.c | 19 static INLINE unsigned int horizontal_long_add_16x8(const uint16x8_t vec_lo, in horizontal_long_add_16x8() 20 const uint16x8_t vec_hi) { in horizontal_long_add_16x8() 39 uint16x8_t *vec_sum_ref_lo, in sad_neon_64() 40 uint16x8_t *vec_sum_ref_hi) { in sad_neon_64() 68 uint16x8_t *vec_sum_ref_lo, in sad_neon_32() 69 uint16x8_t *vec_sum_ref_hi) { in sad_neon_32() 87 uint16x8_t vec_sum_ref0_lo = vdupq_n_u16(0); in aom_sad64x64x4d_neon() 88 uint16x8_t vec_sum_ref0_hi = vdupq_n_u16(0); in aom_sad64x64x4d_neon() 89 uint16x8_t vec_sum_ref1_lo = vdupq_n_u16(0); in aom_sad64x64x4d_neon() 90 uint16x8_t vec_sum_ref1_hi = vdupq_n_u16(0); in aom_sad64x64x4d_neon() [all …]
|
D | sad_neon.c | 21 uint16x8_t q12; in aom_sad8x16_neon() 52 uint16x8_t q12; in aom_sad4x4_neon() 80 uint16x8_t q12, q13; in aom_sad16x8_neon() 111 static INLINE unsigned int horizontal_long_add_16x8(const uint16x8_t vec_lo, in horizontal_long_add_16x8() 112 const uint16x8_t vec_hi) { in horizontal_long_add_16x8() 123 static INLINE unsigned int horizontal_add_16x8(const uint16x8_t vec_16x8) { in horizontal_add_16x8() 134 uint16x8_t vec_accum_lo = vdupq_n_u16(0); in aom_sad64x64_neon() 135 uint16x8_t vec_accum_hi = vdupq_n_u16(0); in aom_sad64x64_neon() 170 uint16x8_t vec_accum_lo = vdupq_n_u16(0); in aom_sad32x32_neon() 171 uint16x8_t vec_accum_hi = vdupq_n_u16(0); in aom_sad32x32_neon() [all …]
|
D | subtract_neon.c | 31 const uint16x8_t v_diff_lo_00 = in aom_subtract_block_neon() 33 const uint16x8_t v_diff_hi_00 = in aom_subtract_block_neon() 35 const uint16x8_t v_diff_lo_16 = in aom_subtract_block_neon() 37 const uint16x8_t v_diff_hi_16 = in aom_subtract_block_neon() 52 const uint16x8_t v_diff_lo = in aom_subtract_block_neon() 54 const uint16x8_t v_diff_hi = in aom_subtract_block_neon() 66 const uint16x8_t v_diff = vsubl_u8(v_src, v_pred); in aom_subtract_block_neon()
|
D | intrapred_neon.c | 25 uint16x8_t sum_top; in dc_4x4() 26 uint16x8_t sum_left; in dc_4x4() 44 const uint16x8_t sum = vaddq_u16(sum_left, sum_top); in dc_4x4() 93 uint16x8_t sum_top; in dc_8x8() 94 uint16x8_t sum_left; in dc_8x8() 114 const uint16x8_t sum = vaddq_u16(sum_left, sum_top); in dc_8x8() 164 uint16x8_t sum_top; in dc_16x16() 165 uint16x8_t sum_left; in dc_16x16() 170 const uint16x8_t p0 = vpaddlq_u8(A); // cascading summation of the top in dc_16x16() 179 const uint16x8_t p0 = vpaddlq_u8(L); // cascading summation of the left in dc_16x16() [all …]
|
D | subpel_variance_neon.c | 36 const uint16x8_t a = vmull_u8(src_0, f0); in var_filter_block2d_bil_w8() 37 const uint16x8_t b = vmlal_u8(a, src_1, f1); in var_filter_block2d_bil_w8() 60 const uint16x8_t a = vmull_u8(vget_low_u8(src_0), f0); in var_filter_block2d_bil_w16() 61 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1); in var_filter_block2d_bil_w16() 63 const uint16x8_t c = vmull_u8(vget_high_u8(src_0), f0); in var_filter_block2d_bil_w16() 64 const uint16x8_t d = vmlal_u8(c, vget_high_u8(src_1), f1); in var_filter_block2d_bil_w16()
|
/external/libaom/libaom/av1/common/arm/ |
D | cfl_neon.c | 23 static INLINE uint16x8_t vldaddq_u16(const uint16_t *buf, size_t offset) { in vldaddq_u16() 58 const uint16x8_t top = vpaddlq_u8(vld1q_u8(input)); in cfl_luma_subsampling_420_lbd_neon() 59 const uint16x8_t sum = vpadalq_u8(top, vld1q_u8(input + input_stride)); in cfl_luma_subsampling_420_lbd_neon() 65 const uint16x8_t top_0 = vaddl_u8(top.val[0], top.val[1]); in cfl_luma_subsampling_420_lbd_neon() 67 const uint16x8_t bot_0 = vaddl_u8(bot.val[0], bot.val[1]); in cfl_luma_subsampling_420_lbd_neon() 69 const uint16x8_t top_1 = vaddl_u8(top.val[2], top.val[3]); in cfl_luma_subsampling_420_lbd_neon() 71 const uint16x8_t bot_1 = vaddl_u8(bot.val[2], bot.val[3]); in cfl_luma_subsampling_420_lbd_neon() 94 const uint16x8_t top = vpaddlq_u8(vld1q_u8(input)); in cfl_luma_subsampling_422_lbd_neon() 115 const uint16x8_t top = vshll_n_u8(vldh_dup_u8(input), 3); in cfl_luma_subsampling_444_lbd_neon() 118 const uint16x8_t top = vshll_n_u8(vld1_u8(input), 3); in cfl_luma_subsampling_444_lbd_neon() [all …]
|
D | mem_neon.h | 93 uint16x8_t *const s0, uint16x8_t *const s1, in load_u16_8x4() 94 uint16x8_t *const s2, uint16x8_t *const s3) { in load_u16_8x4() 193 const uint16x8_t s0, const uint16x8_t s1, in store_u16_8x8() 194 const uint16x8_t s2, const uint16x8_t s3, in store_u16_8x8() 195 const uint16x8_t s4, const uint16x8_t s5, in store_u16_8x8() 196 const uint16x8_t s6, const uint16x8_t s7) { in store_u16_8x8() 227 const uint16x8_t s0, const uint16x8_t s1, in store_u16_8x4() 228 const uint16x8_t s2, const uint16x8_t s3) { in store_u16_8x4()
|
/external/libhevc/encoder/arm/ |
D | ihevce_coarse_layer_sad_neon.c | 190 uint16x8_t abs_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_speed_neon() 191 uint16x8_t abs_23 = vdupq_n_u16(0); in hme_store_4x4_sads_high_speed_neon() 211 uint16x8_t abs_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_speed_neon() 231 uint16x8_t abs = vabdl_u8(vget_low_u8(src), vget_low_u8(ref)); in hme_store_4x4_sads_high_speed_neon() 341 uint16x8_t abs_a_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon() 342 uint16x8_t abs_a_23 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon() 343 uint16x8_t abs_b_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon() 344 uint16x8_t abs_b_23 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon() 372 uint16x8_t abs_a_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon() 373 uint16x8_t abs_b_01 = vdupq_n_u16(0); in hme_store_4x4_sads_high_quality_neon() [all …]
|
D | ihevce_ssd_and_sad_calculator_neon.c | 71 const uint16x8_t sq_abs_l = vmull_u8(abs_l, abs_l); in ihevce_ssd_and_sad_calculator_neon() 72 const uint16x8_t sq_abs_h = vmull_u8(abs_h, abs_h); in ihevce_ssd_and_sad_calculator_neon() 73 uint16x8_t abs_sum; in ihevce_ssd_and_sad_calculator_neon() 93 uint16x8_t abs_sum = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon() 95 uint16x8_t abs, sqabs; in ihevce_ssd_and_sad_calculator_neon() 125 uint16x8_t abs_sum_l = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon() 126 uint16x8_t abs_sum_h = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon() 129 uint16x8_t abs_l, abs_h; in ihevce_ssd_and_sad_calculator_neon() 130 uint16x8_t sqabs_l, sqabs_h; in ihevce_ssd_and_sad_calculator_neon() 172 uint16x8_t abs_sum = vdupq_n_u16(0); in ihevce_ssd_and_sad_calculator_neon() [all …]
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-3v.c | 87 uint16x8_t test_vandq_u16(uint16x8_t a, uint16x8_t b) { in test_vandq_u16() 199 uint16x8_t test_vorrq_u16(uint16x8_t a, uint16x8_t b) { in test_vorrq_u16() 311 uint16x8_t test_veorq_u16(uint16x8_t a, uint16x8_t b) { in test_veorq_u16() 435 uint16x8_t test_vbicq_u16(uint16x8_t a, uint16x8_t b) { in test_vbicq_u16() 563 uint16x8_t test_vornq_u16(uint16x8_t a, uint16x8_t b) { in test_vornq_u16()
|
D | arm_neon_intrinsics.c | 135 uint16x8_t test_vabaq_u16(uint16x8_t a, uint16x8_t b, uint16x8_t c) { in test_vabaq_u16() 202 uint16x8_t test_vabal_u8(uint16x8_t a, uint8x8_t b, uint8x8_t c) { in test_vabal_u8() 369 uint16x8_t test_vabdq_u16(uint16x8_t a, uint16x8_t b) { in test_vabdq_u16() 444 uint16x8_t test_vabdl_u8(uint8x8_t a, uint8x8_t b) { in test_vabdl_u8() 658 uint16x8_t test_vaddq_u16(uint16x8_t a, uint16x8_t b) { in test_vaddq_u16() 725 uint8x8_t test_vaddhn_u16(uint16x8_t a, uint16x8_t b) { in test_vaddhn_u16() 796 uint16x8_t test_vaddl_u8(uint8x8_t a, uint8x8_t b) { in test_vaddl_u8() 859 uint16x8_t test_vaddw_u8(uint16x8_t a, uint8x8_t b) { in test_vaddw_u8() 978 uint16x8_t test_vandq_u16(uint16x8_t a, uint16x8_t b) { in test_vandq_u16() 1105 uint16x8_t test_vbicq_u16(uint16x8_t a, uint16x8_t b) { in test_vbicq_u16() [all …]
|
D | aarch64-neon-2velem.c | 195 uint16x8_t test_vmulq_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmulq_lane_u16() 251 uint16x4_t test_vmul_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmul_laneq_u16() 259 uint16x8_t test_vmulq_laneq_u16(uint16x8_t a, uint16x8_t v) { in test_vmulq_laneq_u16() 1011 uint32x4_t test_vmull_high_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmull_high_lane_u16() 1060 uint32x4_t test_vmull_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmull_laneq_u16() 1111 uint32x4_t test_vmull_high_laneq_u16(uint16x8_t a, uint16x8_t v) { in test_vmull_high_laneq_u16() 1817 uint16x8_t test_vmulq_lane_u16_0(uint16x8_t a, uint16x4_t v) { in test_vmulq_lane_u16_0() 1873 uint16x4_t test_vmul_laneq_u16_0(uint16x4_t a, uint16x8_t v) { in test_vmul_laneq_u16_0() 1881 uint16x8_t test_vmulq_laneq_u16_0(uint16x8_t a, uint16x8_t v) { in test_vmulq_laneq_u16_0() 2561 uint32x4_t test_vmull_high_lane_u16_0(uint16x8_t a, uint16x4_t v) { in test_vmull_high_lane_u16_0() [all …]
|
D | aarch64-neon-intrinsics.c | 125 uint16x8_t test_vaddq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vaddq_u16() 256 uint16x8_t test_vsubq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vsubq_u16() 355 uint16x8_t test_vmulq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vmulq_u16() 500 uint16x8_t test_vmlaq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vmlaq_u16() 619 uint16x8_t test_vmlsq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vmlsq_u16() 851 uint16x8_t test_vabaq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vabaq_u16() 979 uint16x8_t test_vabdq_u16(uint16x8_t v1, uint16x8_t v2) { in test_vabdq_u16() 1217 int16x8_t test_vbslq_s16(uint16x8_t v1, int16x8_t v2, int16x8_t v3) { in test_vbslq_s16() 1275 uint16x8_t test_vbslq_u16(uint16x8_t v1, uint16x8_t v2, uint16x8_t v3) { in test_vbslq_u16() 1350 poly16x8_t test_vbslq_p16(uint16x8_t v1, poly16x8_t v2, poly16x8_t v3) { in test_vbslq_p16() [all …]
|
D | aarch64-neon-perm.c | 80 uint16x8_t test_vuzp1q_u16(uint16x8_t a, uint16x8_t b) { in test_vuzp1q_u16() 227 uint16x8_t test_vuzp2q_u16(uint16x8_t a, uint16x8_t b) { in test_vuzp2q_u16() 374 uint16x8_t test_vzip1q_u16(uint16x8_t a, uint16x8_t b) { in test_vzip1q_u16() 521 uint16x8_t test_vzip2q_u16(uint16x8_t a, uint16x8_t b) { in test_vzip2q_u16() 668 uint16x8_t test_vtrn1q_u16(uint16x8_t a, uint16x8_t b) { in test_vtrn1q_u16() 815 uint16x8_t test_vtrn2q_u16(uint16x8_t a, uint16x8_t b) { in test_vtrn2q_u16() 1245 uint16x8x2_t test_vuzpq_u16(uint16x8_t a, uint16x8_t b) { in test_vuzpq_u16() 1709 uint16x8x2_t test_vzipq_u16(uint16x8_t a, uint16x8_t b) { in test_vzipq_u16() 2173 uint16x8x2_t test_vtrnq_u16(uint16x8_t a, uint16x8_t b) { in test_vtrnq_u16()
|
D | aarch64-neon-misc.c | 81 uint16x8_t test_vceqzq_s16(int16x8_t a) { in test_vceqzq_s16() 147 uint16x8_t test_vceqzq_u16(uint16x8_t a) { in test_vceqzq_u16() 233 uint16x8_t test_vceqzq_p16(poly16x8_t a) { in test_vceqzq_p16() 309 uint16x8_t test_vcgezq_s16(int16x8_t a) { in test_vcgezq_s16() 425 uint16x8_t test_vclezq_s16(int16x8_t a) { in test_vclezq_s16() 541 uint16x8_t test_vcgtzq_s16(int16x8_t a) { in test_vcgtzq_s16() 657 uint16x8_t test_vcltzq_s16(int16x8_t a) { in test_vcltzq_s16() 829 uint16x8_t test_vrev32q_u16(uint16x8_t a) { in test_vrev32q_u16() 941 uint16x8_t test_vrev64q_u16(uint16x8_t a) { in test_vrev64q_u16() 1051 uint16x8_t test_vpaddlq_u8(uint8x16_t a) { in test_vpaddlq_u8() [all …]
|
D | struct-matching-constraint.c | 4 typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t; typedef 6 void b(uint16x8_t sat, uint16x8_t luma) in b()
|
/external/libvpx/libvpx/vpx_dsp/ppc/ |
D | types_vsx.h | 19 typedef vector unsigned short uint16x8_t; typedef 47 (uint16x8_t) vec_mergeh(vec_splat_u8(0), (uint8x16_t)v) 49 (uint16x8_t) vec_mergel(vec_splat_u8(0), (uint8x16_t)v) 59 (uint16x8_t) vec_mergeh((uint8x16_t)v, vec_splat_u8(0)) 61 (uint16x8_t) vec_mergel((uint8x16_t)v, vec_splat_u8(0)) 87 static const uint16x8_t vec_ones_u16 = { 1, 1, 1, 1, 1, 1, 1, 1 }; 91 static const uint16x8_t vec_shift_sign_s16 = { 15, 15, 15, 15, 15, 15, 15, 15 };
|
/external/webp/src/dsp/ |
D | alpha_processing_neon.c | 27 const uint16x8_t r1 = vmull_u8((V).val[1], alpha); \ 28 const uint16x8_t g1 = vmull_u8((V).val[2], alpha); \ 29 const uint16x8_t b1 = vmull_u8((V).val[(OTHER)], alpha); \ 31 const uint16x8_t r2 = vsraq_n_u16(r1, r1, 8); \ 32 const uint16x8_t g2 = vsraq_n_u16(g1, g1, 8); \ 33 const uint16x8_t b2 = vsraq_n_u16(b1, b1, 8); \ 34 const uint16x8_t r3 = vaddq_u16(r2, kOne); \ 35 const uint16x8_t g3 = vaddq_u16(g2, kOne); \ 36 const uint16x8_t b3 = vaddq_u16(b2, kOne); \ 44 const uint16x8_t kOne = vdupq_n_u16(1u); in ApplyAlphaMultiply_NEON()
|