Home
last modified time | relevance | path

Searched refs:vaddl_u8 (Results 1 – 25 of 32) sorted by relevance

12

/external/libaom/libaom/av1/common/arm/
Dwiener_convolve_neon.c118 res0 = vreinterpretq_s16_u16(vaddl_u8(t0, t6)); in av1_wiener_convolve_add_src_neon()
119 res1 = vreinterpretq_s16_u16(vaddl_u8(t1, t5)); in av1_wiener_convolve_add_src_neon()
120 res2 = vreinterpretq_s16_u16(vaddl_u8(t2, t4)); in av1_wiener_convolve_add_src_neon()
125 res0 = vreinterpretq_s16_u16(vaddl_u8(t1, t7)); in av1_wiener_convolve_add_src_neon()
126 res1 = vreinterpretq_s16_u16(vaddl_u8(t2, t6)); in av1_wiener_convolve_add_src_neon()
127 res2 = vreinterpretq_s16_u16(vaddl_u8(t3, t5)); in av1_wiener_convolve_add_src_neon()
132 res0 = vreinterpretq_s16_u16(vaddl_u8(t2, t8)); in av1_wiener_convolve_add_src_neon()
133 res1 = vreinterpretq_s16_u16(vaddl_u8(t3, t7)); in av1_wiener_convolve_add_src_neon()
134 res2 = vreinterpretq_s16_u16(vaddl_u8(t4, t6)); in av1_wiener_convolve_add_src_neon()
139 res0 = vreinterpretq_s16_u16(vaddl_u8(t3, t9)); in av1_wiener_convolve_add_src_neon()
[all …]
Dcfl_neon.c65 const uint16x8_t top_0 = vaddl_u8(top.val[0], top.val[1]); in cfl_luma_subsampling_420_lbd_neon()
67 const uint16x8_t bot_0 = vaddl_u8(bot.val[0], bot.val[1]); in cfl_luma_subsampling_420_lbd_neon()
69 const uint16x8_t top_1 = vaddl_u8(top.val[2], top.val[3]); in cfl_luma_subsampling_420_lbd_neon()
71 const uint16x8_t bot_1 = vaddl_u8(bot.val[2], bot.val[3]); in cfl_luma_subsampling_420_lbd_neon()
100 sum.val[0] = vshlq_n_u16(vaddl_u8(top.val[0], top.val[1]), 2); in cfl_luma_subsampling_422_lbd_neon()
101 sum.val[1] = vshlq_n_u16(vaddl_u8(top.val[2], top.val[3]), 2); in cfl_luma_subsampling_422_lbd_neon()
/external/libgav1/libgav1/src/dsp/arm/
Dintra_edge_neon.cc71 uint16x8_t sum_lo = vaddl_u8(vget_low_u8(src_0), vget_low_u8(src_2)); in IntraEdgeFilter_NEON()
74 uint16x8_t sum_hi = vaddl_u8(vget_high_u8(src_0), vget_high_u8(src_2)); in IntraEdgeFilter_NEON()
96 uint16x8_t sum_lo = vaddl_u8(vget_low_u8(src_0), vget_low_u8(src_2)); in IntraEdgeFilter_NEON()
99 uint16x8_t sum_hi = vaddl_u8(vget_high_u8(src_0), vget_high_u8(src_2)); in IntraEdgeFilter_NEON()
151 vshlq_n_u16(vaddl_u8(vget_low_u8(src_0), vget_low_u8(src_4)), 1); in IntraEdgeFilter_NEON()
153 vaddl_u8(vget_low_u8(src_1), vget_low_u8(src_2)), vget_low_u8(src_3)); in IntraEdgeFilter_NEON()
157 vshlq_n_u16(vaddl_u8(vget_high_u8(src_0), vget_high_u8(src_4)), 1); in IntraEdgeFilter_NEON()
159 vaddw_u8(vaddl_u8(vget_high_u8(src_1), vget_high_u8(src_2)), in IntraEdgeFilter_NEON()
182 vshlq_n_u16(vaddl_u8(vget_low_u8(src_0), vget_low_u8(src_4)), 1); in IntraEdgeFilter_NEON()
184 vaddl_u8(vget_low_u8(src_1), vget_low_u8(src_2)), vget_low_u8(src_3)); in IntraEdgeFilter_NEON()
[all …]
Dloop_filter_neon.cc293 const uint16x8_t p2q2_double = vaddl_u8(p2q2, p2q2); in Filter6()
300 sum = vaddq_u16(vaddl_u8(p1q1, p1q1), sum); in Filter6()
306 sum = vaddq_u16(vaddl_u8(p0q0, p0q0), sum); in Filter6()
322 sum = vaddq_u16(vaddl_u8(q0p0, q1p1), sum); in Filter6()
530 uint16x8_t sum = vaddw_u8(vaddl_u8(p3q3, p3q3), p3q3); in Filter8()
536 sum = vaddq_u16(vaddl_u8(p2q2, p2q2), sum); in Filter8()
542 sum = vaddq_u16(vaddl_u8(p1q1, p0q0), sum); in Filter8()
556 sum = vsubq_u16(sum, vaddl_u8(p3q3, p2q2)); in Filter8()
558 sum = vaddq_u16(vaddl_u8(p1q1, q1p1), sum); in Filter8()
565 sum = vsubq_u16(sum, vaddl_u8(p3q3, p1q1)); in Filter8()
[all …]
Dconvolve_neon.cc2852 row[0] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2855 row[1] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2858 row[2] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2860 row[3] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2863 row[4] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2865 row[5] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2867 row[6] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2869 row[7] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2872 row[8] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
2874 row[9] = vaddl_u8(vld1_u8(src), vld1_u8(src + 1)); in IntraBlockCopy2D()
[all …]
Dcdef_neon.cc93 *partial_lo = vaddl_u8(v_src[0], vext_u8(v_zero, v_src[1], 7)); in AddPartial_D0_D4()
100 vaddl_u8(vext_u8(v_src[1], v_zero, 7), vext_u8(v_src[2], v_zero, 6)); in AddPartial_D0_D4()
203 v_pair_add[0] = vaddl_u8(v_src[0], v_src[1]); in AddPartial_D5_D7()
204 v_pair_add[1] = vaddl_u8(v_src[2], v_src[3]); in AddPartial_D5_D7()
205 v_pair_add[2] = vaddl_u8(v_src[4], v_src[5]); in AddPartial_D5_D7()
206 v_pair_add[3] = vaddl_u8(v_src[6], v_src[7]); in AddPartial_D5_D7()
289 partial_lo[6] = vaddl_u8(v_zero, v_src[0]); in AddPartial()
Dintrapred_cfl_neon.cc175 vaddl_u8(vget_low_u8(row_masked_00), vget_low_u8(row_masked_01)); in CflSubsampler420_NEON()
183 vaddl_u8(vget_high_u8(row_masked_00), vget_high_u8(row_masked_01)); in CflSubsampler420_NEON()
Dloop_restoration_neon.cc67 const int16x8_t ss = vreinterpretq_s16_u16(vaddl_u8(s0, s1)); in WienerHorizontal2()
87 const int16x8_t s_0_2 = vreinterpretq_s16_u16(vaddl_u8(s[0], s[2])); in WienerHorizontalSum()
641 const uint16x8_t sum = vaddl_u8(src[0], src[1]); in Sum3W_16()
700 const uint16x8_t sum01 = vaddl_u8(s[0], s[1]); in Sum5Horizontal()
701 const uint16x8_t sum23 = vaddl_u8(s[2], s[3]); in Sum5Horizontal()
728 const uint16x8_t sum04 = vaddl_u8(s[0], s[4]); in SumHorizontal()
729 const uint16x8_t sum12 = vaddl_u8(s[1], s[2]); in SumHorizontal()
Dintrapred_neon.cc342 vabdq_u16(vaddl_u8(top, left), top_left_x2); in Paeth4Or8xN_NEON()
411 vaddl_u8(vget_low_u8(top[num]), vget_low_u8(left)), top_left_x2); \
413 vaddl_u8(vget_high_u8(top[num]), vget_low_u8(left)), top_left_x2)
/external/XNNPACK/src/q8-avgpool/
Dmp9p8q-neon.c68 const uint16x8_t vsum018 = vaddw_u8(vaddl_u8(vi0, vi1), vi8); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
69 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
70 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
71 const uint16x8_t vsum67 = vaddl_u8(vi6, vi7); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
109 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
110 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
111 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
112 const uint16x8_t vsum67 = vaddl_u8(vi6, vi7); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
172 const int16x8_t vsum01 = vreinterpretq_s16_u16(vaddl_u8(vi0, vi1)); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
173 const int16x8_t vsum23 = vreinterpretq_s16_u16(vaddl_u8(vi2, vi3)); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
[all …]
Dup9-neon.c92 const uint16x8_t vsum018 = vaddw_u8(vaddl_u8(vi0, vi1), vi8); in xnn_q8_avgpool_ukernel_up9__neon()
93 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_avgpool_ukernel_up9__neon()
94 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_avgpool_ukernel_up9__neon()
95 const uint16x8_t vsum67 = vaddl_u8(vi6, vi7); in xnn_q8_avgpool_ukernel_up9__neon()
165 const uint16x8_t vsum018 = vaddw_u8(vaddl_u8(vi0, vi1), vi8); in xnn_q8_avgpool_ukernel_up9__neon()
166 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_avgpool_ukernel_up9__neon()
167 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_avgpool_ukernel_up9__neon()
168 const uint16x8_t vsum67 = vaddl_u8(vi6, vi7); in xnn_q8_avgpool_ukernel_up9__neon()
/external/libaom/libaom/aom_dsp/arm/
Dblend_a64_mask_neon.c234 vreinterpretq_s16_u16(vaddl_u8(vget_low_u8(t0), vget_low_u8(t1))); in aom_lowbd_blend_a64_d16_mask_neon()
236 vreinterpretq_s16_u16(vaddl_u8(vget_low_u8(t2), vget_low_u8(t3))); in aom_lowbd_blend_a64_d16_mask_neon()
238 vreinterpretq_s16_u16(vaddl_u8(vget_low_u8(t4), vget_low_u8(t5))); in aom_lowbd_blend_a64_d16_mask_neon()
240 vreinterpretq_s16_u16(vaddl_u8(vget_low_u8(t6), vget_low_u8(t7))); in aom_lowbd_blend_a64_d16_mask_neon()
243 vaddl_u8(vget_high_u8(t0), vget_high_u8(t1))); in aom_lowbd_blend_a64_d16_mask_neon()
245 vaddl_u8(vget_high_u8(t2), vget_high_u8(t3))); in aom_lowbd_blend_a64_d16_mask_neon()
247 vaddl_u8(vget_high_u8(t4), vget_high_u8(t5))); in aom_lowbd_blend_a64_d16_mask_neon()
249 vaddl_u8(vget_high_u8(t6), vget_high_u8(t7))); in aom_lowbd_blend_a64_d16_mask_neon()
291 mask0 = vreinterpretq_s16_u16(vaddl_u8(mask0_l, mask1_l)); in aom_lowbd_blend_a64_d16_mask_neon()
292 mask1 = vreinterpretq_s16_u16(vaddl_u8(mask2_l, mask3_l)); in aom_lowbd_blend_a64_d16_mask_neon()
[all …]
Davg_neon.c21 const uint16x8_t c = vaddl_u8(vget_low_u8(b), vget_high_u8(b)); in aom_avg_4x4_neon()
38 sum = vaddl_u8(b, c); in aom_avg_8x8_neon()
Dloopfilter_neon.c231 out = vaddl_u8(*p3q3, *p2q2); in lpf_14_neon()
275 p6q6_2 = vaddl_u8(*p6q6, *p6q6); in lpf_14_neon()
286 qp_sum = vaddl_u8(q2p2, q1p1); in lpf_14_neon()
410 out = vaddl_u8(*p3q3, *p2q2); in lpf_8_neon()
524 out = vaddl_u8(*p0q0, *p1q1); in lpf_6_neon()
/external/XNNPACK/src/q8-gavgpool/
Dmp7p7q-neon.c52 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
53 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
54 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
89 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
90 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
91 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
151 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
152 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
153 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
223 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
[all …]
Dup7-neon.c75 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_gavgpool_ukernel_up7__neon()
76 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_gavgpool_ukernel_up7__neon()
77 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_gavgpool_ukernel_up7__neon()
145 const uint16x8_t vsum01 = vaddl_u8(vi0, vi1); in xnn_q8_gavgpool_ukernel_up7__neon()
146 const uint16x8_t vsum23 = vaddl_u8(vi2, vi3); in xnn_q8_gavgpool_ukernel_up7__neon()
147 const uint16x8_t vsum45 = vaddl_u8(vi4, vi5); in xnn_q8_gavgpool_ukernel_up7__neon()
/external/libhevc/encoder/arm/
Dihevce_scale_by_2_neon.c90 q = vreinterpretq_s16_u16(vaddl_u8(l0, r0)); in ihevce_horz_scale_neon_w16()
92 r = vreinterpretq_s16_u16(vaddl_u8(r3, vget_low_u8(src.val[0]))); in ihevce_horz_scale_neon_w16()
164 q = vreinterpretq_s16_u16(vaddl_u8(vget_low_u8(src[t1]), vget_low_u8(src[b1]))); in ihevce_vert_scale_neon_w16()
167 r = vreinterpretq_s16_u16(vaddl_u8(vget_low_u8(src[t2]), vget_low_u8(src[b2]))); in ihevce_vert_scale_neon_w16()
182 q = vreinterpretq_s16_u16(vaddl_u8(vget_high_u8(src[t1]), vget_high_u8(src[b1]))); in ihevce_vert_scale_neon_w16()
185 r = vreinterpretq_s16_u16(vaddl_u8(vget_high_u8(src[t2]), vget_high_u8(src[b2]))); in ihevce_vert_scale_neon_w16()
Dihevce_ssd_and_sad_calculator_neon.c78 abs_sum = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
193 abs_sum_l = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
204 abs_sum_h = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
256 abs_sum_0 = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
267 abs_sum_1 = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
278 abs_sum_2 = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
289 abs_sum_3 = vaddl_u8(abs_l, abs_h); in ihevce_ssd_and_sad_calculator_neon()
/external/libvpx/libvpx/vpx_dsp/arm/
Davg_neon.c24 const uint16x8_t c = vaddl_u8(vget_low_u8(b), vget_high_u8(b)); in vpx_avg_4x4_neon()
38 sum = vaddl_u8(b, c); in vpx_avg_8x8_neon()
Ddeblock_neon.c354 y = vreinterpretq_s16_u16(vaddl_u8(right_context, left_context)); in vpx_mbpost_proc_across_ip_neon()
451 y = vreinterpretq_s16_u16(vaddl_u8(below_context, above_context[0])); in vpx_mbpost_proc_down_neon()
/external/webp/src/dsp/
Dupsampling_neon.c37 const uint16x8_t ad = vaddl_u8(a, d); \
38 const uint16x8_t bc = vaddl_u8(b, c); \
Dfilters_neon.c157 const int16x8_t C = vreinterpretq_s16_u16(vaddl_u8(A, B)); in GradientPredictDirect_NEON()
Denc_neon.c603 const int16x8_t q_a0 = vreinterpretq_s16_u16(vaddl_u8(q4_in.val[0], in DistoVerticalPass_NEON()
605 const int16x8_t q_a1 = vreinterpretq_s16_u16(vaddl_u8(q4_in.val[1], in DistoVerticalPass_NEON()
/external/libpng/arm/
Dfilter_neon_intrinsics.c266 p1 = vaddl_u8(a, b); /* a + b */ in paeth()
267 pc = vaddl_u8(c, c); /* c * 2 */ in paeth()
/external/pdfium/third_party/libpng16/arm/
Dfilter_neon_intrinsics.c266 p1 = vaddl_u8(a, b); /* a + b */ in paeth()
267 pc = vaddl_u8(c, c); /* c * 2 */ in paeth()

12