Searched refs:vpaddlq_s16 (Results 1 – 19 of 19) sorted by relevance
251 m0_32 = vpaddlq_s16(mask0); in aom_lowbd_blend_a64_d16_mask_neon()252 m1_32 = vpaddlq_s16(mask1); in aom_lowbd_blend_a64_d16_mask_neon()253 m2_32 = vpaddlq_s16(mask2); in aom_lowbd_blend_a64_d16_mask_neon()254 m3_32 = vpaddlq_s16(mask3); in aom_lowbd_blend_a64_d16_mask_neon()256 m4_32 = vpaddlq_s16(mask4); in aom_lowbd_blend_a64_d16_mask_neon()257 m5_32 = vpaddlq_s16(mask5); in aom_lowbd_blend_a64_d16_mask_neon()258 m6_32 = vpaddlq_s16(mask6); in aom_lowbd_blend_a64_d16_mask_neon()259 m7_32 = vpaddlq_s16(mask7); in aom_lowbd_blend_a64_d16_mask_neon()296 m0_32 = vpaddlq_s16(mask0); in aom_lowbd_blend_a64_d16_mask_neon()297 m1_32 = vpaddlq_s16(mask1); in aom_lowbd_blend_a64_d16_mask_neon()[all …]
18 const int32x4_t a = vpaddlq_s16(v_16x8); in horizontal_add_s16x8()
309 const int32x4_t a = vpaddlq_s16(sum); in aom_fdct8x8_1_neon()
130 vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in vp8_denoiser_filter_neon()201 vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in vp8_denoiser_filter_neon()347 vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in vp8_denoiser_filter_uv_neon()426 vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in vp8_denoiser_filter_uv_neon()
18 const int32x4_t a = vpaddlq_s16(v_16x8); in HorizontalAddS16x8()178 vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in MbDenoise()249 vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in MbDenoise()
20 const int32x4_t b = vpaddlq_s16(a); in horizontal_add_int16x8()
175 const int32x4_t t0 = vpaddlq_s16(total); // cascading summation of 'total' in vpx_vector_var_neon()
148 int32x4_t sum0 = vpaddlq_s16(vmull_s8(vget_low_s8(lhs), vget_low_s8(rhs))); in vdotq_s32()149 int32x4_t sum1 = vpaddlq_s16(vmull_s8(vget_high_s8(lhs), vget_high_s8(rhs))); in vdotq_s32()170 int32x4_t sum0 = vpaddlq_s16(vmull_s8(vget_low_s8(lhs), lane_rhs)); in vdotq_four_lane_s32()171 int32x4_t sum1 = vpaddlq_s16(vmull_s8(vget_high_s8(lhs), lane_rhs)); in vdotq_four_lane_s32()
170 b01 = vpaddlq_s16(p[0]); in hadamard_sad4x4_4()177 b23 = vpaddlq_s16(p[4]); in hadamard_sad4x4_4()243 b = vpaddlq_s16(p[0]); in hadamard_sad8x8_using4x4()364 b = vpaddlq_s16(a0); in ihevce_HAD_8x8_8bit_plane_neon()439 sad_32_4[0] = vpaddlq_s16(combined_rows[0]); in ihevce_HAD_4x4_8bit_plane_neon()440 sad_32_4[1] = vpaddlq_s16(combined_rows[1]); in ihevce_HAD_4x4_8bit_plane_neon()
450 int64x2_t sad_reg = vpaddlq_s32(vpaddlq_s16(temp_reg)); in sad_cal()
216 t0_s32x4 = vpaddlq_s16( t0_s16x8 ); in silk_LPC_inverse_pred_gain_neon()
103 int64x2_t t = vpaddlq_s32(vaddq_s32(vpaddlq_s16(t1), vpaddlq_s16(t2))); in v128_dotp_su8()130 return vreinterpretq_s64_s32(vpaddlq_s16(vreinterpretq_s16_s64(a))); in v128_padd_s16()
135 int64x2_t r = vpaddlq_s32(vpaddlq_s16(t)); in v64_dotp_su8()316 return vreinterpret_s64_s16(vqmovn_s32(vpaddlq_s16(t))); in v64_madd_us8()
25 const int32x4_t fedc_ba98_7654_3210 = vpaddlq_s16(fe_dc_ba_98_76_54_32_10); in horizontal_add_s8x16()
444 tmp_a = vpaddlq_s16(abs); in ihevc_resi_trans_8x8_neon()888 return vpaddlq_s16(abs); in diff()
171 sum_val = vaddq_s32(sum_val,vpaddlq_s16(abs)); in ihevc_resi_trans_32x32_neon()
1036 return vpaddlq_s16(a); in test_vpaddlq_s16()
11367 return vpaddlq_s16(a); in test_vpaddlq_s16()
914 _NEON2SSESTORAGE int32x4_t vpaddlq_s16(int16x8_t a); // VPADDL.S16 q0,q06430 _NEON2SSESTORAGE int32x4_t vpaddlq_s16(int16x8_t a); // VPADDL.S16 q0,q06431 _NEON2SSE_INLINE int32x4_t vpaddlq_s16(int16x8_t a) // VPADDL.S16 q0,q0 in vpaddlq_s16() function6552 pad = vpaddlq_s16(b); in vpadalq_s16()