/external/libvpx/libvpx/vpx_dsp/arm/ |
D | fdct_partial_neon.c | 80 sum = vadd_s32(horizontal_add_int16x8(left), horizontal_add_int16x8(right)); in vpx_fdct16x16_1_neon() 108 sum = vadd_s32(horizontal_add_int16x8(a0), horizontal_add_int16x8(a1)); in vpx_fdct32x32_1_neon() 109 sum = vadd_s32(sum, horizontal_add_int16x8(a2)); in vpx_fdct32x32_1_neon() 110 sum = vadd_s32(sum, horizontal_add_int16x8(a3)); in vpx_fdct32x32_1_neon()
|
D | avg_neon.c | 71 const int32x2_t s1 = vadd_s32(vreinterpret_s32_s64(vget_low_s64(s0)), in vpx_satd_neon() 176 const int32x2_t t1 = vadd_s32(vget_low_s32(t0), vget_high_s32(t0)); in vpx_vector_var_neon() 180 const int32x2_t s1 = vadd_s32(vreinterpret_s32_s64(vget_low_s64(s0)), in vpx_vector_var_neon()
|
D | sum_neon.h | 22 return vadd_s32(vreinterpret_s32_s64(vget_low_s64(c)), in horizontal_add_int16x8()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | sum_neon.h | 20 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in horizontal_add_s16x8() 27 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in horizontal_add_s32x4()
|
D | avg_neon.c | 69 const int32x2_t s1 = vadd_s32(vreinterpret_s32_s64(vget_low_s64(s0)), in aom_satd_lp_neon()
|
D | fwd_txfm_neon.c | 311 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in aom_fdct8x8_1_neon()
|
/external/libhevc/common/arm/ |
D | ihevc_resi_trans_neon_32x32.c | 1035 sad = vadd_s32(vreinterpret_s32_s64(vget_low_s64(tmp_a)), in ihevc_resi_trans_32x32_neon() 1197 vadd_s32(vget_low_s32(eee), vrev64_s32(vget_high_s32(eee))), in ihevc_resi_trans_32x32_neon() 1255 int32x2_t val_2 = vadd_s32(vget_low_s32(a[2]), vget_high_s32(a[2])); in ihevc_resi_trans_32x32_neon() 1256 int32x2_t val_6 = vadd_s32(vget_low_s32(a[6]), vget_high_s32(a[6])); in ihevc_resi_trans_32x32_neon() 1260 vadd_s32(vget_low_s32(a[10]), vget_high_s32(a[10])); in ihevc_resi_trans_32x32_neon() 1262 vadd_s32(vget_low_s32(a[14]), vget_high_s32(a[14])); in ihevc_resi_trans_32x32_neon() 1286 vadd_s32(vget_low_s32(a[18]), vget_high_s32(a[18])); in ihevc_resi_trans_32x32_neon() 1288 vadd_s32(vget_low_s32(a[22]), vget_high_s32(a[22])); in ihevc_resi_trans_32x32_neon() 1291 vadd_s32(vget_low_s32(a[26]), vget_high_s32(a[26])); in ihevc_resi_trans_32x32_neon() 1293 vadd_s32(vget_low_s32(a[30]), vget_high_s32(a[30])); in ihevc_resi_trans_32x32_neon() [all …]
|
D | ihevc_resi_trans_neon.c | 446 sad_v = vadd_s32(vreinterpret_s32_s64(vget_low_s64(tmp_b)), in ihevc_resi_trans_8x8_neon() 1320 sad_v = vadd_s32(vreinterpret_s32_s64(vget_low_s64(tmp_c)), in ihevc_resi_trans_16x16_neon()
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
D | lattice_neon.c | 157 tmp2_tail = vadd_s32(tmp0_tail, ptr2v_tail); in WebRtcIsacfix_FilterMaLoopNeon() 170 ptr1v_tail = vadd_s32(tmp1_tail, tmp3_tail); in WebRtcIsacfix_FilterMaLoopNeon()
|
/external/libopus/silk/arm/ |
D | biquad_alt_neon_intr.c | 43 …*out32_Q14_s32x2 = vadd_s32( vget_low_s32( *S_s32x4 ), t_s32x2 ); /* silk_SMLAWB( S{0… in silk_biquad_alt_stride2_kernel() 139 …out32_Q14_s32x2 = vadd_s32( out32_Q14_s32x2, offset_s32x2 ); /*… in silk_biquad_alt_stride2_neon()
|
/external/webrtc/webrtc/modules/video_processing/util/ |
D | denoiser_filter_neon.cc | 20 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in HorizontalAddS16x8() 27 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in HorizontalAddS32x4()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/integer_ops/ |
D | softmax.h | 140 vadd_s32(vget_low_s32(sum_of_exps_reduced_4), in Softmax()
|
/external/libhevc/encoder/arm/ |
D | ihevce_had_compute_neon.c | 182 d01 = vadd_s32(d01, d23); in hadamard_sad4x4_4() 442 sad_32_2 = vadd_s32(vget_high_s32(sad_32_4[2]), vget_low_s32(sad_32_4[2])); in ihevce_HAD_4x4_8bit_plane_neon()
|
D | ihevce_common_utils_neon.c | 453 vadd_s32( in sad_cal()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 447 _NEON2SSESTORAGE int32x2_t vadd_s32(int32x2_t a, int32x2_t b); // VADD.I32 d0,d0,d0 2801 _NEON2SSESTORAGE int32x2_t vadd_s32(int32x2_t a, int32x2_t b); // VADD.I32 d0,d0,d0 2802 _NEON2SSE_INLINE int32x2_t vadd_s32(int32x2_t a, int32x2_t b) in vadd_s32() function 2835 #define vadd_u32 vadd_s32 7963 return vadd_s32( a, shift); 8097 return vadd_s32( a, shift);
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | legacy_optimized_ops.h | 4091 vadd_s32(vget_low_s32(sum_of_exps_reduced_4), in Softmax()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 28 return vadd_s32(v1, v2); in test_vadd_s32()
|
D | arm_neon_intrinsics.c | 568 return vadd_s32(a, b); in test_vadd_s32()
|