/external/libaom/libaom/aom_dsp/simd/ |
D | v128_intrinsics_arm.h | 22 return v64_low_u32(vget_low_s64(a)); in v128_low_u32() 25 SIMD_INLINE v64 v128_low_v64(v128 a) { return vget_low_s64(a); } in v128_low_v64() 93 vmovl_s8(vreinterpret_s8_s64(vget_low_s64(a))), in v128_dotp_su8() 94 vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_s64(vget_low_s64(b))))); in v128_dotp_su8() 102 return (int64_t)vget_high_s64(t) + (int64_t)vget_low_s64(t); in v128_dotp_su8() 108 v64_dotp_s16(vget_low_s64(a), vget_low_s64(b)); in v128_dotp_s16() 114 return (int64_t)vget_high_s64(t) + (int64_t)vget_low_s64(t); in v128_dotp_s32() 151 r.lo = v64_sad_u8(s.lo, vget_low_s64(a), vget_low_s64(b)); in v128_sad_u8() 179 r.lo = v64_ssd_u8(s.lo, vget_low_s64(a), vget_low_s64(b)); in v128_ssd_u8() 288 vreinterpretq_s16_s32(vmull_s16(vreinterpret_s16_s64(vget_low_s64(a)), in v128_mulhi_s16() [all …]
|
D | v64_intrinsics_arm.h | 132 return (int64_t)vadd_s64(vget_high_s64(r), vget_low_s64(r)); in v64_dotp_su8() 143 return (int64_t)(vget_high_s64(r) + vget_low_s64(r)); in v64_dotp_s16() 289 return vget_low_s64(vreinterpretq_s64_s16(vuzp2q_s16(t, t))); in v64_mulhi_s16() 304 vpadd_s32(vreinterpret_s32_s64(vget_low_s64(vreinterpretq_s64_s32(t))), in v64_madd_s16()
|
D | v256_intrinsics_v128.h | 633 vtbl4_u8(p, vreinterpret_u8_s64(vget_low_s64(pattern.val[1])))), in v256_shuffle_8() 637 vtbl4_u8(p, vreinterpret_u8_s64(vget_low_s64(pattern.val[0]))))); in v256_shuffle_8() 679 p, vreinterpret_u8_s64(vget_low_s64(p32.val[1])))), in v256_wideshuffle_8() 683 p, vreinterpret_u8_s64(vget_low_s64(p32.val[0]))))); in v256_wideshuffle_8() 688 q, vreinterpret_u8_s64(vget_low_s64(pattern.val[1])))), in v256_wideshuffle_8() 692 q, vreinterpret_u8_s64(vget_low_s64(pattern.val[0]))))); in v256_wideshuffle_8()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | variance_neon.c | 23 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in horizontal_add_s16x8() 30 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in horizontal_add_s32x4() 212 d0s64 = vadd_s64(vget_low_s64(q0s64), vget_high_s64(q0s64)); in aom_variance16x8_neon() 213 d1s64 = vadd_s64(vget_low_s64(q1s64), vget_high_s64(q1s64)); in aom_variance16x8_neon() 218 d10u32 = vshr_n_u32(vreinterpret_u32_s64(vget_low_s64(q5s64)), 7); in aom_variance16x8_neon() 274 d0s64 = vadd_s64(vget_low_s64(q0s64), vget_high_s64(q0s64)); in aom_variance8x16_neon() 275 d1s64 = vadd_s64(vget_low_s64(q1s64), vget_high_s64(q1s64)); in aom_variance8x16_neon() 280 d10u32 = vshr_n_u32(vreinterpret_u32_s64(vget_low_s64(q5s64)), 7); in aom_variance8x16_neon() 343 d0s64 = vadd_s64(vget_low_s64(q1s64), vget_high_s64(q1s64)); in aom_mse16x16_neon() 397 d0s64 = vadd_s64(vget_low_s64(q1s64), vget_high_s64(q1s64)); in aom_get4x4sse_cs_neon()
|
D | fwd_txfm_neon.c | 217 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in aom_fdct8x8_1_neon()
|
/external/webrtc/webrtc/modules/video_processing/util/ |
D | denoiser_filter_neon.cc | 20 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in HorizontalAddS16x8() 27 const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), in HorizontalAddS32x4() 193 vget_low_s64(v_sum_diff_total)); in MbDenoise() 263 vget_low_s64(v_sum_diff_total)); in MbDenoise()
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
D | filters_neon.c | 50 prod = vget_lane_s64(vadd_s64(vget_low_s64(tmpb_v), vget_high_s64(tmpb_v)), in WebRtcIsacfix_AutocorrNeon() 96 prod = vget_lane_s64(vadd_s64(vget_low_s64(tmpb_v), vget_high_s64(tmpb_v)), in WebRtcIsacfix_AutocorrNeon()
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | denoising_neon.c | 146 vget_low_s64(v_sum_diff_total)); in vp8_denoiser_filter_neon() 215 vget_low_s64(v_sum_diff_total)); in vp8_denoiser_filter_neon() 363 vget_low_s64(v_sum_diff_total)); in vp8_denoiser_filter_uv_neon() 440 vget_low_s64(v_sum_diff_total)); in vp8_denoiser_filter_uv_neon()
|
/external/libopus/silk/arm/ |
D | NSQ_neon.c | 62 int64x1_t f = vadd_s64(vget_low_s64(e), vget_high_s64(e)); in silk_noise_shape_quantizer_short_prediction_neon() 102 int64x1_t c = vadd_s64(vget_low_s64(b3), vget_high_s64(b3)); in silk_NSQ_noise_shape_feedback_loop_neon()
|
/external/libopus/celt/arm/ |
D | pitch_neon_intr.c | 61 xy_s64x1 = vadd_s64(vget_low_s64(xy_s64x2), vget_high_s64(xy_s64x2)); in celt_inner_prod_neon() 107 xy01_s64x1 = vadd_s64(vget_low_s64(xy01_s64x2), vget_high_s64(xy01_s64x2)); in dual_inner_prod_neon() 108 xy02_s64x1 = vadd_s64(vget_low_s64(xy02_s64x2), vget_high_s64(xy02_s64x2)); in dual_inner_prod_neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | sum_neon.h | 22 return vadd_s32(vreinterpret_s32_s64(vget_low_s64(c)), in horizontal_add_int16x8()
|
D | avg_neon.c | 71 const int32x2_t s1 = vadd_s32(vreinterpret_s32_s64(vget_low_s64(s0)), in vpx_satd_neon() 180 const int32x2_t s1 = vadd_s32(vreinterpret_s32_s64(vget_low_s64(s0)), in vpx_vector_var_neon()
|
D | variance_neon.c | 324 d0s64 = vadd_s64(vget_low_s64(q1s64), vget_high_s64(q1s64)); in vpx_mse16x16_neon() 377 d0s64 = vadd_s64(vget_low_s64(q1s64), vget_high_s64(q1s64)); in vpx_get4x4sse_cs_neon()
|
/external/clang/test/CodeGen/ |
D | arm-neon-vget.c | 45 return vget_low_s64(a); in low_s64()
|
D | aarch64-neon-vget-hilo.c | 131 return vget_low_s64(a); in test_vget_low_s64()
|
D | arm_neon_intrinsics.c | 3796 return vget_low_s64(a); in test_vget_low_s64()
|
/external/webrtc/webrtc/common_audio/signal_processing/ |
D | cross_correlation_neon.c | 59 int64x1_t sum2 = vadd_s64(vget_low_s64(sum0), vget_high_s64(sum0)); in DotProductWithScaleNeon()
|
/external/libopus/silk/fixed/arm/ |
D | warped_autocorrelation_FIX_neon_intr.c | 212 t_s64x1 = vadd_s64( vget_low_s64( t_s64x2 ), vget_high_s64( t_s64x2 ) ); in silk_warped_autocorrelation_FIX_neon()
|
/external/libvpx/libvpx/vp9/encoder/arm/neon/ |
D | vp9_denoiser_neon.c | 28 vget_low_s64(fedcba98_76543210)); in horizontal_add_s8x16()
|
/external/libhevc/encoder/arm/ |
D | ihevce_had_compute_neon.c | 245 satd = vget_lane_s64(vadd_s64(vget_low_s64(c), vget_high_s64(c)), 0); in hadamard_sad8x8_using4x4() 366 satd = vget_lane_s64(vadd_s64(vget_low_s64(c), vget_high_s64(c)), 0); in ihevce_HAD_8x8_8bit_plane_neon()
|
D | ihevce_common_utils_neon.c | 454 vreinterpret_s32_s64(vget_low_s64(sad_reg)), in sad_cal()
|
/external/libhevc/common/arm/ |
D | ihevc_resi_trans_neon.c | 444 sad_v = vadd_s32(vreinterpret_s32_s64(vget_low_s64(tmp_b)), in ihevc_resi_trans_8x8_neon() 1317 sad_v = vadd_s32(vreinterpret_s32_s64(vget_low_s64(tmp_c)), in ihevc_resi_trans_16x16_neon()
|
D | ihevc_resi_trans_neon_32x32.c | 1033 sad = vadd_s32(vreinterpret_s32_s64(vget_low_s64(tmp_a)), in ihevc_resi_trans_32x32_neon()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 1760 _NEON2SSESTORAGE int64x1_t vget_low_s64(int64x2_t a); // VMOV d0,d0 12649 _NEON2SSESTORAGE int64x1_t vget_low_s64(int64x2_t a); // VMOV d0,d0 12650 _NEON2SSE_INLINE int64x1_t vget_low_s64(int64x2_t a) // VMOV d0,d0 in vget_low_s64() function 12677 #define vget_low_u64 vget_low_s64
|