/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | shortfdct_neon.c | 50 d6s16 = vsub_s16(v2tmp0.val[1], v2tmp1.val[0]); in vp8_short_fdct4x4_neon() 51 d7s16 = vsub_s16(v2tmp0.val[0], v2tmp1.val[1]); in vp8_short_fdct4x4_neon() 59 d2s16 = vsub_s16(d4s16, d5s16); in vp8_short_fdct4x4_neon() 79 d6s16 = vsub_s16(v2tmp0.val[1], v2tmp1.val[0]); in vp8_short_fdct4x4_neon() 80 d7s16 = vsub_s16(v2tmp0.val[0], v2tmp1.val[1]); in vp8_short_fdct4x4_neon() 86 d2s16 = vsub_s16(d4s16, d5s16); in vp8_short_fdct4x4_neon() 102 d1s16 = vsub_s16(d1s16, vreinterpret_s16_u16(d4u16)); in vp8_short_fdct4x4_neon() 248 d1s16 = vsub_s16(d1s16, vreinterpret_s16_u16(d28u16)); in vp8_short_fdct8x4_neon() 249 d5s16 = vsub_s16(d5s16, vreinterpret_s16_u16(d29u16)); in vp8_short_fdct8x4_neon()
|
D | vp8_shortwalsh4x4_neon.c | 53 d6s16 = vsub_s16(v2tmp0.val[1], v2tmp1.val[1]); in vp8_short_walsh4x4_neon() 54 d7s16 = vsub_s16(v2tmp0.val[0], v2tmp1.val[0]); in vp8_short_walsh4x4_neon() 65 d3s16 = vsub_s16(d4s16, d5s16); in vp8_short_walsh4x4_neon() 67 d2s16 = vsub_s16(d7s16, d6s16); in vp8_short_walsh4x4_neon() 69 d0s16 = vsub_s16(d0s16, vreinterpret_s16_u16(d16u16)); in vp8_short_walsh4x4_neon()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | iwalsh_neon.c | 28 d5s16 = vsub_s16(vget_low_s16(q0s16), vget_high_s16(q1s16)); in vp8_short_inv_walsh4x4_neon() 29 d7s16 = vsub_s16(vget_high_s16(q0s16), vget_low_s16(q1s16)); in vp8_short_inv_walsh4x4_neon() 49 d5s16 = vsub_s16(v2tmp0.val[0], v2tmp1.val[1]); in vp8_short_inv_walsh4x4_neon() 50 d7s16 = vsub_s16(v2tmp0.val[1], v2tmp1.val[0]); in vp8_short_inv_walsh4x4_neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | idct16x16_add_neon.c | 506 step1[10] = vsub_s16(step2[9], step2[10]); in vpx_idct16x16_10_add_half1d_pass1() 507 step1[11] = vsub_s16(step2[8], step2[11]); in vpx_idct16x16_10_add_half1d_pass1() 508 step1[12] = vsub_s16(step2[15], step2[12]); in vpx_idct16x16_10_add_half1d_pass1() 509 step1[13] = vsub_s16(step2[14], step2[13]); in vpx_idct16x16_10_add_half1d_pass1() 518 step2[4] = vsub_s16(step1[3], step1[4]); in vpx_idct16x16_10_add_half1d_pass1() 519 step2[5] = vsub_s16(step1[2], step1[5]); in vpx_idct16x16_10_add_half1d_pass1() 520 step2[6] = vsub_s16(step1[1], step1[6]); in vpx_idct16x16_10_add_half1d_pass1() 521 step2[7] = vsub_s16(step1[0], step1[7]); in vpx_idct16x16_10_add_half1d_pass1() 540 out[8] = vsub_s16(step2[7], step2[8]); in vpx_idct16x16_10_add_half1d_pass1() 541 out[9] = vsub_s16(step2[6], step2[9]); in vpx_idct16x16_10_add_half1d_pass1() [all …]
|
D | idct_neon.h | 340 step2[5] = vsub_s16(step1[4], step1[5]); in idct8x8_12_pass1_bd8() 341 step2[6] = vsub_s16(step1[7], step1[6]); in idct8x8_12_pass1_bd8() 347 step1[2] = vsub_s16(step2[1], step2[2]); in idct8x8_12_pass1_bd8() 348 step1[3] = vsub_s16(step2[1], step2[3]); in idct8x8_12_pass1_bd8() 361 io[4] = vsub_s16(step1[3], step2[4]); in idct8x8_12_pass1_bd8() 362 io[5] = vsub_s16(step1[2], step1[5]); in idct8x8_12_pass1_bd8() 363 io[6] = vsub_s16(step1[1], step1[6]); in idct8x8_12_pass1_bd8() 364 io[7] = vsub_s16(step1[0], step2[7]); in idct8x8_12_pass1_bd8()
|
/external/libhevc/common/arm/ |
D | ihevc_quant_iquant_ssd_neon_intr.c | 219 q0 = vsub_s16(pq0, nq0); in ihevc_quant_iquant_ssd_flat_scale_mat_neon() 220 q1 = vsub_s16(pq1, nq1); in ihevc_quant_iquant_ssd_flat_scale_mat_neon() 221 q2 = vsub_s16(pq2, nq2); in ihevc_quant_iquant_ssd_flat_scale_mat_neon() 222 q3 = vsub_s16(pq3, nq3); in ihevc_quant_iquant_ssd_flat_scale_mat_neon()
|
/external/libhevc/encoder/arm/ |
D | ihevce_had_compute_neon.c | 202 int16x4_t q1 = vsub_s16(vget_low_s16(p0), vget_high_s16(p0)); in hadamard_sad8x8_using4x4() 204 int16x4_t q3 = vsub_s16(vget_low_s16(p1), vget_high_s16(p1)); in hadamard_sad8x8_using4x4() 406 h[2] = vsub_s16(vget_high_s16(res_01), vget_low_s16(res_23)); in ihevce_HAD_4x4_8bit_plane_neon() 407 h[3] = vsub_s16(vget_low_s16(res_01), vget_high_s16(res_23)); in ihevce_HAD_4x4_8bit_plane_neon() 411 v[2] = vsub_s16(h[0], h[1]); in ihevce_HAD_4x4_8bit_plane_neon() 412 v[3] = vsub_s16(h[3], h[2]); in ihevce_HAD_4x4_8bit_plane_neon()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | inverse_transform_neon.cc | 2229 s[3] = vsub_s16(s[3], s[1]); in Wht4_NEON() 2231 s[1] = vsub_s16(e, s[1]); in Wht4_NEON() 2232 s[2] = vsub_s16(e, s[2]); in Wht4_NEON() 2233 s[0] = vsub_s16(s[0], s[1]); in Wht4_NEON() 2243 s[3] = vsub_s16(s[3], s[1]); in Wht4_NEON() 2245 s[1] = vsub_s16(e, s[1]); in Wht4_NEON() 2246 s[2] = vsub_s16(e, s[2]); in Wht4_NEON() 2247 s[0] = vsub_s16(s[0], s[1]); in Wht4_NEON()
|
D | intrapred_cfl_neon.cc | 75 vst1_s16(a[y], vsub_s16(b, vget_low_s16(val_v))); in BlockSubtract()
|
/external/libopus/silk/arm/ |
D | NSQ_del_dec_neon_intr.c | 703 int16x4_t q1_Q10_s16x4 = vsub_s16( r_Q10_s16x4, vdup_n_s16( offset_Q10 ) ); in silk_noise_shape_quantizer_del_dec_neon() 718 …q1_Q0_s16x4 = vbsl_s16( greaterThanRdo, vsub_s16( q1_Q10_s16x4, vdup_n_s16( rdo_offset ) ), q1_Q0_… in silk_noise_shape_quantizer_del_dec_neon() 745 rr_Q10_s16x4 = vsub_s16( r_Q10_s16x4, q1_Q10_s16x4 ); in silk_noise_shape_quantizer_del_dec_neon() 749 rr_Q10_s16x4 = vsub_s16( r_Q10_s16x4, q2_Q10_s16x4 ); in silk_noise_shape_quantizer_del_dec_neon()
|
/external/libaom/libaom/av1/common/arm/ |
D | jnt_convolve_neon.c | 56 tmp0 = vsub_s16(vreinterpret_s16_u16(tmp_u0), sub_const_vec); in compute_avg_4x1() 177 tmp0 = vsub_s16(vreinterpret_s16_u16(tmp_u0), sub_const_vec); in compute_avg_4x4() 178 tmp1 = vsub_s16(vreinterpret_s16_u16(tmp_u1), sub_const_vec); in compute_avg_4x4() 179 tmp2 = vsub_s16(vreinterpret_s16_u16(tmp_u2), sub_const_vec); in compute_avg_4x4() 180 tmp3 = vsub_s16(vreinterpret_s16_u16(tmp_u3), sub_const_vec); in compute_avg_4x4()
|
D | cfl_neon.c | 345 vst1_s16(dst, vsub_s16(vreinterpret_s16_u16(vld1_u16(src)), avg_16x4)); in CFL_GET_SUBSAMPLE_FUNCTION()
|
/external/webp/src/dsp/ |
D | enc_neon.c | 318 const int16x4_t tmp2 = vsub_s16(vget_low_s16(a0a1_2), in FTransform_NEON() 336 const int16x4_t out2 = vshr_n_s16(vsub_s16(a0_k7, vget_high_s16(a0a1)), 4); in FTransform_NEON()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v64_intrinsics_arm.h | 248 vsub_s16(vreinterpret_s16_s64(x), vreinterpret_s16_s64(y))); in v64_sub_16()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 637 _NEON2SSESTORAGE int16x4_t vsub_s16(int16x4_t a, int16x4_t b); // VSUB.I16 d0,d0,d0 4371 _NEON2SSESTORAGE int16x4_t vsub_s16(int16x4_t a, int16x4_t b); // VSUB.I16 d0,d0,d0 4372 _NEON2SSE_INLINE int16x4_t vsub_s16(int16x4_t a, int16x4_t b) in vsub_s16() function 4409 #define vsub_u16 vsub_s16
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 153 return vsub_s16(v1, v2); in test_vsub_s16()
|
D | arm_neon_intrinsics.c | 21890 return vsub_s16(a, b); in test_vsub_s16()
|