/external/libvpx/libvpx/vpx_dsp/arm/ |
D | idct32x32_135_add_neon.c | 134 s2[18] = vsubq_s16(s1[19], s1[18]); in vpx_idct32_12_neon() 137 s2[21] = vsubq_s16(s1[20], s1[21]); in vpx_idct32_12_neon() 138 s2[26] = vsubq_s16(s1[27], s1[26]); in vpx_idct32_12_neon() 141 s2[29] = vsubq_s16(s1[28], s1[29]); in vpx_idct32_12_neon() 147 s3[10] = vsubq_s16(s2[11], s2[10]); in vpx_idct32_12_neon() 150 s3[13] = vsubq_s16(s2[12], s2[13]); in vpx_idct32_12_neon() 189 s4[18] = vsubq_s16(s3[17], s3[18]); in vpx_idct32_12_neon() 190 s4[19] = vsubq_s16(s1[16], s2[19]); in vpx_idct32_12_neon() 191 s4[20] = vsubq_s16(s1[23], s2[20]); in vpx_idct32_12_neon() 192 s4[21] = vsubq_s16(s3[22], s3[21]); in vpx_idct32_12_neon() [all …]
|
D | idct32x32_34_add_neon.c | 103 s2[20] = vsubq_s16(s1[23], s1[20]); in vpx_idct32_6_neon() 104 s2[21] = vsubq_s16(s1[22], s1[21]); in vpx_idct32_6_neon() 109 s2[26] = vsubq_s16(s1[25], s1[26]); in vpx_idct32_6_neon() 110 s2[27] = vsubq_s16(s1[24], s1[27]); in vpx_idct32_6_neon() 141 s2[4] = vsubq_s16(s1[0], s1[4]); in vpx_idct32_6_neon() 142 s2[5] = vsubq_s16(s1[0], s1[5]); in vpx_idct32_6_neon() 143 s2[6] = vsubq_s16(s1[0], s1[6]); in vpx_idct32_6_neon() 144 s2[7] = vsubq_s16(s1[0], s1[7]); in vpx_idct32_6_neon() 156 s2[20] = vsubq_s16(s1[19], s1[20]); in vpx_idct32_6_neon() 157 s2[21] = vsubq_s16(s1[18], s1[21]); in vpx_idct32_6_neon() [all …]
|
D | idct32x32_add_neon.c | 249 q[6] = vsubq_s16(q[3], q[0]); in idct32_bands_end_1st_pass() 250 q[7] = vsubq_s16(q[2], q[1]); in idct32_bands_end_1st_pass() 257 q[4] = vsubq_s16(q[11], q[0]); in idct32_bands_end_1st_pass() 258 q[5] = vsubq_s16(q[10], q[1]); in idct32_bands_end_1st_pass() 263 q[6] = vsubq_s16(q[5], q[0]); in idct32_bands_end_1st_pass() 264 q[7] = vsubq_s16(q[4], q[1]); in idct32_bands_end_1st_pass() 271 q[6] = vsubq_s16(q[3], q[0]); in idct32_bands_end_1st_pass() 272 q[7] = vsubq_s16(q[2], q[1]); in idct32_bands_end_1st_pass() 279 q[4] = vsubq_s16(q[13], q[0]); in idct32_bands_end_1st_pass() 280 q[5] = vsubq_s16(q[12], q[1]); in idct32_bands_end_1st_pass() [all …]
|
D | fdct32x32_neon.c | 48 src[index1] = vshlq_n_s16(vsubq_s16(src[index0], src[index1]), 2); \ 242 return vrshrq_n_s16(vsubq_s16(a, a_sign_s16), 2); in sub_round_shift() 262 a[8] = vsubq_s16(in[7], in[8]); in dct_body_first_pass() 263 a[9] = vsubq_s16(in[6], in[9]); in dct_body_first_pass() 264 a[10] = vsubq_s16(in[5], in[10]); in dct_body_first_pass() 265 a[11] = vsubq_s16(in[4], in[11]); in dct_body_first_pass() 266 a[12] = vsubq_s16(in[3], in[12]); in dct_body_first_pass() 267 a[13] = vsubq_s16(in[2], in[13]); in dct_body_first_pass() 268 a[14] = vsubq_s16(in[1], in[14]); in dct_body_first_pass() 269 a[15] = vsubq_s16(in[0], in[15]); in dct_body_first_pass() [all …]
|
D | fdct16x16_neon.c | 101 b[8] = vshlq_n_s16(vsubq_s16(a[7], a[8]), 2); in cross_input() 102 b[9] = vshlq_n_s16(vsubq_s16(a[6], a[9]), 2); in cross_input() 103 b[10] = vshlq_n_s16(vsubq_s16(a[5], a[10]), 2); in cross_input() 104 b[11] = vshlq_n_s16(vsubq_s16(a[4], a[11]), 2); in cross_input() 105 b[12] = vshlq_n_s16(vsubq_s16(a[3], a[12]), 2); in cross_input() 106 b[13] = vshlq_n_s16(vsubq_s16(a[2], a[13]), 2); in cross_input() 107 b[14] = vshlq_n_s16(vsubq_s16(a[1], a[14]), 2); in cross_input() 108 b[15] = vshlq_n_s16(vsubq_s16(a[0], a[15]), 2); in cross_input() 119 b[8] = vsubq_s16(a[7], a[8]); in cross_input() 120 b[9] = vsubq_s16(a[6], a[9]); in cross_input() [all …]
|
D | hadamard_neon.c | 23 const int16x8_t b1 = vsubq_s16(*a0, *a1); in hadamard8x8_one_pass() 25 const int16x8_t b3 = vsubq_s16(*a2, *a3); in hadamard8x8_one_pass() 27 const int16x8_t b5 = vsubq_s16(*a4, *a5); in hadamard8x8_one_pass() 29 const int16x8_t b7 = vsubq_s16(*a6, *a7); in hadamard8x8_one_pass() 33 const int16x8_t c2 = vsubq_s16(b0, b2); in hadamard8x8_one_pass() 34 const int16x8_t c3 = vsubq_s16(b1, b3); in hadamard8x8_one_pass() 37 const int16x8_t c6 = vsubq_s16(b4, b6); in hadamard8x8_one_pass() 38 const int16x8_t c7 = vsubq_s16(b5, b7); in hadamard8x8_one_pass() 41 *a1 = vsubq_s16(c2, c6); in hadamard8x8_one_pass() 42 *a2 = vsubq_s16(c0, c4); in hadamard8x8_one_pass() [all …]
|
D | idct16x16_add_neon.c | 195 step1[9] = vsubq_s16(step2[8], step2[9]); in vpx_idct16x16_256_add_half1d() 196 step1[10] = vsubq_s16(step2[11], step2[10]); in vpx_idct16x16_256_add_half1d() 199 step1[13] = vsubq_s16(step2[12], step2[13]); in vpx_idct16x16_256_add_half1d() 200 step1[14] = vsubq_s16(step2[15], step2[14]); in vpx_idct16x16_256_add_half1d() 207 step2[5] = vsubq_s16(step1[4], step1[5]); in vpx_idct16x16_256_add_half1d() 208 step2[6] = vsubq_s16(step1[7], step1[6]); in vpx_idct16x16_256_add_half1d() 222 step1[2] = vsubq_s16(step2[1], step2[2]); in vpx_idct16x16_256_add_half1d() 223 step1[3] = vsubq_s16(step2[0], step2[3]); in vpx_idct16x16_256_add_half1d() 229 step1[10] = vsubq_s16(step2[9], step2[10]); in vpx_idct16x16_256_add_half1d() 230 step1[11] = vsubq_s16(step2[8], step2[11]); in vpx_idct16x16_256_add_half1d() [all …]
|
D | fwd_txfm_neon.c | 38 const int16x8_t v_s4 = vsubq_s16(input_3, input_4); in vpx_fdct8x8_neon() 39 const int16x8_t v_s5 = vsubq_s16(input_2, input_5); in vpx_fdct8x8_neon() 40 const int16x8_t v_s6 = vsubq_s16(input_1, input_6); in vpx_fdct8x8_neon() 41 const int16x8_t v_s7 = vsubq_s16(input_0, input_7); in vpx_fdct8x8_neon() 45 int16x8_t v_x2 = vsubq_s16(v_s1, v_s2); in vpx_fdct8x8_neon() 46 int16x8_t v_x3 = vsubq_s16(v_s0, v_s3); in vpx_fdct8x8_neon() 79 v_x0 = vsubq_s16(v_s6, v_s5); in vpx_fdct8x8_neon() 94 v_x1 = vsubq_s16(v_s4, ab); in vpx_fdct8x8_neon() 95 v_x2 = vsubq_s16(v_s7, cd); in vpx_fdct8x8_neon()
|
D | idct_neon.h | 57 return vsubq_s16(a, b); in final_sub() 311 a[1] = vsubq_s16(d[0], d[1]); in idct4x4_16_kernel_bd8() 391 step2[5] = vsubq_s16(step1[4], step1[5]); in idct8x8_12_pass2_bd8() 392 step2[6] = vsubq_s16(step1[7], step1[6]); in idct8x8_12_pass2_bd8() 398 step1[2] = vsubq_s16(step2[1], step2[2]); in idct8x8_12_pass2_bd8() 399 step1[3] = vsubq_s16(step2[1], step2[3]); in idct8x8_12_pass2_bd8() 414 output[4] = vsubq_s16(step1[3], step2[4]); in idct8x8_12_pass2_bd8() 415 output[5] = vsubq_s16(step1[2], step1[5]); in idct8x8_12_pass2_bd8() 416 output[6] = vsubq_s16(step1[1], step1[6]); in idct8x8_12_pass2_bd8() 417 output[7] = vsubq_s16(step1[0], step2[7]); in idct8x8_12_pass2_bd8() [all …]
|
/external/libjpeg-turbo/simd/arm/ |
D | jfdctfst-neon.c | 90 int16x8_t tmp7 = vsubq_s16(col0, col7); in jsimd_fdct_ifast_neon() 92 int16x8_t tmp6 = vsubq_s16(col1, col6); in jsimd_fdct_ifast_neon() 94 int16x8_t tmp5 = vsubq_s16(col2, col5); in jsimd_fdct_ifast_neon() 96 int16x8_t tmp4 = vsubq_s16(col3, col4); in jsimd_fdct_ifast_neon() 100 int16x8_t tmp13 = vsubq_s16(tmp0, tmp3); in jsimd_fdct_ifast_neon() 102 int16x8_t tmp12 = vsubq_s16(tmp1, tmp2); in jsimd_fdct_ifast_neon() 105 col4 = vsubq_s16(tmp10, tmp11); in jsimd_fdct_ifast_neon() 109 col6 = vsubq_s16(tmp13, z1); in jsimd_fdct_ifast_neon() 116 int16x8_t z5 = vqdmulhq_lane_s16(vsubq_s16(tmp10, tmp12), consts, 0); in jsimd_fdct_ifast_neon() 125 int16x8_t z13 = vsubq_s16(tmp7, z3); in jsimd_fdct_ifast_neon() [all …]
|
D | jidctfst-neon.c | 278 int16x8_t tmp11 = vsubq_s16(tmp0, tmp2); in jsimd_idct_ifast_neon() 281 int16x8_t tmp1_sub_tmp3 = vsubq_s16(tmp1, tmp3); in jsimd_idct_ifast_neon() 284 tmp12 = vsubq_s16(tmp12, tmp13); in jsimd_idct_ifast_neon() 287 tmp3 = vsubq_s16(tmp10, tmp13); in jsimd_idct_ifast_neon() 289 tmp2 = vsubq_s16(tmp11, tmp12); in jsimd_idct_ifast_neon() 298 int16x8_t neg_z10 = vsubq_s16(tmp5, tmp6); in jsimd_idct_ifast_neon() 300 int16x8_t z12 = vsubq_s16(tmp4, tmp7); in jsimd_idct_ifast_neon() 303 int16x8_t z11_sub_z13 = vsubq_s16(z11, z13); in jsimd_idct_ifast_neon() 307 int16x8_t z10_add_z12 = vsubq_s16(z12, neg_z10); in jsimd_idct_ifast_neon() 312 tmp10 = vsubq_s16(tmp10, z5); in jsimd_idct_ifast_neon() [all …]
|
D | jfdctint-neon.c | 124 int16x8_t tmp7 = vsubq_s16(col0, col7); in jsimd_fdct_islow_neon() 126 int16x8_t tmp6 = vsubq_s16(col1, col6); in jsimd_fdct_islow_neon() 128 int16x8_t tmp5 = vsubq_s16(col2, col5); in jsimd_fdct_islow_neon() 130 int16x8_t tmp4 = vsubq_s16(col3, col4); in jsimd_fdct_islow_neon() 134 int16x8_t tmp13 = vsubq_s16(tmp0, tmp3); in jsimd_fdct_islow_neon() 136 int16x8_t tmp12 = vsubq_s16(tmp1, tmp2); in jsimd_fdct_islow_neon() 139 col4 = vshlq_n_s16(vsubq_s16(tmp10, tmp11), PASS1_BITS); in jsimd_fdct_islow_neon() 263 tmp7 = vsubq_s16(row0, row7); in jsimd_fdct_islow_neon() 265 tmp6 = vsubq_s16(row1, row6); in jsimd_fdct_islow_neon() 267 tmp5 = vsubq_s16(row2, row5); in jsimd_fdct_islow_neon() [all …]
|
D | jquanti-neon.c | 176 row0 = vsubq_s16(row0, sign_row0); in jsimd_quantize_neon() 178 row1 = vsubq_s16(row1, sign_row1); in jsimd_quantize_neon() 180 row2 = vsubq_s16(row2, sign_row2); in jsimd_quantize_neon() 182 row3 = vsubq_s16(row3, sign_row3); in jsimd_quantize_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | hadamard_neon.c | 22 const int16x8_t b1 = vsubq_s16(*a0, *a1); in hadamard8x8_one_pass() 24 const int16x8_t b3 = vsubq_s16(*a2, *a3); in hadamard8x8_one_pass() 26 const int16x8_t b5 = vsubq_s16(*a4, *a5); in hadamard8x8_one_pass() 28 const int16x8_t b7 = vsubq_s16(*a6, *a7); in hadamard8x8_one_pass() 32 const int16x8_t c2 = vsubq_s16(b0, b2); in hadamard8x8_one_pass() 33 const int16x8_t c3 = vsubq_s16(b1, b3); in hadamard8x8_one_pass() 36 const int16x8_t c6 = vsubq_s16(b4, b6); in hadamard8x8_one_pass() 37 const int16x8_t c7 = vsubq_s16(b5, b7); in hadamard8x8_one_pass() 40 *a1 = vsubq_s16(c2, c6); in hadamard8x8_one_pass() 41 *a2 = vsubq_s16(c0, c4); in hadamard8x8_one_pass() [all …]
|
D | fwd_txfm_neon.c | 39 const int16x8_t s_32 = vsubq_s16(input_01, input_32); in aom_fdct4x4_helper() 129 const int16x8_t v_s4 = vsubq_s16(input_3, input_4); in aom_fdct8x8_neon() 130 const int16x8_t v_s5 = vsubq_s16(input_2, input_5); in aom_fdct8x8_neon() 131 const int16x8_t v_s6 = vsubq_s16(input_1, input_6); in aom_fdct8x8_neon() 132 const int16x8_t v_s7 = vsubq_s16(input_0, input_7); in aom_fdct8x8_neon() 136 int16x8_t v_x2 = vsubq_s16(v_s1, v_s2); in aom_fdct8x8_neon() 137 int16x8_t v_x3 = vsubq_s16(v_s0, v_s3); in aom_fdct8x8_neon() 170 v_x0 = vsubq_s16(v_s6, v_s5); in aom_fdct8x8_neon() 185 v_x1 = vsubq_s16(v_s4, ab); in aom_fdct8x8_neon() 186 v_x2 = vsubq_s16(v_s7, cd); in aom_fdct8x8_neon()
|
/external/libhevc/encoder/arm/ |
D | ihevce_had_compute_neon.c | 89 const int16x8_t a2 = vsubq_s16(*r0, *r2); in hadamard4x4_2_one_pass() 90 const int16x8_t a3 = vsubq_s16(*r1, *r3); in hadamard4x4_2_one_pass() 93 *r1 = vsubq_s16(a0, a1); in hadamard4x4_2_one_pass() 95 *r3 = vsubq_s16(a2, a3); in hadamard4x4_2_one_pass() 199 int16x8_t p1 = vsubq_s16(a[i], a[i + 4]); in hadamard_sad8x8_using4x4() 261 const int16x8_t a4 = vsubq_s16(*r0, *r4); in hadamard8x8_one_pass() 263 const int16x8_t a5 = vsubq_s16(*r1, *r5); in hadamard8x8_one_pass() 265 const int16x8_t a6 = vsubq_s16(*r2, *r6); in hadamard8x8_one_pass() 267 const int16x8_t a7 = vsubq_s16(*r3, *r7); in hadamard8x8_one_pass() 270 const int16x8_t b2 = vsubq_s16(a0, a2); in hadamard8x8_one_pass() [all …]
|
D | ihevce_scale_by_2_neon.c | 96 sum = vsubq_s16(p, prec); in ihevce_horz_scale_neon_w16() 97 q = vsubq_s16(q, r); in ihevce_horz_scale_neon_w16() 171 sum = vsubq_s16(p, prec); in ihevce_vert_scale_neon_w16() 172 q = vsubq_s16(q, r); in ihevce_vert_scale_neon_w16() 189 sum = vsubq_s16(p, prec); in ihevce_vert_scale_neon_w16() 190 q = vsubq_s16(q, r); in ihevce_vert_scale_neon_w16()
|
D | ihevce_common_utils_neon.c | 550 sign_reg0 = vsubq_s16(sign_reg0, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 556 sign_reg1 = vsubq_s16(sign_reg1, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 562 sign_reg2 = vsubq_s16(sign_reg2, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 568 sign_reg3 = vsubq_s16(sign_reg3, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 649 sign_reg0 = vsubq_s16(sign_reg0, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 654 sign_reg1 = vsubq_s16(sign_reg1, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 734 sign_reg0 = vsubq_s16(sign_reg0, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 740 sign_reg1 = vsubq_s16(sign_reg1, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 746 sign_reg2 = vsubq_s16(sign_reg2, sign_reg); in ihevce_get_luma_eo_sao_params_neon() 752 sign_reg3 = vsubq_s16(sign_reg3, sign_reg); in ihevce_get_luma_eo_sao_params_neon() [all …]
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht16x16_add_neon.c | 179 x[4] = vsubq_s16(t[0], t[4]); in vpx_iadst16x16_256_add_half1d() 180 x[5] = vsubq_s16(t[1], t[5]); in vpx_iadst16x16_256_add_half1d() 181 x[6] = vsubq_s16(t[2], t[6]); in vpx_iadst16x16_256_add_half1d() 182 x[7] = vsubq_s16(t[3], t[7]); in vpx_iadst16x16_256_add_half1d() 208 x[2] = vsubq_s16(t[0], t[2]); in vpx_iadst16x16_256_add_half1d() 209 x[3] = vsubq_s16(t[1], t[3]); in vpx_iadst16x16_256_add_half1d() 216 x[10] = vsubq_s16(t[8], t[10]); in vpx_iadst16x16_256_add_half1d() 217 x[11] = vsubq_s16(t[9], t[11]); in vpx_iadst16x16_256_add_half1d()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | mask_blend_neon.cc | 119 int16x8_t pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlending4x4_NEON() 130 pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlending4x4_NEON() 151 int16x8_t pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlending4xH_NEON() 161 pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlending4xH_NEON() 170 pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlending4xH_NEON() 179 pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlending4xH_NEON() 214 const int16x8_t pred_mask_1 = vsubq_s16(mask_inverter, pred_mask_0); in MaskBlend_NEON()
|
D | intrapred_cfl_neon.cc | 53 vst1q_s16(a[y], vsubq_s16(b, val_v)); in BlockSubtract() 57 vst1q_s16(a[y], vsubq_s16(b, val_v)); in BlockSubtract() 58 vst1q_s16(a[y] + 8, vsubq_s16(c, val_v)); in BlockSubtract() 64 vst1q_s16(a[y], vsubq_s16(b, val_v)); in BlockSubtract() 65 vst1q_s16(a[y] + 8, vsubq_s16(c, val_v)); in BlockSubtract() 66 vst1q_s16(a[y] + 16, vsubq_s16(d, val_v)); in BlockSubtract() 67 vst1q_s16(a[y] + 24, vsubq_s16(e, val_v)); in BlockSubtract() 627 vst1q_s16(luma_ptr, vsubq_s16(ssample, averages)); in CflSubsampler444_8xH_NEON() 636 vst1q_s16(luma_ptr, vsubq_s16(ssample, averages)); in CflSubsampler444_8xH_NEON() 731 vst1q_s16(&luma_ptr[x], vsubq_s16(ssamples[idx++], averages)); in CflSubsampler444_WxH_NEON() [all …]
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | shortfdct_neon.c | 150 q13s16 = vsubq_s16(v2tmp0.val[1], v2tmp1.val[0]); in vp8_short_fdct8x4_neon() 151 q14s16 = vsubq_s16(v2tmp0.val[0], v2tmp1.val[1]); in vp8_short_fdct8x4_neon() 159 q2s16 = vsubq_s16(q11s16, q12s16); in vp8_short_fdct8x4_neon() 201 q13s16 = vsubq_s16(v2tmp0.val[1], v2tmp1.val[0]); in vp8_short_fdct8x4_neon() 202 q14s16 = vsubq_s16(v2tmp0.val[0], v2tmp1.val[1]); in vp8_short_fdct8x4_neon() 207 q1s16 = vsubq_s16(q11s16, q12s16); in vp8_short_fdct8x4_neon()
|
/external/libhevc/common/arm/ |
D | ihevc_resi_trans_neon.c | 120 o_32 = vsubq_s16(diff_1, diff_2); in ihevc_resi_trans_4x4_neon() 461 const int16x8_t o3 = vsubq_s16(diff_16[3], diff_16[4]); /*C3 - C4*/ in ihevc_resi_trans_8x8_neon() 462 const int16x8_t o2 = vsubq_s16(diff_16[2], diff_16[5]); /*C2 - C5*/ in ihevc_resi_trans_8x8_neon() 463 const int16x8_t o1 = vsubq_s16(diff_16[1], diff_16[6]); /*C1 - C6*/ in ihevc_resi_trans_8x8_neon() 464 const int16x8_t o0 = vsubq_s16(diff_16[0], diff_16[7]); /*C0 - C7*/ in ihevc_resi_trans_8x8_neon() 472 const int16x8_t eo0 = vsubq_s16(e0, e3); /*C0 - C3 - C4 + C7*/ in ihevc_resi_trans_8x8_neon() 473 const int16x8_t eo1 = vsubq_s16(e1, e2); /*C1 - C2 - C5 + C6*/ in ihevc_resi_trans_8x8_neon() 478 const int16x8_t eeo = vsubq_s16(ee0, ee1); in ihevc_resi_trans_8x8_neon() 843 b[8] = vsubq_s16(a[7], a[8]); in cross_input_16() 844 b[9] = vsubq_s16(a[6], a[9]); in cross_input_16() [all …]
|
/external/libaom/libaom/av1/encoder/arm/neon/ |
D | quantize_neon.c | 69 const int16x8_t v_qcoeff = vsubq_s16(v_qcoeff_a, v_coeff_sign); in av1_quantize_fp_neon() 95 const int16x8_t v_qcoeff = vsubq_s16(v_qcoeff_a, v_coeff_sign); in av1_quantize_fp_neon() 167 const int16x8_t v_qcoeff = vsubq_s16(v_qcoeff_a, v_coeff_sign); in av1_quantize_lp_neon() 192 const int16x8_t v_qcoeff = vsubq_s16(v_qcoeff_a, v_coeff_sign); in av1_quantize_lp_neon()
|
D | av1_error_neon.c | 28 const int16x8_t diff = vsubq_s16(c, d); in av1_block_error_neon() 70 const int16x8_t diff = vsubq_s16(c, d); in av1_block_error_lp_neon()
|