Home
last modified time | relevance | path

Searched refs:vld1q_s16 (Results 1 – 25 of 92) sorted by relevance

1234

/external/libvpx/libvpx/vpx_dsp/arm/
Dfdct_partial_neon.c53 int16x8_t sum = vld1q_s16(&input[0]); in vpx_fdct8x8_1_neon()
56 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in vpx_fdct8x8_1_neon()
67 int16x8_t left = vld1q_s16(input); in vpx_fdct16x16_1_neon()
68 int16x8_t right = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon()
73 const int16x8_t a = vld1q_s16(input); in vpx_fdct16x16_1_neon()
74 const int16x8_t b = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon()
89 int16x8_t a0 = vld1q_s16(input); in vpx_fdct32x32_1_neon()
90 int16x8_t a1 = vld1q_s16(input + 8); in vpx_fdct32x32_1_neon()
91 int16x8_t a2 = vld1q_s16(input + 16); in vpx_fdct32x32_1_neon()
92 int16x8_t a3 = vld1q_s16(input + 24); in vpx_fdct32x32_1_neon()
[all …]
Dsum_squares_neon.c47 s[0] = vld1q_s16(src_t + 0 * stride); in vpx_sum_squares_2d_i16_neon()
48 s[1] = vld1q_s16(src_t + 1 * stride); in vpx_sum_squares_2d_i16_neon()
49 s[2] = vld1q_s16(src_t + 2 * stride); in vpx_sum_squares_2d_i16_neon()
50 s[3] = vld1q_s16(src_t + 3 * stride); in vpx_sum_squares_2d_i16_neon()
51 s[4] = vld1q_s16(src_t + 4 * stride); in vpx_sum_squares_2d_i16_neon()
52 s[5] = vld1q_s16(src_t + 5 * stride); in vpx_sum_squares_2d_i16_neon()
53 s[6] = vld1q_s16(src_t + 6 * stride); in vpx_sum_squares_2d_i16_neon()
54 s[7] = vld1q_s16(src_t + 7 * stride); in vpx_sum_squares_2d_i16_neon()
Dquantize_neon.c51 const int16x8_t zbin = vld1q_s16(zbin_ptr); in vpx_quantize_b_neon()
52 const int16x8_t round = vld1q_s16(round_ptr); in vpx_quantize_b_neon()
53 const int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_neon()
54 const int16x8_t quant_shift = vld1q_s16(quant_shift_ptr); in vpx_quantize_b_neon()
55 const int16x8_t dequant = vld1q_s16(dequant_ptr); in vpx_quantize_b_neon()
58 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vpx_quantize_b_neon()
108 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vpx_quantize_b_neon()
211 const int16x8_t zbin = vrshrq_n_s16(vld1q_s16(zbin_ptr), 1); in vpx_quantize_b_32x32_neon()
212 const int16x8_t round = vrshrq_n_s16(vld1q_s16(round_ptr), 1); in vpx_quantize_b_32x32_neon()
213 const int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_32x32_neon()
[all …]
Didct16x16_add_neon.c69 const int16x8_t cospis0 = vld1q_s16(kCospi); in vpx_idct16x16_256_add_half1d()
70 const int16x8_t cospis1 = vld1q_s16(kCospi + 8); in vpx_idct16x16_256_add_half1d()
113 in[0] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
115 in[8] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
117 in[1] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
119 in[9] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
121 in[2] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
123 in[10] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
125 in[3] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
127 in[11] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
[all …]
Dhadamard_neon.c52 int16x8_t a0 = vld1q_s16(src_diff); in vpx_hadamard_8x8_neon()
53 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in vpx_hadamard_8x8_neon()
54 int16x8_t a2 = vld1q_s16(src_diff + 2 * src_stride); in vpx_hadamard_8x8_neon()
55 int16x8_t a3 = vld1q_s16(src_diff + 3 * src_stride); in vpx_hadamard_8x8_neon()
56 int16x8_t a4 = vld1q_s16(src_diff + 4 * src_stride); in vpx_hadamard_8x8_neon()
57 int16x8_t a5 = vld1q_s16(src_diff + 5 * src_stride); in vpx_hadamard_8x8_neon()
58 int16x8_t a6 = vld1q_s16(src_diff + 6 * src_stride); in vpx_hadamard_8x8_neon()
59 int16x8_t a7 = vld1q_s16(src_diff + 7 * src_stride); in vpx_hadamard_8x8_neon()
Dfdct16x16_neon.c31 b[0] = vld1q_s16(a); in load()
33 b[1] = vld1q_s16(a); in load()
35 b[2] = vld1q_s16(a); in load()
37 b[3] = vld1q_s16(a); in load()
39 b[4] = vld1q_s16(a); in load()
41 b[5] = vld1q_s16(a); in load()
43 b[6] = vld1q_s16(a); in load()
45 b[7] = vld1q_s16(a); in load()
47 b[8] = vld1q_s16(a); in load()
49 b[9] = vld1q_s16(a); in load()
[all …]
Dfwd_txfm_neon.c24 int16x8_t input_0 = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in vpx_fdct8x8_neon()
25 int16x8_t input_1 = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in vpx_fdct8x8_neon()
26 int16x8_t input_2 = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in vpx_fdct8x8_neon()
27 int16x8_t input_3 = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in vpx_fdct8x8_neon()
28 int16x8_t input_4 = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in vpx_fdct8x8_neon()
29 int16x8_t input_5 = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in vpx_fdct8x8_neon()
30 int16x8_t input_6 = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in vpx_fdct8x8_neon()
31 int16x8_t input_7 = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in vpx_fdct8x8_neon()
/external/libaom/libaom/aom_dsp/arm/
Dhadamard_neon.c51 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_8x8_neon()
52 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_8x8_neon()
53 int16x8_t a2 = vld1q_s16(src_diff + 2 * src_stride); in aom_hadamard_8x8_neon()
54 int16x8_t a3 = vld1q_s16(src_diff + 3 * src_stride); in aom_hadamard_8x8_neon()
55 int16x8_t a4 = vld1q_s16(src_diff + 4 * src_stride); in aom_hadamard_8x8_neon()
56 int16x8_t a5 = vld1q_s16(src_diff + 5 * src_stride); in aom_hadamard_8x8_neon()
57 int16x8_t a6 = vld1q_s16(src_diff + 6 * src_stride); in aom_hadamard_8x8_neon()
58 int16x8_t a7 = vld1q_s16(src_diff + 7 * src_stride); in aom_hadamard_8x8_neon()
80 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_lp_8x8_neon()
81 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_lp_8x8_neon()
[all …]
Dfwd_txfm_neon.c115 int16x8_t input_0 = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in aom_fdct8x8_neon()
116 int16x8_t input_1 = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in aom_fdct8x8_neon()
117 int16x8_t input_2 = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in aom_fdct8x8_neon()
118 int16x8_t input_3 = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in aom_fdct8x8_neon()
119 int16x8_t input_4 = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in aom_fdct8x8_neon()
120 int16x8_t input_5 = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in aom_fdct8x8_neon()
121 int16x8_t input_6 = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in aom_fdct8x8_neon()
122 int16x8_t input_7 = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in aom_fdct8x8_neon()
303 int16x8_t sum = vld1q_s16(&input[0]); in aom_fdct8x8_1_neon()
305 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in aom_fdct8x8_1_neon()
/external/libvpx/libvpx/vp8/encoder/arm/neon/
Dfastquantizeb_neon.c20 const int16x8_t one_q = vdupq_n_s16(-1), z0 = vld1q_s16(b->coeff), in vp8_fast_quantize_b_neon()
21 z1 = vld1q_s16(b->coeff + 8), round0 = vld1q_s16(b->round), in vp8_fast_quantize_b_neon()
22 round1 = vld1q_s16(b->round + 8), in vp8_fast_quantize_b_neon()
23 quant0 = vld1q_s16(b->quant_fast), in vp8_fast_quantize_b_neon()
24 quant1 = vld1q_s16(b->quant_fast + 8), in vp8_fast_quantize_b_neon()
25 dequant0 = vld1q_s16(d->dequant), in vp8_fast_quantize_b_neon()
26 dequant1 = vld1q_s16(d->dequant + 8); in vp8_fast_quantize_b_neon()
/external/libvpx/libvpx/vp9/common/arm/neon/
Dvp9_iht16x16_add_neon.c78 in[0] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
80 in[8] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
82 in[1] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
84 in[9] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
86 in[2] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
88 in[10] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
90 in[3] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
92 in[11] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
94 in[4] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
96 in[12] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
[all …]
/external/libjpeg-turbo/simd/arm/
Djidctred-neon.c82 int16x8_t row0 = vld1q_s16(coef_block + 0 * DCTSIZE); in jsimd_idct_2x2_neon()
83 int16x8_t row1 = vld1q_s16(coef_block + 1 * DCTSIZE); in jsimd_idct_2x2_neon()
84 int16x8_t row3 = vld1q_s16(coef_block + 3 * DCTSIZE); in jsimd_idct_2x2_neon()
85 int16x8_t row5 = vld1q_s16(coef_block + 5 * DCTSIZE); in jsimd_idct_2x2_neon()
86 int16x8_t row7 = vld1q_s16(coef_block + 7 * DCTSIZE); in jsimd_idct_2x2_neon()
89 int16x8_t quant_row0 = vld1q_s16(quantptr + 0 * DCTSIZE); in jsimd_idct_2x2_neon()
90 int16x8_t quant_row1 = vld1q_s16(quantptr + 1 * DCTSIZE); in jsimd_idct_2x2_neon()
91 int16x8_t quant_row3 = vld1q_s16(quantptr + 3 * DCTSIZE); in jsimd_idct_2x2_neon()
92 int16x8_t quant_row5 = vld1q_s16(quantptr + 5 * DCTSIZE); in jsimd_idct_2x2_neon()
93 int16x8_t quant_row7 = vld1q_s16(quantptr + 7 * DCTSIZE); in jsimd_idct_2x2_neon()
[all …]
Djquanti-neon.c105 int16x8_t row0 = vld1q_s16(workspace + (i + 0) * DCTSIZE); in jsimd_quantize_neon()
106 int16x8_t row1 = vld1q_s16(workspace + (i + 1) * DCTSIZE); in jsimd_quantize_neon()
107 int16x8_t row2 = vld1q_s16(workspace + (i + 2) * DCTSIZE); in jsimd_quantize_neon()
108 int16x8_t row3 = vld1q_s16(workspace + (i + 3) * DCTSIZE); in jsimd_quantize_neon()
118 int16x8_t shift0 = vld1q_s16(shift_ptr + (i + 0) * DCTSIZE); in jsimd_quantize_neon()
119 int16x8_t shift1 = vld1q_s16(shift_ptr + (i + 1) * DCTSIZE); in jsimd_quantize_neon()
120 int16x8_t shift2 = vld1q_s16(shift_ptr + (i + 2) * DCTSIZE); in jsimd_quantize_neon()
121 int16x8_t shift3 = vld1q_s16(shift_ptr + (i + 3) * DCTSIZE); in jsimd_quantize_neon()
Djidctfst-neon.c69 int16x8_t row0 = vld1q_s16(coef_block + 0 * DCTSIZE); in jsimd_idct_ifast_neon()
70 int16x8_t row1 = vld1q_s16(coef_block + 1 * DCTSIZE); in jsimd_idct_ifast_neon()
71 int16x8_t row2 = vld1q_s16(coef_block + 2 * DCTSIZE); in jsimd_idct_ifast_neon()
72 int16x8_t row3 = vld1q_s16(coef_block + 3 * DCTSIZE); in jsimd_idct_ifast_neon()
73 int16x8_t row4 = vld1q_s16(coef_block + 4 * DCTSIZE); in jsimd_idct_ifast_neon()
74 int16x8_t row5 = vld1q_s16(coef_block + 5 * DCTSIZE); in jsimd_idct_ifast_neon()
75 int16x8_t row6 = vld1q_s16(coef_block + 6 * DCTSIZE); in jsimd_idct_ifast_neon()
76 int16x8_t row7 = vld1q_s16(coef_block + 7 * DCTSIZE); in jsimd_idct_ifast_neon()
79 int16x8_t quant_row0 = vld1q_s16(quantptr + 0 * DCTSIZE); in jsimd_idct_ifast_neon()
263 int16x8_t quant_row1 = vld1q_s16(quantptr + 1 * DCTSIZE); in jsimd_idct_ifast_neon()
[all …]
Djcphuff-neon.c198 int16x8_t row0 = vld1q_s16(values + 0 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
199 int16x8_t row1 = vld1q_s16(values + 1 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
200 int16x8_t row2 = vld1q_s16(values + 2 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
201 int16x8_t row3 = vld1q_s16(values + 3 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
202 int16x8_t row4 = vld1q_s16(values + 4 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
203 int16x8_t row5 = vld1q_s16(values + 5 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
204 int16x8_t row6 = vld1q_s16(values + 6 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
205 int16x8_t row7 = vld1q_s16(values + 7 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon()
443 int16x8_t abs_row0 = vld1q_s16(absvalues + 0 * DCTSIZE); in jsimd_encode_mcu_AC_refine_prepare_neon()
444 int16x8_t abs_row1 = vld1q_s16(absvalues + 1 * DCTSIZE); in jsimd_encode_mcu_AC_refine_prepare_neon()
[all …]
/external/libgav1/libgav1/src/dsp/arm/
Daverage_blend_neon.cc45 const int16x8_t pred0 = vld1q_s16(prediction_0); in AverageBlend8Row()
46 const int16x8_t pred1 = vld1q_s16(prediction_1); in AverageBlend8Row()
56 const int16x8_t pred_00 = vld1q_s16(prediction_0); in AverageBlendLargeRow()
57 const int16x8_t pred_01 = vld1q_s16(prediction_1); in AverageBlendLargeRow()
62 const int16x8_t pred_10 = vld1q_s16(prediction_0); in AverageBlendLargeRow()
63 const int16x8_t pred_11 = vld1q_s16(prediction_1); in AverageBlendLargeRow()
Ddistance_weighted_blend_neon.cc64 const int16x8_t src_00 = vld1q_s16(prediction_0); in DistanceWeightedBlendSmall_NEON()
65 const int16x8_t src_10 = vld1q_s16(prediction_1); in DistanceWeightedBlendSmall_NEON()
70 const int16x8_t src_01 = vld1q_s16(prediction_0); in DistanceWeightedBlendSmall_NEON()
71 const int16x8_t src_11 = vld1q_s16(prediction_1); in DistanceWeightedBlendSmall_NEON()
109 const int16x8_t src0_lo = vld1q_s16(prediction_0 + x); in DistanceWeightedBlendLarge_NEON()
110 const int16x8_t src1_lo = vld1q_s16(prediction_1 + x); in DistanceWeightedBlendLarge_NEON()
114 const int16x8_t src0_hi = vld1q_s16(prediction_0 + x + 8); in DistanceWeightedBlendLarge_NEON()
115 const int16x8_t src1_hi = vld1q_s16(prediction_1 + x + 8); in DistanceWeightedBlendLarge_NEON()
Dintrapred_cfl_neon.cc52 const int16x8_t b = vld1q_s16(a[y]); in BlockSubtract()
55 const int16x8_t b = vld1q_s16(a[y]); in BlockSubtract()
56 const int16x8_t c = vld1q_s16(a[y] + 8); in BlockSubtract()
60 const int16x8_t b = vld1q_s16(a[y]); in BlockSubtract()
61 const int16x8_t c = vld1q_s16(a[y] + 8); in BlockSubtract()
62 const int16x8_t d = vld1q_s16(a[y] + 16); in BlockSubtract()
63 const int16x8_t e = vld1q_s16(a[y] + 24); in BlockSubtract()
304 const int16x8_t luma_row = vld1q_s16(luma[y]); in CflIntraPredictor8xN_NEON()
319 const int16x8_t luma_row_0 = vld1q_s16(luma[y]); in CflIntraPredictor16xN_NEON()
320 const int16x8_t luma_row_1 = vld1q_s16(luma[y] + 8); in CflIntraPredictor16xN_NEON()
[all …]
Dwarp_neon.cc265 vld1q_s16(&intermediate_result_column[y]); in Warp_NEON()
271 const int16x8_t filter = vld1q_s16(kWarpedFilters[offset]); in Warp_NEON()
289 const int16x8_t sum = vld1q_s16(tmp); in Warp_NEON()
298 filter[x] = vld1q_s16(kWarpedFilters[offset]); in Warp_NEON()
400 filter[x] = vld1q_s16(kWarpedFilters[offset]); in Warp_NEON()
407 const int16x8_t intermediate = vld1q_s16(intermediate_result[y + k]); in Warp_NEON()
/external/webp/src/dsp/
Dcost_neon.c25 const int16x8_t coeffs_0 = vld1q_s16(coeffs); in SetResidualCoeffs_NEON()
26 const int16x8_t coeffs_1 = vld1q_s16(coeffs + 8); in SetResidualCoeffs_NEON()
71 const int16x8_t c0 = vld1q_s16(res->coeffs); in GetResidualCost_NEON()
72 const int16x8_t c1 = vld1q_s16(res->coeffs + 8); in GetResidualCost_NEON()
/external/libaom/libaom/av1/common/arm/
Dwiener_convolve_neon.c409 s0 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
411 s1 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
413 s2 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
415 s3 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
417 s4 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
419 s5 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
421 s6 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
433 s7 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
435 s8 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
437 s9 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
[all …]
Dmem_neon.h290 *s0 = vld1q_s16(s); in load_s16_8x8()
292 *s1 = vld1q_s16(s); in load_s16_8x8()
294 *s2 = vld1q_s16(s); in load_s16_8x8()
296 *s3 = vld1q_s16(s); in load_s16_8x8()
298 *s4 = vld1q_s16(s); in load_s16_8x8()
300 *s5 = vld1q_s16(s); in load_s16_8x8()
302 *s6 = vld1q_s16(s); in load_s16_8x8()
304 *s7 = vld1q_s16(s); in load_s16_8x8()
310 *s0 = vld1q_s16(s); in load_s16_8x4()
312 *s1 = vld1q_s16(s); in load_s16_8x4()
[all …]
/external/libvpx/libvpx/vp9/encoder/arm/neon/
Dvp9_quantize_neon.c71 const int16x8_t v_iscan = vld1q_s16(&iscan[0]); in vp9_quantize_fp_neon()
96 const int16x8_t v_iscan = vld1q_s16(&iscan[i]); in vp9_quantize_fp_neon()
150 const int16x8_t round = vrshrq_n_s16(vld1q_s16(round_ptr), 1); in vp9_quantize_fp_32x32_neon()
151 const int16x8_t quant = vld1q_s16(quant_ptr); in vp9_quantize_fp_32x32_neon()
154 const int16x8_t dequant_thresh = vshrq_n_s16(vld1q_s16(dequant_ptr), 2); in vp9_quantize_fp_32x32_neon()
158 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vp9_quantize_fp_32x32_neon()
215 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vp9_quantize_fp_32x32_neon()
/external/libaom/libaom/av1/encoder/arm/neon/
Dquantize_neon.c54 const int16x8_t v_iscan = vld1q_s16(&iscan[0]); in av1_quantize_fp_neon()
80 const int16x8_t v_iscan = vld1q_s16(&iscan[i]); in av1_quantize_fp_neon()
152 const int16x8_t v_iscan = vld1q_s16(&scan[0]); in av1_quantize_lp_neon()
153 const int16x8_t v_coeff = vld1q_s16(coeff_ptr); in av1_quantize_lp_neon()
177 const int16x8_t v_iscan = vld1q_s16(&scan[i]); in av1_quantize_lp_neon()
178 const int16x8_t v_coeff = vld1q_s16(coeff_ptr + i); in av1_quantize_lp_neon()
/external/libvpx/libvpx/vp8/common/arm/neon/
Ddequant_idct_neon.c35 q3 = vld1q_s16(input); in vp8_dequant_idct_add_neon()
38 q4 = vld1q_s16(input); in vp8_dequant_idct_add_neon()
42 q5 = vld1q_s16(dq); in vp8_dequant_idct_add_neon()
44 q6 = vld1q_s16(dq); in vp8_dequant_idct_add_neon()

1234