/external/libvpx/libvpx/vpx_dsp/arm/ |
D | fdct_partial_neon.c | 53 int16x8_t sum = vld1q_s16(&input[0]); in vpx_fdct8x8_1_neon() 56 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in vpx_fdct8x8_1_neon() 67 int16x8_t left = vld1q_s16(input); in vpx_fdct16x16_1_neon() 68 int16x8_t right = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon() 73 const int16x8_t a = vld1q_s16(input); in vpx_fdct16x16_1_neon() 74 const int16x8_t b = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon() 89 int16x8_t a0 = vld1q_s16(input); in vpx_fdct32x32_1_neon() 90 int16x8_t a1 = vld1q_s16(input + 8); in vpx_fdct32x32_1_neon() 91 int16x8_t a2 = vld1q_s16(input + 16); in vpx_fdct32x32_1_neon() 92 int16x8_t a3 = vld1q_s16(input + 24); in vpx_fdct32x32_1_neon() [all …]
|
D | sum_squares_neon.c | 47 s[0] = vld1q_s16(src_t + 0 * stride); in vpx_sum_squares_2d_i16_neon() 48 s[1] = vld1q_s16(src_t + 1 * stride); in vpx_sum_squares_2d_i16_neon() 49 s[2] = vld1q_s16(src_t + 2 * stride); in vpx_sum_squares_2d_i16_neon() 50 s[3] = vld1q_s16(src_t + 3 * stride); in vpx_sum_squares_2d_i16_neon() 51 s[4] = vld1q_s16(src_t + 4 * stride); in vpx_sum_squares_2d_i16_neon() 52 s[5] = vld1q_s16(src_t + 5 * stride); in vpx_sum_squares_2d_i16_neon() 53 s[6] = vld1q_s16(src_t + 6 * stride); in vpx_sum_squares_2d_i16_neon() 54 s[7] = vld1q_s16(src_t + 7 * stride); in vpx_sum_squares_2d_i16_neon()
|
D | quantize_neon.c | 51 const int16x8_t zbin = vld1q_s16(zbin_ptr); in vpx_quantize_b_neon() 52 const int16x8_t round = vld1q_s16(round_ptr); in vpx_quantize_b_neon() 53 const int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_neon() 54 const int16x8_t quant_shift = vld1q_s16(quant_shift_ptr); in vpx_quantize_b_neon() 55 const int16x8_t dequant = vld1q_s16(dequant_ptr); in vpx_quantize_b_neon() 58 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vpx_quantize_b_neon() 108 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vpx_quantize_b_neon() 211 const int16x8_t zbin = vrshrq_n_s16(vld1q_s16(zbin_ptr), 1); in vpx_quantize_b_32x32_neon() 212 const int16x8_t round = vrshrq_n_s16(vld1q_s16(round_ptr), 1); in vpx_quantize_b_32x32_neon() 213 const int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_32x32_neon() [all …]
|
D | idct16x16_add_neon.c | 69 const int16x8_t cospis0 = vld1q_s16(kCospi); in vpx_idct16x16_256_add_half1d() 70 const int16x8_t cospis1 = vld1q_s16(kCospi + 8); in vpx_idct16x16_256_add_half1d() 113 in[0] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 115 in[8] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 117 in[1] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 119 in[9] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 121 in[2] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 123 in[10] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 125 in[3] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 127 in[11] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() [all …]
|
D | hadamard_neon.c | 52 int16x8_t a0 = vld1q_s16(src_diff); in vpx_hadamard_8x8_neon() 53 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in vpx_hadamard_8x8_neon() 54 int16x8_t a2 = vld1q_s16(src_diff + 2 * src_stride); in vpx_hadamard_8x8_neon() 55 int16x8_t a3 = vld1q_s16(src_diff + 3 * src_stride); in vpx_hadamard_8x8_neon() 56 int16x8_t a4 = vld1q_s16(src_diff + 4 * src_stride); in vpx_hadamard_8x8_neon() 57 int16x8_t a5 = vld1q_s16(src_diff + 5 * src_stride); in vpx_hadamard_8x8_neon() 58 int16x8_t a6 = vld1q_s16(src_diff + 6 * src_stride); in vpx_hadamard_8x8_neon() 59 int16x8_t a7 = vld1q_s16(src_diff + 7 * src_stride); in vpx_hadamard_8x8_neon()
|
D | fdct16x16_neon.c | 31 b[0] = vld1q_s16(a); in load() 33 b[1] = vld1q_s16(a); in load() 35 b[2] = vld1q_s16(a); in load() 37 b[3] = vld1q_s16(a); in load() 39 b[4] = vld1q_s16(a); in load() 41 b[5] = vld1q_s16(a); in load() 43 b[6] = vld1q_s16(a); in load() 45 b[7] = vld1q_s16(a); in load() 47 b[8] = vld1q_s16(a); in load() 49 b[9] = vld1q_s16(a); in load() [all …]
|
D | fwd_txfm_neon.c | 24 int16x8_t input_0 = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in vpx_fdct8x8_neon() 25 int16x8_t input_1 = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in vpx_fdct8x8_neon() 26 int16x8_t input_2 = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in vpx_fdct8x8_neon() 27 int16x8_t input_3 = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in vpx_fdct8x8_neon() 28 int16x8_t input_4 = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in vpx_fdct8x8_neon() 29 int16x8_t input_5 = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in vpx_fdct8x8_neon() 30 int16x8_t input_6 = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in vpx_fdct8x8_neon() 31 int16x8_t input_7 = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in vpx_fdct8x8_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | hadamard_neon.c | 51 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_8x8_neon() 52 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_8x8_neon() 53 int16x8_t a2 = vld1q_s16(src_diff + 2 * src_stride); in aom_hadamard_8x8_neon() 54 int16x8_t a3 = vld1q_s16(src_diff + 3 * src_stride); in aom_hadamard_8x8_neon() 55 int16x8_t a4 = vld1q_s16(src_diff + 4 * src_stride); in aom_hadamard_8x8_neon() 56 int16x8_t a5 = vld1q_s16(src_diff + 5 * src_stride); in aom_hadamard_8x8_neon() 57 int16x8_t a6 = vld1q_s16(src_diff + 6 * src_stride); in aom_hadamard_8x8_neon() 58 int16x8_t a7 = vld1q_s16(src_diff + 7 * src_stride); in aom_hadamard_8x8_neon() 80 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_lp_8x8_neon() 81 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_lp_8x8_neon() [all …]
|
D | fwd_txfm_neon.c | 115 int16x8_t input_0 = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in aom_fdct8x8_neon() 116 int16x8_t input_1 = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in aom_fdct8x8_neon() 117 int16x8_t input_2 = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in aom_fdct8x8_neon() 118 int16x8_t input_3 = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in aom_fdct8x8_neon() 119 int16x8_t input_4 = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in aom_fdct8x8_neon() 120 int16x8_t input_5 = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in aom_fdct8x8_neon() 121 int16x8_t input_6 = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in aom_fdct8x8_neon() 122 int16x8_t input_7 = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in aom_fdct8x8_neon() 303 int16x8_t sum = vld1q_s16(&input[0]); in aom_fdct8x8_1_neon() 305 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in aom_fdct8x8_1_neon()
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | fastquantizeb_neon.c | 20 const int16x8_t one_q = vdupq_n_s16(-1), z0 = vld1q_s16(b->coeff), in vp8_fast_quantize_b_neon() 21 z1 = vld1q_s16(b->coeff + 8), round0 = vld1q_s16(b->round), in vp8_fast_quantize_b_neon() 22 round1 = vld1q_s16(b->round + 8), in vp8_fast_quantize_b_neon() 23 quant0 = vld1q_s16(b->quant_fast), in vp8_fast_quantize_b_neon() 24 quant1 = vld1q_s16(b->quant_fast + 8), in vp8_fast_quantize_b_neon() 25 dequant0 = vld1q_s16(d->dequant), in vp8_fast_quantize_b_neon() 26 dequant1 = vld1q_s16(d->dequant + 8); in vp8_fast_quantize_b_neon()
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht16x16_add_neon.c | 78 in[0] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 80 in[8] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 82 in[1] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 84 in[9] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 86 in[2] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 88 in[10] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 90 in[3] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 92 in[11] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 94 in[4] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 96 in[12] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() [all …]
|
/external/libjpeg-turbo/simd/arm/ |
D | jidctred-neon.c | 82 int16x8_t row0 = vld1q_s16(coef_block + 0 * DCTSIZE); in jsimd_idct_2x2_neon() 83 int16x8_t row1 = vld1q_s16(coef_block + 1 * DCTSIZE); in jsimd_idct_2x2_neon() 84 int16x8_t row3 = vld1q_s16(coef_block + 3 * DCTSIZE); in jsimd_idct_2x2_neon() 85 int16x8_t row5 = vld1q_s16(coef_block + 5 * DCTSIZE); in jsimd_idct_2x2_neon() 86 int16x8_t row7 = vld1q_s16(coef_block + 7 * DCTSIZE); in jsimd_idct_2x2_neon() 89 int16x8_t quant_row0 = vld1q_s16(quantptr + 0 * DCTSIZE); in jsimd_idct_2x2_neon() 90 int16x8_t quant_row1 = vld1q_s16(quantptr + 1 * DCTSIZE); in jsimd_idct_2x2_neon() 91 int16x8_t quant_row3 = vld1q_s16(quantptr + 3 * DCTSIZE); in jsimd_idct_2x2_neon() 92 int16x8_t quant_row5 = vld1q_s16(quantptr + 5 * DCTSIZE); in jsimd_idct_2x2_neon() 93 int16x8_t quant_row7 = vld1q_s16(quantptr + 7 * DCTSIZE); in jsimd_idct_2x2_neon() [all …]
|
D | jquanti-neon.c | 105 int16x8_t row0 = vld1q_s16(workspace + (i + 0) * DCTSIZE); in jsimd_quantize_neon() 106 int16x8_t row1 = vld1q_s16(workspace + (i + 1) * DCTSIZE); in jsimd_quantize_neon() 107 int16x8_t row2 = vld1q_s16(workspace + (i + 2) * DCTSIZE); in jsimd_quantize_neon() 108 int16x8_t row3 = vld1q_s16(workspace + (i + 3) * DCTSIZE); in jsimd_quantize_neon() 118 int16x8_t shift0 = vld1q_s16(shift_ptr + (i + 0) * DCTSIZE); in jsimd_quantize_neon() 119 int16x8_t shift1 = vld1q_s16(shift_ptr + (i + 1) * DCTSIZE); in jsimd_quantize_neon() 120 int16x8_t shift2 = vld1q_s16(shift_ptr + (i + 2) * DCTSIZE); in jsimd_quantize_neon() 121 int16x8_t shift3 = vld1q_s16(shift_ptr + (i + 3) * DCTSIZE); in jsimd_quantize_neon()
|
D | jidctfst-neon.c | 69 int16x8_t row0 = vld1q_s16(coef_block + 0 * DCTSIZE); in jsimd_idct_ifast_neon() 70 int16x8_t row1 = vld1q_s16(coef_block + 1 * DCTSIZE); in jsimd_idct_ifast_neon() 71 int16x8_t row2 = vld1q_s16(coef_block + 2 * DCTSIZE); in jsimd_idct_ifast_neon() 72 int16x8_t row3 = vld1q_s16(coef_block + 3 * DCTSIZE); in jsimd_idct_ifast_neon() 73 int16x8_t row4 = vld1q_s16(coef_block + 4 * DCTSIZE); in jsimd_idct_ifast_neon() 74 int16x8_t row5 = vld1q_s16(coef_block + 5 * DCTSIZE); in jsimd_idct_ifast_neon() 75 int16x8_t row6 = vld1q_s16(coef_block + 6 * DCTSIZE); in jsimd_idct_ifast_neon() 76 int16x8_t row7 = vld1q_s16(coef_block + 7 * DCTSIZE); in jsimd_idct_ifast_neon() 79 int16x8_t quant_row0 = vld1q_s16(quantptr + 0 * DCTSIZE); in jsimd_idct_ifast_neon() 263 int16x8_t quant_row1 = vld1q_s16(quantptr + 1 * DCTSIZE); in jsimd_idct_ifast_neon() [all …]
|
D | jcphuff-neon.c | 198 int16x8_t row0 = vld1q_s16(values + 0 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 199 int16x8_t row1 = vld1q_s16(values + 1 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 200 int16x8_t row2 = vld1q_s16(values + 2 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 201 int16x8_t row3 = vld1q_s16(values + 3 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 202 int16x8_t row4 = vld1q_s16(values + 4 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 203 int16x8_t row5 = vld1q_s16(values + 5 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 204 int16x8_t row6 = vld1q_s16(values + 6 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 205 int16x8_t row7 = vld1q_s16(values + 7 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() 443 int16x8_t abs_row0 = vld1q_s16(absvalues + 0 * DCTSIZE); in jsimd_encode_mcu_AC_refine_prepare_neon() 444 int16x8_t abs_row1 = vld1q_s16(absvalues + 1 * DCTSIZE); in jsimd_encode_mcu_AC_refine_prepare_neon() [all …]
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | average_blend_neon.cc | 45 const int16x8_t pred0 = vld1q_s16(prediction_0); in AverageBlend8Row() 46 const int16x8_t pred1 = vld1q_s16(prediction_1); in AverageBlend8Row() 56 const int16x8_t pred_00 = vld1q_s16(prediction_0); in AverageBlendLargeRow() 57 const int16x8_t pred_01 = vld1q_s16(prediction_1); in AverageBlendLargeRow() 62 const int16x8_t pred_10 = vld1q_s16(prediction_0); in AverageBlendLargeRow() 63 const int16x8_t pred_11 = vld1q_s16(prediction_1); in AverageBlendLargeRow()
|
D | distance_weighted_blend_neon.cc | 64 const int16x8_t src_00 = vld1q_s16(prediction_0); in DistanceWeightedBlendSmall_NEON() 65 const int16x8_t src_10 = vld1q_s16(prediction_1); in DistanceWeightedBlendSmall_NEON() 70 const int16x8_t src_01 = vld1q_s16(prediction_0); in DistanceWeightedBlendSmall_NEON() 71 const int16x8_t src_11 = vld1q_s16(prediction_1); in DistanceWeightedBlendSmall_NEON() 109 const int16x8_t src0_lo = vld1q_s16(prediction_0 + x); in DistanceWeightedBlendLarge_NEON() 110 const int16x8_t src1_lo = vld1q_s16(prediction_1 + x); in DistanceWeightedBlendLarge_NEON() 114 const int16x8_t src0_hi = vld1q_s16(prediction_0 + x + 8); in DistanceWeightedBlendLarge_NEON() 115 const int16x8_t src1_hi = vld1q_s16(prediction_1 + x + 8); in DistanceWeightedBlendLarge_NEON()
|
D | intrapred_cfl_neon.cc | 52 const int16x8_t b = vld1q_s16(a[y]); in BlockSubtract() 55 const int16x8_t b = vld1q_s16(a[y]); in BlockSubtract() 56 const int16x8_t c = vld1q_s16(a[y] + 8); in BlockSubtract() 60 const int16x8_t b = vld1q_s16(a[y]); in BlockSubtract() 61 const int16x8_t c = vld1q_s16(a[y] + 8); in BlockSubtract() 62 const int16x8_t d = vld1q_s16(a[y] + 16); in BlockSubtract() 63 const int16x8_t e = vld1q_s16(a[y] + 24); in BlockSubtract() 304 const int16x8_t luma_row = vld1q_s16(luma[y]); in CflIntraPredictor8xN_NEON() 319 const int16x8_t luma_row_0 = vld1q_s16(luma[y]); in CflIntraPredictor16xN_NEON() 320 const int16x8_t luma_row_1 = vld1q_s16(luma[y] + 8); in CflIntraPredictor16xN_NEON() [all …]
|
D | warp_neon.cc | 265 vld1q_s16(&intermediate_result_column[y]); in Warp_NEON() 271 const int16x8_t filter = vld1q_s16(kWarpedFilters[offset]); in Warp_NEON() 289 const int16x8_t sum = vld1q_s16(tmp); in Warp_NEON() 298 filter[x] = vld1q_s16(kWarpedFilters[offset]); in Warp_NEON() 400 filter[x] = vld1q_s16(kWarpedFilters[offset]); in Warp_NEON() 407 const int16x8_t intermediate = vld1q_s16(intermediate_result[y + k]); in Warp_NEON()
|
/external/webp/src/dsp/ |
D | cost_neon.c | 25 const int16x8_t coeffs_0 = vld1q_s16(coeffs); in SetResidualCoeffs_NEON() 26 const int16x8_t coeffs_1 = vld1q_s16(coeffs + 8); in SetResidualCoeffs_NEON() 71 const int16x8_t c0 = vld1q_s16(res->coeffs); in GetResidualCost_NEON() 72 const int16x8_t c1 = vld1q_s16(res->coeffs + 8); in GetResidualCost_NEON()
|
/external/libaom/libaom/av1/common/arm/ |
D | wiener_convolve_neon.c | 409 s0 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 411 s1 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 413 s2 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 415 s3 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 417 s4 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 419 s5 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 421 s6 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 433 s7 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 435 s8 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 437 s9 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() [all …]
|
D | mem_neon.h | 290 *s0 = vld1q_s16(s); in load_s16_8x8() 292 *s1 = vld1q_s16(s); in load_s16_8x8() 294 *s2 = vld1q_s16(s); in load_s16_8x8() 296 *s3 = vld1q_s16(s); in load_s16_8x8() 298 *s4 = vld1q_s16(s); in load_s16_8x8() 300 *s5 = vld1q_s16(s); in load_s16_8x8() 302 *s6 = vld1q_s16(s); in load_s16_8x8() 304 *s7 = vld1q_s16(s); in load_s16_8x8() 310 *s0 = vld1q_s16(s); in load_s16_8x4() 312 *s1 = vld1q_s16(s); in load_s16_8x4() [all …]
|
/external/libvpx/libvpx/vp9/encoder/arm/neon/ |
D | vp9_quantize_neon.c | 71 const int16x8_t v_iscan = vld1q_s16(&iscan[0]); in vp9_quantize_fp_neon() 96 const int16x8_t v_iscan = vld1q_s16(&iscan[i]); in vp9_quantize_fp_neon() 150 const int16x8_t round = vrshrq_n_s16(vld1q_s16(round_ptr), 1); in vp9_quantize_fp_32x32_neon() 151 const int16x8_t quant = vld1q_s16(quant_ptr); in vp9_quantize_fp_32x32_neon() 154 const int16x8_t dequant_thresh = vshrq_n_s16(vld1q_s16(dequant_ptr), 2); in vp9_quantize_fp_32x32_neon() 158 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vp9_quantize_fp_32x32_neon() 215 vreinterpretq_u16_s16(vaddq_s16(vld1q_s16(iscan), one)); in vp9_quantize_fp_32x32_neon()
|
/external/libaom/libaom/av1/encoder/arm/neon/ |
D | quantize_neon.c | 54 const int16x8_t v_iscan = vld1q_s16(&iscan[0]); in av1_quantize_fp_neon() 80 const int16x8_t v_iscan = vld1q_s16(&iscan[i]); in av1_quantize_fp_neon() 152 const int16x8_t v_iscan = vld1q_s16(&scan[0]); in av1_quantize_lp_neon() 153 const int16x8_t v_coeff = vld1q_s16(coeff_ptr); in av1_quantize_lp_neon() 177 const int16x8_t v_iscan = vld1q_s16(&scan[i]); in av1_quantize_lp_neon() 178 const int16x8_t v_coeff = vld1q_s16(coeff_ptr + i); in av1_quantize_lp_neon()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | dequant_idct_neon.c | 35 q3 = vld1q_s16(input); in vp8_dequant_idct_add_neon() 38 q4 = vld1q_s16(input); in vp8_dequant_idct_add_neon() 42 q5 = vld1q_s16(dq); in vp8_dequant_idct_add_neon() 44 q6 = vld1q_s16(dq); in vp8_dequant_idct_add_neon()
|