/external/libvpx/vpx_dsp/arm/ |
D | fdct_partial_neon.c | 42 int16x8_t sum = vld1q_s16(&input[0]); in vpx_fdct8x8_1_neon() 45 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in vpx_fdct8x8_1_neon() 56 int16x8_t left = vld1q_s16(input); in vpx_fdct16x16_1_neon() 57 int16x8_t right = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon() 62 const int16x8_t a = vld1q_s16(input); in vpx_fdct16x16_1_neon() 63 const int16x8_t b = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon() 78 int16x8_t a0 = vld1q_s16(input); in vpx_fdct32x32_1_neon() 79 int16x8_t a1 = vld1q_s16(input + 8); in vpx_fdct32x32_1_neon() 80 int16x8_t a2 = vld1q_s16(input + 16); in vpx_fdct32x32_1_neon() 81 int16x8_t a3 = vld1q_s16(input + 24); in vpx_fdct32x32_1_neon() [all …]
|
D | fdct8x8_neon.c | 26 in[0] = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in vpx_fdct8x8_neon() 27 in[1] = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in vpx_fdct8x8_neon() 28 in[2] = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in vpx_fdct8x8_neon() 29 in[3] = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in vpx_fdct8x8_neon() 30 in[4] = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in vpx_fdct8x8_neon() 31 in[5] = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in vpx_fdct8x8_neon() 32 in[6] = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in vpx_fdct8x8_neon() 33 in[7] = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in vpx_fdct8x8_neon() 77 in[0] = vld1q_s16(input + 0 * stride); in vpx_highbd_fdct8x8_neon() 78 in[1] = vld1q_s16(input + 1 * stride); in vpx_highbd_fdct8x8_neon() [all …]
|
D | fdct16x16_neon.h | 19 b[0] = vld1q_s16(a); in load() 21 b[1] = vld1q_s16(a); in load() 23 b[2] = vld1q_s16(a); in load() 25 b[3] = vld1q_s16(a); in load() 27 b[4] = vld1q_s16(a); in load() 29 b[5] = vld1q_s16(a); in load() 31 b[6] = vld1q_s16(a); in load() 33 b[7] = vld1q_s16(a); in load() 35 b[8] = vld1q_s16(a); in load() 37 b[9] = vld1q_s16(a); in load() [all …]
|
D | quantize_neon.c | 82 int16x8_t zbin = vld1q_s16(zbin_ptr); in vpx_quantize_b_neon() 83 int16x8_t round = vld1q_s16(round_ptr); in vpx_quantize_b_neon() 84 int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_neon() 85 int16x8_t quant_shift = vld1q_s16(quant_shift_ptr); in vpx_quantize_b_neon() 86 int16x8_t dequant = vld1q_s16(dequant_ptr); in vpx_quantize_b_neon() 90 const uint16x8_t v_iscan = vreinterpretq_u16_s16(vld1q_s16(iscan)); in vpx_quantize_b_neon() 116 const uint16x8_t v_iscan = vreinterpretq_u16_s16(vld1q_s16(iscan)); in vpx_quantize_b_neon() 229 int16x8_t zbin = vrshrq_n_s16(vld1q_s16(zbin_ptr), 1); in vpx_quantize_b_32x32_neon() 230 int16x8_t round = vrshrq_n_s16(vld1q_s16(round_ptr), 1); in vpx_quantize_b_32x32_neon() 231 int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_32x32_neon() [all …]
|
D | idct16x16_add_neon.c | 69 const int16x8_t cospis0 = vld1q_s16(kCospi); in vpx_idct16x16_256_add_half1d() 70 const int16x8_t cospis1 = vld1q_s16(kCospi + 8); in vpx_idct16x16_256_add_half1d() 113 in[0] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 115 in[8] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 117 in[1] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 119 in[9] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 121 in[2] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 123 in[10] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 125 in[3] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() 127 in[11] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d() [all …]
|
D | sum_squares_neon.c | 47 s[0] = vld1q_s16(src_t + 0 * stride); in vpx_sum_squares_2d_i16_neon() 48 s[1] = vld1q_s16(src_t + 1 * stride); in vpx_sum_squares_2d_i16_neon() 49 s[2] = vld1q_s16(src_t + 2 * stride); in vpx_sum_squares_2d_i16_neon() 50 s[3] = vld1q_s16(src_t + 3 * stride); in vpx_sum_squares_2d_i16_neon() 51 s[4] = vld1q_s16(src_t + 4 * stride); in vpx_sum_squares_2d_i16_neon() 52 s[5] = vld1q_s16(src_t + 5 * stride); in vpx_sum_squares_2d_i16_neon() 53 s[6] = vld1q_s16(src_t + 6 * stride); in vpx_sum_squares_2d_i16_neon() 54 s[7] = vld1q_s16(src_t + 7 * stride); in vpx_sum_squares_2d_i16_neon()
|
/external/XNNPACK/src/s16-window/gen/ |
D | neon-shift15-x32.c | 40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 43 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 45 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 46 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 47 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 48 const int16x8_t vw3 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 63 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32() 64 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32() [all …]
|
D | neon-shift15-x24.c | 40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 44 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 45 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 46 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 59 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 60 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24() 68 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift15_ukernel__neon_x24() 69 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift15_ukernel__neon_x24()
|
D | neon-shift12-x32.c | 40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 43 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 45 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 46 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 47 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 48 const int16x8_t vw3 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 81 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32() 82 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32() [all …]
|
D | neon-x32.c | 41 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32() 42 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32() 43 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32() 44 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32() 46 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32() 47 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32() 48 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32() 49 const int16x8_t vw3 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32() 82 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32() 83 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32() [all …]
|
D | neon-shift15-x16.c | 40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x16() 41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x16() 43 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x16() 44 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x16() 55 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x16() 56 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x16() 64 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift15_ukernel__neon_x16() 65 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift15_ukernel__neon_x16()
|
D | neon-shift12-x24.c | 40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 44 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 45 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 46 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 73 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 74 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24() 86 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift12_ukernel__neon_x24() 87 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift12_ukernel__neon_x24()
|
D | neon-x24.c | 41 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24() 42 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24() 43 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24() 45 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24() 46 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24() 47 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24() 74 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24() 75 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24() 87 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_ukernel__neon_x24() 88 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_ukernel__neon_x24()
|
D | neon-shift12-x16.c | 40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x16() 41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x16() 43 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x16() 44 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x16() 65 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x16() 66 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x16() 78 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift12_ukernel__neon_x16() 79 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift12_ukernel__neon_x16()
|
D | neon-x16.c | 41 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x16() 42 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x16() 44 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x16() 45 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x16() 66 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x16() 67 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x16() 79 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_ukernel__neon_x16() 80 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_ukernel__neon_x16()
|
/external/libaom/aom_dsp/arm/ |
D | hadamard_neon.c | 51 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_8x8_neon() 52 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_8x8_neon() 53 int16x8_t a2 = vld1q_s16(src_diff + 2 * src_stride); in aom_hadamard_8x8_neon() 54 int16x8_t a3 = vld1q_s16(src_diff + 3 * src_stride); in aom_hadamard_8x8_neon() 55 int16x8_t a4 = vld1q_s16(src_diff + 4 * src_stride); in aom_hadamard_8x8_neon() 56 int16x8_t a5 = vld1q_s16(src_diff + 5 * src_stride); in aom_hadamard_8x8_neon() 57 int16x8_t a6 = vld1q_s16(src_diff + 6 * src_stride); in aom_hadamard_8x8_neon() 58 int16x8_t a7 = vld1q_s16(src_diff + 7 * src_stride); in aom_hadamard_8x8_neon() 80 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_lp_8x8_neon() 81 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_lp_8x8_neon() [all …]
|
/external/libavc/decoder/arm/svc/ |
D | isvcd_pred_residual_recon_neon.c | 108 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_pred_residual_recon_4x4_neonintr() 109 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_pred_residual_recon_4x4_neonintr() 110 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_pred_residual_recon_4x4_neonintr() 111 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_pred_residual_recon_4x4_neonintr() 208 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_pred_residual_recon_8x8_neonintr() 209 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_pred_residual_recon_8x8_neonintr() 210 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_pred_residual_recon_8x8_neonintr() 211 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_pred_residual_recon_8x8_neonintr() 212 resd4_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 4)); in isvcd_pred_residual_recon_8x8_neonintr() 213 resd5_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 5)); in isvcd_pred_residual_recon_8x8_neonintr() [all …]
|
D | isvcd_iquant_itrans_residual_neon.c | 496 pred0 = vld1q_s16((int16_t *) pi2_pred); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 497 pred1 = vld1q_s16((int16_t *) pi2_pred + pred_strd); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 498 pred2 = vld1q_s16((int16_t *) pi2_pred + (pred_strd * 2)); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 499 pred3 = vld1q_s16((int16_t *) pi2_pred + (pred_strd * 3)); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 531 i4_out_horz_16x8_r0 = vld1q_s16(pi2_out); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 532 i4_out_horz_16x8_r1 = vld1q_s16(pi2_out + out_strd); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 533 i4_out_horz_16x8_r2 = vld1q_s16(pi2_out + out_strd * 2); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 534 i4_out_horz_16x8_r3 = vld1q_s16(pi2_out + out_strd * 3); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr() 598 pred0 = vld1q_s16((int16_t *) pi2_pred); in isvcd_iquant_itrans_residual_chroma_4x4_dc_neonintr() 599 pred1 = vld1q_s16((int16_t *) pi2_pred + pred_strd); in isvcd_iquant_itrans_residual_chroma_4x4_dc_neonintr() [all …]
|
D | isvcd_iquant_itrans_residual_recon_neon.c | 130 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr() 131 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr() 132 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr() 133 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr() 240 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr() 241 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr() 242 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr() 243 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr() 618 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_neonintr() 619 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_neonintr() [all …]
|
/external/libaom/av1/common/arm/ |
D | wiener_convolve_neon.c | 37 s7 = vld1q_s16(s); \ 251 s0 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 253 s1 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 255 s2 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 257 s3 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 259 s4 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 261 s5 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 263 s6 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 276 s7 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() 278 s8 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon() [all …]
|
/external/libvpx/vp8/encoder/arm/neon/ |
D | fastquantizeb_neon.c | 20 const int16x8_t one_q = vdupq_n_s16(-1), z0 = vld1q_s16(b->coeff), in vp8_fast_quantize_b_neon() 21 z1 = vld1q_s16(b->coeff + 8), round0 = vld1q_s16(b->round), in vp8_fast_quantize_b_neon() 22 round1 = vld1q_s16(b->round + 8), in vp8_fast_quantize_b_neon() 23 quant0 = vld1q_s16(b->quant_fast), in vp8_fast_quantize_b_neon() 24 quant1 = vld1q_s16(b->quant_fast + 8), in vp8_fast_quantize_b_neon() 25 dequant0 = vld1q_s16(d->dequant), in vp8_fast_quantize_b_neon() 26 dequant1 = vld1q_s16(d->dequant + 8); in vp8_fast_quantize_b_neon()
|
/external/libvpx/vp9/common/arm/neon/ |
D | vp9_iht16x16_add_neon.c | 78 in[0] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 80 in[8] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 82 in[1] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 84 in[9] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 86 in[2] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 88 in[10] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 90 in[3] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 92 in[11] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 94 in[4] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() 96 in[12] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d() [all …]
|
/external/libjpeg-turbo/simd/arm/ |
D | jidctred-neon.c | 82 int16x8_t row0 = vld1q_s16(coef_block + 0 * DCTSIZE); in jsimd_idct_2x2_neon() 83 int16x8_t row1 = vld1q_s16(coef_block + 1 * DCTSIZE); in jsimd_idct_2x2_neon() 84 int16x8_t row3 = vld1q_s16(coef_block + 3 * DCTSIZE); in jsimd_idct_2x2_neon() 85 int16x8_t row5 = vld1q_s16(coef_block + 5 * DCTSIZE); in jsimd_idct_2x2_neon() 86 int16x8_t row7 = vld1q_s16(coef_block + 7 * DCTSIZE); in jsimd_idct_2x2_neon() 89 int16x8_t quant_row0 = vld1q_s16(quantptr + 0 * DCTSIZE); in jsimd_idct_2x2_neon() 90 int16x8_t quant_row1 = vld1q_s16(quantptr + 1 * DCTSIZE); in jsimd_idct_2x2_neon() 91 int16x8_t quant_row3 = vld1q_s16(quantptr + 3 * DCTSIZE); in jsimd_idct_2x2_neon() 92 int16x8_t quant_row5 = vld1q_s16(quantptr + 5 * DCTSIZE); in jsimd_idct_2x2_neon() 93 int16x8_t quant_row7 = vld1q_s16(quantptr + 7 * DCTSIZE); in jsimd_idct_2x2_neon() [all …]
|
/external/XNNPACK/src/s16-vlshift/gen/ |
D | neon-x32.c | 34 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32() 35 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32() 36 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32() 37 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32() 52 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32() 61 const int16x8_t vi = vld1q_s16(input); in xnn_s16_vlshift_ukernel__neon_x32()
|
D | neon-x24.c | 34 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24() 35 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24() 36 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24() 49 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24() 58 const int16x8_t vi = vld1q_s16(input); in xnn_s16_vlshift_ukernel__neon_x24()
|