Home
last modified time | relevance | path

Searched refs:vld1q_s16 (Results 1 – 25 of 141) sorted by relevance

123456

/external/libvpx/vpx_dsp/arm/
Dfdct_partial_neon.c42 int16x8_t sum = vld1q_s16(&input[0]); in vpx_fdct8x8_1_neon()
45 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in vpx_fdct8x8_1_neon()
56 int16x8_t left = vld1q_s16(input); in vpx_fdct16x16_1_neon()
57 int16x8_t right = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon()
62 const int16x8_t a = vld1q_s16(input); in vpx_fdct16x16_1_neon()
63 const int16x8_t b = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon()
78 int16x8_t a0 = vld1q_s16(input); in vpx_fdct32x32_1_neon()
79 int16x8_t a1 = vld1q_s16(input + 8); in vpx_fdct32x32_1_neon()
80 int16x8_t a2 = vld1q_s16(input + 16); in vpx_fdct32x32_1_neon()
81 int16x8_t a3 = vld1q_s16(input + 24); in vpx_fdct32x32_1_neon()
[all …]
Dfdct8x8_neon.c26 in[0] = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in vpx_fdct8x8_neon()
27 in[1] = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in vpx_fdct8x8_neon()
28 in[2] = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in vpx_fdct8x8_neon()
29 in[3] = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in vpx_fdct8x8_neon()
30 in[4] = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in vpx_fdct8x8_neon()
31 in[5] = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in vpx_fdct8x8_neon()
32 in[6] = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in vpx_fdct8x8_neon()
33 in[7] = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in vpx_fdct8x8_neon()
77 in[0] = vld1q_s16(input + 0 * stride); in vpx_highbd_fdct8x8_neon()
78 in[1] = vld1q_s16(input + 1 * stride); in vpx_highbd_fdct8x8_neon()
[all …]
Dfdct16x16_neon.h19 b[0] = vld1q_s16(a); in load()
21 b[1] = vld1q_s16(a); in load()
23 b[2] = vld1q_s16(a); in load()
25 b[3] = vld1q_s16(a); in load()
27 b[4] = vld1q_s16(a); in load()
29 b[5] = vld1q_s16(a); in load()
31 b[6] = vld1q_s16(a); in load()
33 b[7] = vld1q_s16(a); in load()
35 b[8] = vld1q_s16(a); in load()
37 b[9] = vld1q_s16(a); in load()
[all …]
Dquantize_neon.c82 int16x8_t zbin = vld1q_s16(zbin_ptr); in vpx_quantize_b_neon()
83 int16x8_t round = vld1q_s16(round_ptr); in vpx_quantize_b_neon()
84 int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_neon()
85 int16x8_t quant_shift = vld1q_s16(quant_shift_ptr); in vpx_quantize_b_neon()
86 int16x8_t dequant = vld1q_s16(dequant_ptr); in vpx_quantize_b_neon()
90 const uint16x8_t v_iscan = vreinterpretq_u16_s16(vld1q_s16(iscan)); in vpx_quantize_b_neon()
116 const uint16x8_t v_iscan = vreinterpretq_u16_s16(vld1q_s16(iscan)); in vpx_quantize_b_neon()
229 int16x8_t zbin = vrshrq_n_s16(vld1q_s16(zbin_ptr), 1); in vpx_quantize_b_32x32_neon()
230 int16x8_t round = vrshrq_n_s16(vld1q_s16(round_ptr), 1); in vpx_quantize_b_32x32_neon()
231 int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_32x32_neon()
[all …]
Didct16x16_add_neon.c69 const int16x8_t cospis0 = vld1q_s16(kCospi); in vpx_idct16x16_256_add_half1d()
70 const int16x8_t cospis1 = vld1q_s16(kCospi + 8); in vpx_idct16x16_256_add_half1d()
113 in[0] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
115 in[8] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
117 in[1] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
119 in[9] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
121 in[2] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
123 in[10] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
125 in[3] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
127 in[11] = vld1q_s16(inputT); in vpx_idct16x16_256_add_half1d()
[all …]
Dsum_squares_neon.c47 s[0] = vld1q_s16(src_t + 0 * stride); in vpx_sum_squares_2d_i16_neon()
48 s[1] = vld1q_s16(src_t + 1 * stride); in vpx_sum_squares_2d_i16_neon()
49 s[2] = vld1q_s16(src_t + 2 * stride); in vpx_sum_squares_2d_i16_neon()
50 s[3] = vld1q_s16(src_t + 3 * stride); in vpx_sum_squares_2d_i16_neon()
51 s[4] = vld1q_s16(src_t + 4 * stride); in vpx_sum_squares_2d_i16_neon()
52 s[5] = vld1q_s16(src_t + 5 * stride); in vpx_sum_squares_2d_i16_neon()
53 s[6] = vld1q_s16(src_t + 6 * stride); in vpx_sum_squares_2d_i16_neon()
54 s[7] = vld1q_s16(src_t + 7 * stride); in vpx_sum_squares_2d_i16_neon()
/external/XNNPACK/src/s16-window/gen/
Dneon-shift15-x32.c40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
43 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
45 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
46 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
47 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
48 const int16x8_t vw3 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
63 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
64 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x32()
[all …]
Dneon-shift15-x24.c40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
44 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
45 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
46 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
59 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
60 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x24()
68 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift15_ukernel__neon_x24()
69 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift15_ukernel__neon_x24()
Dneon-shift12-x32.c40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
43 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
45 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
46 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
47 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
48 const int16x8_t vw3 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
81 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
82 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x32()
[all …]
Dneon-x32.c41 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32()
42 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32()
43 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32()
44 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32()
46 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32()
47 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32()
48 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32()
49 const int16x8_t vw3 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32()
82 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x32()
83 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x32()
[all …]
Dneon-shift15-x16.c40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x16()
41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x16()
43 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x16()
44 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x16()
55 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift15_ukernel__neon_x16()
56 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift15_ukernel__neon_x16()
64 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift15_ukernel__neon_x16()
65 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift15_ukernel__neon_x16()
Dneon-shift12-x24.c40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
42 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
44 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
45 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
46 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
73 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
74 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x24()
86 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift12_ukernel__neon_x24()
87 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift12_ukernel__neon_x24()
Dneon-x24.c41 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24()
42 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24()
43 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24()
45 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24()
46 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24()
47 const int16x8_t vw2 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24()
74 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x24()
75 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x24()
87 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_ukernel__neon_x24()
88 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_ukernel__neon_x24()
Dneon-shift12-x16.c40 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x16()
41 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x16()
43 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x16()
44 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x16()
65 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_shift12_ukernel__neon_x16()
66 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_shift12_ukernel__neon_x16()
78 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_shift12_ukernel__neon_x16()
79 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_shift12_ukernel__neon_x16()
Dneon-x16.c41 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x16()
42 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x16()
44 const int16x8_t vw0 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x16()
45 const int16x8_t vw1 = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x16()
66 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_window_ukernel__neon_x16()
67 const int16x8_t vw = vld1q_s16(w); w += 8; in xnn_s16_window_ukernel__neon_x16()
79 const int16x8_t vi = vld1q_s16(input); input = (const int16_t*) ((uintptr_t) input + n); in xnn_s16_window_ukernel__neon_x16()
80 const int16x8_t vw = vld1q_s16(w); in xnn_s16_window_ukernel__neon_x16()
/external/libaom/aom_dsp/arm/
Dhadamard_neon.c51 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_8x8_neon()
52 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_8x8_neon()
53 int16x8_t a2 = vld1q_s16(src_diff + 2 * src_stride); in aom_hadamard_8x8_neon()
54 int16x8_t a3 = vld1q_s16(src_diff + 3 * src_stride); in aom_hadamard_8x8_neon()
55 int16x8_t a4 = vld1q_s16(src_diff + 4 * src_stride); in aom_hadamard_8x8_neon()
56 int16x8_t a5 = vld1q_s16(src_diff + 5 * src_stride); in aom_hadamard_8x8_neon()
57 int16x8_t a6 = vld1q_s16(src_diff + 6 * src_stride); in aom_hadamard_8x8_neon()
58 int16x8_t a7 = vld1q_s16(src_diff + 7 * src_stride); in aom_hadamard_8x8_neon()
80 int16x8_t a0 = vld1q_s16(src_diff); in aom_hadamard_lp_8x8_neon()
81 int16x8_t a1 = vld1q_s16(src_diff + src_stride); in aom_hadamard_lp_8x8_neon()
[all …]
/external/libavc/decoder/arm/svc/
Disvcd_pred_residual_recon_neon.c108 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_pred_residual_recon_4x4_neonintr()
109 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_pred_residual_recon_4x4_neonintr()
110 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_pred_residual_recon_4x4_neonintr()
111 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_pred_residual_recon_4x4_neonintr()
208 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_pred_residual_recon_8x8_neonintr()
209 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_pred_residual_recon_8x8_neonintr()
210 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_pred_residual_recon_8x8_neonintr()
211 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_pred_residual_recon_8x8_neonintr()
212 resd4_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 4)); in isvcd_pred_residual_recon_8x8_neonintr()
213 resd5_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 5)); in isvcd_pred_residual_recon_8x8_neonintr()
[all …]
Disvcd_iquant_itrans_residual_neon.c496 pred0 = vld1q_s16((int16_t *) pi2_pred); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
497 pred1 = vld1q_s16((int16_t *) pi2_pred + pred_strd); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
498 pred2 = vld1q_s16((int16_t *) pi2_pred + (pred_strd * 2)); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
499 pred3 = vld1q_s16((int16_t *) pi2_pred + (pred_strd * 3)); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
531 i4_out_horz_16x8_r0 = vld1q_s16(pi2_out); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
532 i4_out_horz_16x8_r1 = vld1q_s16(pi2_out + out_strd); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
533 i4_out_horz_16x8_r2 = vld1q_s16(pi2_out + out_strd * 2); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
534 i4_out_horz_16x8_r3 = vld1q_s16(pi2_out + out_strd * 3); in isvcd_iquant_itrans_residual_chroma_4x4_neonintr()
598 pred0 = vld1q_s16((int16_t *) pi2_pred); in isvcd_iquant_itrans_residual_chroma_4x4_dc_neonintr()
599 pred1 = vld1q_s16((int16_t *) pi2_pred + pred_strd); in isvcd_iquant_itrans_residual_chroma_4x4_dc_neonintr()
[all …]
Disvcd_iquant_itrans_residual_recon_neon.c130 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr()
131 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr()
132 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr()
133 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_iquant_itrans_residual_recon_4x4_dc_neonintr()
240 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr()
241 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr()
242 resd2_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 2)); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr()
243 resd3_in = vld1q_s16((int16_t *) pi2_rsd + (rsd_strd * 3)); in isvcd_iquant_itrans_residual_recon_chroma_4x4_dc_neonintr()
618 resd0_in = vld1q_s16((int16_t *) pi2_rsd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_neonintr()
619 resd1_in = vld1q_s16((int16_t *) pi2_rsd + rsd_strd); in isvcd_iquant_itrans_residual_recon_chroma_4x4_neonintr()
[all …]
/external/libaom/av1/common/arm/
Dwiener_convolve_neon.c37 s7 = vld1q_s16(s); \
251 s0 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
253 s1 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
255 s2 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
257 s3 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
259 s4 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
261 s5 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
263 s6 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
276 s7 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
278 s8 = vld1q_s16(s); in av1_wiener_convolve_add_src_neon()
[all …]
/external/libvpx/vp8/encoder/arm/neon/
Dfastquantizeb_neon.c20 const int16x8_t one_q = vdupq_n_s16(-1), z0 = vld1q_s16(b->coeff), in vp8_fast_quantize_b_neon()
21 z1 = vld1q_s16(b->coeff + 8), round0 = vld1q_s16(b->round), in vp8_fast_quantize_b_neon()
22 round1 = vld1q_s16(b->round + 8), in vp8_fast_quantize_b_neon()
23 quant0 = vld1q_s16(b->quant_fast), in vp8_fast_quantize_b_neon()
24 quant1 = vld1q_s16(b->quant_fast + 8), in vp8_fast_quantize_b_neon()
25 dequant0 = vld1q_s16(d->dequant), in vp8_fast_quantize_b_neon()
26 dequant1 = vld1q_s16(d->dequant + 8); in vp8_fast_quantize_b_neon()
/external/libvpx/vp9/common/arm/neon/
Dvp9_iht16x16_add_neon.c78 in[0] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
80 in[8] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
82 in[1] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
84 in[9] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
86 in[2] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
88 in[10] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
90 in[3] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
92 in[11] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
94 in[4] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
96 in[12] = vld1q_s16(inputT); in vpx_iadst16x16_256_add_half1d()
[all …]
/external/libjpeg-turbo/simd/arm/
Djidctred-neon.c82 int16x8_t row0 = vld1q_s16(coef_block + 0 * DCTSIZE); in jsimd_idct_2x2_neon()
83 int16x8_t row1 = vld1q_s16(coef_block + 1 * DCTSIZE); in jsimd_idct_2x2_neon()
84 int16x8_t row3 = vld1q_s16(coef_block + 3 * DCTSIZE); in jsimd_idct_2x2_neon()
85 int16x8_t row5 = vld1q_s16(coef_block + 5 * DCTSIZE); in jsimd_idct_2x2_neon()
86 int16x8_t row7 = vld1q_s16(coef_block + 7 * DCTSIZE); in jsimd_idct_2x2_neon()
89 int16x8_t quant_row0 = vld1q_s16(quantptr + 0 * DCTSIZE); in jsimd_idct_2x2_neon()
90 int16x8_t quant_row1 = vld1q_s16(quantptr + 1 * DCTSIZE); in jsimd_idct_2x2_neon()
91 int16x8_t quant_row3 = vld1q_s16(quantptr + 3 * DCTSIZE); in jsimd_idct_2x2_neon()
92 int16x8_t quant_row5 = vld1q_s16(quantptr + 5 * DCTSIZE); in jsimd_idct_2x2_neon()
93 int16x8_t quant_row7 = vld1q_s16(quantptr + 7 * DCTSIZE); in jsimd_idct_2x2_neon()
[all …]
/external/XNNPACK/src/s16-vlshift/gen/
Dneon-x32.c34 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32()
35 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32()
36 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32()
37 const int16x8_t vi3 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32()
52 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x32()
61 const int16x8_t vi = vld1q_s16(input); in xnn_s16_vlshift_ukernel__neon_x32()
Dneon-x24.c34 const int16x8_t vi0 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24()
35 const int16x8_t vi1 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24()
36 const int16x8_t vi2 = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24()
49 const int16x8_t vi = vld1q_s16(input); input += 8; in xnn_s16_vlshift_ukernel__neon_x24()
58 const int16x8_t vi = vld1q_s16(input); in xnn_s16_vlshift_ukernel__neon_x24()

123456