Home
last modified time | relevance | path

Searched refs:vmovl_u8 (Results 1 – 25 of 80) sorted by relevance

1234

/external/libvpx/libvpx/vpx_dsp/arm/
Dvpx_convolve8_neon.c87 tt0 = vreinterpretq_s16_u16(vmovl_u8(t0)); in vpx_convolve8_horiz_neon()
88 tt1 = vreinterpretq_s16_u16(vmovl_u8(t1)); in vpx_convolve8_horiz_neon()
89 tt2 = vreinterpretq_s16_u16(vmovl_u8(t2)); in vpx_convolve8_horiz_neon()
90 tt3 = vreinterpretq_s16_u16(vmovl_u8(t3)); in vpx_convolve8_horiz_neon()
107 tt0 = vreinterpretq_s16_u16(vmovl_u8(t0)); in vpx_convolve8_horiz_neon()
108 tt1 = vreinterpretq_s16_u16(vmovl_u8(t1)); in vpx_convolve8_horiz_neon()
109 tt2 = vreinterpretq_s16_u16(vmovl_u8(t2)); in vpx_convolve8_horiz_neon()
110 tt3 = vreinterpretq_s16_u16(vmovl_u8(t3)); in vpx_convolve8_horiz_neon()
161 s0 = vreinterpretq_s16_u16(vmovl_u8(t0)); in vpx_convolve8_horiz_neon()
162 s1 = vreinterpretq_s16_u16(vmovl_u8(t1)); in vpx_convolve8_horiz_neon()
[all …]
Dvpx_convolve8_neon.h125 ss[0] = vreinterpretq_s16_u16(vmovl_u8(s[0])); in scale_filter_8()
126 ss[1] = vreinterpretq_s16_u16(vmovl_u8(s[1])); in scale_filter_8()
127 ss[2] = vreinterpretq_s16_u16(vmovl_u8(s[2])); in scale_filter_8()
128 ss[3] = vreinterpretq_s16_u16(vmovl_u8(s[3])); in scale_filter_8()
129 ss[4] = vreinterpretq_s16_u16(vmovl_u8(s[4])); in scale_filter_8()
130 ss[5] = vreinterpretq_s16_u16(vmovl_u8(s[5])); in scale_filter_8()
131 ss[6] = vreinterpretq_s16_u16(vmovl_u8(s[6])); in scale_filter_8()
132 ss[7] = vreinterpretq_s16_u16(vmovl_u8(s[7])); in scale_filter_8()
Dvpx_scaled_convolve8_neon.c50 ss[0] = vreinterpretq_s16_u16(vmovl_u8(s[0])); in scaledconvolve_horiz_w4()
51 ss[1] = vreinterpretq_s16_u16(vmovl_u8(s[1])); in scaledconvolve_horiz_w4()
52 ss[2] = vreinterpretq_s16_u16(vmovl_u8(s[2])); in scaledconvolve_horiz_w4()
53 ss[3] = vreinterpretq_s16_u16(vmovl_u8(s[3])); in scaledconvolve_horiz_w4()
177 t[0] = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(s[0]))); in scaledconvolve_vert_w4()
178 t[1] = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(s[1]))); in scaledconvolve_vert_w4()
179 t[2] = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(s[2]))); in scaledconvolve_vert_w4()
180 t[3] = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(s[3]))); in scaledconvolve_vert_w4()
181 t[4] = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(s[4]))); in scaledconvolve_vert_w4()
182 t[5] = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(s[5]))); in scaledconvolve_vert_w4()
[all …]
/external/libjpeg-turbo/simd/arm/
Djdsample-neon.c88 vmlal_u8(vmovl_u8(vget_low_u8(s1)), vget_low_u8(s0), three_u8); in jsimd_h2v1_fancy_upsample_neon()
90 vmlal_u8(vmovl_u8(vget_high_u8(s1)), vget_high_u8(s0), three_u8); in jsimd_h2v1_fancy_upsample_neon()
92 vmlal_u8(vmovl_u8(vget_low_u8(s0)), vget_low_u8(s1), three_u8); in jsimd_h2v1_fancy_upsample_neon()
94 vmlal_u8(vmovl_u8(vget_high_u8(s0)), vget_high_u8(s1), three_u8); in jsimd_h2v1_fancy_upsample_neon()
127 vmlal_u8(vmovl_u8(vget_low_u8(s1)), vget_low_u8(s0), three_u8); in jsimd_h2v1_fancy_upsample_neon()
129 vmlal_u8(vmovl_u8(vget_high_u8(s1)), vget_high_u8(s0), three_u8); in jsimd_h2v1_fancy_upsample_neon()
131 vmlal_u8(vmovl_u8(vget_low_u8(s0)), vget_low_u8(s1), three_u8); in jsimd_h2v1_fancy_upsample_neon()
133 vmlal_u8(vmovl_u8(vget_high_u8(s0)), vget_high_u8(s1), three_u8); in jsimd_h2v1_fancy_upsample_neon()
247 uint16x8_t s0colsum0_l = vmlal_u8(vmovl_u8(vget_low_u8(s0A)), in jsimd_h2v2_fancy_upsample_neon()
249 uint16x8_t s0colsum0_h = vmlal_u8(vmovl_u8(vget_high_u8(s0A)), in jsimd_h2v2_fancy_upsample_neon()
[all …]
Djcgryext-neon.c69 uint16x8_t r_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_RED])); in jsimd_rgb_gray_convert_neon()
70 uint16x8_t r_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_RED])); in jsimd_rgb_gray_convert_neon()
71 uint16x8_t g_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_GREEN])); in jsimd_rgb_gray_convert_neon()
72 uint16x8_t g_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_GREEN])); in jsimd_rgb_gray_convert_neon()
73 uint16x8_t b_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_BLUE])); in jsimd_rgb_gray_convert_neon()
74 uint16x8_t b_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_BLUE])); in jsimd_rgb_gray_convert_neon()
/external/libaom/libaom/av1/common/arm/
Dconvolve_neon.c237 s0 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t0))); in av1_convolve_x_sr_neon()
238 s1 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t1))); in av1_convolve_x_sr_neon()
239 s2 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t2))); in av1_convolve_x_sr_neon()
240 s3 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t3))); in av1_convolve_x_sr_neon()
241 s4 = vget_high_s16(vreinterpretq_s16_u16(vmovl_u8(t0))); in av1_convolve_x_sr_neon()
242 s5 = vget_high_s16(vreinterpretq_s16_u16(vmovl_u8(t1))); in av1_convolve_x_sr_neon()
243 s6 = vget_high_s16(vreinterpretq_s16_u16(vmovl_u8(t2))); in av1_convolve_x_sr_neon()
254 s7 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t0))); in av1_convolve_x_sr_neon()
255 s8 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t1))); in av1_convolve_x_sr_neon()
256 s9 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t2))); in av1_convolve_x_sr_neon()
[all …]
Djnt_convolve_neon.c355 tt0 = vreinterpretq_s16_u16(vmovl_u8(t0)); in dist_wtd_convolve_2d_horiz_neon()
356 tt1 = vreinterpretq_s16_u16(vmovl_u8(t1)); in dist_wtd_convolve_2d_horiz_neon()
357 tt2 = vreinterpretq_s16_u16(vmovl_u8(t2)); in dist_wtd_convolve_2d_horiz_neon()
358 tt3 = vreinterpretq_s16_u16(vmovl_u8(t3)); in dist_wtd_convolve_2d_horiz_neon()
374 tt0 = vreinterpretq_s16_u16(vmovl_u8(t0)); in dist_wtd_convolve_2d_horiz_neon()
375 tt1 = vreinterpretq_s16_u16(vmovl_u8(t1)); in dist_wtd_convolve_2d_horiz_neon()
376 tt2 = vreinterpretq_s16_u16(vmovl_u8(t2)); in dist_wtd_convolve_2d_horiz_neon()
377 tt3 = vreinterpretq_s16_u16(vmovl_u8(t3)); in dist_wtd_convolve_2d_horiz_neon()
404 tt0 = vreinterpretq_s16_u16(vmovl_u8(t0)); // a0 a1 a2 a3 a4 a5 a6 a7 in dist_wtd_convolve_2d_horiz_neon()
412 s7 = vget_low_s16(vreinterpretq_s16_u16(vmovl_u8(t0))); in dist_wtd_convolve_2d_horiz_neon()
[all …]
Dwiener_convolve_neon.c121 res3 = vreinterpretq_s16_u16(vmovl_u8(t3)); in av1_wiener_convolve_add_src_neon()
128 res3 = vreinterpretq_s16_u16(vmovl_u8(t4)); in av1_wiener_convolve_add_src_neon()
135 res3 = vreinterpretq_s16_u16(vmovl_u8(t5)); in av1_wiener_convolve_add_src_neon()
142 res3 = vreinterpretq_s16_u16(vmovl_u8(t6)); in av1_wiener_convolve_add_src_neon()
149 res3 = vreinterpretq_s16_u16(vmovl_u8(t7)); in av1_wiener_convolve_add_src_neon()
156 res3 = vreinterpretq_s16_u16(vmovl_u8(t8)); in av1_wiener_convolve_add_src_neon()
163 res3 = vreinterpretq_s16_u16(vmovl_u8(t9)); in av1_wiener_convolve_add_src_neon()
170 res3 = vreinterpretq_s16_u16(vmovl_u8(t10)); in av1_wiener_convolve_add_src_neon()
225 res3 = vreinterpretq_s16_u16(vmovl_u8(t3)); in av1_wiener_convolve_add_src_neon()
262 tt0 = vreinterpretq_s16_u16(vmovl_u8(t0)); in av1_wiener_convolve_add_src_neon()
[all …]
/external/libhevc/common/arm/
Dihevc_inter_pred_chroma_copy_w16out.s141 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp)
149 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp)
152 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp)
157 vmovl.u8 q13,d26 @vmovl_u8(vld1_u8(pu1_src_tmp)
183 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp)
191 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp)
194 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp)
220 vmovl.u8 q8,d8 @vmovl_u8(vld1_u8(pu1_src_tmp))
221 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
222 vmovl.u8 q10,d12 @vmovl_u8(vld1_u8(pu1_src_tmp)
[all …]
Dihevc_inter_pred_luma_copy_w16out.s109 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp)
117 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp)
120 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp)
125 vmovl.u8 q13,d26 @vmovl_u8(vld1_u8(pu1_src_tmp)
158 vmovl.u8 q8,d8 @vmovl_u8(vld1_u8(pu1_src_tmp))
159 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
160 vmovl.u8 q10,d12 @vmovl_u8(vld1_u8(pu1_src_tmp)
161 vmovl.u8 q11,d14 @vmovl_u8(vld1_u8(pu1_src_tmp)
188 vmovl.u8 q8,d8 @vmovl_u8(vld1_u8(pu1_src_tmp))
191 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
[all …]
Dihevc_sao_edge_offset_class1_chroma.s170 … @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
173 … @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
192 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
204 …D11 @pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
223 … @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
264 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
270 …D11 @pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
325 … @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
349 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
400 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
Dihevc_sao_edge_offset_class1.s166 … @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
169 … @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
188 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
197 …11 @pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
247 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
253 …D11 @pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_…
308 … @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
325 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
365 …D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_c…
/external/XNNPACK/src/qu8-dwconv/
Dup8x9-minmax-neon.c83 const int16x8_t vxi0 = vreinterpretq_s16_u16(vmovl_u8(vi0)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
90 const int16x8_t vxi1 = vreinterpretq_s16_u16(vmovl_u8(vi1)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
97 const int16x8_t vxi2 = vreinterpretq_s16_u16(vmovl_u8(vi2)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
104 const int16x8_t vxi3 = vreinterpretq_s16_u16(vmovl_u8(vi3)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
111 const int16x8_t vxi4 = vreinterpretq_s16_u16(vmovl_u8(vi4)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
118 const int16x8_t vxi5 = vreinterpretq_s16_u16(vmovl_u8(vi5)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
125 const int16x8_t vxi6 = vreinterpretq_s16_u16(vmovl_u8(vi6)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
132 const int16x8_t vxi7 = vreinterpretq_s16_u16(vmovl_u8(vi7)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
139 const int16x8_t vxi8 = vreinterpretq_s16_u16(vmovl_u8(vi8)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
174 const int16x8_t vxi0 = vreinterpretq_s16_u16(vmovl_u8(vi0)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
[all …]
/external/libjpeg-turbo/simd/arm/aarch64/
Djccolext-neon.c76 uint16x8_t r_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_RED])); in jsimd_rgb_ycc_convert_neon()
77 uint16x8_t g_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_GREEN])); in jsimd_rgb_ycc_convert_neon()
78 uint16x8_t b_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_BLUE])); in jsimd_rgb_ycc_convert_neon()
79 uint16x8_t r_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_RED])); in jsimd_rgb_ycc_convert_neon()
80 uint16x8_t g_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_GREEN])); in jsimd_rgb_ycc_convert_neon()
81 uint16x8_t b_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_BLUE])); in jsimd_rgb_ycc_convert_neon()
175 uint16x8_t r_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_RED])); in jsimd_rgb_ycc_convert_neon()
176 uint16x8_t g_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_GREEN])); in jsimd_rgb_ycc_convert_neon()
177 uint16x8_t b_l = vmovl_u8(vget_low_u8(input_pixels.val[RGB_BLUE])); in jsimd_rgb_ycc_convert_neon()
178 uint16x8_t r_h = vmovl_u8(vget_high_u8(input_pixels.val[RGB_RED])); in jsimd_rgb_ycc_convert_neon()
[all …]
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Ddepthwiseconv_uint8.h48 filter[i] = vaddq_s16(vreinterpretq_s16_u16(vmovl_u8(filter_u8.val[i])),
62 const int16x8_t input_s16 = vreinterpretq_s16_u16(vmovl_u8(input_u8));
91 const int16x8_t filter_s16 = vreinterpretq_s16_u16(vmovl_u8(filter_u8));
110 input[i] = vreinterpretq_s16_u16(vmovl_u8(input_u8[i]));
138 const int16x8_t input_s16 = vreinterpretq_s16_u16(vmovl_u8(input_u8));
159 const int16x8_t filter_s16 = vreinterpretq_s16_u16(vmovl_u8(filter_u8));
173 const int16x8_t input_s16 = vreinterpretq_s16_u16(vmovl_u8(input_u8));
205 vreinterpret_s16_u16(vget_low_u16(vmovl_u8(input_u8)));
231 const int16x8_t filter_s16 = vreinterpretq_s16_u16(vmovl_u8(filter_u8));
250 vreinterpret_s16_u16(vget_low_u16(vmovl_u8(input_u8)));
[all …]
/external/skqp/src/core/
DSkBlitRow_D32.cpp166 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32_blend()
188 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32_blend()
214 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32a_blend()
217 vdst_wide = vmovl_u8(vdst); in blit_row_s32a_blend()
247 vdst_scale = vmovl_u8(vsrc_alphas); in blit_row_s32a_blend()
257 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32a_blend()
260 vdst_wide = vmovl_u8(vdst); in blit_row_s32a_blend()
/external/skia/src/core/
DSkBlitRow_D32.cpp165 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32_blend()
187 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32_blend()
213 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32a_blend()
216 vdst_wide = vmovl_u8(vdst); in blit_row_s32a_blend()
246 vdst_scale = vmovl_u8(vsrc_alphas); in blit_row_s32a_blend()
256 vsrc_wide = vmovl_u8(vsrc); in blit_row_s32a_blend()
259 vdst_wide = vmovl_u8(vdst); in blit_row_s32a_blend()
/external/libjpeg-turbo/simd/arm/aarch32/
Djccolext-neon.c93 uint16x8_t r = vmovl_u8(input_pixels.val[RGB_RED]); in jsimd_rgb_ycc_convert_neon()
94 uint16x8_t g = vmovl_u8(input_pixels.val[RGB_GREEN]); in jsimd_rgb_ycc_convert_neon()
95 uint16x8_t b = vmovl_u8(input_pixels.val[RGB_BLUE]); in jsimd_rgb_ycc_convert_neon()
/external/libhevc/encoder/arm/
Dihevce_hme_utils_neon.c137 src0_8x16b = vreinterpretq_s16_u16(vmovl_u8(src0_8x8b)); in ihevce_get_wt_inp_4x8_neon()
138 src1_8x16b = vreinterpretq_s16_u16(vmovl_u8(src1_8x8b)); in ihevce_get_wt_inp_4x8_neon()
139 src2_8x16b = vreinterpretq_s16_u16(vmovl_u8(src2_8x8b)); in ihevce_get_wt_inp_4x8_neon()
140 src3_8x16b = vreinterpretq_s16_u16(vmovl_u8(src3_8x8b)); in ihevce_get_wt_inp_4x8_neon()
567 src0_8x16b = vreinterpretq_s16_u16(vmovl_u8(src0_8x8b)); in hme_get_wt_inp_ctb_neon()
568 src1_8x16b = vreinterpretq_s16_u16(vmovl_u8(src1_8x8b)); in hme_get_wt_inp_ctb_neon()
569 src2_8x16b = vreinterpretq_s16_u16(vmovl_u8(src2_8x8b)); in hme_get_wt_inp_ctb_neon()
570 src3_8x16b = vreinterpretq_s16_u16(vmovl_u8(src3_8x8b)); in hme_get_wt_inp_ctb_neon()
700 src0_8x16b = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(src0_16x8b))); in hme_get_wt_inp_ctb_neon()
701 src1_8x16b = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(src0_16x8b))); in hme_get_wt_inp_ctb_neon()
/external/skqp/src/opts/
DSk4px_NEON.h11 return Sk16h(vmovl_u8(vget_low_u8 (this->fVec)), in widen()
12 vmovl_u8(vget_high_u8(this->fVec))); in widen()
/external/skia/src/opts/
DSk4px_NEON.h11 return Sk16h(vmovl_u8(vget_low_u8 (this->fVec)), in widen()
12 vmovl_u8(vget_high_u8(this->fVec))); in widen()
/external/tensorflow/tensorflow/core/kernels/
Dquantized_instance_norm.cc71 const uint16x8_t v_high_u16 = vmovl_u8(v_high); in ColMeanAndVariance()
72 const uint16x8_t v_low_u16 = vmovl_u8(v_low); in ColMeanAndVariance()
165 const uint16x8_t v_high = vmovl_u8(vget_high_u8(v)); in MinAndMax()
166 const uint16x8_t v_low = vmovl_u8(vget_low_u8(v)); in MinAndMax()
221 const uint16x8_t v_high = vmovl_u8(vget_high_u8(v)); in InstanceNorm()
222 const uint16x8_t v_low = vmovl_u8(vget_low_u8(v)); in InstanceNorm()
/external/XNNPACK/src/qu8-gemm/
D8x8-minmax-neon.c102 const int16x8_t vxa0 = vreinterpretq_s16_u16(vmovl_u8(va0)); a0 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
104 const int16x8_t vxa1 = vreinterpretq_s16_u16(vmovl_u8(va1)); a1 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
106 const int16x8_t vxa2 = vreinterpretq_s16_u16(vmovl_u8(va2)); a2 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
108 const int16x8_t vxa3 = vreinterpretq_s16_u16(vmovl_u8(va3)); a3 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
110 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); a4 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
112 const int16x8_t vxa5 = vreinterpretq_s16_u16(vmovl_u8(va5)); a5 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
114 const int16x8_t vxa6 = vreinterpretq_s16_u16(vmovl_u8(va6)); a6 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
116 const int16x8_t vxa7 = vreinterpretq_s16_u16(vmovl_u8(va7)); a7 += 8; in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
282 const int16x8_t vxa0 = vreinterpretq_s16_u16(vmovl_u8(va0)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
284 const int16x8_t vxa1 = vreinterpretq_s16_u16(vmovl_u8(va1)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
[all …]
/external/XNNPACK/src/qu8-igemm/
D8x8-minmax-neon.c134 const int16x8_t vxa0 = vreinterpretq_s16_u16(vmovl_u8(va0)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
135 const int16x8_t vxa1 = vreinterpretq_s16_u16(vmovl_u8(va1)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
136 const int16x8_t vxa2 = vreinterpretq_s16_u16(vmovl_u8(va2)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
137 const int16x8_t vxa3 = vreinterpretq_s16_u16(vmovl_u8(va3)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
138 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
139 const int16x8_t vxa5 = vreinterpretq_s16_u16(vmovl_u8(va5)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
140 const int16x8_t vxa6 = vreinterpretq_s16_u16(vmovl_u8(va6)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
141 const int16x8_t vxa7 = vreinterpretq_s16_u16(vmovl_u8(va7)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
330 const int16x8_t vxa0 = vreinterpretq_s16_u16(vmovl_u8(va0)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
331 const int16x8_t vxa1 = vreinterpretq_s16_u16(vmovl_u8(va1)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
[all …]
D4x8-minmax-neon.c90 const int16x8_t vxa0 = vreinterpretq_s16_u16(vmovl_u8(va0)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
91 const int16x8_t vxa1 = vreinterpretq_s16_u16(vmovl_u8(va1)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
92 const int16x8_t vxa2 = vreinterpretq_s16_u16(vmovl_u8(va2)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
93 const int16x8_t vxa3 = vreinterpretq_s16_u16(vmovl_u8(va3)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
214 const int16x8_t vxa0 = vreinterpretq_s16_u16(vmovl_u8(va0)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
215 const int16x8_t vxa1 = vreinterpretq_s16_u16(vmovl_u8(va1)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
216 const int16x8_t vxa2 = vreinterpretq_s16_u16(vmovl_u8(va2)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
217 const int16x8_t vxa3 = vreinterpretq_s16_u16(vmovl_u8(va3)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon()

1234