Home
last modified time | relevance | path

Searched refs:vm1 (Results 1 – 25 of 84) sorted by relevance

1234

/external/XNNPACK/src/f32-argmaxpool/
D4x-wasmsimd-c4.c63 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__wasmsimd_c4() local
64 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_4x__wasmsimd_c4()
65 vidx = wasm_v128_bitselect(wasm_i32x4_const_splat(1), vidx, vm1); in xnn_f32_argmaxpool_ukernel_4x__wasmsimd_c4()
89 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__wasmsimd_c4() local
90 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_4x__wasmsimd_c4()
91 vidx = wasm_v128_bitselect(wasm_i32x4_const_splat(1), vidx, vm1); in xnn_f32_argmaxpool_ukernel_4x__wasmsimd_c4()
D4x-neon-c4.c58 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() local
59 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4()
60 vidx = vbslq_u32(vm1, vmovq_n_u32(1), vidx); in xnn_f32_argmaxpool_ukernel_4x__neon_c4()
82 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() local
83 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4()
84 vidx = vbslq_u32(vm1, vmovq_n_u32(1), vidx); in xnn_f32_argmaxpool_ukernel_4x__neon_c4()
D4x-sse2-c4.c62 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() local
64 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
88 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() local
90 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
D9p8x-wasmsimd-c4.c78 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local
79 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4()
80 vidx = wasm_v128_bitselect(wasm_i32x4_const_splat(1), vidx, vm1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4()
167 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local
169 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4()
170 vidx = wasm_v128_bitselect(vidx1, vidx, vm1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4()
282 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local
284 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4()
285 vidx = wasm_v128_bitselect(vidx1, vidx, vm1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4()
339 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local
[all …]
D9p8x-neon-c4.c69 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local
70 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4()
71 vidx = vbslq_u32(vm1, vmovq_n_u32(1), vidx); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4()
148 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local
150 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4()
151 vidx = vbslq_u32(vm1, vidx1, vidx); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4()
251 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local
253 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4()
254 vidx = vbslq_u32(vm1, vidx1, vidx); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4()
306 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local
[all …]
D9p8x-sse2-c4.c78 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local
80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
167 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local
170 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
282 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local
285 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
339 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local
342 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
D9x-neon-c4.c88 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() local
89 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4()
90 vidx = vbslq_u32(vm1, vmovq_n_u32(1), vidx); in xnn_f32_argmaxpool_ukernel_9x__neon_c4()
137 const uint32x4_t vm1 = vcgtq_f32(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() local
138 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4()
139 vidx = vbslq_u32(vm1, vmovq_n_u32(1), vidx); in xnn_f32_argmaxpool_ukernel_9x__neon_c4()
D9x-wasmsimd-c4.c98 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4() local
99 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4()
100 vidx = wasm_v128_bitselect(wasm_i32x4_const_splat(1), vidx, vm1); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4()
149 const v128_t vm1 = wasm_f32x4_gt(vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4() local
150 vmax = wasm_v128_bitselect(vi1, vmax, vm1); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4()
151 vidx = wasm_v128_bitselect(wasm_i32x4_const_splat(1), vidx, vm1); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4()
D9x-sse2-c4.c97 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() local
99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
148 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() local
150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
/external/XNNPACK/src/qs8-vcvt/gen/
Dvcvt-ssse3-x32.c41 const __m128i vm1 = _mm_cmpgt_epi8(_mm_setzero_si128(), vx1); in xnn_qs8_vcvt_ukernel__ssse3_x32() local
42 __m128i vacc2 = _mm_unpacklo_epi8(vx1, vm1); in xnn_qs8_vcvt_ukernel__ssse3_x32()
43 __m128i vacc3 = _mm_unpackhi_epi8(vx1, vm1); in xnn_qs8_vcvt_ukernel__ssse3_x32()
Dvcvt-sse2-x32.c40 const __m128i vm1 = _mm_cmpgt_epi8(_mm_setzero_si128(), vx1); in xnn_qs8_vcvt_ukernel__sse2_x32() local
41 const __m128i vextx2 = _mm_unpacklo_epi8(vx1, vm1); in xnn_qs8_vcvt_ukernel__sse2_x32()
42 const __m128i vextx3 = _mm_unpackhi_epi8(vx1, vm1); in xnn_qs8_vcvt_ukernel__sse2_x32()
/external/clang/test/CXX/special/class.copy/
Dp11.0x.copy.cpp77 } vm1, vm2(vm1); // expected-error {{deleted}} variable
/external/XNNPACK/src/bf16-gemm/gen/
D5x4c8-minmax-neonbf16-bfdot.c136 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot() local
150 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot()
152 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot()
154 … const bfloat16x8_t va2x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va2), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot()
156 … const bfloat16x8_t va3x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va3), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot()
158 … const bfloat16x8_t va4x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va4), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot()
D4x4c8-minmax-neonbf16-bfdot.c120 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot() local
132 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot()
134 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot()
136 … const bfloat16x8_t va2x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va2), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot()
138 … const bfloat16x8_t va3x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va3), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot()
D3x4c8-minmax-neonbf16-bfdot.c104 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfdot() local
114 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfdot()
116 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfdot()
118 … const bfloat16x8_t va2x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va2), vm1)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfdot()
D5x4c8-minmax-neonbf16-bfmlal.c157 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal() local
176 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal()
179 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal()
182 … const bfloat16x8_t va2x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va2), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal()
185 … const bfloat16x8_t va3x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va3), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal()
188 … const bfloat16x8_t va4x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va4), vm1)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal()
D4x4c8-minmax-neonbf16-bfmlal.c137 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal() local
153 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal()
156 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal()
159 … const bfloat16x8_t va2x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va2), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal()
162 … const bfloat16x8_t va3x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va3), vm1)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal()
D2x4c8-minmax-neonbf16-bfdot.c88 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfdot() local
96 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfdot()
98 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfdot()
D3x4c8-minmax-neonbf16-bfmlal.c117 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfmlal() local
130 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfmlal()
133 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfmlal()
136 … const bfloat16x8_t va2x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va2), vm1)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfmlal()
D2x4c8-minmax-neonbf16-bfmlal.c97 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfmlal() local
107 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfmlal()
110 … const bfloat16x8_t va1x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va1), vm1)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfmlal()
D1x4c8-minmax-neonbf16-bfdot.c72 const uint16x8_t vm1 = vceqq_u16(vreinterpretq_u16_bf16(vb1), vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_1x4c8__neonbf16_bfdot() local
78 … const bfloat16x8_t va0x1 = vreinterpretq_bf16_u16(vbicq_u16(vreinterpretq_u16_bf16(va0), vm1)); in xnn_bf16_gemm_minmax_ukernel_1x4c8__neonbf16_bfdot()
D4x4c8-minmax-neonfma-zip.c158 const uint16x8_t vm1 = vceqq_u16(vb1, vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip() local
171 const uint16x8_t va0x1 = vbicq_u16(va0, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip()
172 const uint16x8_t va1x1 = vbicq_u16(va1, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip()
173 const uint16x8_t va2x1 = vbicq_u16(va2, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip()
174 const uint16x8_t va3x1 = vbicq_u16(va3, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip()
D4x4c8-minmax-neonfma-shland.c158 const uint16x8_t vm1 = vceqq_u16(vb1, vmovq_n_u16(0)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland() local
171 const uint16x8_t va0x1 = vbicq_u16(va0, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland()
172 const uint16x8_t va1x1 = vbicq_u16(va1, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland()
173 const uint16x8_t va2x1 = vbicq_u16(va2, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland()
174 const uint16x8_t va3x1 = vbicq_u16(va3, vm1); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland()
/external/XNNPACK/src/qs8-vlrelu/gen/
Dvlrelu-ssse3-x32.c42 const __m128i vm1 = _mm_cmpgt_epi8(_mm_setzero_si128(), vx1); in xnn_qs8_vlrelu_ukernel__ssse3_x32() local
43 __m128i vacc2 = _mm_unpacklo_epi8(vx1, vm1); in xnn_qs8_vlrelu_ukernel__ssse3_x32()
44 __m128i vacc3 = _mm_unpackhi_epi8(vx1, vm1); in xnn_qs8_vlrelu_ukernel__ssse3_x32()
Dvlrelu-sse2-x32.c43 const __m128i vm1 = _mm_cmpgt_epi8(_mm_setzero_si128(), vx1); in xnn_qs8_vlrelu_ukernel__sse2_x32() local
44 __m128i vextx2 = _mm_unpacklo_epi8(vx1, vm1); in xnn_qs8_vlrelu_ukernel__sse2_x32()
45 __m128i vextx3 = _mm_unpackhi_epi8(vx1, vm1); in xnn_qs8_vlrelu_ukernel__sse2_x32()

1234