Home
last modified time | relevance | path

Searched refs:vsum1xAB (Results 1 – 25 of 66) sorted by relevance

123

/external/XNNPACK/src/qs8-igemm/gen/
D2x16c8-minmax-rndnu-neon-mull.c204 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mull() local
214 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mull()
264 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mull() local
265 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mull()
D2x16c16-minmax-rndnu-neon-mlal.c236 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c16__neon_mlal() local
245 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c16__neon_mlal()
295 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c16__neon_mlal() local
296 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c16__neon_mlal()
D2x16c8-minmax-rndnu-neon-mlal.c343 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local
353 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
403 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local
404 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
D3x16c8-minmax-rndnu-neon-mull.c261 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local
279 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull()
333 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local
334 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull()
D3x16c16-minmax-rndnu-neon-mlal.c309 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local
326 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal()
380 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local
381 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal()
D2x16c4s2-minmax-rndnu-neon-mull.c204 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull() local
205 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull()
D4x16c8-minmax-rndnu-neon-mull.c318 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local
344 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull()
402 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local
403 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull()
D3x16c8-minmax-rndnu-neon-mlal.c450 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local
468 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
522 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local
523 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
D2x16c4-minmax-rndnu-neon-mull-ld2r.c258 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r() local
259 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r()
D4x16c16-minmax-rndnu-neon-mlal.c382 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local
407 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal()
465 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local
466 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal()
D2x16c4-minmax-rndnu-neon-mull-dup.c258 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup() local
259 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup()
D2x16c4-minmax-rndnu-neon-mull-ld1r.c260 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r() local
261 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r()
/external/XNNPACK/src/qs8-gemm/gen/
D2x16c8-minmax-rndnu-neon-mull.c188 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mull() local
198 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mull()
248 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mull() local
249 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mull()
D2x16c16-minmax-rndnu-neon-mlal.c220 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c16__neon_mlal() local
229 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c16__neon_mlal()
279 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c16__neon_mlal() local
280 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c16__neon_mlal()
D3x16c8-minmax-rndnu-neon-mull.c243 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local
261 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull()
315 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local
316 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull()
D2x16c8-minmax-rndnu-neon-mlal.c327 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local
337 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
387 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local
388 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
D3x16c16-minmax-rndnu-neon-mlal.c291 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local
308 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal()
362 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local
363 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal()
D2x16c4s2-minmax-rndnu-neon-mull.c188 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull() local
189 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull()
D4x16c8-minmax-rndnu-neon-mull.c298 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local
324 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull()
382 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local
383 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull()
D3x16c8-minmax-rndnu-neon-mlal.c432 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local
450 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
504 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local
505 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
D2x16c4-minmax-rndnu-neon-mull-dup.c243 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup() local
244 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup()
D2x16c4-minmax-rndnu-neon-mull-ld1r.c245 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r() local
246 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r()
D2x16c4-minmax-rndnu-neon-mull-ld2r.c243 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r() local
244 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r()
D4x16c16-minmax-rndnu-neon-mlal.c362 const int32x4_t vsum1xAB = vpaddq_s32(vacc1x10, vacc1x11); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local
387 int32x4_t vacc1x89AB = vpaddq_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal()
445 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local
446 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB ); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal()
D3x16c4s2-minmax-rndnu-neon-mull.c240 const int32x2_t vsum1xAB = vpadd_s32(vget_low_s32(vacc1xAB), vget_high_s32(vacc1xAB)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4s2__neon_mull() local
241 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4s2__neon_mull()

123