/external/XNNPACK/src/qs8-igemm/gen/ |
D | 2x16c8-minmax-rndnu-neon-mull.c | 259 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mull() local 263 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mull()
|
D | 2x16c16-minmax-rndnu-neon-mlal.c | 290 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c16__neon_mlal() local 294 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c16__neon_mlal()
|
D | 2x16c8-minmax-rndnu-neon-mlal.c | 398 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local 402 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
|
D | 3x16c8-minmax-rndnu-neon-mull.c | 328 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local 332 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull()
|
D | 3x16c16-minmax-rndnu-neon-mlal.c | 375 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local 379 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal()
|
D | 4x16c8-minmax-rndnu-neon-mull.c | 397 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local 401 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mull()
|
D | 3x16c8-minmax-rndnu-neon-mlal.c | 517 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local 521 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
|
D | 4x16c16-minmax-rndnu-neon-mlal.c | 460 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local 464 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c16__neon_mlal()
|
D | 4x16c8-minmax-rndnu-neon-mlal.c | 636 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local 640 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 2x16c8-minmax-rndnu-neon-mull.c | 243 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mull() local 247 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mull()
|
D | 2x16c16-minmax-rndnu-neon-mlal.c | 274 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c16__neon_mlal() local 278 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c16__neon_mlal()
|
D | 3x16c8-minmax-rndnu-neon-mull.c | 310 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local 314 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull()
|
D | 2x16c8-minmax-rndnu-neon-mlal.c | 382 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local 386 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
|
D | 3x16c16-minmax-rndnu-neon-mlal.c | 357 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local 361 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal()
|
D | 4x16c8-minmax-rndnu-neon-mull.c | 377 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull() local 381 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mull()
|
D | 3x16c8-minmax-rndnu-neon-mlal.c | 499 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local 503 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
|
D | 4x16c16-minmax-rndnu-neon-mlal.c | 440 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal() local 444 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c16__neon_mlal()
|
D | 4x16c8-minmax-rndnu-neon-mlal.c | 616 const int32x2_t vpsum1x8 = vadd_s32(vget_low_s32(vacc1x8), vget_high_s32(vacc1x8)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local 620 const int32x2_t vsum1x89 = vpadd_s32(vpsum1x8, vpsum1x9); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal()
|