/external/XNNPACK/src/qs8-igemm/gen/ |
D | 2x16c8-minmax-neon-mull-padal.c | 260 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() local 263 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal()
|
D | 2x16c16-minmax-neon-mlal-padal.c | 292 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_2x16c16__neon_mlal_padal() local 295 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_2x16c16__neon_mlal_padal()
|
D | 3x16c8-minmax-neon-mull-padal.c | 329 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() local 332 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal()
|
D | 2x16c8-minmax-neon-mlal-padal.c | 399 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mlal_padal() local 402 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mlal_padal()
|
D | 3x16c16-minmax-neon-mlal-padal.c | 377 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() local 380 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal()
|
D | 4x16c8-minmax-neon-mull-padal.c | 398 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() local 401 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal()
|
D | 4x16c16-minmax-neon-mlal-padal.c | 462 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() local 465 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal()
|
D | 3x16c8-minmax-neon-mlal-padal.c | 518 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal() local 521 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
|
D | 4x16c8-minmax-neon-mlal-padal.c | 637 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal() local 640 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 2x16c8-minmax-neon-mull-padal.c | 244 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() local 247 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal()
|
D | 2x16c16-minmax-neon-mlal-padal.c | 276 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() local 279 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal()
|
D | 3x16c8-minmax-neon-mull-padal.c | 311 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() local 314 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal()
|
D | 2x16c8-minmax-neon-mlal-padal.c | 383 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mlal_padal() local 386 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mlal_padal()
|
D | 3x16c16-minmax-neon-mlal-padal.c | 359 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() local 362 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal()
|
D | 4x16c8-minmax-neon-mull-padal.c | 378 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() local 381 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal()
|
D | 3x16c8-minmax-neon-mlal-padal.c | 500 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal() local 503 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
|
D | 4x16c16-minmax-neon-mlal-padal.c | 442 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() local 445 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal()
|
D | 4x16c8-minmax-neon-mlal-padal.c | 617 const int32x2_t vpsum1xA = vadd_s32(vget_low_s32(vacc1x10), vget_high_s32(vacc1x10)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal() local 620 const int32x2_t vsum1xAB = vpadd_s32(vpsum1xA, vpsum1xB); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
|