/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x8c8-minmax-rndnu-neon-mull.c | 178 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c8__neon_mull() local 227 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c8__neon_mull() local
|
D | 3x8c16-minmax-rndnu-neon-mlal.c | 202 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c16__neon_mlal() local 250 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c16__neon_mlal() local
|
D | 4x8c16-minmax-rndnu-neon-mlal.c | 243 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c16__neon_mlal() local 297 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c16__neon_mlal() local
|
D | 4x8c8-minmax-rndnu-neon-mull.c | 211 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c8__neon_mull() local 266 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c8__neon_mull() local
|
D | 3x8c8-minmax-rndnu-neon-mlal.c | 279 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c8__neon_mlal() local 328 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c8__neon_mlal() local
|
D | 3x8c4s2-minmax-rndnu-neon-mull.c | 180 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4s2__neon_mull() local
|
D | 4x8c8-minmax-rndnu-neon-mlal.c | 338 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c8__neon_mlal() local 393 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c8__neon_mlal() local
|
D | 3x16c16-minmax-rndnu-neon-mlal.c | 314 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local 400 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local
|
D | 3x16c8-minmax-rndnu-neon-mull.c | 266 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local 353 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local
|
D | 4x8c4s2-minmax-rndnu-neon-mull.c | 212 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c4s2__neon_mull() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld1r.c | 228 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld1r() local
|
D | 3x8c4-minmax-rndnu-neon-mull-dup.c | 225 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_dup() local
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 3x8c8-minmax-rndnu-neon-mull.c | 160 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c8__neon_mull() local 209 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c8__neon_mull() local
|
D | 3x8c16-minmax-rndnu-neon-mlal.c | 184 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c16__neon_mlal() local 232 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c16__neon_mlal() local
|
D | 4x8c16-minmax-rndnu-neon-mlal.c | 223 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c16__neon_mlal() local 277 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c16__neon_mlal() local
|
D | 3x8c8-minmax-rndnu-neon-mlal.c | 261 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c8__neon_mlal() local 310 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c8__neon_mlal() local
|
D | 4x8c8-minmax-rndnu-neon-mull.c | 191 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c8__neon_mull() local 246 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c8__neon_mull() local
|
D | 3x8c4s2-minmax-rndnu-neon-mull.c | 162 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4s2__neon_mull() local
|
D | 4x8c8-minmax-rndnu-neon-mlal.c | 318 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c8__neon_mlal() local 373 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c8__neon_mlal() local
|
D | 3x16c16-minmax-rndnu-neon-mlal.c | 296 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local 382 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c16__neon_mlal() local
|
D | 3x16c8-minmax-rndnu-neon-mull.c | 248 const int32x4_t vsum2x45 = vpaddq_s32(vacc2x4, vacc2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local 335 const int32x2_t vsum2x45 = vpadd_s32(vpsum2x4, vpsum2x5); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mull() local
|
D | 4x8c4s2-minmax-rndnu-neon-mull.c | 192 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c4s2__neon_mull() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld2r.c | 208 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld2r() local
|
D | 3x8c4-minmax-rndnu-neon-mull-dup.c | 208 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_dup() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld1r.c | 211 const int32x2_t vsum2x45 = vpadd_s32(vget_low_s32(vacc2x45), vget_high_s32(vacc2x45)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld1r() local
|