Home
last modified time | relevance | path

Searched refs:vacc2x12 (Results 1 – 12 of 12) sorted by relevance

/external/XNNPACK/src/qs8-igemm/gen/
D3x16c8-minmax-neon-mull-padal.c99 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() local
218 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal()
270 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal()
362 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal()
D3x16c8-minmax-neon-mlal-padal.c99 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal() local
276 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
407 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
459 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
551 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
D3x16c16-minmax-neon-mlal-padal.c99 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() local
260 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal()
318 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal()
410 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal()
D4x16c8-minmax-neon-mull-padal.c103 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() local
268 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal()
327 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal()
431 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal()
D4x16c8-minmax-neon-mlal-padal.c103 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal() local
340 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
507 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
566 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
670 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
D4x16c16-minmax-neon-mlal-padal.c103 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() local
323 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal()
391 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal()
495 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal()
/external/XNNPACK/src/qs8-gemm/gen/
D3x16c8-minmax-neon-mull-padal.c100 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() local
203 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal()
252 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal()
344 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal()
D3x16c8-minmax-neon-mlal-padal.c100 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal() local
261 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
392 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
441 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
533 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
D3x16c16-minmax-neon-mlal-padal.c100 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() local
245 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal()
300 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal()
392 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal()
D4x16c8-minmax-neon-mull-padal.c106 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() local
251 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal()
307 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal()
411 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal()
D4x16c8-minmax-neon-mlal-padal.c106 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal() local
323 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
490 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
546 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
650 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
D4x16c16-minmax-neon-mlal-padal.c106 int32x4_t vacc2x12 = vacc0x12; in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() local
306 vacc2x12 = vpadalq_s16(vacc2x12, vprod2x12); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal()
371 const int32x4_t vsum2xCD = vpaddq_s32(vacc2x12, vacc2x13); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal()
475 const int32x2_t vpsum2xC = vadd_s32(vget_low_s32(vacc2x12), vget_high_s32(vacc2x12)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal()