/external/XNNPACK/src/qs8-gemm/gen/ |
D | 3x16c4-minmax-rndnu-neon-mlal-ld2r.c | 167 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld2r() local 344 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld2r() local 475 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld2r() local
|
D | 3x16c4-minmax-rndnu-neon-mlal-dup.c | 167 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_dup() local 344 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_dup() local 475 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_dup() local
|
D | 3x16c4-minmax-rndnu-neon-mlal-ld1r.c | 173 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld1r() local 353 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld1r() local 484 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld1r() local
|
D | 3x16c4-minmax-rndnu-neon-mull-dup.c | 142 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mull_dup() local 273 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mull_dup() local
|
D | 3x16c4-minmax-rndnu-neon-mull-ld2r.c | 142 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld2r() local 273 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld2r() local
|
D | 3x16c4-minmax-rndnu-neon-mull-ld1r.c | 145 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld1r() local 276 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld1r() local
|
D | 4x16c4-minmax-rndnu-neon-mlal-ld1r.c | 208 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local 436 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local 591 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local
|
D | 4x16c4-minmax-rndnu-neon-mlal-dup.c | 200 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local 424 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local 579 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local
|
D | 4x16c4-minmax-rndnu-neon-mlal-ld2r.c | 200 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local 424 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local 579 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local
|
D | 3x16c4s2-minmax-rndnu-neon-mlal.c | 159 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4s2__neon_mlal() local 330 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4s2__neon_mlal() local
|
D | 4x16c4-minmax-rndnu-neon-mull-ld2r.c | 168 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld2r() local 323 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld2r() local
|
D | 4x16c4-minmax-rndnu-neon-mull-dup.c | 168 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_dup() local 323 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_dup() local
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x16c4-minmax-rndnu-neon-mlal-ld1r.c | 188 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld1r() local 368 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld1r() local 499 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld1r() local
|
D | 3x16c4-minmax-rndnu-neon-mlal-dup.c | 182 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_dup() local 359 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_dup() local 490 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_dup() local
|
D | 3x16c4-minmax-rndnu-neon-mlal-ld2r.c | 182 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld2r() local 359 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld2r() local 490 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mlal_ld2r() local
|
D | 3x16c4-minmax-rndnu-neon-mull-dup.c | 157 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mull_dup() local 288 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mull_dup() local
|
D | 3x16c4-minmax-rndnu-neon-mull-ld2r.c | 157 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld2r() local 288 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld2r() local
|
D | 3x16c4-minmax-rndnu-neon-mull-ld1r.c | 160 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld1r() local 291 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4__neon_mull_ld1r() local
|
D | 4x16c4-minmax-rndnu-neon-mlal-dup.c | 217 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local 441 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local 596 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_dup() local
|
D | 4x16c4-minmax-rndnu-neon-mlal-ld2r.c | 217 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local 441 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local 596 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld2r() local
|
D | 4x16c4-minmax-rndnu-neon-mlal-ld1r.c | 225 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2c0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local 453 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local 608 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local
|
D | 3x16c4s2-minmax-rndnu-neon-mlal.c | 174 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4s2__neon_mlal() local 345 int16x8_t vprod2xABc0 = vmull_s8(vbABc0x0, va2x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c4s2__neon_mlal() local
|
D | 4x16c4-minmax-rndnu-neon-mull-dup.c | 185 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_dup() local 340 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_dup() local
|
D | 4x16c4-minmax-rndnu-neon-mull-ld2r.c | 185 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld2r() local 340 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld2r() local
|
D | 4x16c4-minmax-rndnu-neon-mull-ld1r.c | 189 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r() local 344 const int16x8_t vprod2xABc0 = vmull_s8(vbABc0, va2c0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r() local
|