Searched refs:vb15x0 (Results 1 – 8 of 8) sorted by relevance
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x16c8-minmax-rndnu-neon-mlal.c | 145 … const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local 298 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() 299 int16x8_t vprod1x15 = vmull_s8(vb15x0, va1x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() 300 int16x8_t vprod2x15 = vmull_s8(vb15x0, va2x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
|
D | 2x16c8-minmax-rndnu-neon-mlal.c | 119 … const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local 227 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() 228 int16x8_t vprod1x15 = vmull_s8(vb15x0, va1x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
|
D | 1x16c8-minmax-rndnu-neon-mlal.c | 93 … const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() local 156 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal()
|
D | 4x16c8-minmax-rndnu-neon-mlal.c | 171 … const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local 369 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() 370 int16x8_t vprod1x15 = vmull_s8(vb15x0, va1x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() 371 int16x8_t vprod2x15 = vmull_s8(vb15x0, va2x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() 372 int16x8_t vprod3x15 = vmull_s8(vb15x0, va3x0); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c8__neon_mlal()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 3x16c8-minmax-rndnu-neon-mlal.c | 130 const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() local 283 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() 284 int16x8_t vprod1x15 = vmull_s8(vb15x0, va1x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal() 285 int16x8_t vprod2x15 = vmull_s8(vb15x0, va2x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c8__neon_mlal()
|
D | 2x16c8-minmax-rndnu-neon-mlal.c | 106 const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() local 214 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal() 215 int16x8_t vprod1x15 = vmull_s8(vb15x0, va1x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c8__neon_mlal()
|
D | 1x16c8-minmax-rndnu-neon-mlal.c | 82 const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() local 145 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal()
|
D | 4x16c8-minmax-rndnu-neon-mlal.c | 154 const int8x8_t vb15x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() local 352 int16x8_t vprod0x15 = vmull_s8(vb15x0, va0x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() 353 int16x8_t vprod1x15 = vmull_s8(vb15x0, va1x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() 354 int16x8_t vprod2x15 = vmull_s8(vb15x0, va2x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal() 355 int16x8_t vprod3x15 = vmull_s8(vb15x0, va3x0); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c8__neon_mlal()
|