/external/XNNPACK/src/qs8-gemm/gen/ |
D | 1x8-minmax-rndnu-neon-mull-addw-dup.c | 82 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8__neon_mull_addw_dup() local 140 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8__neon_mull_addw_dup() local
|
D | 2x8-minmax-rndnu-neon-mull-addw-dup.c | 109 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8__neon_mull_addw_dup() local 192 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8__neon_mull_addw_dup() local
|
D | 1x16-minmax-rndnu-neon-mull-addw-dup.c | 114 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16__neon_mull_addw_dup() local 207 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16__neon_mull_addw_dup() local
|
D | 3x8-minmax-rndnu-neon-mull-addw-dup.c | 136 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8__neon_mull_addw_dup() local 244 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8__neon_mull_addw_dup() local
|
D | 4x8-minmax-rndnu-neon-mull-addw-dup.c | 163 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8__neon_mull_addw_dup() local 296 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8__neon_mull_addw_dup() local
|
D | 2x16-minmax-rndnu-neon-mull-addw-dup.c | 161 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16__neon_mull_addw_dup() local 303 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16__neon_mull_addw_dup() local
|
D | 3x16-minmax-rndnu-neon-mull-addw-dup.c | 208 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16__neon_mull_addw_dup() local 399 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16__neon_mull_addw_dup() local
|
D | 4x16-minmax-rndnu-neon-mull-addw-dup.c | 255 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16__neon_mull_addw_dup() local 495 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16__neon_mull_addw_dup() local
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 1x8-minmax-rndnu-neon-mull-addw-dup.c | 93 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8__neon_mull_addw_dup() local 151 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8__neon_mull_addw_dup() local
|
D | 2x8-minmax-rndnu-neon-mull-addw-dup.c | 122 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8__neon_mull_addw_dup() local 205 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8__neon_mull_addw_dup() local
|
D | 1x16-minmax-rndnu-neon-mull-addw-dup.c | 125 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16__neon_mull_addw_dup() local 218 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16__neon_mull_addw_dup() local
|
D | 3x8-minmax-rndnu-neon-mull-addw-dup.c | 151 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8__neon_mull_addw_dup() local 259 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8__neon_mull_addw_dup() local
|
D | 4x8-minmax-rndnu-neon-mull-addw-dup.c | 180 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8__neon_mull_addw_dup() local 313 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8__neon_mull_addw_dup() local
|
D | 2x16-minmax-rndnu-neon-mull-addw-dup.c | 174 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16__neon_mull_addw_dup() local 316 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16__neon_mull_addw_dup() local
|
D | 3x16-minmax-rndnu-neon-mull-addw-dup.c | 223 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16__neon_mull_addw_dup() local 414 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16__neon_mull_addw_dup() local
|
D | 4x16-minmax-rndnu-neon-mull-addw-dup.c | 272 const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16__neon_mull_addw_dup() local 512 … const int16x8_t vprod0x01234567c6 = vmull_s8(vb01234567c6, vdup_lane_s8(va0, 6)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16__neon_mull_addw_dup() local
|