/external/XNNPACK/src/bf16-gemm/gen/ |
D | 2x4c8-minmax-neonbf16-bfdot.c | 124 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfdot() local 129 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfdot()
|
D | 2x4c8-minmax-neonbf16-bfmlal.c | 141 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfmlal() local 146 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonbf16_bfmlal()
|
D | 3x4c8-minmax-neonbf16-bfdot.c | 153 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfdot() local 160 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfdot()
|
D | 2x4c8-minmax-neonfma-shland.c | 188 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonfma_shland() local 193 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonfma_shland()
|
D | 2x4c8-minmax-neonfma-zip.c | 188 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonfma_zip() local 193 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_2x4c8__neonfma_zip()
|
D | 3x4c8-minmax-neonbf16-bfmlal.c | 178 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfmlal() local 185 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonbf16_bfmlal()
|
D | 4x4c8-minmax-neonbf16-bfdot.c | 182 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot() local 191 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfdot()
|
D | 3x4c8-minmax-neonfma-zip.c | 235 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonfma_zip() local 242 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonfma_zip()
|
D | 4x4c8-minmax-neonbf16-bfmlal.c | 215 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal() local 224 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonbf16_bfmlal()
|
D | 5x4c8-minmax-neonbf16-bfdot.c | 211 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot() local 222 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfdot()
|
D | 3x4c8-minmax-neonfma-shland.c | 235 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonfma_shland() local 242 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_3x4c8__neonfma_shland()
|
D | 5x4c8-minmax-neonbf16-bfmlal.c | 252 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal() local 263 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonbf16_bfmlal()
|
D | 4x4c8-minmax-neonfma-shland.c | 282 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland() local 291 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_shland()
|
D | 4x4c8-minmax-neonfma-zip.c | 282 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip() local 291 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_4x4c8__neonfma_zip()
|
D | 5x4c8-minmax-neonfma-zip.c | 329 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonfma_zip() local 340 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonfma_zip()
|
D | 5x4c8-minmax-neonfma-shland.c | 329 const float32x2_t vsum1x2 = vadd_f32(vget_low_f32(vacc1x2), vget_high_f32(vacc1x2)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonfma_shland() local 340 float32x4_t vacc1x0123 = vcombine_f32(vpadd_f32(vsum1x0, vsum1x1), vpadd_f32(vsum1x2, vsum1x3)); in xnn_bf16_gemm_minmax_ukernel_5x4c8__neonfma_shland()
|