/external/libvpx/libvpx/vpx_dsp/arm/ |
D | highbd_idct16x16_add_neon.c | 56 t[0].val[1] = vmull_lane_s32(vget_high_s32(s0.val[0]), in highbd_idct_cospi_2_30() 60 t[1].val[1] = vmull_lane_s32(vget_high_s32(s0.val[1]), in highbd_idct_cospi_2_30() 64 t[2].val[1] = vmull_lane_s32(vget_high_s32(s1.val[0]), in highbd_idct_cospi_2_30() 68 t[3].val[1] = vmull_lane_s32(vget_high_s32(s1.val[1]), in highbd_idct_cospi_2_30() 72 t[0].val[1] = vmlsl_lane_s32(t[0].val[1], vget_high_s32(s1.val[0]), in highbd_idct_cospi_2_30() 76 t[1].val[1] = vmlsl_lane_s32(t[1].val[1], vget_high_s32(s1.val[1]), in highbd_idct_cospi_2_30() 80 t[2].val[1] = vmlal_lane_s32(t[2].val[1], vget_high_s32(s0.val[0]), in highbd_idct_cospi_2_30() 84 t[3].val[1] = vmlal_lane_s32(t[3].val[1], vget_high_s32(s0.val[1]), in highbd_idct_cospi_2_30() 97 vget_high_s32(cospi_4_12_20N_28), 1); in highbd_idct_cospi_4_28() 98 t[0].val[1] = vmull_lane_s32(vget_high_s32(s0.val[0]), in highbd_idct_cospi_4_28() [all …]
|
D | highbd_idct_neon.h | 45 b0 = vmulq_lane_s32(b0, vget_high_s32(cospis), 0); in idct4x4_16_kernel_bd10() 46 b1 = vmulq_lane_s32(b1, vget_high_s32(cospis), 0); in idct4x4_16_kernel_bd10() 47 b2 = vmulq_lane_s32(a[1], vget_high_s32(cospis), 1); in idct4x4_16_kernel_bd10() 50 b3 = vmlaq_lane_s32(b3, a[3], vget_high_s32(cospis), 1); in idct4x4_16_kernel_bd10() 69 c[0] = vmull_lane_s32(vget_low_s32(b0), vget_high_s32(cospis), 0); in idct4x4_16_kernel_bd12() 70 c[1] = vmull_lane_s32(vget_high_s32(b0), vget_high_s32(cospis), 0); in idct4x4_16_kernel_bd12() 71 c[2] = vmull_lane_s32(vget_low_s32(b1), vget_high_s32(cospis), 0); in idct4x4_16_kernel_bd12() 72 c[3] = vmull_lane_s32(vget_high_s32(b1), vget_high_s32(cospis), 0); in idct4x4_16_kernel_bd12() 73 c[4] = vmull_lane_s32(vget_low_s32(a[1]), vget_high_s32(cospis), 1); in idct4x4_16_kernel_bd12() 74 c[5] = vmull_lane_s32(vget_high_s32(a[1]), vget_high_s32(cospis), 1); in idct4x4_16_kernel_bd12() [all …]
|
D | highbd_idct8x8_add_neon.c | 81 step1[4] = vmulq_lane_s32(*io1, vget_high_s32(cospis1), 1); in idct8x8_12_half1d_bd10() 82 step1[5] = vmulq_lane_s32(*io3, vget_high_s32(cospis1), 0); in idct8x8_12_half1d_bd10() 91 step2[1] = vmulq_lane_s32(*io0, vget_high_s32(cospis0), 0); in idct8x8_12_half1d_bd10() 92 step2[2] = vmulq_lane_s32(*io2, vget_high_s32(cospis0), 1); in idct8x8_12_half1d_bd10() 109 step1[6] = vmulq_lane_s32(step2[6], vget_high_s32(cospis0), 0); in idct8x8_12_half1d_bd10() 110 step1[5] = vmlsq_lane_s32(step1[6], step2[5], vget_high_s32(cospis0), 0); in idct8x8_12_half1d_bd10() 111 step1[6] = vmlaq_lane_s32(step1[6], step2[5], vget_high_s32(cospis0), 0); in idct8x8_12_half1d_bd10() 141 input1h = vget_high_s32(*io1); in idct8x8_12_half1d_bd12() 143 input3h = vget_high_s32(*io3); in idct8x8_12_half1d_bd12() 145 step1h[0] = vget_high_s32(*io0); in idct8x8_12_half1d_bd12() [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_resi_trans_neon_32x32.c | 298 vst1_s32(pi4_temp + 256, vget_high_s32(val_8.val[0])); /*Value 8*/ in ihevc_resi_trans_32x32_neon() 300 vst1_s32(pi4_temp + 768, vget_high_s32(val_8.val[1])); /*Value 24*/ in ihevc_resi_trans_32x32_neon() 335 a[4].val[1] = vcombine_s32(vget_high_s32(val_4_0.val[0]), in ihevc_resi_trans_32x32_neon() 336 vget_high_s32(val_20_0.val[0])); //q2 in ihevc_resi_trans_32x32_neon() 340 a[12].val[1] = vcombine_s32(vget_high_s32(val_4_0.val[1]), in ihevc_resi_trans_32x32_neon() 341 vget_high_s32(val_20_0.val[1])); //q8 in ihevc_resi_trans_32x32_neon() 350 a[20].val[1] = vcombine_s32(vget_high_s32(val_4_1.val[0]), in ihevc_resi_trans_32x32_neon() 351 vget_high_s32(val_20_1.val[0])); //q13 in ihevc_resi_trans_32x32_neon() 355 a[28].val[1] = vcombine_s32(vget_high_s32(val_4_1.val[1]), in ihevc_resi_trans_32x32_neon() 356 vget_high_s32(val_20_1.val[1])); //q9 in ihevc_resi_trans_32x32_neon() [all …]
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_highbd_iht4x4_add_neon.c | 30 s[0].val[1] = vmull_lane_s32(vget_high_s32(io[0]), vget_low_s32(sinpi), 0); in highbd_iadst4() 32 s[1].val[1] = vmull_lane_s32(vget_high_s32(io[0]), vget_low_s32(sinpi), 1); in highbd_iadst4() 33 s[2].val[0] = vmull_lane_s32(vget_low_s32(io[1]), vget_high_s32(sinpi), 0); in highbd_iadst4() 34 s[2].val[1] = vmull_lane_s32(vget_high_s32(io[1]), vget_high_s32(sinpi), 0); in highbd_iadst4() 35 s[3].val[0] = vmull_lane_s32(vget_low_s32(io[2]), vget_high_s32(sinpi), 1); in highbd_iadst4() 36 s[3].val[1] = vmull_lane_s32(vget_high_s32(io[2]), vget_high_s32(sinpi), 1); in highbd_iadst4() 38 s[4].val[1] = vmull_lane_s32(vget_high_s32(io[2]), vget_low_s32(sinpi), 0); in highbd_iadst4() 40 s[5].val[1] = vmull_lane_s32(vget_high_s32(io[3]), vget_low_s32(sinpi), 1); in highbd_iadst4() 41 s[6].val[0] = vmull_lane_s32(vget_low_s32(io[3]), vget_high_s32(sinpi), 1); in highbd_iadst4() 42 s[6].val[1] = vmull_lane_s32(vget_high_s32(io[3]), vget_high_s32(sinpi), 1); in highbd_iadst4() [all …]
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 4x16c8-minmax-neon-mull-padal.c | 354 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 355 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 356 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 357 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 361 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 362 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 363 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 364 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 368 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() 369 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mull_padal() [all …]
|
D | 3x16c8-minmax-neon-mull-padal.c | 285 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 286 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 287 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 288 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 292 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 293 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 294 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 295 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 299 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() 300 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mull_padal() [all …]
|
D | 4x8c8-minmax-neon-mull-padal.c | 226 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 227 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 228 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 229 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 233 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 234 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 235 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 236 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 240 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 241 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() [all …]
|
D | 2x16c8-minmax-neon-mull-padal.c | 216 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 217 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 218 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 219 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 223 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 224 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 225 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 226 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 230 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() 231 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mull_padal() [all …]
|
D | 4x16c16-minmax-neon-mlal-padal.c | 418 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 419 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 420 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 421 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 425 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 426 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 427 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 428 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 432 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() 433 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_igemm_minmax_ukernel_4x16c16__neon_mlal_padal() [all …]
|
D | 3x16c16-minmax-neon-mlal-padal.c | 333 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 334 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 335 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 336 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 340 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 341 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 342 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 343 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 347 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() 348 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_igemm_minmax_ukernel_3x16c16__neon_mlal_padal() [all …]
|
D | 3x8c8-minmax-neon-mull-padal.c | 187 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 188 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 189 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 190 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 194 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 195 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 196 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 197 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 201 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() 202 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__neon_mull_padal() [all …]
|
D | 4x8c16-minmax-neon-mlal-padal.c | 258 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 259 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 260 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 261 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 265 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 266 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 267 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 268 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 272 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() 273 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_igemm_minmax_ukernel_4x8c16__neon_mlal_padal() [all …]
|
D | 3x8c16-minmax-neon-mlal-padal.c | 211 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 212 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 213 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 214 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 218 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 219 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 220 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 221 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 225 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() 226 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_igemm_minmax_ukernel_3x8c16__neon_mlal_padal() [all …]
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 4x16c8-minmax-neon-mull-padal.c | 334 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 335 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 336 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 337 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 341 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 342 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 343 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 344 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 348 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() 349 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mull_padal() [all …]
|
D | 3x16c8-minmax-neon-mull-padal.c | 267 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 268 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 269 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 270 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 274 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 275 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 276 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 277 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 281 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() 282 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mull_padal() [all …]
|
D | 4x8c8-minmax-neon-mull-padal.c | 206 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 207 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 208 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 209 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 213 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 214 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 215 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 216 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 220 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 221 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() [all …]
|
D | 2x16c8-minmax-neon-mull-padal.c | 200 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 201 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 202 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 203 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 207 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 208 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 209 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 210 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 214 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() 215 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mull_padal() [all …]
|
D | 4x16c16-minmax-neon-mlal-padal.c | 398 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 399 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 400 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 401 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 405 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 406 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 407 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 408 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 412 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() 413 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_gemm_minmax_ukernel_4x16c16__neon_mlal_padal() [all …]
|
D | 3x16c16-minmax-neon-mlal-padal.c | 315 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 316 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 317 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 318 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 322 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 323 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 324 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 325 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 329 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() 330 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_gemm_minmax_ukernel_3x16c16__neon_mlal_padal() [all …]
|
D | 3x8c8-minmax-neon-mull-padal.c | 169 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 170 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 171 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 172 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 176 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 177 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 178 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 179 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 183 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() 184 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__neon_mull_padal() [all …]
|
D | 2x16c16-minmax-neon-mlal-padal.c | 232 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 233 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 234 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 235 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 239 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 240 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 241 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 242 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 246 const int32x2_t vpsum0x8 = vadd_s32(vget_low_s32(vacc0x8), vget_high_s32(vacc0x8)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() 247 const int32x2_t vpsum0x9 = vadd_s32(vget_low_s32(vacc0x9), vget_high_s32(vacc0x9)); in xnn_qs8_gemm_minmax_ukernel_2x16c16__neon_mlal_padal() [all …]
|
D | 4x8c16-minmax-neon-mlal-padal.c | 238 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 239 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 240 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 241 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 245 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 246 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 247 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 248 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 252 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() 253 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_gemm_minmax_ukernel_4x8c16__neon_mlal_padal() [all …]
|
D | 2x8c8-minmax-neon-mull-padal.c | 132 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 133 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 134 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 135 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 139 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 140 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 141 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 142 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 146 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() 147 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__neon_mull_padal() [all …]
|
D | 3x8c16-minmax-neon-mlal-padal.c | 193 const int32x2_t vpsum0x0 = vadd_s32(vget_low_s32(vacc0x0), vget_high_s32(vacc0x0)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 194 const int32x2_t vpsum0x1 = vadd_s32(vget_low_s32(vacc0x1), vget_high_s32(vacc0x1)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 195 const int32x2_t vpsum0x2 = vadd_s32(vget_low_s32(vacc0x2), vget_high_s32(vacc0x2)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 196 const int32x2_t vpsum0x3 = vadd_s32(vget_low_s32(vacc0x3), vget_high_s32(vacc0x3)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 200 const int32x2_t vpsum0x4 = vadd_s32(vget_low_s32(vacc0x4), vget_high_s32(vacc0x4)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 201 const int32x2_t vpsum0x5 = vadd_s32(vget_low_s32(vacc0x5), vget_high_s32(vacc0x5)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 202 const int32x2_t vpsum0x6 = vadd_s32(vget_low_s32(vacc0x6), vget_high_s32(vacc0x6)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 203 const int32x2_t vpsum0x7 = vadd_s32(vget_low_s32(vacc0x7), vget_high_s32(vacc0x7)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 207 const int32x2_t vpsum1x0 = vadd_s32(vget_low_s32(vacc1x0), vget_high_s32(vacc1x0)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() 208 const int32x2_t vpsum1x1 = vadd_s32(vget_low_s32(vacc1x1), vget_high_s32(vacc1x1)); in xnn_qs8_gemm_minmax_ukernel_3x8c16__neon_mlal_padal() [all …]
|