/external/XNNPACK/src/qs8-igemm/gen/ |
D | 4x8c2-minmax-rndnu-neon-mull-ld2r.c | 97 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld2r() local 112 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld2r() 133 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld2r()
|
D | 4x8c2-minmax-rndnu-neon-mlal-ld2r.c | 267 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld2r() local 282 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld2r() 303 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld2r()
|
D | 4x16c2-minmax-rndnu-neon-mull-ld2r.c | 105 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld2r() local 128 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld2r() 165 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld2r()
|
D | 4x8c2-minmax-rndnu-neon-mull-ld1r.c | 103 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r() local 120 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r()
|
D | 4x8c4-minmax-rndnu-neon-mull-ld1r.c | 106 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c4__neon_mull_ld1r() local 121 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c4__neon_mull_ld1r()
|
D | 4x8c4-minmax-rndnu-neon-mlal-ld1r.c | 258 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c4__neon_mlal_ld1r() local 273 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c4__neon_mlal_ld1r()
|
D | 4x8c2-minmax-rndnu-neon-mlal-ld1r.c | 289 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld1r() local 306 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mlal-ld2r.c | 387 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld2r() local 410 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld2r() 447 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld2r()
|
D | 4x16c4-minmax-rndnu-neon-mull-ld1r.c | 122 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r() local 145 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mull-ld1r.c | 111 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r() local 136 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r()
|
D | 4x16c4-minmax-rndnu-neon-mlal-ld1r.c | 386 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local 409 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mlal-ld1r.c | 409 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld1r() local 434 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld1r()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 4x8c2-minmax-rndnu-neon-mull-ld2r.c | 80 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld2r() local 95 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld2r() 116 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld2r()
|
D | 4x16c2-minmax-rndnu-neon-mull-ld2r.c | 88 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld2r() local 111 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld2r() 148 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld2r()
|
D | 4x8c2-minmax-rndnu-neon-mlal-ld2r.c | 249 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld2r() local 264 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld2r() 285 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld2r()
|
D | 4x8c2-minmax-rndnu-neon-mull-ld1r.c | 86 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r() local 103 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r()
|
D | 4x8c4-minmax-rndnu-neon-mull-ld1r.c | 89 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c4__neon_mull_ld1r() local 104 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c4__neon_mull_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mull-ld1r.c | 94 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r() local 119 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r()
|
D | 4x8c2-minmax-rndnu-neon-mlal-ld1r.c | 271 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld1r() local 288 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mlal_ld1r()
|
D | 4x8c4-minmax-rndnu-neon-mlal-ld1r.c | 241 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c4__neon_mlal_ld1r() local 256 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c4__neon_mlal_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mlal-ld2r.c | 369 const int16x4x2_t va30 = vld2_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld2r() local 392 const int8x8_t va3c0 = vreinterpret_s8_s16(va30.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld2r() 429 const int8x8_t va3c1 = vreinterpret_s8_s16(va30.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld2r()
|
D | 4x16c4-minmax-rndnu-neon-mull-ld1r.c | 105 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r() local 128 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mull_ld1r()
|
D | 4x16c4-minmax-rndnu-neon-mlal-ld1r.c | 369 const int32x2_t va30 = vld1_dup_s32((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r() local 392 const int8x8_t va3c0 = vreinterpret_s8_s32(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c4__neon_mlal_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mlal-ld1r.c | 391 const int16x4_t va30 = vld1_dup_s16((const void*)a3); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld1r() local 416 const int8x8_t va3c0 = vreinterpret_s8_s16(va30); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mlal_ld1r()
|