/external/XNNPACK/src/qc8-igemm/gen/ |
D | 1x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 68 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r() local 113 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r()
|
D | 1x8c2-minmax-fp32-neon-mlal-ld1r.c | 67 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r() local 112 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 77 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 154 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 78 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 155 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 1x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 57 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r() local 101 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r()
|
D | 1x8c2-minmax-rndnu-neon-mlal-ld1r.c | 56 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld1r() local 100 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld1r()
|
D | 1x8c2-minmax-fp32-neon-mlal-ld1r.c | 56 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r() local 100 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 64 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 140 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-rndnu-neon-mlal-ld1r.c | 64 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r() local 140 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 65 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 141 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
D | 1x16c2-minmax-rndnu-neon-mlal-ld1r.c | 58 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() local 134 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r()
|
D | 3x8c2-minmax-rndnu-neon-mlal-ld1r.c | 72 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r() local 180 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r()
|
/external/XNNPACK/src/qc8-gemm/gen/ |
D | 1x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 57 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r() local 101 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r()
|
D | 1x8c2-minmax-fp32-neon-mlal-ld1r.c | 56 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r() local 100 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 64 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 140 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 65 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 141 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 1x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 68 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r() local 113 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r()
|
D | 1x8c2-minmax-rndnu-neon-mlal-ld1r.c | 67 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld1r() local 112 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld1r()
|
D | 1x8c2-minmax-fp32-neon-mlal-ld1r.c | 67 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r() local 112 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 78 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 155 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
D | 2x8c2-minmax-rndnu-neon-mlal-ld1r.c | 77 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r() local 154 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 1x16c2-minmax-rndnu-neon-mlal-ld1r.c | 69 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() local 146 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 77 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 154 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 3x8c2-minmax-rndnu-neon-mlal-ld1r.c | 87 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r() local 196 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r()
|
D | 2x16c2-minmax-rndnu-neon-mlal-ld1r.c | 81 const int16x4_t va03x1 = vld1_dup_s16((const void*)(a0 + 6)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mlal_ld1r() local 208 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va03x1); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mlal_ld1r()
|