Home
last modified time | relevance | path

Searched refs:va01x1 (Results 1 – 25 of 96) sorted by relevance

1234

/external/XNNPACK/src/qs8-igemm/gen/
D1x8c2-minmax-fp32-neon-mlal-ld2r.c63 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r() local
97 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
108 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
D1x8c2-minmax-fp32-neonv8-mlal-ld2r.c64 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r() local
98 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
109 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
D1x8c2-minmax-rndnu-neon-mlal-ld2r.c63 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld2r() local
97 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld2r()
108 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld2r()
D2x8c2-minmax-fp32-neonv8-mlal-ld2r.c74 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r() local
128 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
147 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
D2x8c2-minmax-rndnu-neon-mlal-ld2r.c73 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld2r() local
127 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld2r()
146 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld2r()
D2x8c2-minmax-fp32-neon-mlal-ld2r.c73 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r() local
127 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
146 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
D1x16c2-minmax-rndnu-neon-mlal-ld2r.c65 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() local
123 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r()
142 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r()
/external/XNNPACK/src/qs8-gemm/gen/
D1x8c2-minmax-rndnu-neon-mlal-ld2r.c52 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld2r() local
85 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld2r()
96 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c2__neon_mlal_ld2r()
D1x8c2-minmax-fp32-neon-mlal-ld2r.c52 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r() local
85 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
96 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
D1x8c2-minmax-fp32-neonv8-mlal-ld2r.c53 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r() local
86 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
97 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
D2x8c2-minmax-fp32-neonv8-mlal-ld2r.c61 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r() local
114 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
133 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
D2x8c2-minmax-rndnu-neon-mlal-ld2r.c60 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld2r() local
113 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld2r()
132 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld2r()
D2x8c2-minmax-fp32-neon-mlal-ld2r.c60 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r() local
113 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
132 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
D1x8c4-minmax-fp32-neonv8-mlal-ld1r.c56 const int32x2_t va01x1 = vld1_dup_s32((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c4__neonv8_mlal_ld1r() local
87 const int8x8_t va0c1x1 = vreinterpret_s8_s32(va01x1); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c4__neonv8_mlal_ld1r()
D1x16c2-minmax-rndnu-neon-mlal-ld2r.c54 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() local
111 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r()
130 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r()
D1x8c2-minmax-fp32-neonv8-mlal-ld1r.c55 const int16x4_t va01x1 = vld1_dup_s16((const void*)(a0 + 2)); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r() local
79 const int8x8_t va0c1x1 = vreinterpret_s8_s16(va01x1); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r()
/external/XNNPACK/src/qc8-igemm/gen/
D1x8c2-minmax-fp32-neonv8-mlal-ld2r.c64 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r() local
98 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
109 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
D1x8c2-minmax-fp32-neon-mlal-ld2r.c63 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r() local
97 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
108 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
D2x8c2-minmax-fp32-neon-mlal-ld2r.c73 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r() local
127 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
146 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
D2x8c2-minmax-fp32-neonv8-mlal-ld2r.c74 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r() local
128 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
147 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
D1x8c2-minmax-fp32-neonv8-mlal-ld1r.c66 const int16x4_t va01x1 = vld1_dup_s16((const void*)(a0 + 2)); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r() local
91 const int8x8_t va0c1x1 = vreinterpret_s8_s16(va01x1); in xnn_qc8_igemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld1r()
/external/XNNPACK/src/qc8-gemm/gen/
D1x8c2-minmax-fp32-neon-mlal-ld2r.c52 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r() local
85 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
96 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neon_mlal_ld2r()
D1x8c2-minmax-fp32-neonv8-mlal-ld2r.c53 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r() local
86 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
97 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_gemm_minmax_fp32_ukernel_1x8c2__neonv8_mlal_ld2r()
D2x8c2-minmax-fp32-neon-mlal-ld2r.c60 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r() local
113 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
132 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld2r()
D2x8c2-minmax-fp32-neonv8-mlal-ld2r.c61 const int16x4x2_t va01x1 = vld2_dup_s16((const void*)(a0 + 4)); a0 += 8; in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r() local
114 const int8x8_t va0c2x1 = vreinterpret_s8_s16(va01x1.val[0]); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()
133 const int8x8_t va0c3x1 = vreinterpret_s8_s16(va01x1.val[1]); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld2r()

1234