/external/XNNPACK/src/qs8-gemm/gen/ |
D | 2x16c4s2-minmax-rndnu-neon-mull.c | 166 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull() local 189 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull() local
|
D | 2x16c8-minmax-fp32-avx512skx.c | 62 __m512i vacc1x89AB = vacc0x89AB; in xnn_qs8_gemm_minmax_fp32_ukernel_2x16c8__avx512skx() local
|
D | 3x16c8-minmax-fp32-avx512skx.c | 68 __m512i vacc1x89AB = vacc0x89AB; in xnn_qs8_gemm_minmax_fp32_ukernel_3x16c8__avx512skx() local
|
D | 2x16c4-minmax-rndnu-neon-mull-dup.c | 221 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup() local 244 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup() local
|
D | 2x16c4-minmax-rndnu-neon-mull-ld2r.c | 221 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r() local 244 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r() local
|
D | 2x16c4-minmax-rndnu-neon-mull-ld1r.c | 223 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r() local 246 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r() local
|
D | 3x16c4s2-minmax-rndnu-neon-mull.c | 214 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4s2__neon_mull() local 241 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c4s2__neon_mull() local
|
D | 4x16c8-minmax-fp32-avx512skx.c | 74 __m512i vacc1x89AB = vacc0x89AB; in xnn_qs8_gemm_minmax_fp32_ukernel_4x16c8__avx512skx() local
|
/external/XNNPACK/src/f32-prelu/gen/ |
D | wasmsimd-minmax-2x16.c | 78 v128_t vacc1x89AB = wasm_i32x4_max(vi1x89AB, vzero); in xnn_f32_prelu_ukernel__wasmsimd_minmax_2x16() local
|
D | wasmsimd-bitselect-2x16.c | 77 v128_t vacc1x89AB = wasm_f32x4_mul(vi1x89AB, vw89AB); in xnn_f32_prelu_ukernel__wasmsimd_bitselect_2x16() local
|
D | neon-2x16.c | 74 float32x4_t vacc1x89AB = vmulq_f32(vi1x89AB, vw89AB); in xnn_f32_prelu_ukernel__neon_2x16() local
|
/external/XNNPACK/src/qu8-gemm/gen/ |
D | 2x16c8-minmax-fp32-avx512skx.c | 62 __m512i vacc1x89AB = vacc0x89AB; in xnn_qu8_gemm_minmax_fp32_ukernel_2x16c8__avx512skx() local
|
D | 3x16c8-minmax-fp32-avx512skx.c | 68 __m512i vacc1x89AB = vacc0x89AB; in xnn_qu8_gemm_minmax_fp32_ukernel_3x16c8__avx512skx() local
|
/external/XNNPACK/src/qc8-gemm/gen/ |
D | 2x16c8-minmax-fp32-avx512skx.c | 61 __m512i vacc1x89AB = vacc0x89AB; in xnn_qc8_gemm_minmax_fp32_ukernel_2x16c8__avx512skx() local
|
D | 3x16c8-minmax-fp32-avx512skx.c | 67 __m512i vacc1x89AB = vacc0x89AB; in xnn_qc8_gemm_minmax_fp32_ukernel_3x16c8__avx512skx() local
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 2x16c4s2-minmax-rndnu-neon-mull.c | 182 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull() local 205 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4s2__neon_mull() local
|
D | 2x16c8-minmax-fp32-avx512skx.c | 61 __m512i vacc1x89AB = vacc0x89AB; in xnn_qs8_igemm_minmax_fp32_ukernel_2x16c8__avx512skx() local
|
D | 3x16c8-minmax-fp32-avx512skx.c | 65 __m512i vacc1x89AB = vacc0x89AB; in xnn_qs8_igemm_minmax_fp32_ukernel_3x16c8__avx512skx() local
|
D | 2x16c4-minmax-rndnu-neon-mull-dup.c | 236 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup() local 259 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_dup() local
|
D | 2x16c4-minmax-rndnu-neon-mull-ld2r.c | 236 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r() local 259 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld2r() local
|
D | 2x16c4-minmax-rndnu-neon-mull-ld1r.c | 238 int32x4_t vacc1x89AB = vpaddq_s32(vacc1x89, vacc1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r() local 261 int32x4_t vacc1x89AB = vcombine_s32(vsum1x89, vsum1xAB); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c4__neon_mull_ld1r() local
|
/external/XNNPACK/src/qu8-igemm/gen/ |
D | 2x16c8-minmax-fp32-avx512skx.c | 61 __m512i vacc1x89AB = vacc0x89AB; in xnn_qu8_igemm_minmax_fp32_ukernel_2x16c8__avx512skx() local
|
D | 3x16c8-minmax-fp32-avx512skx.c | 65 __m512i vacc1x89AB = vacc0x89AB; in xnn_qu8_igemm_minmax_fp32_ukernel_3x16c8__avx512skx() local
|
/external/XNNPACK/src/qc8-igemm/gen/ |
D | 2x16c8-minmax-fp32-avx512skx.c | 60 __m512i vacc1x89AB = vacc0x89AB; in xnn_qc8_igemm_minmax_fp32_ukernel_2x16c8__avx512skx() local
|
D | 3x16c8-minmax-fp32-avx512skx.c | 64 __m512i vacc1x89AB = vacc0x89AB; in xnn_qc8_igemm_minmax_fp32_ukernel_3x16c8__avx512skx() local
|