/external/XNNPACK/src/qs8-gemm/gen/ |
D | 1x8c4s2-minmax-rndnu-neon-mull.c | 85 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4s2__neon_mull() local 92 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4s2__neon_mull() local
|
D | 1x8c4-minmax-rndnu-neon-mull-ld2r.c | 109 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld2r() local 116 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld2r() local
|
D | 1x8c4-minmax-rndnu-neon-mull-dup.c | 109 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4__neon_mull_dup() local 116 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4__neon_mull_dup() local
|
D | 1x8c4-minmax-rndnu-neon-mull-ld1r.c | 110 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld1r() local 117 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld1r() local
|
D | 1x8c16-minmax-rndnu-neon-mlal.c | 101 int32x4_t vacc0x4567 = vpaddq_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c16__neon_mlal() local 116 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67 ); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c16__neon_mlal() local
|
D | 1x8c8-minmax-rndnu-neon-mull.c | 94 int32x4_t vacc0x4567 = vpaddq_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c8__neon_mull() local 109 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67 ); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c8__neon_mull() local
|
D | 2x8c4s2-minmax-rndnu-neon-mull.c | 113 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c4s2__neon_mull() local 122 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c4s2__neon_mull() local
|
D | 1x8c4s2-minmax-fp32-neonv8-mlal.c | 135 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c4s2__neonv8_mlal() local 142 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_fp32_ukernel_1x8c4s2__neonv8_mlal() local
|
D | 1x8c4s2-minmax-rndnu-neon-mlal.c | 134 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4s2__neon_mlal() local 141 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x8c4s2__neon_mlal() local
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 1x8c4s2-minmax-rndnu-neon-mull.c | 99 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4s2__neon_mull() local 106 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4s2__neon_mull() local
|
D | 1x8c4-minmax-rndnu-neon-mull-dup.c | 122 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4__neon_mull_dup() local 129 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4__neon_mull_dup() local
|
D | 1x8c4-minmax-rndnu-neon-mull-ld1r.c | 123 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld1r() local 130 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld1r() local
|
D | 1x8c4-minmax-rndnu-neon-mull-ld2r.c | 122 int32x4_t vacc0x4567 = vpaddq_s32(vacc0x45, vacc0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld2r() local 129 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c4__neon_mull_ld2r() local
|
D | 1x8c8-minmax-rndnu-neon-mull.c | 108 int32x4_t vacc0x4567 = vpaddq_s32(vsum0x45, vsum0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c8__neon_mull() local 123 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67 ); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c8__neon_mull() local
|
D | 1x8c16-minmax-rndnu-neon-mlal.c | 115 int32x4_t vacc0x4567 = vpaddq_s32(vsum0x45, vsum0x67); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c16__neon_mlal() local 130 int32x4_t vacc0x4567 = vcombine_s32(vsum0x45, vsum0x67 ); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x8c16__neon_mlal() local
|
/external/XNNPACK/src/f32-gemm/gen/ |
D | 1x8-wasmsimd-loadsplat.c | 43 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_ukernel_1x8__wasmsimd_loadsplat() local
|
D | 1x8-wasmrelaxedsimd-fma-loadsplat.c | 43 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_ukernel_1x8__wasmrelaxedsimd_fma_loadsplat() local
|
D | 1x8-relu-wasmsimd-loadsplat.c | 43 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_relu_ukernel_1x8__wasmsimd_loadsplat() local
|
D | 1x8-relu-wasmrelaxedsimd-fma-loadsplat.c | 43 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_relu_ukernel_1x8__wasmrelaxedsimd_fma_loadsplat() local
|
D | 1x8-minmax-wasmrelaxedsimd-fma-loadsplat.c | 45 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_minmax_ukernel_1x8__wasmrelaxedsimd_fma_loadsplat() local
|
D | 1x8-minmax-wasmsimd-x86-loadsplat.c | 45 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_minmax_ukernel_1x8__wasmsimd_x86_loadsplat() local
|
D | 1x8-minmax-wasmsimd-arm-loadsplat.c | 45 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_minmax_ukernel_1x8__wasmsimd_arm_loadsplat() local
|
D | 1x8-minmax-sse-load1.c | 43 __m128 vacc0x4567 = _mm_load_ps(w + 4); in xnn_f32_gemm_minmax_ukernel_1x8__sse_load1() local
|
D | 1x8-minmax-wasmrelaxedsimd-loadsplat.c | 45 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_gemm_minmax_ukernel_1x8__wasmrelaxedsimd_loadsplat() local
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 1x8-wasmsimd-loadsplat.c | 47 v128_t vacc0x4567 = wasm_v128_load(w + 4); in xnn_f32_igemm_ukernel_1x8__wasmsimd_loadsplat() local
|