/external/XNNPACK/src/qs8-gemm/gen/ |
D | 3x8c2-minmax-rndnu-neon-mull-dup.c | 81 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_dup() local 165 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_dup() local
|
D | 3x8c2-minmax-rndnu-neon-mull-ld2r.c | 84 const int8x8_t va2c0 = vreinterpret_s8_s16(va20.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld2r() local 168 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld2r() local
|
D | 3x8c2-minmax-rndnu-neon-mull-ld4r.c | 81 const int8x8_t va2c0 = vreinterpret_s8_s16(va2.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld4r() local 165 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld4r() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld2r.c | 88 const int8x8_t va2c0 = vreinterpret_s8_s32(va2.val[0]); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld2r() local 174 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld2r() local
|
D | 3x8c2-minmax-rndnu-neon-mull-ld1r.c | 90 const int8x8_t va2c0 = vreinterpret_s8_s16(va20); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local 174 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local
|
D | 3x8c4-minmax-rndnu-neon-mull-dup.c | 88 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_dup() local 174 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_dup() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld1r.c | 91 const int8x8_t va2c0 = vreinterpret_s8_s32(va20); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld1r() local 177 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld1r() local
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x8c2-minmax-rndnu-neon-mull-dup.c | 96 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_dup() local 180 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_dup() local
|
D | 3x8c2-minmax-rndnu-neon-mull-ld4r.c | 96 const int8x8_t va2c0 = vreinterpret_s8_s16(va2.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld4r() local 180 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld4r() local
|
D | 3x8c2-minmax-rndnu-neon-mull-ld2r.c | 99 const int8x8_t va2c0 = vreinterpret_s8_s16(va20.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld2r() local 183 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld2r() local
|
D | 3x8c2-minmax-rndnu-neon-mull-ld1r.c | 105 const int8x8_t va2c0 = vreinterpret_s8_s16(va20); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local 189 const int8x8_t va2c0 = vreinterpret_s8_s16(vdup_lane_s16(vreinterpret_s16_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld1r.c | 106 const int8x8_t va2c0 = vreinterpret_s8_s32(va20); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld1r() local 192 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld1r() local
|
D | 3x8c4-minmax-rndnu-neon-mull-dup.c | 103 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_dup() local 189 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_dup() local
|
D | 3x8c4-minmax-rndnu-neon-mull-ld2r.c | 103 const int8x8_t va2c0 = vreinterpret_s8_s32(va2.val[0]); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld2r() local 189 const int8x8_t va2c0 = vreinterpret_s8_s32(vdup_lane_s32(vreinterpret_s32_s8(va2), 0)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c4__neon_mull_ld2r() local
|
/external/XNNPACK/src/f32-gemm/gen/ |
D | 4x2-minmax-neonfma-lane-ld64.c | 83 const float32x2_t va2c0 = vdup_lane_f32(va2, 0); in xnn_f32_gemm_minmax_ukernel_4x2__neonfma_lane_ld64() local
|
D | 3x8-wasmsimd-splat.c | 73 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemm_ukernel_3x8__wasmsimd_splat() local
|
D | 3x8-wasmrelaxedsimd-fma-splat.c | 73 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemm_ukernel_3x8__wasmrelaxedsimd_fma_splat() local
|
D | 3x8-relu-wasmrelaxedsimd-fma-splat.c | 73 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat() local
|
D | 3x8-relu-wasmsimd-splat.c | 73 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemm_relu_ukernel_3x8__wasmsimd_splat() local
|
D | 3x8-minmax-wasmrelaxedsimd-fma-splat.c | 75 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() local
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 4x2-minmax-neonfma-lane-ld64.c | 105 const float32x2_t va2c0 = vdup_lane_f32(va2, 0); in xnn_f32_igemm_minmax_ukernel_4x2__neonfma_lane_ld64() local
|
D | 3x8-wasmsimd-splat.c | 92 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_igemm_ukernel_3x8__wasmsimd_splat() local
|
D | 3x8-wasmrelaxedsimd-fma-splat.c | 92 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_igemm_ukernel_3x8__wasmrelaxedsimd_fma_splat() local
|
/external/XNNPACK/src/f32-gemm/gen-inc/ |
D | 3x8inc-minmax-wasmsimd-arm-splat.c | 77 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat() local
|
D | 3x8inc-minmax-wasmrelaxedsimd-splat.c | 77 const v128_t va2c0 = wasm_v32x4_shuffle(va2, va2, 0, 0, 0, 0); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_splat() local
|