/external/XNNPACK/src/math/ |
D | expm1minus-wasmsimd-rr2-p6-max.c | 22 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 24 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 25 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 27 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E440p-1f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 28 const v128_t vminus_ln2_lo = wasm_f32x4_splat(0x1.0105C6p-21f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 32 const v128_t vc6 = wasm_f32x4_splat(0x1.6b7338p-10f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 33 const v128_t vc5 = wasm_f32x4_splat(0x1.12278Ep-7f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 34 const v128_t vc4 = wasm_f32x4_splat(0x1.555716p-5f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 35 const v128_t vc3 = wasm_f32x4_splat(0x1.5554B0p-3f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFFFEp-2f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() [all …]
|
D | expm1minus-wasmsimd-rr2-p6-andnot.c | 22 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 23 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 25 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 27 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E440p-1f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 28 const v128_t vminus_ln2_lo = wasm_f32x4_splat(0x1.0105C6p-21f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 32 const v128_t vc6 = wasm_f32x4_splat(0x1.6b7338p-10f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 33 const v128_t vc5 = wasm_f32x4_splat(0x1.12278Ep-7f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 34 const v128_t vc4 = wasm_f32x4_splat(0x1.555716p-5f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 35 const v128_t vc3 = wasm_f32x4_splat(0x1.5554B0p-3f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFFFEp-2f); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() [all …]
|
D | sigmoid-wasmsimd-rr2-p5-div.c | 23 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 24 const v128_t vminus_log2e = wasm_f32x4_splat(-0x1.715476p+0f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 26 const v128_t vln2_hi = wasm_f32x4_splat(0x1.62E400p-1f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 27 const v128_t vln2_lo = wasm_f32x4_splat(0x1.7F7D1Cp-20f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 30 const v128_t vc5 = wasm_f32x4_splat(-0x1.0F9F9Cp-7f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 31 const v128_t vc4 = wasm_f32x4_splat( 0x1.573A1Ap-5f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 32 const v128_t vc3 = wasm_f32x4_splat(-0x1.555A80p-3f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 33 const v128_t vc2 = wasm_f32x4_splat( 0x1.FFFDC6p-2f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 34 const v128_t vc1 = wasm_f32x4_splat(-0x1.FFFFF6p-1f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() 35 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_math_f32_sigmoid__wasmsimd_rr2_p5_div() [all …]
|
D | expm1minus-wasmsimd-rr2-lut16-p3-max.c | 26 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 28 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.800000p19f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 29 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 33 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 34 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 38 const v128_t vc3 = wasm_f32x4_splat(0x1.55561Cp-3f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 39 const v128_t vc2 = wasm_f32x4_splat(0x1.0001ECp-1f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() 40 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max()
|
D | expm1minus-wasmsimd-rr2-lut16-p3-andnot.c | 26 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.800000p19f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 27 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 31 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 33 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 34 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 38 const v128_t vc3 = wasm_f32x4_splat(0x1.55561Cp-3f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 39 const v128_t vc2 = wasm_f32x4_splat(0x1.0001ECp-1f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() 40 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot()
|
D | sigmoid-wasmsimd-rr2-lut64-p2-div.c | 26 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.800000p17f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div() 27 const v128_t vminus_log2e = wasm_f32x4_splat(-0x1.715476p0f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div() 31 const v128_t vln2_hi = wasm_f32x4_splat(0x1.630000p-1f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div() 32 const v128_t vln2_lo = wasm_f32x4_splat(-0x1.BD0106p-13f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div() 34 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFF0Ap-2f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div() 35 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div() 38 const v128_t vdenorm_cutoff = wasm_f32x4_splat(0x1.5D589Ep+6f); in xnn_math_f32_sigmoid__wasmsimd_rr2_lut64_p2_div()
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | wasmsimd-p5-div-x4.c | 26 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 27 const v128_t vminus_log2e = wasm_f32x4_splat(-0x1.715476p+0f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 28 const v128_t vln2_hi = wasm_f32x4_splat(0x1.62E400p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 29 const v128_t vln2_lo = wasm_f32x4_splat(0x1.7F7D1Cp-20f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 30 const v128_t vc5 = wasm_f32x4_splat(-0x1.0F9F9Cp-7f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 31 const v128_t vc4 = wasm_f32x4_splat( 0x1.573A1Ap-5f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 32 const v128_t vc3 = wasm_f32x4_splat(-0x1.555A80p-3f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 33 const v128_t vc2 = wasm_f32x4_splat( 0x1.FFFDC6p-2f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 34 const v128_t vc1 = wasm_f32x4_splat(-0x1.FFFFF6p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() 35 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x4() [all …]
|
D | wasmsimd-p5-div-x8.c | 26 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 27 const v128_t vminus_log2e = wasm_f32x4_splat(-0x1.715476p+0f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 28 const v128_t vln2_hi = wasm_f32x4_splat(0x1.62E400p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 29 const v128_t vln2_lo = wasm_f32x4_splat(0x1.7F7D1Cp-20f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 30 const v128_t vc5 = wasm_f32x4_splat(-0x1.0F9F9Cp-7f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 31 const v128_t vc4 = wasm_f32x4_splat( 0x1.573A1Ap-5f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 32 const v128_t vc3 = wasm_f32x4_splat(-0x1.555A80p-3f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 33 const v128_t vc2 = wasm_f32x4_splat( 0x1.FFFDC6p-2f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 34 const v128_t vc1 = wasm_f32x4_splat(-0x1.FFFFF6p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() 35 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x8() [all …]
|
D | wasmsimd-p5-div-x12.c | 26 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 27 const v128_t vminus_log2e = wasm_f32x4_splat(-0x1.715476p+0f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 28 const v128_t vln2_hi = wasm_f32x4_splat(0x1.62E400p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 29 const v128_t vln2_lo = wasm_f32x4_splat(0x1.7F7D1Cp-20f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 30 const v128_t vc5 = wasm_f32x4_splat(-0x1.0F9F9Cp-7f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 31 const v128_t vc4 = wasm_f32x4_splat( 0x1.573A1Ap-5f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 32 const v128_t vc3 = wasm_f32x4_splat(-0x1.555A80p-3f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 33 const v128_t vc2 = wasm_f32x4_splat( 0x1.FFFDC6p-2f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 34 const v128_t vc1 = wasm_f32x4_splat(-0x1.FFFFF6p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() 35 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_f32_sigmoid_ukernel__wasmsimd_p5_div_x12() [all …]
|
D | wasmsimd-lut64-p2-div-x4.c | 28 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.800000p17f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() 29 const v128_t vminus_log2e = wasm_f32x4_splat(-0x1.715476p0f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() 31 const v128_t vln2_hi = wasm_f32x4_splat(0x1.630000p-1f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() 32 const v128_t vln2_lo = wasm_f32x4_splat(-0x1.BD0106p-13f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() 33 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFF0Ap-2f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() 34 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() 35 const v128_t vdenorm_cutoff = wasm_f32x4_splat(0x1.5D589Ep+6f); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-wasmsimd-arm-rr2-p6-x4.c | 33 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 34 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 35 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 36 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E440p-1f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 37 const v128_t vminus_ln2_lo = wasm_f32x4_splat(0x1.0105C6p-21f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 38 const v128_t vc6 = wasm_f32x4_splat(0x1.6b7338p-10f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.12278Ep-7f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 40 const v128_t vc4 = wasm_f32x4_splat(0x1.555716p-5f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 41 const v128_t vc3 = wasm_f32x4_splat(0x1.5554B0p-3f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() 42 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFFFEp-2f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x4() [all …]
|
D | velu-wasmsimd-x86-rr2-p6-x4.c | 33 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 34 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 35 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 36 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E440p-1f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 37 const v128_t vminus_ln2_lo = wasm_f32x4_splat(0x1.0105C6p-21f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 38 const v128_t vc6 = wasm_f32x4_splat(0x1.6b7338p-10f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.12278Ep-7f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 40 const v128_t vc4 = wasm_f32x4_splat(0x1.555716p-5f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 41 const v128_t vc3 = wasm_f32x4_splat(0x1.5554B0p-3f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() 42 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFFFEp-2f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x4() [all …]
|
D | velu-wasmsimd-arm-rr2-p6-x8.c | 33 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 34 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 35 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 36 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E440p-1f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 37 const v128_t vminus_ln2_lo = wasm_f32x4_splat(0x1.0105C6p-21f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 38 const v128_t vc6 = wasm_f32x4_splat(0x1.6b7338p-10f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.12278Ep-7f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 40 const v128_t vc4 = wasm_f32x4_splat(0x1.555716p-5f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 41 const v128_t vc3 = wasm_f32x4_splat(0x1.5554B0p-3f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() 42 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFFFEp-2f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_p6_x8() [all …]
|
D | velu-wasmsimd-x86-rr2-p6-x8.c | 33 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 34 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 35 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 36 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E440p-1f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 37 const v128_t vminus_ln2_lo = wasm_f32x4_splat(0x1.0105C6p-21f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 38 const v128_t vc6 = wasm_f32x4_splat(0x1.6b7338p-10f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.12278Ep-7f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 40 const v128_t vc4 = wasm_f32x4_splat(0x1.555716p-5f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 41 const v128_t vc3 = wasm_f32x4_splat(0x1.5554B0p-3f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() 42 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFFFEp-2f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_p6_x8() [all …]
|
D | velu-wasmsimd-arm-rr2-lut16-p3-x4.c | 35 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 36 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.800000p19f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 37 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 39 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 40 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 41 const v128_t vc3 = wasm_f32x4_splat(0x1.55561Cp-3f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 42 const v128_t vc2 = wasm_f32x4_splat(0x1.0001ECp-1f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() 43 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
|
D | velu-wasmsimd-x86-rr2-lut16-p3-x4.c | 35 const v128_t vsat_cutoff = wasm_f32x4_splat(-0x1.154246p+4f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 36 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.800000p19f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 37 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 39 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 40 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 41 const v128_t vc3 = wasm_f32x4_splat(0x1.55561Cp-3f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 42 const v128_t vc2 = wasm_f32x4_splat(0x1.0001ECp-1f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() 43 const v128_t vone = wasm_f32x4_splat(1.0f); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | wasmsimd-p5-x4.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x4() [all …]
|
D | wasmsimd-p5-x8.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8() [all …]
|
D | wasmsimd-p5-x8-acc2.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x8_acc2() [all …]
|
D | wasmsimd-p5-x12.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12() [all …]
|
D | wasmsimd-p5-x12-acc2.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc2() [all …]
|
D | wasmsimd-p5-x12-acc3.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x12_acc3() [all …]
|
D | wasmsimd-p5-x16.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16() [all …]
|
D | wasmsimd-p5-x16-acc2.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc2() [all …]
|
D | wasmsimd-p5-x16-acc4.c | 27 const v128_t vmagic_bias = wasm_f32x4_splat(0x1.8000FEp23f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 29 const v128_t vdenorm_cutoff = wasm_f32x4_splat(-0x1.5D589Ep6f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 30 const v128_t vlog2e = wasm_f32x4_splat(0x1.715476p+0f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 32 const v128_t vminus_ln2_hi = wasm_f32x4_splat(-0x1.62E400p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 33 const v128_t vminus_ln2_lo = wasm_f32x4_splat(-0x1.7F7D1Cp-20f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 35 const v128_t vc1 = wasm_f32x4_splat(0x1.FFFFF6p-1f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 36 const v128_t vc2 = wasm_f32x4_splat(0x1.FFFDC6p-2f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 37 const v128_t vc3 = wasm_f32x4_splat(0x1.555A80p-3f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 38 const v128_t vc4 = wasm_f32x4_splat(0x1.573A1Ap-5f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() 39 const v128_t vc5 = wasm_f32x4_splat(0x1.0F9F9Cp-7f); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_p5_x16_acc4() [all …]
|