/external/XNNPACK/src/math/ |
D | expm1minus-scalar-rr2-p5.c | 82 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_p5() local 84 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_p5()
|
D | expm1minus-scalar-rr2-lut16-p3.c | 94 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_lut16_p3() local 96 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_lut16_p3()
|
D | expm1minus-scalar-rr2-p6.c | 84 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_p6() local 86 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_p6()
|
D | expm1minus-scalar-rr2-lut8-p3.c | 94 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_lut8_p3() local 96 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_lut8_p3()
|
D | expm1minus-scalar-rr2-lut8-p4.c | 96 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_lut8_p4() local 98 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_lut8_p4()
|
D | expm1minus-scalar-rr2-lut4-p4.c | 96 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_lut4_p4() local 98 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_lut4_p4()
|
D | expm1minus-scalar-rr2-lut16-p4.c | 96 const float vsm1 = vs - vone; in xnn_math_f32_expm1minus__scalar_rr2_lut16_p4() local 98 const float vf = vp + vsm1; in xnn_math_f32_expm1minus__scalar_rr2_lut16_p4()
|
D | expm1minus-avx512f-rr1-lut16-p3-perm.c | 91 const __m512 vsm1 = _mm512_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx512f_rr1_lut16_p3_perm() local 93 const __m512 vf = _mm512_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx512f_rr1_lut16_p3_perm()
|
D | expm1minus-avx2-rr1-lut4-p4-perm.c | 92 const __m256 vsm1 = _mm256_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx2_rr1_lut4_p4_perm() local 94 const __m256 vf = _mm256_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx2_rr1_lut4_p4_perm()
|
D | expm1minus-avx512f-rr1-p6.c | 82 const __m512 vsm1 = _mm512_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx512f_rr1_p6() local 84 const __m512 vf = _mm512_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx512f_rr1_p6()
|
D | expm1minus-avx2-rr1-lut16-p3-gather.c | 92 const __m256 vsm1 = _mm256_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx2_rr1_lut16_p3_gather() local 94 const __m256 vf = _mm256_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx2_rr1_lut16_p3_gather()
|
D | expm1minus-avx2-rr1-lut8-p4-perm.c | 91 const __m256 vsm1 = _mm256_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx2_rr1_lut8_p4_perm() local 93 const __m256 vf = _mm256_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx2_rr1_lut8_p4_perm()
|
D | expm1minus-neonfma-rr1-p6.c | 80 const float32x4_t vsm1 = vsubq_f32(vs, vone); in xnn_math_f32_expm1minus__neonfma_rr1_p6() local 82 const float32x4_t vf = vaddq_f32(vp, vsm1); in xnn_math_f32_expm1minus__neonfma_rr1_p6()
|
D | expm1minus-avx2-rr1-p6.c | 82 const __m256 vsm1 = _mm256_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx2_rr1_p6() local 84 const __m256 vf = _mm256_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx2_rr1_p6()
|
D | expm1minus-neon-rr2-p6.c | 84 const float32x4_t vsm1 = vsubq_f32(vs, vone); in xnn_math_f32_expm1minus__neon_rr2_p6() local 86 const float32x4_t vf = vaddq_f32(vp, vsm1); in xnn_math_f32_expm1minus__neon_rr2_p6()
|
D | expm1minus-sse2-rr2-p6.c | 85 const __m128 vsm1 = _mm_sub_ps(vs, vone); in xnn_math_f32_expm1minus__sse2_rr2_p6() local 87 const __m128 vf = _mm_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__sse2_rr2_p6()
|
D | expm1minus-wasmsimd-rr2-p6-max.c | 84 const v128_t vsm1 = wasm_f32x4_sub(vs, vone); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max() local 86 const v128_t vf = wasm_f32x4_add(vp, vsm1); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_max()
|
D | expm1minus-wasmsimd-rr2-p6-andnot.c | 85 const v128_t vsm1 = wasm_f32x4_sub(vs, vone); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot() local 87 const v128_t vf = wasm_f32x4_add(vp, vsm1); in xnn_math_f32_expm1minus__wasmsimd_rr2_p6_andnot()
|
D | expm1minus-avx-rr2-p6.c | 87 const __m256 vsm1 = _mm256_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx_rr2_p6() local 89 const __m256 vf = _mm256_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx_rr2_p6()
|
D | expm1minus-avx-rr2-lut4-p4-perm.c | 99 const __m256 vsm1 = _mm256_sub_ps(vs, vone); in xnn_math_f32_expm1minus__avx_rr2_lut4_p4_perm() local 101 const __m256 vf = _mm256_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__avx_rr2_lut4_p4_perm()
|
D | expm1minus-wasmsimd-rr2-lut16-p3-max.c | 99 const v128_t vsm1 = wasm_f32x4_sub(vs, vone); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max() local 101 const v128_t vf = wasm_f32x4_add(vp, vsm1); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_max()
|
D | expm1minus-wasmsimd-rr2-lut16-p3-andnot.c | 100 const v128_t vsm1 = wasm_f32x4_sub(vs, vone); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot() local 102 const v128_t vf = wasm_f32x4_add(vp, vsm1); in xnn_math_f32_expm1minus__wasmsimd_rr2_lut16_p3_andnot()
|
D | expm1minus-neonfma-rr1-lut16-p3.c | 96 const float32x4_t vsm1 = vsubq_f32(vs, vone); in xnn_math_f32_expm1minus__neonfma_rr1_lut16_p3() local 98 const float32x4_t vf = vaddq_f32(vp, vsm1); in xnn_math_f32_expm1minus__neonfma_rr1_lut16_p3()
|
D | expm1minus-neon-rr2-lut16-p3.c | 100 const float32x4_t vsm1 = vsubq_f32(vs, vone); in xnn_math_f32_expm1minus__neon_rr2_lut16_p3() local 102 const float32x4_t vf = vaddq_f32(vp, vsm1); in xnn_math_f32_expm1minus__neon_rr2_lut16_p3()
|
D | expm1minus-sse2-rr2-lut16-p3.c | 111 const __m128 vsm1 = _mm_sub_ps(vs, vone); in xnn_math_f32_expm1minus__sse2_rr2_lut16_p3() local 113 const __m128 vf = _mm_add_ps(vp, vsm1); in xnn_math_f32_expm1minus__sse2_rr2_lut16_p3()
|