/external/XNNPACK/src/f32-sigmoid/gen/ |
D | psimd-p5-div-x16.c | 59 const psimd_f32 vzCDEF = psimd_abs_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x16() local 70 psimd_f32 vnCDEF = psimd_qfma_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x16() 90 psimd_f32 vtCDEF = psimd_qfma_f32(vzCDEF, vnCDEF, vln2_hi); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x16() 144 vfCDEF = psimd_andnotmask_f32(vzCDEF > vdenorm_cutoff, vfCDEF); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x16()
|
D | psimd-p5-div-x20.c | 60 const psimd_f32 vzCDEF = psimd_abs_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x20() local 72 psimd_f32 vnCDEF = psimd_qfma_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x20() 95 psimd_f32 vtCDEF = psimd_qfma_f32(vzCDEF, vnCDEF, vln2_hi); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x20() 158 vfCDEF = psimd_andnotmask_f32(vzCDEF > vdenorm_cutoff, vfCDEF); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x20()
|
D | sse41-p5-div-x16.c | 59 const __m128 vzCDEF = _mm_or_ps(vxCDEF, vsign_mask); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() local 70 __m128 vnCDEF = _mm_add_ps(_mm_mul_ps(vzCDEF, vlog2e), vmagic_bias); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 90 __m128 vtCDEF = _mm_add_ps(_mm_mul_ps(vnCDEF, vminus_ln2_hi), vzCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 149 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16()
|
D | psimd-p5-div-x24.c | 61 const psimd_f32 vzCDEF = psimd_abs_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x24() local 74 psimd_f32 vnCDEF = psimd_qfma_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x24() 100 psimd_f32 vtCDEF = psimd_qfma_f32(vzCDEF, vnCDEF, vln2_hi); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x24() 172 vfCDEF = psimd_andnotmask_f32(vzCDEF > vdenorm_cutoff, vfCDEF); in xnn_f32_sigmoid_ukernel__psimd_p5_div_x24()
|
D | sse2-p5-div-x16.c | 59 const __m128 vzCDEF = _mm_or_ps(vxCDEF, vsign_mask); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() local 70 __m128 vnCDEF = _mm_add_ps(_mm_mul_ps(vzCDEF, vlog2e), vmagic_bias); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 90 __m128 vtCDEF = _mm_add_ps(_mm_mul_ps(vnCDEF, vminus_ln2_hi), vzCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 149 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
|
D | sse41-p5-div-x20.c | 60 const __m128 vzCDEF = _mm_or_ps(vxCDEF, vsign_mask); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() local 72 __m128 vnCDEF = _mm_add_ps(_mm_mul_ps(vzCDEF, vlog2e), vmagic_bias); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 95 __m128 vtCDEF = _mm_add_ps(_mm_mul_ps(vnCDEF, vminus_ln2_hi), vzCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 164 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20()
|
D | sse2-p5-div-x20.c | 60 const __m128 vzCDEF = _mm_or_ps(vxCDEF, vsign_mask); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() local 72 __m128 vnCDEF = _mm_add_ps(_mm_mul_ps(vzCDEF, vlog2e), vmagic_bias); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 95 __m128 vtCDEF = _mm_add_ps(_mm_mul_ps(vnCDEF, vminus_ln2_hi), vzCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 164 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
|
D | sse41-p5-div-x24.c | 61 const __m128 vzCDEF = _mm_or_ps(vxCDEF, vsign_mask); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() local 74 __m128 vnCDEF = _mm_add_ps(_mm_mul_ps(vzCDEF, vlog2e), vmagic_bias); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 100 __m128 vtCDEF = _mm_add_ps(_mm_mul_ps(vnCDEF, vminus_ln2_hi), vzCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 179 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24()
|
D | neonfma-rr1-p5-div-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() local 67 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() 86 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16()
|
D | sse2-p5-div-x24.c | 61 const __m128 vzCDEF = _mm_or_ps(vxCDEF, vsign_mask); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() local 74 __m128 vnCDEF = _mm_add_ps(_mm_mul_ps(vzCDEF, vlog2e), vmagic_bias); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 100 __m128 vtCDEF = _mm_add_ps(_mm_mul_ps(vnCDEF, vminus_ln2_hi), vzCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 179 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
|
D | neonfma-rr1-p5-div-x20.c | 57 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() local 69 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 91 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20()
|
D | neonfma-rr1-p5-nr2recps-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x16() local 67 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x16() 86 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x16()
|
D | neon-rr2-p5-nr2recps-x16.c | 58 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_p5_nr2recps_x16() local 69 float32x4_t vnCDEF = vmlaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neon_rr2_p5_nr2recps_x16() 89 float32x4_t vtCDEF = vmlaq_f32(vzCDEF, vnCDEF, vln2_hi); in xnn_f32_sigmoid_ukernel__neon_rr2_p5_nr2recps_x16()
|
D | neonfma-rr1-p5-nr1recps1fma-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x16() local 67 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x16() 86 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x16()
|
D | neonfma-rr1-p5-nr2fma-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2fma_x16() local 67 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2fma_x16() 86 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2fma_x16()
|
D | neonfma-rr1-p5-nr1recps1fma-x20.c | 57 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x20() local 69 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x20() 91 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x20()
|
D | neonfma-rr1-p5-nr2recps-x20.c | 57 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x20() local 69 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x20() 91 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x20()
|
D | neonfma-rr1-lut2048-p1-div-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() local 68 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e_x2048); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() 137 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2_o2048); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16()
|
D | neonfma-rr1-p5-div-x24.c | 58 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() local 71 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 96 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24()
|
D | neonfma-rr1-lut64-p2-div-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() local 68 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e_x64); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() 137 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2_o64); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16()
|
D | neon-rr2-p5-nr2recps-x20.c | 59 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_p5_nr2recps_x20() local 71 float32x4_t vnCDEF = vmlaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neon_rr2_p5_nr2recps_x20() 94 float32x4_t vtCDEF = vmlaq_f32(vzCDEF, vnCDEF, vln2_hi); in xnn_f32_sigmoid_ukernel__neon_rr2_p5_nr2recps_x20()
|
D | neonfma-rr1-p5-nr2fma-x20.c | 57 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2fma_x20() local 69 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2fma_x20() 91 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2fma_x20()
|
D | neonfma-rr1-lut2048-p1-nr2recps-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x16() local 68 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e_x2048); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x16() 137 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2_o2048); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x16()
|
D | neonfma-rr1-lut2048-p1-nr1recps1fma-x16.c | 56 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x16() local 68 float32x4_t vnCDEF = vfmaq_f32(vmagic_bias, vzCDEF, vminus_log2e_x2048); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x16() 137 float32x4_t vtCDEF = vfmaq_f32(vzCDEF, vnCDEF, vln2_o2048); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x16()
|
D | neon-rr2-lut64-p2-nr2recps-x16.c | 58 const float32x4_t vzCDEF = vabsq_f32(vxCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() local 70 float32x4_t vnCDEF = vmlaq_f32(vmagic_bias, vzCDEF, vminus_log2e_x64); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 140 float32x4_t vtCDEF = vmlaq_f32(vzCDEF, vnCDEF, vln2_o64_hi); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16()
|