Home
last modified time | relevance | path

Searched refs:vz8 (Results 1 – 19 of 19) sorted by relevance

/external/XNNPACK/src/f32-vsigmoid/gen/
Dvsigmoid-avx2-rr1-p5-div-x72.c58 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72() local
68 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72()
98 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72()
188 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72()
Dvsigmoid-avx2-rr1-p5-nr1fma-x72.c58 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x72() local
68 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x72()
98 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x72()
209 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x72()
Dvsigmoid-avx2-rr1-p5-div-x80.c59 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80() local
70 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80()
103 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80()
202 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80()
Dvsigmoid-avx2-rr1-p5-nr2fma-x80.c59 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() local
70 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
103 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
235 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
Dvsigmoid-avx2-rr1-p5-nr1fma-x80.c59 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80() local
70 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
103 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
225 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
Dvsigmoid-avx-rr2-p5-div-x72.c59 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x72() local
69 __m256 vn8 = _mm256_add_ps(_mm256_mul_ps(vz8, vlog2e), vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x72()
117 __m256 vt8 = _mm256_add_ps(_mm256_mul_ps(vn8, vminus_ln2_hi), vz8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x72()
217 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x72()
Dvsigmoid-avx2-rr1-p5-nr2fma-x72.c58 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() local
68 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x72()
98 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x72()
218 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x72()
Dvsigmoid-avx-rr2-p5-div-x80.c60 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80() local
71 __m256 vn8 = _mm256_add_ps(_mm256_mul_ps(vz8, vlog2e), vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80()
124 __m256 vt8 = _mm256_add_ps(_mm256_mul_ps(vn8, vminus_ln2_hi), vz8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80()
234 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80()
Dvsigmoid-avx-rr2-p5-nr2-x72.c60 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x72() local
70 __m256 vn8 = _mm256_add_ps(_mm256_mul_ps(vz8, vlog2e), vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x72()
118 __m256 vt8 = _mm256_add_ps(_mm256_mul_ps(vn8, vminus_ln2_hi), vz8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x72()
247 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x72()
Dvsigmoid-avx-rr2-p5-nr2-x80.c61 const __m256 vz8 = _mm256_or_ps(vx8, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x80() local
72 __m256 vn8 = _mm256_add_ps(_mm256_mul_ps(vz8, vlog2e), vmagic_bias); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x80()
125 __m256 vt8 = _mm256_add_ps(_mm256_mul_ps(vn8, vminus_ln2_hi), vz8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x80()
267 vf8 = _mm256_andnot_ps(_mm256_cmp_ps(vz8, vdenorm_cutoff, _CMP_LT_OS), vf8); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_nr2_x80()
/external/XNNPACK/src/f32-velu/gen/
Dvelu-avx2-rr1-p6-x72.c59 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() local
69 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72()
98 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72()
Dvelu-avx2-rr1-lut16-p3-gather-x72.c59 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72() local
69 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72()
126 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72()
Dvelu-avx2-rr1-lut4-p4-perm-x72.c58 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72() local
68 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72()
115 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72()
Dvelu-avx2-rr1-lut8-p4-perm-x72.c58 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72() local
68 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72()
115 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72()
Dvelu-avx2-rr1-p6-x80.c60 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80() local
71 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
103 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
Dvelu-avx2-rr1-lut16-p3-gather-x80.c60 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() local
71 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
133 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
Dvelu-avx2-rr1-lut4-p4-perm-x80.c59 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80() local
70 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
121 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
Dvelu-avx2-rr1-lut8-p4-perm-x80.c59 const __m256 vz8 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx8, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80() local
70 __m256 vn8 = _mm256_fmadd_ps(vz8, vlog2e, vmagic_bias); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
121 __m256 vt8 = _mm256_fmadd_ps(vn8, vminus_ln2, vz8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
/external/hyphenation-patterns/pl/
Dhyph-pl.pat.txt2311 .vz8