Home
last modified time | relevance | path

Searched refs:ve6 (Results 1 – 25 of 60) sorted by relevance

123

/external/XNNPACK/src/f32-vscaleextexp/gen/
Davx2-p5-x56.c147 __m256 ve6 = _mm256_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x56() local
159 ve6 = _mm256_max_ps(ve6, vmin_exponent); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x56()
170 …const __m256 vs6 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve6, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x56()
Davx2-p5-x64.c157 __m256 ve6 = _mm256_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x64() local
170 ve6 = _mm256_max_ps(ve6, vmin_exponent); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x64()
182 …const __m256 vs6 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve6, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x64()
Davx2-p5-x72.c167 __m256 ve6 = _mm256_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x72() local
181 ve6 = _mm256_max_ps(ve6, vmin_exponent); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x72()
194 …const __m256 vs6 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve6, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x72()
Davx2-p5-x80.c177 __m256 ve6 = _mm256_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x80() local
192 ve6 = _mm256_max_ps(ve6, vmin_exponent); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x80()
206 …const __m256 vs6 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve6, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x80()
Davx2-p5-x88.c187 __m256 ve6 = _mm256_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x88() local
203 ve6 = _mm256_max_ps(ve6, vmin_exponent); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x88()
218 …const __m256 vs6 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve6, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x88()
Davx2-p5-x96.c197 __m256 ve6 = _mm256_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x96() local
214 ve6 = _mm256_max_ps(ve6, vmin_exponent); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x96()
230 …const __m256 vs6 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve6, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x96()
Davx512f-p5-scalef-x112.c141 const __m512 ve6 = _mm512_add_ps(vn6, vscalee); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112() local
150 vf6 = _mm512_scalef_ps(vf6, ve6); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112()
/external/XNNPACK/src/f32-sigmoid/gen/
Davx512f-rr1-lut16-p3-perm-scalef-div-x112.c128 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() local
136 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112()
144 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112()
Davx512f-rr1-p5-scalef-div-x112.c125 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() local
133 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
141 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
Davx512f-rr2-lut32-p2-perm2-scalef-div-x112.c134 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() local
142 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112()
150 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112()
Davx512f-rr1-p5-scalef-div-x128.c135 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() local
144 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128()
153 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128()
Davx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x112.c134 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x112() local
142 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x112()
166 __m512 vf6 = _mm512_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x112()
Davx512f-rr1-lut16-p3-perm-scalef-nr1fma-x112.c128 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x112() local
136 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x112()
160 __m512 vf6 = _mm512_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x112()
Davx512f-rr1-p5-scalef-nr1fma-x112.c125 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x112() local
133 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x112()
157 __m512 vf6 = _mm512_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x112()
Davx512f-rr2-lut32-p2-perm2-scalef-div-x128.c144 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() local
153 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128()
162 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128()
Davx512f-rr1-lut16-p3-perm-scalef-div-x128.c138 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() local
147 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128()
156 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128()
Davx2-rr1-p5-div-x56.c136 const __m256 ve6 = _mm256_fmadd_ps(vt6, vp6, vs6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x56() local
144 const __m256 vd6 = _mm256_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x56()
152 __m256 vf6 = _mm256_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x56()
Davx512f-rr1-lut16-p3-perm-scalef-nr1fma-x128.c138 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128() local
147 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128()
174 __m512 vf6 = _mm512_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128()
Davx2-rr1-p5-div-x64.c147 const __m256 ve6 = _mm256_fmadd_ps(vt6, vp6, vs6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x64() local
156 const __m256 vd6 = _mm256_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x64()
165 __m256 vf6 = _mm256_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x64()
Davx512f-rr1-p5-scalef-nr1fma-x128.c135 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128() local
144 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128()
171 __m512 vf6 = _mm512_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128()
Davx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x128.c144 const __m512 ve6 = _mm512_scalef_ps(vp6, vn6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128() local
153 const __m512 vd6 = _mm512_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128()
180 __m512 vf6 = _mm512_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128()
Davx2-rr1-p5-nr1fma-x56.c136 const __m256 ve6 = _mm256_fmadd_ps(vt6, vp6, vs6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x56() local
144 const __m256 vd6 = _mm256_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x56()
169 __m256 vf6 = _mm256_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x56()
Davx-rr2-p5-div-x56.c159 const __m256 ve6 = _mm256_add_ps(_mm256_mul_ps(vt6, vp6), vs6); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_div_x56() local
167 const __m256 vd6 = _mm256_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_div_x56()
175 __m256 vf6 = _mm256_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_div_x56()
Davx2-rr1-p5-nr2fma-x56.c136 const __m256 ve6 = _mm256_fmadd_ps(vt6, vp6, vs6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56() local
144 const __m256 vd6 = _mm256_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56()
176 __m256 vf6 = _mm256_mul_ps(ve6, vr6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56()
Davx2-rr1-p5-div-x72.c158 const __m256 ve6 = _mm256_fmadd_ps(vt6, vp6, vs6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x72() local
168 const __m256 vd6 = _mm256_add_ps(ve6, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x72()
178 __m256 vf6 = _mm256_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x72()

123