Home
last modified time | relevance | path

Searched refs:vs9 (Results 1 – 25 of 37) sorted by relevance

12

/external/XNNPACK/src/f32-velu/gen/
Dvelu-avx2-rr1-lut4-p4-perm-x80.c122 __m256 vs9 = _mm256_castsi256_ps(_mm256_add_epi32(vl9, ven9)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80() local
166 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
186 vs9 = _mm256_fmsub_ps(vs9, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
207 const __m256 ve9 = _mm256_fmadd_ps(vp9, valpha, vs9); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
Dvelu-avx2-rr1-lut16-p3-gather-x80.c134 __m256 vs9 = _mm256_castsi256_ps(_mm256_add_epi32(vl9, ven9)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() local
167 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
187 vs9 = _mm256_fmsub_ps(vs9, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
208 const __m256 ve9 = _mm256_fmadd_ps(vp9, valpha, vs9); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
Dvelu-avx2-rr1-lut8-p4-perm-x80.c122 __m256 vs9 = _mm256_castsi256_ps(_mm256_add_epi32(vl9, ven9)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80() local
166 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
186 vs9 = _mm256_fmsub_ps(vs9, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
207 const __m256 ve9 = _mm256_fmadd_ps(vp9, valpha, vs9); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
Dvelu-avx2-rr1-p6-x80.c92 __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80() local
169 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
189 vs9 = _mm256_fmsub_ps(vs9, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
210 const __m256 ve9 = _mm256_fmadd_ps(vp9, valpha, vs9); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/
Davx2-p5-x80.c94 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80() local
190 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80()
201 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80()
Davx2-p5-x88.c97 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88() local
201 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88()
213 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88()
Davx2-p5-x96.c100 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x96() local
212 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x96()
225 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x96()
/external/XNNPACK/src/f32-raddexpminusmax/gen/
Davx2-p5-x80.c92 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80() local
188 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80()
199 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80()
Davx2-p5-x80-acc2.c93 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2() local
189 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2()
200 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2()
Davx2-p5-x80-acc5.c96 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5() local
192 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5()
203 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5()
Davx2-p5-x96-acc2.c99 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2() local
211 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2()
224 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2()
Davx2-p5-x96.c98 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96() local
210 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96()
223 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96()
Davx2-p5-x96-acc3.c100 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc3() local
212 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc3()
225 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc3()
Davx2-p5-x96-acc6.c103 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc6() local
215 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc6()
228 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc6()
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/
Davx2-rr1-p5-x80-acc2.c84 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2() local
161 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2()
172 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2()
Davx2-rr1-p5-x80.c83 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80() local
160 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80()
171 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80()
Davx2-rr1-p5-x80-acc5.c87 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5() local
164 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5()
175 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5()
Davx2-rr1-p5-x96-acc3.c91 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3() local
182 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3()
195 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3()
Davx2-rr1-p5-x96-acc6.c94 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc6() local
185 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc6()
198 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc6()
Davx2-rr1-p5-x96-acc2.c90 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc2() local
181 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc2()
194 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc2()
Davx2-rr1-p5-x96.c89 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96() local
180 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96()
193 __m256 vf9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96()
/external/XNNPACK/src/f32-vsigmoid/gen/
Dvsigmoid-avx2-rr1-p5-div-x80.c82 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80() local
159 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80()
170 const __m256 ve9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x80()
Dvsigmoid-avx2-rr1-p5-nr2fma-x80.c82 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() local
159 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
170 const __m256 ve9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
Dvsigmoid-avx2-rr1-p5-nr1fma-x80.c82 const __m256 vs9 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn9), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80() local
159 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
170 const __m256 ve9 = _mm256_fmadd_ps(vt9, vp9, vs9); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
Dvsigmoid-avx-rr2-p5-div-x80.c103 const __m256 vs9 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs_lo9), vs_hi9, 1); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80() local
191 vt9 = _mm256_mul_ps(vt9, vs9); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80()
202 const __m256 ve9 = _mm256_add_ps(_mm256_mul_ps(vt9, vp9), vs9); in xnn_f32_vsigmoid_ukernel__avx_rr2_p5_div_x80()

12