Home
last modified time | relevance | path

Searched refs:vs8 (Results 1 – 25 of 54) sorted by relevance

123

/external/XNNPACK/src/f32-velu/gen/
Dvelu-avx2-rr1-p6-x72.c87 __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() local
157 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72()
175 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72()
194 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72()
Dvelu-avx2-rr1-lut4-p4-perm-x72.c114 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72() local
154 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72()
172 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72()
191 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72()
Dvelu-avx2-rr1-lut8-p4-perm-x72.c114 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72() local
154 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72()
172 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72()
191 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72()
Dvelu-avx2-rr1-lut16-p3-gather-x72.c125 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72() local
155 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72()
173 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72()
192 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72()
Dvelu-avx2-rr1-lut4-p4-perm-x80.c120 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80() local
164 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
184 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
205 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
Dvelu-avx2-rr1-lut16-p3-gather-x80.c132 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() local
165 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
185 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
206 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
Dvelu-avx2-rr1-lut8-p4-perm-x80.c120 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80() local
164 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
184 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
205 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
Dvelu-avx2-rr1-p6-x80.c90 __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80() local
167 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
187 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
208 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
/external/libaom/libaom/build/
D.gitattributes1 *-vs8/*.rules -crlf
/external/XNNPACK/src/f32-raddexpminusmax/gen/
Davx2-p5-x72.c88 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72() local
176 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72()
186 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72()
Davx2-p5-x72-acc3.c90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3() local
178 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3()
188 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3()
Davx2-p5-x80.c91 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80() local
187 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80()
198 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80()
Davx2-p5-x80-acc2.c92 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2() local
188 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2()
199 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2()
Davx2-p5-x80-acc5.c95 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5() local
191 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5()
202 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5()
Davx2-p5-x96-acc2.c98 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2() local
210 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2()
223 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2()
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/
Davx2-p5-x72.c90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72() local
178 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72()
188 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72()
Davx2-p5-x80.c93 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80() local
189 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80()
200 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80()
Davx2-p5-x88.c96 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88() local
200 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88()
212 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88()
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/
Davx2-rr1-p5-x72.c79 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72() local
149 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72()
159 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72()
Davx2-rr1-p5-x72-acc3.c81 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3() local
151 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3()
161 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3()
Davx2-rr1-p5-x80-acc2.c83 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2() local
160 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2()
171 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2()
Davx2-rr1-p5-x80.c82 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80() local
159 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80()
170 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80()
Davx2-rr1-p5-x80-acc5.c86 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5() local
163 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5()
174 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5()
Davx2-rr1-p5-x96-acc3.c90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3() local
181 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3()
194 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3()
/external/XNNPACK/src/f32-vsigmoid/gen/
Dvsigmoid-avx2-rr1-p5-div-x72.c78 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72() local
148 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72()
158 const __m256 ve8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72()

123