/external/XNNPACK/src/f32-velu/gen/ |
D | velu-avx2-rr1-p6-x72.c | 87 __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() local 157 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() 175 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72() 194 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x72()
|
D | velu-avx2-rr1-lut4-p4-perm-x72.c | 114 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72() local 154 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72() 172 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72() 191 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x72()
|
D | velu-avx2-rr1-lut8-p4-perm-x72.c | 114 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72() local 154 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72() 172 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72() 191 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x72()
|
D | velu-avx2-rr1-lut16-p3-gather-x72.c | 125 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72() local 155 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72() 173 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72() 192 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72()
|
D | velu-avx2-rr1-lut4-p4-perm-x80.c | 120 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80() local 164 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80() 184 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80() 205 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x80()
|
D | velu-avx2-rr1-lut16-p3-gather-x80.c | 132 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() local 165 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() 185 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() 206 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80()
|
D | velu-avx2-rr1-lut8-p4-perm-x80.c | 120 __m256 vs8 = _mm256_castsi256_ps(_mm256_add_epi32(vl8, ven8)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80() local 164 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80() 184 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80() 205 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x80()
|
D | velu-avx2-rr1-p6-x80.c | 90 __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80() local 167 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80() 187 vs8 = _mm256_fmsub_ps(vs8, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80() 208 const __m256 ve8 = _mm256_fmadd_ps(vp8, valpha, vs8); in xnn_f32_velu_ukernel__avx2_rr1_p6_x80()
|
/external/libaom/libaom/build/ |
D | .gitattributes | 1 *-vs8/*.rules -crlf
|
/external/XNNPACK/src/f32-raddexpminusmax/gen/ |
D | avx2-p5-x72.c | 88 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72() local 176 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72() 186 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72()
|
D | avx2-p5-x72-acc3.c | 90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3() local 178 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3() 188 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x72_acc3()
|
D | avx2-p5-x80.c | 91 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80() local 187 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80() 198 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80()
|
D | avx2-p5-x80-acc2.c | 92 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2() local 188 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2() 199 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc2()
|
D | avx2-p5-x80-acc5.c | 95 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5() local 191 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5() 202 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x80_acc5()
|
D | avx2-p5-x96-acc2.c | 98 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2() local 210 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2() 223 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddexpminusmax_ukernel__avx2_p5_x96_acc2()
|
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
D | avx2-p5-x72.c | 90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72() local 178 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72() 188 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x72()
|
D | avx2-p5-x80.c | 93 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80() local 189 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80() 200 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x80()
|
D | avx2-p5-x88.c | 96 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88() local 200 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88() 212 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x88()
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | avx2-rr1-p5-x72.c | 79 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72() local 149 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72() 159 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72()
|
D | avx2-rr1-p5-x72-acc3.c | 81 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3() local 151 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3() 161 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x72_acc3()
|
D | avx2-rr1-p5-x80-acc2.c | 83 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2() local 160 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2() 171 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc2()
|
D | avx2-rr1-p5-x80.c | 82 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80() local 159 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80() 170 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80()
|
D | avx2-rr1-p5-x80-acc5.c | 86 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5() local 163 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5() 174 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x80_acc5()
|
D | avx2-rr1-p5-x96-acc3.c | 90 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3() local 181 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3() 194 __m256 vf8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_raddstoreexpminusmax_ukernel__avx2_rr1_p5_x96_acc3()
|
/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-avx2-rr1-p5-div-x72.c | 78 const __m256 vs8 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn8), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72() local 148 vt8 = _mm256_mul_ps(vt8, vs8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72() 158 const __m256 ve8 = _mm256_fmadd_ps(vt8, vp8, vs8); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x72()
|