/external/XNNPACK/src/f32-velu/gen/ |
D | velu-wasm-rr2-p6-x5.c | 54 …const float vz4 = __builtin_wasm_min_f32(__builtin_wasm_max_f32(vx4 * vprescale, vsat_cutoff), 0.0… in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() local
|
D | velu-scalar-rr2-p6-x5.c | 54 const float vz4 = vx4 * vprescale; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local
|
D | velu-wasm-rr2-p6-x6.c | 55 …const float vz4 = __builtin_wasm_min_f32(__builtin_wasm_max_f32(vx4 * vprescale, vsat_cutoff), 0.0… in xnn_f32_velu_ukernel__wasm_rr2_p6_x6() local
|
D | velu-scalar-rr2-p6-x6.c | 55 const float vz4 = vx4 * vprescale; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
|
D | velu-avx2-rr1-p6-x40.c | 51 const __m256 vz4 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx4, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x40() local
|
D | velu-avx512f-rr1-p6-x80.c | 53 const __m512 vz4 = _mm512_max_ps(vsat_cutoff, _mm512_mul_ps(vx4, vprescale)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x80() local
|
D | velu-wasm-rr2-lut16-p3-x5.c | 54 …const float vz4 = __builtin_wasm_min_f32(__builtin_wasm_max_f32(vx4 * vprescale, vsat_cutoff), 0.0… in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x5() local
|
D | velu-scalar-rr2-lut16-p3-x5.c | 54 const float vz4 = vx4 * vprescale; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5() local
|
D | velu-avx2-rr1-p6-x48.c | 52 const __m256 vz4 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx4, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48() local
|
D | velu-avx512f-rr1-p6-x96.c | 54 const __m512 vz4 = _mm512_max_ps(vsat_cutoff, _mm512_mul_ps(vx4, vprescale)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96() local
|
D | velu-wasm-rr2-lut16-p3-x6.c | 55 …const float vz4 = __builtin_wasm_min_f32(__builtin_wasm_max_f32(vx4 * vprescale, vsat_cutoff), 0.0… in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x6() local
|
D | velu-scalar-rr2-lut16-p3-x6.c | 55 const float vz4 = vx4 * vprescale; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6() local
|
D | velu-avx512f-rr1-lut16-p3-perm-x80.c | 51 const __m512 vz4 = _mm512_max_ps(vsat_cutoff, _mm512_mul_ps(vx4, vprescale)); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x80() local
|
D | velu-avx2-rr1-lut16-p3-gather-x40.c | 51 const __m256 vz4 = _mm256_max_ps(vsat_cutoff, _mm256_mul_ps(vx4, vprescale)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x40() local
|
/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-avx512f-rr1-p5-scalef-div-x80.c | 49 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-div-x96.c | 50 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() local
|
D | vsigmoid-avx512f-rr2-lut32-p2-perm2-scalef-div-x80.c | 50 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-nr1fma-x80.c | 49 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x80() local
|
D | vsigmoid-avx512f-rr1-lut16-p3-perm-scalef-div-x80.c | 48 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() local
|
D | vsigmoid-avx2-rr1-p5-div-x40.c | 50 const __m256 vz4 = _mm256_or_ps(vx4, vsign_mask); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x40() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-nr1fma-x96.c | 50 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96() local
|
D | vsigmoid-avx512f-rr1-lut16-p3-perm-scalef-div-x96.c | 49 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() local
|
D | vsigmoid-avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x80.c | 48 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x80() local
|
D | vsigmoid-avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x80.c | 50 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x80() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-div-x112.c | 51 const __m512 vz4 = _mm512_castsi512_ps(_mm512_or_epi32(_mm512_castps_si512(vx4), vsign_mask)); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() local
|