/external/XNNPACK/src/f32-vscaleextexp/gen/ |
D | avx512f-p5-scalef-x80.c | 119 const __m512 ve4 = _mm512_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x80() local
|
D | avx512f-p5-scalef-x96.c | 129 const __m512 ve4 = _mm512_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96() local
|
D | avx2-p5-x40.c | 125 __m256 ve4 = _mm256_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x40() local
|
D | avx512f-p5-scalef-x112.c | 139 const __m512 ve4 = _mm512_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112() local
|
D | avx512f-p5-scalef-x128.c | 149 const __m512 ve4 = _mm512_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128() local
|
D | avx2-p5-x48.c | 135 __m256 ve4 = _mm256_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48() local
|
D | avx512f-p5-scalef-x144.c | 159 const __m512 ve4 = _mm512_add_ps(vn4, vscalee); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144() local
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-wasm-rr2-p6-x5.c | 141 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() local
|
D | velu-scalar-rr2-p6-x5.c | 161 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local
|
D | velu-wasm-rr2-p6-x6.c | 156 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6() local
|
D | velu-scalar-rr2-p6-x6.c | 180 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
|
D | velu-avx2-rr1-p6-x40.c | 130 const __m256 ve4 = _mm256_fmadd_ps(vp4, valpha, vs4); in xnn_f32_velu_ukernel__avx2_rr1_p6_x40() local
|
D | velu-wasm-rr2-lut16-p3-x5.c | 132 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x5() local
|
D | velu-scalar-rr2-lut16-p3-x5.c | 152 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5() local
|
D | velu-avx2-rr1-p6-x48.c | 144 const __m256 ve4 = _mm256_fmadd_ps(vp4, valpha, vs4); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48() local
|
D | velu-wasm-rr2-lut16-p3-x6.c | 146 const float ve4 = (vp4 + vs4) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x6() local
|
/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-avx512f-rr1-p5-scalef-div-x80.c | 103 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-div-x96.c | 113 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() local
|
D | vsigmoid-avx512f-rr2-lut32-p2-perm2-scalef-div-x80.c | 104 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-nr1fma-x80.c | 103 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x80() local
|
D | vsigmoid-avx512f-rr1-lut16-p3-perm-scalef-div-x80.c | 102 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() local
|
D | vsigmoid-avx2-rr1-p5-div-x40.c | 110 const __m256 ve4 = _mm256_fmadd_ps(vt4, vp4, vs4); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x40() local
|
D | vsigmoid-avx512f-rr1-p5-scalef-nr1fma-x96.c | 113 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96() local
|
D | vsigmoid-avx512f-rr1-lut16-p3-perm-scalef-div-x96.c | 112 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() local
|
D | vsigmoid-avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x80.c | 102 const __m512 ve4 = _mm512_scalef_ps(vp4, vn4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x80() local
|