/external/XNNPACK/src/f32-sigmoid/gen/ |
D | avx2-rr1-p5-nr2fma-x48.c | 138 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() local 145 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() 152 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() 159 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48()
|
D | avx2-rr1-p5-nr2fma-x56.c | 151 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56() local 159 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56() 167 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56() 175 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x56()
|
D | avx2-rr1-p5-nr2fma-x64.c | 164 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64() local 173 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64() 182 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64() 191 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64()
|
D | avx2-rr1-p5-nr2fma-x72.c | 177 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() local 187 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() 197 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() 207 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72()
|
D | avx2-rr1-p5-nr2fma-x80.c | 190 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() local 201 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() 212 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() 223 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
|
D | avx512f-rr1-p5-scalef-nr1fma-x96.c | 128 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96() local 135 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96() 142 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x96.c | 137 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x96() local 144 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x96() 151 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x96()
|
D | avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x96.c | 131 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x96() local 138 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x96() 145 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x96()
|
D | avx-rr2-p5-nr2-x48.c | 159 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x48() local 171 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x48() 172 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x48() 179 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x48()
|
D | avx2-rr1-p5-nr1fma-x48.c | 138 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x48() local 145 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x48() 153 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x48()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x112.c | 149 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x112() local 157 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x112() 165 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x112()
|
D | avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x112.c | 143 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x112() local 151 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x112() 159 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x112()
|
D | avx512f-rr1-p5-scalef-nr1fma-x112.c | 140 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x112() local 148 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x112() 156 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x112()
|
D | avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x128.c | 155 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128() local 164 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128() 173 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128()
|
D | avx512f-rr1-p5-scalef-nr1fma-x128.c | 152 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128() local 161 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128() 170 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128()
|
D | avx-rr2-p5-nr2-x56.c | 175 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x56() local 188 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x56() 189 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x56() 198 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x56()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x128.c | 161 __m512 vr5 = _mm512_rcp14_ps(vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128() local 170 vr5 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128() 179 __m512 vf5 = _mm512_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128()
|
D | avx2-rr1-p5-nr1fma-x56.c | 151 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x56() local 159 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x56() 168 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x56()
|
D | avx-rr2-p5-nr2-x64.c | 191 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64() local 205 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64() 206 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64() 217 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64()
|
D | avx2-rr1-p5-nr1fma-x64.c | 164 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x64() local 173 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x64() 183 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x64()
|
D | avx-rr2-p5-nr2-x72.c | 207 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72() local 222 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72() 223 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72() 236 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72()
|
D | avx2-rr1-p5-nr1fma-x72.c | 177 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x72() local 187 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x72() 198 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x72()
|
D | avx2-rr1-p5-nr1fma-x80.c | 190 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x80() local 201 vr5 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr5, vd5, vone), vr5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x80() 213 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
|
D | avx-rr2-p5-nr2-x80.c | 223 __m256 vr5 = _mm256_rcp_ps(vd5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80() local 239 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80() 240 vr5 = _mm256_mul_ps(vr5, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr5, vd5))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80() 255 __m256 vf5 = _mm256_mul_ps(ve5, vr5); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80()
|
/external/llvm-project/lldb/source/Plugins/Process/Utility/ |
D | RegisterInfos_ppc64.h | 146 DEFINE_VMX_PPC64(vr5, LLDB_INVALID_REGNUM), \ 283 uint32_t vr5[4]; member
|