/external/XNNPACK/src/f32-sigmoid/gen/ |
D | avx2-rr1-p5-nr2fma-x64.c | 166 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64() local 175 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64() 184 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64() 193 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x64()
|
D | avx2-rr1-p5-nr2fma-x72.c | 179 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() local 189 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() 199 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72() 209 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x72()
|
D | avx2-rr1-p5-nr2fma-x80.c | 192 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() local 203 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() 214 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80() 225 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x80()
|
D | avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x128.c | 157 __m512 vr7 = _mm512_rcp14_ps(vd7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128() local 166 vr7 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128() 175 __m512 vf7 = _mm512_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x128()
|
D | avx512f-rr1-p5-scalef-nr1fma-x128.c | 154 __m512 vr7 = _mm512_rcp14_ps(vd7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128() local 163 vr7 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128() 172 __m512 vf7 = _mm512_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x128()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x128.c | 163 __m512 vr7 = _mm512_rcp14_ps(vd7); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128() local 172 vr7 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128() 181 __m512 vf7 = _mm512_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x128()
|
D | avx-rr2-p5-nr2-x64.c | 193 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64() local 209 vr7 = _mm256_mul_ps(vr7, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr7, vd7))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64() 210 vr7 = _mm256_mul_ps(vr7, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr7, vd7))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64() 219 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x64()
|
D | avx2-rr1-p5-nr1fma-x64.c | 166 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x64() local 175 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x64() 185 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x64()
|
D | avx-rr2-p5-nr2-x72.c | 209 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72() local 226 vr7 = _mm256_mul_ps(vr7, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr7, vd7))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72() 227 vr7 = _mm256_mul_ps(vr7, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr7, vd7))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72() 238 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x72()
|
D | avx2-rr1-p5-nr1fma-x72.c | 179 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x72() local 189 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x72() 200 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x72()
|
D | avx2-rr1-p5-nr1fma-x80.c | 192 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x80() local 203 vr7 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr7, vd7, vone), vr7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x80() 215 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x80()
|
D | avx-rr2-p5-nr2-x80.c | 225 __m256 vr7 = _mm256_rcp_ps(vd7); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80() local 243 vr7 = _mm256_mul_ps(vr7, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr7, vd7))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80() 244 vr7 = _mm256_mul_ps(vr7, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr7, vd7))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80() 257 __m256 vf7 = _mm256_mul_ps(ve7, vr7); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x80()
|
/external/llvm-project/lldb/source/Plugins/Process/Utility/ |
D | RegisterInfos_ppc64.h | 148 DEFINE_VMX_PPC64(vr7, LLDB_INVALID_REGNUM), \ 285 uint32_t vr7[4]; member
|
D | RegisterInfos_ppc64le.h | 159 DEFINE_VMX(vr7, LLDB_INVALID_REGNUM), \ 363 uint32_t vr7[4]; member
|
/external/llvm-project/llvm/lib/Target/CSKY/ |
D | CSKYRegisterInfo.td | 106 def F7_32 : CSKYFReg32<7, "fr7", ["vr7"]>, DwarfRegNum<[39]>;
|
/external/elfutils/tests/ |
D | run-allregs.sh | 1151 1131: vr7 (vr7), unsigned 128 bits 2174 1131: vr7 (vr7), unsigned 128 bits
|
D | run-addrcfi.sh | 1293 vector reg1131 (vr7): undefined 2315 vector reg1131 (vr7): undefined 3343 vector reg1131 (vr7): undefined
|