/external/XNNPACK/src/f32-sigmoid/gen/ |
D | neon-rr2-lut64-p2-nr2recps-x24.c | 188 vp0123 = vmlsq_f32(vt0123, vp0123, vt0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 189 vp4567 = vmlsq_f32(vt4567, vp4567, vt4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 190 vp89AB = vmlsq_f32(vt89AB, vp89AB, vt89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 191 vpCDEF = vmlsq_f32(vtCDEF, vpCDEF, vtCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 192 vpGHIJ = vmlsq_f32(vtGHIJ, vpGHIJ, vtGHIJ); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 193 vpKLMN = vmlsq_f32(vtKLMN, vpKLMN, vtKLMN); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 200 const float32x4_t vy0123 = vmlsq_f32(vs0123, vs0123, vp0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 201 const float32x4_t vy4567 = vmlsq_f32(vs4567, vs4567, vp4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 202 const float32x4_t vy89AB = vmlsq_f32(vs89AB, vs89AB, vp89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 203 const float32x4_t vyCDEF = vmlsq_f32(vsCDEF, vsCDEF, vpCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() [all …]
|
D | neon-rr2-lut64-p2-nr2recps-x20.c | 171 vp0123 = vmlsq_f32(vt0123, vp0123, vt0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 172 vp4567 = vmlsq_f32(vt4567, vp4567, vt4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 173 vp89AB = vmlsq_f32(vt89AB, vp89AB, vt89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 174 vpCDEF = vmlsq_f32(vtCDEF, vpCDEF, vtCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 175 vpGHIJ = vmlsq_f32(vtGHIJ, vpGHIJ, vtGHIJ); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 182 const float32x4_t vy0123 = vmlsq_f32(vs0123, vs0123, vp0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 183 const float32x4_t vy4567 = vmlsq_f32(vs4567, vs4567, vp4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 184 const float32x4_t vy89AB = vmlsq_f32(vs89AB, vs89AB, vp89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 185 const float32x4_t vyCDEF = vmlsq_f32(vsCDEF, vsCDEF, vpCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() 186 const float32x4_t vyGHIJ = vmlsq_f32(vsGHIJ, vsGHIJ, vpGHIJ); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() [all …]
|
D | neon-rr2-lut64-p2-nr2recps-x16.c | 154 vp0123 = vmlsq_f32(vt0123, vp0123, vt0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 155 vp4567 = vmlsq_f32(vt4567, vp4567, vt4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 156 vp89AB = vmlsq_f32(vt89AB, vp89AB, vt89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 157 vpCDEF = vmlsq_f32(vtCDEF, vpCDEF, vtCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 164 const float32x4_t vy0123 = vmlsq_f32(vs0123, vs0123, vp0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 165 const float32x4_t vy4567 = vmlsq_f32(vs4567, vs4567, vp4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 166 const float32x4_t vy89AB = vmlsq_f32(vs89AB, vs89AB, vp89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 167 const float32x4_t vyCDEF = vmlsq_f32(vsCDEF, vsCDEF, vpCDEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 278 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() 285 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x16() [all …]
|
D | neon-rr2-lut64-p2-nr2recps-x12.c | 137 vp0123 = vmlsq_f32(vt0123, vp0123, vt0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 138 vp4567 = vmlsq_f32(vt4567, vp4567, vt4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 139 vp89AB = vmlsq_f32(vt89AB, vp89AB, vt89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 146 const float32x4_t vy0123 = vmlsq_f32(vs0123, vs0123, vp0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 147 const float32x4_t vy4567 = vmlsq_f32(vs4567, vs4567, vp4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 148 const float32x4_t vy89AB = vmlsq_f32(vs89AB, vs89AB, vp89AB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 250 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 257 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 340 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12() 347 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x12()
|
D | neon-rr2-lut64-p2-nr2recps-x8.c | 120 vp0123 = vmlsq_f32(vt0123, vp0123, vt0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 121 vp4567 = vmlsq_f32(vt4567, vp4567, vt4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 128 const float32x4_t vy0123 = vmlsq_f32(vs0123, vs0123, vp0123); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 129 const float32x4_t vy4567 = vmlsq_f32(vs4567, vs4567, vp4567); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 222 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 229 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 312 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() 319 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8()
|
D | neon-rr2-lut64-p2-nr2recps-x4.c | 98 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x4() 105 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x4() 188 vp = vmlsq_f32(vt, vp, vt); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x4() 195 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x4()
|
/external/XNNPACK/src/math/ |
D | sigmoid-neon-rr1-lut64-p2-nr2recps.c | 91 vp = vmlsq_f32(vt, vp, vt); in xnn_math_f32_sigmoid__neon_rr1_lut64_p2_nr2recps() 98 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_math_f32_sigmoid__neon_rr1_lut64_p2_nr2recps()
|
D | sigmoid-neon-rr2-lut64-p2-nr2recps.c | 95 vp = vmlsq_f32(vt, vp, vt); in xnn_math_f32_sigmoid__neon_rr2_lut64_p2_nr2recps() 102 const float32x4_t vy = vmlsq_f32(vs, vs, vp); in xnn_math_f32_sigmoid__neon_rr2_lut64_p2_nr2recps()
|
/external/webrtc/webrtc/modules/audio_processing/aec/ |
D | aec_rdft_neon.c | 59 x0v = vmlsq_f32(x1v, x3w, vec_swap_sign); in cft1st_128_neon() 103 vmlsq_f32(x1r0_1i0_1r1_x1i1, vec_swap_sign, x3i0_3r0_3i1_x3r1); in cftmdl_128_neon() 161 vmlsq_f32(x1r0_1i0_1r1_x1i1, vec_swap_sign, x3i0_3r0_3i1_x3r1); in cftmdl_128_neon()
|
D | aec_core_neon.c | 62 const float32x4_t e = vmlsq_f32(a, x_fft_buf_im, h_fft_buf_im); in FilterFarNEON() 219 const float32x4_t f = vmlsq_f32(c, x_fft_buf_im, e_fft_re); in FilterAdaptationNEON() 558 vec_dfwefw0110 = vmlsq_f32(vec_dfwefw0110, vec_dfw1, vec_efw0); in SmoothedPSD() 571 vec_dfwxfw0110 = vmlsq_f32(vec_dfwxfw0110, vec_dfw1, vec_xfw0); in SmoothedPSD()
|
/external/XNNPACK/src/f32-sigmoid/ |
D | neon-lut64-p2.c.in | 12 $VMULSUBQ_F32 = "vfmsq_f32" if FMA else "vmlsq_f32"
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 596 _NEON2SSESTORAGE float32x4_t vmlsq_f32(float32x4_t a, float32x4_t b, float32x4_t c); // VMLS.F32 q0… 4110 _NEON2SSESTORAGE float32x4_t vmlsq_f32(float32x4_t a, float32x4_t b, float32x4_t c); // VMLS.F32 q0… 4111 _NEON2SSE_INLINE float32x4_t vmlsq_f32(float32x4_t a, float32x4_t b, float32x4_t c) // VMLS.F32 q0,… in vmlsq_f32() function 13533 return vmlsq_f32(a,b,c); 14217 return vmlsq_f32(a,b,vc); in vmlsq_n_f32()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 605 return vmlsq_f32(v1, v2, v3); in test_vmlsq_f32()
|
D | arm_neon_intrinsics.c | 9399 return vmlsq_f32(a, b, c); in test_vmlsq_f32()
|