Searched refs:vcvtq_s32_f32 (Results 1 – 25 of 27) sorted by relevance
12
33 const int32x4_t vintx0123 = vcvtq_s32_f32(vx0123); in xnn_f32_vrndz_ukernel__neon_x8()34 const int32x4_t vintx4567 = vcvtq_s32_f32(vx4567); in xnn_f32_vrndz_ukernel__neon_x8()53 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndz_ukernel__neon_x8()62 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndz_ukernel__neon_x8()
34 const int32x4_t vintx0123 = vcvtq_s32_f32(vx0123); in xnn_f32_vrndd_ukernel__neon_x8()35 const int32x4_t vintx4567 = vcvtq_s32_f32(vx4567); in xnn_f32_vrndd_ukernel__neon_x8()63 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndd_ukernel__neon_x8()75 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndd_ukernel__neon_x8()
34 const int32x4_t vintx0123 = vcvtq_s32_f32(vx0123); in xnn_f32_vrndu_ukernel__neon_x8()35 const int32x4_t vintx4567 = vcvtq_s32_f32(vx4567); in xnn_f32_vrndu_ukernel__neon_x8()66 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndu_ukernel__neon_x8()79 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndu_ukernel__neon_x8()
32 const int32x4_t vintx0123 = vcvtq_s32_f32(vx0123); in xnn_f32_vrndz_ukernel__neon_x4()46 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndz_ukernel__neon_x4()
33 const int32x4_t vintx0123 = vcvtq_s32_f32(vx0123); in xnn_f32_vrndd_ukernel__neon_x4()53 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndd_ukernel__neon_x4()
33 const int32x4_t vintx0123 = vcvtq_s32_f32(vx0123); in xnn_f32_vrndu_ukernel__neon_x4()55 const int32x4_t vintx = vcvtq_s32_f32(vx); in xnn_f32_vrndu_ukernel__neon_x4()
33 const int32x4_t vintx${ABC[N:N+4]} = vcvtq_s32_f32(vx${ABC[N:N+4]});53 const int32x4_t vintx = vcvtq_s32_f32(vx);62 const int32x4_t vintx = vcvtq_s32_f32(vx);
34 const int32x4_t vintx${ABC[N:N+4]} = vcvtq_s32_f32(vx${ABC[N:N+4]});63 const int32x4_t vintx = vcvtq_s32_f32(vx);75 const int32x4_t vintx = vcvtq_s32_f32(vx);
34 const int32x4_t vintx${ABC[N:N+4]} = vcvtq_s32_f32(vx${ABC[N:N+4]});66 const int32x4_t vintx = vcvtq_s32_f32(vx);79 const int32x4_t vintx = vcvtq_s32_f32(vx);
47 tmp = vcvtq_f32_s32(vcvtq_s32_f32(fx));78 mm = vcvtq_s32_f32(fx);
33 const float32x4_t vrndx = vcvtq_f32_s32(vcvtq_s32_f32(vx)); in xnn_math_f32_roundz__neon_cvt()
35 const float32x4_t vprerndx = vcvtq_f32_s32(vcvtq_s32_f32(vx)); in xnn_math_f32_roundd__neon_cvt()
35 const float32x4_t vprerndx = vcvtq_f32_s32(vcvtq_s32_f32(vx)); in xnn_math_f32_roundu__neon_cvt()
20 auto roundtrip = vcvtq_f32_s32(vcvtq_s32_f32(v)); in emulate_vrndmq_f32()610 return vcvtq_s32_f32(src.fVec);708 return vcvtq_s32_f32((x + 0.5f).fVec); in Sk4f_round()
80 return vcvtq_s32_f32(a); in test_vcvtq_s32_f32()
20 auto roundtrip = vcvtq_f32_s32(vcvtq_s32_f32(v)); in emulate_vrndmq_f32()632 return vcvtq_s32_f32(src.fVec);730 return vcvtq_s32_f32((x + 0.5f).fVec); in Sk4f_round()
235 vcvtq_s32_f32(vmulq_n_f32(vsubq_f32(normed, out_min), out_scale)); in InstanceNorm()
2325 return vcvtq_s32_f32(a); in test_vcvtq_s32_f32()
2395 return vcvtq_s32_f32(a); in test_vcvtq_s32_f32()
2662 return vcvtq_s32_f32(a); in test_vcvtq_s32_f32()
4440 return vreinterpretq_m128i_s32(vcvtq_s32_f32(vreinterpretq_f32_m128(a))); in _mm_cvttps_epi32()4583 int32x4_t r_normal = vcvtq_s32_f32(vaddq_f32( in _mm_cvtps_epi32()4586 vcvtq_s32_f32(vreinterpretq_f32_m128(a)); /* truncate to integer: [a] */ in _mm_cvtps_epi32()
2245 return vcvtq_s32_f32(vaddq_f32(input, round)); in RoundToNearest()
210 F roundtrip = vcvtq_f32_s32(vcvtq_s32_f32(v));
211 F roundtrip = vcvtq_f32_s32(vcvtq_s32_f32(v));
1773 _NEON2SSESTORAGE int32x4_t vcvtq_s32_f32(float32x4_t a); // VCVT.S32.F32 q0, q012708 _NEON2SSESTORAGE int32x4_t vcvtq_s32_f32(float32x4_t a); // VCVT.S32.F32 q0, q012709 _NEON2SSE_INLINE int32x4_t vcvtq_s32_f32(float32x4_t a) in vcvtq_s32_f32() function12738 res2 = vcvtq_s32_f32(max); in vcvtq_u32_f32()