/external/XNNPACK/src/f32-vrnd/gen/ |
D | vrndz-neon-x4.c | 34 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x4() local 38 vrndmask0123 = vbicq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndz_ukernel__neon_x4() 40 const float32x4_t vy0123 = vbslq_f32(vrndmask0123, vrndx0123, vx0123); in xnn_f32_vrndz_ukernel__neon_x4()
|
D | vrndne-neon-x4.c | 33 uint32x4_t vrndmask0123 = vcaltq_f32(vmagic_number, vx0123); in xnn_f32_vrndne_ukernel__neon_x4() local 37 vrndmask0123 = vorrq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndne_ukernel__neon_x4() 41 const float32x4_t vy0123 = vbslq_f32(vrndmask0123, vx0123, vrndabsx0123); in xnn_f32_vrndne_ukernel__neon_x4()
|
D | vrndd-neon-x4.c | 35 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x4() local 39 vrndmask0123 = vbicq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndd_ukernel__neon_x4() 41 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndd_ukernel__neon_x4()
|
D | vrndu-neon-x4.c | 35 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x4() local 39 vrndmask0123 = vbicq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndu_ukernel__neon_x4() 41 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x4()
|
D | vrndz-neon-x8.c | 36 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8() local 42 vrndmask0123 = vbicq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndz_ukernel__neon_x8() 45 const float32x4_t vy0123 = vbslq_f32(vrndmask0123, vrndx0123, vx0123); in xnn_f32_vrndz_ukernel__neon_x8()
|
D | vrndne-neon-x8.c | 34 uint32x4_t vrndmask0123 = vcaltq_f32(vmagic_number, vx0123); in xnn_f32_vrndne_ukernel__neon_x8() local 41 vrndmask0123 = vorrq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndne_ukernel__neon_x8() 47 const float32x4_t vy0123 = vbslq_f32(vrndmask0123, vx0123, vrndabsx0123); in xnn_f32_vrndne_ukernel__neon_x8()
|
D | vrndz-sse2-x4.c | 35 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndz_ukernel__sse2_x4() local 39 …const __m128 vy0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, vrnd… in xnn_f32_vrndz_ukernel__sse2_x4()
|
D | vrndne-sse2-x4.c | 35 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndne_ukernel__sse2_x4() local 39 …const __m128 vy0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, vrnd… in xnn_f32_vrndne_ukernel__sse2_x4()
|
D | vrndd-neon-x8.c | 37 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8() local 43 vrndmask0123 = vbicq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndd_ukernel__neon_x8() 46 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndd_ukernel__neon_x8()
|
D | vrndd-sse2-x4.c | 36 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndd_ukernel__sse2_x4() local 40 …const __m128 vrndx0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, v… in xnn_f32_vrndd_ukernel__sse2_x4()
|
D | vrndu-neon-x8.c | 37 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8() local 43 vrndmask0123 = vbicq_u32(vrndmask0123, vmovq_n_u32(UINT32_C(0x80000000))); in xnn_f32_vrndu_ukernel__neon_x8() 46 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x8()
|
D | vrndu-sse2-x4.c | 36 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndu_ukernel__sse2_x4() local 40 …const __m128 vrndx0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, v… in xnn_f32_vrndu_ukernel__sse2_x4()
|
D | vrndz-sse2-x8.c | 37 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndz_ukernel__sse2_x8() local 43 …const __m128 vy0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, vrnd… in xnn_f32_vrndz_ukernel__sse2_x8()
|
D | vrndne-sse2-x8.c | 37 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndne_ukernel__sse2_x8() local 43 …const __m128 vy0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, vrnd… in xnn_f32_vrndne_ukernel__sse2_x8()
|
D | vrndd-sse2-x8.c | 38 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndd_ukernel__sse2_x8() local 44 …const __m128 vrndx0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, v… in xnn_f32_vrndd_ukernel__sse2_x8()
|
D | vrndu-sse2-x8.c | 38 …const __m128 vrndmask0123 = _mm_castsi128_ps(_mm_or_si128(vmagic, _mm_cmpeq_epi32(vintx0123, vmagi… in xnn_f32_vrndu_ukernel__sse2_x8() local 44 …const __m128 vrndx0123 = _mm_or_ps(_mm_and_ps(vx0123, vrndmask0123), _mm_andnot_ps(vrndmask0123, v… in xnn_f32_vrndu_ukernel__sse2_x8()
|
D | vrndne-wasmsimd-addsub-x8.c | 38 const v128_t vrndmask0123 = wasm_v128_or(vsign_mask, wasm_f32x4_le(vmagic_number, vabsx0123)); in xnn_f32_vrndne_ukernel__wasmsimd_addsub_x8() local 44 const v128_t vy0123 = wasm_v128_bitselect(vx0123, vrndabsx0123, vrndmask0123); in xnn_f32_vrndne_ukernel__wasmsimd_addsub_x8()
|
D | vrndz-wasmsimd-cvt-x8.c | 41 … const v128_t vrndmask0123 = wasm_v128_andnot(wasm_f32x4_lt(vabsx0123, vmagic_number), vsign_mask); in xnn_f32_vrndz_ukernel__wasmsimd_cvt_x8() local 45 const v128_t vy0123 = wasm_v128_bitselect(vrndx0123, vx0123, vrndmask0123); in xnn_f32_vrndz_ukernel__wasmsimd_cvt_x8()
|
D | vrndd-wasmsimd-addsub-x8.c | 39 const v128_t vrndmask0123 = wasm_v128_or(vsign_mask, wasm_f32x4_le(vmagic_number, vabsx0123)); in xnn_f32_vrndd_ukernel__wasmsimd_addsub_x8() local 45 const v128_t vrndx0123 = wasm_v128_bitselect(vx0123, vrndabsx0123, vrndmask0123); in xnn_f32_vrndd_ukernel__wasmsimd_addsub_x8()
|
D | vrndz-wasmsimd-addsub-x8.c | 39 const v128_t vrndmask0123 = wasm_v128_or(vsign_mask, wasm_f32x4_le(vmagic_number, vabsx0123)); in xnn_f32_vrndz_ukernel__wasmsimd_addsub_x8() local 51 const v128_t vy0123 = wasm_v128_bitselect(vx0123, vflrabsx0123, vrndmask0123); in xnn_f32_vrndz_ukernel__wasmsimd_addsub_x8()
|
D | vrndd-wasmsimd-cvt-x8.c | 42 … const v128_t vrndmask0123 = wasm_v128_andnot(wasm_f32x4_lt(vabsx0123, vmagic_number), vsign_mask); in xnn_f32_vrndd_ukernel__wasmsimd_cvt_x8() local 46 const v128_t vrndx0123 = wasm_v128_bitselect(vprerndx0123, vx0123, vrndmask0123); in xnn_f32_vrndd_ukernel__wasmsimd_cvt_x8()
|
D | vrndu-wasmsimd-cvt-x8.c | 42 … const v128_t vrndmask0123 = wasm_v128_andnot(wasm_f32x4_lt(vabsx0123, vmagic_number), vsign_mask); in xnn_f32_vrndu_ukernel__wasmsimd_cvt_x8() local 46 const v128_t vrndx0123 = wasm_v128_bitselect(vprerndx0123, vx0123, vrndmask0123); in xnn_f32_vrndu_ukernel__wasmsimd_cvt_x8()
|
D | vrndu-wasmsimd-addsub-x8.c | 39 const v128_t vrndmask0123 = wasm_v128_or(vsign_mask, wasm_f32x4_le(vmagic_number, vabsx0123)); in xnn_f32_vrndu_ukernel__wasmsimd_addsub_x8() local 45 const v128_t vrndx0123 = wasm_v128_bitselect(vx0123, vrndabsx0123, vrndmask0123); in xnn_f32_vrndu_ukernel__wasmsimd_addsub_x8()
|