/external/XNNPACK/src/f16-f32-vcvt/gen/ |
D | vcvt-wasmsimd-int32-x24.c | 55 const v128_t vsign5 = wasm_v128_and(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__wasmsimd_int32_x24() local 62 const v128_t vnonsign5 = wasm_v128_xor(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__wasmsimd_int32_x24() 90 const v128_t vf5 = wasm_v128_or(vsign5, wasm_v128_bitselect(vnorm5, vdenorm5, vxmask5)); in xnn_f16_f32_vcvt_ukernel__wasmsimd_int32_x24()
|
D | vcvt-sse41-int32-x24.c | 54 const __m128i vsign5 = _mm_and_si128(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__sse41_int32_x24() local 61 const __m128i vnonsign5 = _mm_xor_si128(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__sse41_int32_x24() 89 const __m128i vf5 = _mm_or_si128(vsign5, _mm_blendv_epi8(vdenorm5, vnorm5, vmask5)); in xnn_f16_f32_vcvt_ukernel__sse41_int32_x24()
|
D | vcvt-neon-int32-x24.c | 53 const uint32x4_t vsign5 = vandq_u32(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__neon_int32_x24() local 60 const uint32x4_t vnonsign5 = veorq_u32(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__neon_int32_x24() 88 …const uint32x4_t vf5 = vorrq_u32(vsign5, vreinterpretq_u32_f32(vbslq_f32(vxmask5, vnorm5, vdenorm5… in xnn_f16_f32_vcvt_ukernel__neon_int32_x24()
|
D | vcvt-avx-int32-x24.c | 54 const __m128i vsign5 = _mm_and_si128(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__avx_int32_x24() local 61 const __m128i vnonsign5 = _mm_xor_si128(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__avx_int32_x24() 89 const __m128i vf5 = _mm_or_si128(vsign5, _mm_blendv_epi8(vdenorm5, vnorm5, vmask5)); in xnn_f16_f32_vcvt_ukernel__avx_int32_x24()
|
D | vcvt-sse2-int32-x24.c | 54 const __m128i vsign5 = _mm_and_si128(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__sse2_int32_x24() local 61 const __m128i vnonsign5 = _mm_xor_si128(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__sse2_int32_x24() 94 const __m128i vf5 = _mm_or_si128(vsign5, in xnn_f16_f32_vcvt_ukernel__sse2_int32_x24()
|
D | vcvt-neon-int32-x32.c | 56 const uint32x4_t vsign5 = vandq_u32(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__neon_int32_x32() local 65 const uint32x4_t vnonsign5 = veorq_u32(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__neon_int32_x32() 101 …const uint32x4_t vf5 = vorrq_u32(vsign5, vreinterpretq_u32_f32(vbslq_f32(vxmask5, vnorm5, vdenorm5… in xnn_f16_f32_vcvt_ukernel__neon_int32_x32()
|
D | vcvt-wasmsimd-int32-x32.c | 58 const v128_t vsign5 = wasm_v128_and(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__wasmsimd_int32_x32() local 67 const v128_t vnonsign5 = wasm_v128_xor(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__wasmsimd_int32_x32() 103 const v128_t vf5 = wasm_v128_or(vsign5, wasm_v128_bitselect(vnorm5, vdenorm5, vxmask5)); in xnn_f16_f32_vcvt_ukernel__wasmsimd_int32_x32()
|
D | vcvt-avx-int32-x32.c | 57 const __m128i vsign5 = _mm_and_si128(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__avx_int32_x32() local 66 const __m128i vnonsign5 = _mm_xor_si128(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__avx_int32_x32() 102 const __m128i vf5 = _mm_or_si128(vsign5, _mm_blendv_epi8(vdenorm5, vnorm5, vmask5)); in xnn_f16_f32_vcvt_ukernel__avx_int32_x32()
|
D | vcvt-sse41-int32-x32.c | 57 const __m128i vsign5 = _mm_and_si128(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__sse41_int32_x32() local 66 const __m128i vnonsign5 = _mm_xor_si128(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__sse41_int32_x32() 102 const __m128i vf5 = _mm_or_si128(vsign5, _mm_blendv_epi8(vdenorm5, vnorm5, vmask5)); in xnn_f16_f32_vcvt_ukernel__sse41_int32_x32()
|
D | vcvt-sse2-int32-x32.c | 57 const __m128i vsign5 = _mm_and_si128(vw5, vsign_mask); in xnn_f16_f32_vcvt_ukernel__sse2_int32_x32() local 66 const __m128i vnonsign5 = _mm_xor_si128(vw5, vsign5); in xnn_f16_f32_vcvt_ukernel__sse2_int32_x32() 107 const __m128i vf5 = _mm_or_si128(vsign5, in xnn_f16_f32_vcvt_ukernel__sse2_int32_x32()
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-avx512f-rr1-lut16-p3-perm-x96.c | 141 const __mmask16 vsign5 = _mm512_cmp_ps_mask(vx5, vzero, _CMP_NLT_US); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x96() local 148 vy5 = _mm512_mask_mul_ps(vy5, vsign5, vx5, vbeta); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x96()
|
D | velu-avx512f-rr1-p6-x96.c | 151 const __mmask16 vsign5 = _mm512_cmp_ps_mask(vx5, vzero, _CMP_NLT_US); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96() local 158 vy5 = _mm512_mask_mul_ps(vy5, vsign5, vx5, vbeta); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
|
D | velu-avx512f-rr1-lut16-p3-perm-x112.c | 154 const __mmask16 vsign5 = _mm512_cmp_ps_mask(vx5, vzero, _CMP_NLT_US); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x112() local 163 vy5 = _mm512_mask_mul_ps(vy5, vsign5, vx5, vbeta); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x112()
|
D | velu-avx512f-rr1-p6-x112.c | 165 const __mmask16 vsign5 = _mm512_cmp_ps_mask(vx5, vzero, _CMP_NLT_US); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x112() local 174 vy5 = _mm512_mask_mul_ps(vy5, vsign5, vx5, vbeta); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x112()
|
D | velu-avx512f-rr1-lut16-p3-perm-x128.c | 167 const __mmask16 vsign5 = _mm512_cmp_ps_mask(vx5, vzero, _CMP_NLT_US); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x128() local 178 vy5 = _mm512_mask_mul_ps(vy5, vsign5, vx5, vbeta); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x128()
|
D | velu-avx512f-rr1-p6-x128.c | 179 const __mmask16 vsign5 = _mm512_cmp_ps_mask(vx5, vzero, _CMP_NLT_US); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x128() local 190 vy5 = _mm512_mask_mul_ps(vy5, vsign5, vx5, vbeta); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x128()
|