/external/XNNPACK/src/f32-vbinary/gen/ |
D | vmax-wasmsimd-x86-x16.c | 47 const v128_t vm89AB = wasm_f32x4_le(va89AB, vb89AB); in xnn_f32_vmax_ukernel__wasmsimd_x86_x16() local 52 v128_t vy89AB = wasm_v128_bitselect(vb89AB, va89AB, vm89AB); in xnn_f32_vmax_ukernel__wasmsimd_x86_x16()
|
D | vmin-wasmsimd-x86-x16.c | 47 const v128_t vm89AB = wasm_f32x4_lt(va89AB, vb89AB); in xnn_f32_vmin_ukernel__wasmsimd_x86_x16() local 52 v128_t vy89AB = wasm_v128_bitselect(va89AB, vb89AB, vm89AB); in xnn_f32_vmin_ukernel__wasmsimd_x86_x16()
|
/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-sse2-rr2-p5-div-x12.c | 106 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x12() local 110 vf89AB = _mm_or_ps(_mm_and_ps(vf89AB, vm89AB), _mm_andnot_ps(vm89AB, _mm_sub_ps(vone, vf89AB))); in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x12()
|
D | vsigmoid-sse2-rr2-p5-div-x16.c | 122 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x16() local 127 vf89AB = _mm_or_ps(_mm_and_ps(vf89AB, vm89AB), _mm_andnot_ps(vm89AB, _mm_sub_ps(vone, vf89AB))); in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x16()
|
D | vsigmoid-sse2-rr2-p5-div-x20.c | 138 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x20() local 144 vf89AB = _mm_or_ps(_mm_and_ps(vf89AB, vm89AB), _mm_andnot_ps(vm89AB, _mm_sub_ps(vone, vf89AB))); in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x20()
|
D | vsigmoid-neonfma-rr1-p5-div-x12.c | 100 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_div_x12() local 104 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_div_x12()
|
D | vsigmoid-sse2-rr2-p5-div-x24.c | 154 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x24() local 161 vf89AB = _mm_or_ps(_mm_and_ps(vf89AB, vm89AB), _mm_andnot_ps(vm89AB, _mm_sub_ps(vone, vf89AB))); in xnn_f32_vsigmoid_ukernel__sse2_rr2_p5_div_x24()
|
D | vsigmoid-neonfma-rr1-p5-nr1recps1fma-x12.c | 112 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x12() local 116 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x12()
|
D | vsigmoid-neonfma-rr1-p5-nr2recps-x12.c | 112 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr2recps_x12() local 116 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr2recps_x12()
|
D | vsigmoid-neonfma-rr1-p5-div-x16.c | 115 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_div_x16() local 120 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_div_x16()
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x12.c | 162 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12() local 166 vf89AB = _mm_or_ps(_mm_and_ps(vf89AB, vm89AB), _mm_andnot_ps(vm89AB, _mm_sub_ps(vone, vf89AB))); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12()
|
D | vsigmoid-neonfma-rr1-p5-nr2fma-x12.c | 112 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr2fma_x12() local 116 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr2fma_x12()
|
D | vsigmoid-neon-rr2-p5-nr2recps-x12.c | 117 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neon_rr2_p5_nr2recps_x12() local 121 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neon_rr2_p5_nr2recps_x12()
|
D | vsigmoid-neonfma-rr1-lut64-p2-div-x12.c | 119 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut64_p2_div_x12() local 123 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut64_p2_div_x12()
|
D | vsigmoid-neonfma-rr1-lut2048-p1-div-x12.c | 114 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x12() local 118 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x12()
|
D | vsigmoid-neonfma-rr1-p5-nr1recps1fma-x16.c | 130 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x16() local 135 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_p5_nr1recps1fma_x16()
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-sse2-rr2-p6-x12.c | 113 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_velu_ukernel__sse2_rr2_p6_x12() local 118 const __m128 vy89AB = _mm_or_ps(_mm_and_ps(ve89AB, vm89AB), _mm_andnot_ps(vm89AB, vx89AB)); in xnn_f32_velu_ukernel__sse2_rr2_p6_x12()
|
D | velu-sse2-rr2-p6-x16.c | 129 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_velu_ukernel__sse2_rr2_p6_x16() local 136 const __m128 vy89AB = _mm_or_ps(_mm_and_ps(ve89AB, vm89AB), _mm_andnot_ps(vm89AB, vx89AB)); in xnn_f32_velu_ukernel__sse2_rr2_p6_x16()
|
D | velu-sse2-rr2-p6-x20.c | 145 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_velu_ukernel__sse2_rr2_p6_x20() local 154 const __m128 vy89AB = _mm_or_ps(_mm_and_ps(ve89AB, vm89AB), _mm_andnot_ps(vm89AB, vx89AB)); in xnn_f32_velu_ukernel__sse2_rr2_p6_x20()
|
D | velu-sse2-rr2-p6-x24.c | 161 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_velu_ukernel__sse2_rr2_p6_x24() local 172 const __m128 vy89AB = _mm_or_ps(_mm_and_ps(ve89AB, vm89AB), _mm_andnot_ps(vm89AB, vx89AB)); in xnn_f32_velu_ukernel__sse2_rr2_p6_x24()
|
D | velu-neon-rr2-p6-x12.c | 111 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_velu_ukernel__neon_rr2_p6_x12() local 116 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neon_rr2_p6_x12()
|
D | velu-neonfma-rr1-p6-x12.c | 106 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x12() local 111 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x12()
|
D | velu-neonfma-rr1-p6-x16.c | 121 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x16() local 128 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x16()
|
D | velu-sse2-rr2-lut16-p3-x12.c | 171 …const __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89A… in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12() local 176 const __m128 vy89AB = _mm_or_ps(_mm_and_ps(ve89AB, vm89AB), _mm_andnot_ps(vm89AB, vx89AB)); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12()
|
D | velu-neon-rr2-p6-x16.c | 127 const uint32x4_t vm89AB = vcltq_f32(vx89AB, vmovq_n_f32(0.0f)); in xnn_f32_velu_ukernel__neon_rr2_p6_x16() local 134 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neon_rr2_p6_x16()
|