| /external/XNNPACK/src/f32-vsigmoid/gen/ |
| D | vsigmoid-sse41-rr2-lut64-p2-div-x12.c | 81 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x12() local 109 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x12() local
|
| D | vsigmoid-sse2-rr2-lut64-p2-div-x12.c | 87 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12() local 121 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12() local
|
| D | vsigmoid-sse41-rr2-lut64-p2-div-x16.c | 86 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x16() local 121 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x16() local
|
| D | vsigmoid-sse41-rr2-lut64-p2-div-x20.c | 91 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x20() local 133 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x20() local
|
| D | vsigmoid-sse2-rr2-lut64-p2-div-x16.c | 92 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x16() local 135 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x16() local
|
| D | vsigmoid-sse41-rr2-lut64-p2-div-x24.c | 96 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x24() local 145 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x24() local
|
| D | vsigmoid-sse2-rr2-lut64-p2-div-x20.c | 97 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x20() local 149 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x20() local
|
| D | vsigmoid-neonfma-rr1-lut64-p2-nr1recps1fma-x12.c | 79 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x12() local
|
| D | vsigmoid-neonfma-rr1-lut2048-p1-nr1recps1fma-x12.c | 78 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x12() local
|
| /external/XNNPACK/src/f32-velu/gen/ |
| D | velu-sse41-rr2-lut16-p3-x12.c | 86 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x12() local 114 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x12() local
|
| D | velu-sse41-rr2-lut16-p3-x16.c | 91 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x16() local 126 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x16() local
|
| D | velu-sse2-rr2-lut16-p3-x12.c | 92 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12() local 126 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12() local
|
| D | velu-sse2-rr2-lut16-p3-x16.c | 97 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x16() local 140 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x16() local
|
| D | velu-sse41-rr2-lut16-p3-x20.c | 96 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x20() local 138 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x20() local
|
| D | velu-sse41-rr2-lut16-p3-x24.c | 101 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x24() local 150 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x24() local
|
| D | velu-sse2-rr2-lut16-p3-x20.c | 102 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x20() local 154 const __m128i vl89AB = _mm_unpacklo_epi64(vl89, vlAB); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x20() local
|
| D | velu-neonfma-rr1-lut16-p3-x12.c | 83 const int32x4_t vl89AB = vcombine_s32(vl89, vlAB); in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x12() local
|
| D | velu-neon-rr2-lut16-p3-x12.c | 84 const int32x4_t vl89AB = vcombine_s32(vl89, vlAB); in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x12() local
|
| /external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
| D | neon-rr2-lut64-p2-x12-acc3.c | 84 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_lut64_p2_x12_acc3() local
|
| D | neon-rr2-lut64-p2-x12-acc2.c | 83 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_lut64_p2_x12_acc2() local
|
| D | neonfma-rr1-lut64-p2-x12.c | 81 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_lut64_p2_x12() local
|
| D | neonfma-rr1-lut64-p2-x12-acc3.c | 83 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_lut64_p2_x12_acc3() local
|
| D | neonfma-rr1-lut64-p2-x12-acc2.c | 82 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_lut64_p2_x12_acc2() local
|
| D | neon-rr2-lut64-p2-x12.c | 82 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_lut64_p2_x12() local
|
| D | neonfma-rr1-lut64-p2-x16.c | 90 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_lut64_p2_x16() local
|