/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-sse41-rr2-lut64-p2-div-x8.c | 67 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x8() local 86 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x8() local
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x8.c | 70 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x8() local 93 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x8() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x12.c | 72 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x12() local 98 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x12() local
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x12.c | 75 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12() local 107 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x16.c | 77 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x16() local 110 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x16() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x20.c | 82 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x20() local 122 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x20() local
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x16.c | 80 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x16() local 121 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x16() local
|
D | vsigmoid-neonfma-rr1-lut64-p2-div-x8.c | 59 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx45]); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut64_p2_div_x8() local
|
D | vsigmoid-neonfma-rr1-lut2048-p1-div-x8.c | 58 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx45]); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x8() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x24.c | 87 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x24() local 134 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x24() local
|
D | vsigmoid-neonfma-rr1-lut64-p2-nr1recps1fma-x8.c | 59 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx45]); in xnn_f32_vsigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x8() local
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-sse41-rr2-lut16-p3-x8.c | 72 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x8() local 91 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x8() local
|
D | velu-sse2-rr2-lut16-p3-x8.c | 75 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x8() local 98 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x8() local
|
D | velu-sse41-rr2-lut16-p3-x12.c | 77 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x12() local 103 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x12() local
|
D | velu-sse2-rr2-lut16-p3-x12.c | 80 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12() local 112 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12() local
|
D | velu-sse41-rr2-lut16-p3-x16.c | 82 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x16() local 115 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x16() local
|
D | velu-sse41-rr2-lut16-p3-x20.c | 87 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x20() local 127 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x20() local
|
D | velu-sse2-rr2-lut16-p3-x16.c | 85 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x16() local 126 const __m128i vl45 = _mm_unpacklo_epi32(vl4, vl5); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x16() local
|
D | velu-neon-rr2-lut16-p3-x8.c | 68 …int32x2_t vl45 = vld1_dup_s32((const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint3… in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x8() local
|
D | velu-neonfma-rr1-lut16-p3-x8.c | 67 …int32x2_t vl45 = vld1_dup_s32((const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint3… in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x8() local
|
D | velu-sse41-rr2-lut16-p3-x24.c | 92 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x24() local 139 …const __m128i vl45 = _mm_insert_epi32(vl4, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x24() local
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | neonfma-rr1-lut64-p2-x8.c | 61 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx45]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_lut64_p2_x8() local
|
D | neonfma-rr1-lut64-p2-x8-acc2.c | 62 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx45]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_lut64_p2_x8_acc2() local
|
D | neon-rr2-lut64-p2-x8.c | 62 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx45]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_lut64_p2_x8() local
|
D | neon-rr2-lut64-p2-x8-acc2.c | 63 float32x2_t vl45 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx45]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_lut64_p2_x8_acc2() local
|