/external/XNNPACK/src/f32-sigmoid/gen/ |
D | sse41-lut64-p2-div-x20.c | 103 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_sigmoid_ukernel__sse41_lut64_p2_div_x20() local 149 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_sigmoid_ukernel__sse41_lut64_p2_div_x20() local
|
D | sse41-lut64-p2-div-x24.c | 108 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_sigmoid_ukernel__sse41_lut64_p2_div_x24() local 161 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_sigmoid_ukernel__sse41_lut64_p2_div_x24() local
|
D | sse2-lut64-p2-div-x20.c | 112 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_sigmoid_ukernel__sse2_lut64_p2_div_x20() local 168 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_sigmoid_ukernel__sse2_lut64_p2_div_x20() local
|
D | sse2-lut64-p2-div-x24.c | 117 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_sigmoid_ukernel__sse2_lut64_p2_div_x24() local 182 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_sigmoid_ukernel__sse2_lut64_p2_div_x24() local
|
D | neonfma-rr1-lut2048-p1-div-x20.c | 85 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() local
|
D | neonfma-rr1-lut64-p2-div-x20.c | 86 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() local
|
D | neonfma-rr1-lut64-p2-nr2recps-x20.c | 86 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x20() local
|
D | neonfma-rr1-lut2048-p1-nr2fma-x20.c | 85 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x20() local
|
D | neonfma-rr1-lut64-p2-nr1recps1fma-x20.c | 86 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x20() local
|
D | neonfma-rr1-lut64-p2-nr2fma-x20.c | 86 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x20() local
|
D | neon-rr2-lut64-p2-nr2recps-x20.c | 87 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x20() local
|
D | neon-rr2-lut2048-p1-nr2recps-x20.c | 86 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neon_rr2_lut2048_p1_nr2recps_x20() local
|
D | neonfma-rr1-lut2048-p1-nr2recps-x20.c | 85 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidxGH]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x20() local
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-sse41-rr2-lut16-p3-x20.c | 109 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x20() local 155 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x20() local
|
D | velu-sse41-rr2-lut16-p3-x24.c | 114 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x24() local 167 …const __m128i vlGH = _mm_insert_epi32(vlG, *((const int*) ((uintptr_t) xnn_table_exp2minus_k_over_… in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x24() local
|
D | velu-sse2-rr2-lut16-p3-x20.c | 118 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x20() local 174 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x20() local
|
D | velu-sse2-rr2-lut16-p3-x24.c | 123 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x24() local 188 const __m128i vlGH = _mm_unpacklo_epi32(vlG, vlH); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x24() local
|
D | velu-neon-rr2-lut16-p3-x20.c | 105 …int32x2_t vlGH = vld1_dup_s32((const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint3… in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x20() local
|
D | velu-neonfma-rr1-lut16-p3-x20.c | 104 …int32x2_t vlGH = vld1_dup_s32((const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint3… in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x20() local
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | neonfma-lut64-p2-x20.c | 112 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidxGH]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() local
|
D | neonfma-lut64-p2-x20-acc2.c | 113 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidxGH]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() local
|
D | neon-lut64-p2-x20.c | 113 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidxGH]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() local
|
D | neon-lut64-p2-x20-acc2.c | 114 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidxGH]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() local
|
D | neonfma-lut64-p2-x20-acc5.c | 116 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidxGH]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() local
|
D | neon-lut64-p2-x20-acc5.c | 117 float32x2_t vlGH = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidxGH]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() local
|