/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | scalar-rr2-p5-x4.c | 70 const float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x4() local
|
D | scalar-rr2-p5-x4-acc2.c | 71 const float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x4_acc2() local
|
D | scalar-rr2-p5-x4-acc4.c | 73 const float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x4_acc4() local
|
D | scalar-rr2-lut64-p2-x4-acc2.c | 92 const float vs3 = fp32_from_bits(xnn_table_exp2_k_over_64[vidx3] + ve3); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x4_acc2() local
|
D | scalar-rr2-lut64-p2-x4.c | 91 const float vs3 = fp32_from_bits(xnn_table_exp2_k_over_64[vidx3] + ve3); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x4() local
|
D | scalar-rr2-lut64-p2-x4-acc4.c | 94 const float vs3 = fp32_from_bits(xnn_table_exp2_k_over_64[vidx3] + ve3); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x4_acc4() local
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-wasm-rr2-p6-x4.c | 65 float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() local
|
D | velu-scalar-rr2-p6-x4.c | 65 float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() local
|
D | velu-avx2-rr1-p6-x32.c | 62 __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn3), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x32() local
|
D | velu-scalar-rr2-lut16-p3-x4.c | 79 float vs3 = fp32_from_bits(xnn_table_exp2minus_k_over_16[vidx3] + ven3); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x4() local
|
D | velu-wasm-rr2-p6-x5.c | 68 float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() local
|
D | velu-scalar-rr2-p6-x5.c | 68 float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local
|
D | velu-avx512f-rr1-p6-x64.c | 64 __m512 vs3 = _mm512_castsi512_ps(_mm512_slli_epi32(_mm512_castps_si512(vn3), 23)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x64() local
|
D | velu-wasm-rr2-lut16-p3-x4.c | 79 float vs3 = fp32_from_bits(xnn_table_exp2minus_k_over_16[vidx3] + ven3); in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x4() local
|
D | velu-wasm-rr2-p6-x6.c | 71 float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x6() local
|
D | velu-scalar-rr2-p6-x6.c | 71 float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
|
D | velu-avx2-rr1-p6-x40.c | 65 __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn3), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x40() local
|
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
D | avx2-p5-x32.c | 70 const __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn3), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x32() local
|
D | avx2-p5-x40.c | 73 const __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn3), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x40() local
|
/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-scalar-rr2-p5-div-x4.c | 59 const float vs3 = fp32_from_bits(fp32_to_bits(vn3) << 23); in xnn_f32_vsigmoid_ukernel__scalar_rr2_p5_div_x4() local
|
D | vsigmoid-scalar-rr2-lut2048-p1-div-x4.c | 68 const float vs3 = fp32_from_bits(xnn_table_exp2minus_k_over_2048[vidx3] + ve3); in xnn_f32_vsigmoid_ukernel__scalar_rr2_lut2048_p1_div_x4() local
|
D | vsigmoid-scalar-rr2-lut64-p2-div-x4.c | 68 const float vs3 = fp32_from_bits(xnn_table_exp2minus_k_over_64[vidx3] + ve3); in xnn_f32_vsigmoid_ukernel__scalar_rr2_lut64_p2_div_x4() local
|
D | vsigmoid-avx2-rr1-p5-div-x32.c | 58 const __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn3), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x32() local
|
/external/XNNPACK/src/f32-vscaleextexp/gen/ |
D | avx2-p5-x32.c | 131 …const __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve3, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x32() local
|
D | avx2-p5-x40.c | 143 …const __m256 vs3 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve3, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x40() local
|