| /external/XNNPACK/src/x8-lut/gen/ |
| D | lut-scalar-x8.c | 31 const size_t vx4 = (size_t) x[4]; in xnn_x8_lut_ukernel__scalar_x8() local
|
| D | lut-scalar-x16.c | 31 const size_t vx4 = (size_t) x[4]; in xnn_x8_lut_ukernel__scalar_x16() local
|
| /external/XNNPACK/src/f32-vsqrt/gen/ |
| D | fma3-nr1fma1adj-x40.c | 33 const __m256 vx4 = _mm256_loadu_ps(x + 32); in xnn_f32_vsqrt_ukernel__fma3_nr1fma1adj_x40() local
|
| D | avx512f-nr1fma1adj-x80.c | 34 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vsqrt_ukernel__avx512f_nr1fma1adj_x80() local
|
| D | avx512f-nr1fma1adj-x96.c | 34 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vsqrt_ukernel__avx512f_nr1fma1adj_x96() local
|
| D | fma3-nr1fma1adj-x48.c | 33 const __m256 vx4 = _mm256_loadu_ps(x + 32); in xnn_f32_vsqrt_ukernel__fma3_nr1fma1adj_x48() local
|
| D | fma3-nr1fma1adj-x56.c | 33 const __m256 vx4 = _mm256_loadu_ps(x + 32); in xnn_f32_vsqrt_ukernel__fma3_nr1fma1adj_x56() local
|
| D | avx512f-nr1fma1adj-x112.c | 34 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vsqrt_ukernel__avx512f_nr1fma1adj_x112() local
|
| /external/XNNPACK/src/f32-vscaleextexp/gen/ |
| D | avx512f-p5-scalef-x80.c | 48 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x80() local
|
| D | avx512f-p5-scalef-x96.c | 48 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96() local
|
| D | avx2-p5-x40.c | 54 const __m256 vx4 = _mm256_loadu_ps(x + 32); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x40() local
|
| D | avx512f-p5-scalef-x112.c | 48 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112() local
|
| D | avx512f-p5-scalef-x128.c | 48 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128() local
|
| /external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
| D | avx512f-p5-scalef-x80.c | 55 const __m512 vx4 = _mm512_sub_ps(vi4, vi_max); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x80() local
|
| D | avx512f-p5-scalef-x96.c | 56 const __m512 vx4 = _mm512_sub_ps(vi4, vi_max); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96() local
|
| D | avx2-p5-x40.c | 59 const __m256 vx4 = _mm256_sub_ps(vi4, vi_max); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x40() local
|
| D | avx512f-p5-scalef-x112.c | 57 const __m512 vx4 = _mm512_sub_ps(vi4, vi_max); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112() local
|
| /external/XNNPACK/src/f32-velu/gen/ |
| D | velu-wasm-rr2-p6-x5.c | 47 float vx4 = x[4]; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() local
|
| D | velu-scalar-rr2-p6-x5.c | 47 float vx4 = x[4]; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local
|
| D | velu-wasm-rr2-p6-x6.c | 47 float vx4 = x[4]; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6() local
|
| D | velu-scalar-rr2-p6-x6.c | 47 float vx4 = x[4]; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
|
| D | velu-avx2-rr1-p6-x40.c | 44 __m256 vx4 = _mm256_loadu_ps(x + 32); in xnn_f32_velu_ukernel__avx2_rr1_p6_x40() local
|
| D | velu-avx512f-rr1-p6-x80.c | 46 __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x80() local
|
| /external/XNNPACK/src/f32-vsigmoid/gen/ |
| D | vsigmoid-avx512f-rr1-p5-scalef-div-x80.c | 42 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() local
|
| D | vsigmoid-avx512f-rr1-p5-scalef-div-x96.c | 42 const __m512 vx4 = _mm512_loadu_ps(x + 64); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() local
|