Home
last modified time | relevance | path

Searched refs:vp5 (Results 1 – 25 of 177) sorted by relevance

12345678

/external/XNNPACK/src/f32-velu/gen/
Dvelu-wasm-rr2-p6-x6.c99 float vp5 = vc6 * vt5 + vc5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6() local
106 vp5 = vp5 * vt5 + vc4; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6()
113 vp5 = vp5 * vt5 + vc3; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6()
120 vp5 = vp5 * vt5 + vc2; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6()
127 vp5 *= vt5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6()
147 vp5 = vp5 * vt5 + vt5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6()
159 const float ve5 = (vp5 + vs5) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x6()
Dvelu-scalar-rr2-p6-x6.c123 float vp5 = vc6 * vt5 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
130 vp5 = vp5 * vt5 + vc4; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
137 vp5 = vp5 * vt5 + vc3; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
144 vp5 = vp5 * vt5 + vc2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
151 vp5 *= vt5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
171 vp5 = vp5 * vt5 + vt5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
183 const float ve5 = (vp5 + vs5) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
Dvelu-avx512f-rr1-p6-x96.c90 __m512 vp5 = _mm512_fmadd_ps(vc6, vt5, vc5); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96() local
97 vp5 = _mm512_fmadd_ps(vp5, vt5, vc4); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
104 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
111 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
123 vp5 = _mm512_mul_ps(vp5, vt5); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
138 vp5 = _mm512_fmadd_ps(vp5, vt5, vt5); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
151 __m512 vy5 = _mm512_fmadd_ps(vp5, valpha, vs5); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96()
Dvelu-avx2-rr1-p6-x48.c90 __m256 vp5 = _mm256_fmadd_ps(vc6, vt5, vc5); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48() local
97 vp5 = _mm256_fmadd_ps(vp5, vt5, vc4); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48()
104 vp5 = _mm256_fmadd_ps(vp5, vt5, vc3); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48()
111 vp5 = _mm256_fmadd_ps(vp5, vt5, vc2); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48()
123 vp5 = _mm256_mul_ps(vp5, vt5); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48()
137 vp5 = _mm256_fmadd_ps(vp5, vt5, vt5); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48()
149 const __m256 ve5 = _mm256_fmadd_ps(vp5, valpha, vs5); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48()
/external/XNNPACK/src/f32-vscaleextexp/gen/
Davx512f-p5-scalef-x96.c82 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96() local
89 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96()
96 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96()
103 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96()
110 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96()
123 __m512 vf5 = _mm512_mul_ps(vp5, vscalev); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x96()
Davx512f-p5-scalef-x112.c86 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112() local
94 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112()
102 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112()
110 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112()
118 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112()
132 __m512 vf5 = _mm512_mul_ps(vp5, vscalev); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x112()
Davx512f-p5-scalef-x128.c90 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128() local
99 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128()
108 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128()
117 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128()
126 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128()
141 __m512 vf5 = _mm512_mul_ps(vp5, vscalev); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x128()
Davx2-p5-x48.c88 __m256 vp5 = _mm256_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48() local
95 vp5 = _mm256_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48()
102 vp5 = _mm256_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48()
109 vp5 = _mm256_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48()
116 vp5 = _mm256_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48()
129 __m256 vf5 = _mm256_mul_ps(vp5, vscalev); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48()
Davx512f-p5-scalef-x144.c94 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144() local
104 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
114 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
124 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
134 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
150 __m512 vf5 = _mm512_mul_ps(vp5, vscalev); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/
Davx512f-p5-scalef-x96.c89 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96() local
96 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96()
103 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96()
110 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96()
117 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96()
127 __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x96()
Davx512f-p5-scalef-x112.c94 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112() local
102 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112()
110 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112()
118 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112()
126 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112()
137 __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x112()
Davx512f-p5-scalef-x128.c99 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x128() local
108 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x128()
117 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x128()
126 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x128()
135 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x128()
147 __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x128()
Davx512f-p5-scalef-x144.c104 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x144() local
114 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x144()
124 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x144()
134 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x144()
144 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x144()
157 __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x144()
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/
Davx512f-p5-scalef-x128-acc2.c100 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2() local
109 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
118 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
127 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
136 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
148 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
Davx512f-p5-scalef-x128.c99 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128() local
108 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128()
117 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128()
126 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128()
135 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128()
147 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x128()
Davx512f-p5-scalef-x144.c104 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144() local
114 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
124 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
134 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
144 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
157 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
/external/XNNPACK/src/f32-raddexpminusmax/gen/
Davx512f-p5-scalef-x128.c98 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128() local
107 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128()
116 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128()
125 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128()
134 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128()
146 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128()
Davx512f-p5-scalef-x128-acc2.c99 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2() local
108 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
117 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
126 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
135 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
147 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc2()
Davx512f-p5-scalef-x128-acc4.c101 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc4() local
110 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc4()
119 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc4()
128 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc4()
137 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc4()
149 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x128_acc4()
Davx512f-p5-scalef-x144.c103 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144() local
113 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
123 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
133 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
143 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
156 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
Davx512f-p5-scalef-x144-acc3.c105 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3() local
115 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
125 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
135 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
145 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
158 const __m512 vf5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
/external/XNNPACK/src/f32-sigmoid/gen/
Davx512f-rr1-p5-scalef-div-x96.c79 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() local
86 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96()
93 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96()
100 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96()
107 vp5 = _mm512_fmadd_ps(vp5, vt5, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96()
114 const __m512 ve5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96()
Davx512f-rr1-p5-scalef-div-x112.c84 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() local
92 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
100 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
108 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
116 vp5 = _mm512_fmadd_ps(vp5, vt5, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
124 const __m512 ve5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
Davx512f-rr1-p5-scalef-nr1fma-x96.c79 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96() local
86 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96()
93 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96()
100 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96()
107 vp5 = _mm512_fmadd_ps(vp5, vt5, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96()
114 const __m512 ve5 = _mm512_scalef_ps(vp5, vn5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x96()
/external/XNNPACK/src/f32-raddextexp/gen/
Davx512f-p5-scalef-x128-acc2.c92 __m512 vp5 = _mm512_fmadd_ps(vc5, vt5, vc4); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x128_acc2() local
101 vp5 = _mm512_fmadd_ps(vp5, vt5, vc3); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x128_acc2()
110 vp5 = _mm512_fmadd_ps(vp5, vt5, vc2); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x128_acc2()
119 vp5 = _mm512_fmadd_ps(vp5, vt5, vc1); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x128_acc2()
128 vp5 = _mm512_fmadd_ps(vp5, vt5, vc0); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x128_acc2()
170 vaccv1 = _mm512_add_ps(vaccv1, _mm512_scalef_ps(vp5, vdelta_e5)); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x128_acc2()

12345678