Lines Matching refs:vt
55 __m128 vt = _mm_add_ps(_mm_mul_ps(vn, vminus_ln2_hi), vz); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4() local
56 vt = _mm_add_ps(_mm_mul_ps(vn, vminus_ln2_lo), vt); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
58 __m128 vp = _mm_add_ps(_mm_mul_ps(vc6, vt), vc5); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
59 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vc4); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
60 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vc3); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
61 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vc2); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
62 vp = _mm_mul_ps(vp, vt); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
64 vt = _mm_mul_ps(vt, vs); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
66 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vt); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
84 __m128 vt = _mm_add_ps(_mm_mul_ps(vn, vminus_ln2_hi), vz); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4() local
85 vt = _mm_add_ps(_mm_mul_ps(vn, vminus_ln2_lo), vt); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
87 __m128 vp = _mm_add_ps(_mm_mul_ps(vc6, vt), vc5); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
88 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vc4); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
89 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vc3); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
90 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vc2); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
91 vp = _mm_mul_ps(vp, vt); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
93 vt = _mm_mul_ps(vt, vs); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()
95 vp = _mm_add_ps(_mm_mul_ps(vp, vt), vt); in xnn_f32_velu_ukernel__sse41_rr2_p6_x4()