Home
last modified time | relevance | path

Searched refs:va1c1111 (Results 1 – 18 of 18) sorted by relevance

/external/XNNPACK/src/f32-gemm/gen/
D3x8-minmax-sse-dup.c87 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_gemm_minmax_ukernel_3x8__sse_dup() local
94 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemm_minmax_ukernel_3x8__sse_dup()
97 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemm_minmax_ukernel_3x8__sse_dup()
D3x8-minmax-sse2-dup.c87 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_gemm_minmax_ukernel_3x8__sse2_dup() local
94 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemm_minmax_ukernel_3x8__sse2_dup()
97 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemm_minmax_ukernel_3x8__sse2_dup()
D4x8-minmax-sse2-dup.c100 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_gemm_minmax_ukernel_4x8__sse2_dup() local
108 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemm_minmax_ukernel_4x8__sse2_dup()
112 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemm_minmax_ukernel_4x8__sse2_dup()
D4x8-minmax-sse-dup.c100 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_gemm_minmax_ukernel_4x8__sse_dup() local
108 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemm_minmax_ukernel_4x8__sse_dup()
112 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemm_minmax_ukernel_4x8__sse_dup()
D5x8-minmax-sse-dup.c113 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_gemm_minmax_ukernel_5x8__sse_dup() local
122 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemm_minmax_ukernel_5x8__sse_dup()
127 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemm_minmax_ukernel_5x8__sse_dup()
D5x8-minmax-sse2-dup.c113 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_gemm_minmax_ukernel_5x8__sse2_dup() local
122 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemm_minmax_ukernel_5x8__sse2_dup()
127 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemm_minmax_ukernel_5x8__sse2_dup()
/external/XNNPACK/src/f32-gemm/gen-inc/
D3x8inc-minmax-sse-dup.c89 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_gemminc_minmax_ukernel_3x8__sse_dup() local
96 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemminc_minmax_ukernel_3x8__sse_dup()
99 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemminc_minmax_ukernel_3x8__sse_dup()
D3x8inc-minmax-sse2-dup.c89 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_gemminc_minmax_ukernel_3x8__sse2_dup() local
96 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemminc_minmax_ukernel_3x8__sse2_dup()
99 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemminc_minmax_ukernel_3x8__sse2_dup()
D4x8inc-minmax-sse-dup.c102 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_gemminc_minmax_ukernel_4x8__sse_dup() local
110 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemminc_minmax_ukernel_4x8__sse_dup()
114 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemminc_minmax_ukernel_4x8__sse_dup()
D4x8inc-minmax-sse2-dup.c102 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_gemminc_minmax_ukernel_4x8__sse2_dup() local
110 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemminc_minmax_ukernel_4x8__sse2_dup()
114 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemminc_minmax_ukernel_4x8__sse2_dup()
D5x8inc-minmax-sse2-dup.c115 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_gemminc_minmax_ukernel_5x8__sse2_dup() local
124 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemminc_minmax_ukernel_5x8__sse2_dup()
129 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemminc_minmax_ukernel_5x8__sse2_dup()
D5x8inc-minmax-sse-dup.c115 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_gemminc_minmax_ukernel_5x8__sse_dup() local
124 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_gemminc_minmax_ukernel_5x8__sse_dup()
129 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_gemminc_minmax_ukernel_5x8__sse_dup()
/external/XNNPACK/src/f32-igemm/gen/
D3x8-minmax-sse-dup.c106 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_igemm_minmax_ukernel_3x8__sse_dup() local
113 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_igemm_minmax_ukernel_3x8__sse_dup()
116 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_igemm_minmax_ukernel_3x8__sse_dup()
D3x8-minmax-sse2-dup.c106 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_igemm_minmax_ukernel_3x8__sse2_dup() local
113 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_igemm_minmax_ukernel_3x8__sse2_dup()
116 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_igemm_minmax_ukernel_3x8__sse2_dup()
D4x8-minmax-sse-dup.c122 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_igemm_minmax_ukernel_4x8__sse_dup() local
130 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_igemm_minmax_ukernel_4x8__sse_dup()
134 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_igemm_minmax_ukernel_4x8__sse_dup()
D4x8-minmax-sse2-dup.c122 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_igemm_minmax_ukernel_4x8__sse2_dup() local
130 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_igemm_minmax_ukernel_4x8__sse2_dup()
134 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_igemm_minmax_ukernel_4x8__sse2_dup()
D5x8-minmax-sse-dup.c138 const __m128 va1c1111 = _mm_shuffle_ps(va1, va1, _MM_SHUFFLE(1, 1, 1, 1)); in xnn_f32_igemm_minmax_ukernel_5x8__sse_dup() local
147 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_igemm_minmax_ukernel_5x8__sse_dup()
152 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_igemm_minmax_ukernel_5x8__sse_dup()
D5x8-minmax-sse2-dup.c138 …const __m128 va1c1111 = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(va1), _MM_SHUFFLE(1, 1… in xnn_f32_igemm_minmax_ukernel_5x8__sse2_dup() local
147 vacc1x0123 = _mm_add_ps(vacc1x0123, _mm_mul_ps(va1c1111, vb0123c1)); in xnn_f32_igemm_minmax_ukernel_5x8__sse2_dup()
152 vacc1x4567 = _mm_add_ps(vacc1x4567, _mm_mul_ps(va1c1111, vb4567c1)); in xnn_f32_igemm_minmax_ukernel_5x8__sse2_dup()