Lines Matching refs:_mm_mulhi_epi16
80 const __m128i vprod0_even = _mm_mulhi_epi16(vxi0, vxk0); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
89 const __m128i vprod1_even = _mm_mulhi_epi16(vxi1, vxk1); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
98 const __m128i vprod2_even = _mm_mulhi_epi16(vxi2, vxk2); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
107 const __m128i vprod3_even = _mm_mulhi_epi16(vxi3, vxk3); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
116 const __m128i vprod4_even = _mm_mulhi_epi16(vxi4, vxk4); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
125 const __m128i vprod5_even = _mm_mulhi_epi16(vxi5, vxk5); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
134 const __m128i vprod6_even = _mm_mulhi_epi16(vxi6, vxk6); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
143 const __m128i vprod7_even = _mm_mulhi_epi16(vxi7, vxk7); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
152 const __m128i vprod8_even = _mm_mulhi_epi16(vxi8, vxk8); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
232 const __m128i vprod0_even = _mm_mulhi_epi16(vxi0, vxk0); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
241 const __m128i vprod1_even = _mm_mulhi_epi16(vxi1, vxk1); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
250 const __m128i vprod2_even = _mm_mulhi_epi16(vxi2, vxk2); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
259 const __m128i vprod3_even = _mm_mulhi_epi16(vxi3, vxk3); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
268 const __m128i vprod4_even = _mm_mulhi_epi16(vxi4, vxk4); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
277 const __m128i vprod5_even = _mm_mulhi_epi16(vxi5, vxk5); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
286 const __m128i vprod6_even = _mm_mulhi_epi16(vxi6, vxk6); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
295 const __m128i vprod7_even = _mm_mulhi_epi16(vxi7, vxk7); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()
304 const __m128i vprod8_even = _mm_mulhi_epi16(vxi8, vxk8); in xnn_qu8_dwconv_minmax_ukernel_up8x9__sse2()