Home
last modified time | relevance | path

Searched refs:vxb01 (Results 1 – 9 of 9) sorted by relevance

/external/XNNPACK/src/qs8-gemm/gen/
D3x8c8-minmax-avx2.c92 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2() local
94 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2()
95 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2()
96 vacc2x01 = _mm256_add_epi32(vacc2x01, _mm256_madd_epi16(vxa2, vxb01)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2()
D3x8c8-xw-minmax-avx2.c91 const __m256i vxb01 = _mm256_load_si256((const __m256i*) w); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2() local
93 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2()
94 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2()
95 vacc2x01 = _mm256_add_epi32(vacc2x01, _mm256_madd_epi16(vxa2, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2()
D2x8c8-xw-minmax-avx2.c78 const __m256i vxb01 = _mm256_load_si256((const __m256i*) w); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2() local
80 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2()
81 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2()
D2x8c8-minmax-avx2.c79 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2() local
81 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2()
82 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2()
D1x8c8-xw-minmax-avx2.c65 const __m256i vxb01 = _mm256_load_si256((const __m256i*) w); in xnn_qs8_gemm_xw_minmax_ukernel_1x8c8__avx2() local
67 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_1x8c8__avx2()
D1x8c8-minmax-avx2.c66 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_gemm_minmax_ukernel_1x8c8__avx2() local
68 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_minmax_ukernel_1x8c8__avx2()
/external/XNNPACK/src/qs8-igemm/gen/
D3x8c8-minmax-avx2.c107 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2() local
109 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2()
110 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2()
111 vacc2x01 = _mm256_add_epi32(vacc2x01, _mm256_madd_epi16(vxa2, vxb01)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2()
D2x8c8-minmax-avx2.c92 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2() local
94 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2()
95 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2()
D1x8c8-minmax-avx2.c77 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_igemm_minmax_ukernel_1x8c8__avx2() local
79 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_igemm_minmax_ukernel_1x8c8__avx2()