Searched refs:vxb01 (Results 1 – 9 of 9) sorted by relevance
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 3x8c8-minmax-avx2.c | 92 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2() local 94 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2() 95 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2() 96 vacc2x01 = _mm256_add_epi32(vacc2x01, _mm256_madd_epi16(vxa2, vxb01)); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2()
|
D | 3x8c8-xw-minmax-avx2.c | 91 const __m256i vxb01 = _mm256_load_si256((const __m256i*) w); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2() local 93 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2() 94 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2() 95 vacc2x01 = _mm256_add_epi32(vacc2x01, _mm256_madd_epi16(vxa2, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2()
|
D | 2x8c8-xw-minmax-avx2.c | 78 const __m256i vxb01 = _mm256_load_si256((const __m256i*) w); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2() local 80 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2() 81 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2()
|
D | 2x8c8-minmax-avx2.c | 79 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2() local 81 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2() 82 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2()
|
D | 1x8c8-xw-minmax-avx2.c | 65 const __m256i vxb01 = _mm256_load_si256((const __m256i*) w); in xnn_qs8_gemm_xw_minmax_ukernel_1x8c8__avx2() local 67 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_xw_minmax_ukernel_1x8c8__avx2()
|
D | 1x8c8-minmax-avx2.c | 66 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_gemm_minmax_ukernel_1x8c8__avx2() local 68 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_gemm_minmax_ukernel_1x8c8__avx2()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x8c8-minmax-avx2.c | 107 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2() local 109 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2() 110 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2() 111 vacc2x01 = _mm256_add_epi32(vacc2x01, _mm256_madd_epi16(vxa2, vxb01)); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2()
|
D | 2x8c8-minmax-avx2.c | 92 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2() local 94 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2() 95 vacc1x01 = _mm256_add_epi32(vacc1x01, _mm256_madd_epi16(vxa1, vxb01)); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2()
|
D | 1x8c8-minmax-avx2.c | 77 const __m256i vxb01 = _mm256_cvtepi8_epi16(vb01); in xnn_qs8_igemm_minmax_ukernel_1x8c8__avx2() local 79 vacc0x01 = _mm256_add_epi32(vacc0x01, _mm256_madd_epi16(vxa0, vxb01)); in xnn_qs8_igemm_minmax_ukernel_1x8c8__avx2()
|