• Home
  • Raw
  • Download

Lines Matching refs:_mm_clmulepi64_si128

118 +    *xmm_crc0 = _mm_clmulepi64_si128(*xmm_crc0, xmm_fold4, 0x01);
119 + *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x10);
145 + *xmm_crc1 = _mm_clmulepi64_si128(*xmm_crc1, xmm_fold4, 0x01);
146 + *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x10);
152 + *xmm_crc0 = _mm_clmulepi64_si128(*xmm_crc0, xmm_fold4, 0x01);
153 + *xmm_crc2 = _mm_clmulepi64_si128(*xmm_crc2, xmm_fold4, 0x10);
178 + *xmm_crc2 = _mm_clmulepi64_si128(*xmm_crc2, xmm_fold4, 0x01);
179 + *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x10);
185 + *xmm_crc1 = _mm_clmulepi64_si128(*xmm_crc1, xmm_fold4, 0x01);
186 + *xmm_crc2 = _mm_clmulepi64_si128(*xmm_crc2, xmm_fold4, 0x10);
192 + *xmm_crc0 = _mm_clmulepi64_si128(*xmm_crc0, xmm_fold4, 0x01);
193 + *xmm_crc1 = _mm_clmulepi64_si128(*xmm_crc1, xmm_fold4, 0x10);
222 + *xmm_crc0 = _mm_clmulepi64_si128(*xmm_crc0, xmm_fold4, 0x01);
223 + x_tmp0 = _mm_clmulepi64_si128(x_tmp0, xmm_fold4, 0x10);
228 + *xmm_crc1 = _mm_clmulepi64_si128(*xmm_crc1, xmm_fold4, 0x01);
229 + x_tmp1 = _mm_clmulepi64_si128(x_tmp1, xmm_fold4, 0x10);
234 + *xmm_crc2 = _mm_clmulepi64_si128(*xmm_crc2, xmm_fold4, 0x01);
235 + x_tmp2 = _mm_clmulepi64_si128(x_tmp2, xmm_fold4, 0x10);
240 + *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x01);
241 + x_tmp3 = _mm_clmulepi64_si128(x_tmp3, xmm_fold4, 0x10);
307 + xmm_a0_1 = _mm_clmulepi64_si128(xmm_a0_0, xmm_fold4, 0x10);
308 + xmm_a0_0 = _mm_clmulepi64_si128(xmm_a0_0, xmm_fold4, 0x01);
491 + x_tmp0 = _mm_clmulepi64_si128(xmm_crc0, crc_fold, 0x10);
492 + xmm_crc0 = _mm_clmulepi64_si128(xmm_crc0, crc_fold, 0x01);
496 + x_tmp1 = _mm_clmulepi64_si128(xmm_crc1, crc_fold, 0x10);
497 + xmm_crc1 = _mm_clmulepi64_si128(xmm_crc1, crc_fold, 0x01);
501 + x_tmp2 = _mm_clmulepi64_si128(xmm_crc2, crc_fold, 0x10);
502 + xmm_crc2 = _mm_clmulepi64_si128(xmm_crc2, crc_fold, 0x01);
512 + xmm_crc3 = _mm_clmulepi64_si128(xmm_crc3, crc_fold, 0);
518 + xmm_crc3 = _mm_clmulepi64_si128(xmm_crc3, crc_fold, 0x10);
529 + xmm_crc3 = _mm_clmulepi64_si128(xmm_crc3, crc_fold, 0);
534 + xmm_crc3 = _mm_clmulepi64_si128(xmm_crc3, crc_fold, 0x10);