/crypto/ |
D | sha512_generic.c | 101 u64 a, b, c, d, e, f, g, h, t1, t2; in sha512_transform() local 108 e=state[4]; f=state[5]; g=state[6]; h=state[7]; in sha512_transform() 126 t1 = h + e1(e) + Ch(e,f,g) + sha512_K[i ] + W[(i & 15)]; in sha512_transform() 128 t1 = g + e1(d) + Ch(d,e,f) + sha512_K[i+1] + W[(i & 15) + 1]; in sha512_transform() 130 t1 = f + e1(c) + Ch(c,d,e) + sha512_K[i+2] + W[(i & 15) + 2]; in sha512_transform() 131 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; in sha512_transform() 133 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; in sha512_transform() 135 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; in sha512_transform() 137 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; in sha512_transform() 139 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; in sha512_transform() [all …]
|
D | sm3_generic.c | 44 static inline u32 gg(unsigned int n, u32 e, u32 f, u32 g) in gg() argument 46 return (n < 16) ? (e ^ f ^ g) : ((e & f) | ((~e) & g)); in gg() 78 u32 a, b, c, d, e, f, g, h; in sm3_compress() local 86 f = m[5]; in sm3_compress() 99 tt2 = gg(i, e, f, g) + h + ss1 + *w; in sm3_compress() 107 g = rol32(f, 19); in sm3_compress() 108 f = e; in sm3_compress() 117 m[5] = f ^ m[5]; in sm3_compress() 121 a = b = c = d = e = f = g = h = ss1 = ss2 = tt1 = tt2 = 0; in sm3_compress()
|
D | xor.c | 106 struct xor_block_template *f, *fastest; in calibrate_xor_blocks() local 134 for (f = fastest; f; f = f->next) in calibrate_xor_blocks() 135 if (f->speed > fastest->speed) in calibrate_xor_blocks() 136 fastest = f; in calibrate_xor_blocks()
|
D | poly1305_generic.c | 268 u64 f = 0; in crypto_poly1305_final() local 283 f = (f >> 32) + le32_to_cpu(digest[0]) + dctx->s[0]; in crypto_poly1305_final() 284 put_unaligned_le32(f, dst + 0); in crypto_poly1305_final() 285 f = (f >> 32) + le32_to_cpu(digest[1]) + dctx->s[1]; in crypto_poly1305_final() 286 put_unaligned_le32(f, dst + 4); in crypto_poly1305_final() 287 f = (f >> 32) + le32_to_cpu(digest[2]) + dctx->s[2]; in crypto_poly1305_final() 288 put_unaligned_le32(f, dst + 8); in crypto_poly1305_final() 289 f = (f >> 32) + le32_to_cpu(digest[3]) + dctx->s[3]; in crypto_poly1305_final() 290 put_unaligned_le32(f, dst + 12); in crypto_poly1305_final()
|
D | md5.c | 40 #define MD5STEP(f, w, x, y, z, in, s) \ argument 41 (w += f(x, y, z) + in, w = (w<<s | w>>(32-s)) + x)
|
D | rmd256.c | 40 #define ROUND(a, b, c, d, f, k, x, s) { \ argument 41 (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \
|
D | rmd128.c | 40 #define ROUND(a, b, c, d, f, k, x, s) { \ argument 41 (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \
|
D | rmd160.c | 43 #define ROUND(a, b, c, d, e, f, k, x, s) { \ argument 44 (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \
|
D | rmd320.c | 43 #define ROUND(a, b, c, d, e, f, k, x, s) { \ argument 44 (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \
|