Home
last modified time | relevance | path

Searched refs:a (Results 1 – 25 of 32) sorted by relevance

12

/crypto/
Dmemneq.c68 __crypto_memneq_generic(const void *a, const void *b, size_t size) in __crypto_memneq_generic() argument
74 neq |= *(unsigned long *)a ^ *(unsigned long *)b; in __crypto_memneq_generic()
76 a += sizeof(unsigned long); in __crypto_memneq_generic()
82 neq |= *(unsigned char *)a ^ *(unsigned char *)b; in __crypto_memneq_generic()
84 a += 1; in __crypto_memneq_generic()
92 static inline unsigned long __crypto_memneq_16(const void *a, const void *b) in __crypto_memneq_16() argument
98 neq |= *(unsigned long *)(a) ^ *(unsigned long *)(b); in __crypto_memneq_16()
100 neq |= *(unsigned long *)(a+8) ^ *(unsigned long *)(b+8); in __crypto_memneq_16()
103 neq |= *(unsigned int *)(a) ^ *(unsigned int *)(b); in __crypto_memneq_16()
105 neq |= *(unsigned int *)(a+4) ^ *(unsigned int *)(b+4); in __crypto_memneq_16()
[all …]
Dmd4.c63 #define ROUND1(a,b,c,d,k,s) (a = lshift(a + F(b,c,d) + k, s)) argument
64 #define ROUND2(a,b,c,d,k,s) (a = lshift(a + G(b,c,d) + k + (u32)0x5A827999,s)) argument
65 #define ROUND3(a,b,c,d,k,s) (a = lshift(a + H(b,c,d) + k + (u32)0x6ED9EBA1,s)) argument
69 u32 a, b, c, d; in md4_transform() local
71 a = hash[0]; in md4_transform()
76 ROUND1(a, b, c, d, in[0], 3); in md4_transform()
77 ROUND1(d, a, b, c, in[1], 7); in md4_transform()
78 ROUND1(c, d, a, b, in[2], 11); in md4_transform()
79 ROUND1(b, c, d, a, in[3], 19); in md4_transform()
80 ROUND1(a, b, c, d, in[4], 3); in md4_transform()
[all …]
Dmd5.c45 u32 a, b, c, d; in md5_transform() local
47 a = hash[0]; in md5_transform()
52 MD5STEP(F1, a, b, c, d, in[0] + 0xd76aa478, 7); in md5_transform()
53 MD5STEP(F1, d, a, b, c, in[1] + 0xe8c7b756, 12); in md5_transform()
54 MD5STEP(F1, c, d, a, b, in[2] + 0x242070db, 17); in md5_transform()
55 MD5STEP(F1, b, c, d, a, in[3] + 0xc1bdceee, 22); in md5_transform()
56 MD5STEP(F1, a, b, c, d, in[4] + 0xf57c0faf, 7); in md5_transform()
57 MD5STEP(F1, d, a, b, c, in[5] + 0x4787c62a, 12); in md5_transform()
58 MD5STEP(F1, c, d, a, b, in[6] + 0xa8304613, 17); in md5_transform()
59 MD5STEP(F1, b, c, d, a, in[7] + 0xfd469501, 22); in md5_transform()
[all …]
Dtwofish_generic.c40 #define G1(a) \ argument
41 (ctx->s[0][(a) & 0xFF]) ^ (ctx->s[1][((a) >> 8) & 0xFF]) \
42 ^ (ctx->s[2][((a) >> 16) & 0xFF]) ^ (ctx->s[3][(a) >> 24])
53 #define ENCROUND(n, a, b, c, d) \ argument
54 x = G1 (a); y = G2 (b); \
60 #define DECROUND(n, a, b, c, d) \ argument
61 x = G1 (a); y = G2 (b); \
72 ENCROUND (2 * (n), a, b, c, d); \
73 ENCROUND (2 * (n) + 1, c, d, a, b)
76 DECROUND (2 * (n) + 1, c, d, a, b); \
[all …]
Dtgr192.c34 u64 a, b, c; member
399 u64 a = *ra; in tgr192_round() local
404 a -= sbox1[c & 0xff] ^ sbox2[(c >> 16) & 0xff] in tgr192_round()
410 *ra = a; in tgr192_round()
418 u64 a = *ra; in tgr192_pass() local
422 tgr192_round(&a, &b, &c, x[0], mul); in tgr192_pass()
423 tgr192_round(&b, &c, &a, x[1], mul); in tgr192_pass()
424 tgr192_round(&c, &a, &b, x[2], mul); in tgr192_pass()
425 tgr192_round(&a, &b, &c, x[3], mul); in tgr192_pass()
426 tgr192_round(&b, &c, &a, x[4], mul); in tgr192_pass()
[all …]
Dtwofish_common.c468 #define CALC_S(a, b, c, d, i, w, x, y, z) \ argument
471 (a) ^= exp_to_poly[tmp + (w)]; \
483 #define CALC_SB_2(i, a, b) \ argument
484 ctx->s[0][i] = mds[0][q0[(a) ^ sa] ^ se]; \
486 ctx->s[2][i] = mds[2][q1[(a) ^ sc] ^ sg]; \
491 #define CALC_SB192_2(i, a, b) \ argument
494 ctx->s[2][i] = mds[2][q1[q0[(a) ^ sc] ^ sg] ^ sk]; \
495 ctx->s[3][i] = mds[3][q1[q1[(a) ^ sd] ^ sh] ^ sl];
499 #define CALC_SB256_2(i, a, b) \ argument
501 ctx->s[1][i] = mds[1][q0[q1[q1[(a) ^ sb] ^ sf] ^ sj] ^ sn]; \
[all …]
Dgf128mul.c141 u64 a = be64_to_cpu(x->a); in gf128mul_x8_lle() local
145 x->b = cpu_to_be64((b >> 8) | (a << 56)); in gf128mul_x8_lle()
146 x->a = cpu_to_be64((a >> 8) ^ (_tt << 48)); in gf128mul_x8_lle()
151 u64 a = be64_to_cpu(x->a); in gf128mul_x8_bbe() local
153 u64 _tt = gf128mul_table_be[a >> 56]; in gf128mul_x8_bbe()
155 x->a = cpu_to_be64((a << 8) | (b >> 56)); in gf128mul_x8_bbe()
161 u64 a = le64_to_cpu(x->a); in gf128mul_x8_ble() local
163 u64 _tt = gf128mul_table_be[a >> 56]; in gf128mul_x8_ble()
165 r->a = cpu_to_le64((a << 8) | (b >> 56)); in gf128mul_x8_ble()
312 void gf128mul_64k_bbe(be128 *a, const struct gf128mul_64k *t) in gf128mul_64k_bbe() argument
[all …]
Dsm3_generic.c39 static inline u32 ff(unsigned int n, u32 a, u32 b, u32 c) in ff() argument
41 return (n < 16) ? (a ^ b ^ c) : ((a & b) | (a & c) | (b & c)); in ff()
78 u32 a, b, c, d, e, f, g, h; in sm3_compress() local
81 a = m[0]; in sm3_compress()
92 ss1 = rol32((rol32(a, 12) + e + rol32(t(i), i & 31)), 7); in sm3_compress()
94 ss2 = ss1 ^ rol32(a, 12); in sm3_compress()
96 tt1 = ff(i, a, b, c) + d + ss2 + *wt; in sm3_compress()
104 b = a; in sm3_compress()
105 a = tt1; in sm3_compress()
112 m[0] = a ^ m[0]; in sm3_compress()
[all …]
Dsha512_generic.c101 u64 a, b, c, d, e, f, g, h, t1, t2; in sha512_transform() local
107 a=state[0]; b=state[1]; c=state[2]; d=state[3]; in sha512_transform()
127 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; in sha512_transform()
129 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; in sha512_transform()
131 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; in sha512_transform()
133 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; in sha512_transform()
134 t1 = d + e1(a) + Ch(a,b,c) + sha512_K[i+4] + W[(i & 15) + 4]; in sha512_transform()
136 t1 = c + e1(h) + Ch(h,a,b) + sha512_K[i+5] + W[(i & 15) + 5]; in sha512_transform()
138 t1 = b + e1(g) + Ch(g,h,a) + sha512_K[i+6] + W[(i & 15) + 6]; in sha512_transform()
140 t1 = a + e1(f) + Ch(f,g,h) + sha512_K[i+7] + W[(i & 15) + 7]; in sha512_transform()
[all …]
Decrdsa_defs.h57 .a = cp256a_a,
90 .a = cp256b_a,
127 .a = cp256c_a,
176 .a = tc512a_a,
221 .a = tc512b_a,
DKconfig31 required if you want the system to operate in a FIPS 200
182 This converts an arbitrary crypto algorithm into a parallel
191 This is a generic software asynchronous crypto daemon that
193 into an asynchronous algorithm that executes in a kernel thread.
331 This IV generator generates an IV based on a sequence number by
332 xoring it with a salt. This algorithm is mainly useful for CTR
342 a sequence number xored with a salt. This is the default
385 See: https://csrc.nist.gov/publications/detail/sp/800-38a/addendum/final
402 LRW: Liskov Rivest Wagner, a tweakable, non malleable, non movable
413 OFB: the Output Feedback mode makes a block cipher into a synchronous
[all …]
Dechainiv.c72 u64 a; in echainiv_encrypt() local
74 memcpy(&a, ctx->salt + ivsize - 8, 8); in echainiv_encrypt()
76 a |= 1; in echainiv_encrypt()
77 a *= seqno; in echainiv_encrypt()
79 memcpy(info + ivsize - 8, &a, 8); in echainiv_encrypt()
Decc_curve_defs.h27 .a = nist_p192_a,
53 .a = nist_p256_a,
Dadiantum.c200 r->a = cpu_to_le64(le64_to_cpu(v1->a) + le64_to_cpu(v2->a) + in le128_add()
211 r->a = cpu_to_le64(le64_to_cpu(v1->a) - le64_to_cpu(v2->a) - in le128_sub()
Decc.c367 static uint128_t add_128_128(uint128_t a, uint128_t b) in add_128_128() argument
371 result.m_low = a.m_low + b.m_low; in add_128_128()
372 result.m_high = a.m_high + b.m_high + (result.m_low < a.m_low); in add_128_128()
861 u64 a[ECC_MAX_DIGITS], b[ECC_MAX_DIGITS]; in vli_mod_inv() local
871 vli_set(a, input, ndigits); in vli_mod_inv()
877 while ((cmp_result = vli_cmp(a, b, ndigits)) != 0) { in vli_mod_inv()
880 if (EVEN(a)) { in vli_mod_inv()
881 vli_rshift1(a, ndigits); in vli_mod_inv()
899 vli_sub(a, a, b, ndigits); in vli_mod_inv()
900 vli_rshift1(a, ndigits); in vli_mod_inv()
[all …]
Dblowfish_common.c304 #define ROUND(a, b, n) ({ b ^= P[n]; a ^= bf_F(b); }) argument
Dblowfish_generic.c34 #define ROUND(a, b, n) ({ b ^= P[n]; a ^= bf_F(b); }) argument
Dskcipher.c229 unsigned a; in skcipher_next_slow() local
243 a = crypto_tfm_ctx_alignment() - 1; in skcipher_next_slow()
248 a &= (sizeof(*p) ^ (sizeof(*p) - 1)) >> 1; in skcipher_next_slow()
253 n += alignmask & ~a; in skcipher_next_slow()
256 n += (bsize - 1) & ~(alignmask | a); in skcipher_next_slow()
398 unsigned a = crypto_tfm_ctx_alignment() - 1; in skcipher_copy_iv() local
409 size = alignmask & ~a; in skcipher_copy_iv()
417 size += (bs - 1) & ~(alignmask | a); in skcipher_copy_iv()
Decc.h69 u64 *a; member
Drmd256.c40 #define ROUND(a, b, c, d, f, k, x, s) { \ argument
41 (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \
42 (a) = rol32((a), (s)); \
Drmd128.c40 #define ROUND(a, b, c, d, f, k, x, s) { \ argument
41 (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \
42 (a) = rol32((a), (s)); \
/crypto/asymmetric_keys/
DKconfig6 This option provides support for a key type that holds the data for
44 data and provides the ability to instantiate a crypto key from a
54 private key data and provides the ability to instantiate a crypto key
63 private key data and provides the ability to instantiate a crypto key
80 This option provides a type of key that can be loaded up from a
81 PKCS#7 message - provided the message is signed by a trusted key. If
96 This option provides support for verifying the signature(s) on a
Dx509_public_key.c116 bool a = asymmetric_key_id_same(cert->skid, cert->sig->auth_ids[1]); in x509_check_for_self_signed() local
119 if (!a && !b) in x509_check_for_self_signed()
123 if (((a && !b) || (b && !a)) && in x509_check_for_self_signed()
Dverify_pefile.c204 static int pefile_compare_shdrs(const void *a, const void *b) in pefile_compare_shdrs() argument
206 const struct section_header *shdra = a; in pefile_compare_shdrs()
/crypto/async_tx/
Dasync_raid6_recov.c27 u8 *a, *b, *c; in async_sum_product() local
69 a = page_address(srcs[0]); in async_sum_product()
74 ax = amul[*a++]; in async_sum_product()
150 struct page *p, *q, *a, *b; in __2data_recov_4() local
161 a = blocks[faila]; in __2data_recov_4()
178 tx = async_xor(a, srcs, 0, 2, bytes, submit); in __2data_recov_4()

12