/crypto/asymmetric_keys/ |
D | mscode_parser.c | 54 char buffer[50]; in mscode_note_content_type() local 56 sprint_oid(value, vlen, buffer, sizeof(buffer)); in mscode_note_content_type() 57 pr_err("Unknown OID: %s\n", buffer); in mscode_note_content_type() 83 char buffer[50]; in mscode_note_digest_algo() local 102 sprint_oid(value, vlen, buffer, sizeof(buffer)); in mscode_note_digest_algo() 103 pr_err("Unknown OID: %s\n", buffer); in mscode_note_digest_algo()
|
D | x509_cert_parser.c | 129 char buffer[50]; in x509_note_OID() local 130 sprint_oid(value, vlen, buffer, sizeof(buffer)); in x509_note_OID() 132 (unsigned long)value - ctx->data, buffer); in x509_note_OID() 280 char *buffer; in x509_fabricate_name() local 287 buffer = kmalloc(1, GFP_KERNEL); in x509_fabricate_name() 288 if (!buffer) in x509_fabricate_name() 290 buffer[0] = 0; in x509_fabricate_name() 309 buffer = kmalloc(ctx->o_size + 2 + ctx->cn_size + 1, in x509_fabricate_name() 311 if (!buffer) in x509_fabricate_name() 314 memcpy(buffer, in x509_fabricate_name() [all …]
|
D | pkcs7_parser.c | 171 char buffer[50]; in pkcs7_note_OID() local 172 sprint_oid(value, vlen, buffer, sizeof(buffer)); in pkcs7_note_OID() 174 (unsigned long)value - ctx->data, buffer); in pkcs7_note_OID()
|
/crypto/ |
D | wp512.c | 40 u8 buffer[WP512_BLOCK_SIZE]; member 788 const __be64 *buffer = (const __be64 *)wctx->buffer; in wp512_process_buffer() local 791 block[i] = be64_to_cpu(buffer[i]); in wp512_process_buffer() 995 wctx->buffer[0] = 0; in wp512_init() 1013 u8 *buffer = wctx->buffer; in wp512_update() local 1028 buffer[bufferPos++] |= (u8)(b >> bufferRem); in wp512_update() 1034 buffer[bufferPos] = b << (8 - bufferRem); in wp512_update() 1041 buffer[bufferPos] |= b >> bufferRem; in wp512_update() 1055 buffer[bufferPos] = b << (8 - bufferRem); in wp512_update() 1069 u8 *buffer = wctx->buffer; in wp512_final() local [all …]
|
D | cipher.c | 29 u8 *buffer, *alignbuffer; in setkey_unaligned() local 33 buffer = kmalloc(absize, GFP_ATOMIC); in setkey_unaligned() 34 if (!buffer) in setkey_unaligned() 37 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in setkey_unaligned() 41 kfree(buffer); in setkey_unaligned() 70 u8 buffer[size + alignmask]; in cipher_crypt_unaligned() local 71 u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in cipher_crypt_unaligned()
|
D | blkcipher.c | 78 addr = (u8 *)ALIGN((unsigned long)walk->buffer, walk->alignmask + 1); in blkcipher_done_slow() 137 if (walk->buffer != walk->page) in blkcipher_walk_done() 138 kfree(walk->buffer); in blkcipher_walk_done() 154 if (walk->buffer) in blkcipher_next_slow() 157 walk->buffer = walk->page; in blkcipher_next_slow() 158 if (walk->buffer) in blkcipher_next_slow() 163 walk->buffer = kmalloc(n, GFP_ATOMIC); in blkcipher_next_slow() 164 if (!walk->buffer) in blkcipher_next_slow() 168 walk->dst.virt.addr = (u8 *)ALIGN((unsigned long)walk->buffer, in blkcipher_next_slow() 286 walk->buffer = kmalloc(size, GFP_ATOMIC); in blkcipher_copy_iv() [all …]
|
D | rmd256.c | 28 __le32 buffer[16]; member 250 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd256_init() 259 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd256_update() 265 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd256_update() 270 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd256_update() 273 rmd256_transform(rctx->state, rctx->buffer); in rmd256_update() 277 while (len >= sizeof(rctx->buffer)) { in rmd256_update() 278 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd256_update() 279 rmd256_transform(rctx->state, rctx->buffer); in rmd256_update() 280 data += sizeof(rctx->buffer); in rmd256_update() [all …]
|
D | rmd128.c | 28 __le32 buffer[16]; member 231 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd128_init() 240 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd128_update() 246 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd128_update() 251 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd128_update() 254 rmd128_transform(rctx->state, rctx->buffer); in rmd128_update() 258 while (len >= sizeof(rctx->buffer)) { in rmd128_update() 259 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd128_update() 260 rmd128_transform(rctx->state, rctx->buffer); in rmd128_update() 261 data += sizeof(rctx->buffer); in rmd128_update() [all …]
|
D | ghash-generic.c | 31 u8 buffer[GHASH_BLOCK_SIZE]; member 68 u8 *dst = dctx->buffer; in ghash_update() 105 u8 *dst = dctx->buffer; in ghash_flush() 123 u8 *buf = dctx->buffer; in ghash_final()
|
D | rmd160.c | 28 __le32 buffer[16]; member 275 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd160_init() 284 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd160_update() 290 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd160_update() 295 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd160_update() 298 rmd160_transform(rctx->state, rctx->buffer); in rmd160_update() 302 while (len >= sizeof(rctx->buffer)) { in rmd160_update() 303 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd160_update() 304 rmd160_transform(rctx->state, rctx->buffer); in rmd160_update() 305 data += sizeof(rctx->buffer); in rmd160_update() [all …]
|
D | rmd320.c | 28 __le32 buffer[16]; member 299 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd320_init() 308 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd320_update() 314 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd320_update() 319 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd320_update() 322 rmd320_transform(rctx->state, rctx->buffer); in rmd320_update() 326 while (len >= sizeof(rctx->buffer)) { in rmd320_update() 327 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd320_update() 328 rmd320_transform(rctx->state, rctx->buffer); in rmd320_update() 329 data += sizeof(rctx->buffer); in rmd320_update() [all …]
|
D | crct10dif_common.c | 70 __u16 crc_t10dif_generic(__u16 crc, const unsigned char *buffer, size_t len) in crc_t10dif_generic() argument 75 crc = (crc << 8) ^ t10_dif_crc_table[((crc >> 8) ^ buffer[i]) & 0xff]; in crc_t10dif_generic()
|
D | sha1_generic.c | 56 memcpy(sctx->buffer + partial, data, in crypto_sha1_update() 58 src = sctx->buffer; in crypto_sha1_update() 70 memcpy(sctx->buffer + partial, src, len - done); in crypto_sha1_update()
|
D | ahash.c | 193 u8 *buffer, *alignbuffer; in ahash_setkey_unaligned() local 197 buffer = kmalloc(absize, GFP_KERNEL); in ahash_setkey_unaligned() 198 if (!buffer) in ahash_setkey_unaligned() 201 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in ahash_setkey_unaligned() 204 kzfree(buffer); in ahash_setkey_unaligned()
|
D | aead.c | 35 u8 *buffer, *alignbuffer; in setkey_unaligned() local 39 buffer = kmalloc(absize, GFP_ATOMIC); in setkey_unaligned() 40 if (!buffer) in setkey_unaligned() 43 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in setkey_unaligned() 47 kfree(buffer); in setkey_unaligned()
|
D | ablkcipher.c | 311 u8 *buffer, *alignbuffer; in setkey_unaligned() local 315 buffer = kmalloc(absize, GFP_ATOMIC); in setkey_unaligned() 316 if (!buffer) in setkey_unaligned() 319 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in setkey_unaligned() 323 kfree(buffer); in setkey_unaligned()
|
D | shash.c | 40 u8 *buffer, *alignbuffer; in shash_setkey_unaligned() local 44 buffer = kmalloc(absize, GFP_KERNEL); in shash_setkey_unaligned() 45 if (!buffer) in shash_setkey_unaligned() 48 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in shash_setkey_unaligned() 51 kzfree(buffer); in shash_setkey_unaligned()
|
D | Kconfig | 162 tristate "Software async multi-buffer crypto daemon" 169 provides the kernel thread to assist multi-buffer crypto 170 algorithms for submitting jobs and flushing jobs in multi-buffer 171 crypto algorithms. Multi-buffer crypto algorithms are executed 584 using multi-buffer technique. This algorithm computes on
|