Home
last modified time | relevance | path

Searched refs:buffer (Results 1 – 18 of 18) sorted by relevance

/crypto/asymmetric_keys/
Dmscode_parser.c54 char buffer[50]; in mscode_note_content_type() local
56 sprint_oid(value, vlen, buffer, sizeof(buffer)); in mscode_note_content_type()
57 pr_err("Unknown OID: %s\n", buffer); in mscode_note_content_type()
83 char buffer[50]; in mscode_note_digest_algo() local
102 sprint_oid(value, vlen, buffer, sizeof(buffer)); in mscode_note_digest_algo()
103 pr_err("Unknown OID: %s\n", buffer); in mscode_note_digest_algo()
Dx509_cert_parser.c129 char buffer[50]; in x509_note_OID() local
130 sprint_oid(value, vlen, buffer, sizeof(buffer)); in x509_note_OID()
132 (unsigned long)value - ctx->data, buffer); in x509_note_OID()
280 char *buffer; in x509_fabricate_name() local
287 buffer = kmalloc(1, GFP_KERNEL); in x509_fabricate_name()
288 if (!buffer) in x509_fabricate_name()
290 buffer[0] = 0; in x509_fabricate_name()
309 buffer = kmalloc(ctx->o_size + 2 + ctx->cn_size + 1, in x509_fabricate_name()
311 if (!buffer) in x509_fabricate_name()
314 memcpy(buffer, in x509_fabricate_name()
[all …]
Dpkcs7_parser.c171 char buffer[50]; in pkcs7_note_OID() local
172 sprint_oid(value, vlen, buffer, sizeof(buffer)); in pkcs7_note_OID()
174 (unsigned long)value - ctx->data, buffer); in pkcs7_note_OID()
/crypto/
Dwp512.c40 u8 buffer[WP512_BLOCK_SIZE]; member
788 const __be64 *buffer = (const __be64 *)wctx->buffer; in wp512_process_buffer() local
791 block[i] = be64_to_cpu(buffer[i]); in wp512_process_buffer()
995 wctx->buffer[0] = 0; in wp512_init()
1013 u8 *buffer = wctx->buffer; in wp512_update() local
1028 buffer[bufferPos++] |= (u8)(b >> bufferRem); in wp512_update()
1034 buffer[bufferPos] = b << (8 - bufferRem); in wp512_update()
1041 buffer[bufferPos] |= b >> bufferRem; in wp512_update()
1055 buffer[bufferPos] = b << (8 - bufferRem); in wp512_update()
1069 u8 *buffer = wctx->buffer; in wp512_final() local
[all …]
Dcipher.c29 u8 *buffer, *alignbuffer; in setkey_unaligned() local
33 buffer = kmalloc(absize, GFP_ATOMIC); in setkey_unaligned()
34 if (!buffer) in setkey_unaligned()
37 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in setkey_unaligned()
41 kfree(buffer); in setkey_unaligned()
70 u8 buffer[size + alignmask]; in cipher_crypt_unaligned() local
71 u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in cipher_crypt_unaligned()
Dblkcipher.c78 addr = (u8 *)ALIGN((unsigned long)walk->buffer, walk->alignmask + 1); in blkcipher_done_slow()
137 if (walk->buffer != walk->page) in blkcipher_walk_done()
138 kfree(walk->buffer); in blkcipher_walk_done()
154 if (walk->buffer) in blkcipher_next_slow()
157 walk->buffer = walk->page; in blkcipher_next_slow()
158 if (walk->buffer) in blkcipher_next_slow()
163 walk->buffer = kmalloc(n, GFP_ATOMIC); in blkcipher_next_slow()
164 if (!walk->buffer) in blkcipher_next_slow()
168 walk->dst.virt.addr = (u8 *)ALIGN((unsigned long)walk->buffer, in blkcipher_next_slow()
286 walk->buffer = kmalloc(size, GFP_ATOMIC); in blkcipher_copy_iv()
[all …]
Drmd256.c28 __le32 buffer[16]; member
250 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd256_init()
259 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd256_update()
265 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd256_update()
270 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd256_update()
273 rmd256_transform(rctx->state, rctx->buffer); in rmd256_update()
277 while (len >= sizeof(rctx->buffer)) { in rmd256_update()
278 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd256_update()
279 rmd256_transform(rctx->state, rctx->buffer); in rmd256_update()
280 data += sizeof(rctx->buffer); in rmd256_update()
[all …]
Drmd128.c28 __le32 buffer[16]; member
231 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd128_init()
240 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd128_update()
246 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd128_update()
251 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd128_update()
254 rmd128_transform(rctx->state, rctx->buffer); in rmd128_update()
258 while (len >= sizeof(rctx->buffer)) { in rmd128_update()
259 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd128_update()
260 rmd128_transform(rctx->state, rctx->buffer); in rmd128_update()
261 data += sizeof(rctx->buffer); in rmd128_update()
[all …]
Dghash-generic.c31 u8 buffer[GHASH_BLOCK_SIZE]; member
68 u8 *dst = dctx->buffer; in ghash_update()
105 u8 *dst = dctx->buffer; in ghash_flush()
123 u8 *buf = dctx->buffer; in ghash_final()
Drmd160.c28 __le32 buffer[16]; member
275 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd160_init()
284 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd160_update()
290 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd160_update()
295 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd160_update()
298 rmd160_transform(rctx->state, rctx->buffer); in rmd160_update()
302 while (len >= sizeof(rctx->buffer)) { in rmd160_update()
303 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd160_update()
304 rmd160_transform(rctx->state, rctx->buffer); in rmd160_update()
305 data += sizeof(rctx->buffer); in rmd160_update()
[all …]
Drmd320.c28 __le32 buffer[16]; member
299 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd320_init()
308 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd320_update()
314 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd320_update()
319 memcpy((char *)rctx->buffer + (sizeof(rctx->buffer) - avail), in rmd320_update()
322 rmd320_transform(rctx->state, rctx->buffer); in rmd320_update()
326 while (len >= sizeof(rctx->buffer)) { in rmd320_update()
327 memcpy(rctx->buffer, data, sizeof(rctx->buffer)); in rmd320_update()
328 rmd320_transform(rctx->state, rctx->buffer); in rmd320_update()
329 data += sizeof(rctx->buffer); in rmd320_update()
[all …]
Dcrct10dif_common.c70 __u16 crc_t10dif_generic(__u16 crc, const unsigned char *buffer, size_t len) in crc_t10dif_generic() argument
75 crc = (crc << 8) ^ t10_dif_crc_table[((crc >> 8) ^ buffer[i]) & 0xff]; in crc_t10dif_generic()
Dsha1_generic.c56 memcpy(sctx->buffer + partial, data, in crypto_sha1_update()
58 src = sctx->buffer; in crypto_sha1_update()
70 memcpy(sctx->buffer + partial, src, len - done); in crypto_sha1_update()
Dahash.c193 u8 *buffer, *alignbuffer; in ahash_setkey_unaligned() local
197 buffer = kmalloc(absize, GFP_KERNEL); in ahash_setkey_unaligned()
198 if (!buffer) in ahash_setkey_unaligned()
201 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in ahash_setkey_unaligned()
204 kzfree(buffer); in ahash_setkey_unaligned()
Daead.c35 u8 *buffer, *alignbuffer; in setkey_unaligned() local
39 buffer = kmalloc(absize, GFP_ATOMIC); in setkey_unaligned()
40 if (!buffer) in setkey_unaligned()
43 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in setkey_unaligned()
47 kfree(buffer); in setkey_unaligned()
Dablkcipher.c311 u8 *buffer, *alignbuffer; in setkey_unaligned() local
315 buffer = kmalloc(absize, GFP_ATOMIC); in setkey_unaligned()
316 if (!buffer) in setkey_unaligned()
319 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in setkey_unaligned()
323 kfree(buffer); in setkey_unaligned()
Dshash.c40 u8 *buffer, *alignbuffer; in shash_setkey_unaligned() local
44 buffer = kmalloc(absize, GFP_KERNEL); in shash_setkey_unaligned()
45 if (!buffer) in shash_setkey_unaligned()
48 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in shash_setkey_unaligned()
51 kzfree(buffer); in shash_setkey_unaligned()
DKconfig162 tristate "Software async multi-buffer crypto daemon"
169 provides the kernel thread to assist multi-buffer crypto
170 algorithms for submitting jobs and flushing jobs in multi-buffer
171 crypto algorithms. Multi-buffer crypto algorithms are executed
584 using multi-buffer technique. This algorithm computes on