/crypto/ |
D | skcipher.c | 43 static int skcipher_walk_next(struct skcipher_walk *walk); 45 static inline void skcipher_unmap(struct scatter_walk *walk, void *vaddr) in skcipher_unmap() argument 47 if (PageHighMem(scatterwalk_page(walk))) in skcipher_unmap() 51 static inline void *skcipher_map(struct scatter_walk *walk) in skcipher_map() argument 53 struct page *page = scatterwalk_page(walk); in skcipher_map() 56 offset_in_page(walk->offset); in skcipher_map() 59 static inline void skcipher_map_src(struct skcipher_walk *walk) in skcipher_map_src() argument 61 walk->src.virt.addr = skcipher_map(&walk->in); in skcipher_map_src() 64 static inline void skcipher_map_dst(struct skcipher_walk *walk) in skcipher_map_dst() argument 66 walk->dst.virt.addr = skcipher_map(&walk->out); in skcipher_map_dst() [all …]
|
D | cfb.c | 43 static void crypto_cfb_final(struct skcipher_walk *walk, in crypto_cfb_final() argument 49 u8 *src = walk->src.virt.addr; in crypto_cfb_final() 50 u8 *dst = walk->dst.virt.addr; in crypto_cfb_final() 51 u8 *iv = walk->iv; in crypto_cfb_final() 52 unsigned int nbytes = walk->nbytes; in crypto_cfb_final() 58 static int crypto_cfb_encrypt_segment(struct skcipher_walk *walk, in crypto_cfb_encrypt_segment() argument 62 unsigned int nbytes = walk->nbytes; in crypto_cfb_encrypt_segment() 63 u8 *src = walk->src.virt.addr; in crypto_cfb_encrypt_segment() 64 u8 *dst = walk->dst.virt.addr; in crypto_cfb_encrypt_segment() 65 u8 *iv = walk->iv; in crypto_cfb_encrypt_segment() [all …]
|
D | cbc.c | 17 static int crypto_cbc_encrypt_segment(struct skcipher_walk *walk, in crypto_cbc_encrypt_segment() argument 22 unsigned int nbytes = walk->nbytes; in crypto_cbc_encrypt_segment() 23 u8 *src = walk->src.virt.addr; in crypto_cbc_encrypt_segment() 24 u8 *dst = walk->dst.virt.addr; in crypto_cbc_encrypt_segment() 27 u8 *iv = walk->iv; in crypto_cbc_encrypt_segment() 45 static int crypto_cbc_encrypt_inplace(struct skcipher_walk *walk, in crypto_cbc_encrypt_inplace() argument 50 unsigned int nbytes = walk->nbytes; in crypto_cbc_encrypt_inplace() 51 u8 *src = walk->src.virt.addr; in crypto_cbc_encrypt_inplace() 54 u8 *iv = walk->iv; in crypto_cbc_encrypt_inplace() 68 memcpy(walk->iv, iv, bsize); in crypto_cbc_encrypt_inplace() [all …]
|
D | pcbc.c | 21 struct skcipher_walk *walk, in crypto_pcbc_encrypt_segment() argument 25 unsigned int nbytes = walk->nbytes; in crypto_pcbc_encrypt_segment() 26 u8 *src = walk->src.virt.addr; in crypto_pcbc_encrypt_segment() 27 u8 *dst = walk->dst.virt.addr; in crypto_pcbc_encrypt_segment() 28 u8 * const iv = walk->iv; in crypto_pcbc_encrypt_segment() 43 struct skcipher_walk *walk, in crypto_pcbc_encrypt_inplace() argument 47 unsigned int nbytes = walk->nbytes; in crypto_pcbc_encrypt_inplace() 48 u8 *src = walk->src.virt.addr; in crypto_pcbc_encrypt_inplace() 49 u8 * const iv = walk->iv; in crypto_pcbc_encrypt_inplace() 68 struct skcipher_walk walk; in crypto_pcbc_encrypt() local [all …]
|
D | xctr.c | 34 static void crypto_xctr_crypt_final(struct skcipher_walk *walk, in crypto_xctr_crypt_final() argument 38 const u8 *src = walk->src.virt.addr; in crypto_xctr_crypt_final() 39 u8 *dst = walk->dst.virt.addr; in crypto_xctr_crypt_final() 40 unsigned int nbytes = walk->nbytes; in crypto_xctr_crypt_final() 43 crypto_xor(walk->iv, (u8 *)&ctr32, sizeof(ctr32)); in crypto_xctr_crypt_final() 44 crypto_cipher_encrypt_one(tfm, keystream, walk->iv); in crypto_xctr_crypt_final() 46 crypto_xor(walk->iv, (u8 *)&ctr32, sizeof(ctr32)); in crypto_xctr_crypt_final() 49 static int crypto_xctr_crypt_segment(struct skcipher_walk *walk, in crypto_xctr_crypt_segment() argument 54 const u8 *src = walk->src.virt.addr; in crypto_xctr_crypt_segment() 55 u8 *dst = walk->dst.virt.addr; in crypto_xctr_crypt_segment() [all …]
|
D | ahash.c | 41 static int hash_walk_next(struct crypto_hash_walk *walk) in hash_walk_next() argument 43 unsigned int alignmask = walk->alignmask; in hash_walk_next() 44 unsigned int offset = walk->offset; in hash_walk_next() 45 unsigned int nbytes = min(walk->entrylen, in hash_walk_next() 48 walk->data = kmap_atomic(walk->pg); in hash_walk_next() 49 walk->data += offset; in hash_walk_next() 58 walk->entrylen -= nbytes; in hash_walk_next() 62 static int hash_walk_new_entry(struct crypto_hash_walk *walk) in hash_walk_new_entry() argument 66 sg = walk->sg; in hash_walk_new_entry() 67 walk->offset = sg->offset; in hash_walk_new_entry() [all …]
|
D | ofb.c | 23 struct skcipher_walk walk; in crypto_ofb_crypt() local 26 err = skcipher_walk_virt(&walk, req, false); in crypto_ofb_crypt() 28 while (walk.nbytes >= bsize) { in crypto_ofb_crypt() 29 const u8 *src = walk.src.virt.addr; in crypto_ofb_crypt() 30 u8 *dst = walk.dst.virt.addr; in crypto_ofb_crypt() 31 u8 * const iv = walk.iv; in crypto_ofb_crypt() 32 unsigned int nbytes = walk.nbytes; in crypto_ofb_crypt() 41 err = skcipher_walk_done(&walk, nbytes); in crypto_ofb_crypt() 44 if (walk.nbytes) { in crypto_ofb_crypt() 45 crypto_cipher_encrypt_one(cipher, walk.iv, walk.iv); in crypto_ofb_crypt() [all …]
|
D | ctr.c | 28 static void crypto_ctr_crypt_final(struct skcipher_walk *walk, in crypto_ctr_crypt_final() argument 33 u8 *ctrblk = walk->iv; in crypto_ctr_crypt_final() 36 u8 *src = walk->src.virt.addr; in crypto_ctr_crypt_final() 37 u8 *dst = walk->dst.virt.addr; in crypto_ctr_crypt_final() 38 unsigned int nbytes = walk->nbytes; in crypto_ctr_crypt_final() 46 static int crypto_ctr_crypt_segment(struct skcipher_walk *walk, in crypto_ctr_crypt_segment() argument 52 u8 *ctrblk = walk->iv; in crypto_ctr_crypt_segment() 53 u8 *src = walk->src.virt.addr; in crypto_ctr_crypt_segment() 54 u8 *dst = walk->dst.virt.addr; in crypto_ctr_crypt_segment() 55 unsigned int nbytes = walk->nbytes; in crypto_ctr_crypt_segment() [all …]
|
D | scatterwalk.c | 26 void scatterwalk_copychunks(void *buf, struct scatter_walk *walk, in scatterwalk_copychunks() argument 30 unsigned int len_this_page = scatterwalk_pagelen(walk); in scatterwalk_copychunks() 37 vaddr = scatterwalk_map(walk); in scatterwalk_copychunks() 42 scatterwalk_advance(walk, len_this_page); in scatterwalk_copychunks() 50 scatterwalk_pagedone(walk, out & 1, 1); in scatterwalk_copychunks() 58 struct scatter_walk walk; in scatterwalk_map_and_copy() local 66 scatterwalk_start(&walk, sg); in scatterwalk_map_and_copy() 67 scatterwalk_copychunks(buf, &walk, nbytes, out); in scatterwalk_map_and_copy() 68 scatterwalk_done(&walk, out, 0); in scatterwalk_map_and_copy()
|
D | aegis128-core.c | 281 struct scatter_walk walk; in crypto_aegis128_process_ad() local 285 scatterwalk_start(&walk, sg_src); in crypto_aegis128_process_ad() 287 unsigned int size = scatterwalk_clamp(&walk, assoclen); in crypto_aegis128_process_ad() 289 void *mapped = scatterwalk_map(&walk); in crypto_aegis128_process_ad() 312 scatterwalk_advance(&walk, size); in crypto_aegis128_process_ad() 313 scatterwalk_done(&walk, 0, assoclen); in crypto_aegis128_process_ad() 324 struct skcipher_walk *walk, in crypto_aegis128_process_crypt() argument 331 while (walk->nbytes) { in crypto_aegis128_process_crypt() 332 unsigned int nbytes = walk->nbytes; in crypto_aegis128_process_crypt() 334 if (nbytes < walk->total) in crypto_aegis128_process_crypt() [all …]
|
D | arc4.c | 30 struct skcipher_walk walk; in crypto_arc4_crypt() local 33 err = skcipher_walk_virt(&walk, req, false); in crypto_arc4_crypt() 35 while (walk.nbytes > 0) { in crypto_arc4_crypt() 36 arc4_crypt(ctx, walk.dst.virt.addr, walk.src.virt.addr, in crypto_arc4_crypt() 37 walk.nbytes); in crypto_arc4_crypt() 38 err = skcipher_walk_done(&walk, 0); in crypto_arc4_crypt()
|
D | chacha_generic.c | 18 struct skcipher_walk walk; in chacha_stream_xor() local 22 err = skcipher_walk_virt(&walk, req, false); in chacha_stream_xor() 26 while (walk.nbytes > 0) { in chacha_stream_xor() 27 unsigned int nbytes = walk.nbytes; in chacha_stream_xor() 29 if (nbytes < walk.total) in chacha_stream_xor() 32 chacha_crypt_generic(state, walk.dst.virt.addr, in chacha_stream_xor() 33 walk.src.virt.addr, nbytes, ctx->nrounds); in chacha_stream_xor() 34 err = skcipher_walk_done(&walk, walk.nbytes - nbytes); in chacha_stream_xor()
|
D | crypto_null.c | 78 struct skcipher_walk walk; in null_skcipher_crypt() local 81 err = skcipher_walk_virt(&walk, req, false); in null_skcipher_crypt() 83 while (walk.nbytes) { in null_skcipher_crypt() 84 if (walk.src.virt.addr != walk.dst.virt.addr) in null_skcipher_crypt() 85 memcpy(walk.dst.virt.addr, walk.src.virt.addr, in null_skcipher_crypt() 86 walk.nbytes); in null_skcipher_crypt() 87 err = skcipher_walk_done(&walk, 0); in null_skcipher_crypt()
|
D | ecb.c | 21 struct skcipher_walk walk; in crypto_ecb_crypt() local 25 err = skcipher_walk_virt(&walk, req, false); in crypto_ecb_crypt() 27 while ((nbytes = walk.nbytes) != 0) { in crypto_ecb_crypt() 28 const u8 *src = walk.src.virt.addr; in crypto_ecb_crypt() 29 u8 *dst = walk.dst.virt.addr; in crypto_ecb_crypt() 38 err = skcipher_walk_done(&walk, nbytes); in crypto_ecb_crypt()
|
D | shash.c | 269 struct crypto_hash_walk walk; in shash_ahash_update() local 272 for (nbytes = crypto_hash_walk_first(req, &walk); nbytes > 0; in shash_ahash_update() 273 nbytes = crypto_hash_walk_done(&walk, nbytes)) in shash_ahash_update() 274 nbytes = crypto_shash_update(desc, walk.data, nbytes); in shash_ahash_update() 292 struct crypto_hash_walk walk; in shash_ahash_finup() local 295 nbytes = crypto_hash_walk_first(req, &walk); in shash_ahash_finup() 300 nbytes = crypto_hash_walk_last(&walk) ? in shash_ahash_finup() 301 crypto_shash_finup(desc, walk.data, nbytes, in shash_ahash_finup() 303 crypto_shash_update(desc, walk.data, nbytes); in shash_ahash_finup() 304 nbytes = crypto_hash_walk_done(&walk, nbytes); in shash_ahash_finup()
|
D | keywrap.c | 102 static void crypto_kw_scatterlist_ff(struct scatter_walk *walk, in crypto_kw_scatterlist_ff() argument 114 scatterwalk_start(walk, sg); in crypto_kw_scatterlist_ff() 115 scatterwalk_advance(walk, skip); in crypto_kw_scatterlist_ff()
|