Home
last modified time | relevance | path

Searched refs:cryptlen (Results 1 – 22 of 22) sorted by relevance

/crypto/
Dauthencesn.c107 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv_tail() local
113 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail()
116 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail()
141 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv() local
151 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv()
157 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); in crypto_authenc_esn_genicv()
200 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_encrypt() local
220 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_esn_encrypt()
241 unsigned int cryptlen = req->cryptlen - authsize; in crypto_authenc_esn_decrypt_tail() local
252 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_decrypt_tail()
[all …]
Dkeywrap.c129 u64 t = 6 * ((req->cryptlen) >> 3); in crypto_kw_decrypt()
137 if (req->cryptlen < (2 * SEMIBSIZE) || req->cryptlen % SEMIBSIZE) in crypto_kw_decrypt()
153 unsigned int nbytes = req->cryptlen; in crypto_kw_decrypt()
207 if (req->cryptlen < (2 * SEMIBSIZE) || req->cryptlen % SEMIBSIZE) in crypto_kw_encrypt()
226 unsigned int nbytes = req->cryptlen; in crypto_kw_encrypt()
Dchacha20poly1305.c41 __le64 cryptlen; member
61 unsigned int cryptlen; member
103 req->assoclen + rctx->cryptlen, in poly_verify_tag()
115 req->assoclen + rctx->cryptlen, in poly_copy_tag()
133 if (rctx->cryptlen == 0) in chacha_decrypt()
147 rctx->cryptlen, creq->iv); in chacha_decrypt()
160 if (rctx->cryptlen == req->cryptlen) /* encrypting */ in poly_tail_continue()
180 preq->tail.cryptlen = cpu_to_le64(rctx->cryptlen); in poly_tail()
209 padlen = -rctx->cryptlen % POLY1305_BLOCK_SIZE; in poly_cipherpad()
238 if (rctx->cryptlen == req->cryptlen) /* encrypting */ in poly_cipher()
[all …]
Daegis128-core.c349 u64 assoclen, u64 cryptlen) in crypto_aegis128_final() argument
352 u64 cryptbits = cryptlen * 8; in crypto_aegis128_final()
395 unsigned int cryptlen, in crypto_aegis128_crypt() argument
405 crypto_aegis128_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis128_crypt()
418 unsigned int cryptlen = req->cryptlen; in crypto_aegis128_encrypt() local
425 crypto_aegis128_crypt(req, &tag, cryptlen, ops); in crypto_aegis128_encrypt()
427 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen, in crypto_aegis128_encrypt()
443 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis128_decrypt() local
445 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen, in crypto_aegis128_decrypt()
453 crypto_aegis128_crypt(req, &tag, cryptlen, ops); in crypto_aegis128_decrypt()
Dccm.c136 unsigned int cryptlen) in format_input() argument
154 return set_msg_len(info + 16 - l, cryptlen, l); in format_input()
177 unsigned int cryptlen) in crypto_ccm_auth() argument
190 err = format_input(odata, req, cryptlen); in crypto_ccm_auth()
226 cryptlen += ilen; in crypto_ccm_auth()
229 ahash_request_set_crypt(ahreq, plain, pctx->odata, cryptlen); in crypto_ccm_auth()
244 req->assoclen + req->cryptlen, in crypto_ccm_encrypt_done()
300 unsigned int cryptlen = req->cryptlen; in crypto_ccm_encrypt() local
309 err = crypto_ccm_auth(req, sg_next(pctx->src), cryptlen); in crypto_ccm_encrypt()
320 skcipher_request_set_crypt(skreq, pctx->src, dst, cryptlen + 16, iv); in crypto_ccm_encrypt()
[all …]
Dechainiv.c37 if (req->cryptlen < ivsize) in echainiv_encrypt()
51 req->assoclen + req->cryptlen, in echainiv_encrypt()
62 req->cryptlen, info); in echainiv_encrypt()
94 if (req->cryptlen < ivsize) in echainiv_decrypt()
104 req->cryptlen - ivsize, req->iv); in echainiv_decrypt()
Dgcm.c59 unsigned int cryptlen; member
184 unsigned int cryptlen) in crypto_gcm_init_crypt() argument
196 cryptlen + sizeof(pctx->auth_tag), in crypto_gcm_init_crypt()
237 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len()
298 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue()
327 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue()
329 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue()
425 req->assoclen + req->cryptlen, in gcm_enc_copy_hash()
436 gctx->cryptlen = req->cryptlen; in gcm_encrypt_continue()
464 crypto_gcm_init_crypt(req, req->cryptlen); in crypto_gcm_encrypt()
[all …]
Dxts.c91 const bool cts = (req->cryptlen % XTS_BLOCK_SIZE); in xor_tweak()
167 int offset = req->cryptlen & ~(XTS_BLOCK_SIZE - 1); in cts_final()
170 int tail = req->cryptlen % XTS_BLOCK_SIZE; in cts_final()
211 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) { in encrypt_done()
231 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) { in decrypt_done()
247 if (req->cryptlen < XTS_BLOCK_SIZE) in init_crypt()
253 req->cryptlen & ~(XTS_BLOCK_SIZE - 1), NULL); in init_crypt()
272 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0)) in encrypt()
289 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0)) in decrypt()
Dseqiv.c61 if (req->cryptlen < ivsize) in seqiv_aead_encrypt()
77 req->assoclen + req->cryptlen, in seqiv_aead_encrypt()
99 req->cryptlen - ivsize, info); in seqiv_aead_encrypt()
120 if (req->cryptlen < ivsize + crypto_aead_authsize(geniv)) in seqiv_aead_decrypt()
130 req->cryptlen - ivsize, req->iv); in seqiv_aead_decrypt()
Daead.c92 unsigned int cryptlen = req->cryptlen; in crypto_aead_encrypt() local
100 crypto_stats_aead_encrypt(cryptlen, alg, ret); in crypto_aead_encrypt()
109 unsigned int cryptlen = req->cryptlen; in crypto_aead_decrypt() local
115 else if (req->cryptlen < crypto_aead_authsize(aead)) in crypto_aead_decrypt()
119 crypto_stats_aead_decrypt(cryptlen, alg, ret); in crypto_aead_decrypt()
Dauthenc.c135 req->assoclen + req->cryptlen, in authenc_geniv_ahash_done()
159 req->assoclen + req->cryptlen); in crypto_authenc_genicv()
167 scatterwalk_map_and_copy(hash, req->dst, req->assoclen + req->cryptlen, in crypto_authenc_genicv()
210 unsigned int cryptlen = req->cryptlen; in crypto_authenc_encrypt() local
230 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_encrypt()
269 req->cryptlen - authsize, req->iv); in crypto_authenc_decrypt_tail()
306 req->assoclen + req->cryptlen - authsize); in crypto_authenc_decrypt()
Dskcipher.c453 walk->total = req->cryptlen; in skcipher_walk_skcipher()
555 walk->total = req->cryptlen; in skcipher_walk_aead()
564 walk->total = req->cryptlen; in skcipher_walk_aead_encrypt()
575 walk->total = req->cryptlen - crypto_aead_authsize(tfm); in skcipher_walk_aead_decrypt()
635 return crypt(&desc, req->dst, req->src, req->cryptlen); in skcipher_crypt_blkcipher()
731 ablkcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in skcipher_crypt_ablkcipher()
848 unsigned int cryptlen = req->cryptlen; in crypto_skcipher_encrypt() local
856 crypto_stats_skcipher_encrypt(cryptlen, ret, alg); in crypto_skcipher_encrypt()
865 unsigned int cryptlen = req->cryptlen; in crypto_skcipher_decrypt() local
873 crypto_stats_skcipher_decrypt(cryptlen, ret, alg); in crypto_skcipher_decrypt()
Dadiantum.c232 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_header()
261 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_message()
295 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_finish()
335 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_crypt()
340 if (req->cryptlen < BLOCKCIPHER_BLOCK_SIZE) in adiantum_crypt()
382 if (round_up(stream_len, CHACHA_BLOCK_SIZE) <= req->cryptlen) in adiantum_crypt()
Dcts.c114 lastn = req->cryptlen - offset; in cts_cbc_encrypt()
154 unsigned int nbytes = req->cryptlen; in crypto_cts_encrypt()
196 lastn = req->cryptlen - offset; in cts_cbc_decrypt()
248 unsigned int nbytes = req->cryptlen; in crypto_cts_decrypt()
Dalgapi.c1081 void crypto_stats_aead_encrypt(unsigned int cryptlen, struct crypto_alg *alg, in crypto_stats_aead_encrypt() argument
1088 atomic64_add(cryptlen, &alg->stats.aead.encrypt_tlen); in crypto_stats_aead_encrypt()
1094 void crypto_stats_aead_decrypt(unsigned int cryptlen, struct crypto_alg *alg, in crypto_stats_aead_decrypt() argument
1101 atomic64_add(cryptlen, &alg->stats.aead.decrypt_tlen); in crypto_stats_aead_decrypt()
1254 void crypto_stats_skcipher_encrypt(unsigned int cryptlen, int ret, in crypto_stats_skcipher_encrypt() argument
1261 atomic64_add(cryptlen, &alg->stats.cipher.encrypt_tlen); in crypto_stats_skcipher_encrypt()
1267 void crypto_stats_skcipher_decrypt(unsigned int cryptlen, int ret, in crypto_stats_skcipher_decrypt() argument
1274 atomic64_add(cryptlen, &alg->stats.cipher.decrypt_tlen); in crypto_stats_skcipher_decrypt()
Dpcrypt.c103 req->cryptlen, req->iv); in pcrypt_aead_encrypt()
145 req->cryptlen, req->iv); in pcrypt_aead_decrypt()
Ddrbg.c1837 u32 cryptlen = min3(inlen, outlen, (u32)DRBG_OUTSCRATCHLEN); in drbg_kcapi_sym_ctr() local
1841 cryptlen, drbg->V); in drbg_kcapi_sym_ctr()
1849 memcpy(outbuf, drbg->outscratchpad, cryptlen); in drbg_kcapi_sym_ctr()
1850 memzero_explicit(drbg->outscratchpad, cryptlen); in drbg_kcapi_sym_ctr()
1852 outlen -= cryptlen; in drbg_kcapi_sym_ctr()
1853 outbuf += cryptlen; in drbg_kcapi_sym_ctr()
Dessiv.c167 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in essiv_skcipher_crypt()
263 aead_request_set_crypt(subreq, src, req->dst, req->cryptlen, req->iv); in essiv_aead_crypt()
Dlrw.c233 req->cryptlen, req->iv); in init_crypt()
Dctr.c218 req->cryptlen, iv); in crypto_rfc3686_crypt()
Dcryptd.c299 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in cryptd_skcipher_encrypt()
327 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in cryptd_skcipher_decrypt()
Dtestmgr.c1932 if (req->cryptlen != (enc ? vec->plen : vec->clen) || in test_aead_vec_cfg()
1943 if (req->cryptlen != (enc ? vec->plen : vec->clen)) in test_aead_vec_cfg()
2518 if (req->cryptlen != vec->len || in test_skcipher_vec_cfg()
2528 if (req->cryptlen != vec->len) in test_skcipher_vec_cfg()