/crypto/ |
D | authencesn.c | 97 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv_tail() local 103 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail() 106 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail() 131 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv() local 141 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv() 147 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); in crypto_authenc_esn_genicv() 190 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_encrypt() local 210 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_esn_encrypt() 231 unsigned int cryptlen = req->cryptlen - authsize; in crypto_authenc_esn_decrypt_tail() local 242 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_decrypt_tail() [all …]
|
D | keywrap.c | 130 u64 t = 6 * ((req->cryptlen) >> 3); in crypto_kw_decrypt() 138 if (req->cryptlen < (2 * SEMIBSIZE) || req->cryptlen % SEMIBSIZE) in crypto_kw_decrypt() 154 unsigned int nbytes = req->cryptlen; in crypto_kw_decrypt() 208 if (req->cryptlen < (2 * SEMIBSIZE) || req->cryptlen % SEMIBSIZE) in crypto_kw_encrypt() 227 unsigned int nbytes = req->cryptlen; in crypto_kw_encrypt()
|
D | chacha20poly1305.c | 39 __le64 cryptlen; member 59 unsigned int cryptlen; member 101 req->assoclen + rctx->cryptlen, in poly_verify_tag() 113 req->assoclen + rctx->cryptlen, in poly_copy_tag() 131 if (rctx->cryptlen == 0) in chacha_decrypt() 145 rctx->cryptlen, creq->iv); in chacha_decrypt() 158 if (rctx->cryptlen == req->cryptlen) /* encrypting */ in poly_tail_continue() 178 preq->tail.cryptlen = cpu_to_le64(rctx->cryptlen); in poly_tail() 207 padlen = -rctx->cryptlen % POLY1305_BLOCK_SIZE; in poly_cipherpad() 236 if (rctx->cryptlen == req->cryptlen) /* encrypting */ in poly_cipher() [all …]
|
D | aegis128-core.c | 72 u64 assoclen, u64 cryptlen); 350 u64 assoclen, u64 cryptlen) in crypto_aegis128_final() argument 353 u64 cryptbits = cryptlen * 8; in crypto_aegis128_final() 398 unsigned int cryptlen = req->cryptlen; in crypto_aegis128_encrypt() local 409 cryptlen); in crypto_aegis128_encrypt() 415 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen); in crypto_aegis128_encrypt() 418 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen, in crypto_aegis128_encrypt() 429 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis128_decrypt() local 434 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen, in crypto_aegis128_decrypt() 444 cryptlen); in crypto_aegis128_decrypt() [all …]
|
D | ccm.c | 128 unsigned int cryptlen) in format_input() argument 146 return set_msg_len(info + 16 - l, cryptlen, l); in format_input() 169 unsigned int cryptlen) in crypto_ccm_auth() argument 182 err = format_input(odata, req, cryptlen); in crypto_ccm_auth() 218 cryptlen += ilen; in crypto_ccm_auth() 221 ahash_request_set_crypt(ahreq, plain, pctx->odata, cryptlen); in crypto_ccm_auth() 236 req->assoclen + req->cryptlen, in crypto_ccm_encrypt_done() 292 unsigned int cryptlen = req->cryptlen; in crypto_ccm_encrypt() local 301 err = crypto_ccm_auth(req, sg_next(pctx->src), cryptlen); in crypto_ccm_encrypt() 312 skcipher_request_set_crypt(skreq, pctx->src, dst, cryptlen + 16, iv); in crypto_ccm_encrypt() [all …]
|
D | gcm.c | 58 unsigned int cryptlen; member 178 unsigned int cryptlen) in crypto_gcm_init_crypt() argument 190 cryptlen + sizeof(pctx->auth_tag), in crypto_gcm_init_crypt() 231 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len() 292 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue() 321 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue() 323 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue() 419 req->assoclen + req->cryptlen, in gcm_enc_copy_hash() 430 gctx->cryptlen = req->cryptlen; in gcm_encrypt_continue() 458 crypto_gcm_init_crypt(req, req->cryptlen); in crypto_gcm_encrypt() [all …]
|
D | echainiv.c | 37 if (req->cryptlen < ivsize) in echainiv_encrypt() 51 req->assoclen + req->cryptlen, in echainiv_encrypt() 62 req->cryptlen, info); in echainiv_encrypt() 94 if (req->cryptlen < ivsize) in echainiv_decrypt() 104 req->cryptlen - ivsize, req->iv); in echainiv_decrypt()
|
D | aead.c | 87 unsigned int cryptlen = req->cryptlen; in crypto_aead_encrypt() local 95 crypto_stats_aead_encrypt(cryptlen, alg, ret); in crypto_aead_encrypt() 104 unsigned int cryptlen = req->cryptlen; in crypto_aead_decrypt() local 110 else if (req->cryptlen < crypto_aead_authsize(aead)) in crypto_aead_decrypt() 114 crypto_stats_aead_decrypt(cryptlen, alg, ret); in crypto_aead_decrypt()
|
D | aegis128-neon.c | 18 uint64_t cryptlen); 65 u64 assoclen, u64 cryptlen) in crypto_aegis128_final_simd() argument 68 crypto_aegis128_final_neon(state, tag_xor, assoclen, cryptlen); in crypto_aegis128_final_simd()
|
D | seqiv.c | 59 if (req->cryptlen < ivsize) in seqiv_aead_encrypt() 75 req->assoclen + req->cryptlen, in seqiv_aead_encrypt() 97 req->cryptlen - ivsize, info); in seqiv_aead_encrypt() 118 if (req->cryptlen < ivsize + crypto_aead_authsize(geniv)) in seqiv_aead_decrypt() 128 req->cryptlen - ivsize, req->iv); in seqiv_aead_decrypt()
|
D | xts.c | 87 const bool cts = (req->cryptlen % XTS_BLOCK_SIZE); in xts_xor_tweak() 164 int offset = req->cryptlen & ~(XTS_BLOCK_SIZE - 1); in xts_cts_final() 167 int tail = req->cryptlen % XTS_BLOCK_SIZE; in xts_cts_final() 209 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) { in xts_encrypt_done() 229 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) { in xts_decrypt_done() 247 if (req->cryptlen < XTS_BLOCK_SIZE) in xts_init_crypt() 253 req->cryptlen & ~(XTS_BLOCK_SIZE - 1), NULL); in xts_init_crypt() 272 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0)) in xts_encrypt() 289 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0)) in xts_decrypt()
|
D | authenc.c | 125 req->assoclen + req->cryptlen, in authenc_geniv_ahash_done() 149 req->assoclen + req->cryptlen); in crypto_authenc_genicv() 157 scatterwalk_map_and_copy(hash, req->dst, req->assoclen + req->cryptlen, in crypto_authenc_genicv() 200 unsigned int cryptlen = req->cryptlen; in crypto_authenc_encrypt() local 220 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_encrypt() 259 req->cryptlen - authsize, req->iv); in crypto_authenc_decrypt_tail() 296 req->assoclen + req->cryptlen - authsize); in crypto_authenc_decrypt()
|
D | adiantum.c | 224 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_header() 253 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_message() 287 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_finish() 327 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_crypt() 332 if (req->cryptlen < BLOCKCIPHER_BLOCK_SIZE) in adiantum_crypt() 374 if (round_up(stream_len, CHACHA_BLOCK_SIZE) <= req->cryptlen) in adiantum_crypt()
|
D | skcipher.c | 454 walk->total = req->cryptlen; in skcipher_walk_skcipher() 556 walk->total = req->cryptlen; in skcipher_walk_aead_encrypt() 567 walk->total = req->cryptlen - crypto_aead_authsize(tfm); in skcipher_walk_aead_decrypt() 629 unsigned int cryptlen = req->cryptlen; in crypto_skcipher_encrypt() local 637 crypto_stats_skcipher_encrypt(cryptlen, ret, alg); in crypto_skcipher_encrypt() 646 unsigned int cryptlen = req->cryptlen; in crypto_skcipher_decrypt() local 654 crypto_stats_skcipher_decrypt(cryptlen, ret, alg); in crypto_skcipher_decrypt()
|
D | cts.c | 110 lastn = req->cryptlen - offset; in cts_cbc_encrypt() 150 unsigned int nbytes = req->cryptlen; in crypto_cts_encrypt() 192 lastn = req->cryptlen - offset; in cts_cbc_decrypt() 244 unsigned int nbytes = req->cryptlen; in crypto_cts_decrypt()
|
D | algapi.c | 1085 void crypto_stats_aead_encrypt(unsigned int cryptlen, struct crypto_alg *alg, in crypto_stats_aead_encrypt() argument 1092 atomic64_add(cryptlen, &alg->stats.aead.encrypt_tlen); in crypto_stats_aead_encrypt() 1098 void crypto_stats_aead_decrypt(unsigned int cryptlen, struct crypto_alg *alg, in crypto_stats_aead_decrypt() argument 1105 atomic64_add(cryptlen, &alg->stats.aead.decrypt_tlen); in crypto_stats_aead_decrypt() 1258 void crypto_stats_skcipher_encrypt(unsigned int cryptlen, int ret, in crypto_stats_skcipher_encrypt() argument 1265 atomic64_add(cryptlen, &alg->stats.cipher.encrypt_tlen); in crypto_stats_skcipher_encrypt() 1271 void crypto_stats_skcipher_decrypt(unsigned int cryptlen, int ret, in crypto_stats_skcipher_decrypt() argument 1278 atomic64_add(cryptlen, &alg->stats.cipher.decrypt_tlen); in crypto_stats_skcipher_decrypt()
|
D | aegis128-neon-inner.c | 245 uint64_t cryptlen) in crypto_aegis128_final_neon() argument 254 vmov_n_u64(8 * cryptlen)); in crypto_aegis128_final_neon()
|
D | pcrypt.c | 114 req->cryptlen, req->iv); in pcrypt_aead_encrypt() 163 req->cryptlen, req->iv); in pcrypt_aead_decrypt()
|
D | drbg.c | 1831 u32 cryptlen = min3(inlen, outlen, (u32)DRBG_OUTSCRATCHLEN); in drbg_kcapi_sym_ctr() local 1835 cryptlen, drbg->V); in drbg_kcapi_sym_ctr() 1843 memcpy(outbuf, drbg->outscratchpad, cryptlen); in drbg_kcapi_sym_ctr() 1844 memzero_explicit(drbg->outscratchpad, cryptlen); in drbg_kcapi_sym_ctr() 1846 outlen -= cryptlen; in drbg_kcapi_sym_ctr() 1847 outbuf += cryptlen; in drbg_kcapi_sym_ctr()
|
D | essiv.c | 150 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in essiv_skcipher_crypt() 250 aead_request_set_crypt(subreq, src, req->dst, req->cryptlen, req->iv); in essiv_aead_crypt()
|
D | lrw.c | 234 req->cryptlen, req->iv); in lrw_init_crypt()
|
D | ctr.c | 213 req->cryptlen, iv); in crypto_rfc3686_crypt()
|
D | cryptd.c | 271 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in cryptd_skcipher_encrypt() 299 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in cryptd_skcipher_decrypt()
|
D | testmgr.c | 2025 if (req->cryptlen != (enc ? vec->plen : vec->clen) || in test_aead_vec_cfg() 2036 if (req->cryptlen != (enc ? vec->plen : vec->clen)) in test_aead_vec_cfg() 2786 if (req->cryptlen != vec->len || in test_skcipher_vec_cfg() 2796 if (req->cryptlen != vec->len) in test_skcipher_vec_cfg()
|