Home
last modified time | relevance | path

Searched refs:cipher (Results 1 – 25 of 37) sorted by relevance

12

/crypto/
Decb.c16 struct crypto_cipher *cipher, in crypto_ecb_crypt() argument
19 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ecb_crypt()
31 fn(crypto_cipher_tfm(cipher), dst, src); in crypto_ecb_crypt()
46 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_ecb_encrypt() local
48 return crypto_ecb_crypt(req, cipher, in crypto_ecb_encrypt()
49 crypto_cipher_alg(cipher)->cia_encrypt); in crypto_ecb_encrypt()
55 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_ecb_decrypt() local
57 return crypto_ecb_crypt(req, cipher, in crypto_ecb_decrypt()
58 crypto_cipher_alg(cipher)->cia_decrypt); in crypto_ecb_decrypt()
Dcipher.c78 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; in cipher_encrypt_unaligned() local
81 cipher_crypt_unaligned(cipher->cia_encrypt, tfm, dst, src); in cipher_encrypt_unaligned()
85 cipher->cia_encrypt(tfm, dst, src); in cipher_encrypt_unaligned()
92 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; in cipher_decrypt_unaligned() local
95 cipher_crypt_unaligned(cipher->cia_decrypt, tfm, dst, src); in cipher_decrypt_unaligned()
99 cipher->cia_decrypt(tfm, dst, src); in cipher_decrypt_unaligned()
105 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; in crypto_init_cipher_ops() local
109 cipher_encrypt_unaligned : cipher->cia_encrypt; in crypto_init_cipher_ops()
111 cipher_decrypt_unaligned : cipher->cia_decrypt; in crypto_init_cipher_ops()
DKconfig298 Support for the AEAD wrapper using the ChaCha20 stream cipher combined
353 This block cipher algorithm is required for IPSec.
361 This block cipher algorithm is required for TPM2 Cryptography.
370 This block cipher algorithm is required for IPSec.
393 This is the simplest block cipher algorithm. It simply encrypts
403 narrow block cipher mode for dm-crypt. Use it with cipher
406 rest is used to tie each cipher block to its logical position.
413 OFB: the Output Feedback mode makes a block cipher into a synchronous
414 stream cipher. It generates keystream blocks, which are then XORed
426 This block cipher algorithm is required for RxRPC.
[all …]
Dctr.c101 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_ctr_crypt() local
102 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ctr_crypt()
111 nbytes = crypto_ctr_crypt_inplace(&walk, cipher); in crypto_ctr_crypt()
113 nbytes = crypto_ctr_crypt_segment(&walk, cipher); in crypto_ctr_crypt()
119 crypto_ctr_crypt_final(&walk, cipher); in crypto_ctr_crypt()
228 struct crypto_skcipher *cipher; in crypto_rfc3686_init_tfm() local
232 cipher = crypto_spawn_skcipher(spawn); in crypto_rfc3686_init_tfm()
233 if (IS_ERR(cipher)) in crypto_rfc3686_init_tfm()
234 return PTR_ERR(cipher); in crypto_rfc3686_init_tfm()
236 ctx->child = cipher; in crypto_rfc3686_init_tfm()
[all …]
Dofb.c20 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_ofb_crypt() local
21 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ofb_crypt()
34 crypto_cipher_encrypt_one(cipher, iv, iv); in crypto_ofb_crypt()
44 crypto_cipher_encrypt_one(cipher, walk.iv, walk.iv); in crypto_ofb_crypt()
Dpcbc.c66 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_pcbc_encrypt() local
76 cipher); in crypto_pcbc_encrypt()
79 cipher); in crypto_pcbc_encrypt()
133 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_pcbc_decrypt() local
143 cipher); in crypto_pcbc_decrypt()
146 cipher); in crypto_pcbc_decrypt()
Dtestmgr.c133 struct cipher_test_suite cipher; member
2847 const struct cipher_test_suite *suite = &desc->suite.cipher; in alg_test_skcipher()
3238 const struct cipher_test_suite *suite = &desc->suite.cipher; in alg_test_cipher()
3874 .cipher = __VECS(adiantum_xchacha12_aes_tv_template)
3881 .cipher = __VECS(adiantum_xchacha20_aes_tv_template)
4030 .cipher = __VECS(aes_cbc_tv_template)
4036 .cipher = __VECS(anubis_cbc_tv_template)
4042 .cipher = __VECS(bf_cbc_tv_template)
4048 .cipher = __VECS(camellia_cbc_tv_template)
4054 .cipher = __VECS(cast5_cbc_tv_template)
[all …]
Dkeywrap.c126 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_kw_decrypt() local
166 crypto_cipher_decrypt_one(cipher, (u8 *)&block, in crypto_kw_decrypt()
195 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in crypto_kw_encrypt() local
237 crypto_cipher_encrypt_one(cipher, (u8 *)&block, in crypto_kw_encrypt()
Dskcipher.c801 struct skcipher_alg *cipher = crypto_skcipher_alg(tfm); in skcipher_setkey_unaligned() local
813 ret = cipher->setkey(tfm, alignbuffer, keylen); in skcipher_setkey_unaligned()
821 struct skcipher_alg *cipher = crypto_skcipher_alg(tfm); in skcipher_setkey() local
825 if (keylen < cipher->min_keysize || keylen > cipher->max_keysize) { in skcipher_setkey()
833 err = cipher->setkey(tfm, key, keylen); in skcipher_setkey()
1111 struct crypto_cipher *cipher = skcipher_cipher_simple(tfm); in skcipher_setkey_simple() local
1114 crypto_cipher_clear_flags(cipher, CRYPTO_TFM_REQ_MASK); in skcipher_setkey_simple()
1115 crypto_cipher_set_flags(cipher, crypto_skcipher_get_flags(tfm) & in skcipher_setkey_simple()
1117 err = crypto_cipher_setkey(cipher, key, keylen); in skcipher_setkey_simple()
1118 crypto_skcipher_set_flags(tfm, crypto_cipher_get_flags(cipher) & in skcipher_setkey_simple()
[all …]
Dxcbc.c168 struct crypto_cipher *cipher; in xcbc_init_tfm() local
173 cipher = crypto_spawn_cipher(spawn); in xcbc_init_tfm()
174 if (IS_ERR(cipher)) in xcbc_init_tfm()
175 return PTR_ERR(cipher); in xcbc_init_tfm()
177 ctx->child = cipher; in xcbc_init_tfm()
Dcts.c292 struct crypto_skcipher *cipher; in crypto_cts_init_tfm() local
297 cipher = crypto_spawn_skcipher(spawn); in crypto_cts_init_tfm()
298 if (IS_ERR(cipher)) in crypto_cts_init_tfm()
299 return PTR_ERR(cipher); in crypto_cts_init_tfm()
301 ctx->child = cipher; in crypto_cts_init_tfm()
304 bsize = crypto_skcipher_blocksize(cipher); in crypto_cts_init_tfm()
306 crypto_skcipher_reqsize(cipher), in crypto_cts_init_tfm()
Dvmac.c52 struct crypto_cipher *cipher; member
443 err = crypto_cipher_setkey(tctx->cipher, key, keylen); in vmac_setkey()
450 crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in); in vmac_setkey()
460 crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in); in vmac_setkey()
471 crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in); in vmac_setkey()
589 crypto_cipher_encrypt_one(tctx->cipher, dctx->nonce.bytes, in vmac_final()
603 struct crypto_cipher *cipher; in vmac_init_tfm() local
605 cipher = crypto_spawn_cipher(spawn); in vmac_init_tfm()
606 if (IS_ERR(cipher)) in vmac_init_tfm()
607 return PTR_ERR(cipher); in vmac_init_tfm()
[all …]
Dlrw.c269 struct crypto_skcipher *cipher; in init_tfm() local
271 cipher = crypto_spawn_skcipher(spawn); in init_tfm()
272 if (IS_ERR(cipher)) in init_tfm()
273 return PTR_ERR(cipher); in init_tfm()
275 ctx->child = cipher; in init_tfm()
277 crypto_skcipher_set_reqsize(tfm, crypto_skcipher_reqsize(cipher) + in init_tfm()
Dpcrypt.c161 struct crypto_aead *cipher; in pcrypt_aead_init_tfm() local
170 cipher = crypto_spawn_aead(&ictx->spawn); in pcrypt_aead_init_tfm()
172 if (IS_ERR(cipher)) in pcrypt_aead_init_tfm()
173 return PTR_ERR(cipher); in pcrypt_aead_init_tfm()
175 ctx->child = cipher; in pcrypt_aead_init_tfm()
178 crypto_aead_reqsize(cipher)); in pcrypt_aead_init_tfm()
Dblkcipher.c371 struct blkcipher_alg *cipher = &tfm->__crt_alg->cra_blkcipher; in setkey_unaligned() local
384 ret = cipher->setkey(tfm, alignbuffer, keylen); in setkey_unaligned()
392 struct blkcipher_alg *cipher = &tfm->__crt_alg->cra_blkcipher; in setkey() local
395 if (keylen < cipher->min_keysize || keylen > cipher->max_keysize) { in setkey()
403 return cipher->setkey(tfm, key, keylen); in setkey()
442 struct blkcipher_alg *cipher = &alg->cra_blkcipher; in crypto_blkcipher_ctxsize() local
446 cipher->ivsize) { in crypto_blkcipher_ctxsize()
448 len += cipher->ivsize; in crypto_blkcipher_ctxsize()
Dcmac.c202 struct crypto_cipher *cipher; in cmac_init_tfm() local
207 cipher = crypto_spawn_cipher(spawn); in cmac_init_tfm()
208 if (IS_ERR(cipher)) in cmac_init_tfm()
209 return PTR_ERR(cipher); in cmac_init_tfm()
211 ctx->child = cipher; in cmac_init_tfm()
Dablkcipher.c298 struct ablkcipher_alg *cipher = crypto_ablkcipher_alg(tfm); in setkey_unaligned() local
311 ret = cipher->setkey(tfm, alignbuffer, keylen); in setkey_unaligned()
320 struct ablkcipher_alg *cipher = crypto_ablkcipher_alg(tfm); in setkey() local
323 if (keylen < cipher->min_keysize || keylen > cipher->max_keysize) { in setkey()
331 return cipher->setkey(tfm, key, keylen); in setkey()
Dcrypto_user_stat.c56 rcipher.stat_encrypt_cnt = atomic64_read(&alg->stats.cipher.encrypt_cnt); in crypto_report_cipher()
57 rcipher.stat_encrypt_tlen = atomic64_read(&alg->stats.cipher.encrypt_tlen); in crypto_report_cipher()
58 rcipher.stat_decrypt_cnt = atomic64_read(&alg->stats.cipher.decrypt_cnt); in crypto_report_cipher()
59 rcipher.stat_decrypt_tlen = atomic64_read(&alg->stats.cipher.decrypt_tlen); in crypto_report_cipher()
60 rcipher.stat_err_cnt = atomic64_read(&alg->stats.cipher.err_cnt); in crypto_report_cipher()
Dalgapi.c1059 atomic64_inc(&alg->stats.cipher.err_cnt); in crypto_stats_ablkcipher_encrypt()
1061 atomic64_inc(&alg->stats.cipher.encrypt_cnt); in crypto_stats_ablkcipher_encrypt()
1062 atomic64_add(nbytes, &alg->stats.cipher.encrypt_tlen); in crypto_stats_ablkcipher_encrypt()
1072 atomic64_inc(&alg->stats.cipher.err_cnt); in crypto_stats_ablkcipher_decrypt()
1074 atomic64_inc(&alg->stats.cipher.decrypt_cnt); in crypto_stats_ablkcipher_decrypt()
1075 atomic64_add(nbytes, &alg->stats.cipher.decrypt_tlen); in crypto_stats_ablkcipher_decrypt()
1258 atomic64_inc(&alg->stats.cipher.err_cnt); in crypto_stats_skcipher_encrypt()
1260 atomic64_inc(&alg->stats.cipher.encrypt_cnt); in crypto_stats_skcipher_encrypt()
1261 atomic64_add(cryptlen, &alg->stats.cipher.encrypt_tlen); in crypto_stats_skcipher_encrypt()
1271 atomic64_inc(&alg->stats.cipher.err_cnt); in crypto_stats_skcipher_decrypt()
[all …]
Dcryptd.c369 struct crypto_skcipher *cipher; in cryptd_skcipher_init_tfm() local
371 cipher = crypto_spawn_skcipher(spawn); in cryptd_skcipher_init_tfm()
372 if (IS_ERR(cipher)) in cryptd_skcipher_init_tfm()
373 return PTR_ERR(cipher); in cryptd_skcipher_init_tfm()
375 ctx->child = (struct crypto_sync_skcipher *)cipher; in cryptd_skcipher_init_tfm()
833 struct crypto_aead *cipher; in cryptd_aead_init_tfm() local
835 cipher = crypto_spawn_aead(spawn); in cryptd_aead_init_tfm()
836 if (IS_ERR(cipher)) in cryptd_aead_init_tfm()
837 return PTR_ERR(cipher); in cryptd_aead_init_tfm()
839 ctx->child = cipher; in cryptd_aead_init_tfm()
[all …]
Ddes_generic.c99 .cra_u = { .cipher = {
113 .cra_u = { .cipher = {
Dtea.c225 .cra_u = { .cipher = {
239 .cra_u = { .cipher = {
253 .cra_u = { .cipher = {
Dccm.c897 struct crypto_cipher *cipher; in cbcmac_init_tfm() local
902 cipher = crypto_spawn_cipher(spawn); in cbcmac_init_tfm()
903 if (IS_ERR(cipher)) in cbcmac_init_tfm()
904 return PTR_ERR(cipher); in cbcmac_init_tfm()
906 ctx->child = cipher; in cbcmac_init_tfm()
Dblowfish_generic.c113 .cra_u = { .cipher = {
Dsm4_generic.c220 .cipher = {

12