/crypto/ |
D | Kconfig | 164 cipher mode. Only select this option by hand if you expect to load 274 Support for the AEAD wrapper using the ChaCha20 stream cipher combined 307 This block cipher algorithm is required for IPSec. 316 This block cipher algorithm is required for IPSec. 335 This is the simplest block cipher algorithm. It simply encrypts 345 narrow block cipher mode for dm-crypt. Use it with cipher 348 rest is used to tie each cipher block to its logical position. 356 This block cipher algorithm is required for RxRPC. 364 XTS: IEEE1619/D16 narrow block cipher use with aes-xts-plain, 865 tristate "AES cipher algorithms" [all …]
|
D | cipher.c | 82 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; in cipher_encrypt_unaligned() local 85 cipher_crypt_unaligned(cipher->cia_encrypt, tfm, dst, src); in cipher_encrypt_unaligned() 89 cipher->cia_encrypt(tfm, dst, src); in cipher_encrypt_unaligned() 96 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; in cipher_decrypt_unaligned() local 99 cipher_crypt_unaligned(cipher->cia_decrypt, tfm, dst, src); in cipher_decrypt_unaligned() 103 cipher->cia_decrypt(tfm, dst, src); in cipher_decrypt_unaligned() 109 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; in crypto_init_cipher_ops() local 113 cipher_encrypt_unaligned : cipher->cia_encrypt; in crypto_init_cipher_ops() 115 cipher_decrypt_unaligned : cipher->cia_decrypt; in crypto_init_cipher_ops()
|
D | ccm.c | 26 struct crypto_spawn cipher; member 30 struct crypto_cipher *cipher; member 87 struct crypto_cipher *tfm = ctx->cipher; in crypto_ccm_setkey() 257 struct crypto_cipher *cipher = ctx->cipher; in crypto_ccm_auth() local 269 crypto_cipher_encrypt_one(cipher, odata, odata); in crypto_ccm_auth() 274 get_data_to_compute(cipher, pctx, req->src, req->assoclen); in crypto_ccm_auth() 281 get_data_to_compute(cipher, pctx, plain, cryptlen); in crypto_ccm_auth() 458 struct crypto_cipher *cipher; in crypto_ccm_init_tfm() local 463 cipher = crypto_spawn_cipher(&ictx->cipher); in crypto_ccm_init_tfm() 464 if (IS_ERR(cipher)) in crypto_ccm_init_tfm() [all …]
|
D | xts.c | 238 struct crypto_cipher *cipher; in init_tfm() local 244 cipher = crypto_spawn_cipher(spawn); in init_tfm() 245 if (IS_ERR(cipher)) in init_tfm() 246 return PTR_ERR(cipher); in init_tfm() 248 if (crypto_cipher_blocksize(cipher) != XTS_BLOCK_SIZE) { in init_tfm() 250 crypto_free_cipher(cipher); in init_tfm() 254 ctx->child = cipher; in init_tfm() 256 cipher = crypto_spawn_cipher(spawn); in init_tfm() 257 if (IS_ERR(cipher)) { in init_tfm() 259 return PTR_ERR(cipher); in init_tfm() [all …]
|
D | testmgr.c | 137 struct cipher_test_suite cipher; member 1548 if (desc->suite.cipher.enc.vecs) { in alg_test_cipher() 1549 err = test_cipher(tfm, ENCRYPT, desc->suite.cipher.enc.vecs, in alg_test_cipher() 1550 desc->suite.cipher.enc.count); in alg_test_cipher() 1555 if (desc->suite.cipher.dec.vecs) in alg_test_cipher() 1556 err = test_cipher(tfm, DECRYPT, desc->suite.cipher.dec.vecs, in alg_test_cipher() 1557 desc->suite.cipher.dec.count); in alg_test_cipher() 1577 if (desc->suite.cipher.enc.vecs) { in alg_test_skcipher() 1578 err = test_skcipher(tfm, ENCRYPT, desc->suite.cipher.enc.vecs, in alg_test_skcipher() 1579 desc->suite.cipher.enc.count); in alg_test_skcipher() [all …]
|
D | ctr.c | 163 struct crypto_cipher *cipher; in crypto_ctr_init_tfm() local 165 cipher = crypto_spawn_cipher(spawn); in crypto_ctr_init_tfm() 166 if (IS_ERR(cipher)) in crypto_ctr_init_tfm() 167 return PTR_ERR(cipher); in crypto_ctr_init_tfm() 169 ctx->child = cipher; in crypto_ctr_init_tfm() 311 struct crypto_skcipher *cipher; in crypto_rfc3686_init_tfm() local 315 cipher = crypto_spawn_skcipher2(spawn); in crypto_rfc3686_init_tfm() 316 if (IS_ERR(cipher)) in crypto_rfc3686_init_tfm() 317 return PTR_ERR(cipher); in crypto_rfc3686_init_tfm() 319 ctx->child = cipher; in crypto_rfc3686_init_tfm() [all …]
|
D | ecb.c | 102 struct crypto_cipher *cipher; in crypto_ecb_init_tfm() local 104 cipher = crypto_spawn_cipher(spawn); in crypto_ecb_init_tfm() 105 if (IS_ERR(cipher)) in crypto_ecb_init_tfm() 106 return PTR_ERR(cipher); in crypto_ecb_init_tfm() 108 ctx->child = cipher; in crypto_ecb_init_tfm()
|
D | lrw.c | 298 struct crypto_cipher *cipher; in init_tfm() local 304 cipher = crypto_spawn_cipher(spawn); in init_tfm() 305 if (IS_ERR(cipher)) in init_tfm() 306 return PTR_ERR(cipher); in init_tfm() 308 if (crypto_cipher_blocksize(cipher) != LRW_BLOCK_SIZE) { in init_tfm() 310 crypto_free_cipher(cipher); in init_tfm() 314 ctx->child = cipher; in init_tfm()
|
D | xcbc.c | 179 struct crypto_cipher *cipher; in xcbc_init_tfm() local 184 cipher = crypto_spawn_cipher(spawn); in xcbc_init_tfm() 185 if (IS_ERR(cipher)) in xcbc_init_tfm() 186 return PTR_ERR(cipher); in xcbc_init_tfm() 188 ctx->child = cipher; in xcbc_init_tfm()
|
D | cts.c | 288 struct crypto_skcipher *cipher; in crypto_cts_init_tfm() local 293 cipher = crypto_spawn_skcipher2(spawn); in crypto_cts_init_tfm() 294 if (IS_ERR(cipher)) in crypto_cts_init_tfm() 295 return PTR_ERR(cipher); in crypto_cts_init_tfm() 297 ctx->child = cipher; in crypto_cts_init_tfm() 300 bsize = crypto_skcipher_blocksize(cipher); in crypto_cts_init_tfm() 302 crypto_skcipher_reqsize(cipher), in crypto_cts_init_tfm()
|
D | cmac.c | 205 struct crypto_cipher *cipher; in cmac_init_tfm() local 210 cipher = crypto_spawn_cipher(spawn); in cmac_init_tfm() 211 if (IS_ERR(cipher)) in cmac_init_tfm() 212 return PTR_ERR(cipher); in cmac_init_tfm() 214 ctx->child = cipher; in cmac_init_tfm()
|
D | blkcipher.c | 378 struct blkcipher_alg *cipher = &tfm->__crt_alg->cra_blkcipher; in setkey_unaligned() local 391 ret = cipher->setkey(tfm, alignbuffer, keylen); in setkey_unaligned() 399 struct blkcipher_alg *cipher = &tfm->__crt_alg->cra_blkcipher; in setkey() local 402 if (keylen < cipher->min_keysize || keylen > cipher->max_keysize) { in setkey() 410 return cipher->setkey(tfm, key, keylen); in setkey() 449 struct blkcipher_alg *cipher = &alg->cra_blkcipher; in crypto_blkcipher_ctxsize() local 453 cipher->ivsize) { in crypto_blkcipher_ctxsize() 455 len += cipher->ivsize; in crypto_blkcipher_ctxsize()
|
D | keywrap.c | 326 struct crypto_cipher *cipher; in crypto_kw_init_tfm() local 328 cipher = crypto_spawn_cipher(spawn); in crypto_kw_init_tfm() 329 if (IS_ERR(cipher)) in crypto_kw_init_tfm() 330 return PTR_ERR(cipher); in crypto_kw_init_tfm() 332 ctx->child = cipher; in crypto_kw_init_tfm()
|
D | cbc.c | 198 struct crypto_cipher *cipher; in crypto_cbc_init_tfm() local 200 cipher = crypto_spawn_cipher(spawn); in crypto_cbc_init_tfm() 201 if (IS_ERR(cipher)) in crypto_cbc_init_tfm() 202 return PTR_ERR(cipher); in crypto_cbc_init_tfm() 204 ctx->child = cipher; in crypto_cbc_init_tfm()
|
D | pcbc.c | 208 struct crypto_cipher *cipher; in crypto_pcbc_init_tfm() local 210 cipher = crypto_spawn_cipher(spawn); in crypto_pcbc_init_tfm() 211 if (IS_ERR(cipher)) in crypto_pcbc_init_tfm() 212 return PTR_ERR(cipher); in crypto_pcbc_init_tfm() 214 ctx->child = cipher; in crypto_pcbc_init_tfm()
|
D | skcipher.c | 228 struct skcipher_alg *cipher = crypto_skcipher_alg(tfm); in skcipher_setkey_unaligned() local 240 ret = cipher->setkey(tfm, alignbuffer, keylen); in skcipher_setkey_unaligned() 248 struct skcipher_alg *cipher = crypto_skcipher_alg(tfm); in skcipher_setkey() local 251 if (keylen < cipher->min_keysize || keylen > cipher->max_keysize) { in skcipher_setkey() 259 return cipher->setkey(tfm, key, keylen); in skcipher_setkey()
|
D | pcrypt.c | 228 struct crypto_aead *cipher; in pcrypt_aead_init_tfm() local 237 cipher = crypto_spawn_aead(&ictx->spawn); in pcrypt_aead_init_tfm() 239 if (IS_ERR(cipher)) in pcrypt_aead_init_tfm() 240 return PTR_ERR(cipher); in pcrypt_aead_init_tfm() 242 ctx->child = cipher; in pcrypt_aead_init_tfm() 245 crypto_aead_reqsize(cipher)); in pcrypt_aead_init_tfm()
|
D | ablkcipher.c | 307 struct ablkcipher_alg *cipher = crypto_ablkcipher_alg(tfm); in setkey_unaligned() local 320 ret = cipher->setkey(tfm, alignbuffer, keylen); in setkey_unaligned() 329 struct ablkcipher_alg *cipher = crypto_ablkcipher_alg(tfm); in setkey() local 332 if (keylen < cipher->min_keysize || keylen > cipher->max_keysize) { in setkey() 340 return cipher->setkey(tfm, key, keylen); in setkey()
|
D | cryptd.c | 310 struct crypto_blkcipher *cipher; in cryptd_blkcipher_init_tfm() local 312 cipher = crypto_spawn_blkcipher(spawn); in cryptd_blkcipher_init_tfm() 313 if (IS_ERR(cipher)) in cryptd_blkcipher_init_tfm() 314 return PTR_ERR(cipher); in cryptd_blkcipher_init_tfm() 316 ctx->child = cipher; in cryptd_blkcipher_init_tfm() 808 struct crypto_aead *cipher; in cryptd_aead_init_tfm() local 810 cipher = crypto_spawn_aead(spawn); in cryptd_aead_init_tfm() 811 if (IS_ERR(cipher)) in cryptd_aead_init_tfm() 812 return PTR_ERR(cipher); in cryptd_aead_init_tfm() 814 ctx->child = cipher; in cryptd_aead_init_tfm() [all …]
|
D | tea.c | 229 .cra_u = { .cipher = { 242 .cra_u = { .cipher = { 255 .cra_u = { .cipher = {
|
D | vmac.c | 624 struct crypto_cipher *cipher; in vmac_init_tfm() local 629 cipher = crypto_spawn_cipher(spawn); in vmac_init_tfm() 630 if (IS_ERR(cipher)) in vmac_init_tfm() 631 return PTR_ERR(cipher); in vmac_init_tfm() 633 ctx->child = cipher; in vmac_init_tfm()
|
D | speck.c | 260 .cipher = { 277 .cipher = {
|
D | arc4.c | 125 .cipher = {
|
D | crypto_null.c | 119 .cra_u = { .cipher = {
|
D | blowfish_generic.c | 118 .cra_u = { .cipher = {
|