/crypto/ |
D | simd.c | 54 struct crypto_skcipher *child = &ctx->cryptd_tfm->base; in simd_skcipher_setkey() 74 child = &ctx->cryptd_tfm->base; in simd_skcipher_encrypt() 95 child = &ctx->cryptd_tfm->base; in simd_skcipher_decrypt() 131 reqsize = max(reqsize, crypto_skcipher_reqsize(&cryptd_tfm->base)); in simd_skcipher_init() 166 if (snprintf(alg->base.cra_name, CRYPTO_MAX_ALG_NAME, "%s", algname) >= in simd_skcipher_create_compat() 170 if (snprintf(alg->base.cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s", in simd_skcipher_create_compat() 174 alg->base.cra_flags = CRYPTO_ALG_ASYNC | in simd_skcipher_create_compat() 175 (ialg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS); in simd_skcipher_create_compat() 176 alg->base.cra_priority = ialg->base.cra_priority; in simd_skcipher_create_compat() 177 alg->base.cra_blocksize = ialg->base.cra_blocksize; in simd_skcipher_create_compat() [all …]
|
D | chacha_generic.c | 71 .base.cra_name = "chacha20", 72 .base.cra_driver_name = "chacha20-generic", 73 .base.cra_priority = 100, 74 .base.cra_blocksize = 1, 75 .base.cra_ctxsize = sizeof(struct chacha_ctx), 76 .base.cra_module = THIS_MODULE, 86 .base.cra_name = "xchacha20", 87 .base.cra_driver_name = "xchacha20-generic", 88 .base.cra_priority = 100, 89 .base.cra_blocksize = 1, [all …]
|
D | kpp.c | 63 kpp->base.exit = crypto_kpp_exit_tfm; in crypto_kpp_init_tfm() 81 .tfmsize = offsetof(struct crypto_kpp, base), 92 struct crypto_alg *base = &alg->base; in kpp_prepare_alg() local 94 base->cra_type = &crypto_kpp_type; in kpp_prepare_alg() 95 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in kpp_prepare_alg() 96 base->cra_flags |= CRYPTO_ALG_TYPE_KPP; in kpp_prepare_alg() 101 struct crypto_alg *base = &alg->base; in crypto_register_kpp() local 104 return crypto_register_alg(base); in crypto_register_kpp() 110 crypto_unregister_alg(&alg->base); in crypto_unregister_kpp()
|
D | akcipher.c | 64 akcipher->base.exit = crypto_akcipher_exit_tfm; in crypto_akcipher_init_tfm() 90 .tfmsize = offsetof(struct crypto_akcipher, base), 97 spawn->base.frontend = &crypto_akcipher_type; in crypto_grab_akcipher() 98 return crypto_grab_spawn(&spawn->base, inst, name, type, mask); in crypto_grab_akcipher() 111 struct crypto_alg *base = &alg->base; in akcipher_prepare_alg() local 113 base->cra_type = &crypto_akcipher_type; in akcipher_prepare_alg() 114 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in akcipher_prepare_alg() 115 base->cra_flags |= CRYPTO_ALG_TYPE_AKCIPHER; in akcipher_prepare_alg() 131 struct crypto_alg *base = &alg->base; in crypto_register_akcipher() local 145 return crypto_register_alg(base); in crypto_register_akcipher() [all …]
|
D | cryptd.c | 248 rctx->complete(&req->base, err); in cryptd_skcipher_complete() 255 static void cryptd_skcipher_encrypt(struct crypto_async_request *base, in cryptd_skcipher_encrypt() argument 258 struct skcipher_request *req = skcipher_request_cast(base); in cryptd_skcipher_encrypt() 277 req->base.complete = rctx->complete; in cryptd_skcipher_encrypt() 283 static void cryptd_skcipher_decrypt(struct crypto_async_request *base, in cryptd_skcipher_decrypt() argument 286 struct skcipher_request *req = skcipher_request_cast(base); in cryptd_skcipher_decrypt() 305 req->base.complete = rctx->complete; in cryptd_skcipher_decrypt() 319 rctx->complete = req->base.complete; in cryptd_skcipher_enqueue() 320 req->base.complete = compl; in cryptd_skcipher_enqueue() 322 return cryptd_enqueue_request(queue, &req->base); in cryptd_skcipher_enqueue() [all …]
|
D | ahash.c | 131 walk->flags = req->base.flags; in crypto_hash_walk_first() 168 !(alg->base.cra_flags & CRYPTO_ALG_OPTIONAL_KEY)) in ahash_set_needkey() 207 (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? in ahash_save_req() 238 priv->complete = req->base.complete; in ahash_save_req() 239 priv->data = req->base.data; in ahash_save_req() 240 priv->flags = req->base.flags; in ahash_save_req() 249 req->base.complete = cplt; in ahash_save_req() 250 req->base.data = req; in ahash_save_req() 307 areq->base.complete(&areq->base, err); in ahash_op_unaligned_done() 343 struct crypto_alg *alg = tfm->base.__crt_alg; in crypto_ahash_final() [all …]
|
D | aead.c | 86 struct crypto_alg *alg = aead->base.__crt_alg; in crypto_aead_encrypt() 103 struct crypto_alg *alg = aead->base.__crt_alg; in crypto_aead_decrypt() 137 aead->base.exit = crypto_aead_exit_tfm; in crypto_aead_init_tfm() 149 struct aead_alg *aead = container_of(alg, struct aead_alg, base); in crypto_aead_report() 173 struct aead_alg *aead = container_of(alg, struct aead_alg, base); in crypto_aead_show() 202 .tfmsize = offsetof(struct crypto_aead, base), 209 spawn->base.frontend = &crypto_aead_type; in crypto_grab_aead() 210 return crypto_grab_spawn(&spawn->base, inst, name, type, mask); in crypto_grab_aead() 222 struct crypto_alg *base = &alg->base; in aead_prepare_alg() local 229 alg->chunksize = base->cra_blocksize; in aead_prepare_alg() [all …]
|
D | cts.c | 121 skcipher_request_set_callback(subreq, req->base.flags & in cts_cbc_encrypt() 159 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_encrypt() 160 req->base.complete, in crypto_cts_encrypt() 161 req->base.data); in crypto_cts_encrypt() 170 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_encrypt() 214 skcipher_request_set_callback(subreq, req->base.flags & in cts_cbc_decrypt() 254 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_decrypt() 255 req->base.complete, in crypto_cts_decrypt() 256 req->base.data); in crypto_cts_decrypt() 262 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_decrypt() [all …]
|
D | geniv.c | 77 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in aead_geniv_alloc() 78 "%s(%s)", tmpl->name, alg->base.cra_name) >= in aead_geniv_alloc() 81 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in aead_geniv_alloc() 82 "%s(%s)", tmpl->name, alg->base.cra_driver_name) >= in aead_geniv_alloc() 86 inst->alg.base.cra_priority = alg->base.cra_priority; in aead_geniv_alloc() 87 inst->alg.base.cra_blocksize = alg->base.cra_blocksize; in aead_geniv_alloc() 88 inst->alg.base.cra_alignmask = alg->base.cra_alignmask; in aead_geniv_alloc() 89 inst->alg.base.cra_ctxsize = sizeof(struct aead_geniv_ctx); in aead_geniv_alloc()
|
D | blake2s_generic.c | 28 .base.cra_name = name, \ 29 .base.cra_driver_name = driver_name, \ 30 .base.cra_priority = 100, \ 31 .base.cra_flags = CRYPTO_ALG_OPTIONAL_KEY, \ 32 .base.cra_blocksize = BLAKE2S_BLOCK_SIZE, \ 33 .base.cra_ctxsize = sizeof(struct blake2s_tfm_ctx), \ 34 .base.cra_module = THIS_MODULE, \
|
D | seqiv.c | 39 static void seqiv_aead_encrypt_complete(struct crypto_async_request *base, in seqiv_aead_encrypt_complete() argument 42 struct aead_request *req = base->data; in seqiv_aead_encrypt_complete() 64 compl = req->base.complete; in seqiv_aead_encrypt() 65 data = req->base.data; in seqiv_aead_encrypt() 72 skcipher_request_set_callback(nreq, req->base.flags, in seqiv_aead_encrypt() 85 info = kmemdup(req->iv, ivsize, req->base.flags & in seqiv_aead_encrypt() 95 aead_request_set_callback(subreq, req->base.flags, compl, data); in seqiv_aead_encrypt() 123 compl = req->base.complete; in seqiv_aead_decrypt() 124 data = req->base.data; in seqiv_aead_decrypt() 126 aead_request_set_callback(subreq, req->base.flags, compl, data); in seqiv_aead_decrypt() [all …]
|
D | sha3_generic.c | 246 .base.cra_name = "sha3-224", 247 .base.cra_driver_name = "sha3-224-generic", 248 .base.cra_blocksize = SHA3_224_BLOCK_SIZE, 249 .base.cra_module = THIS_MODULE, 256 .base.cra_name = "sha3-256", 257 .base.cra_driver_name = "sha3-256-generic", 258 .base.cra_blocksize = SHA3_256_BLOCK_SIZE, 259 .base.cra_module = THIS_MODULE, 266 .base.cra_name = "sha3-384", 267 .base.cra_driver_name = "sha3-384-generic", [all …]
|
D | rng.c | 33 struct crypto_alg *alg = tfm->base.__crt_alg; in crypto_rng_reset() 64 struct rng_alg *ralg = container_of(alg, struct rng_alg, base); in seedsize() 107 .tfmsize = offsetof(struct crypto_rng, base), 179 struct crypto_alg *base = &alg->base; in crypto_register_rng() local 184 base->cra_type = &crypto_rng_type; in crypto_register_rng() 185 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in crypto_register_rng() 186 base->cra_flags |= CRYPTO_ALG_TYPE_RNG; in crypto_register_rng() 188 return crypto_register_alg(base); in crypto_register_rng() 194 crypto_unregister_alg(&alg->base); in crypto_unregister_rng()
|
D | fips140-alg-registration.c | 149 fips140_load_init_function(&alg->base); in fips140_aead_init_tfm() 151 if (fips140_wait_until_ready(&alg->base)) in fips140_aead_init_tfm() 161 fips140_load_init_function(&halg->base); in fips140_ahash_init_tfm() 163 if (fips140_wait_until_ready(&halg->base)) in fips140_ahash_init_tfm() 172 fips140_load_init_function(&alg->base); in fips140_shash_init_tfm() 174 if (fips140_wait_until_ready(&alg->base)) in fips140_shash_init_tfm() 183 fips140_load_init_function(&alg->base); in fips140_skcipher_init_tfm() 185 if (fips140_wait_until_ready(&alg->base)) in fips140_skcipher_init_tfm() 215 return prepare_alg(alg, &alg->base, init, fips140_aead_init_tfm); in fips140_prepare_aead_alg() 220 return prepare_alg(alg, &alg->halg.base, init_tfm, in fips140_prepare_ahash_alg() [all …]
|
D | gcm.c | 604 if (strcmp(ghash->base.cra_name, "ghash") != 0 || in crypto_gcm_create_common() 616 if (strncmp(ctr->base.cra_name, "ctr(", 4) != 0 || in crypto_gcm_create_common() 618 ctr->base.cra_blocksize != 1) in crypto_gcm_create_common() 622 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in crypto_gcm_create_common() 623 "gcm(%s", ctr->base.cra_name + 4) >= CRYPTO_MAX_ALG_NAME) in crypto_gcm_create_common() 626 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in crypto_gcm_create_common() 627 "gcm_base(%s,%s)", ctr->base.cra_driver_name, in crypto_gcm_create_common() 628 ghash->base.cra_driver_name) >= in crypto_gcm_create_common() 632 inst->alg.base.cra_priority = (ghash->base.cra_priority + in crypto_gcm_create_common() 633 ctr->base.cra_priority) / 2; in crypto_gcm_create_common() [all …]
|
D | ccm.c | 474 if (strncmp(mac->base.cra_name, "cbcmac(", 7) != 0 || in crypto_ccm_create_common() 486 if (strncmp(ctr->base.cra_name, "ctr(", 4) != 0 || in crypto_ccm_create_common() 488 ctr->base.cra_blocksize != 1) in crypto_ccm_create_common() 492 if (strcmp(ctr->base.cra_name + 4, mac->base.cra_name + 7) != 0) in crypto_ccm_create_common() 496 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in crypto_ccm_create_common() 497 "ccm(%s", ctr->base.cra_name + 4) >= CRYPTO_MAX_ALG_NAME) in crypto_ccm_create_common() 500 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in crypto_ccm_create_common() 501 "ccm_base(%s,%s)", ctr->base.cra_driver_name, in crypto_ccm_create_common() 502 mac->base.cra_driver_name) >= CRYPTO_MAX_ALG_NAME) in crypto_ccm_create_common() 505 inst->alg.base.cra_priority = (mac->base.cra_priority + in crypto_ccm_create_common() [all …]
|
D | arc4.c | 57 .base.cra_name = "ecb(arc4)", 58 .base.cra_driver_name = "ecb(arc4)-generic", 59 .base.cra_priority = 100, 60 .base.cra_blocksize = ARC4_BLOCK_SIZE, 61 .base.cra_ctxsize = sizeof(struct arc4_ctx), 62 .base.cra_module = THIS_MODULE,
|
D | dh.c | 37 static int _compute_val(const struct dh_ctx *ctx, MPI base, MPI val) in _compute_val() argument 40 return mpi_powm(val, base, ctx->xa, ctx->p); in _compute_val() 154 MPI base, val = mpi_alloc(0); in dh_compute_value() local 167 base = mpi_read_raw_from_sgl(req->src, req->src_len); in dh_compute_value() 168 if (!base) { in dh_compute_value() 172 ret = dh_is_pubkey_valid(ctx, base); in dh_compute_value() 176 base = ctx->g; in dh_compute_value() 179 ret = _compute_val(ctx, base, val); in dh_compute_value() 228 mpi_free(base); in dh_compute_value() 254 .base = {
|
D | ctr.c | 149 inst->alg.base.cra_blocksize = 1; in crypto_ctr_create() 210 skcipher_request_set_callback(subreq, req->base.flags, in crypto_rfc3686_crypt() 211 req->base.complete, req->base.data); in crypto_rfc3686_crypt() 289 if (alg->base.cra_blocksize != 1) in crypto_rfc3686_create() 293 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in crypto_rfc3686_create() 294 "rfc3686(%s)", alg->base.cra_name) >= CRYPTO_MAX_ALG_NAME) in crypto_rfc3686_create() 296 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in crypto_rfc3686_create() 297 "rfc3686(%s)", alg->base.cra_driver_name) >= in crypto_rfc3686_create() 301 inst->alg.base.cra_priority = alg->base.cra_priority; in crypto_rfc3686_create() 302 inst->alg.base.cra_blocksize = 1; in crypto_rfc3686_create() [all …]
|
D | echainiv.c | 48 skcipher_request_set_callback(nreq, req->base.flags, in echainiv_encrypt() 59 aead_request_set_callback(subreq, req->base.flags, in echainiv_encrypt() 60 req->base.complete, req->base.data); in echainiv_encrypt() 99 compl = req->base.complete; in echainiv_decrypt() 100 data = req->base.data; in echainiv_decrypt() 102 aead_request_set_callback(subreq, req->base.flags, compl, data); in echainiv_decrypt() 133 inst->alg.base.cra_ctxsize = sizeof(struct aead_geniv_ctx); in echainiv_aead_create() 134 inst->alg.base.cra_ctxsize += inst->alg.ivsize; in echainiv_aead_create()
|
D | acompress.c | 74 acomp->base.exit = crypto_acomp_exit_tfm; in crypto_acomp_init_tfm() 102 .tfmsize = offsetof(struct crypto_acomp, base), 152 struct crypto_alg *base = &alg->base; in crypto_register_acomp() local 154 base->cra_type = &crypto_acomp_type; in crypto_register_acomp() 155 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in crypto_register_acomp() 156 base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS; in crypto_register_acomp() 158 return crypto_register_alg(base); in crypto_register_acomp() 164 crypto_unregister_alg(&alg->base); in crypto_unregister_acomp()
|
D | adiantum.c | 381 req->base.flags, in adiantum_crypt() 476 if (strcmp(streamcipher_alg->base.cra_name, "xchacha12") != 0 && in adiantum_supported_algorithms() 477 strcmp(streamcipher_alg->base.cra_name, "xchacha20") != 0) in adiantum_supported_algorithms() 486 if (strcmp(hash_alg->base.cra_name, "nhpoly1305") != 0) in adiantum_supported_algorithms() 543 streamcipher_alg->base.cra_name, in adiantum_create() 544 blockcipher_alg->cra_name, hash_alg->base.cra_name); in adiantum_create() 552 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in adiantum_create() 553 "adiantum(%s,%s)", streamcipher_alg->base.cra_name, in adiantum_create() 556 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in adiantum_create() 558 streamcipher_alg->base.cra_driver_name, in adiantum_create() [all …]
|
D | essiv.c | 451 struct crypto_alg *base, *block_base; in essiv_create() local 484 base = &skcipher_inst->alg.base; in essiv_create() 493 block_base = &skcipher_alg->base; in essiv_create() 503 base = &aead_inst->alg.base; in essiv_create() 512 block_base = &aead_alg->base; in essiv_create() 545 block_base->cra_name, hash_alg->base.cra_name); in essiv_create() 551 strlcpy(ictx->shash_driver_name, hash_alg->base.cra_driver_name, in essiv_create() 557 if (snprintf(base->cra_name, CRYPTO_MAX_ALG_NAME, in essiv_create() 559 hash_alg->base.cra_name) >= CRYPTO_MAX_ALG_NAME) in essiv_create() 561 if (snprintf(base->cra_driver_name, CRYPTO_MAX_ALG_NAME, in essiv_create() [all …]
|
D | curve25519-generic.c | 63 .base.cra_name = "curve25519", 64 .base.cra_driver_name = "curve25519-generic", 65 .base.cra_priority = 100, 66 .base.cra_module = THIS_MODULE, 67 .base.cra_ctxsize = CURVE25519_KEY_SIZE,
|
D | lrw.c | 215 rctx->subreq.base.flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in lrw_crypt_done() 230 skcipher_request_set_callback(subreq, req->base.flags, lrw_crypt_done, in lrw_init_crypt() 342 if (alg->base.cra_blocksize != LRW_BLOCK_SIZE) in lrw_create() 349 &alg->base); in lrw_create() 354 cipher_name = alg->base.cra_name; in lrw_create() 371 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in lrw_create() 379 inst->alg.base.cra_priority = alg->base.cra_priority; in lrw_create() 380 inst->alg.base.cra_blocksize = LRW_BLOCK_SIZE; in lrw_create() 381 inst->alg.base.cra_alignmask = alg->base.cra_alignmask | in lrw_create() 390 inst->alg.base.cra_ctxsize = sizeof(struct lrw_tfm_ctx); in lrw_create()
|