Home
last modified time | relevance | path

Searched refs:base (Results 1 – 25 of 85) sorted by relevance

1234

/crypto/
Dsimd.c54 struct crypto_skcipher *child = &ctx->cryptd_tfm->base; in simd_skcipher_setkey()
74 child = &ctx->cryptd_tfm->base; in simd_skcipher_encrypt()
95 child = &ctx->cryptd_tfm->base; in simd_skcipher_decrypt()
131 reqsize = max(reqsize, crypto_skcipher_reqsize(&cryptd_tfm->base)); in simd_skcipher_init()
166 if (snprintf(alg->base.cra_name, CRYPTO_MAX_ALG_NAME, "%s", algname) >= in simd_skcipher_create_compat()
170 if (snprintf(alg->base.cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s", in simd_skcipher_create_compat()
174 alg->base.cra_flags = CRYPTO_ALG_ASYNC | in simd_skcipher_create_compat()
175 (ialg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS); in simd_skcipher_create_compat()
176 alg->base.cra_priority = ialg->base.cra_priority; in simd_skcipher_create_compat()
177 alg->base.cra_blocksize = ialg->base.cra_blocksize; in simd_skcipher_create_compat()
[all …]
Dchacha_generic.c71 .base.cra_name = "chacha20",
72 .base.cra_driver_name = "chacha20-generic",
73 .base.cra_priority = 100,
74 .base.cra_blocksize = 1,
75 .base.cra_ctxsize = sizeof(struct chacha_ctx),
76 .base.cra_module = THIS_MODULE,
86 .base.cra_name = "xchacha20",
87 .base.cra_driver_name = "xchacha20-generic",
88 .base.cra_priority = 100,
89 .base.cra_blocksize = 1,
[all …]
Dkpp.c63 kpp->base.exit = crypto_kpp_exit_tfm; in crypto_kpp_init_tfm()
81 .tfmsize = offsetof(struct crypto_kpp, base),
92 struct crypto_alg *base = &alg->base; in kpp_prepare_alg() local
94 base->cra_type = &crypto_kpp_type; in kpp_prepare_alg()
95 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in kpp_prepare_alg()
96 base->cra_flags |= CRYPTO_ALG_TYPE_KPP; in kpp_prepare_alg()
101 struct crypto_alg *base = &alg->base; in crypto_register_kpp() local
104 return crypto_register_alg(base); in crypto_register_kpp()
110 crypto_unregister_alg(&alg->base); in crypto_unregister_kpp()
Dakcipher.c64 akcipher->base.exit = crypto_akcipher_exit_tfm; in crypto_akcipher_init_tfm()
90 .tfmsize = offsetof(struct crypto_akcipher, base),
97 spawn->base.frontend = &crypto_akcipher_type; in crypto_grab_akcipher()
98 return crypto_grab_spawn(&spawn->base, inst, name, type, mask); in crypto_grab_akcipher()
111 struct crypto_alg *base = &alg->base; in akcipher_prepare_alg() local
113 base->cra_type = &crypto_akcipher_type; in akcipher_prepare_alg()
114 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in akcipher_prepare_alg()
115 base->cra_flags |= CRYPTO_ALG_TYPE_AKCIPHER; in akcipher_prepare_alg()
131 struct crypto_alg *base = &alg->base; in crypto_register_akcipher() local
145 return crypto_register_alg(base); in crypto_register_akcipher()
[all …]
Dcryptd.c248 rctx->complete(&req->base, err); in cryptd_skcipher_complete()
255 static void cryptd_skcipher_encrypt(struct crypto_async_request *base, in cryptd_skcipher_encrypt() argument
258 struct skcipher_request *req = skcipher_request_cast(base); in cryptd_skcipher_encrypt()
277 req->base.complete = rctx->complete; in cryptd_skcipher_encrypt()
283 static void cryptd_skcipher_decrypt(struct crypto_async_request *base, in cryptd_skcipher_decrypt() argument
286 struct skcipher_request *req = skcipher_request_cast(base); in cryptd_skcipher_decrypt()
305 req->base.complete = rctx->complete; in cryptd_skcipher_decrypt()
319 rctx->complete = req->base.complete; in cryptd_skcipher_enqueue()
320 req->base.complete = compl; in cryptd_skcipher_enqueue()
322 return cryptd_enqueue_request(queue, &req->base); in cryptd_skcipher_enqueue()
[all …]
Dahash.c131 walk->flags = req->base.flags; in crypto_hash_walk_first()
168 !(alg->base.cra_flags & CRYPTO_ALG_OPTIONAL_KEY)) in ahash_set_needkey()
207 (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? in ahash_save_req()
238 priv->complete = req->base.complete; in ahash_save_req()
239 priv->data = req->base.data; in ahash_save_req()
240 priv->flags = req->base.flags; in ahash_save_req()
249 req->base.complete = cplt; in ahash_save_req()
250 req->base.data = req; in ahash_save_req()
307 areq->base.complete(&areq->base, err); in ahash_op_unaligned_done()
343 struct crypto_alg *alg = tfm->base.__crt_alg; in crypto_ahash_final()
[all …]
Daead.c86 struct crypto_alg *alg = aead->base.__crt_alg; in crypto_aead_encrypt()
103 struct crypto_alg *alg = aead->base.__crt_alg; in crypto_aead_decrypt()
137 aead->base.exit = crypto_aead_exit_tfm; in crypto_aead_init_tfm()
149 struct aead_alg *aead = container_of(alg, struct aead_alg, base); in crypto_aead_report()
173 struct aead_alg *aead = container_of(alg, struct aead_alg, base); in crypto_aead_show()
202 .tfmsize = offsetof(struct crypto_aead, base),
209 spawn->base.frontend = &crypto_aead_type; in crypto_grab_aead()
210 return crypto_grab_spawn(&spawn->base, inst, name, type, mask); in crypto_grab_aead()
222 struct crypto_alg *base = &alg->base; in aead_prepare_alg() local
229 alg->chunksize = base->cra_blocksize; in aead_prepare_alg()
[all …]
Dcts.c121 skcipher_request_set_callback(subreq, req->base.flags & in cts_cbc_encrypt()
159 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_encrypt()
160 req->base.complete, in crypto_cts_encrypt()
161 req->base.data); in crypto_cts_encrypt()
170 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_encrypt()
214 skcipher_request_set_callback(subreq, req->base.flags & in cts_cbc_decrypt()
254 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_decrypt()
255 req->base.complete, in crypto_cts_decrypt()
256 req->base.data); in crypto_cts_decrypt()
262 skcipher_request_set_callback(subreq, req->base.flags, in crypto_cts_decrypt()
[all …]
Dgeniv.c77 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in aead_geniv_alloc()
78 "%s(%s)", tmpl->name, alg->base.cra_name) >= in aead_geniv_alloc()
81 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in aead_geniv_alloc()
82 "%s(%s)", tmpl->name, alg->base.cra_driver_name) >= in aead_geniv_alloc()
86 inst->alg.base.cra_priority = alg->base.cra_priority; in aead_geniv_alloc()
87 inst->alg.base.cra_blocksize = alg->base.cra_blocksize; in aead_geniv_alloc()
88 inst->alg.base.cra_alignmask = alg->base.cra_alignmask; in aead_geniv_alloc()
89 inst->alg.base.cra_ctxsize = sizeof(struct aead_geniv_ctx); in aead_geniv_alloc()
Dblake2s_generic.c28 .base.cra_name = name, \
29 .base.cra_driver_name = driver_name, \
30 .base.cra_priority = 100, \
31 .base.cra_flags = CRYPTO_ALG_OPTIONAL_KEY, \
32 .base.cra_blocksize = BLAKE2S_BLOCK_SIZE, \
33 .base.cra_ctxsize = sizeof(struct blake2s_tfm_ctx), \
34 .base.cra_module = THIS_MODULE, \
Dseqiv.c39 static void seqiv_aead_encrypt_complete(struct crypto_async_request *base, in seqiv_aead_encrypt_complete() argument
42 struct aead_request *req = base->data; in seqiv_aead_encrypt_complete()
64 compl = req->base.complete; in seqiv_aead_encrypt()
65 data = req->base.data; in seqiv_aead_encrypt()
72 skcipher_request_set_callback(nreq, req->base.flags, in seqiv_aead_encrypt()
85 info = kmemdup(req->iv, ivsize, req->base.flags & in seqiv_aead_encrypt()
95 aead_request_set_callback(subreq, req->base.flags, compl, data); in seqiv_aead_encrypt()
123 compl = req->base.complete; in seqiv_aead_decrypt()
124 data = req->base.data; in seqiv_aead_decrypt()
126 aead_request_set_callback(subreq, req->base.flags, compl, data); in seqiv_aead_decrypt()
[all …]
Dsha3_generic.c246 .base.cra_name = "sha3-224",
247 .base.cra_driver_name = "sha3-224-generic",
248 .base.cra_blocksize = SHA3_224_BLOCK_SIZE,
249 .base.cra_module = THIS_MODULE,
256 .base.cra_name = "sha3-256",
257 .base.cra_driver_name = "sha3-256-generic",
258 .base.cra_blocksize = SHA3_256_BLOCK_SIZE,
259 .base.cra_module = THIS_MODULE,
266 .base.cra_name = "sha3-384",
267 .base.cra_driver_name = "sha3-384-generic",
[all …]
Drng.c33 struct crypto_alg *alg = tfm->base.__crt_alg; in crypto_rng_reset()
64 struct rng_alg *ralg = container_of(alg, struct rng_alg, base); in seedsize()
107 .tfmsize = offsetof(struct crypto_rng, base),
179 struct crypto_alg *base = &alg->base; in crypto_register_rng() local
184 base->cra_type = &crypto_rng_type; in crypto_register_rng()
185 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in crypto_register_rng()
186 base->cra_flags |= CRYPTO_ALG_TYPE_RNG; in crypto_register_rng()
188 return crypto_register_alg(base); in crypto_register_rng()
194 crypto_unregister_alg(&alg->base); in crypto_unregister_rng()
Dfips140-alg-registration.c149 fips140_load_init_function(&alg->base); in fips140_aead_init_tfm()
151 if (fips140_wait_until_ready(&alg->base)) in fips140_aead_init_tfm()
161 fips140_load_init_function(&halg->base); in fips140_ahash_init_tfm()
163 if (fips140_wait_until_ready(&halg->base)) in fips140_ahash_init_tfm()
172 fips140_load_init_function(&alg->base); in fips140_shash_init_tfm()
174 if (fips140_wait_until_ready(&alg->base)) in fips140_shash_init_tfm()
183 fips140_load_init_function(&alg->base); in fips140_skcipher_init_tfm()
185 if (fips140_wait_until_ready(&alg->base)) in fips140_skcipher_init_tfm()
215 return prepare_alg(alg, &alg->base, init, fips140_aead_init_tfm); in fips140_prepare_aead_alg()
220 return prepare_alg(alg, &alg->halg.base, init_tfm, in fips140_prepare_ahash_alg()
[all …]
Dgcm.c604 if (strcmp(ghash->base.cra_name, "ghash") != 0 || in crypto_gcm_create_common()
616 if (strncmp(ctr->base.cra_name, "ctr(", 4) != 0 || in crypto_gcm_create_common()
618 ctr->base.cra_blocksize != 1) in crypto_gcm_create_common()
622 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in crypto_gcm_create_common()
623 "gcm(%s", ctr->base.cra_name + 4) >= CRYPTO_MAX_ALG_NAME) in crypto_gcm_create_common()
626 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in crypto_gcm_create_common()
627 "gcm_base(%s,%s)", ctr->base.cra_driver_name, in crypto_gcm_create_common()
628 ghash->base.cra_driver_name) >= in crypto_gcm_create_common()
632 inst->alg.base.cra_priority = (ghash->base.cra_priority + in crypto_gcm_create_common()
633 ctr->base.cra_priority) / 2; in crypto_gcm_create_common()
[all …]
Dccm.c474 if (strncmp(mac->base.cra_name, "cbcmac(", 7) != 0 || in crypto_ccm_create_common()
486 if (strncmp(ctr->base.cra_name, "ctr(", 4) != 0 || in crypto_ccm_create_common()
488 ctr->base.cra_blocksize != 1) in crypto_ccm_create_common()
492 if (strcmp(ctr->base.cra_name + 4, mac->base.cra_name + 7) != 0) in crypto_ccm_create_common()
496 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in crypto_ccm_create_common()
497 "ccm(%s", ctr->base.cra_name + 4) >= CRYPTO_MAX_ALG_NAME) in crypto_ccm_create_common()
500 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in crypto_ccm_create_common()
501 "ccm_base(%s,%s)", ctr->base.cra_driver_name, in crypto_ccm_create_common()
502 mac->base.cra_driver_name) >= CRYPTO_MAX_ALG_NAME) in crypto_ccm_create_common()
505 inst->alg.base.cra_priority = (mac->base.cra_priority + in crypto_ccm_create_common()
[all …]
Darc4.c57 .base.cra_name = "ecb(arc4)",
58 .base.cra_driver_name = "ecb(arc4)-generic",
59 .base.cra_priority = 100,
60 .base.cra_blocksize = ARC4_BLOCK_SIZE,
61 .base.cra_ctxsize = sizeof(struct arc4_ctx),
62 .base.cra_module = THIS_MODULE,
Ddh.c37 static int _compute_val(const struct dh_ctx *ctx, MPI base, MPI val) in _compute_val() argument
40 return mpi_powm(val, base, ctx->xa, ctx->p); in _compute_val()
154 MPI base, val = mpi_alloc(0); in dh_compute_value() local
167 base = mpi_read_raw_from_sgl(req->src, req->src_len); in dh_compute_value()
168 if (!base) { in dh_compute_value()
172 ret = dh_is_pubkey_valid(ctx, base); in dh_compute_value()
176 base = ctx->g; in dh_compute_value()
179 ret = _compute_val(ctx, base, val); in dh_compute_value()
228 mpi_free(base); in dh_compute_value()
254 .base = {
Dctr.c149 inst->alg.base.cra_blocksize = 1; in crypto_ctr_create()
210 skcipher_request_set_callback(subreq, req->base.flags, in crypto_rfc3686_crypt()
211 req->base.complete, req->base.data); in crypto_rfc3686_crypt()
289 if (alg->base.cra_blocksize != 1) in crypto_rfc3686_create()
293 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in crypto_rfc3686_create()
294 "rfc3686(%s)", alg->base.cra_name) >= CRYPTO_MAX_ALG_NAME) in crypto_rfc3686_create()
296 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in crypto_rfc3686_create()
297 "rfc3686(%s)", alg->base.cra_driver_name) >= in crypto_rfc3686_create()
301 inst->alg.base.cra_priority = alg->base.cra_priority; in crypto_rfc3686_create()
302 inst->alg.base.cra_blocksize = 1; in crypto_rfc3686_create()
[all …]
Dechainiv.c48 skcipher_request_set_callback(nreq, req->base.flags, in echainiv_encrypt()
59 aead_request_set_callback(subreq, req->base.flags, in echainiv_encrypt()
60 req->base.complete, req->base.data); in echainiv_encrypt()
99 compl = req->base.complete; in echainiv_decrypt()
100 data = req->base.data; in echainiv_decrypt()
102 aead_request_set_callback(subreq, req->base.flags, compl, data); in echainiv_decrypt()
133 inst->alg.base.cra_ctxsize = sizeof(struct aead_geniv_ctx); in echainiv_aead_create()
134 inst->alg.base.cra_ctxsize += inst->alg.ivsize; in echainiv_aead_create()
Dacompress.c74 acomp->base.exit = crypto_acomp_exit_tfm; in crypto_acomp_init_tfm()
102 .tfmsize = offsetof(struct crypto_acomp, base),
152 struct crypto_alg *base = &alg->base; in crypto_register_acomp() local
154 base->cra_type = &crypto_acomp_type; in crypto_register_acomp()
155 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; in crypto_register_acomp()
156 base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS; in crypto_register_acomp()
158 return crypto_register_alg(base); in crypto_register_acomp()
164 crypto_unregister_alg(&alg->base); in crypto_unregister_acomp()
Dadiantum.c381 req->base.flags, in adiantum_crypt()
476 if (strcmp(streamcipher_alg->base.cra_name, "xchacha12") != 0 && in adiantum_supported_algorithms()
477 strcmp(streamcipher_alg->base.cra_name, "xchacha20") != 0) in adiantum_supported_algorithms()
486 if (strcmp(hash_alg->base.cra_name, "nhpoly1305") != 0) in adiantum_supported_algorithms()
543 streamcipher_alg->base.cra_name, in adiantum_create()
544 blockcipher_alg->cra_name, hash_alg->base.cra_name); in adiantum_create()
552 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in adiantum_create()
553 "adiantum(%s,%s)", streamcipher_alg->base.cra_name, in adiantum_create()
556 if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME, in adiantum_create()
558 streamcipher_alg->base.cra_driver_name, in adiantum_create()
[all …]
Dessiv.c451 struct crypto_alg *base, *block_base; in essiv_create() local
484 base = &skcipher_inst->alg.base; in essiv_create()
493 block_base = &skcipher_alg->base; in essiv_create()
503 base = &aead_inst->alg.base; in essiv_create()
512 block_base = &aead_alg->base; in essiv_create()
545 block_base->cra_name, hash_alg->base.cra_name); in essiv_create()
551 strlcpy(ictx->shash_driver_name, hash_alg->base.cra_driver_name, in essiv_create()
557 if (snprintf(base->cra_name, CRYPTO_MAX_ALG_NAME, in essiv_create()
559 hash_alg->base.cra_name) >= CRYPTO_MAX_ALG_NAME) in essiv_create()
561 if (snprintf(base->cra_driver_name, CRYPTO_MAX_ALG_NAME, in essiv_create()
[all …]
Dcurve25519-generic.c63 .base.cra_name = "curve25519",
64 .base.cra_driver_name = "curve25519-generic",
65 .base.cra_priority = 100,
66 .base.cra_module = THIS_MODULE,
67 .base.cra_ctxsize = CURVE25519_KEY_SIZE,
Dlrw.c215 rctx->subreq.base.flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in lrw_crypt_done()
230 skcipher_request_set_callback(subreq, req->base.flags, lrw_crypt_done, in lrw_init_crypt()
342 if (alg->base.cra_blocksize != LRW_BLOCK_SIZE) in lrw_create()
349 &alg->base); in lrw_create()
354 cipher_name = alg->base.cra_name; in lrw_create()
371 if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME, in lrw_create()
379 inst->alg.base.cra_priority = alg->base.cra_priority; in lrw_create()
380 inst->alg.base.cra_blocksize = LRW_BLOCK_SIZE; in lrw_create()
381 inst->alg.base.cra_alignmask = alg->base.cra_alignmask | in lrw_create()
390 inst->alg.base.cra_ctxsize = sizeof(struct lrw_tfm_ctx); in lrw_create()

1234