1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * sun4i-ss-cipher.c - hardware cryptographic accelerator for Allwinner A20 SoC
4 *
5 * Copyright (C) 2013-2015 Corentin LABBE <clabbe.montjoie@gmail.com>
6 *
7 * This file add support for AES cipher with 128,192,256 bits
8 * keysize in CBC and ECB mode.
9 * Add support also for DES and 3DES in CBC and ECB mode.
10 *
11 * You could find the datasheet in Documentation/arm/sunxi.rst
12 */
13 #include "sun4i-ss.h"
14
sun4i_ss_opti_poll(struct skcipher_request * areq)15 static int noinline_for_stack sun4i_ss_opti_poll(struct skcipher_request *areq)
16 {
17 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
18 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
19 struct sun4i_ss_ctx *ss = op->ss;
20 unsigned int ivsize = crypto_skcipher_ivsize(tfm);
21 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq);
22 u32 mode = ctx->mode;
23 /* when activating SS, the default FIFO space is SS_RX_DEFAULT(32) */
24 u32 rx_cnt = SS_RX_DEFAULT;
25 u32 tx_cnt = 0;
26 u32 spaces;
27 u32 v;
28 int err = 0;
29 unsigned int i;
30 unsigned int ileft = areq->cryptlen;
31 unsigned int oleft = areq->cryptlen;
32 unsigned int todo;
33 unsigned long pi = 0, po = 0; /* progress for in and out */
34 bool miter_err;
35 struct sg_mapping_iter mi, mo;
36 unsigned int oi, oo; /* offset for in and out */
37 unsigned long flags;
38 struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
39 struct sun4i_ss_alg_template *algt;
40
41 if (!areq->cryptlen)
42 return 0;
43
44 if (!areq->src || !areq->dst) {
45 dev_err_ratelimited(ss->dev, "ERROR: Some SGs are NULL\n");
46 return -EINVAL;
47 }
48
49 if (areq->iv && ivsize > 0 && mode & SS_DECRYPTION) {
50 scatterwalk_map_and_copy(ctx->backup_iv, areq->src,
51 areq->cryptlen - ivsize, ivsize, 0);
52 }
53
54 if (IS_ENABLED(CONFIG_CRYPTO_DEV_SUN4I_SS_DEBUG)) {
55 algt = container_of(alg, struct sun4i_ss_alg_template, alg.crypto);
56 algt->stat_opti++;
57 algt->stat_bytes += areq->cryptlen;
58 }
59
60 spin_lock_irqsave(&ss->slock, flags);
61
62 for (i = 0; i < op->keylen / 4; i++)
63 writesl(ss->base + SS_KEY0 + i * 4, &op->key[i], 1);
64
65 if (areq->iv) {
66 for (i = 0; i < 4 && i < ivsize / 4; i++) {
67 v = *(u32 *)(areq->iv + i * 4);
68 writesl(ss->base + SS_IV0 + i * 4, &v, 1);
69 }
70 }
71 writel(mode, ss->base + SS_CTL);
72
73
74 ileft = areq->cryptlen / 4;
75 oleft = areq->cryptlen / 4;
76 oi = 0;
77 oo = 0;
78 do {
79 if (ileft) {
80 sg_miter_start(&mi, areq->src, sg_nents(areq->src),
81 SG_MITER_FROM_SG | SG_MITER_ATOMIC);
82 if (pi)
83 sg_miter_skip(&mi, pi);
84 miter_err = sg_miter_next(&mi);
85 if (!miter_err || !mi.addr) {
86 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n");
87 err = -EINVAL;
88 goto release_ss;
89 }
90 todo = min(rx_cnt, ileft);
91 todo = min_t(size_t, todo, (mi.length - oi) / 4);
92 if (todo) {
93 ileft -= todo;
94 writesl(ss->base + SS_RXFIFO, mi.addr + oi, todo);
95 oi += todo * 4;
96 }
97 if (oi == mi.length) {
98 pi += mi.length;
99 oi = 0;
100 }
101 sg_miter_stop(&mi);
102 }
103
104 spaces = readl(ss->base + SS_FCSR);
105 rx_cnt = SS_RXFIFO_SPACES(spaces);
106 tx_cnt = SS_TXFIFO_SPACES(spaces);
107
108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst),
109 SG_MITER_TO_SG | SG_MITER_ATOMIC);
110 if (po)
111 sg_miter_skip(&mo, po);
112 miter_err = sg_miter_next(&mo);
113 if (!miter_err || !mo.addr) {
114 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n");
115 err = -EINVAL;
116 goto release_ss;
117 }
118 todo = min(tx_cnt, oleft);
119 todo = min_t(size_t, todo, (mo.length - oo) / 4);
120 if (todo) {
121 oleft -= todo;
122 readsl(ss->base + SS_TXFIFO, mo.addr + oo, todo);
123 oo += todo * 4;
124 }
125 if (oo == mo.length) {
126 oo = 0;
127 po += mo.length;
128 }
129 sg_miter_stop(&mo);
130 } while (oleft);
131
132 if (areq->iv) {
133 if (mode & SS_DECRYPTION) {
134 memcpy(areq->iv, ctx->backup_iv, ivsize);
135 memzero_explicit(ctx->backup_iv, ivsize);
136 } else {
137 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize,
138 ivsize, 0);
139 }
140 }
141
142 release_ss:
143 writel(0, ss->base + SS_CTL);
144 spin_unlock_irqrestore(&ss->slock, flags);
145 return err;
146 }
147
sun4i_ss_cipher_poll_fallback(struct skcipher_request * areq)148 static int noinline_for_stack sun4i_ss_cipher_poll_fallback(struct skcipher_request *areq)
149 {
150 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
151 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
152 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq);
153 int err;
154 struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
155 struct sun4i_ss_alg_template *algt;
156
157 if (IS_ENABLED(CONFIG_CRYPTO_DEV_SUN4I_SS_DEBUG)) {
158 algt = container_of(alg, struct sun4i_ss_alg_template, alg.crypto);
159 algt->stat_fb++;
160 }
161
162 skcipher_request_set_tfm(&ctx->fallback_req, op->fallback_tfm);
163 skcipher_request_set_callback(&ctx->fallback_req, areq->base.flags,
164 areq->base.complete, areq->base.data);
165 skcipher_request_set_crypt(&ctx->fallback_req, areq->src, areq->dst,
166 areq->cryptlen, areq->iv);
167 if (ctx->mode & SS_DECRYPTION)
168 err = crypto_skcipher_decrypt(&ctx->fallback_req);
169 else
170 err = crypto_skcipher_encrypt(&ctx->fallback_req);
171
172 return err;
173 }
174
175 /* Generic function that support SG with size not multiple of 4 */
sun4i_ss_cipher_poll(struct skcipher_request * areq)176 static int sun4i_ss_cipher_poll(struct skcipher_request *areq)
177 {
178 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
179 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
180 struct sun4i_ss_ctx *ss = op->ss;
181 int no_chunk = 1;
182 struct scatterlist *in_sg = areq->src;
183 struct scatterlist *out_sg = areq->dst;
184 unsigned int ivsize = crypto_skcipher_ivsize(tfm);
185 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq);
186 struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
187 struct sun4i_ss_alg_template *algt;
188 u32 mode = ctx->mode;
189 /* when activating SS, the default FIFO space is SS_RX_DEFAULT(32) */
190 u32 rx_cnt = SS_RX_DEFAULT;
191 u32 tx_cnt = 0;
192 u32 v;
193 u32 spaces;
194 int err = 0;
195 unsigned int i;
196 unsigned int ileft = areq->cryptlen;
197 unsigned int oleft = areq->cryptlen;
198 unsigned int todo;
199 struct sg_mapping_iter mi, mo;
200 unsigned long pi = 0, po = 0; /* progress for in and out */
201 bool miter_err;
202 unsigned int oi, oo; /* offset for in and out */
203 unsigned int ob = 0; /* offset in buf */
204 unsigned int obo = 0; /* offset in bufo*/
205 unsigned int obl = 0; /* length of data in bufo */
206 unsigned long flags;
207 bool need_fallback = false;
208
209 if (!areq->cryptlen)
210 return 0;
211
212 if (!areq->src || !areq->dst) {
213 dev_err_ratelimited(ss->dev, "ERROR: Some SGs are NULL\n");
214 return -EINVAL;
215 }
216
217 algt = container_of(alg, struct sun4i_ss_alg_template, alg.crypto);
218 if (areq->cryptlen % algt->alg.crypto.base.cra_blocksize)
219 need_fallback = true;
220
221 /*
222 * if we have only SGs with size multiple of 4,
223 * we can use the SS optimized function
224 */
225 while (in_sg && no_chunk == 1) {
226 if ((in_sg->length | in_sg->offset) & 3u)
227 no_chunk = 0;
228 in_sg = sg_next(in_sg);
229 }
230 while (out_sg && no_chunk == 1) {
231 if ((out_sg->length | out_sg->offset) & 3u)
232 no_chunk = 0;
233 out_sg = sg_next(out_sg);
234 }
235
236 if (no_chunk == 1 && !need_fallback)
237 return sun4i_ss_opti_poll(areq);
238
239 if (need_fallback)
240 return sun4i_ss_cipher_poll_fallback(areq);
241
242 if (areq->iv && ivsize > 0 && mode & SS_DECRYPTION) {
243 scatterwalk_map_and_copy(ctx->backup_iv, areq->src,
244 areq->cryptlen - ivsize, ivsize, 0);
245 }
246
247 if (IS_ENABLED(CONFIG_CRYPTO_DEV_SUN4I_SS_DEBUG)) {
248 algt->stat_req++;
249 algt->stat_bytes += areq->cryptlen;
250 }
251
252 spin_lock_irqsave(&ss->slock, flags);
253
254 for (i = 0; i < op->keylen / 4; i++)
255 writesl(ss->base + SS_KEY0 + i * 4, &op->key[i], 1);
256
257 if (areq->iv) {
258 for (i = 0; i < 4 && i < ivsize / 4; i++) {
259 v = *(u32 *)(areq->iv + i * 4);
260 writesl(ss->base + SS_IV0 + i * 4, &v, 1);
261 }
262 }
263 writel(mode, ss->base + SS_CTL);
264
265 ileft = areq->cryptlen;
266 oleft = areq->cryptlen;
267 oi = 0;
268 oo = 0;
269
270 while (oleft) {
271 if (ileft) {
272 sg_miter_start(&mi, areq->src, sg_nents(areq->src),
273 SG_MITER_FROM_SG | SG_MITER_ATOMIC);
274 if (pi)
275 sg_miter_skip(&mi, pi);
276 miter_err = sg_miter_next(&mi);
277 if (!miter_err || !mi.addr) {
278 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n");
279 err = -EINVAL;
280 goto release_ss;
281 }
282 /*
283 * todo is the number of consecutive 4byte word that we
284 * can read from current SG
285 */
286 todo = min(rx_cnt, ileft / 4);
287 todo = min_t(size_t, todo, (mi.length - oi) / 4);
288 if (todo && !ob) {
289 writesl(ss->base + SS_RXFIFO, mi.addr + oi,
290 todo);
291 ileft -= todo * 4;
292 oi += todo * 4;
293 } else {
294 /*
295 * not enough consecutive bytes, so we need to
296 * linearize in buf. todo is in bytes
297 * After that copy, if we have a multiple of 4
298 * we need to be able to write all buf in one
299 * pass, so it is why we min() with rx_cnt
300 */
301 todo = min(rx_cnt * 4 - ob, ileft);
302 todo = min_t(size_t, todo, mi.length - oi);
303 memcpy(ss->buf + ob, mi.addr + oi, todo);
304 ileft -= todo;
305 oi += todo;
306 ob += todo;
307 if (!(ob % 4)) {
308 writesl(ss->base + SS_RXFIFO, ss->buf,
309 ob / 4);
310 ob = 0;
311 }
312 }
313 if (oi == mi.length) {
314 pi += mi.length;
315 oi = 0;
316 }
317 sg_miter_stop(&mi);
318 }
319
320 spaces = readl(ss->base + SS_FCSR);
321 rx_cnt = SS_RXFIFO_SPACES(spaces);
322 tx_cnt = SS_TXFIFO_SPACES(spaces);
323
324 if (!tx_cnt)
325 continue;
326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst),
327 SG_MITER_TO_SG | SG_MITER_ATOMIC);
328 if (po)
329 sg_miter_skip(&mo, po);
330 miter_err = sg_miter_next(&mo);
331 if (!miter_err || !mo.addr) {
332 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n");
333 err = -EINVAL;
334 goto release_ss;
335 }
336 /* todo in 4bytes word */
337 todo = min(tx_cnt, oleft / 4);
338 todo = min_t(size_t, todo, (mo.length - oo) / 4);
339
340 if (todo) {
341 readsl(ss->base + SS_TXFIFO, mo.addr + oo, todo);
342 oleft -= todo * 4;
343 oo += todo * 4;
344 if (oo == mo.length) {
345 po += mo.length;
346 oo = 0;
347 }
348 } else {
349 /*
350 * read obl bytes in bufo, we read at maximum for
351 * emptying the device
352 */
353 readsl(ss->base + SS_TXFIFO, ss->bufo, tx_cnt);
354 obl = tx_cnt * 4;
355 obo = 0;
356 do {
357 /*
358 * how many bytes we can copy ?
359 * no more than remaining SG size
360 * no more than remaining buffer
361 * no need to test against oleft
362 */
363 todo = min_t(size_t,
364 mo.length - oo, obl - obo);
365 memcpy(mo.addr + oo, ss->bufo + obo, todo);
366 oleft -= todo;
367 obo += todo;
368 oo += todo;
369 if (oo == mo.length) {
370 po += mo.length;
371 sg_miter_next(&mo);
372 oo = 0;
373 }
374 } while (obo < obl);
375 /* bufo must be fully used here */
376 }
377 sg_miter_stop(&mo);
378 }
379 if (areq->iv) {
380 if (mode & SS_DECRYPTION) {
381 memcpy(areq->iv, ctx->backup_iv, ivsize);
382 memzero_explicit(ctx->backup_iv, ivsize);
383 } else {
384 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize,
385 ivsize, 0);
386 }
387 }
388
389 release_ss:
390 writel(0, ss->base + SS_CTL);
391 spin_unlock_irqrestore(&ss->slock, flags);
392
393 return err;
394 }
395
396 /* CBC AES */
sun4i_ss_cbc_aes_encrypt(struct skcipher_request * areq)397 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq)
398 {
399 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
400 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
401 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
402
403 rctx->mode = SS_OP_AES | SS_CBC | SS_ENABLED | SS_ENCRYPTION |
404 op->keymode;
405 return sun4i_ss_cipher_poll(areq);
406 }
407
sun4i_ss_cbc_aes_decrypt(struct skcipher_request * areq)408 int sun4i_ss_cbc_aes_decrypt(struct skcipher_request *areq)
409 {
410 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
411 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
412 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
413
414 rctx->mode = SS_OP_AES | SS_CBC | SS_ENABLED | SS_DECRYPTION |
415 op->keymode;
416 return sun4i_ss_cipher_poll(areq);
417 }
418
419 /* ECB AES */
sun4i_ss_ecb_aes_encrypt(struct skcipher_request * areq)420 int sun4i_ss_ecb_aes_encrypt(struct skcipher_request *areq)
421 {
422 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
423 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
424 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
425
426 rctx->mode = SS_OP_AES | SS_ECB | SS_ENABLED | SS_ENCRYPTION |
427 op->keymode;
428 return sun4i_ss_cipher_poll(areq);
429 }
430
sun4i_ss_ecb_aes_decrypt(struct skcipher_request * areq)431 int sun4i_ss_ecb_aes_decrypt(struct skcipher_request *areq)
432 {
433 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
434 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
435 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
436
437 rctx->mode = SS_OP_AES | SS_ECB | SS_ENABLED | SS_DECRYPTION |
438 op->keymode;
439 return sun4i_ss_cipher_poll(areq);
440 }
441
442 /* CBC DES */
sun4i_ss_cbc_des_encrypt(struct skcipher_request * areq)443 int sun4i_ss_cbc_des_encrypt(struct skcipher_request *areq)
444 {
445 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
446 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
447 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
448
449 rctx->mode = SS_OP_DES | SS_CBC | SS_ENABLED | SS_ENCRYPTION |
450 op->keymode;
451 return sun4i_ss_cipher_poll(areq);
452 }
453
sun4i_ss_cbc_des_decrypt(struct skcipher_request * areq)454 int sun4i_ss_cbc_des_decrypt(struct skcipher_request *areq)
455 {
456 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
457 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
458 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
459
460 rctx->mode = SS_OP_DES | SS_CBC | SS_ENABLED | SS_DECRYPTION |
461 op->keymode;
462 return sun4i_ss_cipher_poll(areq);
463 }
464
465 /* ECB DES */
sun4i_ss_ecb_des_encrypt(struct skcipher_request * areq)466 int sun4i_ss_ecb_des_encrypt(struct skcipher_request *areq)
467 {
468 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
469 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
470 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
471
472 rctx->mode = SS_OP_DES | SS_ECB | SS_ENABLED | SS_ENCRYPTION |
473 op->keymode;
474 return sun4i_ss_cipher_poll(areq);
475 }
476
sun4i_ss_ecb_des_decrypt(struct skcipher_request * areq)477 int sun4i_ss_ecb_des_decrypt(struct skcipher_request *areq)
478 {
479 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
480 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
481 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
482
483 rctx->mode = SS_OP_DES | SS_ECB | SS_ENABLED | SS_DECRYPTION |
484 op->keymode;
485 return sun4i_ss_cipher_poll(areq);
486 }
487
488 /* CBC 3DES */
sun4i_ss_cbc_des3_encrypt(struct skcipher_request * areq)489 int sun4i_ss_cbc_des3_encrypt(struct skcipher_request *areq)
490 {
491 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
492 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
493 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
494
495 rctx->mode = SS_OP_3DES | SS_CBC | SS_ENABLED | SS_ENCRYPTION |
496 op->keymode;
497 return sun4i_ss_cipher_poll(areq);
498 }
499
sun4i_ss_cbc_des3_decrypt(struct skcipher_request * areq)500 int sun4i_ss_cbc_des3_decrypt(struct skcipher_request *areq)
501 {
502 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
503 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
504 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
505
506 rctx->mode = SS_OP_3DES | SS_CBC | SS_ENABLED | SS_DECRYPTION |
507 op->keymode;
508 return sun4i_ss_cipher_poll(areq);
509 }
510
511 /* ECB 3DES */
sun4i_ss_ecb_des3_encrypt(struct skcipher_request * areq)512 int sun4i_ss_ecb_des3_encrypt(struct skcipher_request *areq)
513 {
514 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
515 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
516 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
517
518 rctx->mode = SS_OP_3DES | SS_ECB | SS_ENABLED | SS_ENCRYPTION |
519 op->keymode;
520 return sun4i_ss_cipher_poll(areq);
521 }
522
sun4i_ss_ecb_des3_decrypt(struct skcipher_request * areq)523 int sun4i_ss_ecb_des3_decrypt(struct skcipher_request *areq)
524 {
525 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
526 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
527 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
528
529 rctx->mode = SS_OP_3DES | SS_ECB | SS_ENABLED | SS_DECRYPTION |
530 op->keymode;
531 return sun4i_ss_cipher_poll(areq);
532 }
533
sun4i_ss_cipher_init(struct crypto_tfm * tfm)534 int sun4i_ss_cipher_init(struct crypto_tfm *tfm)
535 {
536 struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm);
537 struct sun4i_ss_alg_template *algt;
538 const char *name = crypto_tfm_alg_name(tfm);
539 int err;
540
541 memset(op, 0, sizeof(struct sun4i_tfm_ctx));
542
543 algt = container_of(tfm->__crt_alg, struct sun4i_ss_alg_template,
544 alg.crypto.base);
545 op->ss = algt->ss;
546
547 op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
548 if (IS_ERR(op->fallback_tfm)) {
549 dev_err(op->ss->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
550 name, PTR_ERR(op->fallback_tfm));
551 return PTR_ERR(op->fallback_tfm);
552 }
553
554 crypto_skcipher_set_reqsize(__crypto_skcipher_cast(tfm),
555 sizeof(struct sun4i_cipher_req_ctx) +
556 crypto_skcipher_reqsize(op->fallback_tfm));
557
558 err = pm_runtime_resume_and_get(op->ss->dev);
559 if (err < 0)
560 goto error_pm;
561
562 return 0;
563 error_pm:
564 crypto_free_skcipher(op->fallback_tfm);
565 return err;
566 }
567
sun4i_ss_cipher_exit(struct crypto_tfm * tfm)568 void sun4i_ss_cipher_exit(struct crypto_tfm *tfm)
569 {
570 struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm);
571
572 crypto_free_skcipher(op->fallback_tfm);
573 pm_runtime_put(op->ss->dev);
574 }
575
576 /* check and set the AES key, prepare the mode to be used */
sun4i_ss_aes_setkey(struct crypto_skcipher * tfm,const u8 * key,unsigned int keylen)577 int sun4i_ss_aes_setkey(struct crypto_skcipher *tfm, const u8 *key,
578 unsigned int keylen)
579 {
580 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
581 struct sun4i_ss_ctx *ss = op->ss;
582
583 switch (keylen) {
584 case 128 / 8:
585 op->keymode = SS_AES_128BITS;
586 break;
587 case 192 / 8:
588 op->keymode = SS_AES_192BITS;
589 break;
590 case 256 / 8:
591 op->keymode = SS_AES_256BITS;
592 break;
593 default:
594 dev_dbg(ss->dev, "ERROR: Invalid keylen %u\n", keylen);
595 return -EINVAL;
596 }
597 op->keylen = keylen;
598 memcpy(op->key, key, keylen);
599
600 crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
601 crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
602
603 return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
604 }
605
606 /* check and set the DES key, prepare the mode to be used */
sun4i_ss_des_setkey(struct crypto_skcipher * tfm,const u8 * key,unsigned int keylen)607 int sun4i_ss_des_setkey(struct crypto_skcipher *tfm, const u8 *key,
608 unsigned int keylen)
609 {
610 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
611 int err;
612
613 err = verify_skcipher_des_key(tfm, key);
614 if (err)
615 return err;
616
617 op->keylen = keylen;
618 memcpy(op->key, key, keylen);
619
620 crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
621 crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
622
623 return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
624 }
625
626 /* check and set the 3DES key, prepare the mode to be used */
sun4i_ss_des3_setkey(struct crypto_skcipher * tfm,const u8 * key,unsigned int keylen)627 int sun4i_ss_des3_setkey(struct crypto_skcipher *tfm, const u8 *key,
628 unsigned int keylen)
629 {
630 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
631 int err;
632
633 err = verify_skcipher_des3_key(tfm, key);
634 if (err)
635 return err;
636
637 op->keylen = keylen;
638 memcpy(op->key, key, keylen);
639
640 crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
641 crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
642
643 return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
644 }
645