• Home
  • Raw
  • Download

Lines Matching refs:src_nents

146 		       struct scatterlist *dst, int src_nents,  in caam_unmap()  argument
152 if (src_nents) in caam_unmap()
153 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap()
157 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap()
359 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
380 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
381 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
385 return ERR_PTR(src_nents); in aead_edesc_alloc()
396 if (src_nents) { in aead_edesc_alloc()
397 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
413 dma_unmap_sg(dev, req->src, src_nents, in aead_edesc_alloc()
425 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
426 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
430 return ERR_PTR(src_nents); in aead_edesc_alloc()
433 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
473 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in aead_edesc_alloc()
488 caam_unmap(dev, req->src, req->dst, src_nents, in aead_edesc_alloc()
495 edesc->src_nents = src_nents; in aead_edesc_alloc()
512 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
534 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
1116 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in skcipher_edesc_alloc() local
1124 src_nents = sg_nents_for_len(req->src, req->cryptlen); in skcipher_edesc_alloc()
1125 if (unlikely(src_nents < 0)) { in skcipher_edesc_alloc()
1128 return ERR_PTR(src_nents); in skcipher_edesc_alloc()
1139 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1150 dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE); in skcipher_edesc_alloc()
1154 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1183 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1192 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1205 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1211 edesc->src_nents = src_nents; in skcipher_edesc_alloc()
1229 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in skcipher_edesc_alloc()
1261 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap()
1273 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap()
1404 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_encrypt_done()
1442 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_decrypt_done()
3358 if (edesc->src_nents) in ahash_unmap()
3359 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE); in ahash_unmap()
3529 int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index; in ahash_update_ctx() local
3540 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_ctx()
3541 if (src_nents < 0) { in ahash_update_ctx()
3543 return src_nents; in ahash_update_ctx()
3546 if (src_nents) { in ahash_update_ctx()
3547 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3560 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3565 edesc->src_nents = src_nents; in ahash_update_ctx()
3717 int src_nents, mapped_nents; in ahash_finup_ctx() local
3723 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_ctx()
3724 if (src_nents < 0) { in ahash_finup_ctx()
3726 return src_nents; in ahash_finup_ctx()
3729 if (src_nents) { in ahash_finup_ctx()
3730 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_ctx()
3743 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_ctx()
3747 edesc->src_nents = src_nents; in ahash_finup_ctx()
3810 int src_nents, mapped_nents; in ahash_digest() local
3816 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_digest()
3817 if (src_nents < 0) { in ahash_digest()
3819 return src_nents; in ahash_digest()
3822 if (src_nents) { in ahash_digest()
3823 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_digest()
3836 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_digest()
3840 edesc->src_nents = src_nents; in ahash_digest()
3981 int qm_sg_bytes, src_nents, mapped_nents; in ahash_update_no_ctx() local
3992 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_no_ctx()
3993 if (src_nents < 0) { in ahash_update_no_ctx()
3995 return src_nents; in ahash_update_no_ctx()
3998 if (src_nents) { in ahash_update_no_ctx()
3999 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4012 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4017 edesc->src_nents = src_nents; in ahash_update_no_ctx()
4099 int qm_sg_bytes, src_nents, mapped_nents; in ahash_finup_no_ctx() local
4105 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_no_ctx()
4106 if (src_nents < 0) { in ahash_finup_no_ctx()
4108 return src_nents; in ahash_finup_no_ctx()
4111 if (src_nents) { in ahash_finup_no_ctx()
4112 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_no_ctx()
4125 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_no_ctx()
4129 edesc->src_nents = src_nents; in ahash_finup_no_ctx()
4198 int src_nents, mapped_nents; in ahash_update_first() local
4210 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_first()
4211 if (src_nents < 0) { in ahash_update_first()
4213 return src_nents; in ahash_update_first()
4216 if (src_nents) { in ahash_update_first()
4217 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4230 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4235 edesc->src_nents = src_nents; in ahash_update_first()