/drivers/crypto/caam/ |
D | caamalg_qi.c | 802 dma_addr_t iv_dma; member 824 dma_addr_t iv_dma; member 871 int dst_nents, dma_addr_t iv_dma, int ivsize, in caam_unmap() argument 884 if (iv_dma) in caam_unmap() 885 dma_unmap_single(dev, iv_dma, ivsize, iv_dir); in caam_unmap() 898 edesc->iv_dma, ivsize, DMA_TO_DEVICE, edesc->qm_sg_dma, in aead_unmap() 910 edesc->iv_dma, ivsize, DMA_BIDIRECTIONAL, edesc->qm_sg_dma, in skcipher_unmap() 951 dma_addr_t qm_sg_dma, iv_dma = 0; in aead_edesc_alloc() local 1079 iv_dma = dma_map_single(qidev, iv, ivsize, DMA_TO_DEVICE); in aead_edesc_alloc() 1080 if (dma_mapping_error(qidev, iv_dma)) { in aead_edesc_alloc() [all …]
|
D | caamalg_qi2.h | 112 dma_addr_t iv_dma; member 132 dma_addr_t iv_dma; member
|
D | caamalg.c | 913 dma_addr_t iv_dma; member 923 dma_addr_t iv_dma, int ivsize, dma_addr_t sec4_sg_dma, in caam_unmap() argument 935 if (iv_dma) in caam_unmap() 936 dma_unmap_single(dev, iv_dma, ivsize, DMA_BIDIRECTIONAL); in caam_unmap() 959 edesc->iv_dma, ivsize, in skcipher_unmap() 1689 dma_addr_t iv_dma = 0; in skcipher_edesc_alloc() local 1788 iv_dma = dma_map_single(jrdev, iv, ivsize, DMA_BIDIRECTIONAL); in skcipher_edesc_alloc() 1789 if (dma_mapping_error(jrdev, iv_dma)) { in skcipher_edesc_alloc() 1797 dma_to_sec4_sg_one(edesc->sec4_sg, iv_dma, ivsize, 0); in skcipher_edesc_alloc() 1809 mapped_dst_nents, iv_dma, ivsize, 0); in skcipher_edesc_alloc() [all …]
|
D | caamalg_qi2.c | 143 int dst_nents, dma_addr_t iv_dma, int ivsize, in caam_unmap() argument 156 if (iv_dma) in caam_unmap() 157 dma_unmap_single(dev, iv_dma, ivsize, iv_dir); in caam_unmap() 363 dma_addr_t qm_sg_dma, iv_dma = 0; in aead_edesc_alloc() local 486 iv_dma = dma_map_single(dev, iv, ivsize, DMA_TO_DEVICE); in aead_edesc_alloc() 487 if (dma_mapping_error(dev, iv_dma)) { in aead_edesc_alloc() 498 edesc->iv_dma = iv_dma; in aead_edesc_alloc() 514 iv_dma, ivsize, DMA_TO_DEVICE, 0, 0); in aead_edesc_alloc() 522 dma_to_qm_sg_one(sg_table + qm_sg_index, iv_dma, ivsize, 0); in aead_edesc_alloc() 536 iv_dma, ivsize, DMA_TO_DEVICE, 0, 0); in aead_edesc_alloc() [all …]
|
/drivers/crypto/chelsio/ |
D | chcr_crypto.h | 184 dma_addr_t iv_dma; member
|
D | chcr_algo.c | 2474 reqctx->iv_dma = dma_map_single(dev, reqctx->iv, (IV + reqctx->b0_len), in chcr_aead_dma_map() 2476 if (dma_mapping_error(dev, reqctx->iv_dma)) in chcr_aead_dma_map() 2479 reqctx->b0_dma = reqctx->iv_dma + IV; in chcr_aead_dma_map() 2503 dma_unmap_single(dev, reqctx->iv_dma, IV, DMA_BIDIRECTIONAL); in chcr_aead_dma_map() 2521 dma_unmap_single(dev, reqctx->iv_dma, (IV + reqctx->b0_len), in chcr_aead_dma_unmap() 2572 dsgl_walk_add_page(&dsgl_walk, IV + reqctx->b0_len, reqctx->iv_dma); in chcr_add_aead_dst_ent()
|
/drivers/crypto/ |
D | talitos.h | 63 dma_addr_t iv_dma; member
|
D | talitos.c | 1039 dma_unmap_single(dev, edesc->iv_dma, ivsize, DMA_TO_DEVICE); in ipsec_esp_encrypt_done() 1241 to_talitos_ptr(civ_ptr, edesc->iv_dma, ivsize, is_sec1); in ipsec_esp() 1335 dma_addr_t iv_dma = 0; in talitos_edesc_alloc() local 1403 iv_dma = dma_map_single(dev, iv, ivsize, DMA_TO_DEVICE); in talitos_edesc_alloc() 1409 edesc->iv_dma = iv_dma; in talitos_edesc_alloc() 1590 to_talitos_ptr(&desc->ptr[1], edesc->iv_dma, ivsize, is_sec1); in common_nonsnoop()
|