/drivers/dma/ |
D | iop-adma.h | 381 union iop3xx_desc hw_desc = { .ptr = desc->hw_desc, }; in iop_desc_get_byte_count() local 386 return hw_desc.dma->byte_count; in iop_desc_get_byte_count() 388 return hw_desc.aau->byte_count; in iop_desc_get_byte_count() 415 union iop3xx_desc hw_desc = { .ptr = desc->hw_desc, }; in iop_desc_get_src_addr() local 420 return hw_desc.dma->src_addr; in iop_desc_get_src_addr() 428 return hw_desc.aau->src[src_idx]; in iop_desc_get_src_addr() 430 return hw_desc.aau->src_edc[__desc_idx(src_idx)].src_addr; in iop_desc_get_src_addr() 433 static inline void iop3xx_aau_desc_set_src_addr(struct iop3xx_desc_aau *hw_desc, in iop3xx_aau_desc_set_src_addr() argument 437 hw_desc->src[src_idx] = addr; in iop3xx_aau_desc_set_src_addr() 439 hw_desc->src_edc[__desc_idx(src_idx)].src_addr = addr; in iop3xx_aau_desc_set_src_addr() [all …]
|
D | pxa_dma.c | 94 struct pxad_desc_hw *hw_desc[]; /* DMA coherent descriptors */ member 520 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in set_updater_desc() 521 dma_addr_t dma = sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr; in set_updater_desc() 531 sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr = sw_desc->first; in set_updater_desc() 538 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in is_desc_completed() 551 desc1->hw_desc[desc1->nb_desc - 1]->ddadr = dma_to_chain; in pxad_desc_chain() 729 dma = sw_desc->hw_desc[i - 1]->ddadr; in pxad_free_desc() 733 sw_desc->hw_desc[i], dma); in pxad_free_desc() 754 sw_desc->hw_desc[i] = dma_pool_alloc(sw_desc->desc_pool, in pxad_alloc_desc() 756 if (!sw_desc->hw_desc[i]) { in pxad_alloc_desc() [all …]
|
D | mv_xor.c | 52 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_init() local 54 hw_desc->status = XOR_DESC_DMA_OWNED; in mv_desc_init() 55 hw_desc->phy_next_desc = 0; in mv_desc_init() 57 hw_desc->desc_command = (flags & DMA_PREP_INTERRUPT) ? in mv_desc_init() 59 hw_desc->phy_dest_addr = addr; in mv_desc_init() 60 hw_desc->byte_count = byte_count; in mv_desc_init() 65 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_mode() local 70 hw_desc->desc_command |= XOR_DESC_OPERATION_XOR; in mv_desc_set_mode() 73 hw_desc->desc_command |= XOR_DESC_OPERATION_MEMCPY; in mv_desc_set_mode() 84 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_next_desc() local [all …]
|
D | mv_xor.h | 145 void *hw_desc; member 188 container_of(addr_hw_desc, struct mv_xor_desc_slot, hw_desc) 190 #define mv_hw_desc_slot_idx(hw_desc, idx) \ argument 191 ((void *)(((unsigned long)hw_desc) + ((idx) << 5)))
|
D | altera-msgdma.c | 160 struct msgdma_extended_desc hw_desc; member 355 desc = &new->hw_desc; in msgdma_prep_memcpy() 420 desc = &new->hw_desc; in msgdma_prep_slave_sg() 496 void __iomem *hw_desc = mdev->desc; in msgdma_copy_one() local 515 memcpy((void __force *)hw_desc, &desc->hw_desc, in msgdma_copy_one() 516 sizeof(desc->hw_desc) - sizeof(u32)); in msgdma_copy_one() 521 iowrite32(desc->hw_desc.control, hw_desc + in msgdma_copy_one()
|
D | mv_xor_v2.c | 180 struct mv_xor_v2_descriptor hw_desc; member 319 memcpy(dest_hw_desc, &sw_desc->hw_desc, xor_dev->desc_size); in mv_xor_v2_tx_submit() 394 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_memcpy() 450 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_xor() 498 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_interrupt()
|
D | iop-adma.c | 310 iter->idx, iter->hw_desc, in iop_adma_alloc_slots() 419 char *hw_desc; in iop_adma_alloc_chan_resources() local 440 hw_desc = (char *) iop_chan->device->dma_desc_pool_virt; in iop_adma_alloc_chan_resources() 441 slot->hw_desc = (void *) &hw_desc[idx * IOP_ADMA_SLOT_SIZE]; in iop_adma_alloc_chan_resources() 448 hw_desc = (char *) iop_chan->device->dma_desc_pool; in iop_adma_alloc_chan_resources() 450 (dma_addr_t) &hw_desc[idx * IOP_ADMA_SLOT_SIZE]; in iop_adma_alloc_chan_resources()
|
/drivers/dma/ppc4xx/ |
D | adma.c | 177 print_cb(chan, iter->hw_desc); in print_cb_list() 238 p = desc->hw_desc; in ppc440spe_desc_init_interrupt() 239 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_interrupt() 245 memset(desc->hw_desc, 0, sizeof(struct dma_cdb)); in ppc440spe_desc_init_interrupt() 262 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_null_xor() 274 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_xor() local 276 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_xor() 281 hw_desc->cbc = XOR_CBCR_TGT_BIT | src_cnt; in ppc440spe_desc_init_xor() 284 hw_desc->cbc |= XOR_CBCR_CBCE_BIT; in ppc440spe_desc_init_xor() 294 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_dma2pq() local [all …]
|
D | adma.h | 155 void *hw_desc; member
|
/drivers/crypto/ |
D | sahara.c | 205 struct sahara_hw_desc *hw_desc[SAHARA_MAX_HW_DESC]; member 412 dev_dbg(dev->device, "\thdr = 0x%08x\n", dev->hw_desc[i]->hdr); in sahara_dump_descriptors() 413 dev_dbg(dev->device, "\tlen1 = %u\n", dev->hw_desc[i]->len1); in sahara_dump_descriptors() 414 dev_dbg(dev->device, "\tp1 = 0x%08x\n", dev->hw_desc[i]->p1); in sahara_dump_descriptors() 415 dev_dbg(dev->device, "\tlen2 = %u\n", dev->hw_desc[i]->len2); in sahara_dump_descriptors() 416 dev_dbg(dev->device, "\tp2 = 0x%08x\n", dev->hw_desc[i]->p2); in sahara_dump_descriptors() 418 dev->hw_desc[i]->next); in sahara_dump_descriptors() 455 dev->hw_desc[idx]->len1 = AES_BLOCK_SIZE; in sahara_hw_descriptor_create() 456 dev->hw_desc[idx]->p1 = dev->iv_phys_base; in sahara_hw_descriptor_create() 458 dev->hw_desc[idx]->len1 = 0; in sahara_hw_descriptor_create() [all …]
|
/drivers/crypto/caam/ |
D | caamalg.c | 892 u32 hw_desc[]; member 917 u32 hw_desc[0]; member 972 edesc = container_of(desc, struct aead_edesc, hw_desc[0]); in aead_encrypt_done() 993 edesc = container_of(desc, struct aead_edesc, hw_desc[0]); in aead_decrypt_done() 1016 edesc = container_of(desc, struct skcipher_edesc, hw_desc[0]); in skcipher_encrypt_done() 1056 edesc = container_of(desc, struct skcipher_edesc, hw_desc[0]); in skcipher_decrypt_done() 1095 u32 *desc = edesc->hw_desc; in init_aead_job() 1156 u32 *desc = edesc->hw_desc; in init_gcm_job() 1186 u32 *desc = edesc->hw_desc; in init_chachapoly_job() 1226 u32 *desc = edesc->hw_desc; in init_authenc_job() [all …]
|
D | caampkc.c | 126 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_pub_done() 145 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_priv_f1_done() 164 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_priv_f2_done() 183 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_priv_f3_done() 653 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub); in caam_rsa_enc() 655 ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_pub_done, req); in caam_rsa_enc() 686 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1); in caam_rsa_dec_priv_f1() 688 ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f1_done, req); in caam_rsa_dec_priv_f1() 719 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2); in caam_rsa_dec_priv_f2() 721 ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f2_done, req); in caam_rsa_dec_priv_f2() [all …]
|
D | caamrng.c | 64 u32 hw_desc[DESC_JOB_O_LEN]; member 111 bd = container_of(desc, struct buf_data, hw_desc[0]); in rng_done() 130 u32 *desc = bd->hw_desc; in submit_job() 227 u32 *desc = bd->hw_desc; in rng_create_job_desc()
|
D | caamhash.c | 563 u32 hw_desc[DESC_JOB_IO_LEN_MAX / sizeof(u32)] ____cacheline_aligned; member 613 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done() 641 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done_bi() 673 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done_ctx_src() 701 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done_ctx_dst() 738 init_job_desc_shared(edesc->hw_desc, sh_desc_dma, desc_len(sh_desc), in ahash_edesc_alloc() 774 append_seq_in_ptr(edesc->hw_desc, src_dma, first_bytes + to_hash, in ahash_edesc_add_src() 875 desc = edesc->hw_desc; in ahash_update_ctx() 944 desc = edesc->hw_desc; in ahash_final_ctx() 1030 desc = edesc->hw_desc; in ahash_finup_ctx() [all …]
|
D | caampkc.h | 135 u32 hw_desc[]; member
|