Home
last modified time | relevance | path

Searched refs:hw_desc (Results 1 – 15 of 15) sorted by relevance

/drivers/dma/
Diop-adma.h381 union iop3xx_desc hw_desc = { .ptr = desc->hw_desc, }; in iop_desc_get_byte_count() local
386 return hw_desc.dma->byte_count; in iop_desc_get_byte_count()
388 return hw_desc.aau->byte_count; in iop_desc_get_byte_count()
415 union iop3xx_desc hw_desc = { .ptr = desc->hw_desc, }; in iop_desc_get_src_addr() local
420 return hw_desc.dma->src_addr; in iop_desc_get_src_addr()
428 return hw_desc.aau->src[src_idx]; in iop_desc_get_src_addr()
430 return hw_desc.aau->src_edc[__desc_idx(src_idx)].src_addr; in iop_desc_get_src_addr()
433 static inline void iop3xx_aau_desc_set_src_addr(struct iop3xx_desc_aau *hw_desc, in iop3xx_aau_desc_set_src_addr() argument
437 hw_desc->src[src_idx] = addr; in iop3xx_aau_desc_set_src_addr()
439 hw_desc->src_edc[__desc_idx(src_idx)].src_addr = addr; in iop3xx_aau_desc_set_src_addr()
[all …]
Dpxa_dma.c94 struct pxad_desc_hw *hw_desc[]; /* DMA coherent descriptors */ member
520 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in set_updater_desc()
521 dma_addr_t dma = sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr; in set_updater_desc()
531 sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr = sw_desc->first; in set_updater_desc()
538 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in is_desc_completed()
551 desc1->hw_desc[desc1->nb_desc - 1]->ddadr = dma_to_chain; in pxad_desc_chain()
729 dma = sw_desc->hw_desc[i - 1]->ddadr; in pxad_free_desc()
733 sw_desc->hw_desc[i], dma); in pxad_free_desc()
754 sw_desc->hw_desc[i] = dma_pool_alloc(sw_desc->desc_pool, in pxad_alloc_desc()
756 if (!sw_desc->hw_desc[i]) { in pxad_alloc_desc()
[all …]
Dmv_xor.c52 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_init() local
54 hw_desc->status = XOR_DESC_DMA_OWNED; in mv_desc_init()
55 hw_desc->phy_next_desc = 0; in mv_desc_init()
57 hw_desc->desc_command = (flags & DMA_PREP_INTERRUPT) ? in mv_desc_init()
59 hw_desc->phy_dest_addr = addr; in mv_desc_init()
60 hw_desc->byte_count = byte_count; in mv_desc_init()
65 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_mode() local
70 hw_desc->desc_command |= XOR_DESC_OPERATION_XOR; in mv_desc_set_mode()
73 hw_desc->desc_command |= XOR_DESC_OPERATION_MEMCPY; in mv_desc_set_mode()
84 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_next_desc() local
[all …]
Dmv_xor.h145 void *hw_desc; member
188 container_of(addr_hw_desc, struct mv_xor_desc_slot, hw_desc)
190 #define mv_hw_desc_slot_idx(hw_desc, idx) \ argument
191 ((void *)(((unsigned long)hw_desc) + ((idx) << 5)))
Daltera-msgdma.c160 struct msgdma_extended_desc hw_desc; member
355 desc = &new->hw_desc; in msgdma_prep_memcpy()
420 desc = &new->hw_desc; in msgdma_prep_slave_sg()
496 void __iomem *hw_desc = mdev->desc; in msgdma_copy_one() local
515 memcpy((void __force *)hw_desc, &desc->hw_desc, in msgdma_copy_one()
516 sizeof(desc->hw_desc) - sizeof(u32)); in msgdma_copy_one()
521 iowrite32(desc->hw_desc.control, hw_desc + in msgdma_copy_one()
Dmv_xor_v2.c180 struct mv_xor_v2_descriptor hw_desc; member
319 memcpy(dest_hw_desc, &sw_desc->hw_desc, xor_dev->desc_size); in mv_xor_v2_tx_submit()
394 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_memcpy()
450 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_xor()
498 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_interrupt()
Diop-adma.c310 iter->idx, iter->hw_desc, in iop_adma_alloc_slots()
419 char *hw_desc; in iop_adma_alloc_chan_resources() local
440 hw_desc = (char *) iop_chan->device->dma_desc_pool_virt; in iop_adma_alloc_chan_resources()
441 slot->hw_desc = (void *) &hw_desc[idx * IOP_ADMA_SLOT_SIZE]; in iop_adma_alloc_chan_resources()
448 hw_desc = (char *) iop_chan->device->dma_desc_pool; in iop_adma_alloc_chan_resources()
450 (dma_addr_t) &hw_desc[idx * IOP_ADMA_SLOT_SIZE]; in iop_adma_alloc_chan_resources()
/drivers/dma/ppc4xx/
Dadma.c177 print_cb(chan, iter->hw_desc); in print_cb_list()
238 p = desc->hw_desc; in ppc440spe_desc_init_interrupt()
239 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_interrupt()
245 memset(desc->hw_desc, 0, sizeof(struct dma_cdb)); in ppc440spe_desc_init_interrupt()
262 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_null_xor()
274 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_xor() local
276 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_xor()
281 hw_desc->cbc = XOR_CBCR_TGT_BIT | src_cnt; in ppc440spe_desc_init_xor()
284 hw_desc->cbc |= XOR_CBCR_CBCE_BIT; in ppc440spe_desc_init_xor()
294 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_dma2pq() local
[all …]
Dadma.h155 void *hw_desc; member
/drivers/crypto/
Dsahara.c205 struct sahara_hw_desc *hw_desc[SAHARA_MAX_HW_DESC]; member
412 dev_dbg(dev->device, "\thdr = 0x%08x\n", dev->hw_desc[i]->hdr); in sahara_dump_descriptors()
413 dev_dbg(dev->device, "\tlen1 = %u\n", dev->hw_desc[i]->len1); in sahara_dump_descriptors()
414 dev_dbg(dev->device, "\tp1 = 0x%08x\n", dev->hw_desc[i]->p1); in sahara_dump_descriptors()
415 dev_dbg(dev->device, "\tlen2 = %u\n", dev->hw_desc[i]->len2); in sahara_dump_descriptors()
416 dev_dbg(dev->device, "\tp2 = 0x%08x\n", dev->hw_desc[i]->p2); in sahara_dump_descriptors()
418 dev->hw_desc[i]->next); in sahara_dump_descriptors()
455 dev->hw_desc[idx]->len1 = AES_BLOCK_SIZE; in sahara_hw_descriptor_create()
456 dev->hw_desc[idx]->p1 = dev->iv_phys_base; in sahara_hw_descriptor_create()
458 dev->hw_desc[idx]->len1 = 0; in sahara_hw_descriptor_create()
[all …]
/drivers/crypto/caam/
Dcaamalg.c892 u32 hw_desc[]; member
917 u32 hw_desc[0]; member
972 edesc = container_of(desc, struct aead_edesc, hw_desc[0]); in aead_encrypt_done()
993 edesc = container_of(desc, struct aead_edesc, hw_desc[0]); in aead_decrypt_done()
1016 edesc = container_of(desc, struct skcipher_edesc, hw_desc[0]); in skcipher_encrypt_done()
1056 edesc = container_of(desc, struct skcipher_edesc, hw_desc[0]); in skcipher_decrypt_done()
1095 u32 *desc = edesc->hw_desc; in init_aead_job()
1156 u32 *desc = edesc->hw_desc; in init_gcm_job()
1186 u32 *desc = edesc->hw_desc; in init_chachapoly_job()
1226 u32 *desc = edesc->hw_desc; in init_authenc_job()
[all …]
Dcaampkc.c126 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_pub_done()
145 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_priv_f1_done()
164 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_priv_f2_done()
183 edesc = container_of(desc, struct rsa_edesc, hw_desc[0]); in rsa_priv_f3_done()
653 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub); in caam_rsa_enc()
655 ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_pub_done, req); in caam_rsa_enc()
686 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1); in caam_rsa_dec_priv_f1()
688 ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f1_done, req); in caam_rsa_dec_priv_f1()
719 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2); in caam_rsa_dec_priv_f2()
721 ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f2_done, req); in caam_rsa_dec_priv_f2()
[all …]
Dcaamrng.c64 u32 hw_desc[DESC_JOB_O_LEN]; member
111 bd = container_of(desc, struct buf_data, hw_desc[0]); in rng_done()
130 u32 *desc = bd->hw_desc; in submit_job()
227 u32 *desc = bd->hw_desc; in rng_create_job_desc()
Dcaamhash.c563 u32 hw_desc[DESC_JOB_IO_LEN_MAX / sizeof(u32)] ____cacheline_aligned; member
613 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done()
641 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done_bi()
673 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done_ctx_src()
701 edesc = container_of(desc, struct ahash_edesc, hw_desc[0]); in ahash_done_ctx_dst()
738 init_job_desc_shared(edesc->hw_desc, sh_desc_dma, desc_len(sh_desc), in ahash_edesc_alloc()
774 append_seq_in_ptr(edesc->hw_desc, src_dma, first_bytes + to_hash, in ahash_edesc_add_src()
875 desc = edesc->hw_desc; in ahash_update_ctx()
944 desc = edesc->hw_desc; in ahash_final_ctx()
1030 desc = edesc->hw_desc; in ahash_finup_ctx()
[all …]
Dcaampkc.h135 u32 hw_desc[]; member