Home
last modified time | relevance | path

Searched refs:dma_desc (Results 1 – 25 of 38) sorted by relevance

12

/drivers/dma/
Dtegra20-apb-dma.c160 struct tegra_dma_desc *dma_desc; member
277 struct tegra_dma_desc *dma_desc; in tegra_dma_desc_get() local
283 list_for_each_entry(dma_desc, &tdc->free_dma_desc, node) { in tegra_dma_desc_get()
284 if (async_tx_test_ack(&dma_desc->txd)) { in tegra_dma_desc_get()
285 list_del(&dma_desc->node); in tegra_dma_desc_get()
287 dma_desc->txd.flags = 0; in tegra_dma_desc_get()
288 return dma_desc; in tegra_dma_desc_get()
295 dma_desc = kzalloc(sizeof(*dma_desc), GFP_ATOMIC); in tegra_dma_desc_get()
296 if (!dma_desc) { in tegra_dma_desc_get()
301 dma_async_tx_descriptor_init(&dma_desc->txd, &tdc->dma_chan); in tegra_dma_desc_get()
[all …]
Dtimb_dma.c157 static int td_fill_desc(struct timb_dma_chan *td_chan, u8 *dma_desc, in td_fill_desc() argument
173 dma_desc, (unsigned long long)sg_dma_address(sg)); in td_fill_desc()
175 dma_desc[7] = (sg_dma_address(sg) >> 24) & 0xff; in td_fill_desc()
176 dma_desc[6] = (sg_dma_address(sg) >> 16) & 0xff; in td_fill_desc()
177 dma_desc[5] = (sg_dma_address(sg) >> 8) & 0xff; in td_fill_desc()
178 dma_desc[4] = (sg_dma_address(sg) >> 0) & 0xff; in td_fill_desc()
180 dma_desc[3] = (sg_dma_len(sg) >> 8) & 0xff; in td_fill_desc()
181 dma_desc[2] = (sg_dma_len(sg) >> 0) & 0xff; in td_fill_desc()
183 dma_desc[1] = 0x00; in td_fill_desc()
184 dma_desc[0] = 0x21 | (last ? 0x02 : 0); /* tran, valid */ in td_fill_desc()
/drivers/net/ethernet/stmicro/stmmac/
Ddescs_com.h36 static inline void ehn_desc_rx_set_on_ring(struct dma_desc *p, int end) in ehn_desc_rx_set_on_ring()
43 static inline void ehn_desc_tx_set_on_ring(struct dma_desc *p, int end) in ehn_desc_tx_set_on_ring()
49 static inline void enh_desc_end_tx_desc_on_ring(struct dma_desc *p, int ter) in enh_desc_end_tx_desc_on_ring()
54 static inline void enh_set_tx_desc_len_on_ring(struct dma_desc *p, int len) in enh_set_tx_desc_len_on_ring()
64 static inline void ndesc_rx_set_on_ring(struct dma_desc *p, int end) in ndesc_rx_set_on_ring()
71 static inline void ndesc_tx_set_on_ring(struct dma_desc *p, int end) in ndesc_tx_set_on_ring()
77 static inline void ndesc_end_tx_desc_on_ring(struct dma_desc *p, int ter) in ndesc_end_tx_desc_on_ring()
82 static inline void norm_set_tx_desc_len_on_ring(struct dma_desc *p, int len) in norm_set_tx_desc_len_on_ring()
94 static inline void ehn_desc_rx_set_on_chain(struct dma_desc *p, int end) in ehn_desc_rx_set_on_chain()
99 static inline void ehn_desc_tx_set_on_chain(struct dma_desc *p, int end) in ehn_desc_tx_set_on_chain()
[all …]
Dnorm_desc.c30 struct dma_desc *p, void __iomem *ioaddr) in ndesc_get_tx_status()
64 static int ndesc_get_tx_len(struct dma_desc *p) in ndesc_get_tx_len()
74 struct dma_desc *p) in ndesc_get_rx_status()
123 static void ndesc_init_rx_desc(struct dma_desc *p, int disable_rx_ic, int mode, in ndesc_init_rx_desc()
139 static void ndesc_init_tx_desc(struct dma_desc *p, int mode, int end) in ndesc_init_tx_desc()
148 static int ndesc_get_tx_owner(struct dma_desc *p) in ndesc_get_tx_owner()
153 static int ndesc_get_rx_owner(struct dma_desc *p) in ndesc_get_rx_owner()
158 static void ndesc_set_tx_owner(struct dma_desc *p) in ndesc_set_tx_owner()
163 static void ndesc_set_rx_owner(struct dma_desc *p) in ndesc_set_rx_owner()
168 static int ndesc_get_tx_ls(struct dma_desc *p) in ndesc_get_tx_ls()
[all …]
Dcommon.h299 void (*init_rx_desc) (struct dma_desc *p, int disable_rx_ic, int mode,
302 void (*init_tx_desc) (struct dma_desc *p, int mode, int end);
305 void (*prepare_tx_desc) (struct dma_desc *p, int is_fs, int len,
308 void (*set_tx_owner) (struct dma_desc *p);
309 int (*get_tx_owner) (struct dma_desc *p);
311 void (*close_tx_desc) (struct dma_desc *p);
313 void (*release_tx_desc) (struct dma_desc *p, int mode);
316 void (*clear_tx_ic) (struct dma_desc *p);
318 int (*get_tx_ls) (struct dma_desc *p);
321 struct dma_desc *p, void __iomem *ioaddr);
[all …]
Denh_desc.c30 struct dma_desc *p, void __iomem *ioaddr) in enh_desc_get_tx_status()
88 static int enh_desc_get_tx_len(struct dma_desc *p) in enh_desc_get_tx_len()
178 struct dma_desc *p) in enh_desc_get_rx_status()
240 static void enh_desc_init_rx_desc(struct dma_desc *p, int disable_rx_ic, in enh_desc_init_rx_desc()
256 static void enh_desc_init_tx_desc(struct dma_desc *p, int mode, int end) in enh_desc_init_tx_desc()
265 static int enh_desc_get_tx_owner(struct dma_desc *p) in enh_desc_get_tx_owner()
270 static int enh_desc_get_rx_owner(struct dma_desc *p) in enh_desc_get_rx_owner()
275 static void enh_desc_set_tx_owner(struct dma_desc *p) in enh_desc_set_tx_owner()
280 static void enh_desc_set_rx_owner(struct dma_desc *p) in enh_desc_set_rx_owner()
285 static int enh_desc_get_tx_ls(struct dma_desc *p) in enh_desc_get_tx_ls()
[all …]
Dchain_mode.c36 struct dma_desc *desc = priv->dma_tx + entry; in stmmac_jumbo_frm()
120 struct dma_desc *p = (struct dma_desc *)des; in stmmac_init_dma_chain()
122 dma_phy += sizeof(struct dma_desc); in stmmac_init_dma_chain()
130 static void stmmac_refill_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_refill_desc3()
142 sizeof(struct dma_desc)); in stmmac_refill_desc3()
145 static void stmmac_clean_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_clean_desc3()
157 sizeof(struct dma_desc))); in stmmac_clean_desc3()
Dring_mode.c36 struct dma_desc *desc; in stmmac_jumbo_frm()
41 desc = (struct dma_desc *)(priv->dma_etx + entry); in stmmac_jumbo_frm()
68 desc = (struct dma_desc *)(priv->dma_etx + entry); in stmmac_jumbo_frm()
106 static void stmmac_refill_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_refill_desc3()
116 static void stmmac_init_desc3(struct dma_desc *p) in stmmac_init_desc3()
121 static void stmmac_clean_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_clean_desc3()
Dstmmac_main.c375 if (!priv->hw->desc->get_tx_timestamp_status((struct dma_desc *)desc)) in stmmac_get_tx_hwtstamp()
882 struct dma_desc *p = (struct dma_desc *)head; in stmmac_display_ring()
972 static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p, in stmmac_init_rx_buffers()
1049 struct dma_desc *p; in init_dma_desc_rings()
1085 struct dma_desc *p; in init_dma_desc_rings()
1124 struct dma_desc *p; in dma_free_tx_skbufs()
1202 sizeof(struct dma_desc), in alloc_dma_desc_resources()
1209 sizeof(struct dma_desc), in alloc_dma_desc_resources()
1214 sizeof(struct dma_desc), in alloc_dma_desc_resources()
1242 priv->dma_tx_size * sizeof(struct dma_desc), in free_dma_desc_resources()
[all …]
Dstmmac.h45 struct dma_desc *dma_tx;
61 struct dma_desc *dma_rx ____cacheline_aligned_in_smp;
Ddescs.h28 struct dma_desc { struct
170 struct dma_desc basic; argument
/drivers/ata/
Dpata_pxa.c42 struct pxa_dma_desc *dma_desc; member
73 pd->dma_desc[pd->dma_desc_id].ddadr = pd->dma_desc_addr + in pxa_load_dmac()
76 pd->dma_desc[pd->dma_desc_id].dcmd = DCMD_BURST32 | in pxa_load_dmac()
80 pd->dma_desc[pd->dma_desc_id].dsadr = cpu_addr; in pxa_load_dmac()
81 pd->dma_desc[pd->dma_desc_id].dtadr = pd->dma_io_addr; in pxa_load_dmac()
82 pd->dma_desc[pd->dma_desc_id].dcmd |= DCMD_INCSRCADDR | in pxa_load_dmac()
85 pd->dma_desc[pd->dma_desc_id].dsadr = pd->dma_io_addr; in pxa_load_dmac()
86 pd->dma_desc[pd->dma_desc_id].dtadr = cpu_addr; in pxa_load_dmac()
87 pd->dma_desc[pd->dma_desc_id].dcmd |= DCMD_INCTRGADDR | in pxa_load_dmac()
122 pd->dma_desc[pd->dma_desc_id - 1].ddadr = DDADR_STOP; in pxa_qc_prep()
[all …]
/drivers/usb/dwc2/
Dhcd_ddma.c512 struct dwc2_hcd_dma_desc *dma_desc = &qh->desc_list[idx]; in dwc2_fill_host_isoc_dma_desc() local
515 memset(dma_desc, 0, sizeof(*dma_desc)); in dwc2_fill_host_isoc_dma_desc()
523 dma_desc->buf = (u32)(qtd->urb->dma + frame_desc->offset); in dwc2_fill_host_isoc_dma_desc()
524 dma_desc->status = qh->n_bytes[idx] << HOST_DMA_ISOC_NBYTES_SHIFT & in dwc2_fill_host_isoc_dma_desc()
530 dma_desc->status |= HOST_DMA_IOC; in dwc2_fill_host_isoc_dma_desc()
621 struct dwc2_hcd_dma_desc *dma_desc = &qh->desc_list[n_desc]; in dwc2_fill_host_dma_desc() local
641 dma_desc->status = len << HOST_DMA_NBYTES_SHIFT & HOST_DMA_NBYTES_MASK; in dwc2_fill_host_dma_desc()
646 dma_desc->status |= HOST_DMA_SUP; in dwc2_fill_host_dma_desc()
648 dma_desc->buf = (u32)chan->xfer_dma; in dwc2_fill_host_dma_desc()
803 struct dwc2_hcd_dma_desc *dma_desc = &qh->desc_list[idx]; in dwc2_cmpl_host_isoc_dma_desc() local
[all …]
/drivers/soc/tegra/fuse/
Dfuse-tegra20.c61 struct dma_async_tx_descriptor *dma_desc; in tegra20_fuse_readl() local
70 dma_desc = dmaengine_prep_slave_single(apb_dma_chan, apb_buffer_phys, in tegra20_fuse_readl()
73 if (!dma_desc) in tegra20_fuse_readl()
76 dma_desc->callback = apb_dma_complete; in tegra20_fuse_readl()
77 dma_desc->callback_param = NULL; in tegra20_fuse_readl()
83 dmaengine_submit(dma_desc); in tegra20_fuse_readl()
/drivers/usb/musb/
Dmusb_cppi41.c152 struct dma_async_tx_descriptor *dma_desc; in cppi41_trans_done() local
165 dma_desc = dmaengine_prep_slave_single(dc, in cppi41_trans_done()
170 if (WARN_ON(!dma_desc)) in cppi41_trans_done()
173 dma_desc->callback = cppi41_dma_callback; in cppi41_trans_done()
174 dma_desc->callback_param = &cppi41_channel->channel; in cppi41_trans_done()
175 cppi41_channel->cookie = dma_desc->tx_submit(dma_desc); in cppi41_trans_done()
364 struct dma_async_tx_descriptor *dma_desc; in cppi41_configure_channel() local
416 dma_desc = dmaengine_prep_slave_single(dc, dma_addr, len, direction, in cppi41_configure_channel()
418 if (!dma_desc) in cppi41_configure_channel()
421 dma_desc->callback = cppi41_dma_callback; in cppi41_configure_channel()
[all …]
Dux500_dma.c89 struct dma_async_tx_descriptor *dma_desc; in ux500_configure_channel() local
127 dma_desc = dmaengine_prep_slave_sg(dma_chan, &sg, 1, direction, in ux500_configure_channel()
129 if (!dma_desc) in ux500_configure_channel()
132 dma_desc->callback = ux500_dma_callback; in ux500_configure_channel()
133 dma_desc->callback_param = channel; in ux500_configure_channel()
134 ux500_channel->cookie = dma_desc->tx_submit(dma_desc); in ux500_configure_channel()
/drivers/staging/comedi/drivers/
Dgsc_hpdi.c156 struct plx_dma_desc *dma_desc; member
188 start = le32_to_cpu(devpriv->dma_desc[idx].pci_start_addr); in gsc_hpdi_drain_dma()
203 start = le32_to_cpu(devpriv->dma_desc[idx].pci_start_addr); in gsc_hpdi_drain_dma()
458 devpriv->dma_desc[i].pci_start_addr = in gsc_hpdi_setup_dma_descriptors()
460 devpriv->dma_desc[i].local_start_addr = cpu_to_le32(FIFO_REG); in gsc_hpdi_setup_dma_descriptors()
461 devpriv->dma_desc[i].transfer_size = cpu_to_le32(len); in gsc_hpdi_setup_dma_descriptors()
462 devpriv->dma_desc[i].next = cpu_to_le32((phys_addr + in gsc_hpdi_setup_dma_descriptors()
463 (i + 1) * sizeof(devpriv->dma_desc[0])) | next_bits); in gsc_hpdi_setup_dma_descriptors()
476 devpriv->dma_desc[i - 1].next = cpu_to_le32(phys_addr | next_bits); in gsc_hpdi_setup_dma_descriptors()
526 if (devpriv->dma_desc) in gsc_hpdi_free_dma()
[all …]
/drivers/video/fbdev/
Dpxafb.c1065 struct pxafb_dma_descriptor *dma_desc, *pal_desc; in setup_frame_dma() local
1071 dma_desc = &fbi->dma_buff->dma_desc[dma]; in setup_frame_dma()
1072 dma_desc_off = offsetof(struct pxafb_dma_buff, dma_desc[dma]); in setup_frame_dma()
1074 dma_desc->fsadr = start; in setup_frame_dma()
1075 dma_desc->fidr = 0; in setup_frame_dma()
1076 dma_desc->ldcmd = size; in setup_frame_dma()
1079 dma_desc->fdadr = fbi->dma_buff_phys + dma_desc_off; in setup_frame_dma()
1097 dma_desc->fdadr = fbi->dma_buff_phys + pal_desc_off; in setup_frame_dma()
1129 struct pxafb_dma_descriptor *dma_desc; in setup_smart_dma() local
1132 dma_desc = &fbi->dma_buff->dma_desc[DMA_CMD]; in setup_smart_dma()
[all …]
/drivers/net/ethernet/
Dkorina.c90 #define RD_RING_SIZE (KORINA_NUM_RDS * sizeof(struct dma_desc))
91 #define TD_RING_SIZE (KORINA_NUM_TDS * sizeof(struct dma_desc))
105 struct dma_desc *td_ring; /* transmit descriptor ring */
106 struct dma_desc *rd_ring; /* receive descriptor ring */
184 struct dma_desc *rd) in korina_start_rx()
190 struct dma_desc *rd) in korina_chain_rx()
202 struct dma_desc *td; in korina_send_packet()
351 struct dma_desc *rd = &lp->rd_ring[lp->rx_next_done]; in korina_rx()
522 struct dma_desc *td = &lp->td_ring[lp->tx_next_done]; in korina_tx()
1154 lp->td_ring = (struct dma_desc *)KSEG1ADDR(lp->td_ring); in korina_probe()
/drivers/net/ethernet/broadcom/
Dbgmac.c123 struct bgmac_dma_desc *dma_desc; in bgmac_dma_tx_add() local
158 dma_desc = ring->cpu_base; in bgmac_dma_tx_add()
159 dma_desc += ring->end; in bgmac_dma_tx_add()
160 dma_desc->addr_low = cpu_to_le32(lower_32_bits(slot->dma_addr)); in bgmac_dma_tx_add()
161 dma_desc->addr_high = cpu_to_le32(upper_32_bits(slot->dma_addr)); in bgmac_dma_tx_add()
162 dma_desc->ctl0 = cpu_to_le32(ctl0); in bgmac_dma_tx_add()
163 dma_desc->ctl1 = cpu_to_le32(ctl1); in bgmac_dma_tx_add()
314 struct bgmac_dma_desc *dma_desc = ring->cpu_base + desc_idx; in bgmac_dma_rx_setup_desc() local
325 dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[desc_idx].dma_addr)); in bgmac_dma_rx_setup_desc()
326 dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[desc_idx].dma_addr)); in bgmac_dma_rx_setup_desc()
[all …]
Dbcmsysport.h472 struct dma_desc { struct
485 #define WORDS_PER_DESC (sizeof(struct dma_desc) / sizeof(u32)) argument
638 struct dma_desc *desc_cpu; /* CPU view of the descriptor */
Db44.h241 struct dma_desc { struct
358 struct dma_desc *rx_ring, *tx_ring;
Db44.c67 #define B44_RX_RING_BYTES (sizeof(struct dma_desc) * \
71 #define B44_TX_RING_BYTES (sizeof(struct dma_desc) * \
623 cur /= sizeof(struct dma_desc); in b44_tx()
660 struct dma_desc *dp; in b44_alloc_rx_skb()
735 struct dma_desc *src_desc, *dest_desc; in b44_recycle_rx()
786 prod /= sizeof(struct dma_desc); in b44_rx()
861 bw32(bp, B44_DMARX_PTR, cons * sizeof(struct dma_desc)); in b44_rx()
1042 bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc)); in b44_start_xmit()
1044 bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc)); in b44_start_xmit()
1214 struct dma_desc *rx_ring; in b44_alloc_consistent()
[all …]
/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c590 int dma_desc; in tx_free_ring_skbufs() local
594 for (dma_desc = 0; dma_desc < tx_rsize; dma_desc++) { in tx_free_ring_skbufs()
595 struct sxgbe_tx_norm_desc *tdesc = txqueue->dma_tx + dma_desc; in tx_free_ring_skbufs()
597 if (txqueue->tx_skbuff_dma[dma_desc]) in tx_free_ring_skbufs()
599 txqueue->tx_skbuff_dma[dma_desc], in tx_free_ring_skbufs()
603 dev_kfree_skb_any(txqueue->tx_skbuff[dma_desc]); in tx_free_ring_skbufs()
604 txqueue->tx_skbuff[dma_desc] = NULL; in tx_free_ring_skbufs()
605 txqueue->tx_skbuff_dma[dma_desc] = 0; in tx_free_ring_skbufs()
/drivers/mmc/host/
Dmmci.c573 struct dma_async_tx_descriptor **dma_desc) in __mmci_dma_prep_data() argument
625 *dma_desc = desc; in __mmci_dma_prep_data()
650 return __mmci_dma_prep_data(host, data, &nd->dma_chan, &nd->dma_desc); in mmci_dma_prep_next()
692 WARN_ON(!data->host_cookie && (next->dma_desc || next->dma_chan)); in mmci_get_next_data()
694 host->dma_desc_current = next->dma_desc; in mmci_get_next_data()
696 next->dma_desc = NULL; in mmci_get_next_data()
739 next->dma_desc = NULL; in mmci_post_request()

12