/drivers/dma/ |
D | tegra20-apb-dma.c | 160 struct tegra_dma_desc *dma_desc; member 277 struct tegra_dma_desc *dma_desc; in tegra_dma_desc_get() local 283 list_for_each_entry(dma_desc, &tdc->free_dma_desc, node) { in tegra_dma_desc_get() 284 if (async_tx_test_ack(&dma_desc->txd)) { in tegra_dma_desc_get() 285 list_del(&dma_desc->node); in tegra_dma_desc_get() 287 dma_desc->txd.flags = 0; in tegra_dma_desc_get() 288 return dma_desc; in tegra_dma_desc_get() 295 dma_desc = kzalloc(sizeof(*dma_desc), GFP_ATOMIC); in tegra_dma_desc_get() 296 if (!dma_desc) { in tegra_dma_desc_get() 301 dma_async_tx_descriptor_init(&dma_desc->txd, &tdc->dma_chan); in tegra_dma_desc_get() [all …]
|
D | timb_dma.c | 157 static int td_fill_desc(struct timb_dma_chan *td_chan, u8 *dma_desc, in td_fill_desc() argument 173 dma_desc, (unsigned long long)sg_dma_address(sg)); in td_fill_desc() 175 dma_desc[7] = (sg_dma_address(sg) >> 24) & 0xff; in td_fill_desc() 176 dma_desc[6] = (sg_dma_address(sg) >> 16) & 0xff; in td_fill_desc() 177 dma_desc[5] = (sg_dma_address(sg) >> 8) & 0xff; in td_fill_desc() 178 dma_desc[4] = (sg_dma_address(sg) >> 0) & 0xff; in td_fill_desc() 180 dma_desc[3] = (sg_dma_len(sg) >> 8) & 0xff; in td_fill_desc() 181 dma_desc[2] = (sg_dma_len(sg) >> 0) & 0xff; in td_fill_desc() 183 dma_desc[1] = 0x00; in td_fill_desc() 184 dma_desc[0] = 0x21 | (last ? 0x02 : 0); /* tran, valid */ in td_fill_desc()
|
/drivers/net/ethernet/stmicro/stmmac/ |
D | descs_com.h | 36 static inline void ehn_desc_rx_set_on_ring(struct dma_desc *p, int end) in ehn_desc_rx_set_on_ring() 43 static inline void ehn_desc_tx_set_on_ring(struct dma_desc *p, int end) in ehn_desc_tx_set_on_ring() 49 static inline void enh_desc_end_tx_desc_on_ring(struct dma_desc *p, int ter) in enh_desc_end_tx_desc_on_ring() 54 static inline void enh_set_tx_desc_len_on_ring(struct dma_desc *p, int len) in enh_set_tx_desc_len_on_ring() 64 static inline void ndesc_rx_set_on_ring(struct dma_desc *p, int end) in ndesc_rx_set_on_ring() 71 static inline void ndesc_tx_set_on_ring(struct dma_desc *p, int end) in ndesc_tx_set_on_ring() 77 static inline void ndesc_end_tx_desc_on_ring(struct dma_desc *p, int ter) in ndesc_end_tx_desc_on_ring() 82 static inline void norm_set_tx_desc_len_on_ring(struct dma_desc *p, int len) in norm_set_tx_desc_len_on_ring() 94 static inline void ehn_desc_rx_set_on_chain(struct dma_desc *p, int end) in ehn_desc_rx_set_on_chain() 99 static inline void ehn_desc_tx_set_on_chain(struct dma_desc *p, int end) in ehn_desc_tx_set_on_chain() [all …]
|
D | norm_desc.c | 30 struct dma_desc *p, void __iomem *ioaddr) in ndesc_get_tx_status() 64 static int ndesc_get_tx_len(struct dma_desc *p) in ndesc_get_tx_len() 74 struct dma_desc *p) in ndesc_get_rx_status() 123 static void ndesc_init_rx_desc(struct dma_desc *p, int disable_rx_ic, int mode, in ndesc_init_rx_desc() 139 static void ndesc_init_tx_desc(struct dma_desc *p, int mode, int end) in ndesc_init_tx_desc() 148 static int ndesc_get_tx_owner(struct dma_desc *p) in ndesc_get_tx_owner() 153 static int ndesc_get_rx_owner(struct dma_desc *p) in ndesc_get_rx_owner() 158 static void ndesc_set_tx_owner(struct dma_desc *p) in ndesc_set_tx_owner() 163 static void ndesc_set_rx_owner(struct dma_desc *p) in ndesc_set_rx_owner() 168 static int ndesc_get_tx_ls(struct dma_desc *p) in ndesc_get_tx_ls() [all …]
|
D | common.h | 299 void (*init_rx_desc) (struct dma_desc *p, int disable_rx_ic, int mode, 302 void (*init_tx_desc) (struct dma_desc *p, int mode, int end); 305 void (*prepare_tx_desc) (struct dma_desc *p, int is_fs, int len, 308 void (*set_tx_owner) (struct dma_desc *p); 309 int (*get_tx_owner) (struct dma_desc *p); 311 void (*close_tx_desc) (struct dma_desc *p); 313 void (*release_tx_desc) (struct dma_desc *p, int mode); 316 void (*clear_tx_ic) (struct dma_desc *p); 318 int (*get_tx_ls) (struct dma_desc *p); 321 struct dma_desc *p, void __iomem *ioaddr); [all …]
|
D | enh_desc.c | 30 struct dma_desc *p, void __iomem *ioaddr) in enh_desc_get_tx_status() 88 static int enh_desc_get_tx_len(struct dma_desc *p) in enh_desc_get_tx_len() 178 struct dma_desc *p) in enh_desc_get_rx_status() 240 static void enh_desc_init_rx_desc(struct dma_desc *p, int disable_rx_ic, in enh_desc_init_rx_desc() 256 static void enh_desc_init_tx_desc(struct dma_desc *p, int mode, int end) in enh_desc_init_tx_desc() 265 static int enh_desc_get_tx_owner(struct dma_desc *p) in enh_desc_get_tx_owner() 270 static int enh_desc_get_rx_owner(struct dma_desc *p) in enh_desc_get_rx_owner() 275 static void enh_desc_set_tx_owner(struct dma_desc *p) in enh_desc_set_tx_owner() 280 static void enh_desc_set_rx_owner(struct dma_desc *p) in enh_desc_set_rx_owner() 285 static int enh_desc_get_tx_ls(struct dma_desc *p) in enh_desc_get_tx_ls() [all …]
|
D | chain_mode.c | 36 struct dma_desc *desc = priv->dma_tx + entry; in stmmac_jumbo_frm() 120 struct dma_desc *p = (struct dma_desc *)des; in stmmac_init_dma_chain() 122 dma_phy += sizeof(struct dma_desc); in stmmac_init_dma_chain() 130 static void stmmac_refill_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_refill_desc3() 142 sizeof(struct dma_desc)); in stmmac_refill_desc3() 145 static void stmmac_clean_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_clean_desc3() 157 sizeof(struct dma_desc))); in stmmac_clean_desc3()
|
D | ring_mode.c | 36 struct dma_desc *desc; in stmmac_jumbo_frm() 41 desc = (struct dma_desc *)(priv->dma_etx + entry); in stmmac_jumbo_frm() 68 desc = (struct dma_desc *)(priv->dma_etx + entry); in stmmac_jumbo_frm() 106 static void stmmac_refill_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_refill_desc3() 116 static void stmmac_init_desc3(struct dma_desc *p) in stmmac_init_desc3() 121 static void stmmac_clean_desc3(void *priv_ptr, struct dma_desc *p) in stmmac_clean_desc3()
|
D | stmmac_main.c | 375 if (!priv->hw->desc->get_tx_timestamp_status((struct dma_desc *)desc)) in stmmac_get_tx_hwtstamp() 882 struct dma_desc *p = (struct dma_desc *)head; in stmmac_display_ring() 972 static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p, in stmmac_init_rx_buffers() 1049 struct dma_desc *p; in init_dma_desc_rings() 1085 struct dma_desc *p; in init_dma_desc_rings() 1124 struct dma_desc *p; in dma_free_tx_skbufs() 1202 sizeof(struct dma_desc), in alloc_dma_desc_resources() 1209 sizeof(struct dma_desc), in alloc_dma_desc_resources() 1214 sizeof(struct dma_desc), in alloc_dma_desc_resources() 1242 priv->dma_tx_size * sizeof(struct dma_desc), in free_dma_desc_resources() [all …]
|
D | stmmac.h | 45 struct dma_desc *dma_tx; 61 struct dma_desc *dma_rx ____cacheline_aligned_in_smp;
|
D | descs.h | 28 struct dma_desc { struct 170 struct dma_desc basic; argument
|
/drivers/ata/ |
D | pata_pxa.c | 42 struct pxa_dma_desc *dma_desc; member 73 pd->dma_desc[pd->dma_desc_id].ddadr = pd->dma_desc_addr + in pxa_load_dmac() 76 pd->dma_desc[pd->dma_desc_id].dcmd = DCMD_BURST32 | in pxa_load_dmac() 80 pd->dma_desc[pd->dma_desc_id].dsadr = cpu_addr; in pxa_load_dmac() 81 pd->dma_desc[pd->dma_desc_id].dtadr = pd->dma_io_addr; in pxa_load_dmac() 82 pd->dma_desc[pd->dma_desc_id].dcmd |= DCMD_INCSRCADDR | in pxa_load_dmac() 85 pd->dma_desc[pd->dma_desc_id].dsadr = pd->dma_io_addr; in pxa_load_dmac() 86 pd->dma_desc[pd->dma_desc_id].dtadr = cpu_addr; in pxa_load_dmac() 87 pd->dma_desc[pd->dma_desc_id].dcmd |= DCMD_INCTRGADDR | in pxa_load_dmac() 122 pd->dma_desc[pd->dma_desc_id - 1].ddadr = DDADR_STOP; in pxa_qc_prep() [all …]
|
/drivers/usb/dwc2/ |
D | hcd_ddma.c | 512 struct dwc2_hcd_dma_desc *dma_desc = &qh->desc_list[idx]; in dwc2_fill_host_isoc_dma_desc() local 515 memset(dma_desc, 0, sizeof(*dma_desc)); in dwc2_fill_host_isoc_dma_desc() 523 dma_desc->buf = (u32)(qtd->urb->dma + frame_desc->offset); in dwc2_fill_host_isoc_dma_desc() 524 dma_desc->status = qh->n_bytes[idx] << HOST_DMA_ISOC_NBYTES_SHIFT & in dwc2_fill_host_isoc_dma_desc() 530 dma_desc->status |= HOST_DMA_IOC; in dwc2_fill_host_isoc_dma_desc() 621 struct dwc2_hcd_dma_desc *dma_desc = &qh->desc_list[n_desc]; in dwc2_fill_host_dma_desc() local 641 dma_desc->status = len << HOST_DMA_NBYTES_SHIFT & HOST_DMA_NBYTES_MASK; in dwc2_fill_host_dma_desc() 646 dma_desc->status |= HOST_DMA_SUP; in dwc2_fill_host_dma_desc() 648 dma_desc->buf = (u32)chan->xfer_dma; in dwc2_fill_host_dma_desc() 803 struct dwc2_hcd_dma_desc *dma_desc = &qh->desc_list[idx]; in dwc2_cmpl_host_isoc_dma_desc() local [all …]
|
/drivers/soc/tegra/fuse/ |
D | fuse-tegra20.c | 61 struct dma_async_tx_descriptor *dma_desc; in tegra20_fuse_readl() local 70 dma_desc = dmaengine_prep_slave_single(apb_dma_chan, apb_buffer_phys, in tegra20_fuse_readl() 73 if (!dma_desc) in tegra20_fuse_readl() 76 dma_desc->callback = apb_dma_complete; in tegra20_fuse_readl() 77 dma_desc->callback_param = NULL; in tegra20_fuse_readl() 83 dmaengine_submit(dma_desc); in tegra20_fuse_readl()
|
/drivers/usb/musb/ |
D | musb_cppi41.c | 152 struct dma_async_tx_descriptor *dma_desc; in cppi41_trans_done() local 165 dma_desc = dmaengine_prep_slave_single(dc, in cppi41_trans_done() 170 if (WARN_ON(!dma_desc)) in cppi41_trans_done() 173 dma_desc->callback = cppi41_dma_callback; in cppi41_trans_done() 174 dma_desc->callback_param = &cppi41_channel->channel; in cppi41_trans_done() 175 cppi41_channel->cookie = dma_desc->tx_submit(dma_desc); in cppi41_trans_done() 364 struct dma_async_tx_descriptor *dma_desc; in cppi41_configure_channel() local 416 dma_desc = dmaengine_prep_slave_single(dc, dma_addr, len, direction, in cppi41_configure_channel() 418 if (!dma_desc) in cppi41_configure_channel() 421 dma_desc->callback = cppi41_dma_callback; in cppi41_configure_channel() [all …]
|
D | ux500_dma.c | 89 struct dma_async_tx_descriptor *dma_desc; in ux500_configure_channel() local 127 dma_desc = dmaengine_prep_slave_sg(dma_chan, &sg, 1, direction, in ux500_configure_channel() 129 if (!dma_desc) in ux500_configure_channel() 132 dma_desc->callback = ux500_dma_callback; in ux500_configure_channel() 133 dma_desc->callback_param = channel; in ux500_configure_channel() 134 ux500_channel->cookie = dma_desc->tx_submit(dma_desc); in ux500_configure_channel()
|
/drivers/staging/comedi/drivers/ |
D | gsc_hpdi.c | 156 struct plx_dma_desc *dma_desc; member 188 start = le32_to_cpu(devpriv->dma_desc[idx].pci_start_addr); in gsc_hpdi_drain_dma() 203 start = le32_to_cpu(devpriv->dma_desc[idx].pci_start_addr); in gsc_hpdi_drain_dma() 458 devpriv->dma_desc[i].pci_start_addr = in gsc_hpdi_setup_dma_descriptors() 460 devpriv->dma_desc[i].local_start_addr = cpu_to_le32(FIFO_REG); in gsc_hpdi_setup_dma_descriptors() 461 devpriv->dma_desc[i].transfer_size = cpu_to_le32(len); in gsc_hpdi_setup_dma_descriptors() 462 devpriv->dma_desc[i].next = cpu_to_le32((phys_addr + in gsc_hpdi_setup_dma_descriptors() 463 (i + 1) * sizeof(devpriv->dma_desc[0])) | next_bits); in gsc_hpdi_setup_dma_descriptors() 476 devpriv->dma_desc[i - 1].next = cpu_to_le32(phys_addr | next_bits); in gsc_hpdi_setup_dma_descriptors() 526 if (devpriv->dma_desc) in gsc_hpdi_free_dma() [all …]
|
/drivers/video/fbdev/ |
D | pxafb.c | 1065 struct pxafb_dma_descriptor *dma_desc, *pal_desc; in setup_frame_dma() local 1071 dma_desc = &fbi->dma_buff->dma_desc[dma]; in setup_frame_dma() 1072 dma_desc_off = offsetof(struct pxafb_dma_buff, dma_desc[dma]); in setup_frame_dma() 1074 dma_desc->fsadr = start; in setup_frame_dma() 1075 dma_desc->fidr = 0; in setup_frame_dma() 1076 dma_desc->ldcmd = size; in setup_frame_dma() 1079 dma_desc->fdadr = fbi->dma_buff_phys + dma_desc_off; in setup_frame_dma() 1097 dma_desc->fdadr = fbi->dma_buff_phys + pal_desc_off; in setup_frame_dma() 1129 struct pxafb_dma_descriptor *dma_desc; in setup_smart_dma() local 1132 dma_desc = &fbi->dma_buff->dma_desc[DMA_CMD]; in setup_smart_dma() [all …]
|
/drivers/net/ethernet/ |
D | korina.c | 90 #define RD_RING_SIZE (KORINA_NUM_RDS * sizeof(struct dma_desc)) 91 #define TD_RING_SIZE (KORINA_NUM_TDS * sizeof(struct dma_desc)) 105 struct dma_desc *td_ring; /* transmit descriptor ring */ 106 struct dma_desc *rd_ring; /* receive descriptor ring */ 184 struct dma_desc *rd) in korina_start_rx() 190 struct dma_desc *rd) in korina_chain_rx() 202 struct dma_desc *td; in korina_send_packet() 351 struct dma_desc *rd = &lp->rd_ring[lp->rx_next_done]; in korina_rx() 522 struct dma_desc *td = &lp->td_ring[lp->tx_next_done]; in korina_tx() 1154 lp->td_ring = (struct dma_desc *)KSEG1ADDR(lp->td_ring); in korina_probe()
|
/drivers/net/ethernet/broadcom/ |
D | bgmac.c | 123 struct bgmac_dma_desc *dma_desc; in bgmac_dma_tx_add() local 158 dma_desc = ring->cpu_base; in bgmac_dma_tx_add() 159 dma_desc += ring->end; in bgmac_dma_tx_add() 160 dma_desc->addr_low = cpu_to_le32(lower_32_bits(slot->dma_addr)); in bgmac_dma_tx_add() 161 dma_desc->addr_high = cpu_to_le32(upper_32_bits(slot->dma_addr)); in bgmac_dma_tx_add() 162 dma_desc->ctl0 = cpu_to_le32(ctl0); in bgmac_dma_tx_add() 163 dma_desc->ctl1 = cpu_to_le32(ctl1); in bgmac_dma_tx_add() 314 struct bgmac_dma_desc *dma_desc = ring->cpu_base + desc_idx; in bgmac_dma_rx_setup_desc() local 325 dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[desc_idx].dma_addr)); in bgmac_dma_rx_setup_desc() 326 dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[desc_idx].dma_addr)); in bgmac_dma_rx_setup_desc() [all …]
|
D | bcmsysport.h | 472 struct dma_desc { struct 485 #define WORDS_PER_DESC (sizeof(struct dma_desc) / sizeof(u32)) argument 638 struct dma_desc *desc_cpu; /* CPU view of the descriptor */
|
D | b44.h | 241 struct dma_desc { struct 358 struct dma_desc *rx_ring, *tx_ring;
|
D | b44.c | 67 #define B44_RX_RING_BYTES (sizeof(struct dma_desc) * \ 71 #define B44_TX_RING_BYTES (sizeof(struct dma_desc) * \ 623 cur /= sizeof(struct dma_desc); in b44_tx() 660 struct dma_desc *dp; in b44_alloc_rx_skb() 735 struct dma_desc *src_desc, *dest_desc; in b44_recycle_rx() 786 prod /= sizeof(struct dma_desc); in b44_rx() 861 bw32(bp, B44_DMARX_PTR, cons * sizeof(struct dma_desc)); in b44_rx() 1042 bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc)); in b44_start_xmit() 1044 bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc)); in b44_start_xmit() 1214 struct dma_desc *rx_ring; in b44_alloc_consistent() [all …]
|
/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_main.c | 590 int dma_desc; in tx_free_ring_skbufs() local 594 for (dma_desc = 0; dma_desc < tx_rsize; dma_desc++) { in tx_free_ring_skbufs() 595 struct sxgbe_tx_norm_desc *tdesc = txqueue->dma_tx + dma_desc; in tx_free_ring_skbufs() 597 if (txqueue->tx_skbuff_dma[dma_desc]) in tx_free_ring_skbufs() 599 txqueue->tx_skbuff_dma[dma_desc], in tx_free_ring_skbufs() 603 dev_kfree_skb_any(txqueue->tx_skbuff[dma_desc]); in tx_free_ring_skbufs() 604 txqueue->tx_skbuff[dma_desc] = NULL; in tx_free_ring_skbufs() 605 txqueue->tx_skbuff_dma[dma_desc] = 0; in tx_free_ring_skbufs()
|
/drivers/mmc/host/ |
D | mmci.c | 573 struct dma_async_tx_descriptor **dma_desc) in __mmci_dma_prep_data() argument 625 *dma_desc = desc; in __mmci_dma_prep_data() 650 return __mmci_dma_prep_data(host, data, &nd->dma_chan, &nd->dma_desc); in mmci_dma_prep_next() 692 WARN_ON(!data->host_cookie && (next->dma_desc || next->dma_chan)); in mmci_get_next_data() 694 host->dma_desc_current = next->dma_desc; in mmci_get_next_data() 696 next->dma_desc = NULL; in mmci_get_next_data() 739 next->dma_desc = NULL; in mmci_post_request()
|