Lines Matching +full:dma +full:- +full:maxburst
2 * Core driver for the Intel integrated DMA 64-bit
15 #include <linux/dma-mapping.h>
30 /* ---------------------------------------------------------------------- */
34 return &chan->dev->device; in chan2dev()
37 /* ---------------------------------------------------------------------- */
45 channel_clear_bit(idma64, MASK(XFER), idma64->all_chan_mask); in idma64_off()
46 channel_clear_bit(idma64, MASK(BLOCK), idma64->all_chan_mask); in idma64_off()
47 channel_clear_bit(idma64, MASK(SRC_TRAN), idma64->all_chan_mask); in idma64_off()
48 channel_clear_bit(idma64, MASK(DST_TRAN), idma64->all_chan_mask); in idma64_off()
49 channel_clear_bit(idma64, MASK(ERROR), idma64->all_chan_mask); in idma64_off()
53 } while (dma_readl(idma64, CFG) & IDMA64_CFG_DMA_EN && --count); in idma64_off()
61 /* ---------------------------------------------------------------------- */
75 channel_set_bit(idma64, MASK(XFER), idma64c->mask); in idma64_chan_init()
76 channel_set_bit(idma64, MASK(ERROR), idma64c->mask); in idma64_chan_init()
81 * The iDMA is turned off in ->probe() and looses context during system in idma64_chan_init()
90 channel_clear_bit(idma64, CH_EN, idma64c->mask); in idma64_chan_stop()
95 struct idma64_desc *desc = idma64c->desc; in idma64_chan_start()
96 struct idma64_hw_desc *hw = &desc->hw[0]; in idma64_chan_start()
104 channel_writeq(idma64c, LLP, hw->llp); in idma64_chan_start()
106 channel_set_bit(idma64, CH_EN, idma64c->mask); in idma64_chan_start()
111 struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device); in idma64_stop_transfer()
118 struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device); in idma64_start_transfer()
122 vdesc = vchan_next_desc(&idma64c->vchan); in idma64_start_transfer()
124 idma64c->desc = NULL; in idma64_start_transfer()
128 list_del(&vdesc->node); in idma64_start_transfer()
129 idma64c->desc = to_idma64_desc(vdesc); in idma64_start_transfer()
138 /* ---------------------------------------------------------------------- */
143 struct idma64_chan *idma64c = &idma64->chan[c]; in idma64_chan_irq()
147 spin_lock_irqsave(&idma64c->vchan.lock, flags); in idma64_chan_irq()
148 desc = idma64c->desc; in idma64_chan_irq()
151 dma_writel(idma64, CLEAR(ERROR), idma64c->mask); in idma64_chan_irq()
152 desc->status = DMA_ERROR; in idma64_chan_irq()
154 dma_writel(idma64, CLEAR(XFER), idma64c->mask); in idma64_chan_irq()
155 desc->status = DMA_COMPLETE; in idma64_chan_irq()
156 vchan_cookie_complete(&desc->vdesc); in idma64_chan_irq()
160 /* idma64_start_transfer() updates idma64c->desc */ in idma64_chan_irq()
161 if (idma64c->desc == NULL || desc->status == DMA_ERROR) in idma64_chan_irq()
164 spin_unlock_irqrestore(&idma64c->vchan.lock, flags); in idma64_chan_irq()
175 dev_vdbg(idma64->dma.dev, "%s: status=%#x\n", __func__, status); in idma64_irq()
177 /* Check if we have any interrupt from the DMA controller */ in idma64_irq()
184 for (i = 0; i < idma64->dma.chancnt; i++) in idma64_irq()
190 /* ---------------------------------------------------------------------- */
200 desc->hw = kcalloc(ndesc, sizeof(*desc->hw), GFP_NOWAIT); in idma64_alloc_desc()
201 if (!desc->hw) { in idma64_alloc_desc()
214 if (desc->ndesc) { in idma64_desc_free()
215 unsigned int i = desc->ndesc; in idma64_desc_free()
218 hw = &desc->hw[--i]; in idma64_desc_free()
219 dma_pool_free(idma64c->pool, hw->lli, hw->llp); in idma64_desc_free()
223 kfree(desc->hw); in idma64_desc_free()
229 struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan); in idma64_vdesc_free()
238 struct idma64_lli *lli = hw->lli; in idma64_hw_desc_fill()
240 u32 ctlhi = IDMA64C_CTLH_BLOCK_TS(hw->len); in idma64_hw_desc_fill()
245 sar = hw->phys; in idma64_hw_desc_fill()
246 dar = config->dst_addr; in idma64_hw_desc_fill()
249 src_width = __ffs(sar | hw->len | 4); in idma64_hw_desc_fill()
250 dst_width = __ffs(config->dst_addr_width); in idma64_hw_desc_fill()
252 sar = config->src_addr; in idma64_hw_desc_fill()
253 dar = hw->phys; in idma64_hw_desc_fill()
256 src_width = __ffs(config->src_addr_width); in idma64_hw_desc_fill()
257 dst_width = __ffs(dar | hw->len | 4); in idma64_hw_desc_fill()
260 lli->sar = sar; in idma64_hw_desc_fill()
261 lli->dar = dar; in idma64_hw_desc_fill()
263 lli->ctlhi = ctlhi; in idma64_hw_desc_fill()
264 lli->ctllo = ctllo | in idma64_hw_desc_fill()
265 IDMA64C_CTLL_SRC_MSIZE(config->src_maxburst) | in idma64_hw_desc_fill()
266 IDMA64C_CTLL_DST_MSIZE(config->dst_maxburst) | in idma64_hw_desc_fill()
270 lli->llp = llp; in idma64_hw_desc_fill()
276 struct dma_slave_config *config = &idma64c->config; in idma64_desc_fill()
277 unsigned int i = desc->ndesc; in idma64_desc_fill()
278 struct idma64_hw_desc *hw = &desc->hw[i - 1]; in idma64_desc_fill()
279 struct idma64_lli *lli = hw->lli; in idma64_desc_fill()
284 hw = &desc->hw[--i]; in idma64_desc_fill()
285 idma64_hw_desc_fill(hw, config, desc->direction, llp); in idma64_desc_fill()
286 llp = hw->llp; in idma64_desc_fill()
287 desc->length += hw->len; in idma64_desc_fill()
291 lli->ctllo |= IDMA64C_CTLL_INT_EN; in idma64_desc_fill()
294 lli->ctllo &= ~(IDMA64C_CTLL_LLP_S_EN | IDMA64C_CTLL_LLP_D_EN); in idma64_desc_fill()
312 struct idma64_hw_desc *hw = &desc->hw[i]; in idma64_prep_slave_sg()
314 /* Allocate DMA capable memory for hardware descriptor */ in idma64_prep_slave_sg()
315 hw->lli = dma_pool_alloc(idma64c->pool, GFP_NOWAIT, &hw->llp); in idma64_prep_slave_sg()
316 if (!hw->lli) { in idma64_prep_slave_sg()
317 desc->ndesc = i; in idma64_prep_slave_sg()
322 hw->phys = sg_dma_address(sg); in idma64_prep_slave_sg()
323 hw->len = sg_dma_len(sg); in idma64_prep_slave_sg()
326 desc->ndesc = sg_len; in idma64_prep_slave_sg()
327 desc->direction = direction; in idma64_prep_slave_sg()
328 desc->status = DMA_IN_PROGRESS; in idma64_prep_slave_sg()
331 return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags); in idma64_prep_slave_sg()
339 spin_lock_irqsave(&idma64c->vchan.lock, flags); in idma64_issue_pending()
340 if (vchan_issue_pending(&idma64c->vchan) && !idma64c->desc) in idma64_issue_pending()
342 spin_unlock_irqrestore(&idma64c->vchan.lock, flags); in idma64_issue_pending()
347 struct idma64_desc *desc = idma64c->desc; in idma64_active_desc_size()
349 size_t bytes = desc->length; in idma64_active_desc_size()
355 hw = &desc->hw[i]; in idma64_active_desc_size()
356 if (hw->llp == llp) in idma64_active_desc_size()
358 bytes -= hw->len; in idma64_active_desc_size()
359 } while (++i < desc->ndesc); in idma64_active_desc_size()
365 bytes += desc->hw[--i].len; in idma64_active_desc_size()
367 return bytes - IDMA64C_CTLH_BLOCK_TS(ctlhi); in idma64_active_desc_size()
383 spin_lock_irqsave(&idma64c->vchan.lock, flags); in idma64_tx_status()
384 vdesc = vchan_find_desc(&idma64c->vchan, cookie); in idma64_tx_status()
385 if (idma64c->desc && cookie == idma64c->desc->vdesc.tx.cookie) { in idma64_tx_status()
388 status = idma64c->desc->status; in idma64_tx_status()
390 bytes = to_idma64_desc(vdesc)->length; in idma64_tx_status()
393 spin_unlock_irqrestore(&idma64c->vchan.lock, flags); in idma64_tx_status()
398 static void convert_burst(u32 *maxburst) in convert_burst() argument
400 if (*maxburst) in convert_burst()
401 *maxburst = __fls(*maxburst); in convert_burst()
403 *maxburst = 0; in convert_burst()
412 if (!is_slave_direction(config->direction)) in idma64_slave_config()
413 return -EINVAL; in idma64_slave_config()
415 memcpy(&idma64c->config, config, sizeof(idma64c->config)); in idma64_slave_config()
417 convert_burst(&idma64c->config.src_maxburst); in idma64_slave_config()
418 convert_burst(&idma64c->config.dst_maxburst); in idma64_slave_config()
438 } while (!(cfglo & IDMA64C_CFGL_FIFO_EMPTY) && --count); in idma64_chan_deactivate()
454 spin_lock_irqsave(&idma64c->vchan.lock, flags); in idma64_pause()
455 if (idma64c->desc && idma64c->desc->status == DMA_IN_PROGRESS) { in idma64_pause()
457 idma64c->desc->status = DMA_PAUSED; in idma64_pause()
459 spin_unlock_irqrestore(&idma64c->vchan.lock, flags); in idma64_pause()
469 spin_lock_irqsave(&idma64c->vchan.lock, flags); in idma64_resume()
470 if (idma64c->desc && idma64c->desc->status == DMA_PAUSED) { in idma64_resume()
471 idma64c->desc->status = DMA_IN_PROGRESS; in idma64_resume()
474 spin_unlock_irqrestore(&idma64c->vchan.lock, flags); in idma64_resume()
485 spin_lock_irqsave(&idma64c->vchan.lock, flags); in idma64_terminate_all()
488 if (idma64c->desc) { in idma64_terminate_all()
489 idma64_vdesc_free(&idma64c->desc->vdesc); in idma64_terminate_all()
490 idma64c->desc = NULL; in idma64_terminate_all()
492 vchan_get_all_descriptors(&idma64c->vchan, &head); in idma64_terminate_all()
493 spin_unlock_irqrestore(&idma64c->vchan.lock, flags); in idma64_terminate_all()
495 vchan_dma_desc_free_list(&idma64c->vchan, &head); in idma64_terminate_all()
503 vchan_synchronize(&idma64c->vchan); in idma64_synchronize()
511 idma64c->pool = dma_pool_create(dev_name(chan2dev(chan)), in idma64_alloc_chan_resources()
512 chan->device->dev, in idma64_alloc_chan_resources()
514 if (!idma64c->pool) { in idma64_alloc_chan_resources()
516 return -ENOMEM; in idma64_alloc_chan_resources()
527 dma_pool_destroy(idma64c->pool); in idma64_free_chan_resources()
528 idma64c->pool = NULL; in idma64_free_chan_resources()
531 /* ---------------------------------------------------------------------- */
545 idma64 = devm_kzalloc(chip->dev, sizeof(*idma64), GFP_KERNEL); in idma64_probe()
547 return -ENOMEM; in idma64_probe()
549 idma64->regs = chip->regs; in idma64_probe()
550 chip->idma64 = idma64; in idma64_probe()
552 idma64->chan = devm_kcalloc(chip->dev, nr_chan, sizeof(*idma64->chan), in idma64_probe()
554 if (!idma64->chan) in idma64_probe()
555 return -ENOMEM; in idma64_probe()
557 idma64->all_chan_mask = (1 << nr_chan) - 1; in idma64_probe()
562 ret = devm_request_irq(chip->dev, chip->irq, idma64_irq, IRQF_SHARED, in idma64_probe()
563 dev_name(chip->dev), idma64); in idma64_probe()
567 INIT_LIST_HEAD(&idma64->dma.channels); in idma64_probe()
569 struct idma64_chan *idma64c = &idma64->chan[i]; in idma64_probe()
571 idma64c->vchan.desc_free = idma64_vdesc_free; in idma64_probe()
572 vchan_init(&idma64c->vchan, &idma64->dma); in idma64_probe()
574 idma64c->regs = idma64->regs + i * IDMA64_CH_LENGTH; in idma64_probe()
575 idma64c->mask = BIT(i); in idma64_probe()
578 dma_cap_set(DMA_SLAVE, idma64->dma.cap_mask); in idma64_probe()
579 dma_cap_set(DMA_PRIVATE, idma64->dma.cap_mask); in idma64_probe()
581 idma64->dma.device_alloc_chan_resources = idma64_alloc_chan_resources; in idma64_probe()
582 idma64->dma.device_free_chan_resources = idma64_free_chan_resources; in idma64_probe()
584 idma64->dma.device_prep_slave_sg = idma64_prep_slave_sg; in idma64_probe()
586 idma64->dma.device_issue_pending = idma64_issue_pending; in idma64_probe()
587 idma64->dma.device_tx_status = idma64_tx_status; in idma64_probe()
589 idma64->dma.device_config = idma64_slave_config; in idma64_probe()
590 idma64->dma.device_pause = idma64_pause; in idma64_probe()
591 idma64->dma.device_resume = idma64_resume; in idma64_probe()
592 idma64->dma.device_terminate_all = idma64_terminate_all; in idma64_probe()
593 idma64->dma.device_synchronize = idma64_synchronize; in idma64_probe()
595 idma64->dma.src_addr_widths = IDMA64_BUSWIDTHS; in idma64_probe()
596 idma64->dma.dst_addr_widths = IDMA64_BUSWIDTHS; in idma64_probe()
597 idma64->dma.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); in idma64_probe()
598 idma64->dma.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST; in idma64_probe()
600 idma64->dma.dev = chip->sysdev; in idma64_probe()
602 dma_set_max_seg_size(idma64->dma.dev, IDMA64C_CTLH_BLOCK_TS_MASK); in idma64_probe()
604 ret = dma_async_device_register(&idma64->dma); in idma64_probe()
608 dev_info(chip->dev, "Found Intel integrated DMA 64-bit\n"); in idma64_probe()
614 struct idma64 *idma64 = chip->idma64; in idma64_remove()
617 dma_async_device_unregister(&idma64->dma); in idma64_remove()
623 devm_free_irq(chip->dev, chip->irq, idma64); in idma64_remove()
625 for (i = 0; i < idma64->dma.chancnt; i++) { in idma64_remove()
626 struct idma64_chan *idma64c = &idma64->chan[i]; in idma64_remove()
628 tasklet_kill(&idma64c->vchan.task); in idma64_remove()
634 /* ---------------------------------------------------------------------- */
639 struct device *dev = &pdev->dev; in idma64_platform_probe()
640 struct device *sysdev = dev->parent; in idma64_platform_probe()
646 return -ENOMEM; in idma64_platform_probe()
648 chip->irq = platform_get_irq(pdev, 0); in idma64_platform_probe()
649 if (chip->irq < 0) in idma64_platform_probe()
650 return chip->irq; in idma64_platform_probe()
653 chip->regs = devm_ioremap_resource(dev, mem); in idma64_platform_probe()
654 if (IS_ERR(chip->regs)) in idma64_platform_probe()
655 return PTR_ERR(chip->regs); in idma64_platform_probe()
661 chip->dev = dev; in idma64_platform_probe()
662 chip->sysdev = sysdev; in idma64_platform_probe()
685 idma64_off(chip->idma64); in idma64_pm_suspend()
693 idma64_on(chip->idma64); in idma64_pm_resume()