• Home
  • Raw
  • Download

Lines Matching refs:dma

117 	struct knav_dma_device		*dma;  member
257 static void dma_hw_enable_all(struct knav_dma_device *dma) in dma_hw_enable_all() argument
261 for (i = 0; i < dma->max_tx_chan; i++) { in dma_hw_enable_all()
262 writel_relaxed(0, &dma->reg_tx_chan[i].mode); in dma_hw_enable_all()
263 writel_relaxed(DMA_ENABLE, &dma->reg_tx_chan[i].control); in dma_hw_enable_all()
268 static void knav_dma_hw_init(struct knav_dma_device *dma) in knav_dma_hw_init() argument
273 spin_lock(&dma->lock); in knav_dma_hw_init()
274 v = dma->loopback ? DMA_LOOPBACK : 0; in knav_dma_hw_init()
275 writel_relaxed(v, &dma->reg_global->emulation_control); in knav_dma_hw_init()
277 v = readl_relaxed(&dma->reg_global->perf_control); in knav_dma_hw_init()
278 v |= ((dma->rx_timeout & DMA_RX_TIMEOUT_MASK) << DMA_RX_TIMEOUT_SHIFT); in knav_dma_hw_init()
279 writel_relaxed(v, &dma->reg_global->perf_control); in knav_dma_hw_init()
281 v = ((dma->tx_priority << DMA_TX_PRIO_SHIFT) | in knav_dma_hw_init()
282 (dma->rx_priority << DMA_RX_PRIO_SHIFT)); in knav_dma_hw_init()
284 writel_relaxed(v, &dma->reg_global->priority_control); in knav_dma_hw_init()
287 for (i = 0; i < dma->max_rx_chan; i++) in knav_dma_hw_init()
288 writel_relaxed(DMA_ENABLE, &dma->reg_rx_chan[i].control); in knav_dma_hw_init()
290 for (i = 0; i < dma->logical_queue_managers; i++) in knav_dma_hw_init()
291 writel_relaxed(dma->qm_base_address[i], in knav_dma_hw_init()
292 &dma->reg_global->qm_base_address[i]); in knav_dma_hw_init()
293 spin_unlock(&dma->lock); in knav_dma_hw_init()
296 static void knav_dma_hw_destroy(struct knav_dma_device *dma) in knav_dma_hw_destroy() argument
301 spin_lock(&dma->lock); in knav_dma_hw_destroy()
304 for (i = 0; i < dma->max_rx_chan; i++) in knav_dma_hw_destroy()
305 writel_relaxed(v, &dma->reg_rx_chan[i].control); in knav_dma_hw_destroy()
307 for (i = 0; i < dma->max_tx_chan; i++) in knav_dma_hw_destroy()
308 writel_relaxed(v, &dma->reg_tx_chan[i].control); in knav_dma_hw_destroy()
309 spin_unlock(&dma->lock); in knav_dma_hw_destroy()
341 struct knav_dma_device *dma) in dma_debug_show_devices() argument
345 list_for_each_entry(chan, &dma->chan_list, list) { in dma_debug_show_devices()
353 struct knav_dma_device *dma; in dma_debug_show() local
355 list_for_each_entry(dma, &kdev->list, list) { in dma_debug_show()
356 if (atomic_read(&dma->ref_count)) { in dma_debug_show()
358 dma->name, dma->max_tx_chan, dma->max_rx_flow); in dma_debug_show()
359 dma_debug_show_devices(s, dma); in dma_debug_show()
422 struct knav_dma_device *dma; in knav_dma_open_channel() local
450 list_for_each_entry(dma, &kdev->list, list) { in knav_dma_open_channel()
451 if (!strcmp(dma->name, instance)) { in knav_dma_open_channel()
463 list_for_each_entry(chan, &dma->chan_list, list) { in knav_dma_open_channel()
490 if (atomic_inc_return(&chan->dma->ref_count) <= 1) in knav_dma_open_channel()
491 knav_dma_hw_init(chan->dma); in knav_dma_open_channel()
521 if (atomic_dec_return(&chan->dma->ref_count) <= 0) in knav_dma_close_channel()
522 knav_dma_hw_destroy(chan->dma); in knav_dma_close_channel()
525 chan->channel, chan->flow, chan->dma->name); in knav_dma_close_channel()
529 static void __iomem *pktdma_get_regs(struct knav_dma_device *dma, in pktdma_get_regs() argument
557 struct knav_dma_device *dma = chan->dma; in pktdma_init_rx_chan() local
560 chan->reg_rx_flow = dma->reg_rx_flow + flow; in pktdma_init_rx_chan()
569 struct knav_dma_device *dma = chan->dma; in pktdma_init_tx_chan() local
572 chan->reg_chan = dma->reg_tx_chan + channel; in pktdma_init_tx_chan()
573 chan->reg_tx_sched = dma->reg_tx_sched + channel; in pktdma_init_tx_chan()
580 static int pktdma_init_chan(struct knav_dma_device *dma, in pktdma_init_chan() argument
593 chan->dma = dma; in pktdma_init_chan()
608 list_add_tail(&chan->list, &dma->chan_list); in pktdma_init_chan()
617 struct knav_dma_device *dma; in dma_init() local
623 dma = devm_kzalloc(kdev->dev, sizeof(*dma), GFP_KERNEL); in dma_init()
624 if (!dma) { in dma_init()
628 INIT_LIST_HEAD(&dma->list); in dma_init()
629 INIT_LIST_HEAD(&dma->chan_list); in dma_init()
636 dma->logical_queue_managers = len / sizeof(u32); in dma_init()
637 if (dma->logical_queue_managers > DMA_MAX_QMS) { in dma_init()
639 dma->logical_queue_managers); in dma_init()
640 dma->logical_queue_managers = DMA_MAX_QMS; in dma_init()
644 dma->qm_base_address, in dma_init()
645 dma->logical_queue_managers); in dma_init()
651 dma->reg_global = pktdma_get_regs(dma, node, 0, &size); in dma_init()
652 if (!dma->reg_global) in dma_init()
659 dma->reg_tx_chan = pktdma_get_regs(dma, node, 1, &size); in dma_init()
660 if (!dma->reg_tx_chan) in dma_init()
664 dma->reg_rx_chan = pktdma_get_regs(dma, node, 2, &size); in dma_init()
665 if (!dma->reg_rx_chan) in dma_init()
669 dma->reg_tx_sched = pktdma_get_regs(dma, node, 3, &size); in dma_init()
670 if (!dma->reg_tx_sched) in dma_init()
674 dma->reg_rx_flow = pktdma_get_regs(dma, node, 4, &size); in dma_init()
675 if (!dma->reg_rx_flow) in dma_init()
679 dma->rx_priority = DMA_PRIO_DEFAULT; in dma_init()
680 dma->tx_priority = DMA_PRIO_DEFAULT; in dma_init()
682 dma->enable_all = (of_get_property(node, "ti,enable-all", NULL) != NULL); in dma_init()
683 dma->loopback = (of_get_property(node, "ti,loop-back", NULL) != NULL); in dma_init()
692 dma->rx_timeout = timeout; in dma_init()
693 dma->max_rx_chan = max_rx_chan; in dma_init()
694 dma->max_rx_flow = max_rx_flow; in dma_init()
695 dma->max_tx_chan = min(max_tx_chan, max_tx_sched); in dma_init()
696 atomic_set(&dma->ref_count, 0); in dma_init()
697 strcpy(dma->name, node->name); in dma_init()
698 spin_lock_init(&dma->lock); in dma_init()
700 for (i = 0; i < dma->max_tx_chan; i++) { in dma_init()
701 if (pktdma_init_chan(dma, DMA_MEM_TO_DEV, i) >= 0) in dma_init()
705 for (i = 0; i < dma->max_rx_flow; i++) { in dma_init()
706 if (pktdma_init_chan(dma, DMA_DEV_TO_MEM, i) >= 0) in dma_init()
710 list_add_tail(&dma->list, &kdev->list); in dma_init()
716 if (dma->enable_all) { in dma_init()
717 atomic_inc(&dma->ref_count); in dma_init()
718 knav_dma_hw_init(dma); in dma_init()
719 dma_hw_enable_all(dma); in dma_init()
723 dma->name, num_chan, dma->max_rx_flow, in dma_init()
724 dma->max_tx_chan, dma->max_rx_chan, in dma_init()
725 dma->loopback ? ", loopback" : ""); in dma_init()
790 struct knav_dma_device *dma; in knav_dma_remove() local
792 list_for_each_entry(dma, &kdev->list, list) { in knav_dma_remove()
793 if (atomic_dec_return(&dma->ref_count) == 0) in knav_dma_remove()
794 knav_dma_hw_destroy(dma); in knav_dma_remove()