Home
last modified time | relevance | path

Searched refs:dma_rx (Results 1 – 19 of 19) sorted by relevance

/drivers/net/irda/
Dsa1100_ir.c63 struct sa1100_buf dma_rx; member
165 if (si->dma_rx.skb) in sa1100_irda_rx_alloc()
168 si->dma_rx.skb = alloc_skb(HPSIR_MAX_RXLEN + 1, GFP_ATOMIC); in sa1100_irda_rx_alloc()
169 if (!si->dma_rx.skb) { in sa1100_irda_rx_alloc()
178 skb_reserve(si->dma_rx.skb, 1); in sa1100_irda_rx_alloc()
180 sg_set_buf(&si->dma_rx.sg, si->dma_rx.skb->data, HPSIR_MAX_RXLEN); in sa1100_irda_rx_alloc()
181 if (dma_map_sg(si->dma_rx.dev, &si->dma_rx.sg, 1, DMA_FROM_DEVICE) == 0) { in sa1100_irda_rx_alloc()
182 dev_kfree_skb_any(si->dma_rx.skb); in sa1100_irda_rx_alloc()
195 if (!si->dma_rx.skb) { in sa1100_irda_rx_dma_start()
208 dmaengine_terminate_all(si->dma_rx.chan); in sa1100_irda_rx_dma_start()
[all …]
/drivers/spi/
Dspi-omap2-mcspi.c98 struct dma_chan *dma_rx; member
385 if (mcspi_dma->dma_rx) { in omap2_mcspi_rx_dma()
390 dmaengine_slave_config(mcspi_dma->dma_rx, &cfg); in omap2_mcspi_rx_dma()
399 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_rx, &sg, 1, in omap2_mcspi_rx_dma()
411 dma_async_issue_pending(mcspi_dma->dma_rx); in omap2_mcspi_rx_dma()
833 mcspi_dma->dma_rx = dma_request_channel(mask, omap_dma_filter_fn, &sig); in omap2_mcspi_request_dma()
834 if (!mcspi_dma->dma_rx) in omap2_mcspi_request_dma()
840 dma_release_channel(mcspi_dma->dma_rx); in omap2_mcspi_request_dma()
841 mcspi_dma->dma_rx = NULL; in omap2_mcspi_request_dma()
880 if (!mcspi_dma->dma_rx || !mcspi_dma->dma_tx) { in omap2_mcspi_setup()
[all …]
Dspi-ep93xx.c127 struct dma_chan *dma_rx; member
568 chan = espi->dma_rx; in ep93xx_spi_dma_prepare()
654 chan = espi->dma_rx; in ep93xx_spi_dma_finish()
697 dma_async_issue_pending(espi->dma_rx); in ep93xx_spi_dma_transfer()
762 if (espi->dma_rx && t->len > SPI_FIFO_SIZE) in ep93xx_spi_process_transfer()
982 espi->dma_rx = dma_request_channel(mask, ep93xx_spi_dma_filter, in ep93xx_spi_setup_dma()
984 if (!espi->dma_rx) { in ep93xx_spi_setup_dma()
1003 dma_release_channel(espi->dma_rx); in ep93xx_spi_setup_dma()
1004 espi->dma_rx = NULL; in ep93xx_spi_setup_dma()
1013 if (espi->dma_rx) { in ep93xx_spi_release_dma()
[all …]
Dspi-davinci.c135 struct dma_chan *dma_rx; member
590 dmaengine_slave_config(dspi->dma_rx, &dma_rx_conf); in davinci_spi_bufs()
621 rxdesc = dmaengine_prep_slave_sg(dspi->dma_rx, in davinci_spi_bufs()
644 dma_async_issue_pending(dspi->dma_rx); in davinci_spi_bufs()
752 dspi->dma_rx = dma_request_channel(mask, edma_filter_fn, in davinci_spi_request_dma()
754 if (!dspi->dma_rx) { in davinci_spi_request_dma()
771 dma_release_channel(dspi->dma_rx); in davinci_spi_request_dma()
1013 dma_release_channel(dspi->dma_rx); in davinci_spi_probe()
/drivers/net/ethernet/micrel/
Dks8842.c160 ((adapter)->dma_rx.channel != -1))
172 struct ks8842_rx_dma_ctl dma_rx; member
556 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; in __ks8842_start_new_rx_dma()
608 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; in ks8842_rx_frame_dma_tasklet()
849 if (adapter->dma_rx.adesc) in ks8842_dma_rx_cb()
850 tasklet_schedule(&adapter->dma_rx.tasklet); in ks8842_dma_rx_cb()
874 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_stop_dma()
898 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_dealloc_dma_bufs()
930 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_alloc_dma_bufs()
1003 adapter->dma_rx.channel = -1; in ks8842_open()
[all …]
/drivers/net/ethernet/stmicro/stmmac/
Ddwmac100_dma.c36 int burst_len, u32 dma_tx, u32 dma_rx, int atds) in dwmac100_dma_init() argument
64 writel(dma_rx, ioaddr + DMA_RCV_BASE_ADDR); in dwmac100_dma_init()
Ddwmac1000_dma.c34 int burst_len, u32 dma_tx, u32 dma_rx, int atds) in dwmac1000_dma_init() argument
108 writel(dma_rx, ioaddr + DMA_RCV_BASE_ADDR); in dwmac1000_dma_init()
Dstmmac_main.c407 desc = (priv->dma_rx + entry); in stmmac_get_rx_hwtstamp()
913 stmmac_display_ring((void *)priv->dma_rx, rxsize, 0); in stmmac_display_rings()
954 priv->hw->desc->init_rx_desc(&priv->dma_rx[i], in stmmac_clear_descriptors()
1035 priv->dma_rx = dma_alloc_coherent(priv->device, rxsize * in init_dma_desc_rings()
1043 if ((!priv->dma_rx) || (!priv->dma_tx)) in init_dma_desc_rings()
1066 p = priv->dma_rx + i; in init_dma_desc_rings()
1087 priv->hw->chain->init(priv->dma_rx, priv->dma_rx_phy, in init_dma_desc_rings()
1166 priv->dma_rx, priv->dma_rx_phy); in free_dma_desc_resources()
1952 p = priv->dma_rx + entry; in stmmac_rx_refill()
2000 stmmac_display_ring((void *)priv->dma_rx, rxsize, 0); in stmmac_rx()
[all …]
Dstmmac.h55 struct dma_desc *dma_rx ____cacheline_aligned_in_smp;
Dcommon.h356 int burst_len, u32 dma_tx, u32 dma_rx, int atds);
/drivers/net/ethernet/calxeda/
Dxgmac.c369 struct xgmac_dma_desc *dma_rx; member
679 p = priv->dma_rx + entry; in xgmac_rx_refill()
724 priv->dma_rx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init()
729 if (!priv->dma_rx) in xgmac_dma_desc_rings_init()
747 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init()
753 desc_init_rx_desc(priv->dma_rx, DMA_RX_RING_SZ, priv->dma_buf_sz); in xgmac_dma_desc_rings_init()
770 priv->dma_rx, priv->dma_rx_phy); in xgmac_dma_desc_rings_init()
788 p = priv->dma_rx + i; in xgmac_free_rx_skbufs()
836 if (priv->dma_rx) { in xgmac_free_dma_desc_rings()
839 priv->dma_rx, priv->dma_rx_phy); in xgmac_free_dma_desc_rings()
[all …]
/drivers/mmc/host/
Domap.c140 struct dma_chan *dma_rx; member
426 c = host->dma_rx; in mmc_omap_release_dma()
1020 c = host->dma_rx; in mmc_omap_prepare_data()
1089 host->dma_tx : host->dma_rx; in mmc_omap_start_request()
1431 host->dma_rx = dma_request_channel(mask, omap_dma_filter_fn, &sig); in mmc_omap_probe()
1433 if (!host->dma_rx) { in mmc_omap_probe()
1439 if (!host->dma_rx) in mmc_omap_probe()
1483 if (host->dma_rx) in mmc_omap_probe()
1484 dma_release_channel(host->dma_rx); in mmc_omap_probe()
1521 if (host->dma_rx) in mmc_omap_remove()
[all …]
Ddavinci_mmc.c207 struct dma_chan *dma_rx; member
411 sync_dev = host->dma_rx; in davinci_abort_dma()
455 chan = host->dma_rx; in mmc_davinci_send_dma_request()
456 dmaengine_slave_config(host->dma_rx, &dma_rx_conf); in mmc_davinci_send_dma_request()
458 desc = dmaengine_prep_slave_sg(host->dma_rx, in mmc_davinci_send_dma_request()
515 dma_release_channel(host->dma_rx); in davinci_release_dma_channels()
534 host->dma_rx = in davinci_acquire_dma_channels()
537 if (!host->dma_rx) { in davinci_acquire_dma_channels()
/drivers/net/wireless/brcm80211/brcmsmac/
Ddma.h84 int dma_rx(struct dma_pub *pub, struct sk_buff_head *skb_list);
Ddma.c925 int dma_rx(struct dma_pub *pub, struct sk_buff_head *skb_list) in dma_rx() function
Dmain.c7780 morepending = dma_rx(wlc_hw->di[fifo], &recv_frames); in brcms_b_recv()
/drivers/net/wireless/b43legacy/
Ddma.c1309 static void dma_rx(struct b43legacy_dmaring *ring, in dma_rx() function
1420 dma_rx(ring, &slot); in b43legacy_dma_rx()
/drivers/net/ethernet/cirrus/
Dcs89x0.c415 dma_rx(struct net_device *dev) in dma_rx() function
793 dma_rx(dev); in net_interrupt()
/drivers/net/wireless/b43/
Ddma.c1645 static void dma_rx(struct b43_dmaring *ring, int *slot) in dma_rx() function
1762 dma_rx(ring, &slot); in b43_dma_rx()