/drivers/net/ethernet/marvell/octeon_ep/ |
D | octep_rx.c | 34 struct octep_oq_desc_hw *desc_ring = oq->desc_ring; in octep_oq_fill_ring_buffers() local 44 desc_ring[i].buffer_ptr = dma_map_page(oq->dev, page, 0, in octep_oq_fill_ring_buffers() 47 if (dma_mapping_error(oq->dev, desc_ring[i].buffer_ptr)) { in octep_oq_fill_ring_buffers() 63 dma_unmap_page(oq->dev, desc_ring[i].buffer_ptr, PAGE_SIZE, DMA_FROM_DEVICE); in octep_oq_fill_ring_buffers() 81 struct octep_oq_desc_hw *desc_ring = oq->desc_ring; in octep_oq_refill() local 94 desc_ring[refill_idx].buffer_ptr = dma_map_page(oq->dev, page, 0, in octep_oq_refill() 96 if (dma_mapping_error(oq->dev, desc_ring[refill_idx].buffer_ptr)) { in octep_oq_refill() 152 oq->desc_ring = dma_alloc_coherent(oq->dev, desc_ring_size, in octep_setup_oq() 155 if (unlikely(!oq->desc_ring)) { in octep_setup_oq() 182 oq->desc_ring, oq->desc_ring_dma); in octep_setup_oq() [all …]
|
D | octep_tx.c | 200 iq->desc_ring = dma_alloc_coherent(iq->dev, desc_ring_size, in octep_setup_iq() 202 if (unlikely(!iq->desc_ring)) { in octep_setup_iq() 250 iq->desc_ring, iq->desc_ring_dma); in octep_setup_iq() 275 if (iq->desc_ring) in octep_free_iq() 277 iq->desc_ring, iq->desc_ring_dma); in octep_free_iq()
|
D | octep_rx.h | 192 struct octep_oq_desc_hw *desc_ring; member
|
D | octep_tx.h | 178 struct octep_tx_desc_hw *desc_ring; member
|
D | octep_main.c | 656 hw_desc = &iq->desc_ring[wi]; in octep_start_xmit()
|
/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_droq.c | 173 struct octeon_droq_desc *desc_ring = droq->desc_ring; in octeon_droq_setup_ring_buffers() local 187 desc_ring[i].info_ptr = 0; in octeon_droq_setup_ring_buffers() 188 desc_ring[i].buffer_ptr = in octeon_droq_setup_ring_buffers() 208 if (droq->desc_ring) in octeon_delete_droq() 210 droq->desc_ring, droq->desc_ring_dma); in octeon_delete_droq() 271 droq->desc_ring = lio_dma_alloc(oct, desc_ring_size, in octeon_init_droq() 274 if (!droq->desc_ring) { in octeon_init_droq() 281 q_no, droq->desc_ring, droq->desc_ring_dma); in octeon_init_droq() 392 struct octeon_droq_desc *desc_ring) in octeon_droq_refill_pullup_descs() argument 404 desc_ring[droq->refill_idx].buffer_ptr = in octeon_droq_refill_pullup_descs() [all …]
|
D | octeon_droq.h | 257 struct octeon_droq_desc *desc_ring; member
|
/drivers/net/ethernet/google/gve/ |
D | gve_rx_dqo.c | 233 if (rx->dqo.bufq.desc_ring) { in gve_rx_free_ring_dqo() 234 size = sizeof(rx->dqo.bufq.desc_ring[0]) * buffer_queue_slots; in gve_rx_free_ring_dqo() 235 dma_free_coherent(hdev, size, rx->dqo.bufq.desc_ring, in gve_rx_free_ring_dqo() 237 rx->dqo.bufq.desc_ring = NULL; in gve_rx_free_ring_dqo() 240 if (rx->dqo.complq.desc_ring) { in gve_rx_free_ring_dqo() 241 size = sizeof(rx->dqo.complq.desc_ring[0]) * in gve_rx_free_ring_dqo() 243 dma_free_coherent(hdev, size, rx->dqo.complq.desc_ring, in gve_rx_free_ring_dqo() 245 rx->dqo.complq.desc_ring = NULL; in gve_rx_free_ring_dqo() 297 size = sizeof(rx->dqo.complq.desc_ring[0]) * in gve_rx_alloc_ring_dqo() 299 rx->dqo.complq.desc_ring = in gve_rx_alloc_ring_dqo() [all …]
|
D | gve_rx.c | 62 dma_free_coherent(dev, bytes, rx->desc.desc_ring, rx->desc.bus); in gve_rx_free_ring() 63 rx->desc.desc_ring = NULL; in gve_rx_free_ring() 279 rx->desc.desc_ring = dma_alloc_coherent(hdev, bytes, &rx->desc.bus, in gve_rx_alloc_ring() 281 if (!rx->desc.desc_ring) { in gve_rx_alloc_ring() 857 desc = rx->desc.desc_ring + next_idx; in gve_rx_work_pending() 932 struct gve_rx_desc *desc = &rx->desc.desc_ring[idx]; in gve_clean_rx_done() 937 next_desc = &rx->desc.desc_ring[(idx + 1) & rx->mask]; in gve_clean_rx_done() 944 desc = &rx->desc.desc_ring[idx]; in gve_clean_rx_done()
|
D | gve.h | 77 struct gve_rx_desc *desc_ring; /* the descriptor ring */ member 117 struct gve_rx_desc_dqo *desc_ring; member 126 struct gve_rx_compl_desc_dqo *desc_ring; member
|
D | gve_adminq.c | 706 if (priv->rx_desc_cnt * sizeof(priv->rx->desc.desc_ring[0]) in gve_set_desc_cnt()
|
/drivers/dma/ |
D | plx_dma.c | 116 struct plx_dma_desc **desc_ring; member 131 return plxdev->desc_ring[i & (PLX_DMA_RING_COUNT - 1)]; in plx_dma_get_desc() 381 plxdev->desc_ring = kcalloc(PLX_DMA_RING_COUNT, in plx_dma_alloc_desc() 382 sizeof(*plxdev->desc_ring), GFP_KERNEL); in plx_dma_alloc_desc() 383 if (!plxdev->desc_ring) in plx_dma_alloc_desc() 395 plxdev->desc_ring[i] = desc; in plx_dma_alloc_desc() 402 kfree(plxdev->desc_ring[i]); in plx_dma_alloc_desc() 403 kfree(plxdev->desc_ring); in plx_dma_alloc_desc() 480 kfree(plxdev->desc_ring[i]); in plx_dma_free_chan_resources() 482 kfree(plxdev->desc_ring); in plx_dma_free_chan_resources()
|
/drivers/net/ethernet/ti/ |
D | cpmac.c | 192 struct cpmac_desc *desc_ring; member 552 desc = &priv->desc_ring[queue]; in cpmac_start_xmit() 585 desc = &priv->desc_ring[queue]; in cpmac_end_xmit() 704 if (unlikely(!priv->desc_ring)) in cpmac_clear_tx() 707 priv->desc_ring[i].dataflags = 0; in cpmac_clear_tx() 708 if (priv->desc_ring[i].skb) { in cpmac_clear_tx() 709 dev_kfree_skb_any(priv->desc_ring[i].skb); in cpmac_clear_tx() 710 priv->desc_ring[i].skb = NULL; in cpmac_clear_tx() 930 priv->desc_ring = dma_alloc_coherent(&dev->dev, in cpmac_open() 934 if (!priv->desc_ring) { in cpmac_open() [all …]
|
/drivers/net/ethernet/socionext/ |
D | netsec.c | 292 struct netsec_desc_ring desc_ring[NETSEC_RING_MAX]; member 637 struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_TX]; in netsec_clean_tx_dring() 739 struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; in netsec_alloc_rx_data() 761 struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; in netsec_rx_fill() 832 struct netsec_desc_ring *tx_ring = &priv->desc_ring[NETSEC_RING_TX]; in netsec_xdp_queue_one() 861 &priv->desc_ring[NETSEC_RING_RX]; in netsec_xdp_queue_one() 884 struct netsec_desc_ring *tx_ring = &priv->desc_ring[NETSEC_RING_TX]; in netsec_xdp_xmit_back() 901 struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; in netsec_run_xdp() 953 struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; in netsec_process_rx() 1115 struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_TX]; in netsec_check_stop_tx() [all …]
|
/drivers/net/ethernet/cortina/ |
D | gemini.c | 552 struct gmac_txdesc *desc_ring; in gmac_setup_txqs() local 565 desc_ring = dma_alloc_coherent(geth->dev, len * sizeof(*desc_ring), in gmac_setup_txqs() 568 if (!desc_ring) { in gmac_setup_txqs() 575 dma_free_coherent(geth->dev, len * sizeof(*desc_ring), in gmac_setup_txqs() 576 desc_ring, port->txq_dma_base); in gmac_setup_txqs() 585 txq->ring = desc_ring; in gmac_setup_txqs() 596 desc_ring += entries; in gmac_setup_txqs()
|
/drivers/net/ethernet/rdc/ |
D | r6040.c | 294 static void r6040_init_ring_desc(struct r6040_descriptor *desc_ring, in r6040_init_ring_desc() argument 297 struct r6040_descriptor *desc = desc_ring; in r6040_init_ring_desc() 308 desc->vndescp = desc_ring; in r6040_init_ring_desc()
|