Home
last modified time | relevance | path

Searched refs:tx_bufs (Results 1 – 12 of 12) sorted by relevance

/drivers/net/ethernet/sun/
Dsunvnet_common.c1011 if (port->tx_bufs[txi].skb) { in vnet_clean_tx_ring()
1015 BUG_ON(port->tx_bufs[txi].skb->next); in vnet_clean_tx_ring()
1017 port->tx_bufs[txi].skb->next = skb; in vnet_clean_tx_ring()
1018 skb = port->tx_bufs[txi].skb; in vnet_clean_tx_ring()
1019 port->tx_bufs[txi].skb = NULL; in vnet_clean_tx_ring()
1022 port->tx_bufs[txi].cookies, in vnet_clean_tx_ring()
1023 port->tx_bufs[txi].ncookies); in vnet_clean_tx_ring()
1398 BUG_ON(port->tx_bufs[txi].skb); in sunvnet_start_xmit_common()
1404 err = vnet_skb_map(port->vio.lp, skb, port->tx_bufs[txi].cookies, 2, in sunvnet_start_xmit_common()
1411 port->tx_bufs[txi].skb = skb; in sunvnet_start_xmit_common()
[all …]
Dsunvnet_common.h73 struct vnet_tx_entry tx_bufs[VNET_TX_RING_SIZE]; member
/drivers/net/ethernet/aurora/
Dnb8800.c354 txb = &priv->tx_bufs[priv->tx_queue]; in __nb8800_tx_dma_start()
424 txb = &priv->tx_bufs[next]; in nb8800_xmit()
442 desc->n_addr = priv->tx_bufs[next].dma_desc; in nb8800_xmit()
497 struct nb8800_tx_buf *txb = &priv->tx_bufs[done]; in nb8800_tx_done()
760 if (priv->tx_bufs) { in nb8800_dma_free()
762 kfree_skb(priv->tx_bufs[i].skb); in nb8800_dma_free()
764 kfree(priv->tx_bufs); in nb8800_dma_free()
765 priv->tx_bufs = NULL; in nb8800_dma_free()
805 struct nb8800_tx_buf *txb = &priv->tx_bufs[i]; in nb8800_dma_reset()
853 priv->tx_bufs = kcalloc(n_tx, sizeof(*priv->tx_bufs), GFP_KERNEL); in nb8800_dma_init()
[all …]
Dnb8800.h266 struct nb8800_tx_buf *tx_bufs; member
/drivers/net/ethernet/apple/
Dbmac.c73 struct sk_buff *tx_bufs[N_TX_RING]; member
493 if (bp->tx_bufs[i] != NULL) { in bmac_suspend()
494 dev_kfree_skb(bp->tx_bufs[i]); in bmac_suspend()
495 bp->tx_bufs[i] = NULL; in bmac_suspend()
662 bp->tx_bufs[bp->tx_fill] = skb; in bmac_transmit_packet()
779 if (bp->tx_bufs[bp->tx_empty]) { in bmac_txdma_intr()
781 dev_consume_skb_irq(bp->tx_bufs[bp->tx_empty]); in bmac_txdma_intr()
783 bp->tx_bufs[bp->tx_empty] = NULL; in bmac_txdma_intr()
1420 if (bp->tx_bufs[i] != NULL) { in bmac_close()
1421 dev_kfree_skb(bp->tx_bufs[i]); in bmac_close()
[all …]
Dmace.c56 struct sk_buff *tx_bufs[N_TX_RING]; member
422 dev_kfree_skb(mp->tx_bufs[i]); in mace_clean_rings()
557 mp->tx_bufs[fill] = skb; in mace_xmit_start()
765 dev->stats.tx_bytes += mp->tx_bufs[i]->len; in mace_interrupt()
768 dev_consume_skb_irq(mp->tx_bufs[i]); in mace_interrupt()
844 dev_kfree_skb_irq(mp->tx_bufs[i]); in mace_tx_timeout()
/drivers/hsi/clients/
Dcmt_speech.c101 unsigned int tx_bufs; member
788 static void set_buffer_sizes(struct cs_hsi_iface *hi, int rx_bufs, int tx_bufs) in set_buffer_sizes() argument
791 hi->tx_bufs = tx_bufs; in set_buffer_sizes()
793 hi->mmap_cfg->tx_bufs = tx_bufs; in set_buffer_sizes()
814 (buf_cfg->rx_bufs + buf_cfg->tx_bufs); in check_buf_params()
819 buf_cfg->tx_bufs > CS_MAX_BUFFERS) { in check_buf_params()
886 set_buffer_sizes(hi, buf_cfg->rx_bufs, buf_cfg->tx_bufs); in cs_hsi_data_enable()
904 for (i = 0; i < hi->mmap_cfg->tx_bufs; i++) { in cs_hsi_data_enable()
/drivers/net/ethernet/silan/
Dsc92031.c284 void *tx_bufs; member
957 skb_copy_and_csum_dev(skb, priv->tx_bufs + entry * TX_BUF_SIZE); in sc92031_start_xmit()
961 memset(priv->tx_bufs + entry * TX_BUF_SIZE + len, in sc92031_start_xmit()
1004 priv->tx_bufs = dma_alloc_coherent(&pdev->dev, TX_BUF_TOT_LEN, in sc92031_open()
1006 if (unlikely(!priv->tx_bufs)) { in sc92031_open()
1035 dma_free_coherent(&pdev->dev, TX_BUF_TOT_LEN, priv->tx_bufs, in sc92031_open()
1062 dma_free_coherent(&pdev->dev, TX_BUF_TOT_LEN, priv->tx_bufs, in sc92031_stop()
/drivers/net/ethernet/aeroflex/
Dgreth.h105 unsigned char *tx_bufs[GRETH_TXBD_NUM]; member
Dgreth.c223 kfree(greth->tx_bufs[i]); in greth_clean_rings()
297 greth->tx_bufs[i] = kmalloc(MAX_FRAME_SIZE, GFP_KERNEL); in greth_init_rings()
299 if (greth->tx_bufs[i] == NULL) { in greth_init_rings()
306 greth->tx_bufs[i], in greth_init_rings()
/drivers/net/ethernet/realtek/
D8139too.c597 unsigned char *tx_bufs; /* Tx bounce buffer region. */ member
1330 tp->tx_bufs = dma_alloc_coherent(&tp->pci_dev->dev, TX_BUF_TOT_LEN, in rtl8139_open()
1334 if (tp->tx_bufs == NULL || tp->rx_ring == NULL) { in rtl8139_open()
1337 if (tp->tx_bufs) in rtl8139_open()
1339 tp->tx_bufs, tp->tx_bufs_dma); in rtl8139_open()
1427 RTL_W32_F (TxAddr0 + (i * 4), tp->tx_bufs_dma + (tp->tx_buf[i] - tp->tx_bufs)); in rtl8139_hw_start()
1457 tp->tx_buf[i] = &tp->tx_bufs[i * TX_BUF_SIZE]; in rtl8139_init_ring()
2288 tp->tx_bufs, tp->tx_bufs_dma); in rtl8139_close()
2290 tp->tx_bufs = NULL; in rtl8139_close()
/drivers/net/ethernet/via/
Dvia-rhine.c451 unsigned char *tx_bufs; member
1158 rp->tx_bufs = dma_alloc_coherent(hwdev, in alloc_ring()
1162 if (rp->tx_bufs == NULL) { in alloc_ring()
1190 if (rp->tx_bufs) in free_ring()
1192 rp->tx_bufs, rp->tx_bufs_dma); in free_ring()
1194 rp->tx_bufs = NULL; in free_ring()
1319 rp->tx_buf[i] = &rp->tx_bufs[i * PKT_BUF_SZ]; in alloc_tbufs()
1810 rp->tx_bufs)); in rhine_start_tx()