• Home
  • Raw
  • Download

Lines Matching refs:orig

369 	struct ring_desc *orig;  member
1028 if (np->rx_ring.orig) in free_rings()
1030 np->rx_ring.orig, np->ring_addr); in free_rings()
1812 less_rx = np->get_rx.orig; in nv_alloc_rx()
1813 if (less_rx-- == np->first_rx.orig) in nv_alloc_rx()
1814 less_rx = np->last_rx.orig; in nv_alloc_rx()
1816 while (np->put_rx.orig != less_rx) { in nv_alloc_rx()
1830 np->put_rx.orig->buf = cpu_to_le32(np->put_rx_ctx->dma); in nv_alloc_rx()
1832 np->put_rx.orig->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL); in nv_alloc_rx()
1833 if (unlikely(np->put_rx.orig++ == np->last_rx.orig)) in nv_alloc_rx()
1834 np->put_rx.orig = np->first_rx.orig; in nv_alloc_rx()
1908 np->last_rx.orig = &np->rx_ring.orig[np->rx_ring_size-1]; in nv_init_rx()
1916 np->rx_ring.orig[i].flaglen = 0; in nv_init_rx()
1917 np->rx_ring.orig[i].buf = 0; in nv_init_rx()
1937 np->last_tx.orig = &np->tx_ring.orig[np->tx_ring_size-1]; in nv_init_tx()
1950 np->tx_ring.orig[i].flaglen = 0; in nv_init_tx()
1951 np->tx_ring.orig[i].buf = 0; in nv_init_tx()
2013 np->tx_ring.orig[i].flaglen = 0; in nv_drain_tx()
2014 np->tx_ring.orig[i].buf = 0; in nv_drain_tx()
2044 np->rx_ring.orig[i].flaglen = 0; in nv_drain_rx()
2045 np->rx_ring.orig[i].buf = 0; in nv_drain_rx()
2224 start_tx = put_tx = np->put_tx.orig; in nv_start_xmit()
2250 if (unlikely(put_tx++ == np->last_tx.orig)) in nv_start_xmit()
2251 put_tx = np->first_tx.orig; in nv_start_xmit()
2297 if (unlikely(put_tx++ == np->last_tx.orig)) in nv_start_xmit()
2298 put_tx = np->first_tx.orig; in nv_start_xmit()
2325 np->put_tx.orig = put_tx; in nv_start_xmit()
2538 struct ring_desc *orig_get_tx = np->get_tx.orig; in nv_tx_done()
2541 while ((np->get_tx.orig != np->put_tx.orig) && in nv_tx_done()
2542 !((flags = le32_to_cpu(np->get_tx.orig->flaglen)) & NV_TX_VALID) && in nv_tx_done()
2582 if (unlikely(np->get_tx.orig++ == np->last_tx.orig)) in nv_tx_done()
2583 np->get_tx.orig = np->first_tx.orig; in nv_tx_done()
2590 if (unlikely((np->tx_stop == 1) && (np->get_tx.orig != orig_get_tx))) { in nv_tx_done()
2692 le32_to_cpu(np->tx_ring.orig[i].buf), in nv_tx_timeout()
2693 le32_to_cpu(np->tx_ring.orig[i].flaglen), in nv_tx_timeout()
2694 le32_to_cpu(np->tx_ring.orig[i+1].buf), in nv_tx_timeout()
2695 le32_to_cpu(np->tx_ring.orig[i+1].flaglen), in nv_tx_timeout()
2696 le32_to_cpu(np->tx_ring.orig[i+2].buf), in nv_tx_timeout()
2697 le32_to_cpu(np->tx_ring.orig[i+2].flaglen), in nv_tx_timeout()
2698 le32_to_cpu(np->tx_ring.orig[i+3].buf), in nv_tx_timeout()
2699 le32_to_cpu(np->tx_ring.orig[i+3].flaglen)); in nv_tx_timeout()
2808 while ((np->get_rx.orig != np->put_rx.orig) && in nv_rx_process()
2809 !((flags = le32_to_cpu(np->get_rx.orig->flaglen)) & NV_RX_AVAIL) && in nv_rx_process()
2894 if (unlikely(np->get_rx.orig++ == np->last_rx.orig)) in nv_rx_process()
2895 np->get_rx.orig = np->first_rx.orig; in nv_rx_process()
4638 np->rx_ring.orig = (struct ring_desc *)rxtx_ring; in nv_set_ringparam()
4639 np->tx_ring.orig = &np->rx_ring.orig[np->rx_ring_size]; in nv_set_ringparam()
5083 np->tx_ring.orig[0].buf = cpu_to_le32(test_dma_addr); in nv_loopback_test()
5084 np->tx_ring.orig[0].flaglen = cpu_to_le32((pkt_len-1) | np->tx_flags | tx_flags_extra); in nv_loopback_test()
5097 flags = le32_to_cpu(np->rx_ring.orig[0].flaglen); in nv_loopback_test()
5098 len = nv_descr_getlength(&np->rx_ring.orig[0], np->desc_ver); in nv_loopback_test()
5733 np->rx_ring.orig = pci_alloc_consistent(pci_dev, in nv_probe()
5736 if (!np->rx_ring.orig) in nv_probe()
5738 np->tx_ring.orig = &np->rx_ring.orig[np->rx_ring_size]; in nv_probe()