• Home
  • Raw
  • Download

Lines Matching refs:rx_ring

288 struct rx_ring {  struct
492 struct rx_ring rx_ring; member
740 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_enable() local
742 if (rx_ring->fbr[1]->buffsize == 4096) in et131x_rx_dma_enable()
744 else if (rx_ring->fbr[1]->buffsize == 8192) in et131x_rx_dma_enable()
746 else if (rx_ring->fbr[1]->buffsize == 16384) in et131x_rx_dma_enable()
750 if (rx_ring->fbr[0]->buffsize == 256) in et131x_rx_dma_enable()
752 else if (rx_ring->fbr[0]->buffsize == 512) in et131x_rx_dma_enable()
754 else if (rx_ring->fbr[0]->buffsize == 1024) in et131x_rx_dma_enable()
1543 struct rx_ring *rx_local = &adapter->rx_ring; in et131x_config_rx_dma_regs()
1869 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_memory_alloc() local
1873 rx_ring->fbr[0] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1874 if (rx_ring->fbr[0] == NULL) in et131x_rx_dma_memory_alloc()
1876 rx_ring->fbr[1] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1877 if (rx_ring->fbr[1] == NULL) in et131x_rx_dma_memory_alloc()
1898 rx_ring->fbr[0]->buffsize = 256; in et131x_rx_dma_memory_alloc()
1899 rx_ring->fbr[0]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1900 rx_ring->fbr[1]->buffsize = 2048; in et131x_rx_dma_memory_alloc()
1901 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1903 rx_ring->fbr[0]->buffsize = 512; in et131x_rx_dma_memory_alloc()
1904 rx_ring->fbr[0]->num_entries = 1024; in et131x_rx_dma_memory_alloc()
1905 rx_ring->fbr[1]->buffsize = 4096; in et131x_rx_dma_memory_alloc()
1906 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1908 rx_ring->fbr[0]->buffsize = 1024; in et131x_rx_dma_memory_alloc()
1909 rx_ring->fbr[0]->num_entries = 768; in et131x_rx_dma_memory_alloc()
1910 rx_ring->fbr[1]->buffsize = 16384; in et131x_rx_dma_memory_alloc()
1911 rx_ring->fbr[1]->num_entries = 128; in et131x_rx_dma_memory_alloc()
1914 rx_ring->psr_entries = rx_ring->fbr[0]->num_entries + in et131x_rx_dma_memory_alloc()
1915 rx_ring->fbr[1]->num_entries; in et131x_rx_dma_memory_alloc()
1918 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1934 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1974 psr_size = sizeof(struct pkt_stat_desc) * rx_ring->psr_entries; in et131x_rx_dma_memory_alloc()
1976 rx_ring->ps_ring_virtaddr = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1978 &rx_ring->ps_ring_physaddr, in et131x_rx_dma_memory_alloc()
1981 if (!rx_ring->ps_ring_virtaddr) { in et131x_rx_dma_memory_alloc()
1988 rx_ring->rx_status_block = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1990 &rx_ring->rx_status_bus, in et131x_rx_dma_memory_alloc()
1992 if (!rx_ring->rx_status_block) { in et131x_rx_dma_memory_alloc()
1997 rx_ring->num_rfd = NIC_DEFAULT_NUM_RFD; in et131x_rx_dma_memory_alloc()
2002 INIT_LIST_HEAD(&rx_ring->recv_list); in et131x_rx_dma_memory_alloc()
2013 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_memory_free() local
2017 WARN_ON(rx_ring->num_ready_recv != rx_ring->num_rfd); in et131x_rx_dma_memory_free()
2019 while (!list_empty(&rx_ring->recv_list)) { in et131x_rx_dma_memory_free()
2020 rfd = list_entry(rx_ring->recv_list.next, in et131x_rx_dma_memory_free()
2030 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_free()
2060 if (rx_ring->ps_ring_virtaddr) { in et131x_rx_dma_memory_free()
2061 psr_size = sizeof(struct pkt_stat_desc) * rx_ring->psr_entries; in et131x_rx_dma_memory_free()
2064 rx_ring->ps_ring_virtaddr, in et131x_rx_dma_memory_free()
2065 rx_ring->ps_ring_physaddr); in et131x_rx_dma_memory_free()
2067 rx_ring->ps_ring_virtaddr = NULL; in et131x_rx_dma_memory_free()
2071 if (rx_ring->rx_status_block) { in et131x_rx_dma_memory_free()
2074 rx_ring->rx_status_block, in et131x_rx_dma_memory_free()
2075 rx_ring->rx_status_bus); in et131x_rx_dma_memory_free()
2076 rx_ring->rx_status_block = NULL; in et131x_rx_dma_memory_free()
2080 kfree(rx_ring->fbr[0]); in et131x_rx_dma_memory_free()
2081 kfree(rx_ring->fbr[1]); in et131x_rx_dma_memory_free()
2084 rx_ring->num_ready_recv = 0; in et131x_rx_dma_memory_free()
2092 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_init_recv() local
2095 for (rfdct = 0; rfdct < rx_ring->num_rfd; rfdct++) { in et131x_init_recv()
2103 list_add_tail(&rfd->list_node, &rx_ring->recv_list); in et131x_init_recv()
2106 rx_ring->num_ready_recv++; in et131x_init_recv()
2129 struct rx_ring *rx_local = &adapter->rx_ring; in nic_return_rfd()
2188 struct rx_ring *rx_local = &adapter->rx_ring; in nic_rx_pkts()
2302 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_handle_recv_pkts() local
2309 if (list_empty(&rx_ring->recv_list)) { in et131x_handle_recv_pkts()
2310 WARN_ON(rx_ring->num_ready_recv != 0); in et131x_handle_recv_pkts()
2332 if (rx_ring->num_ready_recv < RFD_LOW_WATER_MARK) in et131x_handle_recv_pkts()
2339 rx_ring->unfinished_receives = true; in et131x_handle_recv_pkts()
2344 rx_ring->unfinished_receives = false; in et131x_handle_recv_pkts()
3369 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_isr() local
3403 if (rx_ring->unfinished_receives) in et131x_isr()