/drivers/net/ |
D | sis190.c | 280 u32 rx_buf_sz; member 464 static inline void sis190_give_to_asic(struct RxDesc *desc, u32 rx_buf_sz) in sis190_give_to_asic() argument 469 desc->size = cpu_to_le32((rx_buf_sz & RX_BUF_MASK) | eor); in sis190_give_to_asic() 475 u32 rx_buf_sz) in sis190_map_to_asic() argument 478 sis190_give_to_asic(desc, rx_buf_sz); in sis190_map_to_asic() 493 u32 rx_buf_sz = tp->rx_buf_sz; in sis190_alloc_rx_skb() local 496 skb = netdev_alloc_skb(tp->dev, rx_buf_sz); in sis190_alloc_rx_skb() 500 mapping = pci_map_single(tp->pci_dev, skb->data, tp->rx_buf_sz, in sis190_alloc_rx_skb() 502 sis190_map_to_asic(desc, mapping, rx_buf_sz); in sis190_alloc_rx_skb() 596 sis190_give_to_asic(desc, tp->rx_buf_sz); in sis190_rx_interrupt() [all …]
|
D | hamachi.c | 503 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 1143 hmp->rx_buf_sz, PCI_DMA_FROMDEVICE); in hamachi_tx_timeout() 1150 struct sk_buff *skb = netdev_alloc_skb(dev, hmp->rx_buf_sz); in hamachi_tx_timeout() 1157 skb->data, hmp->rx_buf_sz, PCI_DMA_FROMDEVICE)); in hamachi_tx_timeout() 1159 DescEndPacket | DescIntr | (hmp->rx_buf_sz - 2)); in hamachi_tx_timeout() 1193 hmp->rx_buf_sz = (dev->mtu <= 1492 ? PKT_BUF_SZ : in hamachi_init_ring() 1203 struct sk_buff *skb = dev_alloc_skb(hmp->rx_buf_sz); in hamachi_init_ring() 1210 skb->data, hmp->rx_buf_sz, PCI_DMA_FROMDEVICE)); in hamachi_init_ring() 1213 DescEndPacket | DescIntr | (hmp->rx_buf_sz -2)); in hamachi_init_ring() 1507 hmp->rx_buf_sz, in hamachi_rx() [all …]
|
D | yellowfin.c | 325 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 734 yp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in yellowfin_init_ring() 738 cpu_to_le32(CMD_RX_BUF | INTR_ALWAYS | yp->rx_buf_sz); in yellowfin_init_ring() 744 struct sk_buff *skb = dev_alloc_skb(yp->rx_buf_sz); in yellowfin_init_ring() 751 skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE)); in yellowfin_init_ring() 1065 yp->rx_buf_sz, PCI_DMA_FROMDEVICE); in yellowfin_rx() 1130 yp->rx_buf_sz, in yellowfin_rx() 1142 yp->rx_buf_sz, in yellowfin_rx() 1157 struct sk_buff *skb = dev_alloc_skb(yp->rx_buf_sz); in yellowfin_rx() 1164 skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE)); in yellowfin_rx() [all …]
|
D | dl2k.c | 507 skb = netdev_alloc_skb (dev, np->rx_buf_sz); in rio_timer() 520 (np->pdev, skb->data, np->rx_buf_sz, in rio_timer() 524 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer() 554 np->rx_buf_sz = (dev->mtu <= 1500 ? PACKET_SIZE : dev->mtu + 32); in alloc_list() 578 struct sk_buff *skb = netdev_alloc_skb (dev, np->rx_buf_sz); in alloc_list() 590 np->pdev, skb->data, np->rx_buf_sz, in alloc_list() 592 np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); in alloc_list() 871 np->rx_buf_sz, in receive_packet() 878 np->rx_buf_sz, in receive_packet() 888 np->rx_buf_sz, in receive_packet() [all …]
|
D | sundance.c | 379 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 996 np->rx_buf_sz = (dev->mtu <= 1520 ? PKT_BUF_SZ : dev->mtu + 16); in init_ring() 1009 struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz); in init_ring() 1016 pci_map_single(np->pci_dev, skb->data, np->rx_buf_sz, in init_ring() 1018 np->rx_ring[i].frag[0].length = cpu_to_le32(np->rx_buf_sz | LastFrag); in init_ring() 1340 np->rx_buf_sz, in rx_poll() 1346 np->rx_buf_sz, in rx_poll() 1352 np->rx_buf_sz, in rx_poll() 1394 skb = dev_alloc_skb(np->rx_buf_sz); in refill_rx() 1402 np->rx_buf_sz, PCI_DMA_FROMDEVICE)); in refill_rx() [all …]
|
D | epic100.c | 276 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 927 ep->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in epic_init_ring() 932 ep->rx_ring[i].buflength = ep->rx_buf_sz; in epic_init_ring() 942 struct sk_buff *skb = dev_alloc_skb(ep->rx_buf_sz); in epic_init_ring() 948 skb->data, ep->rx_buf_sz, PCI_DMA_FROMDEVICE); in epic_init_ring() 1213 ep->rx_buf_sz, in epic_rx() 1219 ep->rx_buf_sz, in epic_rx() 1224 ep->rx_buf_sz, PCI_DMA_FROMDEVICE); in epic_rx() 1242 skb = ep->rx_skbuff[entry] = dev_alloc_skb(ep->rx_buf_sz); in epic_rx() 1247 skb->data, ep->rx_buf_sz, PCI_DMA_FROMDEVICE); in epic_rx() [all …]
|
D | via-rhine.c | 398 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 907 rp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in alloc_rbufs() 914 rp->rx_ring[i].desc_length = cpu_to_le32(rp->rx_buf_sz); in alloc_rbufs() 924 struct sk_buff *skb = netdev_alloc_skb(dev, rp->rx_buf_sz); in alloc_rbufs() 931 pci_map_single(rp->pdev, skb->data, rp->rx_buf_sz, in alloc_rbufs() 952 rp->rx_buf_sz, PCI_DMA_FROMDEVICE); in free_rbufs() 1487 rp->rx_buf_sz, in rhine_rx() 1496 rp->rx_buf_sz, in rhine_rx() 1510 rp->rx_buf_sz, in rhine_rx() 1527 skb = netdev_alloc_skb(dev, rp->rx_buf_sz); in rhine_rx() [all …]
|
D | fealnx.c | 410 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 1075 skb = dev_alloc_skb(np->rx_buf_sz); in allocate_rx_buffers() 1085 np->rx_buf_sz, PCI_DMA_FROMDEVICE); in allocate_rx_buffers() 1246 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in init_ring() 1254 np->rx_ring[i].control = np->rx_buf_sz << RBSShift; in init_ring() 1267 struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz); in init_ring() 1278 np->rx_buf_sz, PCI_DMA_FROMDEVICE); in init_ring() 1707 np->rx_buf_sz, in netdev_rx() 1721 np->rx_buf_sz, in netdev_rx() 1726 np->rx_buf_sz, in netdev_rx() [all …]
|
D | 8139cp.c | 354 unsigned rx_buf_sz; member 448 cp->rx_buf_sz = mtu + ETH_HLEN + 8; in cp_set_rxbufsize() 450 cp->rx_buf_sz = PKT_BUF_SZ; in cp_set_rxbufsize() 553 buflen = cp->rx_buf_sz + NET_IP_ALIGN; in cp_rx_poll() 585 cp->rx_buf_sz); in cp_rx_poll() 587 desc->opts1 = cpu_to_le32(DescOwn | cp->rx_buf_sz); in cp_rx_poll() 1061 skb = netdev_alloc_skb(dev, cp->rx_buf_sz + NET_IP_ALIGN); in cp_refill_rx() 1068 cp->rx_buf_sz, PCI_DMA_FROMDEVICE); in cp_refill_rx() 1075 cpu_to_le32(DescOwn | RingEnd | cp->rx_buf_sz); in cp_refill_rx() 1078 cpu_to_le32(DescOwn | cp->rx_buf_sz); in cp_refill_rx() [all …]
|
D | r8169.c | 476 unsigned rx_buf_sz; member 2228 tp->rx_buf_sz = (mtu > RX_BUF_SIZE) ? mtu + ETH_HLEN + 8 : RX_BUF_SIZE; in rtl8169_set_rxbufsize() 2916 pci_unmap_single(pdev, le64_to_cpu(desc->addr), tp->rx_buf_sz, in rtl8169_free_rx_skb() 2923 static inline void rtl8169_mark_to_asic(struct RxDesc *desc, u32 rx_buf_sz) in rtl8169_mark_to_asic() argument 2927 desc->opts1 = cpu_to_le32(DescOwn | eor | rx_buf_sz); in rtl8169_mark_to_asic() 2931 u32 rx_buf_sz) in rtl8169_map_to_asic() argument 2935 rtl8169_mark_to_asic(desc, rx_buf_sz); in rtl8169_map_to_asic() 2940 struct RxDesc *desc, int rx_buf_sz, in rtl8169_alloc_rx_skb() argument 2949 skb = netdev_alloc_skb(dev, rx_buf_sz + pad); in rtl8169_alloc_rx_skb() 2955 mapping = pci_map_single(pdev, skb->data, rx_buf_sz, in rtl8169_alloc_rx_skb() [all …]
|
D | ipg.c | 759 sp->rx_buf_sz, PCI_DMA_FROMDEVICE)); in ipg_get_rxbuff() 782 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in init_rfdlist() 1116 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in ipg_nic_rx_free_skb() 1183 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in ipg_nic_rx_check_error() 1249 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in ipg_nic_rx_with_start() 1466 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in ipg_nic_rx() 1533 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in ipg_nic_rx() 1713 sp->rx_buf_sz, PCI_DMA_FROMDEVICE); in ipg_rx_clear() 1746 sp->rx_buf_sz = sp->rxsupport_size; in ipg_nic_open()
|
D | starfire.c | 578 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 969 writel((np->rx_buf_sz << RxBufferLenShift) | in netdev_open() 1188 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in init_ring() 1192 struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz); in init_ring() 1196 …np->rx_info[i].mapping = pci_map_single(np->pci_dev, skb->data, np->rx_buf_sz, PCI_DMA_FROMDEVICE); in init_ring() 1497 pci_unmap_single(np->pci_dev, np->rx_info[entry].mapping, np->rx_buf_sz, PCI_DMA_FROMDEVICE); in __netdev_rx() 1614 skb = dev_alloc_skb(np->rx_buf_sz); in refill_rx_ring() 1619 pci_map_single(np->pci_dev, skb->data, np->rx_buf_sz, PCI_DMA_FROMDEVICE); in refill_rx_ring() 1984 pci_unmap_single(np->pci_dev, np->rx_info[i].mapping, np->rx_buf_sz, PCI_DMA_FROMDEVICE); in netdev_close()
|
D | sungem.h | 938 #define RX_BUF_ALLOC_SIZE(gp) ((gp)->rx_buf_sz + 28 + RX_OFFSET + 64) 1003 int rx_buf_sz; member
|
D | sh_eth.c | 249 skb = dev_alloc_skb(mdp->rx_buf_sz); in sh_eth_ring_format() 268 rxdesc->buffer_length = (mdp->rx_buf_sz + 16) & ~0x0F; in sh_eth_ring_format() 327 mdp->rx_buf_sz = (ndev->mtu <= 1492 ? PKT_BUF_SZ : in sh_eth_ring_init() 555 rxdesc->buffer_length = (mdp->rx_buf_sz + 16) & ~0x0F; in sh_eth_rx() 558 skb = dev_alloc_skb(mdp->rx_buf_sz); in sh_eth_rx()
|
D | natsemi.c | 561 unsigned int rx_buf_sz; member 1747 if (np->rx_buf_sz > NATSEMI_LONGPKT) in init_registers() 1934 unsigned int buflen = np->rx_buf_sz+NATSEMI_PADDING; in refill_rx() 1944 np->rx_ring[entry].cmd_status = cpu_to_le32(np->rx_buf_sz); in refill_rx() 1957 np->rx_buf_sz = ETH_DATA_LEN + NATSEMI_HEADERS; in set_bufsize() 1959 np->rx_buf_sz = dev->mtu + NATSEMI_HEADERS; in set_bufsize() 2021 unsigned int buflen = np->rx_buf_sz; in drain_rx() 2276 unsigned int buflen = np->rx_buf_sz; in netdev_rx() 2336 } else if (pkt_len > np->rx_buf_sz) { in netdev_rx()
|
D | ipg.h | 715 unsigned int rx_buf_sz; member
|
D | dl2k.h | 503 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
|
D | forcedeth.c | 773 unsigned int rx_buf_sz; member 1700 struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz + NV_RX_ALLOC_PAD); in nv_alloc_rx() 1710 np->put_rx.orig->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL); in nv_alloc_rx() 1732 struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz + NV_RX_ALLOC_PAD); in nv_alloc_rx_optimized() 1743 np->put_rx.ex->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX2_AVAIL); in nv_alloc_rx_optimized() 2874 np->rx_buf_sz = ETH_DATA_LEN + NV_RX_HEADERS; in set_bufsize() 2876 np->rx_buf_sz = dev->mtu + NV_RX_HEADERS; in set_bufsize() 2925 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_change_mtu() 4072 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_do_nic_poll() 4602 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_set_ringparam() [all …]
|
D | sh_eth.h | 648 u32 rx_buf_sz; /* Based on MTU+slack. */ member
|
D | sungem.c | 829 skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_rx() 1681 gp->rx_buf_sz = max(dev->mtu + ETH_HLEN + VLAN_HLEN, in gem_init_rings() 1697 skb_put(skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_init_rings() 1889 writel(0x20000000 | (gp->rx_buf_sz + 4), gp->regs + MAC_MAXFSZ); in gem_init_mac() 1971 int max_frame = (gp->rx_buf_sz + 4 + 64) & ~63; in gem_init_pause_thresholds()
|
/drivers/net/tulip/ |
D | de2104x.c | 296 unsigned rx_buf_sz; member 428 buflen = copying_skb ? (len + RX_OFFSET) : de->rx_buf_sz; in de_rx() 469 cpu_to_le32(RingEnd | de->rx_buf_sz); in de_rx() 471 de->rx_ring[rx_tail].opts2 = cpu_to_le32(de->rx_buf_sz); in de_rx() 1260 skb = dev_alloc_skb(de->rx_buf_sz); in de_refill_rx() 1267 skb->data, de->rx_buf_sz, PCI_DMA_FROMDEVICE); in de_refill_rx() 1273 cpu_to_le32(RingEnd | de->rx_buf_sz); in de_refill_rx() 1275 de->rx_ring[i].opts2 = cpu_to_le32(de->rx_buf_sz); in de_refill_rx() 1321 de->rx_buf_sz, PCI_DMA_FROMDEVICE); in de_clean_rings() 1364 de->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in de_open()
|
D | winbond-840.c | 311 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 806 np->rx_ring[i].length = np->rx_buf_sz; in init_rxtx_rings() 815 struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz); in init_rxtx_rings() 820 np->rx_buf_sz,PCI_DMA_FROMDEVICE); in init_rxtx_rings() 978 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in alloc_ringdesc() 1273 skb = dev_alloc_skb(np->rx_buf_sz); in netdev_rx() 1279 np->rx_buf_sz, PCI_DMA_FROMDEVICE); in netdev_rx()
|
/drivers/atm/ |
D | iphase.c | 767 iadev->rx_buf_sz = IA_RX_BUF_SZ; 778 iadev->rx_buf_sz = IA_RX_BUF_SZ; 790 iadev->rx_buf_sz = IA_RX_BUF_SZ; 795 iadev->rx_buf_sz, iadev->rx_pkt_ram);) 1149 if (len > iadev->rx_buf_sz) { 1150 printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz); 1310 if ((length > iadev->rx_buf_sz) || (length > 1457 writew(iadev->rx_buf_sz, iadev->reass_reg+BUF_SIZE); 1471 rx_pkt_start += iadev->rx_buf_sz; 3133 iadev->num_rx_desc, iadev->rx_buf_sz,
|
D | iphase.h | 1027 u16 num_rx_desc, rx_buf_sz, rxing; member
|