/drivers/net/wireless/ipw2x00/ |
D | libipw_tx.c | 203 txb->frag_size = txb_size; in libipw_alloc_txb() 260 int i, bytes_per_frag, nr_frags, bytes_last_frag, frag_size, in libipw_xmit() local 385 frag_size = MAX_FRAG_THRESHOLD; in libipw_xmit() 387 frag_size = ieee->fts; in libipw_xmit() 393 bytes_per_frag = frag_size - hdr_len; in libipw_xmit() 415 frag_size = bytes + hdr_len; in libipw_xmit() 418 rts_required = (frag_size > ieee->rts in libipw_xmit() 426 txb = libipw_alloc_txb(nr_frags, frag_size, in libipw_xmit() 435 txb->payload_size = frag_size * (nr_frags - 1) + in libipw_xmit()
|
D | libipw.h | 495 u16 frag_size; member
|
D | ipw2100.c | 3143 packet->info.d_struct.txb->frag_size - in ipw2100_tx_send_data()
|
/drivers/infiniband/hw/qib/ |
D | qib_user_sdma.c | 71 u16 frag_size; /* frag size used by PSM */ member 355 if ((pkt->payload_size + newlen) >= pkt->frag_size) { in qib_user_sdma_page_to_frags() 356 newlen = pkt->frag_size - pkt->payload_size; in qib_user_sdma_page_to_frags() 768 if (pkt->frag_size == pkt->bytes_togo && in qib_user_sdma_init_payload() 822 u16 frag_size; in qib_user_sdma_queue_pkts() local 902 frag_size = ((le32_to_cpu(*pbc))>>16) & 0xFFFF; in qib_user_sdma_queue_pkts() 903 if (((frag_size ? frag_size : bytes_togo) + len) > in qib_user_sdma_queue_pkts() 909 if (frag_size) { in qib_user_sdma_queue_pkts() 912 n = npages*((2*PAGE_SIZE/frag_size)+1); in qib_user_sdma_queue_pkts() 934 pkt->frag_size = frag_size; in qib_user_sdma_queue_pkts() [all …]
|
/drivers/staging/rtl8192u/ieee80211/ |
D | ieee80211_tx.c | 254 txb->frag_size = txb_size; in ieee80211_alloc_txb() 604 int i, bytes_per_frag, nr_frags, bytes_last_frag, frag_size; in ieee80211_xmit() local 705 frag_size = MAX_FRAG_THRESHOLD; in ieee80211_xmit() 709 frag_size = ieee->fts;//default:392 in ieee80211_xmit() 728 bytes_per_frag = frag_size - hdr_len; in ieee80211_xmit() 750 txb = ieee80211_alloc_txb(nr_frags, frag_size + ieee->tx_headroom, GFP_ATOMIC); in ieee80211_xmit()
|
D | ieee80211.h | 1185 __le16 frag_size; member
|
/drivers/staging/rtl8192e/ |
D | rtllib_tx.c | 235 txb->frag_size = cpu_to_le16(txb_size); in rtllib_alloc_txb() 558 int i, bytes_per_frag, nr_frags, bytes_last_frag, frag_size; in rtllib_xmit_inter() local 708 frag_size = MAX_FRAG_THRESHOLD; in rtllib_xmit_inter() 711 frag_size = ieee->fts; in rtllib_xmit_inter() 735 bytes_per_frag = frag_size - hdr_len; in rtllib_xmit_inter() 760 txb = rtllib_alloc_txb(nr_frags, frag_size + in rtllib_xmit_inter()
|
D | rtllib.h | 1298 __le16 frag_size; member
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 65 ((PAGE_SIZE << order) < frag_info->frag_size)) in mlx4_alloc_pages() 214 cpu_to_be32(priv->frag_info[i].frag_size); in mlx4_en_init_rx_desc() 566 dma_sync_single_for_cpu(priv->ddev, dma, frag_info->frag_size, in mlx4_en_complete_rx_desc() 571 skb_frag_size_set(&skb_frags_rx[nr], frag_info->frag_size); in mlx4_en_complete_rx_desc() 976 priv->frag_info[i].frag_size = in mlx4_en_calc_rx_buf() 989 buf_size += priv->frag_info[i].frag_size; in mlx4_en_calc_rx_buf() 1003 priv->frag_info[i].frag_size, in mlx4_en_calc_rx_buf()
|
D | mlx4_en.h | 476 u16 frag_size; member
|
/drivers/staging/line6/ |
D | driver.c | 156 int frag_size = min(line6->max_packet_size, size - i); in line6_send_raw_message() local 162 (char *)frag_buf, frag_size, in line6_send_raw_message() 171 done += frag_size; in line6_send_raw_message()
|
/drivers/net/ethernet/marvell/ |
D | mvneta.c | 284 unsigned int frag_size; member 1337 if (likely(pp->frag_size <= PAGE_SIZE)) in mvneta_frag_alloc() 1338 return netdev_alloc_frag(pp->frag_size); in mvneta_frag_alloc() 1340 return kmalloc(pp->frag_size, GFP_ATOMIC); in mvneta_frag_alloc() 1345 if (likely(pp->frag_size <= PAGE_SIZE)) in mvneta_frag_free() 1504 skb = build_skb(data, pp->frag_size > PAGE_SIZE ? 0 : pp->frag_size); in mvneta_rx() 2463 pp->frag_size = SKB_DATA_ALIGN(MVNETA_RX_BUF_SIZE(pp->pkt_size)) + in mvneta_change_mtu() 2628 pp->frag_size = SKB_DATA_ALIGN(MVNETA_RX_BUF_SIZE(pp->pkt_size)) + in mvneta_open()
|
/drivers/scsi/bnx2fc/ |
D | bnx2fc_io.c | 1621 int frag_size, sg_frags; in bnx2fc_split_bd() local 1626 frag_size = BNX2FC_BD_SPLIT_SZ; in bnx2fc_split_bd() 1628 frag_size = sg_len; in bnx2fc_split_bd() 1631 bd[bd_index + sg_frags].buf_len = (u16)frag_size; in bnx2fc_split_bd() 1634 addr += (u64) frag_size; in bnx2fc_split_bd() 1636 sg_len -= frag_size; in bnx2fc_split_bd()
|
/drivers/nfc/ |
D | pn533.c | 2577 int frag_size; in pn533_fill_fragment_skbs() local 2582 frag_size = PN533_CMD_DATAFRAME_MAXLEN; in pn533_fill_fragment_skbs() 2584 frag_size = skb->len; in pn533_fill_fragment_skbs() 2587 frag = pn533_alloc_skb(dev, frag_size); in pn533_fill_fragment_skbs() 2598 if (frag_size == PN533_CMD_DATAFRAME_MAXLEN) in pn533_fill_fragment_skbs() 2605 memcpy(skb_put(frag, frag_size), skb->data, frag_size); in pn533_fill_fragment_skbs() 2608 skb_pull(skb, frag_size); in pn533_fill_fragment_skbs()
|
/drivers/net/ethernet/nvidia/ |
D | forcedeth.c | 2206 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit() local 2208 entries += (frag_size >> NV_TX2_TSO_MAX_SHIFT) + in nv_start_xmit() 2209 ((frag_size & (NV_TX2_TSO_MAX_SIZE-1)) ? 1 : 0); in nv_start_xmit() 2257 u32 frag_size = skb_frag_size(frag); in nv_start_xmit() local 2266 bcnt = (frag_size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : frag_size; in nv_start_xmit() 2294 frag_size -= bcnt; in nv_start_xmit() 2299 } while (frag_size); in nv_start_xmit() 2354 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit_optimized() local 2356 entries += (frag_size >> NV_TX2_TSO_MAX_SHIFT) + in nv_start_xmit_optimized() 2357 ((frag_size & (NV_TX2_TSO_MAX_SIZE-1)) ? 1 : 0); in nv_start_xmit_optimized() [all …]
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_cmn.c | 582 u32 i, frag_len, frag_size; in bnx2x_fill_frag_skb() local 587 frag_size = le16_to_cpu(cqe->pkt_len) - len_on_bd; in bnx2x_fill_frag_skb() 595 if (frag_size) in bnx2x_fill_frag_skb() 617 frag_len = min_t(u32, frag_size, (u32)full_page); in bnx2x_fill_frag_skb() 619 frag_len = min_t(u32, frag_size, (u32)SGE_PAGES); in bnx2x_fill_frag_skb() 656 frag_size -= frag_len; in bnx2x_fill_frag_skb() 947 u16 frag_size, pages; in bnx2x_rx_int() local 975 frag_size = le16_to_cpu(cqe->end_agg_cqe.pkt_len) - in bnx2x_rx_int() 979 pages = (frag_size + tpa_info->full_page - 1) / in bnx2x_rx_int() 982 pages = SGE_PAGE_ALIGN(frag_size) >> in bnx2x_rx_int()
|
/drivers/staging/rtl8712/ |
D | ieee80211.h | 649 u16 frag_size; member
|
/drivers/net/ |
D | virtio_net.c | 306 unsigned int frag_size = min((unsigned)PAGE_SIZE - offset, len); in page_to_skb() local 308 frag_size, truesize); in page_to_skb() 309 len -= frag_size; in page_to_skb()
|
/drivers/net/ethernet/cadence/ |
D | macb.c | 1192 unsigned int count, nr_frags, frag_size, f; in macb_start_xmit() local 1210 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in macb_start_xmit() 1211 count += macb_count_tx_descriptors(bp, frag_size); in macb_start_xmit()
|
/drivers/net/ethernet/emulex/benet/ |
D | be_cmds.h | 534 u8 frag_size; member 2044 u16 cq_id, u16 frag_size, u32 if_id, u32 rss, u8 *rss_id);
|
D | be_cmds.c | 1296 struct be_queue_info *rxq, u16 cq_id, u16 frag_size, in be_cmd_rxq_create() argument 1317 req->frag_size = fls(frag_size) - 1; in be_cmd_rxq_create()
|
/drivers/staging/rtl8188eu/include/ |
D | ieee80211.h | 859 u16 frag_size; member
|
/drivers/net/ethernet/broadcom/ |
D | tg3.c | 6656 unsigned int *frag_size) in tg3_alloc_rx_data() argument 6693 *frag_size = skb_size; in tg3_alloc_rx_data() 6696 *frag_size = 0; in tg3_alloc_rx_data() 6865 unsigned int frag_size; in tg3_rx() local 6868 *post_ptr, &frag_size); in tg3_rx() 6882 skb = build_skb(data, frag_size); in tg3_rx() 6884 tg3_frag_free(frag_size != 0, data); in tg3_rx() 8387 unsigned int frag_size; in tg3_rx_prodring_alloc() local 8390 &frag_size) < 0) { in tg3_rx_prodring_alloc() 8422 unsigned int frag_size; in tg3_rx_prodring_alloc() local [all …]
|
D | bnx2.c | 3044 unsigned int i, frag_len, frag_size, pages; in bnx2_rx_skb() local 3049 frag_size = len + 4 - hdr_len; in bnx2_rx_skb() 3050 pages = PAGE_ALIGN(frag_size) >> PAGE_SHIFT; in bnx2_rx_skb() 3056 frag_len = min(frag_size, (unsigned int) PAGE_SIZE); in bnx2_rx_skb() 3101 frag_size -= frag_len; in bnx2_rx_skb()
|