/drivers/infiniband/hw/nes/ |
D | nes_mgt.c | 158 skb = fpdu_info->frags[i].skb; in nes_download_callback() 159 if (fpdu_info->frags[i].cmplt) { in nes_download_callback() 249 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; in get_fpdu_info() local 276 memset(&frags, 0, sizeof frags); in get_fpdu_info() 278 frags[i].physaddr = cb->busaddr; in get_fpdu_info() 279 frags[i].physaddr += skb->data - cb->data_start; in get_fpdu_info() 280 frags[i].frag_len = min(tmp_len, skb->len); in get_fpdu_info() 281 frags[i].skb = skb; in get_fpdu_info() 282 frags[i].cmplt = (skb->len == frags[i].frag_len); in get_fpdu_info() 283 frag_tot += frags[i].frag_len; in get_fpdu_info() [all …]
|
D | nes_mgt.h | 82 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; member
|
D | nes_nic.c | 440 &skb_shinfo(skb)->frags[skb_fragment_index]; in nes_nic_send() 445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); in nes_nic_send() 560 &skb_shinfo(skb)->frags[tso_frag_count]; in nes_netdev_start_xmit() 635 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index])); in nes_netdev_start_xmit() 639 tso_wqe_length += skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index++]); in nes_netdev_start_xmit()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 47 struct mlx4_en_rx_alloc *frags, in mlx4_en_alloc_frags() argument 82 frags[i] = ring_alloc[i]; in mlx4_en_alloc_frags() 103 struct mlx4_en_rx_alloc *frags, in mlx4_en_free_frag() argument 108 if (frags[i].offset == frag_info->last_offset) { in mlx4_en_free_frag() 109 dma_unmap_page(priv->ddev, frags[i].dma, MLX4_EN_ALLOC_SIZE, in mlx4_en_free_frag() 112 if (frags[i].page) in mlx4_en_free_frag() 113 put_page(frags[i].page); in mlx4_en_free_frag() 200 struct mlx4_en_rx_alloc *frags = ring->rx_info + in mlx4_en_prepare_rx_desc() local 203 return mlx4_en_alloc_frags(priv, rx_desc, frags, ring->page_alloc); in mlx4_en_prepare_rx_desc() 215 struct mlx4_en_rx_alloc *frags; in mlx4_en_free_rx_desc() local [all …]
|
D | en_tx.c | 206 int frags = skb_shinfo(skb)->nr_frags; in mlx4_en_free_tx_desc() local 228 for (i = 0; i < frags; i++) { in mlx4_en_free_tx_desc() 229 frag = &skb_shinfo(skb)->frags[i]; in mlx4_en_free_tx_desc() 255 for (i = 0; i < frags; i++) { in mlx4_en_free_tx_desc() 259 frag = &skb_shinfo(skb)->frags[i]; in mlx4_en_free_tx_desc() 436 ptr = skb_frag_address_safe(&skb_shinfo(skb)->frags[0]); in is_inline() 507 skb_frag_size(&skb_shinfo(skb)->frags[0])); in build_inline_wqe() 527 fragptr, skb_frag_size(&skb_shinfo(skb)->frags[0])); in build_inline_wqe() 723 frag = &skb_shinfo(skb)->frags[i]; in mlx4_en_xmit()
|
/drivers/net/ethernet/tile/ |
D | tilepro.c | 1681 static unsigned int tile_net_tx_frags(lepp_frag_t *frags, in tile_net_tx_frags() argument 1697 frags[n].cpa_lo = cpa; in tile_net_tx_frags() 1698 frags[n].cpa_hi = cpa >> 32; in tile_net_tx_frags() 1699 frags[n].length = b_len; in tile_net_tx_frags() 1700 frags[n].hash_for_home = hash_default; in tile_net_tx_frags() 1706 skb_frag_t *f = &sh->frags[i]; in tile_net_tx_frags() 1721 frags[n].cpa_lo = cpa; in tile_net_tx_frags() 1722 frags[n].cpa_hi = cpa >> 32; in tile_net_tx_frags() 1723 frags[n].length = skb_frag_size(f); in tile_net_tx_frags() 1724 frags[n].hash_for_home = hash_for_home; in tile_net_tx_frags() [all …]
|
D | tilegx.c | 1356 f_size = skb_frag_size(&sh->frags[f_id]); in tso_count_edescs() 1447 f_size = skb_frag_size(&sh->frags[f_id]); in tso_headers_prepare() 1524 f_size = skb_frag_size(&sh->frags[f_id]); in tso_egress() 1525 f_data = tile_net_frag_buf(&sh->frags[f_id]); in tso_egress() 1616 static unsigned int tile_net_tx_frags(struct frag *frags, in tile_net_tx_frags() argument 1625 frags[n].buf = b_data; in tile_net_tx_frags() 1626 frags[n++].length = b_len; in tile_net_tx_frags() 1630 skb_frag_t *f = &sh->frags[i]; in tile_net_tx_frags() 1631 frags[n].buf = tile_net_frag_buf(f); in tile_net_tx_frags() 1632 frags[n++].length = skb_frag_size(f); in tile_net_tx_frags() [all …]
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 297 end = &si->frags[si->nr_frags]; in map_skb() 298 for (fp = si->frags; fp < end; fp++) { in map_skb() 307 while (fp-- > si->frags) in map_skb() 905 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl() 921 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 922 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl() 927 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 1377 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags() 1378 gl->frags[0].offset + offset, in copy_frags() 1379 gl->frags[0].size - offset); in copy_frags() [all …]
|
D | adapter.h | 147 struct page_frag frags[MAX_SKB_FRAGS]; member
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 247 end = &si->frags[si->nr_frags]; in map_skb() 249 for (fp = si->frags; fp < end; fp++) { in map_skb() 258 while (fp-- > si->frags) in map_skb() 276 end = &si->frags[si->nr_frags]; in unmap_skb() 277 for (fp = si->frags; fp < end; fp++) in unmap_skb() 781 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl() 796 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 797 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl() 802 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 1514 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags() [all …]
|
/drivers/net/xen-netback/ |
D | netback.c | 375 unsigned long size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in xen_netbk_count_skb_slots() 376 unsigned long offset = skb_shinfo(skb)->frags[i].page_offset; in xen_netbk_count_skb_slots() 596 skb_frag_page(&skb_shinfo(skb)->frags[i]), in netbk_gop_skb() 597 skb_frag_size(&skb_shinfo(skb)->frags[i]), in netbk_gop_skb() 598 skb_shinfo(skb)->frags[i].page_offset, in netbk_gop_skb() 1055 skb_frag_t *frags = shinfo->frags; in xen_netbk_get_requests() local 1071 start = (frag_get_pending_idx(&shinfo->frags[0]) == pending_idx); in xen_netbk_get_requests() 1153 frag_set_pending_idx(&frags[shinfo->nr_frags], head_idx); in xen_netbk_get_requests() 1163 frag_get_pending_idx(&frags[shinfo->nr_frags]), in xen_netbk_get_requests() 1191 start = (frag_get_pending_idx(&shinfo->frags[0]) == pending_idx); in xen_netbk_tx_check_gop() [all …]
|
/drivers/media/pci/ttpci/ |
D | av7110_av.c | 584 p->frags = 0; in av7110_p2t_init() 594 p->frags = 0; in clear_p2t() 598 static int find_pes_header(u8 const *buf, long int length, int *frags) in find_pes_header() argument 603 *frags = 0; in find_pes_header() 632 *frags = 1; in find_pes_header() 635 *frags = 2; in find_pes_header() 639 *frags = 3; in find_pes_header() 653 if (p->frags){ in av7110_p2t_write() 655 switch(p->frags) { in av7110_p2t_write() 700 p->frags = 0; in av7110_p2t_write() [all …]
|
D | av7110.h | 59 int frags; member
|
/drivers/net/ |
D | xen-netfront.c | 313 pfn = page_to_pfn(skb_frag_page(&skb_shinfo(skb)->frags[0])); in xennet_alloc_rx_buffers() 314 vaddr = page_address(skb_frag_page(&skb_shinfo(skb)->frags[0])); in xennet_alloc_rx_buffers() 419 int frags = skb_shinfo(skb)->nr_frags; in xennet_make_frags() local 453 for (i = 0; i < frags; i++) { in xennet_make_frags() 454 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_make_frags() 517 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_frag_slots() local 520 for (i = 0; i < frags; i++) { in xennet_count_skb_frag_slots() 521 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_frag_slots() 832 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; in xennet_fill_frags() 1015 skb_shinfo(skb)->frags[0].page_offset = rx->offset; in xennet_poll() [all …]
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_ib.c | 113 skb_frag_t *frag = &skb_shinfo(skb)->frags[0]; in ipoib_ud_skb_put_frags() 334 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_map_tx() 346 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in ipoib_dma_map_tx() 372 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_unmap_tx() 511 skb_frag_t *frags = skb_shinfo(skb)->frags; in post_send() local 524 priv->tx_sge[i + off].length = skb_frag_size(&frags[i]); in post_send()
|
D | ipoib_cm.c | 80 static void ipoib_cm_dma_unmap_rx(struct ipoib_dev_priv *priv, int frags, in ipoib_cm_dma_unmap_rx() argument 87 for (i = 0; i < frags; ++i) in ipoib_cm_dma_unmap_rx() 142 int id, int frags, in ipoib_cm_alloc_rx_skb() argument 166 for (i = 0; i < frags; i++) { in ipoib_cm_alloc_rx_skb() 537 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags() 565 int frags; in ipoib_cm_handle_rx_wc() local 639 frags = PAGE_ALIGN(wc->byte_len - min(wc->byte_len, in ipoib_cm_handle_rx_wc() 642 newskb = ipoib_cm_alloc_rx_skb(dev, rx_ring, wr_id, frags, mapping); in ipoib_cm_handle_rx_wc() 653 ipoib_cm_dma_unmap_rx(priv, frags, rx_ring[wr_id].mapping); in ipoib_cm_handle_rx_wc() 654 memcpy(rx_ring[wr_id].mapping, mapping, (frags + 1) * sizeof *mapping); in ipoib_cm_handle_rx_wc()
|
/drivers/scsi/cxgbi/ |
D | libcxgbi.c | 901 int frags = skb_shinfo(skb)->nr_frags + in cxgbi_sock_send_pdus() local 911 if (frags >= SKB_WR_LIST_SIZE) { in cxgbi_sock_send_pdus() 1801 unsigned int dlen, struct page_frag *frags, in sgl_read_to_frags() argument 1826 if (i && page == frags[i - 1].page && in sgl_read_to_frags() 1828 frags[i - 1].offset + frags[i - 1].size) { in sgl_read_to_frags() 1829 frags[i - 1].size += copy; in sgl_read_to_frags() 1837 frags[i].page = page; in sgl_read_to_frags() 1838 frags[i].offset = sg->offset + sgoffset; in sgl_read_to_frags() 1839 frags[i].size = copy; in sgl_read_to_frags() 1948 tdata->frags, MAX_PDU_FRAGS); in cxgbi_conn_init_pdu() [all …]
|
/drivers/isdn/i4l/ |
D | isdn_ppp.c | 1566 lp->netdev->pb->frags = NULL; in isdn_ppp_mp_init() 1655 if ((frag = mp->frags) == NULL || MP_LT(newseq, MP_SEQ(frag))) { in isdn_ppp_mp_receive() 1657 mp->frags = frag = newfrag; in isdn_ppp_mp_receive() 1721 if (mp->frags == frag) in isdn_ppp_mp_receive() 1722 mp->frags = nextf; in isdn_ppp_mp_receive() 1741 mp->frags = nextf; in isdn_ppp_mp_receive() 1781 mp->frags = isdn_ppp_mp_discard(mp, start, nextf); in isdn_ppp_mp_receive() 1790 if (mp->frags == NULL) in isdn_ppp_mp_receive() 1791 mp->frags = frag; in isdn_ppp_mp_receive() 1798 frag = mp->frags->next; in isdn_ppp_mp_receive() [all …]
|
/drivers/net/ethernet/marvell/ |
D | mvneta.c | 1423 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvneta_tx_frag_process() 1481 int frags = 0; in mvneta_tx() local 1487 frags = skb_shinfo(skb)->nr_frags + 1; in mvneta_tx() 1503 frags = 0; in mvneta_tx() 1507 if (frags == 1) { in mvneta_tx() 1526 frags = 0; in mvneta_tx() 1531 txq->count += frags; in mvneta_tx() 1532 mvneta_txq_pend_desc_add(pp, txq, frags); in mvneta_tx() 1538 if (frags > 0) { in mvneta_tx() 1555 if (txq->count == frags && frags > 0) in mvneta_tx()
|
/drivers/net/ethernet/neterion/vxge/ |
D | vxge-traffic.c | 1476 priv->frags = 0; in vxge_hw_fifo_txdl_reserve() 1519 txdp = (struct vxge_hw_fifo_txd *)txdlh + txdl_priv->frags; in vxge_hw_fifo_txdl_buffer_set() 1529 if (txdl_priv->frags) { in vxge_hw_fifo_txdl_buffer_set() 1531 (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_buffer_set() 1542 txdl_priv->frags++; in vxge_hw_fifo_txdl_buffer_set() 1569 txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_post() 1578 txdl_priv->frags - 1, in vxge_hw_fifo_txdl_post()
|
/drivers/net/ethernet/qlogic/ |
D | qla3xxx.c | 2268 static int ql_get_seg_count(struct ql3_adapter *qdev, unsigned short frags) in ql_get_seg_count() argument 2273 if (frags <= 2) in ql_get_seg_count() 2274 return frags + 1; in ql_get_seg_count() 2275 else if (frags <= 6) in ql_get_seg_count() 2276 return frags + 2; in ql_get_seg_count() 2277 else if (frags <= 10) in ql_get_seg_count() 2278 return frags + 3; in ql_get_seg_count() 2279 else if (frags <= 14) in ql_get_seg_count() 2280 return frags + 4; in ql_get_seg_count() 2281 else if (frags <= 18) in ql_get_seg_count() [all …]
|
/drivers/net/ethernet/tundra/ |
D | tsi108_eth.c | 657 int frags = skb_shinfo(skb)->nr_frags + 1; in tsi108_send_packet() local 679 if (data->txfree - frags < MAX_SKB_FRAGS + 1) { in tsi108_send_packet() 685 for (i = 0; i < frags; i++) { in tsi108_send_packet() 711 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in tsi108_send_packet() 720 if (i == frags - 1) in tsi108_send_packet()
|
/drivers/net/ethernet/aeroflex/ |
D | greth.c | 116 skb_frag_address(&skb_shinfo(skb)->frags[i]), in greth_print_tx_packet() 117 skb_shinfo(skb)->frags[i].size, true); in greth_print_tx_packet() 196 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_rings() 513 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_start_xmit_gbit() 711 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_tx_gbit()
|
/drivers/scsi/cxgbi/cxgb3i/ |
D | cxgb3i.c | 407 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); in push_tx_frames() local 408 int wrs_needed = skb_wrs[frags]; in push_tx_frames() 413 WARN_ON(frags >= SKB_WR_LIST_SIZE || wrs_needed < 1); in push_tx_frames() 418 csk, skb->len, skb->data_len, frags, in push_tx_frames() 433 csk, skb->len, skb->data_len, frags, skb->csum, in push_tx_frames()
|
/drivers/net/ethernet/sfc/ |
D | tx.c | 238 fragment = &skb_shinfo(skb)->frags[i]; in efx_enqueue_skb() 1043 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso() 1065 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()
|