Home
last modified time | relevance | path

Searched refs:frags (Results 1 – 25 of 94) sorted by relevance

1234

/drivers/infiniband/hw/nes/
Dnes_mgt.c158 skb = fpdu_info->frags[i].skb; in nes_download_callback()
159 if (fpdu_info->frags[i].cmplt) { in nes_download_callback()
249 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; in get_fpdu_info() local
276 memset(&frags, 0, sizeof frags); in get_fpdu_info()
278 frags[i].physaddr = cb->busaddr; in get_fpdu_info()
279 frags[i].physaddr += skb->data - cb->data_start; in get_fpdu_info()
280 frags[i].frag_len = min(tmp_len, skb->len); in get_fpdu_info()
281 frags[i].skb = skb; in get_fpdu_info()
282 frags[i].cmplt = (skb->len == frags[i].frag_len); in get_fpdu_info()
283 frag_tot += frags[i].frag_len; in get_fpdu_info()
[all …]
Dnes_mgt.h82 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; member
Dnes_nic.c440 &skb_shinfo(skb)->frags[skb_fragment_index]; in nes_nic_send()
445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); in nes_nic_send()
560 &skb_shinfo(skb)->frags[tso_frag_count]; in nes_netdev_start_xmit()
635 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index])); in nes_netdev_start_xmit()
639 tso_wqe_length += skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index++]); in nes_netdev_start_xmit()
/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c47 struct mlx4_en_rx_alloc *frags, in mlx4_en_alloc_frags() argument
82 frags[i] = ring_alloc[i]; in mlx4_en_alloc_frags()
103 struct mlx4_en_rx_alloc *frags, in mlx4_en_free_frag() argument
108 if (frags[i].offset == frag_info->last_offset) { in mlx4_en_free_frag()
109 dma_unmap_page(priv->ddev, frags[i].dma, MLX4_EN_ALLOC_SIZE, in mlx4_en_free_frag()
112 if (frags[i].page) in mlx4_en_free_frag()
113 put_page(frags[i].page); in mlx4_en_free_frag()
200 struct mlx4_en_rx_alloc *frags = ring->rx_info + in mlx4_en_prepare_rx_desc() local
203 return mlx4_en_alloc_frags(priv, rx_desc, frags, ring->page_alloc); in mlx4_en_prepare_rx_desc()
215 struct mlx4_en_rx_alloc *frags; in mlx4_en_free_rx_desc() local
[all …]
Den_tx.c206 int frags = skb_shinfo(skb)->nr_frags; in mlx4_en_free_tx_desc() local
228 for (i = 0; i < frags; i++) { in mlx4_en_free_tx_desc()
229 frag = &skb_shinfo(skb)->frags[i]; in mlx4_en_free_tx_desc()
255 for (i = 0; i < frags; i++) { in mlx4_en_free_tx_desc()
259 frag = &skb_shinfo(skb)->frags[i]; in mlx4_en_free_tx_desc()
436 ptr = skb_frag_address_safe(&skb_shinfo(skb)->frags[0]); in is_inline()
507 skb_frag_size(&skb_shinfo(skb)->frags[0])); in build_inline_wqe()
527 fragptr, skb_frag_size(&skb_shinfo(skb)->frags[0])); in build_inline_wqe()
723 frag = &skb_shinfo(skb)->frags[i]; in mlx4_en_xmit()
/drivers/net/ethernet/tile/
Dtilepro.c1681 static unsigned int tile_net_tx_frags(lepp_frag_t *frags, in tile_net_tx_frags() argument
1697 frags[n].cpa_lo = cpa; in tile_net_tx_frags()
1698 frags[n].cpa_hi = cpa >> 32; in tile_net_tx_frags()
1699 frags[n].length = b_len; in tile_net_tx_frags()
1700 frags[n].hash_for_home = hash_default; in tile_net_tx_frags()
1706 skb_frag_t *f = &sh->frags[i]; in tile_net_tx_frags()
1721 frags[n].cpa_lo = cpa; in tile_net_tx_frags()
1722 frags[n].cpa_hi = cpa >> 32; in tile_net_tx_frags()
1723 frags[n].length = skb_frag_size(f); in tile_net_tx_frags()
1724 frags[n].hash_for_home = hash_for_home; in tile_net_tx_frags()
[all …]
Dtilegx.c1356 f_size = skb_frag_size(&sh->frags[f_id]); in tso_count_edescs()
1447 f_size = skb_frag_size(&sh->frags[f_id]); in tso_headers_prepare()
1524 f_size = skb_frag_size(&sh->frags[f_id]); in tso_egress()
1525 f_data = tile_net_frag_buf(&sh->frags[f_id]); in tso_egress()
1616 static unsigned int tile_net_tx_frags(struct frag *frags, in tile_net_tx_frags() argument
1625 frags[n].buf = b_data; in tile_net_tx_frags()
1626 frags[n++].length = b_len; in tile_net_tx_frags()
1630 skb_frag_t *f = &sh->frags[i]; in tile_net_tx_frags()
1631 frags[n].buf = tile_net_frag_buf(f); in tile_net_tx_frags()
1632 frags[n++].length = skb_frag_size(f); in tile_net_tx_frags()
[all …]
/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c297 end = &si->frags[si->nr_frags]; in map_skb()
298 for (fp = si->frags; fp < end; fp++) { in map_skb()
307 while (fp-- > si->frags) in map_skb()
905 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl()
921 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
922 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl()
927 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
1377 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
1378 gl->frags[0].offset + offset, in copy_frags()
1379 gl->frags[0].size - offset); in copy_frags()
[all …]
Dadapter.h147 struct page_frag frags[MAX_SKB_FRAGS]; member
/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c247 end = &si->frags[si->nr_frags]; in map_skb()
249 for (fp = si->frags; fp < end; fp++) { in map_skb()
258 while (fp-- > si->frags) in map_skb()
276 end = &si->frags[si->nr_frags]; in unmap_skb()
277 for (fp = si->frags; fp < end; fp++) in unmap_skb()
781 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl()
796 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
797 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl()
802 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
1514 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
[all …]
/drivers/net/xen-netback/
Dnetback.c375 unsigned long size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in xen_netbk_count_skb_slots()
376 unsigned long offset = skb_shinfo(skb)->frags[i].page_offset; in xen_netbk_count_skb_slots()
596 skb_frag_page(&skb_shinfo(skb)->frags[i]), in netbk_gop_skb()
597 skb_frag_size(&skb_shinfo(skb)->frags[i]), in netbk_gop_skb()
598 skb_shinfo(skb)->frags[i].page_offset, in netbk_gop_skb()
1055 skb_frag_t *frags = shinfo->frags; in xen_netbk_get_requests() local
1071 start = (frag_get_pending_idx(&shinfo->frags[0]) == pending_idx); in xen_netbk_get_requests()
1153 frag_set_pending_idx(&frags[shinfo->nr_frags], head_idx); in xen_netbk_get_requests()
1163 frag_get_pending_idx(&frags[shinfo->nr_frags]), in xen_netbk_get_requests()
1191 start = (frag_get_pending_idx(&shinfo->frags[0]) == pending_idx); in xen_netbk_tx_check_gop()
[all …]
/drivers/media/pci/ttpci/
Dav7110_av.c584 p->frags = 0; in av7110_p2t_init()
594 p->frags = 0; in clear_p2t()
598 static int find_pes_header(u8 const *buf, long int length, int *frags) in find_pes_header() argument
603 *frags = 0; in find_pes_header()
632 *frags = 1; in find_pes_header()
635 *frags = 2; in find_pes_header()
639 *frags = 3; in find_pes_header()
653 if (p->frags){ in av7110_p2t_write()
655 switch(p->frags) { in av7110_p2t_write()
700 p->frags = 0; in av7110_p2t_write()
[all …]
Dav7110.h59 int frags; member
/drivers/net/
Dxen-netfront.c313 pfn = page_to_pfn(skb_frag_page(&skb_shinfo(skb)->frags[0])); in xennet_alloc_rx_buffers()
314 vaddr = page_address(skb_frag_page(&skb_shinfo(skb)->frags[0])); in xennet_alloc_rx_buffers()
419 int frags = skb_shinfo(skb)->nr_frags; in xennet_make_frags() local
453 for (i = 0; i < frags; i++) { in xennet_make_frags()
454 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_make_frags()
517 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_frag_slots() local
520 for (i = 0; i < frags; i++) { in xennet_count_skb_frag_slots()
521 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_frag_slots()
832 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; in xennet_fill_frags()
1015 skb_shinfo(skb)->frags[0].page_offset = rx->offset; in xennet_poll()
[all …]
/drivers/infiniband/ulp/ipoib/
Dipoib_ib.c113 skb_frag_t *frag = &skb_shinfo(skb)->frags[0]; in ipoib_ud_skb_put_frags()
334 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_map_tx()
346 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in ipoib_dma_map_tx()
372 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_unmap_tx()
511 skb_frag_t *frags = skb_shinfo(skb)->frags; in post_send() local
524 priv->tx_sge[i + off].length = skb_frag_size(&frags[i]); in post_send()
Dipoib_cm.c80 static void ipoib_cm_dma_unmap_rx(struct ipoib_dev_priv *priv, int frags, in ipoib_cm_dma_unmap_rx() argument
87 for (i = 0; i < frags; ++i) in ipoib_cm_dma_unmap_rx()
142 int id, int frags, in ipoib_cm_alloc_rx_skb() argument
166 for (i = 0; i < frags; i++) { in ipoib_cm_alloc_rx_skb()
537 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags()
565 int frags; in ipoib_cm_handle_rx_wc() local
639 frags = PAGE_ALIGN(wc->byte_len - min(wc->byte_len, in ipoib_cm_handle_rx_wc()
642 newskb = ipoib_cm_alloc_rx_skb(dev, rx_ring, wr_id, frags, mapping); in ipoib_cm_handle_rx_wc()
653 ipoib_cm_dma_unmap_rx(priv, frags, rx_ring[wr_id].mapping); in ipoib_cm_handle_rx_wc()
654 memcpy(rx_ring[wr_id].mapping, mapping, (frags + 1) * sizeof *mapping); in ipoib_cm_handle_rx_wc()
/drivers/scsi/cxgbi/
Dlibcxgbi.c901 int frags = skb_shinfo(skb)->nr_frags + in cxgbi_sock_send_pdus() local
911 if (frags >= SKB_WR_LIST_SIZE) { in cxgbi_sock_send_pdus()
1801 unsigned int dlen, struct page_frag *frags, in sgl_read_to_frags() argument
1826 if (i && page == frags[i - 1].page && in sgl_read_to_frags()
1828 frags[i - 1].offset + frags[i - 1].size) { in sgl_read_to_frags()
1829 frags[i - 1].size += copy; in sgl_read_to_frags()
1837 frags[i].page = page; in sgl_read_to_frags()
1838 frags[i].offset = sg->offset + sgoffset; in sgl_read_to_frags()
1839 frags[i].size = copy; in sgl_read_to_frags()
1948 tdata->frags, MAX_PDU_FRAGS); in cxgbi_conn_init_pdu()
[all …]
/drivers/isdn/i4l/
Disdn_ppp.c1566 lp->netdev->pb->frags = NULL; in isdn_ppp_mp_init()
1655 if ((frag = mp->frags) == NULL || MP_LT(newseq, MP_SEQ(frag))) { in isdn_ppp_mp_receive()
1657 mp->frags = frag = newfrag; in isdn_ppp_mp_receive()
1721 if (mp->frags == frag) in isdn_ppp_mp_receive()
1722 mp->frags = nextf; in isdn_ppp_mp_receive()
1741 mp->frags = nextf; in isdn_ppp_mp_receive()
1781 mp->frags = isdn_ppp_mp_discard(mp, start, nextf); in isdn_ppp_mp_receive()
1790 if (mp->frags == NULL) in isdn_ppp_mp_receive()
1791 mp->frags = frag; in isdn_ppp_mp_receive()
1798 frag = mp->frags->next; in isdn_ppp_mp_receive()
[all …]
/drivers/net/ethernet/marvell/
Dmvneta.c1423 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvneta_tx_frag_process()
1481 int frags = 0; in mvneta_tx() local
1487 frags = skb_shinfo(skb)->nr_frags + 1; in mvneta_tx()
1503 frags = 0; in mvneta_tx()
1507 if (frags == 1) { in mvneta_tx()
1526 frags = 0; in mvneta_tx()
1531 txq->count += frags; in mvneta_tx()
1532 mvneta_txq_pend_desc_add(pp, txq, frags); in mvneta_tx()
1538 if (frags > 0) { in mvneta_tx()
1555 if (txq->count == frags && frags > 0) in mvneta_tx()
/drivers/net/ethernet/neterion/vxge/
Dvxge-traffic.c1476 priv->frags = 0; in vxge_hw_fifo_txdl_reserve()
1519 txdp = (struct vxge_hw_fifo_txd *)txdlh + txdl_priv->frags; in vxge_hw_fifo_txdl_buffer_set()
1529 if (txdl_priv->frags) { in vxge_hw_fifo_txdl_buffer_set()
1531 (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_buffer_set()
1542 txdl_priv->frags++; in vxge_hw_fifo_txdl_buffer_set()
1569 txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_post()
1578 txdl_priv->frags - 1, in vxge_hw_fifo_txdl_post()
/drivers/net/ethernet/qlogic/
Dqla3xxx.c2268 static int ql_get_seg_count(struct ql3_adapter *qdev, unsigned short frags) in ql_get_seg_count() argument
2273 if (frags <= 2) in ql_get_seg_count()
2274 return frags + 1; in ql_get_seg_count()
2275 else if (frags <= 6) in ql_get_seg_count()
2276 return frags + 2; in ql_get_seg_count()
2277 else if (frags <= 10) in ql_get_seg_count()
2278 return frags + 3; in ql_get_seg_count()
2279 else if (frags <= 14) in ql_get_seg_count()
2280 return frags + 4; in ql_get_seg_count()
2281 else if (frags <= 18) in ql_get_seg_count()
[all …]
/drivers/net/ethernet/tundra/
Dtsi108_eth.c657 int frags = skb_shinfo(skb)->nr_frags + 1; in tsi108_send_packet() local
679 if (data->txfree - frags < MAX_SKB_FRAGS + 1) { in tsi108_send_packet()
685 for (i = 0; i < frags; i++) { in tsi108_send_packet()
711 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in tsi108_send_packet()
720 if (i == frags - 1) in tsi108_send_packet()
/drivers/net/ethernet/aeroflex/
Dgreth.c116 skb_frag_address(&skb_shinfo(skb)->frags[i]), in greth_print_tx_packet()
117 skb_shinfo(skb)->frags[i].size, true); in greth_print_tx_packet()
196 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_rings()
513 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_start_xmit_gbit()
711 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_tx_gbit()
/drivers/scsi/cxgbi/cxgb3i/
Dcxgb3i.c407 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); in push_tx_frames() local
408 int wrs_needed = skb_wrs[frags]; in push_tx_frames()
413 WARN_ON(frags >= SKB_WR_LIST_SIZE || wrs_needed < 1); in push_tx_frames()
418 csk, skb->len, skb->data_len, frags, in push_tx_frames()
433 csk, skb->len, skb->data_len, frags, skb->csum, in push_tx_frames()
/drivers/net/ethernet/sfc/
Dtx.c238 fragment = &skb_shinfo(skb)->frags[i]; in efx_enqueue_skb()
1043 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()
1065 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()

1234