Home
last modified time | relevance | path

Searched refs:frags (Results 1 – 25 of 195) sorted by relevance

12345678

/drivers/net/ethernet/mellanox/mlx5/core/
Dalloc.c83 buf->frags = kzalloc(sizeof(*buf->frags), GFP_KERNEL); in mlx5_buf_alloc_node()
84 if (!buf->frags) in mlx5_buf_alloc_node()
87 buf->frags->buf = mlx5_dma_zalloc_coherent_node(dev, size, in mlx5_buf_alloc_node()
89 if (!buf->frags->buf) in mlx5_buf_alloc_node()
92 buf->frags->map = t; in mlx5_buf_alloc_node()
101 kfree(buf->frags); in mlx5_buf_alloc_node()
114 dma_free_coherent(mlx5_core_dma_dev(dev), buf->size, buf->frags->buf, in mlx5_buf_free()
115 buf->frags->map); in mlx5_buf_free()
117 kfree(buf->frags); in mlx5_buf_free()
129 buf->frags = kcalloc(buf->npages, sizeof(struct mlx5_buf_list), in mlx5_frag_buf_alloc_node()
[all …]
Dwq.c61 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_cyc_create()
128 mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc); in mlx5_wq_qp_create()
136 mlx5_init_fbc_offset(wq_ctrl->buf.frags, in mlx5_wq_qp_create()
142 mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages, in mlx5_wq_qp_create()
185 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc); in mlx5_cqwq_create()
234 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_ll_create()
/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c77 struct mlx4_en_rx_alloc *frags, in mlx4_en_alloc_frags() argument
82 for (i = 0; i < priv->num_frags; i++, frags++) { in mlx4_en_alloc_frags()
83 if (!frags->page) { in mlx4_en_alloc_frags()
84 if (mlx4_alloc_page(priv, frags, gfp)) in mlx4_en_alloc_frags()
88 rx_desc->data[i].addr = cpu_to_be64(frags->dma + in mlx4_en_alloc_frags()
89 frags->page_offset); in mlx4_en_alloc_frags()
139 struct mlx4_en_rx_alloc *frags = ring->rx_info + in mlx4_en_prepare_rx_desc() local
143 if (!frags->page) { in mlx4_en_prepare_rx_desc()
145 frags->page = ring->page_cache.buf[ring->page_cache.index].page; in mlx4_en_prepare_rx_desc()
146 frags->dma = ring->page_cache.buf[ring->page_cache.index].dma; in mlx4_en_prepare_rx_desc()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
Dtls_rxtx.c52 skb_frag_t frags[MAX_SKB_FRAGS]; member
126 info->frags[i] = record->frags[i]; in mlx5e_tls_get_sync_data()
127 __skb_frag_ref(&info->frags[i]); in mlx5e_tls_get_sync_data()
128 remaining -= skb_frag_size(&info->frags[i]); in mlx5e_tls_get_sync_data()
131 skb_frag_size_add(&info->frags[i], remaining); in mlx5e_tls_get_sync_data()
239 skb_shinfo(nskb)->frags[i] = info.frags[i]; in mlx5e_tls_handle_ooo()
Dktls_tx.c221 skb_frag_t frags[MAX_SKB_FRAGS]; member
270 skb_frag_t *frag = &record->frags[i]; in tx_sync_info_get()
274 info->frags[i++] = *frag; in tx_sync_info_get()
278 skb_frag_size_add(&info->frags[i - 1], remaining); in tx_sync_info_get()
413 skb_frag_t *f = &info.frags[i]; in mlx5e_ktls_tx_handle_ooo()
445 put_page(skb_frag_page(&info.frags[i])); in mlx5e_ktls_tx_handle_ooo()
/drivers/net/wireless/ath/ath10k/
Dhtt_tx.c1415 struct htt_data_tx_desc_frag *frags; in ath10k_htt_tx_32() local
1474 frags = (struct htt_data_tx_desc_frag *) in ath10k_htt_tx_32()
1475 &ext_desc_t[msdu_id].frags; in ath10k_htt_tx_32()
1477 frags[0].tword_addr.paddr_lo = in ath10k_htt_tx_32()
1479 frags[0].tword_addr.paddr_hi = 0; in ath10k_htt_tx_32()
1480 frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len); in ath10k_htt_tx_32()
1485 frags = txbuf->frags; in ath10k_htt_tx_32()
1486 frags[0].dword_addr.paddr = in ath10k_htt_tx_32()
1488 frags[0].dword_addr.len = __cpu_to_le32(msdu->len); in ath10k_htt_tx_32()
1489 frags[1].dword_addr.paddr = 0; in ath10k_htt_tx_32()
[all …]
/drivers/net/wireless/ath/wil6210/
Dtrace.h215 TP_PROTO(u8 vring, u16 index, unsigned int len, u8 frags),
216 TP_ARGS(vring, index, len, frags),
219 __field(u8, frags)
225 __entry->frags = frags;
230 __entry->vring, __entry->index, __entry->len, __entry->frags)
/drivers/net/xen-netback/
Dnetback.c386 skb_frag_t *frags = shinfo->frags; in xenvif_get_requests() local
472 frag_set_pending_idx(&frags[shinfo->nr_frags], pending_idx); in xenvif_get_requests()
485 frags = shinfo->frags; in xenvif_get_requests()
498 frag_set_pending_idx(&frags[shinfo->nr_frags], in xenvif_get_requests()
566 frag_get_pending_idx(&shinfo->frags[0]) == in xenvif_tx_check_gop()
609 pending_idx = frag_get_pending_idx(&shinfo->frags[i]); in xenvif_tx_check_gop()
652 pending_idx = frag_get_pending_idx(&shinfo->frags[j]); in xenvif_tx_check_gop()
663 pending_idx = frag_get_pending_idx(&first_shinfo->frags[j]); in xenvif_tx_check_gop()
694 skb_frag_t *frag = shinfo->frags + i; in xenvif_fill_frags()
1122 skb_frag_t frags[MAX_SKB_FRAGS]; in xenvif_handle_frag_list() local
[all …]
/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c288 end = &si->frags[si->nr_frags]; in map_skb()
289 for (fp = si->frags; fp < end; fp++) { in map_skb()
298 while (fp-- > si->frags) in map_skb()
917 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl()
933 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
934 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl()
939 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
1479 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
1480 gl->frags[0].offset + offset, in copy_frags()
1481 gl->frags[0].size - offset); in copy_frags()
[all …]
/drivers/target/iscsi/cxgbit/
Dcxgbit_main.c219 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags()
220 gl->frags[0].offset + offset, in cxgbit_copy_frags()
221 gl->frags[0].size - offset); in cxgbit_copy_frags()
224 gl->frags[i].page, in cxgbit_copy_frags()
225 gl->frags[i].offset, in cxgbit_copy_frags()
226 gl->frags[i].size); in cxgbit_copy_frags()
231 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags()
301 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl()
Dcxgbit_target.c882 skb_frag_t *dfrag = &ssi->frags[pdu_cb->dfrag_idx]; in cxgbit_handle_immediate_data()
1401 pdu_cb->ddigest, pdu_cb->frags); in cxgbit_lro_skb_dump()
1404 skb, i, skb_frag_off(&ssi->frags[i]), in cxgbit_lro_skb_dump()
1405 skb_frag_size(&ssi->frags[i])); in cxgbit_lro_skb_dump()
1416 put_page(skb_frag_page(&ssi->frags[i])); in cxgbit_lro_hskb_reset()
1441 memcpy(&hssi->frags[hfrag_idx], &ssi->frags[pdu_cb->hfrag_idx], in cxgbit_lro_skb_merge()
1444 get_page(skb_frag_page(&hssi->frags[hfrag_idx])); in cxgbit_lro_skb_merge()
1446 hpdu_cb->frags++; in cxgbit_lro_skb_merge()
1449 len = skb_frag_size(&hssi->frags[hfrag_idx]); in cxgbit_lro_skb_merge()
1463 memcpy(&hssi->frags[dfrag_idx], in cxgbit_lro_skb_merge()
[all …]
Dcxgbit_lro.h41 u8 frags; member
/drivers/net/wireless/intel/iwlwifi/
Diwl-dbg-tlv.c351 struct iwl_dram_data *frag = &fw_mon->frags[i]; in iwl_dbg_tlv_fragments_free()
361 kfree(fw_mon->frags); in iwl_dbg_tlv_fragments_free()
362 fw_mon->frags = NULL; in iwl_dbg_tlv_fragments_free()
551 fw_mon->frags = kcalloc(num_frags, sizeof(*fw_mon->frags), GFP_KERNEL); in iwl_dbg_tlv_alloc_fragments()
552 if (!fw_mon->frags) in iwl_dbg_tlv_alloc_fragments()
562 pages = iwl_dbg_tlv_alloc_fragment(fwrt, &fw_mon->frags[i], in iwl_dbg_tlv_alloc_fragments()
637 struct iwl_buf_alloc_frag *frag = &data.frags[j]; in iwl_dbg_tlv_apply_buffer()
639 &fw_mon->frags[fw_mon_idx++]; in iwl_dbg_tlv_apply_buffer()
/drivers/staging/media/av7110/
Dav7110_av.c570 p->frags = 0; in av7110_p2t_init()
580 p->frags = 0; in clear_p2t()
584 static int find_pes_header(u8 const *buf, long int length, int *frags) in find_pes_header() argument
589 *frags = 0; in find_pes_header()
618 *frags = 1; in find_pes_header()
621 *frags = 2; in find_pes_header()
625 *frags = 3; in find_pes_header()
639 if (p->frags){ in av7110_p2t_write()
641 switch(p->frags) { in av7110_p2t_write()
686 p->frags = 0; in av7110_p2t_write()
[all …]
/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c254 end = &si->frags[si->nr_frags]; in cxgb4_map_skb()
256 for (fp = si->frags; fp < end; fp++) { in cxgb4_map_skb()
265 while (fp-- > si->frags) in cxgb4_map_skb()
283 end = &si->frags[si->nr_frags]; in unmap_skb()
284 for (fp = si->frags; fp < end; fp++) in unmap_skb()
853 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in cxgb4_write_sgl()
869 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl()
870 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in cxgb4_write_sgl()
875 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl()
927 frag = &si->frags[frag_idx]; in cxgb4_write_partial_sgl()
[all …]
/drivers/staging/qlge/
DTODO8 while containing two frags of order-1 allocations, ie. >16K)
12 head room but only puts data in the frags.
/drivers/infiniband/hw/mlx5/
Dodp.c1689 } frags[]; member
1697 mlx5r_deref_odp_mkey(&work->frags[i].mr->mmkey); in destroy_prefetch_work()
1746 ret = pagefault_mr(work->frags[i].mr, work->frags[i].io_virt, in mlx5_ib_prefetch_mr_work()
1747 work->frags[i].length, &bytes_mapped, in mlx5_ib_prefetch_mr_work()
1751 mlx5_update_odp_stats(work->frags[i].mr, prefetch, ret); in mlx5_ib_prefetch_mr_work()
1768 work->frags[i].io_virt = sg_list[i].addr; in init_prefetch_work()
1769 work->frags[i].length = sg_list[i].length; in init_prefetch_work()
1770 work->frags[i].mr = in init_prefetch_work()
1772 if (!work->frags[i].mr) { in init_prefetch_work()
1826 work = kvzalloc(struct_size(work, frags, num_sge), GFP_KERNEL); in mlx5_ib_advise_mr_prefetch()
/drivers/staging/unisys/visornic/
Dvisornic_main.c238 struct phys_info frags[]) in visor_copy_fragsinfo_from_skb() argument
263 frags[count].pi_pfn = in visor_copy_fragsinfo_from_skb()
265 frags[count].pi_off = in visor_copy_fragsinfo_from_skb()
268 PI_PAGE_SIZE - frags[count].pi_off); in visor_copy_fragsinfo_from_skb()
273 frags[count].pi_len = size; in visor_copy_fragsinfo_from_skb()
284 skb_frag_page(&skb_shinfo(skb)->frags[frag])), in visor_copy_fragsinfo_from_skb()
285 skb_frag_off(&skb_shinfo(skb)->frags[frag]), in visor_copy_fragsinfo_from_skb()
286 skb_frag_size(&skb_shinfo(skb)->frags[frag]), in visor_copy_fragsinfo_from_skb()
287 count, frags_max, frags); in visor_copy_fragsinfo_from_skb()
307 &frags[count]); in visor_copy_fragsinfo_from_skb()
[all …]
/drivers/net/ethernet/cavium/liquidio/
Dlio_vf_main.c829 int i, frags, iq; in free_netsgbuf() local
836 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf()
843 while (frags--) { in free_netsgbuf()
844 skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf()
871 int i, frags, iq; in free_netsgbuf_with_resp() local
880 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf_with_resp()
887 while (frags--) { in free_netsgbuf_with_resp()
888 skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf_with_resp()
1513 int i, frags; in liquidio_xmit() local
1543 frags = skb_shinfo(skb)->nr_frags; in liquidio_xmit()
[all …]
/drivers/net/ethernet/aquantia/atlantic/
Daq_ptp.c752 unsigned int frags; in aq_ptp_xmit() local
759 frags = skb_shinfo(skb)->nr_frags + 1; in aq_ptp_xmit()
764 if (frags > AQ_CFG_SKB_FRAGS_MAX || frags > aq_ring_avail_dx(ring)) { in aq_ptp_xmit()
781 frags = aq_nic_map_skb(aq_nic, skb, ring); in aq_ptp_xmit()
783 if (likely(frags)) { in aq_ptp_xmit()
785 ring, frags); in aq_ptp_xmit()
/drivers/infiniband/ulp/ipoib/
Dipoib_cm.c83 static void ipoib_cm_dma_unmap_rx(struct ipoib_dev_priv *priv, int frags, in ipoib_cm_dma_unmap_rx() argument
90 for (i = 0; i < frags; ++i) in ipoib_cm_dma_unmap_rx()
143 int id, int frags, in ipoib_cm_alloc_rx_skb() argument
168 for (i = 0; i < frags; i++) { in ipoib_cm_alloc_rx_skb()
541 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags()
569 int frags; in ipoib_cm_handle_rx_wc() local
643 frags = PAGE_ALIGN(wc->byte_len - in ipoib_cm_handle_rx_wc()
647 newskb = ipoib_cm_alloc_rx_skb(dev, rx_ring, wr_id, frags, in ipoib_cm_handle_rx_wc()
659 ipoib_cm_dma_unmap_rx(priv, frags, rx_ring[wr_id].mapping); in ipoib_cm_handle_rx_wc()
660 memcpy(rx_ring[wr_id].mapping, mapping, (frags + 1) * sizeof(*mapping)); in ipoib_cm_handle_rx_wc()
/drivers/net/ethernet/neterion/vxge/
Dvxge-traffic.c1446 priv->frags = 0; in vxge_hw_fifo_txdl_reserve()
1486 txdp = (struct vxge_hw_fifo_txd *)txdlh + txdl_priv->frags; in vxge_hw_fifo_txdl_buffer_set()
1496 if (txdl_priv->frags) { in vxge_hw_fifo_txdl_buffer_set()
1498 (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_buffer_set()
1509 txdl_priv->frags++; in vxge_hw_fifo_txdl_buffer_set()
1531 txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_post()
1540 txdl_priv->frags - 1, in vxge_hw_fifo_txdl_post()
/drivers/net/ethernet/aquantia/atlantic/hw_atl/
Dhw_atl_b0.h52 unsigned int frags);
/drivers/net/ethernet/hisilicon/
Dhix5hd2_gmac.c236 struct frags_info frags[SG_MAX_SKB_FRAGS + 1]; member
581 addr = le32_to_cpu(desc->frags[i].addr); in hix5hd2_clean_sg_desc()
582 len = le32_to_cpu(desc->frags[i].size); in hix5hd2_clean_sg_desc()
721 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in hix5hd2_fill_sg_desc()
728 desc->frags[i].addr = cpu_to_le32(addr); in hix5hd2_fill_sg_desc()
729 desc->frags[i].size = cpu_to_le32(len); in hix5hd2_fill_sg_desc()
/drivers/net/
Dxen-netfront.c332 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_alloc_rx_buffers()
566 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_slots() local
572 for (i = 0; i < frags; i++) { in xennet_count_skb_slots()
573 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_slots()
829 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xennet_start_xmit()
1087 skb_frag_page(&skb_shinfo(skb)->frags[0]), in xennet_get_responses()
1168 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; in xennet_fill_frags()
1331 skb_frag_off_set(&skb_shinfo(skb)->frags[0], rx->offset); in xennet_poll()
1332 skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status); in xennet_poll()
1459 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_release_rx_bufs()

12345678