/drivers/net/ethernet/mellanox/mlx5/core/ |
D | alloc.c | 83 buf->frags = kzalloc(sizeof(*buf->frags), GFP_KERNEL); in mlx5_buf_alloc_node() 84 if (!buf->frags) in mlx5_buf_alloc_node() 87 buf->frags->buf = mlx5_dma_zalloc_coherent_node(dev, size, in mlx5_buf_alloc_node() 89 if (!buf->frags->buf) in mlx5_buf_alloc_node() 92 buf->frags->map = t; in mlx5_buf_alloc_node() 101 kfree(buf->frags); in mlx5_buf_alloc_node() 114 dma_free_coherent(mlx5_core_dma_dev(dev), buf->size, buf->frags->buf, in mlx5_buf_free() 115 buf->frags->map); in mlx5_buf_free() 117 kfree(buf->frags); in mlx5_buf_free() 129 buf->frags = kcalloc(buf->npages, sizeof(struct mlx5_buf_list), in mlx5_frag_buf_alloc_node() [all …]
|
D | wq.c | 61 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_cyc_create() 128 mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc); in mlx5_wq_qp_create() 136 mlx5_init_fbc_offset(wq_ctrl->buf.frags, in mlx5_wq_qp_create() 142 mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages, in mlx5_wq_qp_create() 185 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc); in mlx5_cqwq_create() 234 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_ll_create()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 77 struct mlx4_en_rx_alloc *frags, in mlx4_en_alloc_frags() argument 82 for (i = 0; i < priv->num_frags; i++, frags++) { in mlx4_en_alloc_frags() 83 if (!frags->page) { in mlx4_en_alloc_frags() 84 if (mlx4_alloc_page(priv, frags, gfp)) in mlx4_en_alloc_frags() 88 rx_desc->data[i].addr = cpu_to_be64(frags->dma + in mlx4_en_alloc_frags() 89 frags->page_offset); in mlx4_en_alloc_frags() 139 struct mlx4_en_rx_alloc *frags = ring->rx_info + in mlx4_en_prepare_rx_desc() local 143 if (!frags->page) { in mlx4_en_prepare_rx_desc() 145 frags->page = ring->page_cache.buf[ring->page_cache.index].page; in mlx4_en_prepare_rx_desc() 146 frags->dma = ring->page_cache.buf[ring->page_cache.index].dma; in mlx4_en_prepare_rx_desc() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | tls_rxtx.c | 52 skb_frag_t frags[MAX_SKB_FRAGS]; member 126 info->frags[i] = record->frags[i]; in mlx5e_tls_get_sync_data() 127 __skb_frag_ref(&info->frags[i]); in mlx5e_tls_get_sync_data() 128 remaining -= skb_frag_size(&info->frags[i]); in mlx5e_tls_get_sync_data() 131 skb_frag_size_add(&info->frags[i], remaining); in mlx5e_tls_get_sync_data() 239 skb_shinfo(nskb)->frags[i] = info.frags[i]; in mlx5e_tls_handle_ooo()
|
D | ktls_tx.c | 221 skb_frag_t frags[MAX_SKB_FRAGS]; member 270 skb_frag_t *frag = &record->frags[i]; in tx_sync_info_get() 274 info->frags[i++] = *frag; in tx_sync_info_get() 278 skb_frag_size_add(&info->frags[i - 1], remaining); in tx_sync_info_get() 413 skb_frag_t *f = &info.frags[i]; in mlx5e_ktls_tx_handle_ooo() 445 put_page(skb_frag_page(&info.frags[i])); in mlx5e_ktls_tx_handle_ooo()
|
/drivers/net/wireless/ath/ath10k/ |
D | htt_tx.c | 1415 struct htt_data_tx_desc_frag *frags; in ath10k_htt_tx_32() local 1474 frags = (struct htt_data_tx_desc_frag *) in ath10k_htt_tx_32() 1475 &ext_desc_t[msdu_id].frags; in ath10k_htt_tx_32() 1477 frags[0].tword_addr.paddr_lo = in ath10k_htt_tx_32() 1479 frags[0].tword_addr.paddr_hi = 0; in ath10k_htt_tx_32() 1480 frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len); in ath10k_htt_tx_32() 1485 frags = txbuf->frags; in ath10k_htt_tx_32() 1486 frags[0].dword_addr.paddr = in ath10k_htt_tx_32() 1488 frags[0].dword_addr.len = __cpu_to_le32(msdu->len); in ath10k_htt_tx_32() 1489 frags[1].dword_addr.paddr = 0; in ath10k_htt_tx_32() [all …]
|
/drivers/net/wireless/ath/wil6210/ |
D | trace.h | 215 TP_PROTO(u8 vring, u16 index, unsigned int len, u8 frags), 216 TP_ARGS(vring, index, len, frags), 219 __field(u8, frags) 225 __entry->frags = frags; 230 __entry->vring, __entry->index, __entry->len, __entry->frags)
|
/drivers/net/xen-netback/ |
D | netback.c | 386 skb_frag_t *frags = shinfo->frags; in xenvif_get_requests() local 472 frag_set_pending_idx(&frags[shinfo->nr_frags], pending_idx); in xenvif_get_requests() 485 frags = shinfo->frags; in xenvif_get_requests() 498 frag_set_pending_idx(&frags[shinfo->nr_frags], in xenvif_get_requests() 566 frag_get_pending_idx(&shinfo->frags[0]) == in xenvif_tx_check_gop() 609 pending_idx = frag_get_pending_idx(&shinfo->frags[i]); in xenvif_tx_check_gop() 652 pending_idx = frag_get_pending_idx(&shinfo->frags[j]); in xenvif_tx_check_gop() 663 pending_idx = frag_get_pending_idx(&first_shinfo->frags[j]); in xenvif_tx_check_gop() 694 skb_frag_t *frag = shinfo->frags + i; in xenvif_fill_frags() 1122 skb_frag_t frags[MAX_SKB_FRAGS]; in xenvif_handle_frag_list() local [all …]
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 288 end = &si->frags[si->nr_frags]; in map_skb() 289 for (fp = si->frags; fp < end; fp++) { in map_skb() 298 while (fp-- > si->frags) in map_skb() 917 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl() 933 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 934 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl() 939 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 1479 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags() 1480 gl->frags[0].offset + offset, in copy_frags() 1481 gl->frags[0].size - offset); in copy_frags() [all …]
|
/drivers/target/iscsi/cxgbit/ |
D | cxgbit_main.c | 219 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags() 220 gl->frags[0].offset + offset, in cxgbit_copy_frags() 221 gl->frags[0].size - offset); in cxgbit_copy_frags() 224 gl->frags[i].page, in cxgbit_copy_frags() 225 gl->frags[i].offset, in cxgbit_copy_frags() 226 gl->frags[i].size); in cxgbit_copy_frags() 231 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags() 301 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl()
|
D | cxgbit_target.c | 882 skb_frag_t *dfrag = &ssi->frags[pdu_cb->dfrag_idx]; in cxgbit_handle_immediate_data() 1401 pdu_cb->ddigest, pdu_cb->frags); in cxgbit_lro_skb_dump() 1404 skb, i, skb_frag_off(&ssi->frags[i]), in cxgbit_lro_skb_dump() 1405 skb_frag_size(&ssi->frags[i])); in cxgbit_lro_skb_dump() 1416 put_page(skb_frag_page(&ssi->frags[i])); in cxgbit_lro_hskb_reset() 1441 memcpy(&hssi->frags[hfrag_idx], &ssi->frags[pdu_cb->hfrag_idx], in cxgbit_lro_skb_merge() 1444 get_page(skb_frag_page(&hssi->frags[hfrag_idx])); in cxgbit_lro_skb_merge() 1446 hpdu_cb->frags++; in cxgbit_lro_skb_merge() 1449 len = skb_frag_size(&hssi->frags[hfrag_idx]); in cxgbit_lro_skb_merge() 1463 memcpy(&hssi->frags[dfrag_idx], in cxgbit_lro_skb_merge() [all …]
|
D | cxgbit_lro.h | 41 u8 frags; member
|
/drivers/net/wireless/intel/iwlwifi/ |
D | iwl-dbg-tlv.c | 351 struct iwl_dram_data *frag = &fw_mon->frags[i]; in iwl_dbg_tlv_fragments_free() 361 kfree(fw_mon->frags); in iwl_dbg_tlv_fragments_free() 362 fw_mon->frags = NULL; in iwl_dbg_tlv_fragments_free() 551 fw_mon->frags = kcalloc(num_frags, sizeof(*fw_mon->frags), GFP_KERNEL); in iwl_dbg_tlv_alloc_fragments() 552 if (!fw_mon->frags) in iwl_dbg_tlv_alloc_fragments() 562 pages = iwl_dbg_tlv_alloc_fragment(fwrt, &fw_mon->frags[i], in iwl_dbg_tlv_alloc_fragments() 637 struct iwl_buf_alloc_frag *frag = &data.frags[j]; in iwl_dbg_tlv_apply_buffer() 639 &fw_mon->frags[fw_mon_idx++]; in iwl_dbg_tlv_apply_buffer()
|
/drivers/staging/media/av7110/ |
D | av7110_av.c | 570 p->frags = 0; in av7110_p2t_init() 580 p->frags = 0; in clear_p2t() 584 static int find_pes_header(u8 const *buf, long int length, int *frags) in find_pes_header() argument 589 *frags = 0; in find_pes_header() 618 *frags = 1; in find_pes_header() 621 *frags = 2; in find_pes_header() 625 *frags = 3; in find_pes_header() 639 if (p->frags){ in av7110_p2t_write() 641 switch(p->frags) { in av7110_p2t_write() 686 p->frags = 0; in av7110_p2t_write() [all …]
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 254 end = &si->frags[si->nr_frags]; in cxgb4_map_skb() 256 for (fp = si->frags; fp < end; fp++) { in cxgb4_map_skb() 265 while (fp-- > si->frags) in cxgb4_map_skb() 283 end = &si->frags[si->nr_frags]; in unmap_skb() 284 for (fp = si->frags; fp < end; fp++) in unmap_skb() 853 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in cxgb4_write_sgl() 869 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl() 870 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in cxgb4_write_sgl() 875 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl() 927 frag = &si->frags[frag_idx]; in cxgb4_write_partial_sgl() [all …]
|
/drivers/staging/qlge/ |
D | TODO | 8 while containing two frags of order-1 allocations, ie. >16K) 12 head room but only puts data in the frags.
|
/drivers/infiniband/hw/mlx5/ |
D | odp.c | 1689 } frags[]; member 1697 mlx5r_deref_odp_mkey(&work->frags[i].mr->mmkey); in destroy_prefetch_work() 1746 ret = pagefault_mr(work->frags[i].mr, work->frags[i].io_virt, in mlx5_ib_prefetch_mr_work() 1747 work->frags[i].length, &bytes_mapped, in mlx5_ib_prefetch_mr_work() 1751 mlx5_update_odp_stats(work->frags[i].mr, prefetch, ret); in mlx5_ib_prefetch_mr_work() 1768 work->frags[i].io_virt = sg_list[i].addr; in init_prefetch_work() 1769 work->frags[i].length = sg_list[i].length; in init_prefetch_work() 1770 work->frags[i].mr = in init_prefetch_work() 1772 if (!work->frags[i].mr) { in init_prefetch_work() 1826 work = kvzalloc(struct_size(work, frags, num_sge), GFP_KERNEL); in mlx5_ib_advise_mr_prefetch()
|
/drivers/staging/unisys/visornic/ |
D | visornic_main.c | 238 struct phys_info frags[]) in visor_copy_fragsinfo_from_skb() argument 263 frags[count].pi_pfn = in visor_copy_fragsinfo_from_skb() 265 frags[count].pi_off = in visor_copy_fragsinfo_from_skb() 268 PI_PAGE_SIZE - frags[count].pi_off); in visor_copy_fragsinfo_from_skb() 273 frags[count].pi_len = size; in visor_copy_fragsinfo_from_skb() 284 skb_frag_page(&skb_shinfo(skb)->frags[frag])), in visor_copy_fragsinfo_from_skb() 285 skb_frag_off(&skb_shinfo(skb)->frags[frag]), in visor_copy_fragsinfo_from_skb() 286 skb_frag_size(&skb_shinfo(skb)->frags[frag]), in visor_copy_fragsinfo_from_skb() 287 count, frags_max, frags); in visor_copy_fragsinfo_from_skb() 307 &frags[count]); in visor_copy_fragsinfo_from_skb() [all …]
|
/drivers/net/ethernet/cavium/liquidio/ |
D | lio_vf_main.c | 829 int i, frags, iq; in free_netsgbuf() local 836 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf() 843 while (frags--) { in free_netsgbuf() 844 skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf() 871 int i, frags, iq; in free_netsgbuf_with_resp() local 880 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf_with_resp() 887 while (frags--) { in free_netsgbuf_with_resp() 888 skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf_with_resp() 1513 int i, frags; in liquidio_xmit() local 1543 frags = skb_shinfo(skb)->nr_frags; in liquidio_xmit() [all …]
|
/drivers/net/ethernet/aquantia/atlantic/ |
D | aq_ptp.c | 752 unsigned int frags; in aq_ptp_xmit() local 759 frags = skb_shinfo(skb)->nr_frags + 1; in aq_ptp_xmit() 764 if (frags > AQ_CFG_SKB_FRAGS_MAX || frags > aq_ring_avail_dx(ring)) { in aq_ptp_xmit() 781 frags = aq_nic_map_skb(aq_nic, skb, ring); in aq_ptp_xmit() 783 if (likely(frags)) { in aq_ptp_xmit() 785 ring, frags); in aq_ptp_xmit()
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_cm.c | 83 static void ipoib_cm_dma_unmap_rx(struct ipoib_dev_priv *priv, int frags, in ipoib_cm_dma_unmap_rx() argument 90 for (i = 0; i < frags; ++i) in ipoib_cm_dma_unmap_rx() 143 int id, int frags, in ipoib_cm_alloc_rx_skb() argument 168 for (i = 0; i < frags; i++) { in ipoib_cm_alloc_rx_skb() 541 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags() 569 int frags; in ipoib_cm_handle_rx_wc() local 643 frags = PAGE_ALIGN(wc->byte_len - in ipoib_cm_handle_rx_wc() 647 newskb = ipoib_cm_alloc_rx_skb(dev, rx_ring, wr_id, frags, in ipoib_cm_handle_rx_wc() 659 ipoib_cm_dma_unmap_rx(priv, frags, rx_ring[wr_id].mapping); in ipoib_cm_handle_rx_wc() 660 memcpy(rx_ring[wr_id].mapping, mapping, (frags + 1) * sizeof(*mapping)); in ipoib_cm_handle_rx_wc()
|
/drivers/net/ethernet/neterion/vxge/ |
D | vxge-traffic.c | 1446 priv->frags = 0; in vxge_hw_fifo_txdl_reserve() 1486 txdp = (struct vxge_hw_fifo_txd *)txdlh + txdl_priv->frags; in vxge_hw_fifo_txdl_buffer_set() 1496 if (txdl_priv->frags) { in vxge_hw_fifo_txdl_buffer_set() 1498 (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_buffer_set() 1509 txdl_priv->frags++; in vxge_hw_fifo_txdl_buffer_set() 1531 txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); in vxge_hw_fifo_txdl_post() 1540 txdl_priv->frags - 1, in vxge_hw_fifo_txdl_post()
|
/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
D | hw_atl_b0.h | 52 unsigned int frags);
|
/drivers/net/ethernet/hisilicon/ |
D | hix5hd2_gmac.c | 236 struct frags_info frags[SG_MAX_SKB_FRAGS + 1]; member 581 addr = le32_to_cpu(desc->frags[i].addr); in hix5hd2_clean_sg_desc() 582 len = le32_to_cpu(desc->frags[i].size); in hix5hd2_clean_sg_desc() 721 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in hix5hd2_fill_sg_desc() 728 desc->frags[i].addr = cpu_to_le32(addr); in hix5hd2_fill_sg_desc() 729 desc->frags[i].size = cpu_to_le32(len); in hix5hd2_fill_sg_desc()
|
/drivers/net/ |
D | xen-netfront.c | 332 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_alloc_rx_buffers() 566 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_slots() local 572 for (i = 0; i < frags; i++) { in xennet_count_skb_slots() 573 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_slots() 829 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xennet_start_xmit() 1087 skb_frag_page(&skb_shinfo(skb)->frags[0]), in xennet_get_responses() 1168 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; in xennet_fill_frags() 1331 skb_frag_off_set(&skb_shinfo(skb)->frags[0], rx->offset); in xennet_poll() 1332 skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status); in xennet_poll() 1459 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_release_rx_bufs()
|