Home
last modified time | relevance | path

Searched refs:dma_len (Results 1 – 25 of 82) sorted by relevance

1234

/drivers/parisc/
Diommu-helpers.h22 unsigned long dma_offset = 0, dma_len = 0; in iommu_fill_pdir() local
47 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); in iommu_fill_pdir()
51 dma_len = sg_dma_len(startsg); in iommu_fill_pdir()
106 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ in iommu_coalesce_chunks() local
120 dma_len = startsg->length; in iommu_coalesce_chunks()
149 if (unlikely(ALIGN(dma_len + dma_offset + startsg->length, IOVP_SIZE) > in iommu_coalesce_chunks()
163 dma_len += startsg->length; in iommu_coalesce_chunks()
171 sg_dma_len(contig_sg) = dma_len; in iommu_coalesce_chunks()
172 dma_len = ALIGN(dma_len + dma_offset, IOVP_SIZE); in iommu_coalesce_chunks()
175 | (iommu_alloc_range(ioc, dev, dma_len) << IOVP_SHIFT) in iommu_coalesce_chunks()
/drivers/net/wireless/broadcom/brcm80211/brcmsmac/
Dbrcms_trace_brcmsmac_tx.h76 u16 dma_len),
78 dma_len),
85 __field(u16, dma_len)
93 __entry->dma_len = dma_len;
98 __entry->ampdu_frames, __entry->dma_len)
Dampdu.c509 session->dma_len = 0; in brcms_c_ampdu_reset_session()
568 session->dma_len += p->len; in brcms_c_ampdu_add_frame()
644 u16 dma_len = session->dma_len; in brcms_c_ampdu_finalize() local
771 dma_len = min(dma_len, f->ampdu_pld_size); in brcms_c_ampdu_finalize()
772 txh->PreloadSize = cpu_to_le16(dma_len); in brcms_c_ampdu_finalize()
/drivers/net/ethernet/mellanox/mlx5/core/en/
Dxdp.h165 u32 dma_len = xdptxd->len; in mlx5e_xdp_mpwqe_add_dseg() local
169 if (session->inline_on && dma_len <= MLX5E_XDP_INLINE_WQE_SZ_THRSD) { in mlx5e_xdp_mpwqe_add_dseg()
172 u16 ds_len = sizeof(*inline_dseg) + dma_len; in mlx5e_xdp_mpwqe_add_dseg()
175 inline_dseg->byte_count = cpu_to_be32(dma_len | MLX5_INLINE_SEG); in mlx5e_xdp_mpwqe_add_dseg()
176 memcpy(inline_dseg->data, xdptxd->data, dma_len); in mlx5e_xdp_mpwqe_add_dseg()
184 dseg->byte_count = cpu_to_be32(dma_len); in mlx5e_xdp_mpwqe_add_dseg()
Dxdp.c313 u32 dma_len = xdptxd->len; in mlx5e_xmit_xdp_frame() local
319 if (unlikely(dma_len < MLX5E_XDP_MIN_INLINE || sq->hw_mtu < dma_len)) { in mlx5e_xmit_xdp_frame()
335 dma_len -= MLX5E_XDP_MIN_INLINE; in mlx5e_xmit_xdp_frame()
342 dseg->byte_count = cpu_to_be32(dma_len); in mlx5e_xmit_xdp_frame()
/drivers/i2c/busses/
Di2c-stm32.c84 dma->dma_len = 0; in stm32_i2c_dma_free()
114 dma->dma_len = len; in stm32_i2c_prep_dma_xfer()
117 dma->dma_buf = dma_map_single(chan_dev, buf, dma->dma_len, in stm32_i2c_prep_dma_xfer()
125 dma->dma_len, in stm32_i2c_prep_dma_xfer()
149 dma_unmap_single(chan_dev, dma->dma_buf, dma->dma_len, in stm32_i2c_prep_dma_xfer()
Di2c-imx.c181 unsigned int dma_len; member
354 dma->dma_len, dma->dma_data_dir); in i2c_imx_dma_callback()
367 dma->dma_len, dma->dma_data_dir); in i2c_imx_dma_xfer()
374 dma->dma_len, dma->dma_transfer_dir, in i2c_imx_dma_xfer()
396 dma->dma_len, dma->dma_data_dir); in i2c_imx_dma_xfer()
406 dma->dma_len = 0; in i2c_imx_dma_free()
623 dma->dma_len = msgs->len - 1; in i2c_imx_dma_write()
687 dma->dma_len = msgs->len - 2; in i2c_imx_dma_read()
/drivers/net/ethernet/sfc/
Dtx_tso.c107 unsigned int dma_len; in efx_tx_queue_insert() local
121 dma_len = tx_queue->efx->type->tx_limit_len(tx_queue, in efx_tx_queue_insert()
125 if (dma_len >= len) in efx_tx_queue_insert()
128 buffer->len = dma_len; in efx_tx_queue_insert()
130 dma_addr += dma_len; in efx_tx_queue_insert()
131 len -= dma_len; in efx_tx_queue_insert()
Dtx.c340 unsigned int dma_len; in efx_tx_map_chunk() local
345 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in efx_tx_map_chunk()
347 buffer->len = dma_len; in efx_tx_map_chunk()
350 len -= dma_len; in efx_tx_map_chunk()
351 dma_addr += dma_len; in efx_tx_map_chunk()
/drivers/scsi/
Dam53c974.c243 static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in pci_esp_dma_length_limit() argument
255 if (dma_len > (1U << dma_limit)) in pci_esp_dma_length_limit()
256 dma_len = (1U << dma_limit); in pci_esp_dma_length_limit()
262 end = base + dma_len; in pci_esp_dma_length_limit()
265 dma_len = end - base; in pci_esp_dma_length_limit()
267 return dma_len; in pci_esp_dma_length_limit()
Dzorro_esp.c219 u32 dma_len) in zorro_esp_dma_length_limit() argument
221 return dma_len > (1U << 16) ? (1U << 16) : dma_len; in zorro_esp_dma_length_limit()
225 u32 dma_len) in fastlane_esp_dma_length_limit() argument
228 return dma_len > 0xfffc ? 0xfffc : dma_len; in fastlane_esp_dma_length_limit()
Dmac53c94.c368 u32 dma_len; in set_dma_cmds() local
382 dma_len = sg_dma_len(scl); in set_dma_cmds()
383 if (dma_len > 0xffff) in set_dma_cmds()
385 total += dma_len; in set_dma_cmds()
386 dcmds->req_count = cpu_to_le16(dma_len); in set_dma_cmds()
Desp_scsi.c519 static u32 esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in esp_dma_length_limit() argument
523 if (dma_len > (1U << 24)) in esp_dma_length_limit()
524 dma_len = (1U << 24); in esp_dma_length_limit()
534 if (dma_len > (1U << 16)) in esp_dma_length_limit()
535 dma_len = (1U << 16); in esp_dma_length_limit()
541 end = base + dma_len; in esp_dma_length_limit()
544 dma_len = end - base; in esp_dma_length_limit()
546 return dma_len; in esp_dma_length_limit()
1747 unsigned int dma_len = esp_cur_dma_len(ent, cmd); in esp_process_event() local
1758 dma_len = esp->ops->dma_length_limit(esp, dma_addr, in esp_process_event()
[all …]
Dmac_esp.c247 static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in mac_esp_dma_length_limit() argument
249 return dma_len > 0xFFFF ? 0xFFFF : dma_len; in mac_esp_dma_length_limit()
/drivers/staging/wusbcore/host/whci/
Dqset.c452 size_t dma_len; in qset_add_urb_sg() local
480 dma_len = dma_remaining; in qset_add_urb_sg()
488 if (std->len + dma_len > QTD_MAX_XFER_SIZE) { in qset_add_urb_sg()
489 dma_len = (QTD_MAX_XFER_SIZE / qset->max_packet) in qset_add_urb_sg()
493 std->len += dma_len; in qset_add_urb_sg()
497 ep = dma_addr + dma_len; in qset_add_urb_sg()
517 dma_remaining -= dma_len; in qset_add_urb_sg()
518 remaining -= dma_len; in qset_add_urb_sg()
/drivers/net/ethernet/chelsio/cxgb/
Dsge.c165 DEFINE_DMA_UNMAP_LEN(dma_len);
171 DEFINE_DMA_UNMAP_LEN(dma_len);
513 dma_unmap_len(ce, dma_len), in free_freelQ_buffers()
626 if (likely(dma_unmap_len(ce, dma_len))) { in free_cmdQ_buffers()
628 dma_unmap_len(ce, dma_len), in free_cmdQ_buffers()
829 unsigned int dma_len = q->rx_buffer_size - q->dma_offset; in refill_free_list() local
840 mapping = pci_map_single(pdev, skb->data, dma_len, in refill_free_list()
846 dma_unmap_len_set(ce, dma_len, dma_len); in refill_free_list()
849 e->len_gen = V_CMD_LEN(dma_len) | V_CMD_GEN1(q->genbit); in refill_free_list()
1054 dma_unmap_len(ce, dma_len), in get_packet()
[all …]
/drivers/misc/mei/
Dclient.c1557 u32 dma_len; in mei_cl_irq_write() local
1601 } else if (dr_slots && hbuf_len >= hdr_len + sizeof(dma_len)) { in mei_cl_irq_write()
1608 mei_hdr.length = sizeof(dma_len); in mei_cl_irq_write()
1609 dma_len = len; in mei_cl_irq_write()
1610 data = &dma_len; in mei_cl_irq_write()
1665 u32 dma_len; in mei_cl_write() local
1728 } else if (dr_slots && hbuf_len >= hdr_len + sizeof(dma_len)) { in mei_cl_write()
1735 mei_hdr.length = sizeof(dma_len); in mei_cl_write()
1736 dma_len = len; in mei_cl_write()
1737 data = &dma_len; in mei_cl_write()
/drivers/net/wireless/mediatek/mt7601u/
Ddma.c116 u16 dma_len = get_unaligned_le16(data); in mt7601u_rx_next_seg_len() local
119 WARN_ON_ONCE(!dma_len) || in mt7601u_rx_next_seg_len()
120 WARN_ON_ONCE(dma_len + MT_DMA_HDRS > data_len) || in mt7601u_rx_next_seg_len()
121 WARN_ON_ONCE(dma_len & 0x3)) in mt7601u_rx_next_seg_len()
124 return MT_DMA_HDRS + dma_len; in mt7601u_rx_next_seg_len()
/drivers/net/ethernet/sfc/falcon/
Dtx.c186 unsigned int dma_len; in ef4_tx_map_chunk() local
191 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in ef4_tx_map_chunk()
193 buffer->len = dma_len; in ef4_tx_map_chunk()
196 len -= dma_len; in ef4_tx_map_chunk()
197 dma_addr += dma_len; in ef4_tx_map_chunk()
/drivers/crypto/
Dtalitos.c335 return ((struct talitos_desc *)(edesc->buf + edesc->dma_len))->hdr1; in get_request_hdr()
495 (edesc->buf + edesc->dma_len))->hdr;
1011 if (edesc->dma_len) in ipsec_esp_unmap()
1012 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, in ipsec_esp_unmap()
1062 oicv = edesc->buf + edesc->dma_len; in ipsec_esp_decrypt_swauth_done()
1217 dma_addr_t dma_icv = edesc->dma_link_tbl + edesc->dma_len - authsize; in ipsec_esp()
1307 edesc->dma_len, in ipsec_esp()
1334 int src_nents, dst_nents, alloc_len, dma_len, src_len, dst_len; in talitos_edesc_alloc() local
1382 dma_len = (src_nents ? src_len : 0) + in talitos_edesc_alloc()
1385 dma_len = (src_nents + dst_nents + 2) * in talitos_edesc_alloc()
[all …]
/drivers/ata/
Dpata_arasan_cf.c434 u32 xfer_cnt, sglen, dma_len, xfer_ctr; in sg_xfer() local
474 dma_len = min(xfer_cnt, FIFO_SIZE); in sg_xfer()
475 ret = dma_xfer(acdev, src, dest, dma_len); in sg_xfer()
482 src += dma_len; in sg_xfer()
484 dest += dma_len; in sg_xfer()
486 sglen -= dma_len; in sg_xfer()
487 xfer_cnt -= dma_len; in sg_xfer()
/drivers/block/rsxx/
Ddma.c633 unsigned int dma_len, in rsxx_queue_dma() argument
649 dma->sub_page.cnt = (dma_len >> 9); in rsxx_queue_dma()
680 unsigned int dma_len; in rsxx_dma_queue_bio() local
720 dma_len = min(bv_len, in rsxx_dma_queue_bio()
725 dma_off, dma_len, in rsxx_dma_queue_bio()
733 addr8 += dma_len; in rsxx_dma_queue_bio()
734 bv_off += dma_len; in rsxx_dma_queue_bio()
735 bv_len -= dma_len; in rsxx_dma_queue_bio()
/drivers/mmc/host/
Domap_hsmmc.c165 unsigned int dma_len; member
183 unsigned int dma_len; member
1203 int dma_len; in omap_hsmmc_pre_dma_transfer() local
1215 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, in omap_hsmmc_pre_dma_transfer()
1219 dma_len = host->next_data.dma_len; in omap_hsmmc_pre_dma_transfer()
1220 host->next_data.dma_len = 0; in omap_hsmmc_pre_dma_transfer()
1224 if (dma_len == 0) in omap_hsmmc_pre_dma_transfer()
1228 next->dma_len = dma_len; in omap_hsmmc_pre_dma_transfer()
1231 host->dma_len = dma_len; in omap_hsmmc_pre_dma_transfer()
Dtifm_sd.c263 unsigned int dma_len, dma_blk_cnt, dma_off; in tifm_sd_set_dma_data() local
279 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]) - host->block_pos; in tifm_sd_set_dma_data()
280 if (!dma_len) { in tifm_sd_set_dma_data()
285 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]); in tifm_sd_set_dma_data()
288 if (dma_len < t_size) { in tifm_sd_set_dma_data()
289 dma_blk_cnt = dma_len / r_data->blksz; in tifm_sd_set_dma_data()
300 else if (dma_len) { in tifm_sd_set_dma_data()
/drivers/net/ethernet/synopsys/
Ddwc-xlgmac-net.c996 skb = napi_alloc_skb(napi, desc_data->rx.hdr.dma_len); in xlgmac_create_skb()
1005 desc_data->rx.hdr.dma_len, in xlgmac_create_skb()
1011 copy_len = min(desc_data->rx.hdr.dma_len, copy_len); in xlgmac_create_skb()
1021 desc_data->rx.buf.dma_len, in xlgmac_create_skb()
1027 len, desc_data->rx.buf.dma_len); in xlgmac_create_skb()
1201 desc_data->rx.buf.dma_len, in xlgmac_rx_poll()
1209 desc_data->rx.buf.dma_len); in xlgmac_rx_poll()

1234