Home
last modified time | relevance | path

Searched refs:dma_addr (Results 1 – 25 of 547) sorted by relevance

12345678910>>...22

/drivers/net/ethernet/cisco/enic/
Denic_res.h44 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_ex() argument
55 (u64)dma_addr | VNIC_PADDR_TARGET, in enic_queue_wq_desc_ex()
65 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop, desc_skip_cnt, in enic_queue_wq_desc_ex()
70 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_cont() argument
73 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_cont()
79 dma_addr_t dma_addr, unsigned int len, int vlan_tag_insert, in enic_queue_wq_desc() argument
82 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc()
89 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum() argument
93 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum()
101 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum_l4() argument
[all …]
/drivers/net/ethernet/apm/xgene-v2/
Dring.c24 next_dma = ring->dma_addr + (offset * XGENE_ENET_DESC_SIZE); in xge_setup_desc()
37 dma_addr_t dma_addr = ring->dma_addr; in xge_update_tx_desc_addr() local
39 xge_wr_csr(pdata, DMATXDESCL, dma_addr); in xge_update_tx_desc_addr()
40 xge_wr_csr(pdata, DMATXDESCH, upper_32_bits(dma_addr)); in xge_update_tx_desc_addr()
49 dma_addr_t dma_addr = ring->dma_addr; in xge_update_rx_desc_addr() local
51 xge_wr_csr(pdata, DMARXDESCL, dma_addr); in xge_update_rx_desc_addr()
52 xge_wr_csr(pdata, DMARXDESCH, upper_32_bits(dma_addr)); in xge_update_rx_desc_addr()
Dmain.c74 dma_addr_t dma_addr; in xge_refill_buffers() local
86 dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE); in xge_refill_buffers()
87 if (dma_mapping_error(dev, dma_addr)) { in xge_refill_buffers()
94 ring->pkt_info[tail].dma_addr = dma_addr; in xge_refill_buffers()
101 upper_32_bits(dma_addr))); in xge_refill_buffers()
104 raw_desc->m0 = cpu_to_le64(SET_BITS(PKT_ADDRL, dma_addr) | in xge_refill_buffers()
178 static dma_addr_t dma_addr; in xge_start_xmit() local
195 pkt_buf = dma_alloc_coherent(dev, XGENE_ENET_STD_MTU, &dma_addr, in xge_start_xmit()
208 upper_32_bits(dma_addr))); in xge_start_xmit()
211 tx_ring->pkt_info[tail].dma_addr = dma_addr; in xge_start_xmit()
[all …]
/drivers/net/ethernet/qualcomm/emac/
Demac-mac.c305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
315 writel(upper_32_bits(adpt->rx_q.rfd.dma_addr), in emac_mac_dma_rings_config()
318 writel(lower_32_bits(adpt->rx_q.rfd.dma_addr), in emac_mac_dma_rings_config()
320 writel(lower_32_bits(adpt->rx_q.rrd.dma_addr), in emac_mac_dma_rings_config()
598 if (tpbuf->dma_addr) { in emac_tx_q_descs_free()
600 tpbuf->dma_addr, tpbuf->length, in emac_tx_q_descs_free()
602 tpbuf->dma_addr = 0; in emac_tx_q_descs_free()
635 if (rfbuf->dma_addr) { in emac_rx_q_free_descs()
636 dma_unmap_single(dev, rfbuf->dma_addr, rfbuf->length, in emac_rx_q_free_descs()
[all …]
/drivers/net/ethernet/sfc/
Dtx_tso.c65 dma_addr_t dma_addr; member
103 dma_addr_t dma_addr, unsigned int len, in efx_tx_queue_insert() argument
119 buffer->dma_addr = dma_addr; in efx_tx_queue_insert()
122 dma_addr, len); in efx_tx_queue_insert()
130 dma_addr += dma_len; in efx_tx_queue_insert()
175 dma_addr_t dma_addr; in tso_start() local
198 dma_addr = dma_map_single(dma_dev, skb->data, in tso_start()
200 st->header_dma_addr = dma_addr; in tso_start()
202 st->dma_addr = dma_addr + header_len; in tso_start()
205 return unlikely(dma_mapping_error(dma_dev, dma_addr)) ? -ENOMEM : 0; in tso_start()
[all …]
Dtx_common.c153 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer()
290 dma_addr_t dma_addr, size_t len) in efx_tx_map_chunk() argument
301 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in efx_tx_map_chunk()
306 buffer->dma_addr = dma_addr; in efx_tx_map_chunk()
309 dma_addr += dma_len; in efx_tx_map_chunk()
337 dma_addr_t dma_addr, unmap_addr; in efx_tx_map_data() local
346 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); in efx_tx_map_data()
349 unmap_addr = dma_addr; in efx_tx_map_data()
351 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in efx_tx_map_data()
362 efx_tx_map_chunk(tx_queue, dma_addr, header_len); in efx_tx_map_data()
[all …]
Dmcdi_functions.c80 dma_addr_t dma_addr; in efx_mcdi_ev_init() local
115 dma_addr = channel->eventq.buf.dma_addr; in efx_mcdi_ev_init()
117 MCDI_SET_ARRAY_QWORD(inbuf, INIT_EVQ_IN_DMA_ADDR, i, dma_addr); in efx_mcdi_ev_init()
118 dma_addr += EFX_BUF_SIZE; in efx_mcdi_ev_init()
172 dma_addr_t dma_addr; in efx_mcdi_tx_init() local
185 dma_addr = tx_queue->txd.buf.dma_addr; in efx_mcdi_tx_init()
188 tx_queue->queue, entries, (u64)dma_addr); in efx_mcdi_tx_init()
191 MCDI_SET_ARRAY_QWORD(inbuf, INIT_TXQ_IN_DMA_ADDR, i, dma_addr); in efx_mcdi_tx_init()
192 dma_addr += EFX_BUF_SIZE; in efx_mcdi_tx_init()
285 dma_addr_t dma_addr; in efx_mcdi_rx_init() local
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en/
Dxdp.c65 dma_addr_t dma_addr; in mlx5e_xmit_xdp_buff() local
89 dma_addr = dma_map_single(sq->pdev, xdptxd.data, xdptxd.len, in mlx5e_xmit_xdp_buff()
91 if (dma_mapping_error(sq->pdev, dma_addr)) { in mlx5e_xmit_xdp_buff()
96 xdptxd.dma_addr = dma_addr; in mlx5e_xmit_xdp_buff()
98 xdpi.frame.dma_addr = dma_addr; in mlx5e_xmit_xdp_buff()
108 dma_addr = di->addr + (xdpf->data - (void *)xdpf); in mlx5e_xmit_xdp_buff()
109 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd.len, in mlx5e_xmit_xdp_buff()
112 xdptxd.dma_addr = dma_addr; in mlx5e_xmit_xdp_buff()
323 dma_addr_t dma_addr = xdptxd->dma_addr; in mlx5e_xmit_xdp_frame() local
347 dma_addr += MLX5E_XDP_MIN_INLINE; in mlx5e_xmit_xdp_frame()
[all …]
/drivers/bus/mhi/host/
Dboot.c33 bhi_vec->dma_addr = mhi_buf->dma_addr; in mhi_rddm_prepare()
40 upper_32_bits(mhi_buf->dma_addr)); in mhi_rddm_prepare()
43 lower_32_bits(mhi_buf->dma_addr)); in mhi_rddm_prepare()
53 &mhi_buf->dma_addr, mhi_buf->len, sequence_id); in mhi_rddm_prepare()
198 upper_32_bits(mhi_buf->dma_addr)); in mhi_fw_load_bhie()
201 lower_32_bits(mhi_buf->dma_addr)); in mhi_fw_load_bhie()
227 dma_addr_t dma_addr, in mhi_fw_load_bhi() argument
257 upper_32_bits(dma_addr)); in mhi_fw_load_bhi()
259 lower_32_bits(dma_addr)); in mhi_fw_load_bhi()
306 mhi_buf->buf, mhi_buf->dma_addr); in mhi_free_bhie_table()
[all …]
/drivers/media/platform/mtk-jpeg/
Dmtk_jpeg_enc_hw.c59 dma_addr_t dma_addr; in mtk_jpeg_set_enc_src() local
62 dma_addr = vb2_dma_contig_plane_dma_addr(src_buf, i) + in mtk_jpeg_set_enc_src()
65 writel(dma_addr, base + JPEG_ENC_SRC_LUMA_ADDR); in mtk_jpeg_set_enc_src()
67 writel(dma_addr, base + JPEG_ENC_SRC_CHROMA_ADDR); in mtk_jpeg_set_enc_src()
74 dma_addr_t dma_addr; in mtk_jpeg_set_enc_dst() local
79 dma_addr = vb2_dma_contig_plane_dma_addr(dst_buf, 0); in mtk_jpeg_set_enc_dst()
81 dma_addr_offsetmask = dma_addr & JPEG_ENC_DST_ADDR_OFFSET_MASK; in mtk_jpeg_set_enc_dst()
86 writel(dma_addr & ~0xf, base + JPEG_ENC_DST_ADDR0); in mtk_jpeg_set_enc_dst()
87 writel((dma_addr + size) & ~0xf, base + JPEG_ENC_STALL_ADDR0); in mtk_jpeg_set_enc_dst()
/drivers/media/pci/solo6x10/
Dsolo6x10-p2m.c32 dma_addr_t dma_addr; in solo_p2m_dma() local
40 dma_addr = dma_map_single(&solo_dev->pdev->dev, sys_addr, size, in solo_p2m_dma()
42 if (dma_mapping_error(&solo_dev->pdev->dev, dma_addr)) in solo_p2m_dma()
45 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, in solo_p2m_dma()
48 dma_unmap_single(&solo_dev->pdev->dev, dma_addr, size, in solo_p2m_dma()
93 desc[1].dma_addr); in solo_p2m_dma_desc()
125 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_fill_desc() argument
128 WARN_ON_ONCE(dma_addr & 0x03); in solo_p2m_fill_desc()
141 desc->dma_addr = dma_addr; in solo_p2m_fill_desc()
146 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_dma_t() argument
[all …]
/drivers/net/ethernet/ti/
Dk3-cppi-desc-pool.c19 dma_addr_t dma_addr; member
38 pool->dma_addr); in k3_cppi_desc_pool_destroy()
77 &pool->dma_addr, GFP_KERNEL); in k3_cppi_desc_pool_create_name()
83 (phys_addr_t)pool->dma_addr, pool->mem_size, in k3_cppi_desc_pool_create_name()
94 pool->dma_addr); in k3_cppi_desc_pool_create_name()
105 return addr ? pool->dma_addr + (addr - pool->cpumem) : 0; in k3_cppi_desc_pool_virt2dma()
110 return dma ? pool->cpumem + (dma - pool->dma_addr) : NULL; in k3_cppi_desc_pool_dma2virt()
/drivers/tty/serial/cpm_uart/
Dcpm_uart_cpm1.c67 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local
86 dma_addr = (u32)cpm_dpram_phys(mem_addr); in cpm_uart_allocbuf()
88 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf()
100 pinfo->dma_addr = dma_addr; /* physical address*/ in cpm_uart_allocbuf()
119 pinfo->dma_addr); in cpm_uart_freebuf()
Dcpm_uart_cpm2.c102 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local
120 dma_addr = virt_to_bus(mem_addr); in cpm_uart_allocbuf()
123 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf()
135 pinfo->dma_addr = dma_addr; in cpm_uart_allocbuf()
154 pinfo->dma_addr); in cpm_uart_freebuf()
/drivers/net/ethernet/sfc/falcon/
Dtx.c38 buffer->dma_addr = page_buf->dma_addr + offset; in ef4_tx_get_copy_buffer()
58 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in ef4_dequeue_buffer()
181 dma_addr_t dma_addr, in ef4_tx_map_chunk() argument
191 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in ef4_tx_map_chunk()
194 buffer->dma_addr = dma_addr; in ef4_tx_map_chunk()
197 dma_addr += dma_len; in ef4_tx_map_chunk()
211 dma_addr_t dma_addr, unmap_addr; in ef4_tx_map_data() local
220 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); in ef4_tx_map_data()
223 unmap_addr = dma_addr; in ef4_tx_map_data()
225 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in ef4_tx_map_data()
[all …]
/drivers/net/ethernet/pensando/ionic/
Dionic_txrx.c60 buf_info->dma_addr = dma_map_page(dev, page, 0, in ionic_rx_page_alloc()
62 if (unlikely(dma_mapping_error(dev, buf_info->dma_addr))) { in ionic_rx_page_alloc()
91 dma_unmap_page(dev, buf_info->dma_addr, IONIC_PAGE_SIZE, DMA_FROM_DEVICE); in ionic_rx_page_free()
159 buf_info->dma_addr + buf_info->page_offset, in ionic_rx_frags()
167 dma_unmap_page(dev, buf_info->dma_addr, in ionic_rx_frags()
209 dma_sync_single_for_cpu(dev, buf_info->dma_addr + buf_info->page_offset, in ionic_rx_copybreak()
212 dma_sync_single_for_device(dev, buf_info->dma_addr + buf_info->page_offset, in ionic_rx_copybreak()
391 desc->addr = cpu_to_le64(buf_info->dma_addr + buf_info->page_offset); in ionic_rx_fill()
411 sg_elem->addr = cpu_to_le64(buf_info->dma_addr + buf_info->page_offset); in ionic_rx_fill()
610 dma_addr_t dma_addr; in ionic_tx_map_single() local
[all …]
/drivers/crypto/caam/
Dkey_gen.c50 dma_addr_t dma_addr; in gen_split_key() local
75 dma_addr = dma_map_single(jrdev, key_out, local_max, DMA_BIDIRECTIONAL); in gen_split_key()
76 if (dma_mapping_error(jrdev, dma_addr)) { in gen_split_key()
82 append_key(desc, dma_addr, keylen, CLASS_2 | KEY_DEST_CLASS_REG); in gen_split_key()
100 append_fifo_store(desc, dma_addr, adata->keylen, in gen_split_key()
121 dma_unmap_single(jrdev, dma_addr, local_max, DMA_BIDIRECTIONAL); in gen_split_key()
/drivers/net/ethernet/freescale/dpaa2/
Ddpaa2-eth-trace.h100 dma_addr_t dma_addr,
106 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid),
116 __field(dma_addr_t, dma_addr)
128 __entry->dma_addr = dma_addr;
141 &__entry->dma_addr,
/drivers/xen/
Dswiotlb-xen.c73 dma_addr_t dma_addr) in xen_dma_to_phys() argument
75 return xen_bus_to_phys(dev, dma_to_phys(dev, dma_addr)); in xen_dma_to_phys()
92 static int is_xen_swiotlb_buffer(struct device *dev, dma_addr_t dma_addr) in is_xen_swiotlb_buffer() argument
94 unsigned long bfn = XEN_PFN_DOWN(dma_to_phys(dev, dma_addr)); in is_xen_swiotlb_buffer()
437 xen_swiotlb_sync_single_for_cpu(struct device *dev, dma_addr_t dma_addr, in xen_swiotlb_sync_single_for_cpu() argument
440 phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu()
443 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_cpu()
446 xen_dma_sync_for_cpu(dev, dma_addr, size, dir); in xen_swiotlb_sync_single_for_cpu()
449 if (is_xen_swiotlb_buffer(dev, dma_addr)) in xen_swiotlb_sync_single_for_cpu()
454 xen_swiotlb_sync_single_for_device(struct device *dev, dma_addr_t dma_addr, in xen_swiotlb_sync_single_for_device() argument
[all …]
/drivers/usb/musb/
Dtusb6010_omap.c41 dma_addr_t dma_addr; member
138 buf = phys_to_virt((u32)chdat->dma_addr) + chdat->transfer_len; in tusb_omap_dma_cb()
140 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb()
145 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb()
182 u8 rndis_mode, dma_addr_t dma_addr, u32 len) in tusb_omap_dma_program() argument
202 if (unlikely(dma_addr & 0x1) || (len < 32) || (len > packet_sz)) in tusb_omap_dma_program()
211 if (dma_addr & 0x2) in tusb_omap_dma_program()
256 chdat->dma_addr = dma_addr; in tusb_omap_dma_program()
262 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program()
266 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program()
[all …]
/drivers/vdpa/vdpa_sim/
Dvdpa_sim.c148 dma_addr_t dma_addr; in vdpasim_map_range() local
157 dma_addr = iova_dma_addr(&vdpasim->iova, iova); in vdpasim_map_range()
160 ret = vhost_iotlb_add_range(vdpasim->iommu, (u64)dma_addr, in vdpasim_map_range()
161 (u64)dma_addr + size - 1, (u64)paddr, perm); in vdpasim_map_range()
169 return dma_addr; in vdpasim_map_range()
172 static void vdpasim_unmap_range(struct vdpasim *vdpasim, dma_addr_t dma_addr, in vdpasim_unmap_range() argument
176 vhost_iotlb_del_range(vdpasim->iommu, (u64)dma_addr, in vdpasim_unmap_range()
177 (u64)dma_addr + size - 1); in vdpasim_unmap_range()
180 free_iova(&vdpasim->iova, iova_pfn(&vdpasim->iova, dma_addr)); in vdpasim_unmap_range()
198 static void vdpasim_unmap_page(struct device *dev, dma_addr_t dma_addr, in vdpasim_unmap_page() argument
[all …]
/drivers/crypto/marvell/octeontx2/
Dotx2_cptvf_reqmgr.c31 (void *) req->in[i].dma_addr); in otx2_cpt_dump_sg_list()
41 (void *) req->out[i].dma_addr); in otx2_cpt_dump_sg_list()
103 list[i].dma_addr = dma_map_single(&pdev->dev, list[i].vptr, in setup_sgio_components()
106 if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) { in setup_sgio_components()
119 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components()
120 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components()
121 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components()
122 sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr); in setup_sgio_components()
130 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components()
134 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components()
[all …]
/drivers/crypto/cavium/cpt/
Dcptvf_reqmanager.c59 list[i].dma_addr = dma_map_single(&pdev->dev, in setup_sgio_components()
64 list[i].dma_addr))) { in setup_sgio_components()
80 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components()
81 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components()
82 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components()
83 sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr); in setup_sgio_components()
92 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components()
96 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components()
100 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components()
110 if (list[j].dma_addr) { in setup_sgio_components()
[all …]
/drivers/scsi/fnic/
Dfnic_res.h31 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_desc() argument
40 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_desc()
51 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); in fnic_queue_wq_desc()
55 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_eth_desc() argument
64 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_eth_desc()
76 vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1); in fnic_queue_wq_eth_desc()
224 void *os_buf, dma_addr_t dma_addr, in fnic_queue_rq_desc() argument
230 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_rq_desc()
234 vnic_rq_post(rq, os_buf, 0, dma_addr, len); in fnic_queue_rq_desc()
/drivers/media/platform/mtk-vcodec/
Dmtk_vcodec_util.c43 mem->va = dma_alloc_coherent(dev, size, &mem->dma_addr, GFP_KERNEL); in mtk_vcodec_mem_alloc()
52 (unsigned long)mem->dma_addr); in mtk_vcodec_mem_alloc()
74 (unsigned long)mem->dma_addr); in mtk_vcodec_mem_free()
77 dma_free_coherent(dev, size, mem->va, mem->dma_addr); in mtk_vcodec_mem_free()
79 mem->dma_addr = 0; in mtk_vcodec_mem_free()

12345678910>>...22