/drivers/net/ethernet/cisco/enic/ |
D | enic_res.h | 44 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_ex() argument 55 (u64)dma_addr | VNIC_PADDR_TARGET, in enic_queue_wq_desc_ex() 65 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop, desc_skip_cnt, in enic_queue_wq_desc_ex() 70 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_cont() argument 73 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_cont() 79 dma_addr_t dma_addr, unsigned int len, int vlan_tag_insert, in enic_queue_wq_desc() argument 82 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc() 89 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum() argument 93 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum() 101 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum_l4() argument [all …]
|
/drivers/net/ethernet/apm/xgene-v2/ |
D | ring.c | 24 next_dma = ring->dma_addr + (offset * XGENE_ENET_DESC_SIZE); in xge_setup_desc() 37 dma_addr_t dma_addr = ring->dma_addr; in xge_update_tx_desc_addr() local 39 xge_wr_csr(pdata, DMATXDESCL, dma_addr); in xge_update_tx_desc_addr() 40 xge_wr_csr(pdata, DMATXDESCH, upper_32_bits(dma_addr)); in xge_update_tx_desc_addr() 49 dma_addr_t dma_addr = ring->dma_addr; in xge_update_rx_desc_addr() local 51 xge_wr_csr(pdata, DMARXDESCL, dma_addr); in xge_update_rx_desc_addr() 52 xge_wr_csr(pdata, DMARXDESCH, upper_32_bits(dma_addr)); in xge_update_rx_desc_addr()
|
D | main.c | 74 dma_addr_t dma_addr; in xge_refill_buffers() local 86 dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE); in xge_refill_buffers() 87 if (dma_mapping_error(dev, dma_addr)) { in xge_refill_buffers() 94 ring->pkt_info[tail].dma_addr = dma_addr; in xge_refill_buffers() 101 upper_32_bits(dma_addr))); in xge_refill_buffers() 104 raw_desc->m0 = cpu_to_le64(SET_BITS(PKT_ADDRL, dma_addr) | in xge_refill_buffers() 178 static dma_addr_t dma_addr; in xge_start_xmit() local 195 pkt_buf = dma_alloc_coherent(dev, XGENE_ENET_STD_MTU, &dma_addr, in xge_start_xmit() 208 upper_32_bits(dma_addr))); in xge_start_xmit() 211 tx_ring->pkt_info[tail].dma_addr = dma_addr; in xge_start_xmit() [all …]
|
/drivers/net/ethernet/qualcomm/emac/ |
D | emac-mac.c | 305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config() 308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config() 315 writel(upper_32_bits(adpt->rx_q.rfd.dma_addr), in emac_mac_dma_rings_config() 318 writel(lower_32_bits(adpt->rx_q.rfd.dma_addr), in emac_mac_dma_rings_config() 320 writel(lower_32_bits(adpt->rx_q.rrd.dma_addr), in emac_mac_dma_rings_config() 598 if (tpbuf->dma_addr) { in emac_tx_q_descs_free() 600 tpbuf->dma_addr, tpbuf->length, in emac_tx_q_descs_free() 602 tpbuf->dma_addr = 0; in emac_tx_q_descs_free() 635 if (rfbuf->dma_addr) { in emac_rx_q_free_descs() 636 dma_unmap_single(dev, rfbuf->dma_addr, rfbuf->length, in emac_rx_q_free_descs() [all …]
|
/drivers/net/ethernet/sfc/ |
D | tx_tso.c | 65 dma_addr_t dma_addr; member 103 dma_addr_t dma_addr, unsigned int len, in efx_tx_queue_insert() argument 119 buffer->dma_addr = dma_addr; in efx_tx_queue_insert() 122 dma_addr, len); in efx_tx_queue_insert() 130 dma_addr += dma_len; in efx_tx_queue_insert() 175 dma_addr_t dma_addr; in tso_start() local 198 dma_addr = dma_map_single(dma_dev, skb->data, in tso_start() 200 st->header_dma_addr = dma_addr; in tso_start() 202 st->dma_addr = dma_addr + header_len; in tso_start() 205 return unlikely(dma_mapping_error(dma_dev, dma_addr)) ? -ENOMEM : 0; in tso_start() [all …]
|
D | tx_common.c | 153 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer() 290 dma_addr_t dma_addr, size_t len) in efx_tx_map_chunk() argument 301 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in efx_tx_map_chunk() 306 buffer->dma_addr = dma_addr; in efx_tx_map_chunk() 309 dma_addr += dma_len; in efx_tx_map_chunk() 337 dma_addr_t dma_addr, unmap_addr; in efx_tx_map_data() local 346 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); in efx_tx_map_data() 349 unmap_addr = dma_addr; in efx_tx_map_data() 351 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in efx_tx_map_data() 362 efx_tx_map_chunk(tx_queue, dma_addr, header_len); in efx_tx_map_data() [all …]
|
D | mcdi_functions.c | 80 dma_addr_t dma_addr; in efx_mcdi_ev_init() local 115 dma_addr = channel->eventq.buf.dma_addr; in efx_mcdi_ev_init() 117 MCDI_SET_ARRAY_QWORD(inbuf, INIT_EVQ_IN_DMA_ADDR, i, dma_addr); in efx_mcdi_ev_init() 118 dma_addr += EFX_BUF_SIZE; in efx_mcdi_ev_init() 172 dma_addr_t dma_addr; in efx_mcdi_tx_init() local 185 dma_addr = tx_queue->txd.buf.dma_addr; in efx_mcdi_tx_init() 188 tx_queue->queue, entries, (u64)dma_addr); in efx_mcdi_tx_init() 191 MCDI_SET_ARRAY_QWORD(inbuf, INIT_TXQ_IN_DMA_ADDR, i, dma_addr); in efx_mcdi_tx_init() 192 dma_addr += EFX_BUF_SIZE; in efx_mcdi_tx_init() 285 dma_addr_t dma_addr; in efx_mcdi_rx_init() local [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | xdp.c | 65 dma_addr_t dma_addr; in mlx5e_xmit_xdp_buff() local 89 dma_addr = dma_map_single(sq->pdev, xdptxd.data, xdptxd.len, in mlx5e_xmit_xdp_buff() 91 if (dma_mapping_error(sq->pdev, dma_addr)) { in mlx5e_xmit_xdp_buff() 96 xdptxd.dma_addr = dma_addr; in mlx5e_xmit_xdp_buff() 98 xdpi.frame.dma_addr = dma_addr; in mlx5e_xmit_xdp_buff() 108 dma_addr = di->addr + (xdpf->data - (void *)xdpf); in mlx5e_xmit_xdp_buff() 109 dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd.len, in mlx5e_xmit_xdp_buff() 112 xdptxd.dma_addr = dma_addr; in mlx5e_xmit_xdp_buff() 323 dma_addr_t dma_addr = xdptxd->dma_addr; in mlx5e_xmit_xdp_frame() local 347 dma_addr += MLX5E_XDP_MIN_INLINE; in mlx5e_xmit_xdp_frame() [all …]
|
/drivers/bus/mhi/host/ |
D | boot.c | 33 bhi_vec->dma_addr = mhi_buf->dma_addr; in mhi_rddm_prepare() 40 upper_32_bits(mhi_buf->dma_addr)); in mhi_rddm_prepare() 43 lower_32_bits(mhi_buf->dma_addr)); in mhi_rddm_prepare() 53 &mhi_buf->dma_addr, mhi_buf->len, sequence_id); in mhi_rddm_prepare() 198 upper_32_bits(mhi_buf->dma_addr)); in mhi_fw_load_bhie() 201 lower_32_bits(mhi_buf->dma_addr)); in mhi_fw_load_bhie() 227 dma_addr_t dma_addr, in mhi_fw_load_bhi() argument 257 upper_32_bits(dma_addr)); in mhi_fw_load_bhi() 259 lower_32_bits(dma_addr)); in mhi_fw_load_bhi() 306 mhi_buf->buf, mhi_buf->dma_addr); in mhi_free_bhie_table() [all …]
|
/drivers/media/platform/mtk-jpeg/ |
D | mtk_jpeg_enc_hw.c | 59 dma_addr_t dma_addr; in mtk_jpeg_set_enc_src() local 62 dma_addr = vb2_dma_contig_plane_dma_addr(src_buf, i) + in mtk_jpeg_set_enc_src() 65 writel(dma_addr, base + JPEG_ENC_SRC_LUMA_ADDR); in mtk_jpeg_set_enc_src() 67 writel(dma_addr, base + JPEG_ENC_SRC_CHROMA_ADDR); in mtk_jpeg_set_enc_src() 74 dma_addr_t dma_addr; in mtk_jpeg_set_enc_dst() local 79 dma_addr = vb2_dma_contig_plane_dma_addr(dst_buf, 0); in mtk_jpeg_set_enc_dst() 81 dma_addr_offsetmask = dma_addr & JPEG_ENC_DST_ADDR_OFFSET_MASK; in mtk_jpeg_set_enc_dst() 86 writel(dma_addr & ~0xf, base + JPEG_ENC_DST_ADDR0); in mtk_jpeg_set_enc_dst() 87 writel((dma_addr + size) & ~0xf, base + JPEG_ENC_STALL_ADDR0); in mtk_jpeg_set_enc_dst()
|
/drivers/media/pci/solo6x10/ |
D | solo6x10-p2m.c | 32 dma_addr_t dma_addr; in solo_p2m_dma() local 40 dma_addr = dma_map_single(&solo_dev->pdev->dev, sys_addr, size, in solo_p2m_dma() 42 if (dma_mapping_error(&solo_dev->pdev->dev, dma_addr)) in solo_p2m_dma() 45 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, in solo_p2m_dma() 48 dma_unmap_single(&solo_dev->pdev->dev, dma_addr, size, in solo_p2m_dma() 93 desc[1].dma_addr); in solo_p2m_dma_desc() 125 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_fill_desc() argument 128 WARN_ON_ONCE(dma_addr & 0x03); in solo_p2m_fill_desc() 141 desc->dma_addr = dma_addr; in solo_p2m_fill_desc() 146 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_dma_t() argument [all …]
|
/drivers/net/ethernet/ti/ |
D | k3-cppi-desc-pool.c | 19 dma_addr_t dma_addr; member 38 pool->dma_addr); in k3_cppi_desc_pool_destroy() 77 &pool->dma_addr, GFP_KERNEL); in k3_cppi_desc_pool_create_name() 83 (phys_addr_t)pool->dma_addr, pool->mem_size, in k3_cppi_desc_pool_create_name() 94 pool->dma_addr); in k3_cppi_desc_pool_create_name() 105 return addr ? pool->dma_addr + (addr - pool->cpumem) : 0; in k3_cppi_desc_pool_virt2dma() 110 return dma ? pool->cpumem + (dma - pool->dma_addr) : NULL; in k3_cppi_desc_pool_dma2virt()
|
/drivers/tty/serial/cpm_uart/ |
D | cpm_uart_cpm1.c | 67 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local 86 dma_addr = (u32)cpm_dpram_phys(mem_addr); in cpm_uart_allocbuf() 88 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf() 100 pinfo->dma_addr = dma_addr; /* physical address*/ in cpm_uart_allocbuf() 119 pinfo->dma_addr); in cpm_uart_freebuf()
|
D | cpm_uart_cpm2.c | 102 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local 120 dma_addr = virt_to_bus(mem_addr); in cpm_uart_allocbuf() 123 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf() 135 pinfo->dma_addr = dma_addr; in cpm_uart_allocbuf() 154 pinfo->dma_addr); in cpm_uart_freebuf()
|
/drivers/net/ethernet/sfc/falcon/ |
D | tx.c | 38 buffer->dma_addr = page_buf->dma_addr + offset; in ef4_tx_get_copy_buffer() 58 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in ef4_dequeue_buffer() 181 dma_addr_t dma_addr, in ef4_tx_map_chunk() argument 191 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in ef4_tx_map_chunk() 194 buffer->dma_addr = dma_addr; in ef4_tx_map_chunk() 197 dma_addr += dma_len; in ef4_tx_map_chunk() 211 dma_addr_t dma_addr, unmap_addr; in ef4_tx_map_data() local 220 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); in ef4_tx_map_data() 223 unmap_addr = dma_addr; in ef4_tx_map_data() 225 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in ef4_tx_map_data() [all …]
|
/drivers/net/ethernet/pensando/ionic/ |
D | ionic_txrx.c | 60 buf_info->dma_addr = dma_map_page(dev, page, 0, in ionic_rx_page_alloc() 62 if (unlikely(dma_mapping_error(dev, buf_info->dma_addr))) { in ionic_rx_page_alloc() 91 dma_unmap_page(dev, buf_info->dma_addr, IONIC_PAGE_SIZE, DMA_FROM_DEVICE); in ionic_rx_page_free() 159 buf_info->dma_addr + buf_info->page_offset, in ionic_rx_frags() 167 dma_unmap_page(dev, buf_info->dma_addr, in ionic_rx_frags() 209 dma_sync_single_for_cpu(dev, buf_info->dma_addr + buf_info->page_offset, in ionic_rx_copybreak() 212 dma_sync_single_for_device(dev, buf_info->dma_addr + buf_info->page_offset, in ionic_rx_copybreak() 391 desc->addr = cpu_to_le64(buf_info->dma_addr + buf_info->page_offset); in ionic_rx_fill() 411 sg_elem->addr = cpu_to_le64(buf_info->dma_addr + buf_info->page_offset); in ionic_rx_fill() 610 dma_addr_t dma_addr; in ionic_tx_map_single() local [all …]
|
/drivers/crypto/caam/ |
D | key_gen.c | 50 dma_addr_t dma_addr; in gen_split_key() local 75 dma_addr = dma_map_single(jrdev, key_out, local_max, DMA_BIDIRECTIONAL); in gen_split_key() 76 if (dma_mapping_error(jrdev, dma_addr)) { in gen_split_key() 82 append_key(desc, dma_addr, keylen, CLASS_2 | KEY_DEST_CLASS_REG); in gen_split_key() 100 append_fifo_store(desc, dma_addr, adata->keylen, in gen_split_key() 121 dma_unmap_single(jrdev, dma_addr, local_max, DMA_BIDIRECTIONAL); in gen_split_key()
|
/drivers/net/ethernet/freescale/dpaa2/ |
D | dpaa2-eth-trace.h | 100 dma_addr_t dma_addr, 106 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid), 116 __field(dma_addr_t, dma_addr) 128 __entry->dma_addr = dma_addr; 141 &__entry->dma_addr,
|
/drivers/xen/ |
D | swiotlb-xen.c | 73 dma_addr_t dma_addr) in xen_dma_to_phys() argument 75 return xen_bus_to_phys(dev, dma_to_phys(dev, dma_addr)); in xen_dma_to_phys() 92 static int is_xen_swiotlb_buffer(struct device *dev, dma_addr_t dma_addr) in is_xen_swiotlb_buffer() argument 94 unsigned long bfn = XEN_PFN_DOWN(dma_to_phys(dev, dma_addr)); in is_xen_swiotlb_buffer() 437 xen_swiotlb_sync_single_for_cpu(struct device *dev, dma_addr_t dma_addr, in xen_swiotlb_sync_single_for_cpu() argument 440 phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu() 443 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_cpu() 446 xen_dma_sync_for_cpu(dev, dma_addr, size, dir); in xen_swiotlb_sync_single_for_cpu() 449 if (is_xen_swiotlb_buffer(dev, dma_addr)) in xen_swiotlb_sync_single_for_cpu() 454 xen_swiotlb_sync_single_for_device(struct device *dev, dma_addr_t dma_addr, in xen_swiotlb_sync_single_for_device() argument [all …]
|
/drivers/usb/musb/ |
D | tusb6010_omap.c | 41 dma_addr_t dma_addr; member 138 buf = phys_to_virt((u32)chdat->dma_addr) + chdat->transfer_len; in tusb_omap_dma_cb() 140 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb() 145 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb() 182 u8 rndis_mode, dma_addr_t dma_addr, u32 len) in tusb_omap_dma_program() argument 202 if (unlikely(dma_addr & 0x1) || (len < 32) || (len > packet_sz)) in tusb_omap_dma_program() 211 if (dma_addr & 0x2) in tusb_omap_dma_program() 256 chdat->dma_addr = dma_addr; in tusb_omap_dma_program() 262 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program() 266 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program() [all …]
|
/drivers/vdpa/vdpa_sim/ |
D | vdpa_sim.c | 148 dma_addr_t dma_addr; in vdpasim_map_range() local 157 dma_addr = iova_dma_addr(&vdpasim->iova, iova); in vdpasim_map_range() 160 ret = vhost_iotlb_add_range(vdpasim->iommu, (u64)dma_addr, in vdpasim_map_range() 161 (u64)dma_addr + size - 1, (u64)paddr, perm); in vdpasim_map_range() 169 return dma_addr; in vdpasim_map_range() 172 static void vdpasim_unmap_range(struct vdpasim *vdpasim, dma_addr_t dma_addr, in vdpasim_unmap_range() argument 176 vhost_iotlb_del_range(vdpasim->iommu, (u64)dma_addr, in vdpasim_unmap_range() 177 (u64)dma_addr + size - 1); in vdpasim_unmap_range() 180 free_iova(&vdpasim->iova, iova_pfn(&vdpasim->iova, dma_addr)); in vdpasim_unmap_range() 198 static void vdpasim_unmap_page(struct device *dev, dma_addr_t dma_addr, in vdpasim_unmap_page() argument [all …]
|
/drivers/crypto/marvell/octeontx2/ |
D | otx2_cptvf_reqmgr.c | 31 (void *) req->in[i].dma_addr); in otx2_cpt_dump_sg_list() 41 (void *) req->out[i].dma_addr); in otx2_cpt_dump_sg_list() 103 list[i].dma_addr = dma_map_single(&pdev->dev, list[i].vptr, in setup_sgio_components() 106 if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) { in setup_sgio_components() 119 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components() 120 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components() 121 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components() 122 sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr); in setup_sgio_components() 130 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components() 134 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components() [all …]
|
/drivers/crypto/cavium/cpt/ |
D | cptvf_reqmanager.c | 59 list[i].dma_addr = dma_map_single(&pdev->dev, in setup_sgio_components() 64 list[i].dma_addr))) { in setup_sgio_components() 80 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components() 81 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components() 82 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components() 83 sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr); in setup_sgio_components() 92 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components() 96 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components() 100 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components() 110 if (list[j].dma_addr) { in setup_sgio_components() [all …]
|
/drivers/scsi/fnic/ |
D | fnic_res.h | 31 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_desc() argument 40 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_desc() 51 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); in fnic_queue_wq_desc() 55 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_eth_desc() argument 64 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_eth_desc() 76 vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1); in fnic_queue_wq_eth_desc() 224 void *os_buf, dma_addr_t dma_addr, in fnic_queue_rq_desc() argument 230 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_rq_desc() 234 vnic_rq_post(rq, os_buf, 0, dma_addr, len); in fnic_queue_rq_desc()
|
/drivers/media/platform/mtk-vcodec/ |
D | mtk_vcodec_util.c | 43 mem->va = dma_alloc_coherent(dev, size, &mem->dma_addr, GFP_KERNEL); in mtk_vcodec_mem_alloc() 52 (unsigned long)mem->dma_addr); in mtk_vcodec_mem_alloc() 74 (unsigned long)mem->dma_addr); in mtk_vcodec_mem_free() 77 dma_free_coherent(dev, size, mem->va, mem->dma_addr); in mtk_vcodec_mem_free() 79 mem->dma_addr = 0; in mtk_vcodec_mem_free()
|