/drivers/net/ethernet/cisco/enic/ |
D | enic_res.h | 44 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_ex() argument 52 (u64)dma_addr | VNIC_PADDR_TARGET, in enic_queue_wq_desc_ex() 62 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); in enic_queue_wq_desc_ex() 66 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_cont() argument 69 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_cont() 75 dma_addr_t dma_addr, unsigned int len, int vlan_tag_insert, in enic_queue_wq_desc() argument 78 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc() 85 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum() argument 89 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum() 97 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum_l4() argument [all …]
|
D | vnic_wq.h | 55 dma_addr_t dma_addr; member 104 void *os_buf, dma_addr_t dma_addr, in vnic_wq_post() argument 111 buf->dma_addr = dma_addr; in vnic_wq_post()
|
/drivers/infiniband/hw/amso1100/ |
D | c2_alloc.c | 44 dma_addr_t dma_addr; in c2_alloc_mqsp_chunk() local 47 &dma_addr, gfp_mask); in c2_alloc_mqsp_chunk() 51 new_head->dma_addr = dma_addr; in c2_alloc_mqsp_chunk() 52 dma_unmap_addr_set(new_head, mapping, new_head->dma_addr); in c2_alloc_mqsp_chunk() 90 dma_addr_t *dma_addr, gfp_t gfp_mask) in c2_alloc_mqsp() argument 112 *dma_addr = head->dma_addr + in c2_alloc_mqsp() 116 &(head->shared_ptr[mqsp]), (unsigned long long) *dma_addr); in c2_alloc_mqsp()
|
/drivers/gpu/drm/exynos/ |
D | exynos_drm_buf.c | 49 if (buf->dma_addr) { in lowlevel_buffer_allocate() 80 &buf->dma_addr, GFP_KERNEL); in lowlevel_buffer_allocate() 95 start_addr = buf->dma_addr; in lowlevel_buffer_allocate() 108 (unsigned long)buf->dma_addr, in lowlevel_buffer_allocate() 114 (dma_addr_t)buf->dma_addr); in lowlevel_buffer_allocate() 115 buf->dma_addr = (dma_addr_t)NULL; in lowlevel_buffer_allocate() 139 if (!buf->dma_addr) { in lowlevel_buffer_deallocate() 146 (unsigned long)buf->dma_addr, in lowlevel_buffer_deallocate() 158 (dma_addr_t)buf->dma_addr); in lowlevel_buffer_deallocate() 159 buf->dma_addr = (dma_addr_t)NULL; in lowlevel_buffer_deallocate()
|
D | exynos_mixer.c | 46 dma_addr_t dma_addr; member 377 if (!win_data->dma_addr) in vp_video_buffer() 402 luma_addr[0] = win_data->dma_addr; in vp_video_buffer() 405 luma_addr[0] = win_data->dma_addr; in vp_video_buffer() 406 chroma_addr[0] = win_data->dma_addr in vp_video_buffer() 491 dma_addr_t dma_addr; in mixer_graph_buffer() local 513 dma_addr = win_data->dma_addr; in mixer_graph_buffer() 530 dma_addr = dma_addr in mixer_graph_buffer() 568 mixer_reg_write(res, MXR_GRAPHIC_BASE(win), dma_addr); in mixer_graph_buffer() 652 win_data->dma_addr = overlay->dma_addr[0]; in mixer_win_mode_set() [all …]
|
/drivers/net/ethernet/sfc/ |
D | tx.c | 40 dma_addr_t unmap_addr = (buffer->dma_addr + buffer->len - in efx_dequeue_buffer() 80 dma_addr_t dma_addr; member 105 efx_max_tx_len(struct efx_nic *efx, dma_addr_t dma_addr) in efx_max_tx_len() argument 113 unsigned len = (~dma_addr & (EFX_PAGE_SIZE - 1)) + 1; in efx_max_tx_len() 116 if (EFX_WORKAROUND_5391(efx) && (dma_addr & 0xf)) in efx_max_tx_len() 117 len = min_t(unsigned, len, 512 - (dma_addr & 0xf)); in efx_max_tx_len() 164 dma_addr_t dma_addr, unmap_addr = 0; in efx_enqueue_skb() local 194 dma_addr = pci_map_single(pci_dev, skb->data, len, PCI_DMA_TODEVICE); in efx_enqueue_skb() 198 if (unlikely(pci_dma_mapping_error(pci_dev, dma_addr))) in efx_enqueue_skb() 204 unmap_addr = dma_addr; in efx_enqueue_skb() [all …]
|
D | rx.c | 159 rx_buf->dma_addr = pci_map_single(efx->pci_dev, in efx_init_rx_buffers_skb() 163 rx_buf->dma_addr))) { in efx_init_rx_buffers_skb() 194 dma_addr_t dma_addr; in efx_init_rx_buffers_page() local 205 dma_addr = pci_map_page(efx->pci_dev, page, 0, in efx_init_rx_buffers_page() 208 if (unlikely(pci_dma_mapping_error(efx->pci_dev, dma_addr))) { in efx_init_rx_buffers_page() 215 state->dma_addr = dma_addr; in efx_init_rx_buffers_page() 218 dma_addr += sizeof(struct efx_rx_page_state); in efx_init_rx_buffers_page() 224 rx_buf->dma_addr = dma_addr + EFX_PAGE_IP_ALIGN; in efx_init_rx_buffers_page() 236 dma_addr += (PAGE_SIZE >> 1); in efx_init_rx_buffers_page() 257 state->dma_addr, in efx_unmap_rx_buffer() [all …]
|
/drivers/usb/host/whci/ |
D | qset.c | 234 qtd->page_list_ptr = cpu_to_le64(std->dma_addr); in qset_add_qtds() 304 dma_addr_t dma_addr; in qset_free_std() local 307 dma_addr = le64_to_cpu(std->pl_virt[0].buf_ptr); in qset_free_std() 309 dma_addr = std->dma_addr; in qset_free_std() 311 dma_unmap_single(whc->wusbhc.dev, dma_addr, in qset_free_std() 318 if (std->dma_addr) in qset_free_std() 319 dma_unmap_single(whc->wusbhc.dev, std->dma_addr, in qset_free_std() 360 dma_addr_t dma_addr = std->dma_addr; in qset_fill_page_list() local 371 sp = dma_addr & ~(WHCI_PAGE_SIZE-1); in qset_fill_page_list() 372 ep = dma_addr + std->len; in qset_fill_page_list() [all …]
|
D | hw.c | 51 dma_addr_t dma_addr; in whc_do_gencmd() local 71 dma_addr = whc->gen_cmd_buf_dma; in whc_do_gencmd() 73 dma_addr = 0; in whc_do_gencmd() 79 le_writeq(dma_addr, whc->base + WUSBGENADDR); in whc_do_gencmd()
|
/drivers/media/video/ |
D | videobuf2-dma-contig.c | 28 dma_addr_t dma_addr; member 46 buf->vaddr = dma_alloc_coherent(conf->dev, size, &buf->dma_addr, in vb2_dma_contig_alloc() 73 buf->dma_addr); in vb2_dma_contig_put() 82 return &buf->dma_addr; in vb2_dma_contig_cookie() 110 return vb2_mmap_pfn_range(vma, buf->dma_addr, buf->size, in vb2_dma_contig_mmap() 119 dma_addr_t dma_addr = 0; in vb2_dma_contig_get_userptr() local 126 ret = vb2_get_contig_userptr(vaddr, size, &vma, &dma_addr); in vb2_dma_contig_get_userptr() 135 buf->dma_addr = dma_addr; in vb2_dma_contig_get_userptr()
|
/drivers/tty/serial/cpm_uart/ |
D | cpm_uart_cpm1.c | 81 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local 100 dma_addr = (u32)cpm_dpram_phys(mem_addr); in cpm_uart_allocbuf() 102 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf() 114 pinfo->dma_addr = dma_addr; /* physical address*/ in cpm_uart_allocbuf() 133 pinfo->dma_addr); in cpm_uart_freebuf()
|
D | cpm_uart_cpm2.c | 117 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local 135 dma_addr = virt_to_bus(mem_addr); in cpm_uart_allocbuf() 138 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf() 150 pinfo->dma_addr = dma_addr; in cpm_uart_allocbuf() 169 pinfo->dma_addr); in cpm_uart_freebuf()
|
/drivers/infiniband/hw/ipath/ |
D | ipath_user_sdma.c | 151 void *kvaddr, dma_addr_t dma_addr) in ipath_user_sdma_init_frag() argument 159 pkt->addr[i].addr = dma_addr; in ipath_user_sdma_init_frag() 166 void *kvaddr, dma_addr_t dma_addr) in ipath_user_sdma_init_header() argument 171 kvaddr, dma_addr); in ipath_user_sdma_init_header() 185 dma_addr_t dma_addr; in ipath_user_sdma_coalesce() local 208 dma_addr = dma_map_page(&dd->pcidev->dev, page, 0, len, in ipath_user_sdma_coalesce() 210 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in ipath_user_sdma_coalesce() 216 dma_addr); in ipath_user_sdma_coalesce() 300 dma_addr_t dma_addr = in ipath_user_sdma_pin_pages() local 305 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in ipath_user_sdma_pin_pages() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_user_sdma.c | 150 void *kvaddr, dma_addr_t dma_addr) in qib_user_sdma_init_frag() argument 158 pkt->addr[i].addr = dma_addr; in qib_user_sdma_init_frag() 165 void *kvaddr, dma_addr_t dma_addr) in qib_user_sdma_init_header() argument 170 kvaddr, dma_addr); in qib_user_sdma_init_header() 185 dma_addr_t dma_addr; in qib_user_sdma_coalesce() local 208 dma_addr = dma_map_page(&dd->pcidev->dev, page, 0, len, in qib_user_sdma_coalesce() 210 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in qib_user_sdma_coalesce() 216 dma_addr); in qib_user_sdma_coalesce() 303 dma_addr_t dma_addr = in qib_user_sdma_pin_pages() local 308 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in qib_user_sdma_pin_pages() [all …]
|
/drivers/usb/musb/ |
D | tusb6010_omap.c | 43 dma_addr_t dma_addr; member 176 buf = phys_to_virt((u32)chdat->dma_addr) + chdat->transfer_len; in tusb_omap_dma_cb() 178 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb() 183 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb() 226 u8 rndis_mode, dma_addr_t dma_addr, u32 len) in tusb_omap_dma_program() argument 244 if (unlikely(dma_addr & 0x1) || (len < 32) || (len > packet_sz)) in tusb_omap_dma_program() 253 if (dma_addr & 0x2) in tusb_omap_dma_program() 307 chdat->dma_addr = dma_addr; in tusb_omap_dma_program() 312 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program() 315 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program() [all …]
|
D | musbhsdma.c | 121 dma_addr_t dma_addr, u32 len) in configure_channel() argument 131 channel, packet_sz, dma_addr, len, mode); in configure_channel() 148 musb_write_hsdma_addr(mbase, bchannel, dma_addr); in configure_channel() 159 dma_addr_t dma_addr, u32 len) in dma_channel_program() argument 168 packet_sz, dma_addr, len, mode); in dma_channel_program() 176 packet_sz, &mode, &dma_addr, &len); in dma_channel_program() 190 if ((musb->hwvers >= MUSB_HWVERS_1800) && (dma_addr % 4)) in dma_channel_program() 194 musb_channel->start_addr = dma_addr; in dma_channel_program() 199 configure_channel(channel, packet_sz, mode, dma_addr, len); in dma_channel_program()
|
D | ux500_dma.c | 81 dma_addr_t dma_addr, u32 len) in ux500_configure_channel() argument 97 packet_sz, mode, dma_addr, len, ux500_channel->is_tx); in ux500_configure_channel() 102 sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len, in ux500_configure_channel() 103 offset_in_page(dma_addr)); in ux500_configure_channel() 104 sg_dma_address(&sg) = dma_addr; in ux500_configure_channel() 204 dma_addr_t dma_addr, u32 len) in ux500_dma_channel_program() argument 211 if (!ux500_dma_is_compatible(channel, packet_sz, (void *)dma_addr, len)) in ux500_dma_channel_program() 216 ret = ux500_configure_channel(channel, packet_sz, mode, dma_addr, len); in ux500_dma_channel_program()
|
/drivers/scsi/fnic/ |
D | fnic_res.h | 31 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_desc() argument 40 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_desc() 51 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); in fnic_queue_wq_desc() 55 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_eth_desc() argument 64 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_eth_desc() 76 vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1); in fnic_queue_wq_eth_desc() 224 void *os_buf, dma_addr_t dma_addr, in fnic_queue_rq_desc() argument 230 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_rq_desc() 234 vnic_rq_post(rq, os_buf, 0, dma_addr, len); in fnic_queue_rq_desc()
|
D | vnic_wq.h | 69 dma_addr_t dma_addr; member 114 void *os_buf, dma_addr_t dma_addr, in vnic_wq_post() argument 121 buf->dma_addr = dma_addr; in vnic_wq_post()
|
/drivers/staging/media/solo6x10/ |
D | p2m.c | 30 dma_addr_t dma_addr; in solo_p2m_dma() local 39 dma_addr = pci_map_single(solo_dev->pdev, sys_addr, size, in solo_p2m_dma() 42 ret = solo_p2m_dma_t(solo_dev, id, wr, dma_addr, ext_addr, size); in solo_p2m_dma() 44 pci_unmap_single(solo_dev->pdev, dma_addr, size, in solo_p2m_dma() 51 dma_addr_t dma_addr, u32 ext_addr, u32 size) in solo_p2m_dma_t() argument 59 solo_p2m_push_desc(&desc[1], wr, dma_addr, ext_addr, size, 0, 0); in solo_p2m_dma_t() 66 void solo_p2m_push_desc(struct p2m_desc *desc, int wr, dma_addr_t dma_addr, in solo_p2m_push_desc() argument 69 desc->ta = cpu_to_le32(dma_addr); in solo_p2m_push_desc()
|
/drivers/net/ethernet/aeroflex/ |
D | greth.c | 234 u32 dma_addr; in greth_init_rings() local 251 dma_addr = dma_map_single(greth->dev, in greth_init_rings() 256 if (dma_mapping_error(greth->dev, dma_addr)) { in greth_init_rings() 262 greth_write_bd(&rx_bd[i].addr, dma_addr); in greth_init_rings() 279 dma_addr = dma_map_single(greth->dev, in greth_init_rings() 284 if (dma_mapping_error(greth->dev, dma_addr)) { in greth_init_rings() 289 greth_write_bd(&rx_bd[i].addr, dma_addr); in greth_init_rings() 302 dma_addr = dma_map_single(greth->dev, in greth_init_rings() 307 if (dma_mapping_error(greth->dev, dma_addr)) { in greth_init_rings() 312 greth_write_bd(&tx_bd[i].addr, dma_addr); in greth_init_rings() [all …]
|
/drivers/message/i2o/ |
D | memory.c | 78 dma_addr_t dma_addr; in i2o_dma_map_single() local 91 dma_addr = dma_map_single(&c->pdev->dev, ptr, size, direction); in i2o_dma_map_single() 92 if (!dma_mapping_error(&c->pdev->dev, dma_addr)) { in i2o_dma_map_single() 101 *mptr++ = cpu_to_le32(i2o_dma_low(dma_addr)); in i2o_dma_map_single() 104 *mptr++ = cpu_to_le32(i2o_dma_high(dma_addr)); in i2o_dma_map_single() 108 return dma_addr; in i2o_dma_map_single()
|
/drivers/net/ethernet/ibm/ |
D | ibmveth.c | 165 pool->dma_addr = kmalloc(sizeof(dma_addr_t) * pool->size, GFP_KERNEL); in ibmveth_alloc_buffer_pool() 166 if (!pool->dma_addr) { in ibmveth_alloc_buffer_pool() 175 kfree(pool->dma_addr); in ibmveth_alloc_buffer_pool() 176 pool->dma_addr = NULL; in ibmveth_alloc_buffer_pool() 183 memset(pool->dma_addr, 0, sizeof(dma_addr_t) * pool->size); in ibmveth_alloc_buffer_pool() 216 dma_addr_t dma_addr; in ibmveth_replenish_buffer_pool() local 241 dma_addr = dma_map_single(&adapter->vdev->dev, skb->data, in ibmveth_replenish_buffer_pool() 244 if (dma_mapping_error(&adapter->vdev->dev, dma_addr)) in ibmveth_replenish_buffer_pool() 248 pool->dma_addr[index] = dma_addr; in ibmveth_replenish_buffer_pool() 255 desc.fields.address = dma_addr; in ibmveth_replenish_buffer_pool() [all …]
|
/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 159 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc() 176 u64 dma_addr; in iser_alloc_rx_descriptors() local 189 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors() 191 if (ib_dma_mapping_error(device->ib_device, dma_addr)) in iser_alloc_rx_descriptors() 194 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors() 197 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors() 208 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_alloc_rx_descriptors() 228 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_free_rx_descriptors() 461 rx_dma = rx_desc->dma_addr; in iser_rcv_completion() 505 ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr, in iser_snd_completion()
|
/drivers/atm/ |
D | fore200e.c | 91 #define FORE200E_DMA_INDEX(dma_addr, type, index) ((dma_addr) + (index) * sizeof(type)) argument 198 …chunk->dma_addr = fore200e->bus->dma_map(fore200e, chunk->align_addr, chunk->align_size, direction… in fore200e_chunk_alloc() 209 fore200e->bus->dma_unmap(fore200e, chunk->dma_addr, chunk->dma_size, chunk->direction); in fore200e_chunk_free() 428 u32 dma_addr = pci_map_single((struct pci_dev*)fore200e->bus_dev, virt_addr, size, direction); in fore200e_pca_dma_map() local 431 virt_addr, size, direction, dma_addr); in fore200e_pca_dma_map() 433 return dma_addr; in fore200e_pca_dma_map() 438 fore200e_pca_dma_unmap(struct fore200e* fore200e, u32 dma_addr, int size, int direction) in fore200e_pca_dma_unmap() argument 441 dma_addr, size, direction); in fore200e_pca_dma_unmap() 443 pci_unmap_single((struct pci_dev*)fore200e->bus_dev, dma_addr, size, direction); in fore200e_pca_dma_unmap() 448 fore200e_pca_dma_sync_for_cpu(struct fore200e* fore200e, u32 dma_addr, int size, int direction) in fore200e_pca_dma_sync_for_cpu() argument [all …]
|