Home
last modified time | relevance | path

Searched refs:dma_addr (Results 1 – 25 of 250) sorted by relevance

12345678910

/drivers/net/ethernet/cisco/enic/
Denic_res.h44 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_ex() argument
52 (u64)dma_addr | VNIC_PADDR_TARGET, in enic_queue_wq_desc_ex()
62 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); in enic_queue_wq_desc_ex()
66 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_cont() argument
69 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_cont()
75 dma_addr_t dma_addr, unsigned int len, int vlan_tag_insert, in enic_queue_wq_desc() argument
78 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc()
85 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum() argument
89 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum()
97 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum_l4() argument
[all …]
Dvnic_wq.h55 dma_addr_t dma_addr; member
104 void *os_buf, dma_addr_t dma_addr, in vnic_wq_post() argument
111 buf->dma_addr = dma_addr; in vnic_wq_post()
/drivers/infiniband/hw/amso1100/
Dc2_alloc.c44 dma_addr_t dma_addr; in c2_alloc_mqsp_chunk() local
47 &dma_addr, gfp_mask); in c2_alloc_mqsp_chunk()
51 new_head->dma_addr = dma_addr; in c2_alloc_mqsp_chunk()
52 dma_unmap_addr_set(new_head, mapping, new_head->dma_addr); in c2_alloc_mqsp_chunk()
90 dma_addr_t *dma_addr, gfp_t gfp_mask) in c2_alloc_mqsp() argument
112 *dma_addr = head->dma_addr + in c2_alloc_mqsp()
116 &(head->shared_ptr[mqsp]), (unsigned long long) *dma_addr); in c2_alloc_mqsp()
/drivers/gpu/drm/exynos/
Dexynos_drm_buf.c49 if (buf->dma_addr) { in lowlevel_buffer_allocate()
80 &buf->dma_addr, GFP_KERNEL); in lowlevel_buffer_allocate()
95 start_addr = buf->dma_addr; in lowlevel_buffer_allocate()
108 (unsigned long)buf->dma_addr, in lowlevel_buffer_allocate()
114 (dma_addr_t)buf->dma_addr); in lowlevel_buffer_allocate()
115 buf->dma_addr = (dma_addr_t)NULL; in lowlevel_buffer_allocate()
139 if (!buf->dma_addr) { in lowlevel_buffer_deallocate()
146 (unsigned long)buf->dma_addr, in lowlevel_buffer_deallocate()
158 (dma_addr_t)buf->dma_addr); in lowlevel_buffer_deallocate()
159 buf->dma_addr = (dma_addr_t)NULL; in lowlevel_buffer_deallocate()
Dexynos_mixer.c46 dma_addr_t dma_addr; member
377 if (!win_data->dma_addr) in vp_video_buffer()
402 luma_addr[0] = win_data->dma_addr; in vp_video_buffer()
405 luma_addr[0] = win_data->dma_addr; in vp_video_buffer()
406 chroma_addr[0] = win_data->dma_addr in vp_video_buffer()
491 dma_addr_t dma_addr; in mixer_graph_buffer() local
513 dma_addr = win_data->dma_addr; in mixer_graph_buffer()
530 dma_addr = dma_addr in mixer_graph_buffer()
568 mixer_reg_write(res, MXR_GRAPHIC_BASE(win), dma_addr); in mixer_graph_buffer()
652 win_data->dma_addr = overlay->dma_addr[0]; in mixer_win_mode_set()
[all …]
/drivers/net/ethernet/sfc/
Dtx.c40 dma_addr_t unmap_addr = (buffer->dma_addr + buffer->len - in efx_dequeue_buffer()
80 dma_addr_t dma_addr; member
105 efx_max_tx_len(struct efx_nic *efx, dma_addr_t dma_addr) in efx_max_tx_len() argument
113 unsigned len = (~dma_addr & (EFX_PAGE_SIZE - 1)) + 1; in efx_max_tx_len()
116 if (EFX_WORKAROUND_5391(efx) && (dma_addr & 0xf)) in efx_max_tx_len()
117 len = min_t(unsigned, len, 512 - (dma_addr & 0xf)); in efx_max_tx_len()
164 dma_addr_t dma_addr, unmap_addr = 0; in efx_enqueue_skb() local
194 dma_addr = pci_map_single(pci_dev, skb->data, len, PCI_DMA_TODEVICE); in efx_enqueue_skb()
198 if (unlikely(pci_dma_mapping_error(pci_dev, dma_addr))) in efx_enqueue_skb()
204 unmap_addr = dma_addr; in efx_enqueue_skb()
[all …]
Drx.c159 rx_buf->dma_addr = pci_map_single(efx->pci_dev, in efx_init_rx_buffers_skb()
163 rx_buf->dma_addr))) { in efx_init_rx_buffers_skb()
194 dma_addr_t dma_addr; in efx_init_rx_buffers_page() local
205 dma_addr = pci_map_page(efx->pci_dev, page, 0, in efx_init_rx_buffers_page()
208 if (unlikely(pci_dma_mapping_error(efx->pci_dev, dma_addr))) { in efx_init_rx_buffers_page()
215 state->dma_addr = dma_addr; in efx_init_rx_buffers_page()
218 dma_addr += sizeof(struct efx_rx_page_state); in efx_init_rx_buffers_page()
224 rx_buf->dma_addr = dma_addr + EFX_PAGE_IP_ALIGN; in efx_init_rx_buffers_page()
236 dma_addr += (PAGE_SIZE >> 1); in efx_init_rx_buffers_page()
257 state->dma_addr, in efx_unmap_rx_buffer()
[all …]
/drivers/usb/host/whci/
Dqset.c234 qtd->page_list_ptr = cpu_to_le64(std->dma_addr); in qset_add_qtds()
304 dma_addr_t dma_addr; in qset_free_std() local
307 dma_addr = le64_to_cpu(std->pl_virt[0].buf_ptr); in qset_free_std()
309 dma_addr = std->dma_addr; in qset_free_std()
311 dma_unmap_single(whc->wusbhc.dev, dma_addr, in qset_free_std()
318 if (std->dma_addr) in qset_free_std()
319 dma_unmap_single(whc->wusbhc.dev, std->dma_addr, in qset_free_std()
360 dma_addr_t dma_addr = std->dma_addr; in qset_fill_page_list() local
371 sp = dma_addr & ~(WHCI_PAGE_SIZE-1); in qset_fill_page_list()
372 ep = dma_addr + std->len; in qset_fill_page_list()
[all …]
Dhw.c51 dma_addr_t dma_addr; in whc_do_gencmd() local
71 dma_addr = whc->gen_cmd_buf_dma; in whc_do_gencmd()
73 dma_addr = 0; in whc_do_gencmd()
79 le_writeq(dma_addr, whc->base + WUSBGENADDR); in whc_do_gencmd()
/drivers/media/video/
Dvideobuf2-dma-contig.c28 dma_addr_t dma_addr; member
46 buf->vaddr = dma_alloc_coherent(conf->dev, size, &buf->dma_addr, in vb2_dma_contig_alloc()
73 buf->dma_addr); in vb2_dma_contig_put()
82 return &buf->dma_addr; in vb2_dma_contig_cookie()
110 return vb2_mmap_pfn_range(vma, buf->dma_addr, buf->size, in vb2_dma_contig_mmap()
119 dma_addr_t dma_addr = 0; in vb2_dma_contig_get_userptr() local
126 ret = vb2_get_contig_userptr(vaddr, size, &vma, &dma_addr); in vb2_dma_contig_get_userptr()
135 buf->dma_addr = dma_addr; in vb2_dma_contig_get_userptr()
/drivers/tty/serial/cpm_uart/
Dcpm_uart_cpm1.c81 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local
100 dma_addr = (u32)cpm_dpram_phys(mem_addr); in cpm_uart_allocbuf()
102 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf()
114 pinfo->dma_addr = dma_addr; /* physical address*/ in cpm_uart_allocbuf()
133 pinfo->dma_addr); in cpm_uart_freebuf()
Dcpm_uart_cpm2.c117 dma_addr_t dma_addr = 0; in cpm_uart_allocbuf() local
135 dma_addr = virt_to_bus(mem_addr); in cpm_uart_allocbuf()
138 mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr, in cpm_uart_allocbuf()
150 pinfo->dma_addr = dma_addr; in cpm_uart_allocbuf()
169 pinfo->dma_addr); in cpm_uart_freebuf()
/drivers/infiniband/hw/ipath/
Dipath_user_sdma.c151 void *kvaddr, dma_addr_t dma_addr) in ipath_user_sdma_init_frag() argument
159 pkt->addr[i].addr = dma_addr; in ipath_user_sdma_init_frag()
166 void *kvaddr, dma_addr_t dma_addr) in ipath_user_sdma_init_header() argument
171 kvaddr, dma_addr); in ipath_user_sdma_init_header()
185 dma_addr_t dma_addr; in ipath_user_sdma_coalesce() local
208 dma_addr = dma_map_page(&dd->pcidev->dev, page, 0, len, in ipath_user_sdma_coalesce()
210 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in ipath_user_sdma_coalesce()
216 dma_addr); in ipath_user_sdma_coalesce()
300 dma_addr_t dma_addr = in ipath_user_sdma_pin_pages() local
305 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in ipath_user_sdma_pin_pages()
[all …]
/drivers/infiniband/hw/qib/
Dqib_user_sdma.c150 void *kvaddr, dma_addr_t dma_addr) in qib_user_sdma_init_frag() argument
158 pkt->addr[i].addr = dma_addr; in qib_user_sdma_init_frag()
165 void *kvaddr, dma_addr_t dma_addr) in qib_user_sdma_init_header() argument
170 kvaddr, dma_addr); in qib_user_sdma_init_header()
185 dma_addr_t dma_addr; in qib_user_sdma_coalesce() local
208 dma_addr = dma_map_page(&dd->pcidev->dev, page, 0, len, in qib_user_sdma_coalesce()
210 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in qib_user_sdma_coalesce()
216 dma_addr); in qib_user_sdma_coalesce()
303 dma_addr_t dma_addr = in qib_user_sdma_pin_pages() local
308 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { in qib_user_sdma_pin_pages()
[all …]
/drivers/usb/musb/
Dtusb6010_omap.c43 dma_addr_t dma_addr; member
176 buf = phys_to_virt((u32)chdat->dma_addr) + chdat->transfer_len; in tusb_omap_dma_cb()
178 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb()
183 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb()
226 u8 rndis_mode, dma_addr_t dma_addr, u32 len) in tusb_omap_dma_program() argument
244 if (unlikely(dma_addr & 0x1) || (len < 32) || (len > packet_sz)) in tusb_omap_dma_program()
253 if (dma_addr & 0x2) in tusb_omap_dma_program()
307 chdat->dma_addr = dma_addr; in tusb_omap_dma_program()
312 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program()
315 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program()
[all …]
Dmusbhsdma.c121 dma_addr_t dma_addr, u32 len) in configure_channel() argument
131 channel, packet_sz, dma_addr, len, mode); in configure_channel()
148 musb_write_hsdma_addr(mbase, bchannel, dma_addr); in configure_channel()
159 dma_addr_t dma_addr, u32 len) in dma_channel_program() argument
168 packet_sz, dma_addr, len, mode); in dma_channel_program()
176 packet_sz, &mode, &dma_addr, &len); in dma_channel_program()
190 if ((musb->hwvers >= MUSB_HWVERS_1800) && (dma_addr % 4)) in dma_channel_program()
194 musb_channel->start_addr = dma_addr; in dma_channel_program()
199 configure_channel(channel, packet_sz, mode, dma_addr, len); in dma_channel_program()
Dux500_dma.c81 dma_addr_t dma_addr, u32 len) in ux500_configure_channel() argument
97 packet_sz, mode, dma_addr, len, ux500_channel->is_tx); in ux500_configure_channel()
102 sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len, in ux500_configure_channel()
103 offset_in_page(dma_addr)); in ux500_configure_channel()
104 sg_dma_address(&sg) = dma_addr; in ux500_configure_channel()
204 dma_addr_t dma_addr, u32 len) in ux500_dma_channel_program() argument
211 if (!ux500_dma_is_compatible(channel, packet_sz, (void *)dma_addr, len)) in ux500_dma_channel_program()
216 ret = ux500_configure_channel(channel, packet_sz, mode, dma_addr, len); in ux500_dma_channel_program()
/drivers/scsi/fnic/
Dfnic_res.h31 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_desc() argument
40 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_desc()
51 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); in fnic_queue_wq_desc()
55 void *os_buf, dma_addr_t dma_addr, in fnic_queue_wq_eth_desc() argument
64 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_wq_eth_desc()
76 vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1); in fnic_queue_wq_eth_desc()
224 void *os_buf, dma_addr_t dma_addr, in fnic_queue_rq_desc() argument
230 (u64)dma_addr | VNIC_PADDR_TARGET, in fnic_queue_rq_desc()
234 vnic_rq_post(rq, os_buf, 0, dma_addr, len); in fnic_queue_rq_desc()
Dvnic_wq.h69 dma_addr_t dma_addr; member
114 void *os_buf, dma_addr_t dma_addr, in vnic_wq_post() argument
121 buf->dma_addr = dma_addr; in vnic_wq_post()
/drivers/staging/media/solo6x10/
Dp2m.c30 dma_addr_t dma_addr; in solo_p2m_dma() local
39 dma_addr = pci_map_single(solo_dev->pdev, sys_addr, size, in solo_p2m_dma()
42 ret = solo_p2m_dma_t(solo_dev, id, wr, dma_addr, ext_addr, size); in solo_p2m_dma()
44 pci_unmap_single(solo_dev->pdev, dma_addr, size, in solo_p2m_dma()
51 dma_addr_t dma_addr, u32 ext_addr, u32 size) in solo_p2m_dma_t() argument
59 solo_p2m_push_desc(&desc[1], wr, dma_addr, ext_addr, size, 0, 0); in solo_p2m_dma_t()
66 void solo_p2m_push_desc(struct p2m_desc *desc, int wr, dma_addr_t dma_addr, in solo_p2m_push_desc() argument
69 desc->ta = cpu_to_le32(dma_addr); in solo_p2m_push_desc()
/drivers/net/ethernet/aeroflex/
Dgreth.c234 u32 dma_addr; in greth_init_rings() local
251 dma_addr = dma_map_single(greth->dev, in greth_init_rings()
256 if (dma_mapping_error(greth->dev, dma_addr)) { in greth_init_rings()
262 greth_write_bd(&rx_bd[i].addr, dma_addr); in greth_init_rings()
279 dma_addr = dma_map_single(greth->dev, in greth_init_rings()
284 if (dma_mapping_error(greth->dev, dma_addr)) { in greth_init_rings()
289 greth_write_bd(&rx_bd[i].addr, dma_addr); in greth_init_rings()
302 dma_addr = dma_map_single(greth->dev, in greth_init_rings()
307 if (dma_mapping_error(greth->dev, dma_addr)) { in greth_init_rings()
312 greth_write_bd(&tx_bd[i].addr, dma_addr); in greth_init_rings()
[all …]
/drivers/message/i2o/
Dmemory.c78 dma_addr_t dma_addr; in i2o_dma_map_single() local
91 dma_addr = dma_map_single(&c->pdev->dev, ptr, size, direction); in i2o_dma_map_single()
92 if (!dma_mapping_error(&c->pdev->dev, dma_addr)) { in i2o_dma_map_single()
101 *mptr++ = cpu_to_le32(i2o_dma_low(dma_addr)); in i2o_dma_map_single()
104 *mptr++ = cpu_to_le32(i2o_dma_high(dma_addr)); in i2o_dma_map_single()
108 return dma_addr; in i2o_dma_map_single()
/drivers/net/ethernet/ibm/
Dibmveth.c165 pool->dma_addr = kmalloc(sizeof(dma_addr_t) * pool->size, GFP_KERNEL); in ibmveth_alloc_buffer_pool()
166 if (!pool->dma_addr) { in ibmveth_alloc_buffer_pool()
175 kfree(pool->dma_addr); in ibmveth_alloc_buffer_pool()
176 pool->dma_addr = NULL; in ibmveth_alloc_buffer_pool()
183 memset(pool->dma_addr, 0, sizeof(dma_addr_t) * pool->size); in ibmveth_alloc_buffer_pool()
216 dma_addr_t dma_addr; in ibmveth_replenish_buffer_pool() local
241 dma_addr = dma_map_single(&adapter->vdev->dev, skb->data, in ibmveth_replenish_buffer_pool()
244 if (dma_mapping_error(&adapter->vdev->dev, dma_addr)) in ibmveth_replenish_buffer_pool()
248 pool->dma_addr[index] = dma_addr; in ibmveth_replenish_buffer_pool()
255 desc.fields.address = dma_addr; in ibmveth_replenish_buffer_pool()
[all …]
/drivers/infiniband/ulp/iser/
Diser_initiator.c159 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc()
176 u64 dma_addr; in iser_alloc_rx_descriptors() local
189 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors()
191 if (ib_dma_mapping_error(device->ib_device, dma_addr)) in iser_alloc_rx_descriptors()
194 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors()
197 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors()
208 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_alloc_rx_descriptors()
228 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_free_rx_descriptors()
461 rx_dma = rx_desc->dma_addr; in iser_rcv_completion()
505 ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr, in iser_snd_completion()
/drivers/atm/
Dfore200e.c91 #define FORE200E_DMA_INDEX(dma_addr, type, index) ((dma_addr) + (index) * sizeof(type)) argument
198 …chunk->dma_addr = fore200e->bus->dma_map(fore200e, chunk->align_addr, chunk->align_size, direction… in fore200e_chunk_alloc()
209 fore200e->bus->dma_unmap(fore200e, chunk->dma_addr, chunk->dma_size, chunk->direction); in fore200e_chunk_free()
428 u32 dma_addr = pci_map_single((struct pci_dev*)fore200e->bus_dev, virt_addr, size, direction); in fore200e_pca_dma_map() local
431 virt_addr, size, direction, dma_addr); in fore200e_pca_dma_map()
433 return dma_addr; in fore200e_pca_dma_map()
438 fore200e_pca_dma_unmap(struct fore200e* fore200e, u32 dma_addr, int size, int direction) in fore200e_pca_dma_unmap() argument
441 dma_addr, size, direction); in fore200e_pca_dma_unmap()
443 pci_unmap_single((struct pci_dev*)fore200e->bus_dev, dma_addr, size, direction); in fore200e_pca_dma_unmap()
448 fore200e_pca_dma_sync_for_cpu(struct fore200e* fore200e, u32 dma_addr, int size, int direction) in fore200e_pca_dma_sync_for_cpu() argument
[all …]

12345678910