Home
last modified time | relevance | path

Searched refs:vaddr (Results 1 – 25 of 284) sorted by relevance

12345678910>>...12

/drivers/media/v4l2-core/
Dvideobuf2-vmalloc.c25 void *vaddr; member
47 buf->vaddr = vmalloc_user(buf->size); in vb2_vmalloc_alloc()
53 if (!buf->vaddr) { in vb2_vmalloc_alloc()
68 vfree(buf->vaddr); in vb2_vmalloc_put()
73 static void *vb2_vmalloc_get_userptr(struct device *dev, unsigned long vaddr, in vb2_vmalloc_get_userptr() argument
87 offset = vaddr & ~PAGE_MASK; in vb2_vmalloc_get_userptr()
89 vec = vb2_create_framevec(vaddr, size, dma_dir == DMA_FROM_DEVICE); in vb2_vmalloc_get_userptr()
106 buf->vaddr = (__force void *) in vb2_vmalloc_get_userptr()
109 buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1, in vb2_vmalloc_get_userptr()
113 if (!buf->vaddr) in vb2_vmalloc_get_userptr()
[all …]
Dvideobuf-vmalloc.c105 __func__, i, mem->vaddr); in videobuf_vm_close()
107 vfree(mem->vaddr); in videobuf_vm_close()
108 mem->vaddr = NULL; in videobuf_vm_close()
173 if (!mem->vaddr) { in __videobuf_iolock()
192 mem->vaddr = vmalloc_user(pages); in __videobuf_iolock()
193 if (!mem->vaddr) { in __videobuf_iolock()
198 mem->vaddr, pages); in __videobuf_iolock()
257 mem->vaddr = vmalloc_user(pages); in __videobuf_mmap_mapper()
258 if (!mem->vaddr) { in __videobuf_mmap_mapper()
262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); in __videobuf_mmap_mapper()
[all …]
Dvideobuf2-dma-sg.c35 void *vaddr; member
114 buf->vaddr = NULL; in vb2_dma_sg_alloc()
184 if (buf->vaddr) in vb2_dma_sg_put()
185 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put()
220 static void *vb2_dma_sg_get_userptr(struct device *dev, unsigned long vaddr, in vb2_dma_sg_get_userptr() argument
235 buf->vaddr = NULL; in vb2_dma_sg_get_userptr()
238 buf->offset = vaddr & ~PAGE_MASK; in vb2_dma_sg_get_userptr()
241 vec = vb2_create_framevec(vaddr, size, buf->dma_dir == DMA_FROM_DEVICE); in vb2_dma_sg_get_userptr()
290 if (buf->vaddr) in vb2_dma_sg_put_userptr()
291 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put_userptr()
[all …]
Dvideobuf-dma-contig.c28 void *vaddr; member
45 mem->vaddr = dma_alloc_coherent(dev, mem->size, in __videobuf_dc_alloc()
48 if (!mem->vaddr) { in __videobuf_dc_alloc()
53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); in __videobuf_dc_alloc()
61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); in __videobuf_dc_free()
63 mem->vaddr = NULL; in __videobuf_dc_free()
117 i, mem->vaddr); in videobuf_vm_close()
120 mem->vaddr = NULL; in videobuf_vm_close()
233 return mem->vaddr; in __videobuf_to_vaddr()
250 if (!mem->vaddr) { in __videobuf_iolock()
[all …]
Dvideobuf2-dma-contig.c26 void *vaddr; member
79 if (!buf->vaddr && buf->db_attach) in vb2_dc_vaddr()
80 buf->vaddr = dma_buf_vmap(buf->db_attach->dmabuf); in vb2_dc_vaddr()
82 return buf->vaddr; in vb2_dc_vaddr()
162 buf->vaddr = buf->cookie; in vb2_dc_alloc()
338 return buf->vaddr ? buf->vaddr + pgnum * PAGE_SIZE : NULL; in vb2_dc_dmabuf_ops_kmap()
345 return buf->vaddr; in vb2_dc_dmabuf_ops_vmap()
476 static void *vb2_dc_get_userptr(struct device *dev, unsigned long vaddr, in vb2_dc_get_userptr() argument
489 if (!IS_ALIGNED(vaddr | size, dma_align)) { in vb2_dc_get_userptr()
509 offset = vaddr & ~PAGE_MASK; in vb2_dc_get_userptr()
[all …]
/drivers/lguest/
Dpage_tables.c83 static pgd_t *spgd_addr(struct lg_cpu *cpu, u32 i, unsigned long vaddr) in spgd_addr() argument
85 unsigned int index = pgd_index(vaddr); in spgd_addr()
97 static pmd_t *spmd_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spmd_addr() argument
99 unsigned int index = pmd_index(vaddr); in spmd_addr()
115 static pte_t *spte_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spte_addr() argument
118 pmd_t *pmd = spmd_addr(cpu, spgd, vaddr); in spte_addr()
129 return &page[pte_index(vaddr)]; in spte_addr()
136 static unsigned long gpgd_addr(struct lg_cpu *cpu, unsigned long vaddr) in gpgd_addr() argument
138 unsigned int index = vaddr >> (PGDIR_SHIFT); in gpgd_addr()
144 static unsigned long gpmd_addr(pgd_t gpgd, unsigned long vaddr) in gpmd_addr() argument
[all …]
/drivers/net/ethernet/freescale/fman/
Dfman_muram.c47 unsigned long vaddr) in fman_muram_vbase_to_offset() argument
49 return vaddr - (unsigned long)muram->vbase; in fman_muram_vbase_to_offset()
68 void __iomem *vaddr; in fman_muram_init() local
81 vaddr = ioremap(base, size); in fman_muram_init()
82 if (!vaddr) { in fman_muram_init()
87 ret = gen_pool_add_virt(muram->pool, (unsigned long)vaddr, in fman_muram_init()
91 iounmap(vaddr); in fman_muram_init()
95 memset_io(vaddr, 0, (int)size); in fman_muram_init()
97 muram->vbase = vaddr; in fman_muram_init()
134 unsigned long vaddr; in fman_muram_alloc() local
[all …]
/drivers/base/
Ddma-mapping.c21 void *vaddr; member
29 dma_free_coherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_coherent_release()
36 dma_free_noncoherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_noncoherent_release()
43 if (this->vaddr == match->vaddr) { in dmam_match()
68 void *vaddr; in dmam_alloc_coherent() local
74 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent()
75 if (!vaddr) { in dmam_alloc_coherent()
80 dr->vaddr = vaddr; in dmam_alloc_coherent()
86 return vaddr; in dmam_alloc_coherent()
99 void dmam_free_coherent(struct device *dev, size_t size, void *vaddr, in dmam_free_coherent() argument
[all …]
/drivers/block/
Dcciss.h66 void __iomem *vaddr; member
226 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command()
227 readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command()
243 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
244 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
249 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
250 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
263 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask()
264 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask()
269 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask()
[all …]
Dsmart1,2.h44 writel(c->busaddr, h->vaddr + S42XX_REQUEST_PORT_OFFSET); in smart4_submit_command()
56 writel(0, h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask()
60 h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask()
72 return (!readl(h->vaddr + S42XX_REQUEST_PORT_OFFSET)); in smart4_fifo_full()
82 = readl(h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed()
90 writel(0, h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed()
103 readl(h->vaddr + S42XX_INTR_STATUS); in smart4_intr_pending()
123 writel(c->busaddr, h->vaddr + COMMAND_FIFO); in smart2_submit_command()
128 writel(val, h->vaddr + INTR_MASK); in smart2_intr_mask()
133 return readl(h->vaddr + COMMAND_FIFO); in smart2_fifo_full()
[all …]
/drivers/misc/sgi-gru/
Dgrufault.c61 struct vm_area_struct *gru_find_vma(unsigned long vaddr) in gru_find_vma() argument
65 vma = find_vma(current->mm, vaddr); in gru_find_vma()
66 if (vma && vma->vm_start <= vaddr && vma->vm_ops == &gru_vm_ops) in gru_find_vma()
79 static struct gru_thread_state *gru_find_lock_gts(unsigned long vaddr) in gru_find_lock_gts() argument
86 vma = gru_find_vma(vaddr); in gru_find_lock_gts()
88 gts = gru_find_thread_state(vma, TSID(vaddr, vma)); in gru_find_lock_gts()
96 static struct gru_thread_state *gru_alloc_locked_gts(unsigned long vaddr) in gru_alloc_locked_gts() argument
103 vma = gru_find_vma(vaddr); in gru_alloc_locked_gts()
107 gts = gru_alloc_thread_state(vma, TSID(vaddr, vma)); in gru_alloc_locked_gts()
191 unsigned long vaddr, int write, in non_atomic_pte_lookup() argument
[all …]
Dgruhandles.c152 unsigned long vaddr, unsigned long vaddrmask, in tgh_invalidate() argument
156 tgh->vaddr = vaddr; in tgh_invalidate()
170 unsigned long vaddr, int asid, int dirty, in tfh_write_only() argument
174 tfh->fillvaddr = vaddr; in tfh_write_only()
186 unsigned long vaddr, int asid, int dirty, in tfh_write_restart() argument
190 tfh->fillvaddr = vaddr; in tfh_write_restart()
/drivers/hwtracing/coresight/
Dcoresight-tmc-etr.c37 void __iomem *vaddr; member
47 memset(drvdata->vaddr, 0, drvdata->size); in tmc_etr_enable_hw()
91 drvdata->buf = drvdata->vaddr + rwp - drvdata->paddr; in tmc_etr_dump_hw()
94 drvdata->buf = drvdata->vaddr; in tmc_etr_dump_hw()
120 void __iomem *vaddr = NULL; in tmc_enable_etr_sink_sysfs() local
130 if (!drvdata->vaddr) { in tmc_enable_etr_sink_sysfs()
138 vaddr = dma_alloc_coherent(drvdata->dev, drvdata->size, in tmc_enable_etr_sink_sysfs()
140 if (!vaddr) in tmc_enable_etr_sink_sysfs()
167 drvdata->vaddr = vaddr; in tmc_enable_etr_sink_sysfs()
169 drvdata->buf = drvdata->vaddr; in tmc_enable_etr_sink_sysfs()
[all …]
/drivers/scsi/
Dhpsa.h166 void __iomem *vaddr; member
415 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command()
416 (void) readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command()
422 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_no_read()
428 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_ioaccel2()
440 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
441 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
445 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
446 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
454 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_performant_intr_mask()
[all …]
Dsun3_scsi.c309 unsigned char *vaddr; in sun3scsi_dma_finish() local
311 vaddr = (unsigned char *)dvma_vmetov(sun3_dma_orig_addr); in sun3scsi_dma_finish()
313 vaddr += (sun3_dma_orig_count - fifo); in sun3scsi_dma_finish()
314 vaddr--; in sun3scsi_dma_finish()
318 *vaddr = (dregs->bpack_lo & 0xff00) >> 8; in sun3scsi_dma_finish()
319 vaddr--; in sun3scsi_dma_finish()
322 *vaddr = (dregs->bpack_hi & 0x00ff); in sun3scsi_dma_finish()
323 vaddr--; in sun3scsi_dma_finish()
326 *vaddr = (dregs->bpack_hi & 0xff00) >> 8; in sun3scsi_dma_finish()
358 unsigned char *vaddr; in sun3scsi_dma_finish() local
[all …]
/drivers/xen/xenbus/
Dxenbus_client.c78 void **vaddr);
79 int (*unmap)(struct xenbus_device *dev, void *vaddr);
380 int xenbus_grant_ring(struct xenbus_device *dev, void *vaddr, in xenbus_grant_ring() argument
388 virt_to_gfn(vaddr), 0); in xenbus_grant_ring()
396 vaddr = vaddr + XEN_PAGE_SIZE; in xenbus_grant_ring()
470 unsigned int nr_grefs, void **vaddr) in xenbus_map_ring_valloc() argument
472 return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr); in xenbus_map_ring_valloc()
544 void **vaddr) in xenbus_map_ring_valloc_pv() argument
554 *vaddr = NULL; in xenbus_map_ring_valloc_pv()
586 *vaddr = area->addr; in xenbus_map_ring_valloc_pv()
[all …]
/drivers/staging/android/ion/
Dion_test.c69 void *vaddr = vmap(&page, 1, VM_MAP, pgprot); in ion_handle_test_dma() local
73 if (!vaddr) { in ion_handle_test_dma()
79 ret = copy_from_user(vaddr + offset, ptr, to_copy); in ion_handle_test_dma()
81 ret = copy_to_user(ptr, vaddr + offset, to_copy); in ion_handle_test_dma()
83 vunmap(vaddr); in ion_handle_test_dma()
119 void *vaddr = dma_buf_kmap(dma_buf, page_offset); in ion_handle_test_kernel() local
121 if (!vaddr) in ion_handle_test_kernel()
127 ret = copy_from_user(vaddr + copy_offset, ptr, to_copy); in ion_handle_test_kernel()
129 ret = copy_to_user(ptr, vaddr + copy_offset, to_copy); in ion_handle_test_kernel()
131 dma_buf_kunmap(dma_buf, page_offset, vaddr); in ion_handle_test_kernel()
/drivers/acpi/apei/
Dghes.c154 unsigned long vaddr; in ghes_ioremap_pfn_nmi() local
156 vaddr = (unsigned long)GHES_IOREMAP_NMI_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_nmi()
157 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_nmi()
160 return (void __iomem *)vaddr; in ghes_ioremap_pfn_nmi()
165 unsigned long vaddr, paddr; in ghes_ioremap_pfn_irq() local
168 vaddr = (unsigned long)GHES_IOREMAP_IRQ_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_irq()
173 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, paddr, prot); in ghes_ioremap_pfn_irq()
175 return (void __iomem *)vaddr; in ghes_ioremap_pfn_irq()
180 unsigned long vaddr = (unsigned long __force)vaddr_ptr; in ghes_iounmap_nmi() local
183 BUG_ON(vaddr != (unsigned long)GHES_IOREMAP_NMI_PAGE(base)); in ghes_iounmap_nmi()
[all …]
/drivers/gpu/drm/i915/
Di915_gem_execbuffer.c324 unsigned long vaddr; member
333 cache->vaddr = 0; in reloc_cache_init()
353 void *vaddr; in reloc_cache_fini() local
355 if (!cache->vaddr) in reloc_cache_fini()
358 vaddr = unmask_page(cache->vaddr); in reloc_cache_fini()
359 if (cache->vaddr & KMAP) { in reloc_cache_fini()
360 if (cache->vaddr & CLFLUSH_AFTER) in reloc_cache_fini()
363 kunmap_atomic(vaddr); in reloc_cache_fini()
367 io_mapping_unmap_atomic((void __iomem *)vaddr); in reloc_cache_fini()
386 void *vaddr; in reloc_kmap() local
[all …]
/drivers/net/wireless/ath/ath10k/
Dhtt_tx.c91 ar->htt.tx_q_state.vaddr->count[tid][peer_id] = count; in __ath10k_htt_tx_txq_recalc()
92 ar->htt.tx_q_state.vaddr->map[tid][idx] &= ~bit; in __ath10k_htt_tx_txq_recalc()
93 ar->htt.tx_q_state.vaddr->map[tid][idx] |= count ? bit : 0; in __ath10k_htt_tx_txq_recalc()
112 seq = le32_to_cpu(ar->htt.tx_q_state.vaddr->seq); in __ath10k_htt_tx_txq_sync()
114 ar->htt.tx_q_state.vaddr->seq = cpu_to_le32(seq); in __ath10k_htt_tx_txq_sync()
119 size = sizeof(*ar->htt.tx_q_state.vaddr); in __ath10k_htt_tx_txq_sync()
236 if (!htt->frag_desc.vaddr) in ath10k_htt_tx_free_cont_frag_desc()
243 htt->frag_desc.vaddr, in ath10k_htt_tx_free_cont_frag_desc()
256 htt->frag_desc.vaddr = dma_alloc_coherent(ar->dev, size, in ath10k_htt_tx_alloc_cont_frag_desc()
259 if (!htt->frag_desc.vaddr) { in ath10k_htt_tx_alloc_cont_frag_desc()
[all …]
/drivers/i2c/busses/
Di2c-ibm_iic.c85 volatile struct iic_regs __iomem *iic = dev->vaddr; in dump_iic_regs()
129 out_8(&dev->vaddr->intmsk, enable ? INTRMSK_EIMTC : 0); in iic_interrupt_mode()
137 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_init()
182 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_reset()
243 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_smbus_quick()
329 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_handler()
347 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_result()
380 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_abort_xfer()
412 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_wait_for_tc()
463 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_bytes()
[all …]
/drivers/net/wireless/ath/ath6kl/
Dtarget.h333 #define AR6003_VTOP(vaddr) ((vaddr) & 0x001fffff) argument
334 #define AR6004_VTOP(vaddr) (vaddr) argument
336 #define TARG_VTOP(target_type, vaddr) \ argument
337 (((target_type) == TARGET_TYPE_AR6003) ? AR6003_VTOP(vaddr) : \
338 (((target_type) == TARGET_TYPE_AR6004) ? AR6004_VTOP(vaddr) : 0))
/drivers/media/platform/coda/
Dcoda-jpeg.c172 coda_memcpy_parabuf(ctx->parabuf.vaddr, huff + i); in coda_jpeg_write_tables()
176 coda_memcpy_parabuf(ctx->parabuf.vaddr, qmat + i); in coda_jpeg_write_tables()
183 void *vaddr = vb2_plane_vaddr(vb, 0); in coda_jpeg_check_buffer() local
187 soi = be16_to_cpup((__be16 *)vaddr); in coda_jpeg_check_buffer()
192 vaddr += len - 2; in coda_jpeg_check_buffer()
194 eoi = be16_to_cpup((__be16 *)(vaddr - i)); in coda_jpeg_check_buffer()
/drivers/vfio/
Dvfio_iommu_type1.c75 unsigned long vaddr; /* Process virtual addr */ member
210 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument
216 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { in vaddr_get_pfn()
223 vma = find_vma_intersection(current->mm, vaddr, vaddr + 1); in vaddr_get_pfn()
226 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn()
241 static long vfio_pin_pages(unsigned long vaddr, long npage, in vfio_pin_pages() argument
252 ret = vaddr_get_pfn(vaddr, prot, pfn_base); in vfio_pin_pages()
269 for (vaddr += PAGE_SIZE; i < npage; i++, vaddr += PAGE_SIZE) { in vfio_pin_pages()
270 ret = vaddr_get_pfn(vaddr, prot, &pfn); in vfio_pin_pages()
551 unsigned long vaddr = map->vaddr; in vfio_dma_do_map() local
[all …]
/drivers/gpu/drm/
Ddrm_gem_cma_helper.c112 cma_obj->vaddr = dma_alloc_wc(drm->dev, size, &cma_obj->paddr, in drm_gem_cma_create()
114 if (!cma_obj->vaddr) { in drm_gem_cma_create()
188 if (cma_obj->vaddr) { in drm_gem_cma_free_object()
190 cma_obj->vaddr, cma_obj->paddr); in drm_gem_cma_free_object()
321 ret = dma_mmap_wc(cma_obj->base.dev->dev, vma, cma_obj->vaddr, in drm_gem_cma_mmap_obj()
380 off, &cma_obj->paddr, cma_obj->vaddr, obj->size); in drm_gem_cma_describe()
409 ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr, in drm_gem_cma_prime_get_sg_table()
508 return cma_obj->vaddr; in drm_gem_cma_prime_vmap()
523 void drm_gem_cma_prime_vunmap(struct drm_gem_object *obj, void *vaddr) in drm_gem_cma_prime_vunmap() argument

12345678910>>...12