/drivers/media/v4l2-core/ |
D | videobuf2-vmalloc.c | 25 void *vaddr; member 47 buf->vaddr = vmalloc_user(buf->size); in vb2_vmalloc_alloc() 53 if (!buf->vaddr) { in vb2_vmalloc_alloc() 68 vfree(buf->vaddr); in vb2_vmalloc_put() 73 static void *vb2_vmalloc_get_userptr(struct device *dev, unsigned long vaddr, in vb2_vmalloc_get_userptr() argument 87 offset = vaddr & ~PAGE_MASK; in vb2_vmalloc_get_userptr() 89 vec = vb2_create_framevec(vaddr, size, dma_dir == DMA_FROM_DEVICE); in vb2_vmalloc_get_userptr() 106 buf->vaddr = (__force void *) in vb2_vmalloc_get_userptr() 109 buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1, in vb2_vmalloc_get_userptr() 113 if (!buf->vaddr) in vb2_vmalloc_get_userptr() [all …]
|
D | videobuf-vmalloc.c | 105 __func__, i, mem->vaddr); in videobuf_vm_close() 107 vfree(mem->vaddr); in videobuf_vm_close() 108 mem->vaddr = NULL; in videobuf_vm_close() 173 if (!mem->vaddr) { in __videobuf_iolock() 192 mem->vaddr = vmalloc_user(pages); in __videobuf_iolock() 193 if (!mem->vaddr) { in __videobuf_iolock() 198 mem->vaddr, pages); in __videobuf_iolock() 257 mem->vaddr = vmalloc_user(pages); in __videobuf_mmap_mapper() 258 if (!mem->vaddr) { in __videobuf_mmap_mapper() 262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); in __videobuf_mmap_mapper() [all …]
|
D | videobuf2-dma-sg.c | 35 void *vaddr; member 114 buf->vaddr = NULL; in vb2_dma_sg_alloc() 184 if (buf->vaddr) in vb2_dma_sg_put() 185 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put() 220 static void *vb2_dma_sg_get_userptr(struct device *dev, unsigned long vaddr, in vb2_dma_sg_get_userptr() argument 235 buf->vaddr = NULL; in vb2_dma_sg_get_userptr() 238 buf->offset = vaddr & ~PAGE_MASK; in vb2_dma_sg_get_userptr() 241 vec = vb2_create_framevec(vaddr, size, buf->dma_dir == DMA_FROM_DEVICE); in vb2_dma_sg_get_userptr() 290 if (buf->vaddr) in vb2_dma_sg_put_userptr() 291 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put_userptr() [all …]
|
D | videobuf-dma-contig.c | 28 void *vaddr; member 45 mem->vaddr = dma_alloc_coherent(dev, mem->size, in __videobuf_dc_alloc() 48 if (!mem->vaddr) { in __videobuf_dc_alloc() 53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); in __videobuf_dc_alloc() 61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); in __videobuf_dc_free() 63 mem->vaddr = NULL; in __videobuf_dc_free() 117 i, mem->vaddr); in videobuf_vm_close() 120 mem->vaddr = NULL; in videobuf_vm_close() 233 return mem->vaddr; in __videobuf_to_vaddr() 250 if (!mem->vaddr) { in __videobuf_iolock() [all …]
|
D | videobuf2-dma-contig.c | 26 void *vaddr; member 79 if (!buf->vaddr && buf->db_attach) in vb2_dc_vaddr() 80 buf->vaddr = dma_buf_vmap(buf->db_attach->dmabuf); in vb2_dc_vaddr() 82 return buf->vaddr; in vb2_dc_vaddr() 162 buf->vaddr = buf->cookie; in vb2_dc_alloc() 338 return buf->vaddr ? buf->vaddr + pgnum * PAGE_SIZE : NULL; in vb2_dc_dmabuf_ops_kmap() 345 return buf->vaddr; in vb2_dc_dmabuf_ops_vmap() 476 static void *vb2_dc_get_userptr(struct device *dev, unsigned long vaddr, in vb2_dc_get_userptr() argument 489 if (!IS_ALIGNED(vaddr | size, dma_align)) { in vb2_dc_get_userptr() 509 offset = vaddr & ~PAGE_MASK; in vb2_dc_get_userptr() [all …]
|
/drivers/lguest/ |
D | page_tables.c | 83 static pgd_t *spgd_addr(struct lg_cpu *cpu, u32 i, unsigned long vaddr) in spgd_addr() argument 85 unsigned int index = pgd_index(vaddr); in spgd_addr() 97 static pmd_t *spmd_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spmd_addr() argument 99 unsigned int index = pmd_index(vaddr); in spmd_addr() 115 static pte_t *spte_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spte_addr() argument 118 pmd_t *pmd = spmd_addr(cpu, spgd, vaddr); in spte_addr() 129 return &page[pte_index(vaddr)]; in spte_addr() 136 static unsigned long gpgd_addr(struct lg_cpu *cpu, unsigned long vaddr) in gpgd_addr() argument 138 unsigned int index = vaddr >> (PGDIR_SHIFT); in gpgd_addr() 144 static unsigned long gpmd_addr(pgd_t gpgd, unsigned long vaddr) in gpmd_addr() argument [all …]
|
/drivers/net/ethernet/freescale/fman/ |
D | fman_muram.c | 47 unsigned long vaddr) in fman_muram_vbase_to_offset() argument 49 return vaddr - (unsigned long)muram->vbase; in fman_muram_vbase_to_offset() 68 void __iomem *vaddr; in fman_muram_init() local 81 vaddr = ioremap(base, size); in fman_muram_init() 82 if (!vaddr) { in fman_muram_init() 87 ret = gen_pool_add_virt(muram->pool, (unsigned long)vaddr, in fman_muram_init() 91 iounmap(vaddr); in fman_muram_init() 95 memset_io(vaddr, 0, (int)size); in fman_muram_init() 97 muram->vbase = vaddr; in fman_muram_init() 134 unsigned long vaddr; in fman_muram_alloc() local [all …]
|
/drivers/base/ |
D | dma-mapping.c | 21 void *vaddr; member 29 dma_free_coherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_coherent_release() 36 dma_free_noncoherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_noncoherent_release() 43 if (this->vaddr == match->vaddr) { in dmam_match() 68 void *vaddr; in dmam_alloc_coherent() local 74 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent() 75 if (!vaddr) { in dmam_alloc_coherent() 80 dr->vaddr = vaddr; in dmam_alloc_coherent() 86 return vaddr; in dmam_alloc_coherent() 99 void dmam_free_coherent(struct device *dev, size_t size, void *vaddr, in dmam_free_coherent() argument [all …]
|
/drivers/block/ |
D | cciss.h | 66 void __iomem *vaddr; member 226 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 227 readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 243 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 244 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 249 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 250 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 263 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() 264 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() 269 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() [all …]
|
D | smart1,2.h | 44 writel(c->busaddr, h->vaddr + S42XX_REQUEST_PORT_OFFSET); in smart4_submit_command() 56 writel(0, h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask() 60 h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask() 72 return (!readl(h->vaddr + S42XX_REQUEST_PORT_OFFSET)); in smart4_fifo_full() 82 = readl(h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed() 90 writel(0, h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed() 103 readl(h->vaddr + S42XX_INTR_STATUS); in smart4_intr_pending() 123 writel(c->busaddr, h->vaddr + COMMAND_FIFO); in smart2_submit_command() 128 writel(val, h->vaddr + INTR_MASK); in smart2_intr_mask() 133 return readl(h->vaddr + COMMAND_FIFO); in smart2_fifo_full() [all …]
|
/drivers/misc/sgi-gru/ |
D | grufault.c | 61 struct vm_area_struct *gru_find_vma(unsigned long vaddr) in gru_find_vma() argument 65 vma = find_vma(current->mm, vaddr); in gru_find_vma() 66 if (vma && vma->vm_start <= vaddr && vma->vm_ops == &gru_vm_ops) in gru_find_vma() 79 static struct gru_thread_state *gru_find_lock_gts(unsigned long vaddr) in gru_find_lock_gts() argument 86 vma = gru_find_vma(vaddr); in gru_find_lock_gts() 88 gts = gru_find_thread_state(vma, TSID(vaddr, vma)); in gru_find_lock_gts() 96 static struct gru_thread_state *gru_alloc_locked_gts(unsigned long vaddr) in gru_alloc_locked_gts() argument 103 vma = gru_find_vma(vaddr); in gru_alloc_locked_gts() 107 gts = gru_alloc_thread_state(vma, TSID(vaddr, vma)); in gru_alloc_locked_gts() 191 unsigned long vaddr, int write, in non_atomic_pte_lookup() argument [all …]
|
D | gruhandles.c | 152 unsigned long vaddr, unsigned long vaddrmask, in tgh_invalidate() argument 156 tgh->vaddr = vaddr; in tgh_invalidate() 170 unsigned long vaddr, int asid, int dirty, in tfh_write_only() argument 174 tfh->fillvaddr = vaddr; in tfh_write_only() 186 unsigned long vaddr, int asid, int dirty, in tfh_write_restart() argument 190 tfh->fillvaddr = vaddr; in tfh_write_restart()
|
/drivers/hwtracing/coresight/ |
D | coresight-tmc-etr.c | 37 void __iomem *vaddr; member 47 memset(drvdata->vaddr, 0, drvdata->size); in tmc_etr_enable_hw() 91 drvdata->buf = drvdata->vaddr + rwp - drvdata->paddr; in tmc_etr_dump_hw() 94 drvdata->buf = drvdata->vaddr; in tmc_etr_dump_hw() 120 void __iomem *vaddr = NULL; in tmc_enable_etr_sink_sysfs() local 130 if (!drvdata->vaddr) { in tmc_enable_etr_sink_sysfs() 138 vaddr = dma_alloc_coherent(drvdata->dev, drvdata->size, in tmc_enable_etr_sink_sysfs() 140 if (!vaddr) in tmc_enable_etr_sink_sysfs() 167 drvdata->vaddr = vaddr; in tmc_enable_etr_sink_sysfs() 169 drvdata->buf = drvdata->vaddr; in tmc_enable_etr_sink_sysfs() [all …]
|
/drivers/scsi/ |
D | hpsa.h | 166 void __iomem *vaddr; member 415 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 416 (void) readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 422 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_no_read() 428 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_ioaccel2() 440 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 441 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 445 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 446 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 454 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_performant_intr_mask() [all …]
|
D | sun3_scsi.c | 309 unsigned char *vaddr; in sun3scsi_dma_finish() local 311 vaddr = (unsigned char *)dvma_vmetov(sun3_dma_orig_addr); in sun3scsi_dma_finish() 313 vaddr += (sun3_dma_orig_count - fifo); in sun3scsi_dma_finish() 314 vaddr--; in sun3scsi_dma_finish() 318 *vaddr = (dregs->bpack_lo & 0xff00) >> 8; in sun3scsi_dma_finish() 319 vaddr--; in sun3scsi_dma_finish() 322 *vaddr = (dregs->bpack_hi & 0x00ff); in sun3scsi_dma_finish() 323 vaddr--; in sun3scsi_dma_finish() 326 *vaddr = (dregs->bpack_hi & 0xff00) >> 8; in sun3scsi_dma_finish() 358 unsigned char *vaddr; in sun3scsi_dma_finish() local [all …]
|
/drivers/xen/xenbus/ |
D | xenbus_client.c | 78 void **vaddr); 79 int (*unmap)(struct xenbus_device *dev, void *vaddr); 380 int xenbus_grant_ring(struct xenbus_device *dev, void *vaddr, in xenbus_grant_ring() argument 388 virt_to_gfn(vaddr), 0); in xenbus_grant_ring() 396 vaddr = vaddr + XEN_PAGE_SIZE; in xenbus_grant_ring() 470 unsigned int nr_grefs, void **vaddr) in xenbus_map_ring_valloc() argument 472 return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr); in xenbus_map_ring_valloc() 544 void **vaddr) in xenbus_map_ring_valloc_pv() argument 554 *vaddr = NULL; in xenbus_map_ring_valloc_pv() 586 *vaddr = area->addr; in xenbus_map_ring_valloc_pv() [all …]
|
/drivers/staging/android/ion/ |
D | ion_test.c | 69 void *vaddr = vmap(&page, 1, VM_MAP, pgprot); in ion_handle_test_dma() local 73 if (!vaddr) { in ion_handle_test_dma() 79 ret = copy_from_user(vaddr + offset, ptr, to_copy); in ion_handle_test_dma() 81 ret = copy_to_user(ptr, vaddr + offset, to_copy); in ion_handle_test_dma() 83 vunmap(vaddr); in ion_handle_test_dma() 119 void *vaddr = dma_buf_kmap(dma_buf, page_offset); in ion_handle_test_kernel() local 121 if (!vaddr) in ion_handle_test_kernel() 127 ret = copy_from_user(vaddr + copy_offset, ptr, to_copy); in ion_handle_test_kernel() 129 ret = copy_to_user(ptr, vaddr + copy_offset, to_copy); in ion_handle_test_kernel() 131 dma_buf_kunmap(dma_buf, page_offset, vaddr); in ion_handle_test_kernel()
|
/drivers/acpi/apei/ |
D | ghes.c | 154 unsigned long vaddr; in ghes_ioremap_pfn_nmi() local 156 vaddr = (unsigned long)GHES_IOREMAP_NMI_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_nmi() 157 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_nmi() 160 return (void __iomem *)vaddr; in ghes_ioremap_pfn_nmi() 165 unsigned long vaddr, paddr; in ghes_ioremap_pfn_irq() local 168 vaddr = (unsigned long)GHES_IOREMAP_IRQ_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_irq() 173 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, paddr, prot); in ghes_ioremap_pfn_irq() 175 return (void __iomem *)vaddr; in ghes_ioremap_pfn_irq() 180 unsigned long vaddr = (unsigned long __force)vaddr_ptr; in ghes_iounmap_nmi() local 183 BUG_ON(vaddr != (unsigned long)GHES_IOREMAP_NMI_PAGE(base)); in ghes_iounmap_nmi() [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_gem_execbuffer.c | 324 unsigned long vaddr; member 333 cache->vaddr = 0; in reloc_cache_init() 353 void *vaddr; in reloc_cache_fini() local 355 if (!cache->vaddr) in reloc_cache_fini() 358 vaddr = unmask_page(cache->vaddr); in reloc_cache_fini() 359 if (cache->vaddr & KMAP) { in reloc_cache_fini() 360 if (cache->vaddr & CLFLUSH_AFTER) in reloc_cache_fini() 363 kunmap_atomic(vaddr); in reloc_cache_fini() 367 io_mapping_unmap_atomic((void __iomem *)vaddr); in reloc_cache_fini() 386 void *vaddr; in reloc_kmap() local [all …]
|
/drivers/net/wireless/ath/ath10k/ |
D | htt_tx.c | 91 ar->htt.tx_q_state.vaddr->count[tid][peer_id] = count; in __ath10k_htt_tx_txq_recalc() 92 ar->htt.tx_q_state.vaddr->map[tid][idx] &= ~bit; in __ath10k_htt_tx_txq_recalc() 93 ar->htt.tx_q_state.vaddr->map[tid][idx] |= count ? bit : 0; in __ath10k_htt_tx_txq_recalc() 112 seq = le32_to_cpu(ar->htt.tx_q_state.vaddr->seq); in __ath10k_htt_tx_txq_sync() 114 ar->htt.tx_q_state.vaddr->seq = cpu_to_le32(seq); in __ath10k_htt_tx_txq_sync() 119 size = sizeof(*ar->htt.tx_q_state.vaddr); in __ath10k_htt_tx_txq_sync() 236 if (!htt->frag_desc.vaddr) in ath10k_htt_tx_free_cont_frag_desc() 243 htt->frag_desc.vaddr, in ath10k_htt_tx_free_cont_frag_desc() 256 htt->frag_desc.vaddr = dma_alloc_coherent(ar->dev, size, in ath10k_htt_tx_alloc_cont_frag_desc() 259 if (!htt->frag_desc.vaddr) { in ath10k_htt_tx_alloc_cont_frag_desc() [all …]
|
/drivers/i2c/busses/ |
D | i2c-ibm_iic.c | 85 volatile struct iic_regs __iomem *iic = dev->vaddr; in dump_iic_regs() 129 out_8(&dev->vaddr->intmsk, enable ? INTRMSK_EIMTC : 0); in iic_interrupt_mode() 137 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_init() 182 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_reset() 243 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_smbus_quick() 329 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_handler() 347 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_result() 380 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_abort_xfer() 412 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_wait_for_tc() 463 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_bytes() [all …]
|
/drivers/net/wireless/ath/ath6kl/ |
D | target.h | 333 #define AR6003_VTOP(vaddr) ((vaddr) & 0x001fffff) argument 334 #define AR6004_VTOP(vaddr) (vaddr) argument 336 #define TARG_VTOP(target_type, vaddr) \ argument 337 (((target_type) == TARGET_TYPE_AR6003) ? AR6003_VTOP(vaddr) : \ 338 (((target_type) == TARGET_TYPE_AR6004) ? AR6004_VTOP(vaddr) : 0))
|
/drivers/media/platform/coda/ |
D | coda-jpeg.c | 172 coda_memcpy_parabuf(ctx->parabuf.vaddr, huff + i); in coda_jpeg_write_tables() 176 coda_memcpy_parabuf(ctx->parabuf.vaddr, qmat + i); in coda_jpeg_write_tables() 183 void *vaddr = vb2_plane_vaddr(vb, 0); in coda_jpeg_check_buffer() local 187 soi = be16_to_cpup((__be16 *)vaddr); in coda_jpeg_check_buffer() 192 vaddr += len - 2; in coda_jpeg_check_buffer() 194 eoi = be16_to_cpup((__be16 *)(vaddr - i)); in coda_jpeg_check_buffer()
|
/drivers/vfio/ |
D | vfio_iommu_type1.c | 75 unsigned long vaddr; /* Process virtual addr */ member 210 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument 216 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { in vaddr_get_pfn() 223 vma = find_vma_intersection(current->mm, vaddr, vaddr + 1); in vaddr_get_pfn() 226 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn() 241 static long vfio_pin_pages(unsigned long vaddr, long npage, in vfio_pin_pages() argument 252 ret = vaddr_get_pfn(vaddr, prot, pfn_base); in vfio_pin_pages() 269 for (vaddr += PAGE_SIZE; i < npage; i++, vaddr += PAGE_SIZE) { in vfio_pin_pages() 270 ret = vaddr_get_pfn(vaddr, prot, &pfn); in vfio_pin_pages() 551 unsigned long vaddr = map->vaddr; in vfio_dma_do_map() local [all …]
|
/drivers/gpu/drm/ |
D | drm_gem_cma_helper.c | 112 cma_obj->vaddr = dma_alloc_wc(drm->dev, size, &cma_obj->paddr, in drm_gem_cma_create() 114 if (!cma_obj->vaddr) { in drm_gem_cma_create() 188 if (cma_obj->vaddr) { in drm_gem_cma_free_object() 190 cma_obj->vaddr, cma_obj->paddr); in drm_gem_cma_free_object() 321 ret = dma_mmap_wc(cma_obj->base.dev->dev, vma, cma_obj->vaddr, in drm_gem_cma_mmap_obj() 380 off, &cma_obj->paddr, cma_obj->vaddr, obj->size); in drm_gem_cma_describe() 409 ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr, in drm_gem_cma_prime_get_sg_table() 508 return cma_obj->vaddr; in drm_gem_cma_prime_vmap() 523 void drm_gem_cma_prime_vunmap(struct drm_gem_object *obj, void *vaddr) in drm_gem_cma_prime_vunmap() argument
|