Home
last modified time | relevance | path

Searched refs:vm_start (Results 1 – 25 of 123) sorted by relevance

12345

/drivers/gpu/drm/
Ddrm_vm.c78 if (efi_range_is_wc(vma->vm_start, vma->vm_end - in drm_io_prot()
79 vma->vm_start)) in drm_io_prot()
139 resource_size_t offset = vmf->address - vma->vm_start; in drm_vm_fault()
210 offset = vmf->address - vma->vm_start; in drm_vm_shm_fault()
240 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_shm_close()
314 offset = vmf->address - vma->vm_start; in drm_vm_dma_fault()
351 offset = vmf->address - vma->vm_start; in drm_vm_sg_fault()
395 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_open_locked()
421 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_close_locked()
465 unsigned long length = vma->vm_end - vma->vm_start; in drm_mmap_dma()
[all …]
/drivers/media/v4l2-core/
Dvideobuf-vmalloc.c56 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open()
68 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close()
227 buf->baddr = vma->vm_start; in __videobuf_mmap_mapper()
233 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); in __videobuf_mmap_mapper()
254 map, q, vma->vm_start, vma->vm_end, in __videobuf_mmap_mapper()
Dvideobuf-dma-contig.c67 map, map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open()
79 map, map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close()
290 buf->baddr = vma->vm_start; in __videobuf_mmap_mapper()
321 map, q, vma->vm_start, vma->vm_end, in __videobuf_mmap_mapper()
Dvideobuf-dma-sg.c389 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open()
402 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close()
440 vmf->address, vma->vm_start, vma->vm_end); in videobuf_vm_fault()
630 q->bufs[i]->baddr = vma->vm_start + size; in __videobuf_mmap_mapper()
641 map, q, vma->vm_start, vma->vm_end, vma->vm_pgoff, first, last); in __videobuf_mmap_mapper()
/drivers/char/
Dmspec.c75 unsigned long vm_start; /* Original (unsplit) base. */ member
114 last_index = (vdata->vm_end - vdata->vm_start) >> PAGE_SHIFT; in mspec_close()
202 vdata->vm_start = vma->vm_start; in mspec_mmap()
Duv_mmtimer.c151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) in uv_mmtimer_mmap()
166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, in uv_mmtimer_mmap()
Dbsr.c113 unsigned long size = vma->vm_end - vma->vm_start; in bsr_mmap()
121 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, in bsr_mmap()
124 ret = io_remap_pfn_range(vma, vma->vm_start, in bsr_mmap()
/drivers/soc/qcom/
Drmtfs_mem.c136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap()
139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap()
140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap()
146 vma->vm_start, in qcom_rmtfs_mem_mmap()
148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
/drivers/xen/xenfs/
Dxenstored.c36 size_t size = vma->vm_end - vma->vm_start; in xsd_kva_mmap()
41 if (remap_pfn_range(vma, vma->vm_start, in xsd_kva_mmap()
/drivers/media/common/videobuf2/
Dvideobuf2-memops.c93 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_open()
111 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_close()
/drivers/gpu/drm/amd/amdkfd/
Dkfd_doorbell.c136 if (vma->vm_end - vma->vm_start != kfd_doorbell_process_slice(dev)) in kfd_doorbell_mmap()
155 (unsigned long long) vma->vm_start, address, vma->vm_flags, in kfd_doorbell_mmap()
160 vma->vm_start, in kfd_doorbell_mmap()
Dkfd_events.c788 get_order(vma->vm_end - vma->vm_start)) { in kfd_event_mmap()
807 pr_debug(" start user address == 0x%08lx\n", vma->vm_start); in kfd_event_mmap()
812 vma->vm_end - vma->vm_start); in kfd_event_mmap()
814 page->user_address = (uint64_t __user *)vma->vm_start; in kfd_event_mmap()
817 ret = remap_pfn_range(vma, vma->vm_start, pfn, in kfd_event_mmap()
818 vma->vm_end - vma->vm_start, vma->vm_page_prot); in kfd_event_mmap()
820 p->signal_mapped_size = vma->vm_end - vma->vm_start; in kfd_event_mmap()
912 if (vma && address >= vma->vm_start) { in kfd_signal_iommu_event()
/drivers/sbus/char/
Dflash.c66 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) in flash_mmap()
67 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); in flash_mmap()
71 if (io_remap_pfn_range(vma, vma->vm_start, addr, size, vma->vm_page_prot)) in flash_mmap()
/drivers/xen/xenbus/
Dxenbus_dev_backend.c94 size_t size = vma->vm_end - vma->vm_start; in xenbus_backend_mmap()
102 if (remap_pfn_range(vma, vma->vm_start, in xenbus_backend_mmap()
/drivers/pci/
Dmmap.c68 return io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_resource_range()
69 vma->vm_end - vma->vm_start, in pci_mmap_resource_range()
/drivers/gpu/drm/i915/gem/
Di915_gem_mman.c32 return vma->vm_start == addr && in __vma_matches()
33 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches()
274 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu()
308 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; in vm_fault_gtt()
381 area->vm_start + (vma->ggtt_view.partial.offset << PAGE_SHIFT), in vm_fault_gtt()
383 min_t(u64, vma->size, area->vm_end - area->vm_start), in vm_fault_gtt()
441 addr -= area->vm_start; in vm_access()
/drivers/dax/
Ddevice.c35 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma()
38 current->comm, func, vma->vm_start, vma->vm_end, in check_vma()
130 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault()
172 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault()
207 vmf->vma->vm_start, vmf->vma->vm_end, pe_size); in dev_dax_huge_fault()
/drivers/infiniband/core/
Dib_core_uverbs.c77 if (vma->vm_end - vma->vm_start != size) in rdma_user_mmap_io()
91 if (io_remap_pfn_range(vma, vma->vm_start, pfn, size, prot)) { in rdma_user_mmap_io()
171 if (entry->npages * PAGE_SIZE != vma->vm_end - vma->vm_start) { in rdma_user_mmap_entry_get()
/drivers/infiniband/hw/qib/
Dqib_file_ops.c715 if ((vma->vm_end - vma->vm_start) > len) { in qib_mmap_mem()
718 vma->vm_end - vma->vm_start, len); in qib_mmap_mem()
740 ret = remap_pfn_range(vma, vma->vm_start, pfn, in qib_mmap_mem()
763 if ((vma->vm_end - vma->vm_start) > sz) { in mmap_ureg()
766 vma->vm_end - vma->vm_start); in mmap_ureg()
773 ret = io_remap_pfn_range(vma, vma->vm_start, in mmap_ureg()
775 vma->vm_end - vma->vm_start, in mmap_ureg()
795 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->palign)) { in mmap_piobufs()
798 vma->vm_end - vma->vm_start); in mmap_piobufs()
820 ret = io_remap_pfn_range(vma, vma->vm_start, phys >> PAGE_SHIFT, in mmap_piobufs()
[all …]
/drivers/infiniband/hw/cxgb4/
Dprovider.c126 int len = vma->vm_end - vma->vm_start; in c4iw_mmap()
137 if (vma->vm_start & (PAGE_SIZE-1)) in c4iw_mmap()
157 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
177 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
185 ret = remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
/drivers/vfio/pci/
Dvfio_pci_core.c1335 zap_vma_ptes(vma, vma->vm_start, in vfio_pci_zap_and_vma_lock()
1336 vma->vm_end - vma->vm_start); in vfio_pci_zap_and_vma_lock()
1392 zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start); in vfio_pci_mmap_open()
1437 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in vfio_pci_mmap_fault()
1438 vma->vm_end - vma->vm_start, in vfio_pci_mmap_fault()
1441 zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start); in vfio_pci_mmap_fault()
1447 zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start); in vfio_pci_mmap_fault()
1475 if (vma->vm_end < vma->vm_start) in vfio_pci_core_mmap()
1494 req_len = vma->vm_end - vma->vm_start; in vfio_pci_core_mmap()
/drivers/gpu/drm/virtio/
Dvirtgpu_vram.c37 unsigned long vm_size = vma->vm_end - vma->vm_start; in virtio_gpu_vram_mmap()
61 ret = io_remap_pfn_range(vma, vma->vm_start, in virtio_gpu_vram_mmap()
/drivers/xen/
Dprivcmd.c288 if (!vma || (msg->va != vma->vm_start) || vma->vm_private_data) in privcmd_ioctl_mmap()
293 state.va = vma->vm_start; in privcmd_ioctl_mmap()
520 if (m.addr != vma->vm_start || in privcmd_ioctl_mmap_batch()
532 if (m.addr < vma->vm_start || in privcmd_ioctl_mmap_batch()
904 int numgfns = (vma->vm_end - vma->vm_start) >> XEN_PAGE_SHIFT; in privcmd_close()
922 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, in privcmd_fault()
/drivers/infiniband/hw/hfi1/
Dfile_ops.c396 if ((vma->vm_end - vma->vm_start) != memlen) { in hfi1_file_mmap()
398 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap()
407 addr = vma->vm_start; in hfi1_file_mmap()
523 if ((vma->vm_end - vma->vm_start) != memlen) { in hfi1_file_mmap()
526 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap()
535 vma->vm_end - vma->vm_start, vma->vm_flags); in hfi1_file_mmap()
541 ret = io_remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap()
546 ret = remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap()
551 ret = remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap()
/drivers/misc/habanalabs/common/
Dcommand_buffer.c471 new_mmap_size = cb->mmap_size - (vma->vm_end - vma->vm_start); in cb_vm_close()
512 user_cb_size = vma->vm_end - vma->vm_start; in hl_cb_mmap()
516 vma->vm_end - vma->vm_start, cb->size); in hl_cb_mmap()
521 if (!access_ok((void __user *) (uintptr_t) vma->vm_start, in hl_cb_mmap()
525 vma->vm_start); in hl_cb_mmap()

12345