/drivers/gpu/drm/ |
D | drm_vm.c | 78 if (efi_range_is_wc(vma->vm_start, vma->vm_end - in drm_io_prot() 79 vma->vm_start)) in drm_io_prot() 139 resource_size_t offset = vmf->address - vma->vm_start; in drm_vm_fault() 210 offset = vmf->address - vma->vm_start; in drm_vm_shm_fault() 240 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_shm_close() 314 offset = vmf->address - vma->vm_start; in drm_vm_dma_fault() 351 offset = vmf->address - vma->vm_start; in drm_vm_sg_fault() 395 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_open_locked() 421 vma->vm_start, vma->vm_end - vma->vm_start); in drm_vm_close_locked() 465 unsigned long length = vma->vm_end - vma->vm_start; in drm_mmap_dma() [all …]
|
/drivers/media/v4l2-core/ |
D | videobuf-vmalloc.c | 56 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open() 68 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close() 227 buf->baddr = vma->vm_start; in __videobuf_mmap_mapper() 233 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); in __videobuf_mmap_mapper() 254 map, q, vma->vm_start, vma->vm_end, in __videobuf_mmap_mapper()
|
D | videobuf-dma-contig.c | 67 map, map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open() 79 map, map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close() 290 buf->baddr = vma->vm_start; in __videobuf_mmap_mapper() 321 map, q, vma->vm_start, vma->vm_end, in __videobuf_mmap_mapper()
|
D | videobuf-dma-sg.c | 389 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_open() 402 map->count, vma->vm_start, vma->vm_end); in videobuf_vm_close() 440 vmf->address, vma->vm_start, vma->vm_end); in videobuf_vm_fault() 630 q->bufs[i]->baddr = vma->vm_start + size; in __videobuf_mmap_mapper() 641 map, q, vma->vm_start, vma->vm_end, vma->vm_pgoff, first, last); in __videobuf_mmap_mapper()
|
/drivers/char/ |
D | mspec.c | 75 unsigned long vm_start; /* Original (unsplit) base. */ member 114 last_index = (vdata->vm_end - vdata->vm_start) >> PAGE_SHIFT; in mspec_close() 202 vdata->vm_start = vma->vm_start; in mspec_mmap()
|
D | uv_mmtimer.c | 151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) in uv_mmtimer_mmap() 166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, in uv_mmtimer_mmap()
|
D | bsr.c | 113 unsigned long size = vma->vm_end - vma->vm_start; in bsr_mmap() 121 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, in bsr_mmap() 124 ret = io_remap_pfn_range(vma, vma->vm_start, in bsr_mmap()
|
/drivers/soc/qcom/ |
D | rmtfs_mem.c | 136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap() 139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap() 140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap() 146 vma->vm_start, in qcom_rmtfs_mem_mmap() 148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
|
/drivers/xen/xenfs/ |
D | xenstored.c | 36 size_t size = vma->vm_end - vma->vm_start; in xsd_kva_mmap() 41 if (remap_pfn_range(vma, vma->vm_start, in xsd_kva_mmap()
|
/drivers/media/common/videobuf2/ |
D | videobuf2-memops.c | 93 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_open() 111 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_close()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_doorbell.c | 136 if (vma->vm_end - vma->vm_start != kfd_doorbell_process_slice(dev)) in kfd_doorbell_mmap() 155 (unsigned long long) vma->vm_start, address, vma->vm_flags, in kfd_doorbell_mmap() 160 vma->vm_start, in kfd_doorbell_mmap()
|
D | kfd_events.c | 788 get_order(vma->vm_end - vma->vm_start)) { in kfd_event_mmap() 807 pr_debug(" start user address == 0x%08lx\n", vma->vm_start); in kfd_event_mmap() 812 vma->vm_end - vma->vm_start); in kfd_event_mmap() 814 page->user_address = (uint64_t __user *)vma->vm_start; in kfd_event_mmap() 817 ret = remap_pfn_range(vma, vma->vm_start, pfn, in kfd_event_mmap() 818 vma->vm_end - vma->vm_start, vma->vm_page_prot); in kfd_event_mmap() 820 p->signal_mapped_size = vma->vm_end - vma->vm_start; in kfd_event_mmap() 912 if (vma && address >= vma->vm_start) { in kfd_signal_iommu_event()
|
/drivers/sbus/char/ |
D | flash.c | 66 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) in flash_mmap() 67 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); in flash_mmap() 71 if (io_remap_pfn_range(vma, vma->vm_start, addr, size, vma->vm_page_prot)) in flash_mmap()
|
/drivers/xen/xenbus/ |
D | xenbus_dev_backend.c | 94 size_t size = vma->vm_end - vma->vm_start; in xenbus_backend_mmap() 102 if (remap_pfn_range(vma, vma->vm_start, in xenbus_backend_mmap()
|
/drivers/pci/ |
D | mmap.c | 68 return io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_resource_range() 69 vma->vm_end - vma->vm_start, in pci_mmap_resource_range()
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_mman.c | 32 return vma->vm_start == addr && in __vma_matches() 33 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches() 274 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu() 308 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; in vm_fault_gtt() 381 area->vm_start + (vma->ggtt_view.partial.offset << PAGE_SHIFT), in vm_fault_gtt() 383 min_t(u64, vma->size, area->vm_end - area->vm_start), in vm_fault_gtt() 441 addr -= area->vm_start; in vm_access()
|
/drivers/dax/ |
D | device.c | 35 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma() 38 current->comm, func, vma->vm_start, vma->vm_end, in check_vma() 130 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault() 172 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault() 207 vmf->vma->vm_start, vmf->vma->vm_end, pe_size); in dev_dax_huge_fault()
|
/drivers/infiniband/core/ |
D | ib_core_uverbs.c | 77 if (vma->vm_end - vma->vm_start != size) in rdma_user_mmap_io() 91 if (io_remap_pfn_range(vma, vma->vm_start, pfn, size, prot)) { in rdma_user_mmap_io() 171 if (entry->npages * PAGE_SIZE != vma->vm_end - vma->vm_start) { in rdma_user_mmap_entry_get()
|
/drivers/infiniband/hw/qib/ |
D | qib_file_ops.c | 715 if ((vma->vm_end - vma->vm_start) > len) { in qib_mmap_mem() 718 vma->vm_end - vma->vm_start, len); in qib_mmap_mem() 740 ret = remap_pfn_range(vma, vma->vm_start, pfn, in qib_mmap_mem() 763 if ((vma->vm_end - vma->vm_start) > sz) { in mmap_ureg() 766 vma->vm_end - vma->vm_start); in mmap_ureg() 773 ret = io_remap_pfn_range(vma, vma->vm_start, in mmap_ureg() 775 vma->vm_end - vma->vm_start, in mmap_ureg() 795 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->palign)) { in mmap_piobufs() 798 vma->vm_end - vma->vm_start); in mmap_piobufs() 820 ret = io_remap_pfn_range(vma, vma->vm_start, phys >> PAGE_SHIFT, in mmap_piobufs() [all …]
|
/drivers/infiniband/hw/cxgb4/ |
D | provider.c | 126 int len = vma->vm_end - vma->vm_start; in c4iw_mmap() 137 if (vma->vm_start & (PAGE_SIZE-1)) in c4iw_mmap() 157 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap() 177 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap() 185 ret = remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
|
/drivers/vfio/pci/ |
D | vfio_pci_core.c | 1335 zap_vma_ptes(vma, vma->vm_start, in vfio_pci_zap_and_vma_lock() 1336 vma->vm_end - vma->vm_start); in vfio_pci_zap_and_vma_lock() 1392 zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start); in vfio_pci_mmap_open() 1437 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in vfio_pci_mmap_fault() 1438 vma->vm_end - vma->vm_start, in vfio_pci_mmap_fault() 1441 zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start); in vfio_pci_mmap_fault() 1447 zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start); in vfio_pci_mmap_fault() 1475 if (vma->vm_end < vma->vm_start) in vfio_pci_core_mmap() 1494 req_len = vma->vm_end - vma->vm_start; in vfio_pci_core_mmap()
|
/drivers/gpu/drm/virtio/ |
D | virtgpu_vram.c | 37 unsigned long vm_size = vma->vm_end - vma->vm_start; in virtio_gpu_vram_mmap() 61 ret = io_remap_pfn_range(vma, vma->vm_start, in virtio_gpu_vram_mmap()
|
/drivers/xen/ |
D | privcmd.c | 288 if (!vma || (msg->va != vma->vm_start) || vma->vm_private_data) in privcmd_ioctl_mmap() 293 state.va = vma->vm_start; in privcmd_ioctl_mmap() 520 if (m.addr != vma->vm_start || in privcmd_ioctl_mmap_batch() 532 if (m.addr < vma->vm_start || in privcmd_ioctl_mmap_batch() 904 int numgfns = (vma->vm_end - vma->vm_start) >> XEN_PAGE_SHIFT; in privcmd_close() 922 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, in privcmd_fault()
|
/drivers/infiniband/hw/hfi1/ |
D | file_ops.c | 396 if ((vma->vm_end - vma->vm_start) != memlen) { in hfi1_file_mmap() 398 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap() 407 addr = vma->vm_start; in hfi1_file_mmap() 523 if ((vma->vm_end - vma->vm_start) != memlen) { in hfi1_file_mmap() 526 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap() 535 vma->vm_end - vma->vm_start, vma->vm_flags); in hfi1_file_mmap() 541 ret = io_remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap() 546 ret = remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap() 551 ret = remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap()
|
/drivers/misc/habanalabs/common/ |
D | command_buffer.c | 471 new_mmap_size = cb->mmap_size - (vma->vm_end - vma->vm_start); in cb_vm_close() 512 user_cb_size = vma->vm_end - vma->vm_start; in hl_cb_mmap() 516 vma->vm_end - vma->vm_start, cb->size); in hl_cb_mmap() 521 if (!access_ok((void __user *) (uintptr_t) vma->vm_start, in hl_cb_mmap() 525 vma->vm_start); in hl_cb_mmap()
|