/drivers/base/ |
D | dma-mapping.c | 281 unsigned long vm_flags, pgprot_t prot, in dma_common_pages_remap() argument 286 area = get_vm_area_caller(size, vm_flags, caller); in dma_common_pages_remap() 306 unsigned long vm_flags, in dma_common_contiguous_remap() argument 321 ptr = dma_common_pages_remap(pages, size, vm_flags, prot, caller); in dma_common_contiguous_remap() 331 void dma_common_free_remap(void *cpu_addr, size_t size, unsigned long vm_flags) in dma_common_free_remap() argument 335 if (!area || (area->flags & vm_flags) != vm_flags) { in dma_common_free_remap()
|
/drivers/gpu/drm/udl/ |
D | udl_gem.c | 67 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in update_vm_cache_attr() 70 pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); in update_vm_cache_attr() 73 pgprot_noncached(vm_get_page_prot(vma->vm_flags)); in update_vm_cache_attr() 95 vma->vm_flags &= ~VM_PFNMAP; in udl_drm_gem_mmap() 96 vma->vm_flags |= VM_MIXEDMAP; in udl_drm_gem_mmap()
|
/drivers/gpu/drm/ttm/ |
D | ttm_bo_vm.c | 211 cvma.vm_page_prot = vm_get_page_prot(cvma.vm_flags); in ttm_bo_vm_fault() 249 if (vma->vm_flags & VM_MIXEDMAP) in ttm_bo_vm_fault() 363 vma->vm_flags |= VM_MIXEDMAP; in ttm_bo_mmap() 364 vma->vm_flags |= VM_IO | VM_DONTEXPAND | VM_DONTDUMP; in ttm_bo_mmap() 379 vma->vm_flags |= VM_MIXEDMAP; in ttm_fbdev_mmap() 380 vma->vm_flags |= VM_IO | VM_DONTEXPAND; in ttm_fbdev_mmap()
|
/drivers/staging/lustre/lustre/llite/ |
D | llite_mmap.c | 71 vma->vm_flags & VM_SHARED) { in our_vma() 127 fio->ft_executable = vma->vm_flags & VM_EXEC; in ll_fault_io_init() 135 *ra_flags = vma->vm_flags & (VM_RAND_READ | VM_SEQ_READ); in ll_fault_io_init() 136 vma->vm_flags &= ~VM_SEQ_READ; in ll_fault_io_init() 137 vma->vm_flags |= VM_RAND_READ; in ll_fault_io_init() 139 CDEBUG(D_MMAP, "vm_flags: %lx (%lu %d)\n", vma->vm_flags, in ll_fault_io_init() 327 vma->vm_flags |= ra_flags; in ll_fault0()
|
/drivers/sbus/char/ |
D | flash.c | 46 if ((vma->vm_flags & VM_READ) && in flash_mmap() 47 (vma->vm_flags & VM_WRITE)) { in flash_mmap() 51 if (vma->vm_flags & VM_READ) { in flash_mmap() 54 } else if (vma->vm_flags & VM_WRITE) { in flash_mmap()
|
/drivers/gpu/drm/vgem/ |
D | vgem_drv.c | 201 unsigned long flags = vma->vm_flags; in vgem_mmap() 211 vma->vm_flags = flags | VM_DONTEXPAND | VM_DONTDUMP; in vgem_mmap() 296 vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; in vgem_prime_mmap() 297 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); in vgem_prime_mmap()
|
/drivers/gpu/drm/ |
D | drm_vm.c | 59 pgprot_t tmp = vm_get_page_prot(vma->vm_flags); in drm_io_prot() 80 pgprot_t tmp = vm_get_page_prot(vma->vm_flags); in drm_dma_prot() 488 vma->vm_flags &= ~(VM_WRITE | VM_MAYWRITE); in drm_mmap_dma() 504 vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; in drm_mmap_dma() 572 vma->vm_flags &= ~(VM_WRITE | VM_MAYWRITE); in drm_mmap_locked() 640 vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; in drm_mmap_locked()
|
/drivers/char/ |
D | mspec.c | 258 if ((vma->vm_flags & VM_SHARED) == 0) in mspec_mmap() 261 if ((vma->vm_flags & VM_WRITE) == 0) in mspec_mmap() 280 vma->vm_flags |= VM_IO | VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP; in mspec_mmap()
|
D | uv_mmtimer.c | 154 if (vma->vm_flags & VM_WRITE) in uv_mmtimer_mmap()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_doorbell.c | 140 vma->vm_flags |= VM_IO | VM_DONTCOPY | VM_DONTEXPAND | VM_NORESERVE | in kfd_doorbell_mmap() 151 (unsigned long long) vma->vm_start, address, vma->vm_flags, in kfd_doorbell_mmap()
|
D | kfd_events.c | 836 vma->vm_flags |= VM_IO | VM_DONTCOPY | VM_DONTEXPAND | VM_NORESERVE in kfd_event_mmap() 843 pr_debug(" vm_flags == 0x%08lX\n", vma->vm_flags); in kfd_event_mmap() 929 if (is_write_requested && !(vma->vm_flags & VM_WRITE)) in kfd_signal_iommu_event() 933 if (is_execute_requested && !(vma->vm_flags & VM_EXEC)) in kfd_signal_iommu_event()
|
/drivers/gpu/drm/exynos/ |
D | exynos_drm_gem.c | 347 vma->vm_flags &= ~VM_PFNMAP; in exynos_drm_gem_mmap_buffer() 503 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in exynos_drm_gem_mmap_obj() 506 pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); in exynos_drm_gem_mmap_obj() 509 pgprot_noncached(vm_get_page_prot(vma->vm_flags)); in exynos_drm_gem_mmap_obj()
|
/drivers/infiniband/hw/qib/ |
D | qib_file_ops.c | 749 if (vma->vm_flags & VM_WRITE) { in qib_mmap_mem() 757 vma->vm_flags &= ~VM_MAYWRITE; in qib_mmap_mem() 793 vma->vm_flags |= VM_DONTCOPY | VM_DONTEXPAND; in mmap_ureg() 834 vma->vm_flags &= ~VM_MAYREAD; in mmap_piobufs() 835 vma->vm_flags |= VM_DONTCOPY | VM_DONTEXPAND; in mmap_piobufs() 868 if (vma->vm_flags & VM_WRITE) { in mmap_rcvegrbufs() 871 vma->vm_flags); in mmap_rcvegrbufs() 876 vma->vm_flags &= ~VM_MAYWRITE; in mmap_rcvegrbufs() 957 if (vma->vm_flags & VM_WRITE) { in mmap_kvaddr() 960 vma->vm_flags); in mmap_kvaddr() [all …]
|
/drivers/video/fbdev/core/ |
D | fb_defio.c | 170 vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; in fb_deferred_io_mmap() 172 vma->vm_flags |= VM_IO; in fb_deferred_io_mmap()
|
/drivers/gpu/drm/tegra/ |
D | gem.c | 471 vma->vm_flags &= ~VM_PFNMAP; in tegra_drm_mmap() 483 pgprot_t prot = vm_get_page_prot(vma->vm_flags); in tegra_drm_mmap() 485 vma->vm_flags |= VM_MIXEDMAP; in tegra_drm_mmap() 486 vma->vm_flags &= ~VM_PFNMAP; in tegra_drm_mmap()
|
/drivers/staging/android/ |
D | ashmem.c | 387 if (unlikely((vma->vm_flags & ~calc_vm_prot_bits(asma->prot_mask, 0)) & in ashmem_mmap() 392 vma->vm_flags &= ~calc_vm_may_flags(~asma->prot_mask); in ashmem_mmap() 402 vmfile = shmem_file_setup(name, asma->size, vma->vm_flags); in ashmem_mmap() 412 if (vma->vm_flags & VM_SHARED) in ashmem_mmap()
|
/drivers/gpu/drm/msm/ |
D | msm_gem.c | 164 vma->vm_flags &= ~VM_PFNMAP; in msm_gem_mmap_obj() 165 vma->vm_flags |= VM_MIXEDMAP; in msm_gem_mmap_obj() 168 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); in msm_gem_mmap_obj() 170 vma->vm_page_prot = pgprot_noncached(vm_get_page_prot(vma->vm_flags)); in msm_gem_mmap_obj() 182 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in msm_gem_mmap_obj()
|
/drivers/misc/mic/scif/ |
D | scif_mmap.c | 574 req.prot = vma->vm_flags & (VM_READ | VM_WRITE); in scif_munmap() 646 req.prot = vma->vm_flags & (VM_READ | VM_WRITE); in scif_mmap() 675 vma->vm_flags |= VM_DONTCOPY | VM_DONTEXPAND | VM_DONTDUMP; in scif_mmap() 678 vma->vm_flags |= VM_IO | VM_PFNMAP; in scif_mmap()
|
/drivers/xen/ |
D | gntdev.c | 987 if ((vma->vm_flags & VM_WRITE) && !(vma->vm_flags & VM_SHARED)) in gntdev_mmap() 1008 vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP | VM_MIXEDMAP; in gntdev_mmap() 1011 vma->vm_flags |= VM_DONTCOPY; in gntdev_mmap() 1019 if ((vma->vm_flags & VM_WRITE) && in gntdev_mmap() 1024 if (!(vma->vm_flags & VM_WRITE)) in gntdev_mmap()
|
/drivers/gpu/drm/i915/ |
D | i915_mm.c | 67 GEM_BUG_ON((vma->vm_flags & in remap_io_mapping()
|
/drivers/gpu/drm/vc4/ |
D | vc4_bo.c | 396 if (bo->validated_shader && (vma->vm_flags & VM_WRITE)) { in vc4_mmap() 406 vma->vm_flags &= ~VM_PFNMAP; in vc4_mmap() 421 if (bo->validated_shader && (vma->vm_flags & VM_WRITE)) { in vc4_prime_mmap()
|
/drivers/infiniband/hw/hfi1/ |
D | file_ops.c | 461 !(vma->vm_flags & VM_SHARED)) { in hfi1_file_mmap() 474 flags = vma->vm_flags; in hfi1_file_mmap() 539 if (vma->vm_flags & VM_WRITE) { in hfi1_file_mmap() 543 vma->vm_flags &= ~VM_MAYWRITE; in hfi1_file_mmap() 665 vma->vm_flags = flags; in hfi1_file_mmap() 669 vma->vm_end - vma->vm_start, vma->vm_flags); in hfi1_file_mmap()
|
/drivers/vfio/platform/ |
D | vfio_platform_common.c | 599 if (!(vma->vm_flags & VM_SHARED)) in vfio_platform_mmap() 612 && (vma->vm_flags & VM_READ)) in vfio_platform_mmap() 616 && (vma->vm_flags & VM_WRITE)) in vfio_platform_mmap()
|
/drivers/gpu/drm/omapdrm/ |
D | omap_gem.c | 602 vma->vm_flags &= ~VM_PFNMAP; in omap_gem_mmap_obj() 603 vma->vm_flags |= VM_MIXEDMAP; in omap_gem_mmap_obj() 606 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); in omap_gem_mmap_obj() 608 vma->vm_page_prot = pgprot_noncached(vm_get_page_prot(vma->vm_flags)); in omap_gem_mmap_obj() 627 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in omap_gem_mmap_obj()
|
/drivers/dax/ |
D | dax.c | 275 if ((vma->vm_flags & VM_MAYSHARE) != VM_MAYSHARE) { in check_vma() 290 && (vma->vm_flags & VM_DONTCOPY) == 0) { in check_vma() 485 vma->vm_flags |= VM_MIXEDMAP | VM_HUGEPAGE; in dax_mmap()
|