/arch/metag/kernel/ |
D | dma.c | 101 unsigned long vm_start; member 109 .vm_start = CONSISTENT_START, 117 unsigned long addr = head->vm_start, end = head->vm_end - size; in metag_vm_region_alloc() 130 if ((addr + size) <= c->vm_start) in metag_vm_region_alloc() 142 new->vm_start = addr; in metag_vm_region_alloc() 162 if (c->vm_active && c->vm_start == addr) in metag_vm_region_find() 225 unsigned long vaddr = c->vm_start; in dma_alloc_coherent() 258 return (void *)c->vm_start; in dma_alloc_coherent() 287 if ((c->vm_end - c->vm_start) != size) { in dma_free_coherent() 289 __func__, c->vm_end - c->vm_start, size); in dma_free_coherent() [all …]
|
/arch/powerpc/mm/ |
D | dma-noncoherent.c | 86 unsigned long vm_start; member 92 .vm_start = CONSISTENT_BASE, 99 unsigned long addr = head->vm_start, end = head->vm_end - size; in ppc_vm_region_alloc() 112 if ((addr + size) <= c->vm_start) in ppc_vm_region_alloc() 124 new->vm_start = addr; in ppc_vm_region_alloc() 142 if (c->vm_start == addr) in ppc_vm_region_find() 218 unsigned long vaddr = c->vm_start; in __dma_alloc_coherent() 244 return (void *)c->vm_start; in __dma_alloc_coherent() 270 if ((c->vm_end - c->vm_start) != size) { in __dma_free_coherent() 272 __func__, c->vm_end - c->vm_start, size); in __dma_free_coherent() [all …]
|
/arch/ia64/include/asm/ |
D | fb.h | 12 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in fb_pgprotect()
|
/arch/sparc/include/asm/ |
D | tlb_32.h | 6 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 11 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/avr32/include/asm/ |
D | tlb.h | 12 flush_cache_range(vma, vma->vm_start, vma->vm_end) 15 flush_tlb_range(vma, vma->vm_start, vma->vm_end)
|
/arch/arc/include/asm/ |
D | tlb.h | 32 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 39 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/metag/include/asm/ |
D | tlb.h | 14 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 20 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/xtensa/include/asm/ |
D | tlb.h | 29 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 35 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/parisc/include/asm/ |
D | tlb.h | 11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 16 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/x86/um/ |
D | mem_32.c | 20 gate_vma.vm_start = FIXADDR_USER_START; in gate_vma_init() 52 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
D | mem_64.c | 6 if (vma->vm_mm && vma->vm_start == um_vdso_addr) in arch_vma_name()
|
/arch/arc/kernel/ |
D | arc_hostlink.c | 25 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in arc_hl_mmap() 26 vma->vm_end - vma->vm_start, in arc_hl_mmap()
|
D | troubleshoot.c | 99 if (vma && (vma->vm_start <= address)) { in show_faulting_vma() 109 vma->vm_start < TASK_UNMAPPED_BASE ? in show_faulting_vma() 110 address : address - vma->vm_start, in show_faulting_vma() 111 nm, vma->vm_start, vma->vm_end); in show_faulting_vma()
|
/arch/parisc/mm/ |
D | fault.c | 126 if (tree->vm_start > addr) { 132 if (prev->vm_next->vm_start > addr) 196 vma->vm_start, vma->vm_end); in show_signal_msg() 229 if (!vma || address < vma->vm_start) in do_page_fault() 306 address < vma->vm_start || address >= vma->vm_end) { in do_page_fault()
|
/arch/powerpc/kernel/ |
D | proc_powerpc.c | 46 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) in page_map_mmap() 49 remap_pfn_range(vma, vma->vm_start, in page_map_mmap()
|
/arch/um/drivers/ |
D | mmapper_kern.c | 56 size = vma->vm_end - vma->vm_start; in mmapper_mmap() 64 if (remap_pfn_range(vma, vma->vm_start, p_buf >> PAGE_SHIFT, size, in mmapper_mmap()
|
/arch/microblaze/kernel/ |
D | dma.c | 163 unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; in dma_direct_mmap_coherent() 177 return remap_pfn_range(vma, vma->vm_start, pfn + off, in dma_direct_mmap_coherent() 178 vma->vm_end - vma->vm_start, vma->vm_page_prot); in dma_direct_mmap_coherent()
|
/arch/tile/mm/ |
D | elf.c | 78 if (vma->vm_start == (ELF_ET_DYN_BASE & PAGE_MASK)) { in notify_exec() 82 snprintf(buf, sizeof(buf), "0x%lx:@", vma->vm_start); in notify_exec()
|
/arch/cris/arch-v32/drivers/pci/ |
D | bios.c | 31 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_page_range() 32 vma->vm_end - vma->vm_start, in pci_mmap_page_range()
|
/arch/powerpc/include/asm/ |
D | fb.h | 12 vma->vm_end - vma->vm_start, in fb_pgprotect()
|
/arch/nios2/include/asm/ |
D | tlb.h | 25 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/ia64/pci/ |
D | pci.c | 425 unsigned long size = vma->vm_end - vma->vm_start; in pci_mmap_page_range() 455 efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in pci_mmap_page_range() 460 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_page_range() 461 vma->vm_end - vma->vm_start, vma->vm_page_prot)) in pci_mmap_page_range() 496 unsigned long size = vma->vm_end - vma->vm_start; in pci_mmap_legacy_page_range() 520 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_legacy_page_range()
|
/arch/mips/include/asm/ |
D | tlb.h | 11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
|
/arch/metag/mm/ |
D | hugetlbpage.c | 50 if (addr + len > vma->vm_start) in prepare_hugepage_range() 53 (ALIGN_HUGEPT(addr + len) > vma->vm_start)) in prepare_hugepage_range() 157 <= vma->vm_start)) { in hugetlb_get_unmapped_area_existing()
|
/arch/nios2/kernel/ |
D | sys_nios2.c | 42 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) in sys_cacheflush()
|