Home
last modified time | relevance | path

Searched refs:vm_start (Results 1 – 25 of 108) sorted by relevance

12345

/arch/metag/kernel/
Ddma.c101 unsigned long vm_start; member
109 .vm_start = CONSISTENT_START,
117 unsigned long addr = head->vm_start, end = head->vm_end - size; in metag_vm_region_alloc()
130 if ((addr + size) <= c->vm_start) in metag_vm_region_alloc()
142 new->vm_start = addr; in metag_vm_region_alloc()
162 if (c->vm_active && c->vm_start == addr) in metag_vm_region_find()
225 unsigned long vaddr = c->vm_start; in dma_alloc_coherent()
258 return (void *)c->vm_start; in dma_alloc_coherent()
287 if ((c->vm_end - c->vm_start) != size) { in dma_free_coherent()
289 __func__, c->vm_end - c->vm_start, size); in dma_free_coherent()
[all …]
/arch/powerpc/mm/
Ddma-noncoherent.c86 unsigned long vm_start; member
92 .vm_start = CONSISTENT_BASE,
99 unsigned long addr = head->vm_start, end = head->vm_end - size; in ppc_vm_region_alloc()
112 if ((addr + size) <= c->vm_start) in ppc_vm_region_alloc()
124 new->vm_start = addr; in ppc_vm_region_alloc()
142 if (c->vm_start == addr) in ppc_vm_region_find()
218 unsigned long vaddr = c->vm_start; in __dma_alloc_coherent()
244 return (void *)c->vm_start; in __dma_alloc_coherent()
270 if ((c->vm_end - c->vm_start) != size) { in __dma_free_coherent()
272 __func__, c->vm_end - c->vm_start, size); in __dma_free_coherent()
[all …]
/arch/ia64/include/asm/
Dfb.h12 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in fb_pgprotect()
/arch/sparc/include/asm/
Dtlb_32.h6 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
11 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/arch/avr32/include/asm/
Dtlb.h12 flush_cache_range(vma, vma->vm_start, vma->vm_end)
15 flush_tlb_range(vma, vma->vm_start, vma->vm_end)
/arch/arc/include/asm/
Dtlb.h32 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
39 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/arch/metag/include/asm/
Dtlb.h14 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
20 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/arch/xtensa/include/asm/
Dtlb.h29 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
35 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/arch/parisc/include/asm/
Dtlb.h11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
16 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/arch/x86/um/
Dmem_32.c20 gate_vma.vm_start = FIXADDR_USER_START; in gate_vma_init()
52 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
Dmem_64.c6 if (vma->vm_mm && vma->vm_start == um_vdso_addr) in arch_vma_name()
/arch/arc/kernel/
Darc_hostlink.c25 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in arc_hl_mmap()
26 vma->vm_end - vma->vm_start, in arc_hl_mmap()
Dtroubleshoot.c99 if (vma && (vma->vm_start <= address)) { in show_faulting_vma()
109 vma->vm_start < TASK_UNMAPPED_BASE ? in show_faulting_vma()
110 address : address - vma->vm_start, in show_faulting_vma()
111 nm, vma->vm_start, vma->vm_end); in show_faulting_vma()
/arch/parisc/mm/
Dfault.c126 if (tree->vm_start > addr) {
132 if (prev->vm_next->vm_start > addr)
196 vma->vm_start, vma->vm_end); in show_signal_msg()
229 if (!vma || address < vma->vm_start) in do_page_fault()
306 address < vma->vm_start || address >= vma->vm_end) { in do_page_fault()
/arch/powerpc/kernel/
Dproc_powerpc.c46 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) in page_map_mmap()
49 remap_pfn_range(vma, vma->vm_start, in page_map_mmap()
/arch/um/drivers/
Dmmapper_kern.c56 size = vma->vm_end - vma->vm_start; in mmapper_mmap()
64 if (remap_pfn_range(vma, vma->vm_start, p_buf >> PAGE_SHIFT, size, in mmapper_mmap()
/arch/microblaze/kernel/
Ddma.c163 unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; in dma_direct_mmap_coherent()
177 return remap_pfn_range(vma, vma->vm_start, pfn + off, in dma_direct_mmap_coherent()
178 vma->vm_end - vma->vm_start, vma->vm_page_prot); in dma_direct_mmap_coherent()
/arch/tile/mm/
Delf.c78 if (vma->vm_start == (ELF_ET_DYN_BASE & PAGE_MASK)) { in notify_exec()
82 snprintf(buf, sizeof(buf), "0x%lx:@", vma->vm_start); in notify_exec()
/arch/cris/arch-v32/drivers/pci/
Dbios.c31 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_page_range()
32 vma->vm_end - vma->vm_start, in pci_mmap_page_range()
/arch/powerpc/include/asm/
Dfb.h12 vma->vm_end - vma->vm_start, in fb_pgprotect()
/arch/nios2/include/asm/
Dtlb.h25 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
/arch/ia64/pci/
Dpci.c425 unsigned long size = vma->vm_end - vma->vm_start; in pci_mmap_page_range()
455 efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in pci_mmap_page_range()
460 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_page_range()
461 vma->vm_end - vma->vm_start, vma->vm_page_prot)) in pci_mmap_page_range()
496 unsigned long size = vma->vm_end - vma->vm_start; in pci_mmap_legacy_page_range()
520 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_legacy_page_range()
/arch/mips/include/asm/
Dtlb.h11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
/arch/metag/mm/
Dhugetlbpage.c50 if (addr + len > vma->vm_start) in prepare_hugepage_range()
53 (ALIGN_HUGEPT(addr + len) > vma->vm_start)) in prepare_hugepage_range()
157 <= vma->vm_start)) { in hugetlb_get_unmapped_area_existing()
/arch/nios2/kernel/
Dsys_nios2.c42 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) in sys_cacheflush()

12345