Lines Matching refs:end
75 unsigned long start, unsigned long end);
488 unsigned long end, struct vm_area_struct **pprev, in find_vma_links() argument
504 if (vma_tmp->vm_start < end) in find_vma_links()
522 unsigned long addr, unsigned long end) in count_vma_pages_range() argument
528 vma = find_vma_intersection(mm, addr, end); in count_vma_pages_range()
532 nr_pages = (min(end, vma->vm_end) - in count_vma_pages_range()
539 if (vma->vm_start > end) in count_vma_pages_range()
542 overlap_len = min(end, vma->vm_end) - vma->vm_start; in count_vma_pages_range()
679 unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert, in __vma_adjust() argument
695 if (end >= next->vm_end) { in __vma_adjust()
707 VM_WARN_ON(end != next->vm_end); in __vma_adjust()
722 remove_next = 1 + (end > next->vm_end); in __vma_adjust()
724 end != next->vm_next->vm_end); in __vma_adjust()
726 end != next->vm_end); in __vma_adjust()
728 end = next->vm_end; in __vma_adjust()
741 } else if (end > next->vm_start) { in __vma_adjust()
746 adjust_next = (end - next->vm_start) >> PAGE_SHIFT; in __vma_adjust()
750 } else if (end < vma->vm_end) { in __vma_adjust()
756 adjust_next = -((vma->vm_end - end) >> PAGE_SHIFT); in __vma_adjust()
777 vma_adjust_trans_huge(orig_vma, start, end, adjust_next); in __vma_adjust()
822 if (end != vma->vm_end) { in __vma_adjust()
823 vma->vm_end = end; in __vma_adjust()
931 end = next->vm_end; in __vma_adjust()
1103 unsigned long end, unsigned long vm_flags, in vma_merge() argument
1109 pgoff_t pglen = (end - addr) >> PAGE_SHIFT; in vma_merge()
1125 if (area && area->vm_end == end) /* cases 6, 7, 8 */ in vma_merge()
1130 VM_WARN_ON(area && end > area->vm_end); in vma_merge()
1131 VM_WARN_ON(addr >= end); in vma_merge()
1145 if (next && end == next->vm_start && in vma_merge()
1160 end, prev->vm_pgoff, NULL, prev); in vma_merge()
1170 if (next && end == next->vm_start && in vma_merge()
2497 unsigned long start, unsigned long end) in unmap_region() argument
2503 tlb_gather_mmu(&tlb, mm, start, end); in unmap_region()
2505 unmap_vmas(&tlb, vma, start, end); in unmap_region()
2508 tlb_finish_mmu(&tlb, start, end); in unmap_region()
2517 struct vm_area_struct *prev, unsigned long end) in detach_vmas_to_be_unmapped() argument
2529 } while (vma && vma->vm_start < end); in detach_vmas_to_be_unmapped()
2631 unsigned long end; in do_munmap() local
2649 end = start + len; in do_munmap()
2650 if (vma->vm_start >= end) in do_munmap()
2668 if (end < vma->vm_end && mm->map_count >= sysctl_max_map_count) in do_munmap()
2678 last = find_vma(mm, end); in do_munmap()
2679 if (last && end > last->vm_start) { in do_munmap()
2680 int error = __split_vma(mm, last, end, 1); in do_munmap()
2691 while (tmp && tmp->vm_start < end) { in do_munmap()
2703 detach_vmas_to_be_unmapped(mm, vma, prev, end); in do_munmap()
2704 unmap_region(mm, vma, prev, start, end); in do_munmap()
2706 arch_unmap(mm, vma, start, end); in do_munmap()