Lines Matching refs:next
286 struct vm_area_struct *next = vma->vm_next; in remove_vma() local
295 return next; in remove_vma()
305 struct vm_area_struct *next; in SYSCALL_DEFINE1() local
350 next = find_vma(mm, oldbrk); in SYSCALL_DEFINE1()
351 if (next && newbrk + PAGE_SIZE > vm_start_gap(next)) in SYSCALL_DEFINE1()
740 struct vm_area_struct *next; in __vma_unlink() local
743 prev->vm_next = next = vma->vm_next; in __vma_unlink()
744 if (next) in __vma_unlink()
745 next->vm_prev = prev; in __vma_unlink()
762 struct vm_area_struct *next = vma->vm_next; in vma_adjust() local
772 if (next && !insert) { in vma_adjust()
775 if (end >= next->vm_end) { in vma_adjust()
780 again: remove_next = 1 + (end > next->vm_end); in vma_adjust()
781 end = next->vm_end; in vma_adjust()
782 exporter = next; in vma_adjust()
784 } else if (end > next->vm_start) { in vma_adjust()
789 adjust_next = (end - next->vm_start) >> PAGE_SHIFT; in vma_adjust()
790 exporter = next; in vma_adjust()
800 importer = next; in vma_adjust()
825 uprobe_munmap(next, next->vm_start, in vma_adjust()
826 next->vm_end); in vma_adjust()
845 anon_vma = next->anon_vma; in vma_adjust()
847 VM_BUG_ON_VMA(adjust_next && next->anon_vma && in vma_adjust()
848 anon_vma != next->anon_vma, next); in vma_adjust()
852 anon_vma_interval_tree_pre_update_vma(next); in vma_adjust()
859 vma_interval_tree_remove(next, root); in vma_adjust()
872 next->vm_start += adjust_next << PAGE_SHIFT; in vma_adjust()
873 next->vm_pgoff += adjust_next; in vma_adjust()
878 vma_interval_tree_insert(next, root); in vma_adjust()
888 __vma_unlink(mm, next, vma); in vma_adjust()
890 __remove_shared_vm_struct(next, file, mapping); in vma_adjust()
902 if (!next) in vma_adjust()
905 vma_gap_update(next); in vma_adjust()
912 anon_vma_interval_tree_post_update_vma(next); in vma_adjust()
922 uprobe_mmap(next); in vma_adjust()
927 uprobe_munmap(next, next->vm_start, next->vm_end); in vma_adjust()
930 if (next->anon_vma) in vma_adjust()
931 anon_vma_merge(vma, next); in vma_adjust()
933 mpol_put(vma_policy(next)); in vma_adjust()
934 kmem_cache_free(vm_area_cachep, next); in vma_adjust()
940 next = vma->vm_next; in vma_adjust()
943 else if (next) in vma_adjust()
944 vma_gap_update(next); in vma_adjust()
1080 struct vm_area_struct *area, *next; in vma_merge() local
1091 next = prev->vm_next; in vma_merge()
1093 next = mm->mmap; in vma_merge()
1094 area = next; in vma_merge()
1095 if (next && next->vm_end == end) /* cases 6, 7, 8 */ in vma_merge()
1096 next = next->vm_next; in vma_merge()
1108 if (next && end == next->vm_start && in vma_merge()
1109 mpol_equal(policy, vma_policy(next)) && in vma_merge()
1110 can_vma_merge_before(next, vm_flags, anon_vma, in vma_merge()
1113 next->anon_vma, NULL)) { in vma_merge()
1116 next->vm_end, prev->vm_pgoff, NULL); in vma_merge()
1129 if (next && end == next->vm_start && in vma_merge()
1130 mpol_equal(policy, vma_policy(next)) && in vma_merge()
1131 can_vma_merge_before(next, vm_flags, anon_vma, in vma_merge()
1137 err = vma_adjust(area, addr, next->vm_end, in vma_merge()
1138 next->vm_pgoff - pglen, NULL); in vma_merge()
2192 struct vm_area_struct *next; in expand_upwards() local
2212 next = vma->vm_next; in expand_upwards()
2213 if (next && next->vm_start < gap_addr && in expand_upwards()
2214 (next->vm_flags & (VM_WRITE|VM_READ|VM_EXEC))) { in expand_upwards()
2215 if (!(next->vm_flags & VM_GROWSUP)) in expand_upwards()
2451 struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; in unmap_region() local
2459 next ? next->vm_start : USER_PGTABLES_CEILING); in unmap_region()