Lines Matching refs:tmp
486 struct vm_area_struct *mpnt, *tmp, *prev, **pprev; in dup_mmap() local
546 tmp = vm_area_dup(mpnt); in dup_mmap()
547 if (!tmp) in dup_mmap()
549 retval = vma_dup_policy(mpnt, tmp); in dup_mmap()
552 tmp->vm_mm = mm; in dup_mmap()
553 retval = dup_userfaultfd(tmp, &uf); in dup_mmap()
556 if (tmp->vm_flags & VM_WIPEONFORK) { in dup_mmap()
558 tmp->anon_vma = NULL; in dup_mmap()
559 if (anon_vma_prepare(tmp)) in dup_mmap()
561 } else if (anon_vma_fork(tmp, mpnt)) in dup_mmap()
563 tmp->vm_flags &= ~(VM_LOCKED | VM_LOCKONFAULT); in dup_mmap()
564 tmp->vm_next = tmp->vm_prev = NULL; in dup_mmap()
565 file = tmp->vm_file; in dup_mmap()
571 if (tmp->vm_flags & VM_DENYWRITE) in dup_mmap()
574 if (tmp->vm_flags & VM_SHARED) in dup_mmap()
578 vma_interval_tree_insert_after(tmp, mpnt, in dup_mmap()
589 if (is_vm_hugetlb_page(tmp)) in dup_mmap()
590 reset_vma_resv_huge_pages(tmp); in dup_mmap()
595 *pprev = tmp; in dup_mmap()
596 pprev = &tmp->vm_next; in dup_mmap()
597 tmp->vm_prev = prev; in dup_mmap()
598 prev = tmp; in dup_mmap()
600 __vma_link_rb(mm, tmp, rb_link, rb_parent); in dup_mmap()
601 rb_link = &tmp->vm_rb.rb_right; in dup_mmap()
602 rb_parent = &tmp->vm_rb; in dup_mmap()
605 if (!(tmp->vm_flags & VM_WIPEONFORK)) in dup_mmap()
608 if (tmp->vm_ops && tmp->vm_ops->open) in dup_mmap()
609 tmp->vm_ops->open(tmp); in dup_mmap()
625 mpol_put(vma_policy(tmp)); in dup_mmap()
627 vm_area_free(tmp); in dup_mmap()