Lines Matching refs:src_mm
700 copy_nonpresent_pte(struct mm_struct *dst_mm, struct mm_struct *src_mm, in copy_nonpresent_pte() argument
718 &src_mm->mmlist); in copy_nonpresent_pte()
739 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte()
770 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte()
804 struct mm_struct *src_mm = src_vma->vm_mm; in copy_present_page() local
823 if (likely(!atomic_read(&src_mm->has_pinned))) in copy_present_page()
873 struct mm_struct *src_mm = src_vma->vm_mm; in copy_present_pte() local
897 ptep_set_wrprotect(src_mm, addr, src_pte); in copy_present_pte()
917 page_copy_prealloc(struct mm_struct *src_mm, struct vm_area_struct *vma, in page_copy_prealloc() argument
926 if (mem_cgroup_charge(new_page, src_mm, GFP_KERNEL)) { in page_copy_prealloc()
941 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pte_range() local
960 src_ptl = pte_lockptr(src_mm, src_pmd); in copy_pte_range()
982 entry.val = copy_nonpresent_pte(dst_mm, src_mm, in copy_pte_range()
1028 prealloc = page_copy_prealloc(src_mm, src_vma, addr); in copy_pte_range()
1048 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pmd_range() local
1062 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range()
1085 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pud_range() local
1099 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range()
1148 struct mm_struct *src_mm = src_vma->vm_mm; in copy_page_range() local
1164 return copy_hugetlb_page_range(dst_mm, src_mm, src_vma); in copy_page_range()
1186 0, src_vma, src_mm, addr, end); in copy_page_range()
1195 mmap_assert_write_locked(src_mm); in copy_page_range()
1196 raw_write_seqcount_begin(&src_mm->write_protect_seq); in copy_page_range()
1201 src_pgd = pgd_offset(src_mm, addr); in copy_page_range()
1214 raw_write_seqcount_end(&src_mm->write_protect_seq); in copy_page_range()