• Home
  • Raw
  • Download

Lines Matching refs:address

139 static int follow_pfn_pte(struct vm_area_struct *vma, unsigned long address,  in follow_pfn_pte()  argument
154 set_pte_at(vma->vm_mm, address, pte, entry); in follow_pfn_pte()
155 update_mmu_cache(vma, address, pte); in follow_pfn_pte()
174 unsigned long address, pmd_t *pmd, unsigned int flags, in follow_page_pte() argument
186 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); in follow_page_pte()
203 migration_entry_wait(mm, pmd, address); in follow_page_pte()
213 page = vm_normal_page(vma, address, pte); in follow_page_pte()
236 ret = follow_pfn_pte(vma, address, ptep, flags); in follow_page_pte()
309 unsigned long address, pud_t *pudp, in follow_pmd_mask() argument
318 pmd = pmd_offset(pudp, address); in follow_pmd_mask()
327 page = follow_huge_pmd(mm, address, pmd, flags); in follow_pmd_mask()
333 page = follow_huge_pd(vma, address, in follow_pmd_mask()
359 page = follow_devmap_pmd(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
365 return follow_page_pte(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
385 return follow_page_pte(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
393 split_huge_pmd(vma, pmd, address); in follow_pmd_mask()
410 split_huge_pmd(vma, pmd, address); in follow_pmd_mask()
415 follow_page_pte(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
417 page = follow_trans_huge_pmd(vma, address, pmd, flags); in follow_pmd_mask()
424 unsigned long address, p4d_t *p4dp, in follow_pud_mask() argument
433 pud = pud_offset(p4dp, address); in follow_pud_mask()
437 page = follow_huge_pud(mm, address, pud, flags); in follow_pud_mask()
443 page = follow_huge_pd(vma, address, in follow_pud_mask()
452 page = follow_devmap_pud(vma, address, pud, flags, &ctx->pgmap); in follow_pud_mask()
460 return follow_pmd_mask(vma, address, pud, flags, ctx); in follow_pud_mask()
464 unsigned long address, pgd_t *pgdp, in follow_p4d_mask() argument
471 p4d = p4d_offset(pgdp, address); in follow_p4d_mask()
479 page = follow_huge_pd(vma, address, in follow_p4d_mask()
486 return follow_pud_mask(vma, address, p4d, flags, ctx); in follow_p4d_mask()
509 unsigned long address, unsigned int flags, in follow_page_mask() argument
519 page = follow_huge_addr(mm, address, flags & FOLL_WRITE); in follow_page_mask()
525 pgd = pgd_offset(mm, address); in follow_page_mask()
531 page = follow_huge_pgd(mm, address, pgd, flags); in follow_page_mask()
537 page = follow_huge_pd(vma, address, in follow_page_mask()
545 return follow_p4d_mask(vma, address, pgd, flags, ctx); in follow_page_mask()
548 struct page *follow_page(struct vm_area_struct *vma, unsigned long address, in follow_page() argument
554 page = follow_page_mask(vma, address, foll_flags, &ctx); in follow_page()
560 static int get_gate_page(struct mm_struct *mm, unsigned long address, in get_gate_page() argument
574 if (address > TASK_SIZE) in get_gate_page()
575 pgd = pgd_offset_k(address); in get_gate_page()
577 pgd = pgd_offset_gate(mm, address); in get_gate_page()
580 p4d = p4d_offset(pgd, address); in get_gate_page()
583 pud = pud_offset(p4d, address); in get_gate_page()
586 pmd = pmd_offset(pud, address); in get_gate_page()
590 pte = pte_offset_map(pmd, address); in get_gate_page()
596 *page = vm_normal_page(*vma, address, *pte); in get_gate_page()
619 unsigned long address, unsigned int *flags, int *nonblocking) in faultin_page() argument
640 ret = handle_mm_fault(vma, address, fault_flags); in faultin_page()
949 unsigned long address, unsigned int fault_flags, in fixup_user_fault() argument
955 address = untagged_addr(address); in fixup_user_fault()
961 vma = find_extend_vma(mm, address); in fixup_user_fault()
962 if (!vma || address < vma->vm_start) in fixup_user_fault()
968 ret = handle_mm_fault(vma, address, fault_flags); in fixup_user_fault()