• Home
  • Raw
  • Download

Lines Matching refs:pages

54 void put_user_pages_dirty_lock(struct page **pages, unsigned long npages,  in put_user_pages_dirty_lock()  argument
66 put_user_pages(pages, npages); in put_user_pages_dirty_lock()
71 struct page *page = compound_head(pages[index]); in put_user_pages_dirty_lock()
108 void put_user_pages(struct page **pages, unsigned long npages) in put_user_pages() argument
118 put_user_page(pages[index]); in put_user_pages()
781 unsigned int gup_flags, struct page **pages, in __get_user_pages() argument
793 VM_BUG_ON(!!pages != !!(gup_flags & FOLL_GET)); in __get_user_pages()
814 pages ? &pages[i] : NULL); in __get_user_pages()
826 i = follow_hugetlb_page(mm, vma, pages, vmas, in __get_user_pages()
871 if (pages) { in __get_user_pages()
872 pages[i] = page; in __get_user_pages()
1002 struct page **pages, in __get_user_pages_locked() argument
1017 if (pages) in __get_user_pages_locked()
1023 ret = __get_user_pages(tsk, mm, start, nr_pages, flags, pages, in __get_user_pages_locked()
1054 if (likely(pages)) in __get_user_pages_locked()
1055 pages += ret; in __get_user_pages_locked()
1067 pages, NULL, NULL); in __get_user_pages_locked()
1078 if (likely(pages)) in __get_user_pages_locked()
1079 pages++; in __get_user_pages_locked()
1151 unsigned int gup_flags, struct page **pages, in get_user_pages_remote() argument
1163 return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, in get_user_pages_remote()
1318 unsigned long nr_pages, struct page **pages, in __get_user_pages_locked() argument
1344 if (pages) { in __get_user_pages_locked()
1345 pages[i] = virt_to_page(start); in __get_user_pages_locked()
1346 if (pages[i]) in __get_user_pages_locked()
1347 get_page(pages[i]); in __get_user_pages_locked()
1437 struct page **pages, in check_and_migrate_cma_pages() argument
1450 struct page *head = compound_head(pages[i]); in check_and_migrate_cma_pages()
1456 step = compound_nr(head) - (pages[i] - head); in check_and_migrate_cma_pages()
1489 put_page(pages[i]); in check_and_migrate_cma_pages()
1508 pages, vmas, NULL, in check_and_migrate_cma_pages()
1524 struct page **pages, in check_and_migrate_cma_pages() argument
1540 struct page **pages, in __gup_longterm_locked() argument
1549 if (!pages) in __gup_longterm_locked()
1562 rc = __get_user_pages_locked(tsk, mm, start, nr_pages, pages, in __gup_longterm_locked()
1572 put_page(pages[i]); in __gup_longterm_locked()
1577 rc = check_and_migrate_cma_pages(tsk, mm, start, rc, pages, in __gup_longterm_locked()
1591 struct page **pages, in __gup_longterm_locked() argument
1595 return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, in __gup_longterm_locked()
1608 unsigned int gup_flags, struct page **pages, in get_user_pages() argument
1612 pages, vmas, gup_flags | FOLL_TOUCH); in get_user_pages()
1638 unsigned int gup_flags, struct page **pages, in get_user_pages_locked() argument
1651 pages, NULL, locked, in get_user_pages_locked()
1672 struct page **pages, unsigned int gup_flags) in get_user_pages_unlocked() argument
1688 ret = __get_user_pages_locked(current, mm, start, nr_pages, pages, NULL, in get_user_pages_unlocked()
1786 struct page **pages) in undo_dev_pagemap() argument
1789 struct page *page = pages[--(*nr)]; in undo_dev_pagemap()
1812 unsigned int flags, struct page **pages, int *nr) in gup_pte_range() argument
1839 undo_dev_pagemap(nr, nr_start, pages); in gup_pte_range()
1860 pages[*nr] = page; in gup_pte_range()
1885 unsigned int flags, struct page **pages, int *nr) in gup_pte_range() argument
1893 unsigned long end, struct page **pages, int *nr) in __gup_device_huge() argument
1903 undo_dev_pagemap(nr, nr_start, pages); in __gup_device_huge()
1907 pages[*nr] = page; in __gup_device_huge()
1919 unsigned long end, struct page **pages, int *nr) in __gup_device_huge_pmd() argument
1925 if (!__gup_device_huge(fault_pfn, addr, end, pages, nr)) in __gup_device_huge_pmd()
1929 undo_dev_pagemap(nr, nr_start, pages); in __gup_device_huge_pmd()
1936 unsigned long end, struct page **pages, int *nr) in __gup_device_huge_pud() argument
1942 if (!__gup_device_huge(fault_pfn, addr, end, pages, nr)) in __gup_device_huge_pud()
1946 undo_dev_pagemap(nr, nr_start, pages); in __gup_device_huge_pud()
1953 unsigned long end, struct page **pages, int *nr) in __gup_device_huge_pmd() argument
1960 unsigned long end, struct page **pages, int *nr) in __gup_device_huge_pud() argument
1977 struct page **pages, int *nr) in gup_hugepte() argument
2002 pages[*nr] = page; in gup_hugepte()
2028 struct page **pages, int *nr) in gup_huge_pd() argument
2037 if (!gup_hugepte(ptep, sz, addr, end, flags, pages, nr)) in gup_huge_pd()
2046 struct page **pages, int *nr) in gup_huge_pd() argument
2054 struct page **pages, int *nr) in gup_huge_pmd() argument
2065 return __gup_device_huge_pmd(orig, pmdp, addr, end, pages, nr); in gup_huge_pmd()
2071 pages[*nr] = page; in gup_huge_pmd()
2095 unsigned long end, unsigned int flags, struct page **pages, int *nr) in gup_huge_pud() argument
2106 return __gup_device_huge_pud(orig, pudp, addr, end, pages, nr); in gup_huge_pud()
2112 pages[*nr] = page; in gup_huge_pud()
2137 struct page **pages, int *nr) in gup_huge_pgd() argument
2149 pages[*nr] = page; in gup_huge_pgd()
2173 unsigned int flags, struct page **pages, int *nr) in gup_pmd_range() argument
2197 pages, nr)) in gup_pmd_range()
2206 PMD_SHIFT, next, flags, pages, nr)) in gup_pmd_range()
2208 } else if (!gup_pte_range(pmd, addr, next, flags, pages, nr)) in gup_pmd_range()
2216 unsigned int flags, struct page **pages, int *nr) in gup_pud_range() argument
2230 pages, nr)) in gup_pud_range()
2234 PUD_SHIFT, next, flags, pages, nr)) in gup_pud_range()
2236 } else if (!gup_pmd_range(pud, addr, next, flags, pages, nr)) in gup_pud_range()
2244 unsigned int flags, struct page **pages, int *nr) in gup_p4d_range() argument
2259 P4D_SHIFT, next, flags, pages, nr)) in gup_p4d_range()
2261 } else if (!gup_pud_range(p4d, addr, next, flags, pages, nr)) in gup_p4d_range()
2269 unsigned int flags, struct page **pages, int *nr) in gup_pgd_range() argument
2283 pages, nr)) in gup_pgd_range()
2287 PGDIR_SHIFT, next, flags, pages, nr)) in gup_pgd_range()
2289 } else if (!gup_p4d_range(pgd, addr, next, flags, pages, nr)) in gup_pgd_range()
2295 unsigned int flags, struct page **pages, int *nr) in gup_pgd_range() argument
2321 struct page **pages) in __get_user_pages_fast() argument
2351 gup_pgd_range(start, end, write ? FOLL_WRITE : 0, pages, &nr); in __get_user_pages_fast()
2360 unsigned int gup_flags, struct page **pages) in __gup_longterm_unlocked() argument
2372 pages, NULL, gup_flags); in __gup_longterm_unlocked()
2376 pages, gup_flags); in __gup_longterm_unlocked()
2399 unsigned int gup_flags, struct page **pages) in get_user_pages_fast() argument
2420 gup_pgd_range(addr, end, gup_flags, pages, &nr); in get_user_pages_fast()
2428 pages += nr; in get_user_pages_fast()
2431 gup_flags, pages); in get_user_pages_fast()