Lines Matching refs:nr_pages
66 unsigned long nr_pages);
454 unsigned long nr_pages; in privcmd_ioctl_mmap_batch() local
482 nr_pages = DIV_ROUND_UP(m.num, XEN_PFN_PER_PAGE); in privcmd_ioctl_mmap_batch()
483 if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) in privcmd_ioctl_mmap_batch()
525 m.addr + (nr_pages << PAGE_SHIFT) != vma->vm_end) { in privcmd_ioctl_mmap_batch()
530 ret = alloc_empty_pages(vma, nr_pages); in privcmd_ioctl_mmap_batch()
537 m.addr + (nr_pages << PAGE_SHIFT) > vma->vm_end) { in privcmd_ioctl_mmap_batch()
541 if (privcmd_vma_range_is_mapped(vma, m.addr, nr_pages)) { in privcmd_ioctl_mmap_batch()
586 struct page *pages[], unsigned int nr_pages) in lock_pages() argument
597 if (requested > nr_pages) in lock_pages()
606 nr_pages -= pinned; in lock_pages()
613 static void unlock_pages(struct page *pages[], unsigned int nr_pages) in unlock_pages() argument
620 for (i = 0; i < nr_pages; i++) { in unlock_pages()
631 unsigned int nr_pages = 0; in privcmd_ioctl_dm_op() local
672 nr_pages += DIV_ROUND_UP( in privcmd_ioctl_dm_op()
677 pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL); in privcmd_ioctl_dm_op()
689 rc = lock_pages(kbufs, kdata.num, pages, nr_pages); in privcmd_ioctl_dm_op()
703 unlock_pages(pages, nr_pages); in privcmd_ioctl_dm_op()
847 unsigned long nr_pages) in privcmd_vma_range_is_mapped() argument
849 return apply_to_page_range(vma->vm_mm, addr, nr_pages << PAGE_SHIFT, in privcmd_vma_range_is_mapped()