/kernel/linux/linux-5.10/include/linux/ |
D | pagevec.h | 26 void __pagevec_release(struct pagevec *pvec); 27 void __pagevec_lru_add(struct pagevec *pvec); 28 unsigned pagevec_lookup_entries(struct pagevec *pvec, 32 void pagevec_remove_exceptionals(struct pagevec *pvec); 33 unsigned pagevec_lookup_range(struct pagevec *pvec, 36 static inline unsigned pagevec_lookup(struct pagevec *pvec, in pagevec_lookup() argument 40 return pagevec_lookup_range(pvec, mapping, start, (pgoff_t)-1); in pagevec_lookup() 43 unsigned pagevec_lookup_range_tag(struct pagevec *pvec, 46 unsigned pagevec_lookup_range_nr_tag(struct pagevec *pvec, 49 static inline unsigned pagevec_lookup_tag(struct pagevec *pvec, in pagevec_lookup_tag() argument [all …]
|
/kernel/linux/linux-5.10/mm/ |
D | swap.c | 51 struct pagevec pvec; member 206 static void pagevec_lru_move_fn(struct pagevec *pvec, in pagevec_lru_move_fn() argument 215 for (i = 0; i < pagevec_count(pvec); i++) { in pagevec_lru_move_fn() 216 struct page *page = pvec->pages[i]; in pagevec_lru_move_fn() 231 release_pages(pvec->pages, pvec->nr); in pagevec_lru_move_fn() 232 pagevec_reinit(pvec); in pagevec_lru_move_fn() 252 static void pagevec_move_tail(struct pagevec *pvec) in pagevec_move_tail() argument 256 pagevec_lru_move_fn(pvec, pagevec_move_tail_fn, &pgmoved); in pagevec_move_tail() 269 struct pagevec *pvec; in rotate_reclaimable_page() local 274 pvec = this_cpu_ptr(&lru_rotate.pvec); in rotate_reclaimable_page() [all …]
|
D | truncate.c | 60 struct pagevec *pvec, pgoff_t *indices, in truncate_exceptional_pvec_entries() argument 70 for (j = 0; j < pagevec_count(pvec); j++) in truncate_exceptional_pvec_entries() 71 if (xa_is_value(pvec->pages[j])) in truncate_exceptional_pvec_entries() 74 if (j == pagevec_count(pvec)) in truncate_exceptional_pvec_entries() 82 for (i = j; i < pagevec_count(pvec); i++) { in truncate_exceptional_pvec_entries() 83 struct page *page = pvec->pages[i]; in truncate_exceptional_pvec_entries() 87 pvec->pages[j++] = page; in truncate_exceptional_pvec_entries() 104 pvec->nr = j; in truncate_exceptional_pvec_entries() 295 struct pagevec pvec; in truncate_inode_pages_range() local 324 pagevec_init(&pvec); in truncate_inode_pages_range() [all …]
|
D | mlock.c | 251 static bool __putback_lru_fast_prepare(struct page *page, struct pagevec *pvec, in __putback_lru_fast_prepare() argument 258 pagevec_add(pvec, page); in __putback_lru_fast_prepare() 274 static void __putback_lru_fast(struct pagevec *pvec, int pgrescued) in __putback_lru_fast() argument 276 count_vm_events(UNEVICTABLE_PGMUNLOCKED, pagevec_count(pvec)); in __putback_lru_fast() 281 __pagevec_lru_add(pvec); in __putback_lru_fast() 295 static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone) in __munlock_pagevec() argument 298 int nr = pagevec_count(pvec); in __munlock_pagevec() 308 struct page *page = pvec->pages[i]; in __munlock_pagevec() 329 pagevec_add(&pvec_putback, pvec->pages[i]); in __munlock_pagevec() 330 pvec->pages[i] = NULL; in __munlock_pagevec() [all …]
|
D | shmem.c | 849 struct pagevec pvec; in shmem_unlock_mapping() local 853 pagevec_init(&pvec); in shmem_unlock_mapping() 862 pvec.nr = find_get_entries(mapping, index, in shmem_unlock_mapping() 863 PAGEVEC_SIZE, pvec.pages, indices); in shmem_unlock_mapping() 864 if (!pvec.nr) in shmem_unlock_mapping() 866 index = indices[pvec.nr - 1] + 1; in shmem_unlock_mapping() 867 pagevec_remove_exceptionals(&pvec); in shmem_unlock_mapping() 868 check_move_unevictable_pages(&pvec); in shmem_unlock_mapping() 869 pagevec_release(&pvec); in shmem_unlock_mapping() 913 struct pagevec pvec; in shmem_undo_range() local [all …]
|
D | filemap.c | 296 struct pagevec *pvec) in page_cache_delete_batch() argument 298 XA_STATE(xas, &mapping->i_pages, pvec->pages[0]->index); in page_cache_delete_batch() 305 if (i >= pagevec_count(pvec)) in page_cache_delete_batch() 318 if (page != pvec->pages[i]) { in page_cache_delete_batch() 319 VM_BUG_ON_PAGE(page->index > pvec->pages[i]->index, in page_cache_delete_batch() 344 struct pagevec *pvec) in delete_from_page_cache_batch() argument 349 if (!pagevec_count(pvec)) in delete_from_page_cache_batch() 353 for (i = 0; i < pagevec_count(pvec); i++) { in delete_from_page_cache_batch() 354 trace_mm_filemap_delete_from_page_cache(pvec->pages[i]); in delete_from_page_cache_batch() 356 unaccount_page_cache_page(mapping, pvec->pages[i]); in delete_from_page_cache_batch() [all …]
|
D | page-writeback.c | 2170 struct pagevec pvec; in write_cache_pages() local 2178 pagevec_init(&pvec); in write_cache_pages() 2198 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages() 2204 struct page *page = pvec.pages[i]; in write_cache_pages() 2280 pagevec_release(&pvec); in write_cache_pages()
|
/kernel/linux/linux-5.10/fs/nilfs2/ |
D | page.c | 243 struct pagevec pvec; in nilfs_copy_dirty_pages() local 248 pagevec_init(&pvec); in nilfs_copy_dirty_pages() 250 if (!pagevec_lookup_tag(&pvec, smap, &index, PAGECACHE_TAG_DIRTY)) in nilfs_copy_dirty_pages() 253 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_dirty_pages() 254 struct page *page = pvec.pages[i], *dpage; in nilfs_copy_dirty_pages() 278 pagevec_release(&pvec); in nilfs_copy_dirty_pages() 297 struct pagevec pvec; in nilfs_copy_back_pages() local 301 pagevec_init(&pvec); in nilfs_copy_back_pages() 303 n = pagevec_lookup(&pvec, smap, &index); in nilfs_copy_back_pages() 307 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_back_pages() [all …]
|
D | segment.c | 707 struct pagevec pvec; in nilfs_lookup_dirty_data_buffers() local 721 pagevec_init(&pvec); in nilfs_lookup_dirty_data_buffers() 724 !pagevec_lookup_range_tag(&pvec, mapping, &index, last, in nilfs_lookup_dirty_data_buffers() 728 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_data_buffers() 730 struct page *page = pvec.pages[i]; in nilfs_lookup_dirty_data_buffers() 750 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers() 756 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers() 766 struct pagevec pvec; in nilfs_lookup_dirty_node_buffers() local 774 pagevec_init(&pvec); in nilfs_lookup_dirty_node_buffers() 776 while (pagevec_lookup_tag(&pvec, btnc_inode->i_mapping, &index, in nilfs_lookup_dirty_node_buffers() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/ |
D | i915_gem_shmem.c | 21 static void check_release_pagevec(struct pagevec *pvec) in check_release_pagevec() argument 23 check_move_unevictable_pages(pvec); in check_release_pagevec() 24 __pagevec_release(pvec); in check_release_pagevec() 194 struct pagevec pvec; in shmem_get_pages() local 196 pagevec_init(&pvec); in shmem_get_pages() 198 if (!pagevec_add(&pvec, page)) in shmem_get_pages() 199 check_release_pagevec(&pvec); in shmem_get_pages() 201 if (pagevec_count(&pvec)) in shmem_get_pages() 202 check_release_pagevec(&pvec); in shmem_get_pages() 303 struct pagevec pvec; in shmem_put_pages() local [all …]
|
D | i915_gem_userptr.c | 401 struct page **pvec, unsigned long num_pages) in __i915_gem_userptr_alloc_pages() argument 414 sg = __sg_alloc_table_from_pages(st, pvec, num_pages, 0, in __i915_gem_userptr_alloc_pages() 449 struct page **pvec; in __i915_gem_userptr_get_pages_worker() local 455 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in __i915_gem_userptr_get_pages_worker() 456 if (pvec != NULL) { in __i915_gem_userptr_get_pages_worker() 476 pvec + pinned, NULL, &locked); in __i915_gem_userptr_get_pages_worker() 493 pages = __i915_gem_userptr_alloc_pages(obj, pvec, in __i915_gem_userptr_get_pages_worker() 507 unpin_user_pages(pvec, pinned); in __i915_gem_userptr_get_pages_worker() 508 kvfree(pvec); in __i915_gem_userptr_get_pages_worker() 560 struct page **pvec; in i915_gem_userptr_get_pages() local [all …]
|
/kernel/linux/linux-5.10/fs/iomap/ |
D | seek.c | 80 struct pagevec pvec; in page_cache_seek_hole_data() local 85 pagevec_init(&pvec); in page_cache_seek_hole_data() 90 nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, &index, in page_cache_seek_hole_data() 96 struct page *page = pvec.pages[i]; in page_cache_seek_hole_data() 102 pagevec_release(&pvec); in page_cache_seek_hole_data() 115 pagevec_release(&pvec); in page_cache_seek_hole_data()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
D | drm_gem.c | 509 static void drm_gem_check_release_pagevec(struct pagevec *pvec) in drm_gem_check_release_pagevec() argument 511 check_move_unevictable_pages(pvec); in drm_gem_check_release_pagevec() 512 __pagevec_release(pvec); in drm_gem_check_release_pagevec() 545 struct pagevec pvec; in drm_gem_get_pages() local 588 pagevec_init(&pvec); in drm_gem_get_pages() 590 if (!pagevec_add(&pvec, pages[i])) in drm_gem_get_pages() 591 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages() 593 if (pagevec_count(&pvec)) in drm_gem_get_pages() 594 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages() 613 struct pagevec pvec; in drm_gem_put_pages() local [all …]
|
/kernel/linux/linux-5.10/fs/hmdfs/ |
D | client_writeback.c | 97 struct pagevec pvec; in hmdfs_idirty_pages() local 102 pagevec_init(&pvec); in hmdfs_idirty_pages() 104 pagevec_init(&pvec, 0); in hmdfs_idirty_pages() 106 while (pagevec_lookup_tag(&pvec, inode->i_mapping, &index, tag)) { in hmdfs_idirty_pages() 107 nr_dirty_pages += pagevec_count(&pvec); in hmdfs_idirty_pages() 108 pagevec_release(&pvec); in hmdfs_idirty_pages()
|
/kernel/linux/linux-5.10/fs/gfs2/ |
D | aops.c | 234 struct pagevec *pvec, in gfs2_write_jdata_pagevec() argument 249 struct page *page = pvec->pages[i]; in gfs2_write_jdata_pagevec() 332 struct pagevec pvec; in gfs2_write_cache_jdata() local 342 pagevec_init(&pvec); in gfs2_write_cache_jdata() 368 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in gfs2_write_cache_jdata() 373 ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index); in gfs2_write_cache_jdata() 378 pagevec_release(&pvec); in gfs2_write_cache_jdata()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 658 struct page **pvec = NULL; in etnaviv_gem_userptr_get_pages() local 667 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_userptr_get_pages() 668 if (!pvec) in etnaviv_gem_userptr_get_pages() 674 struct page **pages = pvec + pinned; in etnaviv_gem_userptr_get_pages() 679 unpin_user_pages(pvec, pinned); in etnaviv_gem_userptr_get_pages() 680 kvfree(pvec); in etnaviv_gem_userptr_get_pages() 688 etnaviv_obj->pages = pvec; in etnaviv_gem_userptr_get_pages()
|
/kernel/linux/linux-5.10/fs/ceph/ |
D | addr.c | 823 struct pagevec pvec; in ceph_writepages_start() local 847 pagevec_init(&pvec); in ceph_writepages_start() 905 pvec_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, in ceph_writepages_start() 911 page = pvec.pages[i]; in ceph_writepages_start() 1031 pvec.pages[i] = NULL; in ceph_writepages_start() 1043 if (!pvec.pages[j]) in ceph_writepages_start() 1046 pvec.pages[n] = pvec.pages[j]; in ceph_writepages_start() 1049 pvec.nr = n; in ceph_writepages_start() 1054 pagevec_release(&pvec); in ceph_writepages_start() 1179 dout("pagevec_release on %d pages (%p)\n", (int)pvec.nr, in ceph_writepages_start() [all …]
|
/kernel/linux/linux-5.10/fs/hugetlbfs/ |
D | inode.c | 111 static void huge_pagevec_release(struct pagevec *pvec) in huge_pagevec_release() argument 115 for (i = 0; i < pagevec_count(pvec); ++i) in huge_pagevec_release() 116 put_page(pvec->pages[i]); in huge_pagevec_release() 118 pagevec_reinit(pvec); in huge_pagevec_release() 473 struct pagevec pvec; in remove_inode_hugepages() local 480 pagevec_init(&pvec); in remove_inode_hugepages() 486 if (!pagevec_lookup_range(&pvec, mapping, &next, end - 1)) in remove_inode_hugepages() 489 for (i = 0; i < pagevec_count(&pvec); ++i) { in remove_inode_hugepages() 490 struct page *page = pvec.pages[i]; in remove_inode_hugepages() 549 huge_pagevec_release(&pvec); in remove_inode_hugepages()
|
/kernel/linux/linux-5.10/fs/f2fs/ |
D | node.c | 1460 struct pagevec pvec; in last_fsync_dnode() local 1464 pagevec_init(&pvec); in last_fsync_dnode() 1467 while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, in last_fsync_dnode() 1472 struct page *page = pvec.pages[i]; in last_fsync_dnode() 1476 pagevec_release(&pvec); in last_fsync_dnode() 1507 pagevec_release(&pvec); in last_fsync_dnode() 1676 struct pagevec pvec; in f2fs_fsync_node_pages() local 1690 pagevec_init(&pvec); in f2fs_fsync_node_pages() 1693 while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, in f2fs_fsync_node_pages() 1698 struct page *page = pvec.pages[i]; in f2fs_fsync_node_pages() [all …]
|
D | checkpoint.c | 387 struct pagevec pvec; in f2fs_sync_meta_pages() local 395 pagevec_init(&pvec); in f2fs_sync_meta_pages() 399 while ((nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, in f2fs_sync_meta_pages() 404 struct page *page = pvec.pages[i]; in f2fs_sync_meta_pages() 409 pagevec_release(&pvec); in f2fs_sync_meta_pages() 439 pagevec_release(&pvec); in f2fs_sync_meta_pages()
|
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
D | mmu.c | 1724 static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, in mmu_pages_add() argument 1730 for (i=0; i < pvec->nr; i++) in mmu_pages_add() 1731 if (pvec->page[i].sp == sp) in mmu_pages_add() 1734 pvec->page[pvec->nr].sp = sp; in mmu_pages_add() 1735 pvec->page[pvec->nr].idx = idx; in mmu_pages_add() 1736 pvec->nr++; in mmu_pages_add() 1737 return (pvec->nr == KVM_PAGE_ARRAY_NR); in mmu_pages_add() 1748 struct kvm_mmu_pages *pvec) in __mmu_unsync_walk() argument 1764 if (mmu_pages_add(pvec, child, i)) in __mmu_unsync_walk() 1767 ret = __mmu_unsync_walk(child, pvec); in __mmu_unsync_walk() [all …]
|
/kernel/linux/linux-5.10/fs/fscache/ |
D | page.c | 1219 struct pagevec pvec; in __fscache_uncache_all_inode_pages() local 1230 pagevec_init(&pvec); in __fscache_uncache_all_inode_pages() 1233 if (!pagevec_lookup(&pvec, mapping, &next)) in __fscache_uncache_all_inode_pages() 1235 for (i = 0; i < pagevec_count(&pvec); i++) { in __fscache_uncache_all_inode_pages() 1236 struct page *page = pvec.pages[i]; in __fscache_uncache_all_inode_pages() 1242 pagevec_release(&pvec); in __fscache_uncache_all_inode_pages()
|
/kernel/linux/linux-5.10/drivers/mtd/lpddr/ |
D | lpddr_cmds.c | 404 unsigned long adr, const struct kvec **pvec, in do_write_buffer() argument 436 vec = *pvec; in do_write_buffer() 466 *pvec = vec; in do_write_buffer()
|
/kernel/linux/linux-5.10/fs/ext4/ |
D | inode.c | 1552 struct pagevec pvec; in mpage_release_unused_pages() local 1577 pagevec_init(&pvec); in mpage_release_unused_pages() 1579 nr_pages = pagevec_lookup_range(&pvec, mapping, &index, end); in mpage_release_unused_pages() 1583 struct page *page = pvec.pages[i]; in mpage_release_unused_pages() 1595 pagevec_release(&pvec); in mpage_release_unused_pages() 2347 struct pagevec pvec; in mpage_map_and_submit_buffers() local 2362 pagevec_init(&pvec); in mpage_map_and_submit_buffers() 2364 nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, in mpage_map_and_submit_buffers() 2369 struct page *page = pvec.pages[i]; in mpage_map_and_submit_buffers() 2385 pagevec_release(&pvec); in mpage_map_and_submit_buffers() [all …]
|
/kernel/linux/linux-5.10/fs/ |
D | buffer.c | 1610 struct pagevec pvec; in clean_bdev_aliases() local 1618 pagevec_init(&pvec); in clean_bdev_aliases() 1619 while (pagevec_lookup_range(&pvec, bd_mapping, &index, end)) { in clean_bdev_aliases() 1620 count = pagevec_count(&pvec); in clean_bdev_aliases() 1622 struct page *page = pvec.pages[i]; in clean_bdev_aliases() 1651 pagevec_release(&pvec); in clean_bdev_aliases()
|