Home
last modified time | relevance | path

Searched refs:pvec (Results 1 – 25 of 32) sorted by relevance

12

/kernel/linux/linux-5.10/include/linux/
Dpagevec.h26 void __pagevec_release(struct pagevec *pvec);
27 void __pagevec_lru_add(struct pagevec *pvec);
28 unsigned pagevec_lookup_entries(struct pagevec *pvec,
32 void pagevec_remove_exceptionals(struct pagevec *pvec);
33 unsigned pagevec_lookup_range(struct pagevec *pvec,
36 static inline unsigned pagevec_lookup(struct pagevec *pvec, in pagevec_lookup() argument
40 return pagevec_lookup_range(pvec, mapping, start, (pgoff_t)-1); in pagevec_lookup()
43 unsigned pagevec_lookup_range_tag(struct pagevec *pvec,
46 unsigned pagevec_lookup_range_nr_tag(struct pagevec *pvec,
49 static inline unsigned pagevec_lookup_tag(struct pagevec *pvec, in pagevec_lookup_tag() argument
[all …]
/kernel/linux/linux-5.10/mm/
Dswap.c51 struct pagevec pvec; member
206 static void pagevec_lru_move_fn(struct pagevec *pvec, in pagevec_lru_move_fn() argument
215 for (i = 0; i < pagevec_count(pvec); i++) { in pagevec_lru_move_fn()
216 struct page *page = pvec->pages[i]; in pagevec_lru_move_fn()
231 release_pages(pvec->pages, pvec->nr); in pagevec_lru_move_fn()
232 pagevec_reinit(pvec); in pagevec_lru_move_fn()
252 static void pagevec_move_tail(struct pagevec *pvec) in pagevec_move_tail() argument
256 pagevec_lru_move_fn(pvec, pagevec_move_tail_fn, &pgmoved); in pagevec_move_tail()
269 struct pagevec *pvec; in rotate_reclaimable_page() local
274 pvec = this_cpu_ptr(&lru_rotate.pvec); in rotate_reclaimable_page()
[all …]
Dtruncate.c60 struct pagevec *pvec, pgoff_t *indices, in truncate_exceptional_pvec_entries() argument
70 for (j = 0; j < pagevec_count(pvec); j++) in truncate_exceptional_pvec_entries()
71 if (xa_is_value(pvec->pages[j])) in truncate_exceptional_pvec_entries()
74 if (j == pagevec_count(pvec)) in truncate_exceptional_pvec_entries()
82 for (i = j; i < pagevec_count(pvec); i++) { in truncate_exceptional_pvec_entries()
83 struct page *page = pvec->pages[i]; in truncate_exceptional_pvec_entries()
87 pvec->pages[j++] = page; in truncate_exceptional_pvec_entries()
104 pvec->nr = j; in truncate_exceptional_pvec_entries()
295 struct pagevec pvec; in truncate_inode_pages_range() local
324 pagevec_init(&pvec); in truncate_inode_pages_range()
[all …]
Dmlock.c251 static bool __putback_lru_fast_prepare(struct page *page, struct pagevec *pvec, in __putback_lru_fast_prepare() argument
258 pagevec_add(pvec, page); in __putback_lru_fast_prepare()
274 static void __putback_lru_fast(struct pagevec *pvec, int pgrescued) in __putback_lru_fast() argument
276 count_vm_events(UNEVICTABLE_PGMUNLOCKED, pagevec_count(pvec)); in __putback_lru_fast()
281 __pagevec_lru_add(pvec); in __putback_lru_fast()
295 static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone) in __munlock_pagevec() argument
298 int nr = pagevec_count(pvec); in __munlock_pagevec()
308 struct page *page = pvec->pages[i]; in __munlock_pagevec()
329 pagevec_add(&pvec_putback, pvec->pages[i]); in __munlock_pagevec()
330 pvec->pages[i] = NULL; in __munlock_pagevec()
[all …]
Dshmem.c849 struct pagevec pvec; in shmem_unlock_mapping() local
853 pagevec_init(&pvec); in shmem_unlock_mapping()
862 pvec.nr = find_get_entries(mapping, index, in shmem_unlock_mapping()
863 PAGEVEC_SIZE, pvec.pages, indices); in shmem_unlock_mapping()
864 if (!pvec.nr) in shmem_unlock_mapping()
866 index = indices[pvec.nr - 1] + 1; in shmem_unlock_mapping()
867 pagevec_remove_exceptionals(&pvec); in shmem_unlock_mapping()
868 check_move_unevictable_pages(&pvec); in shmem_unlock_mapping()
869 pagevec_release(&pvec); in shmem_unlock_mapping()
913 struct pagevec pvec; in shmem_undo_range() local
[all …]
Dfilemap.c296 struct pagevec *pvec) in page_cache_delete_batch() argument
298 XA_STATE(xas, &mapping->i_pages, pvec->pages[0]->index); in page_cache_delete_batch()
305 if (i >= pagevec_count(pvec)) in page_cache_delete_batch()
318 if (page != pvec->pages[i]) { in page_cache_delete_batch()
319 VM_BUG_ON_PAGE(page->index > pvec->pages[i]->index, in page_cache_delete_batch()
344 struct pagevec *pvec) in delete_from_page_cache_batch() argument
349 if (!pagevec_count(pvec)) in delete_from_page_cache_batch()
353 for (i = 0; i < pagevec_count(pvec); i++) { in delete_from_page_cache_batch()
354 trace_mm_filemap_delete_from_page_cache(pvec->pages[i]); in delete_from_page_cache_batch()
356 unaccount_page_cache_page(mapping, pvec->pages[i]); in delete_from_page_cache_batch()
[all …]
Dpage-writeback.c2170 struct pagevec pvec; in write_cache_pages() local
2178 pagevec_init(&pvec); in write_cache_pages()
2198 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages()
2204 struct page *page = pvec.pages[i]; in write_cache_pages()
2280 pagevec_release(&pvec); in write_cache_pages()
/kernel/linux/linux-5.10/fs/nilfs2/
Dpage.c243 struct pagevec pvec; in nilfs_copy_dirty_pages() local
248 pagevec_init(&pvec); in nilfs_copy_dirty_pages()
250 if (!pagevec_lookup_tag(&pvec, smap, &index, PAGECACHE_TAG_DIRTY)) in nilfs_copy_dirty_pages()
253 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_dirty_pages()
254 struct page *page = pvec.pages[i], *dpage; in nilfs_copy_dirty_pages()
278 pagevec_release(&pvec); in nilfs_copy_dirty_pages()
297 struct pagevec pvec; in nilfs_copy_back_pages() local
301 pagevec_init(&pvec); in nilfs_copy_back_pages()
303 n = pagevec_lookup(&pvec, smap, &index); in nilfs_copy_back_pages()
307 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_back_pages()
[all …]
Dsegment.c707 struct pagevec pvec; in nilfs_lookup_dirty_data_buffers() local
721 pagevec_init(&pvec); in nilfs_lookup_dirty_data_buffers()
724 !pagevec_lookup_range_tag(&pvec, mapping, &index, last, in nilfs_lookup_dirty_data_buffers()
728 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_data_buffers()
730 struct page *page = pvec.pages[i]; in nilfs_lookup_dirty_data_buffers()
750 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers()
756 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers()
766 struct pagevec pvec; in nilfs_lookup_dirty_node_buffers() local
774 pagevec_init(&pvec); in nilfs_lookup_dirty_node_buffers()
776 while (pagevec_lookup_tag(&pvec, btnc_inode->i_mapping, &index, in nilfs_lookup_dirty_node_buffers()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/
Di915_gem_shmem.c21 static void check_release_pagevec(struct pagevec *pvec) in check_release_pagevec() argument
23 check_move_unevictable_pages(pvec); in check_release_pagevec()
24 __pagevec_release(pvec); in check_release_pagevec()
194 struct pagevec pvec; in shmem_get_pages() local
196 pagevec_init(&pvec); in shmem_get_pages()
198 if (!pagevec_add(&pvec, page)) in shmem_get_pages()
199 check_release_pagevec(&pvec); in shmem_get_pages()
201 if (pagevec_count(&pvec)) in shmem_get_pages()
202 check_release_pagevec(&pvec); in shmem_get_pages()
303 struct pagevec pvec; in shmem_put_pages() local
[all …]
Di915_gem_userptr.c401 struct page **pvec, unsigned long num_pages) in __i915_gem_userptr_alloc_pages() argument
414 sg = __sg_alloc_table_from_pages(st, pvec, num_pages, 0, in __i915_gem_userptr_alloc_pages()
449 struct page **pvec; in __i915_gem_userptr_get_pages_worker() local
455 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in __i915_gem_userptr_get_pages_worker()
456 if (pvec != NULL) { in __i915_gem_userptr_get_pages_worker()
476 pvec + pinned, NULL, &locked); in __i915_gem_userptr_get_pages_worker()
493 pages = __i915_gem_userptr_alloc_pages(obj, pvec, in __i915_gem_userptr_get_pages_worker()
507 unpin_user_pages(pvec, pinned); in __i915_gem_userptr_get_pages_worker()
508 kvfree(pvec); in __i915_gem_userptr_get_pages_worker()
560 struct page **pvec; in i915_gem_userptr_get_pages() local
[all …]
/kernel/linux/linux-5.10/fs/iomap/
Dseek.c80 struct pagevec pvec; in page_cache_seek_hole_data() local
85 pagevec_init(&pvec); in page_cache_seek_hole_data()
90 nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, &index, in page_cache_seek_hole_data()
96 struct page *page = pvec.pages[i]; in page_cache_seek_hole_data()
102 pagevec_release(&pvec); in page_cache_seek_hole_data()
115 pagevec_release(&pvec); in page_cache_seek_hole_data()
/kernel/linux/linux-5.10/drivers/gpu/drm/
Ddrm_gem.c509 static void drm_gem_check_release_pagevec(struct pagevec *pvec) in drm_gem_check_release_pagevec() argument
511 check_move_unevictable_pages(pvec); in drm_gem_check_release_pagevec()
512 __pagevec_release(pvec); in drm_gem_check_release_pagevec()
545 struct pagevec pvec; in drm_gem_get_pages() local
588 pagevec_init(&pvec); in drm_gem_get_pages()
590 if (!pagevec_add(&pvec, pages[i])) in drm_gem_get_pages()
591 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages()
593 if (pagevec_count(&pvec)) in drm_gem_get_pages()
594 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages()
613 struct pagevec pvec; in drm_gem_put_pages() local
[all …]
/kernel/linux/linux-5.10/fs/hmdfs/
Dclient_writeback.c97 struct pagevec pvec; in hmdfs_idirty_pages() local
102 pagevec_init(&pvec); in hmdfs_idirty_pages()
104 pagevec_init(&pvec, 0); in hmdfs_idirty_pages()
106 while (pagevec_lookup_tag(&pvec, inode->i_mapping, &index, tag)) { in hmdfs_idirty_pages()
107 nr_dirty_pages += pagevec_count(&pvec); in hmdfs_idirty_pages()
108 pagevec_release(&pvec); in hmdfs_idirty_pages()
/kernel/linux/linux-5.10/fs/gfs2/
Daops.c234 struct pagevec *pvec, in gfs2_write_jdata_pagevec() argument
249 struct page *page = pvec->pages[i]; in gfs2_write_jdata_pagevec()
332 struct pagevec pvec; in gfs2_write_cache_jdata() local
342 pagevec_init(&pvec); in gfs2_write_cache_jdata()
368 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in gfs2_write_cache_jdata()
373 ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index); in gfs2_write_cache_jdata()
378 pagevec_release(&pvec); in gfs2_write_cache_jdata()
/kernel/linux/linux-5.10/drivers/gpu/drm/etnaviv/
Detnaviv_gem.c658 struct page **pvec = NULL; in etnaviv_gem_userptr_get_pages() local
667 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_userptr_get_pages()
668 if (!pvec) in etnaviv_gem_userptr_get_pages()
674 struct page **pages = pvec + pinned; in etnaviv_gem_userptr_get_pages()
679 unpin_user_pages(pvec, pinned); in etnaviv_gem_userptr_get_pages()
680 kvfree(pvec); in etnaviv_gem_userptr_get_pages()
688 etnaviv_obj->pages = pvec; in etnaviv_gem_userptr_get_pages()
/kernel/linux/linux-5.10/fs/ceph/
Daddr.c823 struct pagevec pvec; in ceph_writepages_start() local
847 pagevec_init(&pvec); in ceph_writepages_start()
905 pvec_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, in ceph_writepages_start()
911 page = pvec.pages[i]; in ceph_writepages_start()
1031 pvec.pages[i] = NULL; in ceph_writepages_start()
1043 if (!pvec.pages[j]) in ceph_writepages_start()
1046 pvec.pages[n] = pvec.pages[j]; in ceph_writepages_start()
1049 pvec.nr = n; in ceph_writepages_start()
1054 pagevec_release(&pvec); in ceph_writepages_start()
1179 dout("pagevec_release on %d pages (%p)\n", (int)pvec.nr, in ceph_writepages_start()
[all …]
/kernel/linux/linux-5.10/fs/hugetlbfs/
Dinode.c111 static void huge_pagevec_release(struct pagevec *pvec) in huge_pagevec_release() argument
115 for (i = 0; i < pagevec_count(pvec); ++i) in huge_pagevec_release()
116 put_page(pvec->pages[i]); in huge_pagevec_release()
118 pagevec_reinit(pvec); in huge_pagevec_release()
473 struct pagevec pvec; in remove_inode_hugepages() local
480 pagevec_init(&pvec); in remove_inode_hugepages()
486 if (!pagevec_lookup_range(&pvec, mapping, &next, end - 1)) in remove_inode_hugepages()
489 for (i = 0; i < pagevec_count(&pvec); ++i) { in remove_inode_hugepages()
490 struct page *page = pvec.pages[i]; in remove_inode_hugepages()
549 huge_pagevec_release(&pvec); in remove_inode_hugepages()
/kernel/linux/linux-5.10/fs/f2fs/
Dnode.c1460 struct pagevec pvec; in last_fsync_dnode() local
1464 pagevec_init(&pvec); in last_fsync_dnode()
1467 while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, in last_fsync_dnode()
1472 struct page *page = pvec.pages[i]; in last_fsync_dnode()
1476 pagevec_release(&pvec); in last_fsync_dnode()
1507 pagevec_release(&pvec); in last_fsync_dnode()
1676 struct pagevec pvec; in f2fs_fsync_node_pages() local
1690 pagevec_init(&pvec); in f2fs_fsync_node_pages()
1693 while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, in f2fs_fsync_node_pages()
1698 struct page *page = pvec.pages[i]; in f2fs_fsync_node_pages()
[all …]
Dcheckpoint.c387 struct pagevec pvec; in f2fs_sync_meta_pages() local
395 pagevec_init(&pvec); in f2fs_sync_meta_pages()
399 while ((nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, in f2fs_sync_meta_pages()
404 struct page *page = pvec.pages[i]; in f2fs_sync_meta_pages()
409 pagevec_release(&pvec); in f2fs_sync_meta_pages()
439 pagevec_release(&pvec); in f2fs_sync_meta_pages()
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/
Dmmu.c1724 static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, in mmu_pages_add() argument
1730 for (i=0; i < pvec->nr; i++) in mmu_pages_add()
1731 if (pvec->page[i].sp == sp) in mmu_pages_add()
1734 pvec->page[pvec->nr].sp = sp; in mmu_pages_add()
1735 pvec->page[pvec->nr].idx = idx; in mmu_pages_add()
1736 pvec->nr++; in mmu_pages_add()
1737 return (pvec->nr == KVM_PAGE_ARRAY_NR); in mmu_pages_add()
1748 struct kvm_mmu_pages *pvec) in __mmu_unsync_walk() argument
1764 if (mmu_pages_add(pvec, child, i)) in __mmu_unsync_walk()
1767 ret = __mmu_unsync_walk(child, pvec); in __mmu_unsync_walk()
[all …]
/kernel/linux/linux-5.10/fs/fscache/
Dpage.c1219 struct pagevec pvec; in __fscache_uncache_all_inode_pages() local
1230 pagevec_init(&pvec); in __fscache_uncache_all_inode_pages()
1233 if (!pagevec_lookup(&pvec, mapping, &next)) in __fscache_uncache_all_inode_pages()
1235 for (i = 0; i < pagevec_count(&pvec); i++) { in __fscache_uncache_all_inode_pages()
1236 struct page *page = pvec.pages[i]; in __fscache_uncache_all_inode_pages()
1242 pagevec_release(&pvec); in __fscache_uncache_all_inode_pages()
/kernel/linux/linux-5.10/drivers/mtd/lpddr/
Dlpddr_cmds.c404 unsigned long adr, const struct kvec **pvec, in do_write_buffer() argument
436 vec = *pvec; in do_write_buffer()
466 *pvec = vec; in do_write_buffer()
/kernel/linux/linux-5.10/fs/ext4/
Dinode.c1552 struct pagevec pvec; in mpage_release_unused_pages() local
1577 pagevec_init(&pvec); in mpage_release_unused_pages()
1579 nr_pages = pagevec_lookup_range(&pvec, mapping, &index, end); in mpage_release_unused_pages()
1583 struct page *page = pvec.pages[i]; in mpage_release_unused_pages()
1595 pagevec_release(&pvec); in mpage_release_unused_pages()
2347 struct pagevec pvec; in mpage_map_and_submit_buffers() local
2362 pagevec_init(&pvec); in mpage_map_and_submit_buffers()
2364 nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, in mpage_map_and_submit_buffers()
2369 struct page *page = pvec.pages[i]; in mpage_map_and_submit_buffers()
2385 pagevec_release(&pvec); in mpage_map_and_submit_buffers()
[all …]
/kernel/linux/linux-5.10/fs/
Dbuffer.c1610 struct pagevec pvec; in clean_bdev_aliases() local
1618 pagevec_init(&pvec); in clean_bdev_aliases()
1619 while (pagevec_lookup_range(&pvec, bd_mapping, &index, end)) { in clean_bdev_aliases()
1620 count = pagevec_count(&pvec); in clean_bdev_aliases()
1622 struct page *page = pvec.pages[i]; in clean_bdev_aliases()
1651 pagevec_release(&pvec); in clean_bdev_aliases()

12