/fs/nilfs2/ |
D | page.c | 243 struct folio_batch fbatch; in nilfs_copy_dirty_pages() local 248 folio_batch_init(&fbatch); in nilfs_copy_dirty_pages() 251 PAGECACHE_TAG_DIRTY, &fbatch)) in nilfs_copy_dirty_pages() 254 for (i = 0; i < folio_batch_count(&fbatch); i++) { in nilfs_copy_dirty_pages() 255 struct folio *folio = fbatch.folios[i], *dfolio; in nilfs_copy_dirty_pages() 279 folio_batch_release(&fbatch); in nilfs_copy_dirty_pages() 298 struct folio_batch fbatch; in nilfs_copy_back_pages() local 302 folio_batch_init(&fbatch); in nilfs_copy_back_pages() 304 n = filemap_get_folios(smap, &start, ~0UL, &fbatch); in nilfs_copy_back_pages() 308 for (i = 0; i < folio_batch_count(&fbatch); i++) { in nilfs_copy_back_pages() [all …]
|
D | segment.c | 702 struct folio_batch fbatch; in nilfs_lookup_dirty_data_buffers() local 716 folio_batch_init(&fbatch); in nilfs_lookup_dirty_data_buffers() 720 PAGECACHE_TAG_DIRTY, &fbatch)) in nilfs_lookup_dirty_data_buffers() 723 for (i = 0; i < folio_batch_count(&fbatch); i++) { in nilfs_lookup_dirty_data_buffers() 725 struct folio *folio = fbatch.folios[i]; in nilfs_lookup_dirty_data_buffers() 748 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers() 754 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers() 764 struct folio_batch fbatch; in nilfs_lookup_dirty_node_buffers() local 771 folio_batch_init(&fbatch); in nilfs_lookup_dirty_node_buffers() 774 (pgoff_t)-1, PAGECACHE_TAG_DIRTY, &fbatch)) { in nilfs_lookup_dirty_node_buffers() [all …]
|
D | btree.c | 2158 struct folio_batch fbatch; in nilfs_btree_lookup_dirty_buffers() local 2168 folio_batch_init(&fbatch); in nilfs_btree_lookup_dirty_buffers() 2171 PAGECACHE_TAG_DIRTY, &fbatch)) { in nilfs_btree_lookup_dirty_buffers() 2172 for (i = 0; i < folio_batch_count(&fbatch); i++) { in nilfs_btree_lookup_dirty_buffers() 2173 bh = head = folio_buffers(fbatch.folios[i]); in nilfs_btree_lookup_dirty_buffers() 2180 folio_batch_release(&fbatch); in nilfs_btree_lookup_dirty_buffers()
|
/fs/ramfs/ |
D | file-nommu.c | 208 struct folio_batch fbatch; in ramfs_nommu_get_unmapped_area() local 224 folio_batch_init(&fbatch); in ramfs_nommu_get_unmapped_area() 228 ULONG_MAX, &fbatch); in ramfs_nommu_get_unmapped_area() 235 ret = (unsigned long) folio_address(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area() 236 pfn = folio_pfn(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area() 240 if (pfn + nr_pages != folio_pfn(fbatch.folios[loop])) { in ramfs_nommu_get_unmapped_area() 244 nr_pages += folio_nr_pages(fbatch.folios[loop]); in ramfs_nommu_get_unmapped_area() 250 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area() 256 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area()
|
/fs/afs/ |
D | write.c | 470 struct folio_batch fbatch; in afs_extend_writeback() local 481 folio_batch_init(&fbatch); in afs_extend_writeback() 540 if (!folio_batch_add(&fbatch, folio)) in afs_extend_writeback() 553 if (!folio_batch_count(&fbatch)) in afs_extend_writeback() 556 for (i = 0; i < folio_batch_count(&fbatch); i++) { in afs_extend_writeback() 557 folio = fbatch.folios[i]; in afs_extend_writeback() 570 folio_batch_release(&fbatch); in afs_extend_writeback() 709 struct folio_batch fbatch; in afs_writepages_region() local 715 folio_batch_init(&fbatch); in afs_writepages_region() 721 PAGECACHE_TAG_DIRTY, &fbatch); in afs_writepages_region() [all …]
|
/fs/gfs2/ |
D | aops.c | 209 struct folio_batch *fbatch, in gfs2_write_jdata_batch() argument 218 int nr_folios = folio_batch_count(fbatch); in gfs2_write_jdata_batch() 221 nr_pages += folio_nr_pages(fbatch->folios[i]); in gfs2_write_jdata_batch() 229 struct folio *folio = fbatch->folios[i]; in gfs2_write_jdata_batch() 312 struct folio_batch fbatch; in gfs2_write_cache_jdata() local 322 folio_batch_init(&fbatch); in gfs2_write_cache_jdata() 349 tag, &fbatch); in gfs2_write_cache_jdata() 353 ret = gfs2_write_jdata_batch(mapping, wbc, &fbatch, in gfs2_write_cache_jdata() 359 folio_batch_release(&fbatch); in gfs2_write_cache_jdata()
|
/fs/ceph/ |
D | addr.c | 935 struct folio_batch fbatch; in ceph_writepages_start() local 965 folio_batch_init(&fbatch); in ceph_writepages_start() 1032 end, tag, &fbatch); in ceph_writepages_start() 1037 page = &fbatch.folios[i]->page; in ceph_writepages_start() 1177 fbatch.folios[i] = NULL; in ceph_writepages_start() 1188 if (!fbatch.folios[j]) in ceph_writepages_start() 1191 fbatch.folios[n] = fbatch.folios[j]; in ceph_writepages_start() 1194 fbatch.nr = n; in ceph_writepages_start() 1199 folio_batch_release(&fbatch); in ceph_writepages_start() 1350 dout("folio_batch release on %d folios (%p)\n", (int)fbatch.nr, in ceph_writepages_start() [all …]
|
/fs/f2fs/ |
D | node.c | 1552 struct folio_batch fbatch; in last_fsync_dnode() local 1556 folio_batch_init(&fbatch); in last_fsync_dnode() 1561 &fbatch))) { in last_fsync_dnode() 1565 struct page *page = &fbatch.folios[i]->page; in last_fsync_dnode() 1569 folio_batch_release(&fbatch); in last_fsync_dnode() 1600 folio_batch_release(&fbatch); in last_fsync_dnode() 1768 struct folio_batch fbatch; in f2fs_fsync_node_pages() local 1782 folio_batch_init(&fbatch); in f2fs_fsync_node_pages() 1787 &fbatch))) { in f2fs_fsync_node_pages() 1791 struct page *page = &fbatch.folios[i]->page; in f2fs_fsync_node_pages() [all …]
|
D | checkpoint.c | 426 struct folio_batch fbatch; in f2fs_sync_meta_pages() local 434 folio_batch_init(&fbatch); in f2fs_sync_meta_pages() 440 PAGECACHE_TAG_DIRTY, &fbatch))) { in f2fs_sync_meta_pages() 444 struct folio *folio = fbatch.folios[i]; in f2fs_sync_meta_pages() 448 folio_nr_pages(fbatch.folios[i-1])) { in f2fs_sync_meta_pages() 449 folio_batch_release(&fbatch); in f2fs_sync_meta_pages() 481 folio_batch_release(&fbatch); in f2fs_sync_meta_pages()
|
D | compress.c | 1956 struct folio_batch fbatch; in f2fs_invalidate_compress_pages() local 1963 folio_batch_init(&fbatch); in f2fs_invalidate_compress_pages() 1968 nr = filemap_get_folios(mapping, &index, end - 1, &fbatch); in f2fs_invalidate_compress_pages() 1973 struct folio *folio = fbatch.folios[i]; in f2fs_invalidate_compress_pages() 1989 folio_batch_release(&fbatch); in f2fs_invalidate_compress_pages()
|
D | data.c | 2993 struct folio_batch fbatch; in f2fs_write_cache_pages() local 3035 folio_batch_init(&fbatch); in f2fs_write_cache_pages() 3065 tag, &fbatch); in f2fs_write_cache_pages() 3073 struct folio *folio = fbatch.folios[i]; in f2fs_write_cache_pages() 3082 folio_batch_release(&fbatch); in f2fs_write_cache_pages() 3088 folio_batch_release(&fbatch); in f2fs_write_cache_pages()
|
/fs/btrfs/tests/ |
D | extent-io-tests.c | 25 struct folio_batch fbatch; in process_page_range() local 32 folio_batch_init(&fbatch); in process_page_range() 36 end_index, &fbatch); in process_page_range() 38 struct folio *folio = fbatch.folios[i]; in process_page_range() 48 folio_batch_release(&fbatch); in process_page_range()
|
/fs/btrfs/ |
D | extent_io.c | 215 struct folio_batch fbatch; in __process_pages_contig() local 218 folio_batch_init(&fbatch); in __process_pages_contig() 223 end_index, &fbatch); in __process_pages_contig() 225 struct folio *folio = fbatch.folios[i]; in __process_pages_contig() 230 folio_batch_release(&fbatch); in __process_pages_contig() 261 struct folio_batch fbatch; in lock_delalloc_pages() local 266 folio_batch_init(&fbatch); in lock_delalloc_pages() 271 end_index, &fbatch); in lock_delalloc_pages() 276 struct page *page = &fbatch.folios[i]->page; in lock_delalloc_pages() 294 folio_batch_release(&fbatch); in lock_delalloc_pages() [all …]
|
D | compression.c | 195 struct folio_batch fbatch; in end_compressed_writeback() local 203 folio_batch_init(&fbatch); in end_compressed_writeback() 206 &fbatch); in end_compressed_writeback() 212 struct folio *folio = fbatch.folios[i]; in end_compressed_writeback() 217 folio_batch_release(&fbatch); in end_compressed_writeback()
|
/fs/hugetlbfs/ |
D | inode.c | 677 struct folio_batch fbatch; in remove_inode_hugepages() local 682 folio_batch_init(&fbatch); in remove_inode_hugepages() 684 while (filemap_get_folios(mapping, &next, end - 1, &fbatch)) { in remove_inode_hugepages() 685 for (i = 0; i < folio_batch_count(&fbatch); ++i) { in remove_inode_hugepages() 686 struct folio *folio = fbatch.folios[i]; in remove_inode_hugepages() 702 folio_batch_release(&fbatch); in remove_inode_hugepages()
|
/fs/ext4/ |
D | inode.c | 1578 struct folio_batch fbatch; in mpage_release_unused_pages() local 1603 folio_batch_init(&fbatch); in mpage_release_unused_pages() 1605 nr = filemap_get_folios(mapping, &index, end, &fbatch); in mpage_release_unused_pages() 1609 struct folio *folio = fbatch.folios[i]; in mpage_release_unused_pages() 1626 folio_batch_release(&fbatch); in mpage_release_unused_pages() 2110 struct folio_batch fbatch; in mpage_map_and_submit_buffers() local 2125 folio_batch_init(&fbatch); in mpage_map_and_submit_buffers() 2127 nr = filemap_get_folios(inode->i_mapping, &start, end, &fbatch); in mpage_map_and_submit_buffers() 2131 struct folio *folio = fbatch.folios[i]; in mpage_map_and_submit_buffers() 2148 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers() [all …]
|
/fs/ |
D | buffer.c | 1704 struct folio_batch fbatch; in clean_bdev_aliases() local 1712 folio_batch_init(&fbatch); in clean_bdev_aliases() 1713 while (filemap_get_folios(bd_mapping, &index, end, &fbatch)) { in clean_bdev_aliases() 1714 count = folio_batch_count(&fbatch); in clean_bdev_aliases() 1716 struct folio *folio = fbatch.folios[i]; in clean_bdev_aliases() 1745 folio_batch_release(&fbatch); in clean_bdev_aliases()
|