/fs/isofs/ |
D | compress.c | 42 struct page **pages, unsigned poffset, in zisofs_uncompress_block() argument 68 if (!pages[i]) in zisofs_uncompress_block() 70 memset(page_address(pages[i]), 0, PAGE_SIZE); in zisofs_uncompress_block() 71 flush_dcache_page(pages[i]); in zisofs_uncompress_block() 72 SetPageUptodate(pages[i]); in zisofs_uncompress_block() 122 if (pages[curpage]) { in zisofs_uncompress_block() 123 stream.next_out = page_address(pages[curpage]) in zisofs_uncompress_block() 175 if (pages[curpage]) { in zisofs_uncompress_block() 176 flush_dcache_page(pages[curpage]); in zisofs_uncompress_block() 177 SetPageUptodate(pages[curpage]); in zisofs_uncompress_block() [all …]
|
/fs/erofs/ |
D | pcpubuf.c | 15 struct page **pages; member 64 struct page **pages, **oldpages; in erofs_pcpubuf_growsize() local 67 pages = kmalloc_array(nrpages, sizeof(*pages), GFP_KERNEL); in erofs_pcpubuf_growsize() 68 if (!pages) { in erofs_pcpubuf_growsize() 74 pages[i] = erofs_allocpage(&pagepool, GFP_KERNEL); in erofs_pcpubuf_growsize() 75 if (!pages[i]) { in erofs_pcpubuf_growsize() 77 oldpages = pages; in erofs_pcpubuf_growsize() 81 ptr = vmap(pages, nrpages, VM_MAP, PAGE_KERNEL); in erofs_pcpubuf_growsize() 84 oldpages = pages; in erofs_pcpubuf_growsize() 90 oldpages = pcb->pages; in erofs_pcpubuf_growsize() [all …]
|
D | zpvec.h | 31 erofs_vtptr_t *pages; member 43 kunmap_atomic(ctor->pages); in z_erofs_pagevec_ctor_exit() 59 const erofs_vtptr_t t = ctor->pages[index]; in z_erofs_pagevec_ctor_next_page() 79 ctor->pages = atomic ? in z_erofs_pagevec_ctor_pagedown() 88 erofs_vtptr_t *pages, in z_erofs_pagevec_ctor_init() argument 93 ctor->pages = pages; in z_erofs_pagevec_ctor_init() 133 ctor->pages[ctor->index++] = tagptr_fold(erofs_vtptr_t, page, type); in z_erofs_pagevec_enqueue() 148 t = ctor->pages[ctor->index]; in z_erofs_pagevec_dequeue() 156 ctor->pages[ctor->index++] = tagptr_fold(erofs_vtptr_t, NULL, 0); in z_erofs_pagevec_dequeue()
|
D | zdata.c | 363 struct page **pages; in preload_compressed_pages() local 369 pages = pcl->compressed_pages; in preload_compressed_pages() 371 for (; index < pcl->obj.index + pcl->pclusterpages; ++index, ++pages) { in preload_compressed_pages() 377 if (READ_ONCE(*pages)) in preload_compressed_pages() 401 if (!cmpxchg_relaxed(pages, NULL, tagptr_cast_ptr(t))) in preload_compressed_pages() 981 struct page **pages, **compressed_pages, *page; in z_erofs_decompress_pcluster() local 996 pages = pages_onstack; in z_erofs_decompress_pcluster() 999 pages = z_pagemap_global; in z_erofs_decompress_pcluster() 1006 pages = kvmalloc_array(nr_pages, sizeof(struct page *), in z_erofs_decompress_pcluster() 1010 if (!pages) { in z_erofs_decompress_pcluster() [all …]
|
/fs/ramfs/ |
D | file-nommu.c | 65 struct page *pages; in ramfs_nommu_expand_for_mapping() local 84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping() 85 if (!pages) in ramfs_nommu_expand_for_mapping() 92 split_page(pages, order); in ramfs_nommu_expand_for_mapping() 96 __free_page(pages + loop); in ramfs_nommu_expand_for_mapping() 100 data = page_address(pages); in ramfs_nommu_expand_for_mapping() 105 struct page *page = pages + loop; in ramfs_nommu_expand_for_mapping() 124 __free_page(pages + loop++); in ramfs_nommu_expand_for_mapping() 208 struct page **pages = NULL, **ptr, *page; in ramfs_nommu_get_unmapped_area() local 224 pages = kcalloc(lpages, sizeof(struct page *), GFP_KERNEL); in ramfs_nommu_get_unmapped_area() [all …]
|
/fs/squashfs/ |
D | page_actor.c | 32 if (actor->next_page == actor->pages) in cache_next_page() 44 int pages, int length) in squashfs_page_actor_init() argument 51 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init() 53 actor->pages = pages; in squashfs_page_actor_init() 70 if ((actor->next_page == actor->pages) || in handle_next_page() 106 struct page **page, int pages, int length) in squashfs_page_actor_init_special() argument 123 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init_special() 125 actor->pages = pages; in squashfs_page_actor_init_special()
|
D | file_direct.c | 33 int i, n, pages, bytes, res = -ENOMEM; in squashfs_readpage_block() local 41 pages = end_index - start_index + 1; in squashfs_readpage_block() 43 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block() 64 pages = i; in squashfs_readpage_block() 70 actor = squashfs_page_actor_init_special(msblk, page, pages, expected); in squashfs_readpage_block() 89 if (page[pages - 1]->index == end_index && bytes) { in squashfs_readpage_block() 90 pageaddr = kmap_local_page(page[pages - 1]); in squashfs_readpage_block() 96 for (i = 0; i < pages; i++) { in squashfs_readpage_block() 112 for (i = 0; i < pages; i++) { in squashfs_readpage_block()
|
D | file.c | 499 unsigned int pages, unsigned int expected) in squashfs_readahead_fragment() argument 514 for (n = 0; n < pages; n++) { in squashfs_readahead_fragment() 544 struct page **pages; in squashfs_readahead() local 550 pages = kmalloc_array(max_pages, sizeof(void *), GFP_KERNEL); in squashfs_readahead() 551 if (!pages) in squashfs_readahead() 567 nr_pages = __readahead_batch(ractl, pages, max_pages); in squashfs_readahead() 574 index = pages[0]->index >> shift; in squashfs_readahead() 576 if ((pages[nr_pages - 1]->index >> shift) != index) in squashfs_readahead() 581 res = squashfs_readahead_fragment(pages, nr_pages, in squashfs_readahead() 592 actor = squashfs_page_actor_init_special(msblk, pages, nr_pages, in squashfs_readahead() [all …]
|
D | cache.c | 206 for (j = 0; j < cache->pages; j++) in squashfs_cache_delete() 245 cache->pages = block_size >> PAGE_SHIFT; in squashfs_cache_init() 246 cache->pages = cache->pages ? cache->pages : 1; in squashfs_cache_init() 258 entry->data = kcalloc(cache->pages, sizeof(void *), GFP_KERNEL); in squashfs_cache_init() 264 for (j = 0; j < cache->pages; j++) { in squashfs_cache_init() 273 cache->pages, 0); in squashfs_cache_init() 408 int pages = (length + PAGE_SIZE - 1) >> PAGE_SHIFT; in squashfs_read_table() local 417 data = kcalloc(pages, sizeof(void *), GFP_KERNEL); in squashfs_read_table() 423 actor = squashfs_page_actor_init(data, pages, length); in squashfs_read_table() 429 for (i = 0; i < pages; i++, buffer += PAGE_SIZE) in squashfs_read_table()
|
/fs/proc/ |
D | meminfo.c | 38 unsigned long pages[NR_LRU_LISTS]; in meminfo_proc_show() local 52 pages[lru] = global_node_page_state(NR_LRU_BASE + lru); in meminfo_proc_show() 64 show_val_kb(m, "Active: ", pages[LRU_ACTIVE_ANON] + in meminfo_proc_show() 65 pages[LRU_ACTIVE_FILE]); in meminfo_proc_show() 66 show_val_kb(m, "Inactive: ", pages[LRU_INACTIVE_ANON] + in meminfo_proc_show() 67 pages[LRU_INACTIVE_FILE]); in meminfo_proc_show() 68 show_val_kb(m, "Active(anon): ", pages[LRU_ACTIVE_ANON]); in meminfo_proc_show() 69 show_val_kb(m, "Inactive(anon): ", pages[LRU_INACTIVE_ANON]); in meminfo_proc_show() 70 show_val_kb(m, "Active(file): ", pages[LRU_ACTIVE_FILE]); in meminfo_proc_show() 71 show_val_kb(m, "Inactive(file): ", pages[LRU_INACTIVE_FILE]); in meminfo_proc_show() [all …]
|
/fs/ntfs/ |
D | compress.c | 502 struct page **pages; in ntfs_read_compressed_block() local 515 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_NOFS); in ntfs_read_compressed_block() 522 if (unlikely(!pages || !bhs || !completed_pages)) { in ntfs_read_compressed_block() 524 kfree(pages); in ntfs_read_compressed_block() 537 pages[xpage] = page; in ntfs_read_compressed_block() 551 kfree(pages); in ntfs_read_compressed_block() 563 pages[i] = grab_cache_page_nowait(mapping, offset); in ntfs_read_compressed_block() 564 page = pages[i]; in ntfs_read_compressed_block() 579 pages[i] = NULL; in ntfs_read_compressed_block() 738 page = pages[cur_page]; in ntfs_read_compressed_block() [all …]
|
D | file.c | 492 pgoff_t index, const unsigned nr_pages, struct page **pages, in __ntfs_grab_cache_pages() argument 500 pages[nr] = find_get_page_flags(mapping, index, FGP_LOCK | in __ntfs_grab_cache_pages() 502 if (!pages[nr]) { in __ntfs_grab_cache_pages() 518 pages[nr] = *cached_page; in __ntfs_grab_cache_pages() 528 unlock_page(pages[--nr]); in __ntfs_grab_cache_pages() 529 put_page(pages[nr]); in __ntfs_grab_cache_pages() 567 static int ntfs_prepare_pages_for_non_resident_write(struct page **pages, in ntfs_prepare_pages_for_non_resident_write() argument 597 BUG_ON(!pages); in ntfs_prepare_pages_for_non_resident_write() 598 BUG_ON(!*pages); in ntfs_prepare_pages_for_non_resident_write() 599 vi = pages[0]->mapping->host; in ntfs_prepare_pages_for_non_resident_write() [all …]
|
/fs/cramfs/ |
D | inode.c | 186 struct page *pages[BLKS_PER_BUF]; in cramfs_blkdev_read() local 224 pages[i] = page; in cramfs_blkdev_read() 228 struct page *page = pages[i]; in cramfs_blkdev_read() 235 pages[i] = NULL; in cramfs_blkdev_read() 247 struct page *page = pages[i]; in cramfs_blkdev_read() 297 static u32 cramfs_get_block_range(struct inode *inode, u32 pgoff, u32 *pages) in cramfs_get_block_range() argument 318 pgoff+i, pgoff + *pages - 1, in cramfs_get_block_range() 324 } while (++i < *pages); in cramfs_get_block_range() 326 *pages = i; in cramfs_get_block_range() 359 unsigned int pages, max_pages, offset; in cramfs_physmem_mmap() local [all …]
|
/fs/crypto/ |
D | bio.c | 125 struct page *pages[16]; /* write up to 16 pages at a time */ in fscrypt_zeroout_range() local 139 BUILD_BUG_ON(ARRAY_SIZE(pages) > BIO_MAX_VECS); in fscrypt_zeroout_range() 140 nr_pages = min_t(u64, ARRAY_SIZE(pages), in fscrypt_zeroout_range() 151 pages[i] = fscrypt_alloc_bounce_page(i == 0 ? GFP_NOFS : in fscrypt_zeroout_range() 153 if (!pages[i]) in fscrypt_zeroout_range() 172 ZERO_PAGE(0), pages[i], in fscrypt_zeroout_range() 182 ret = bio_add_page(bio, pages[i++], offset, 0); in fscrypt_zeroout_range() 200 fscrypt_free_bounce_page(pages[i]); in fscrypt_zeroout_range()
|
/fs/ceph/ |
D | addr.c | 238 ceph_put_page_vector(osd_data->pages, num_pages, false); in finish_netfs_read() 251 struct page **pages; in ceph_netfs_issue_op() local 268 err = iov_iter_get_pages_alloc(&iter, &pages, len, &page_off); in ceph_netfs_issue_op() 278 osd_req_op_extent_osd_data_pages(req, 0, pages, len, 0, false, false); in ceph_netfs_issue_op() 664 page = osd_data->pages[j]; in writepages_finish() 687 release_pages(osd_data->pages, num_pages); in writepages_finish() 697 mempool_free(osd_data->pages, ceph_wb_pagevec_pool); in writepages_finish() 699 kfree(osd_data->pages); in writepages_finish() 788 struct page **pages = NULL, **data_pages; in ceph_writepages_start() local 803 page = pvec.pages[i]; in ceph_writepages_start() [all …]
|
/fs/nfs/ |
D | nfs3acl.c | 50 struct page *pages[NFSACL_MAXPAGES] = { }; in nfs3_get_acl() local 54 .pages = pages, in nfs3_get_acl() 103 for (count = 0; count < NFSACL_MAXPAGES && args.pages[count]; count++) in nfs3_get_acl() 104 __free_page(args.pages[count]); in nfs3_get_acl() 167 struct page *pages[NFSACL_MAXPAGES]; in __nfs3_proc_setacls() local 172 .pages = pages, in __nfs3_proc_setacls() 206 args.pages[args.npages] = alloc_page(GFP_KERNEL); in __nfs3_proc_setacls() 207 if (args.pages[args.npages] == NULL) in __nfs3_proc_setacls() 243 __free_page(args.pages[args.npages]); in __nfs3_proc_setacls()
|
D | pnfs_dev.c | 103 struct page **pages = NULL; in nfs4_get_device_info() local 124 pages = kcalloc(max_pages, sizeof(struct page *), gfp_flags); in nfs4_get_device_info() 125 if (!pages) in nfs4_get_device_info() 129 pages[i] = alloc_page(gfp_flags); in nfs4_get_device_info() 130 if (!pages[i]) in nfs4_get_device_info() 136 pdev->pages = pages; in nfs4_get_device_info() 158 __free_page(pages[i]); in nfs4_get_device_info() 159 kfree(pages); in nfs4_get_device_info()
|
D | nfs42proc.c | 1176 struct page *pages[NFS4XATTR_MAXPAGES]; in _nfs42_proc_setxattr() local 1179 .xattr_pages = pages, in _nfs42_proc_setxattr() 1207 put_page(pages[np - 1]); in _nfs42_proc_setxattr() 1216 void *buf, size_t buflen, struct page **pages, in _nfs42_proc_getxattr() argument 1233 arg.xattr_pages = pages; in _nfs42_proc_getxattr() 1250 nfs4_xattr_cache_add(inode, name, NULL, pages, res.xattr_len); in _nfs42_proc_getxattr() 1255 _copy_from_pages(buf, pages, 0, res.xattr_len); in _nfs42_proc_getxattr() 1265 struct page **pages; in _nfs42_proc_listxattrs() local 1294 pages = kcalloc(np, sizeof(struct page *), GFP_KERNEL); in _nfs42_proc_listxattrs() 1295 if (!pages) in _nfs42_proc_listxattrs() [all …]
|
/fs/orangefs/ |
D | inode.c | 86 struct page **pages; member 93 struct inode *inode = ow->pages[0]->mapping->host; in orangefs_writepages_work() 104 set_page_writeback(ow->pages[i]); in orangefs_writepages_work() 105 ow->bv[i].bv_page = ow->pages[i]; in orangefs_writepages_work() 106 ow->bv[i].bv_len = min(page_offset(ow->pages[i]) + PAGE_SIZE, in orangefs_writepages_work() 108 max(ow->off, page_offset(ow->pages[i])); in orangefs_writepages_work() 111 page_offset(ow->pages[i]); in orangefs_writepages_work() 128 SetPageError(ow->pages[i]); in orangefs_writepages_work() 129 mapping_set_error(ow->pages[i]->mapping, ret); in orangefs_writepages_work() 130 if (PagePrivate(ow->pages[i])) { in orangefs_writepages_work() [all …]
|
/fs/btrfs/ |
D | struct-funcs.c | 74 token->kaddr = page_address(token->eb->pages[idx]); \ 80 token->kaddr = page_address(token->eb->pages[idx + 1]); \ 91 char *kaddr = page_address(eb->pages[idx]); \ 101 kaddr = page_address(eb->pages[idx + 1]); \ 125 token->kaddr = page_address(token->eb->pages[idx]); \ 133 token->kaddr = page_address(token->eb->pages[idx + 1]); \ 143 char *kaddr = page_address(eb->pages[idx]); \ 156 kaddr = page_address(eb->pages[idx + 1]); \
|
/fs/cifs/ |
D | fscache.h | 107 struct list_head *pages, in cifs_readpages_from_fscache() argument 111 return __cifs_readpages_from_fscache(inode, mapping, pages, in cifs_readpages_from_fscache() 124 struct list_head *pages) in cifs_fscache_readpages_cancel() argument 127 return __cifs_fscache_readpages_cancel(inode, pages); in cifs_fscache_readpages_cancel() 167 struct list_head *pages, in cifs_readpages_from_fscache() argument 177 struct list_head *pages) in cifs_fscache_readpages_cancel() argument
|
/fs/ntfs3/ |
D | file.c | 841 struct page **pages, u32 pages_per_frame, in ntfs_get_frame_pages() argument 855 page = pages[npages]; in ntfs_get_frame_pages() 866 pages[npages] = page; in ntfs_get_frame_pages() 887 struct page *page, **pages = NULL; in ntfs_compress_write() local 907 pages = kmalloc_array(pages_per_frame, sizeof(struct page *), GFP_NOFS); in ntfs_compress_write() 908 if (!pages) in ntfs_compress_write() 941 pages, pages_per_frame, in ntfs_compress_write() 947 err = ni_read_frame(ni, frame_vbo, pages, in ntfs_compress_write() 951 page = pages[ip]; in ntfs_compress_write() 962 page = pages[ip]; in ntfs_compress_write() [all …]
|
/fs/9p/ |
D | cache.h | 36 struct list_head *pages, 61 struct list_head *pages, in v9fs_readpages_from_fscache() argument 64 return __v9fs_readpages_from_fscache(inode, mapping, pages, in v9fs_readpages_from_fscache() 117 struct list_head *pages, in v9fs_readpages_from_fscache() argument
|
/fs/fuse/ |
D | ioctl.c | 204 ap.pages = fuse_pages_alloc(fm->fc->max_pages, GFP_KERNEL, &ap.descs); in fuse_do_ioctl() 206 if (!ap.pages || !iov_page) in fuse_do_ioctl() 248 ap.pages[ap.num_pages] = alloc_page(GFP_KERNEL | __GFP_HIGHMEM); in fuse_do_ioctl() 249 if (!ap.pages[ap.num_pages]) in fuse_do_ioctl() 269 c = copy_page_from_iter(ap.pages[i], 0, PAGE_SIZE, &ii); in fuse_do_ioctl() 307 vaddr = kmap_atomic(ap.pages[0]); in fuse_do_ioctl() 336 c = copy_page_to_iter(ap.pages[i], 0, PAGE_SIZE, &ii); in fuse_do_ioctl() 344 __free_page(ap.pages[--ap.num_pages]); in fuse_do_ioctl() 345 kfree(ap.pages); in fuse_do_ioctl()
|
/fs/btrfs/tests/ |
D | extent-io-tests.c | 23 struct page *pages[16]; in process_page_range() local 34 ARRAY_SIZE(pages)), pages); in process_page_range() 37 !PageLocked(pages[i])) in process_page_range() 39 if (flags & PROCESS_UNLOCK && PageLocked(pages[i])) in process_page_range() 40 unlock_page(pages[i]); in process_page_range() 41 put_page(pages[i]); in process_page_range() 43 put_page(pages[i]); in process_page_range()
|