Home
last modified time | relevance | path

Searched refs:pages (Results 1 – 25 of 138) sorted by relevance

123456

/fs/isofs/
Dcompress.c42 struct page **pages, unsigned poffset, in zisofs_uncompress_block() argument
68 if (!pages[i]) in zisofs_uncompress_block()
70 memset(page_address(pages[i]), 0, PAGE_SIZE); in zisofs_uncompress_block()
71 flush_dcache_page(pages[i]); in zisofs_uncompress_block()
72 SetPageUptodate(pages[i]); in zisofs_uncompress_block()
122 if (pages[curpage]) { in zisofs_uncompress_block()
123 stream.next_out = page_address(pages[curpage]) in zisofs_uncompress_block()
175 if (pages[curpage]) { in zisofs_uncompress_block()
176 flush_dcache_page(pages[curpage]); in zisofs_uncompress_block()
177 SetPageUptodate(pages[curpage]); in zisofs_uncompress_block()
[all …]
/fs/erofs/
Dpcpubuf.c15 struct page **pages; member
64 struct page **pages, **oldpages; in erofs_pcpubuf_growsize() local
67 pages = kmalloc_array(nrpages, sizeof(*pages), GFP_KERNEL); in erofs_pcpubuf_growsize()
68 if (!pages) { in erofs_pcpubuf_growsize()
74 pages[i] = erofs_allocpage(&pagepool, GFP_KERNEL); in erofs_pcpubuf_growsize()
75 if (!pages[i]) { in erofs_pcpubuf_growsize()
77 oldpages = pages; in erofs_pcpubuf_growsize()
81 ptr = vmap(pages, nrpages, VM_MAP, PAGE_KERNEL); in erofs_pcpubuf_growsize()
84 oldpages = pages; in erofs_pcpubuf_growsize()
90 oldpages = pcb->pages; in erofs_pcpubuf_growsize()
[all …]
Dzpvec.h31 erofs_vtptr_t *pages; member
43 kunmap_atomic(ctor->pages); in z_erofs_pagevec_ctor_exit()
59 const erofs_vtptr_t t = ctor->pages[index]; in z_erofs_pagevec_ctor_next_page()
79 ctor->pages = atomic ? in z_erofs_pagevec_ctor_pagedown()
88 erofs_vtptr_t *pages, in z_erofs_pagevec_ctor_init() argument
93 ctor->pages = pages; in z_erofs_pagevec_ctor_init()
133 ctor->pages[ctor->index++] = tagptr_fold(erofs_vtptr_t, page, type); in z_erofs_pagevec_enqueue()
148 t = ctor->pages[ctor->index]; in z_erofs_pagevec_dequeue()
156 ctor->pages[ctor->index++] = tagptr_fold(erofs_vtptr_t, NULL, 0); in z_erofs_pagevec_dequeue()
Dzdata.c363 struct page **pages; in preload_compressed_pages() local
369 pages = pcl->compressed_pages; in preload_compressed_pages()
371 for (; index < pcl->obj.index + pcl->pclusterpages; ++index, ++pages) { in preload_compressed_pages()
377 if (READ_ONCE(*pages)) in preload_compressed_pages()
401 if (!cmpxchg_relaxed(pages, NULL, tagptr_cast_ptr(t))) in preload_compressed_pages()
981 struct page **pages, **compressed_pages, *page; in z_erofs_decompress_pcluster() local
996 pages = pages_onstack; in z_erofs_decompress_pcluster()
999 pages = z_pagemap_global; in z_erofs_decompress_pcluster()
1006 pages = kvmalloc_array(nr_pages, sizeof(struct page *), in z_erofs_decompress_pcluster()
1010 if (!pages) { in z_erofs_decompress_pcluster()
[all …]
/fs/ramfs/
Dfile-nommu.c65 struct page *pages; in ramfs_nommu_expand_for_mapping() local
84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
85 if (!pages) in ramfs_nommu_expand_for_mapping()
92 split_page(pages, order); in ramfs_nommu_expand_for_mapping()
96 __free_page(pages + loop); in ramfs_nommu_expand_for_mapping()
100 data = page_address(pages); in ramfs_nommu_expand_for_mapping()
105 struct page *page = pages + loop; in ramfs_nommu_expand_for_mapping()
124 __free_page(pages + loop++); in ramfs_nommu_expand_for_mapping()
208 struct page **pages = NULL, **ptr, *page; in ramfs_nommu_get_unmapped_area() local
224 pages = kcalloc(lpages, sizeof(struct page *), GFP_KERNEL); in ramfs_nommu_get_unmapped_area()
[all …]
/fs/squashfs/
Dpage_actor.c32 if (actor->next_page == actor->pages) in cache_next_page()
44 int pages, int length) in squashfs_page_actor_init() argument
51 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init()
53 actor->pages = pages; in squashfs_page_actor_init()
70 if ((actor->next_page == actor->pages) || in handle_next_page()
106 struct page **page, int pages, int length) in squashfs_page_actor_init_special() argument
123 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init_special()
125 actor->pages = pages; in squashfs_page_actor_init_special()
Dfile_direct.c33 int i, n, pages, bytes, res = -ENOMEM; in squashfs_readpage_block() local
41 pages = end_index - start_index + 1; in squashfs_readpage_block()
43 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block()
64 pages = i; in squashfs_readpage_block()
70 actor = squashfs_page_actor_init_special(msblk, page, pages, expected); in squashfs_readpage_block()
89 if (page[pages - 1]->index == end_index && bytes) { in squashfs_readpage_block()
90 pageaddr = kmap_local_page(page[pages - 1]); in squashfs_readpage_block()
96 for (i = 0; i < pages; i++) { in squashfs_readpage_block()
112 for (i = 0; i < pages; i++) { in squashfs_readpage_block()
Dfile.c499 unsigned int pages, unsigned int expected) in squashfs_readahead_fragment() argument
514 for (n = 0; n < pages; n++) { in squashfs_readahead_fragment()
544 struct page **pages; in squashfs_readahead() local
550 pages = kmalloc_array(max_pages, sizeof(void *), GFP_KERNEL); in squashfs_readahead()
551 if (!pages) in squashfs_readahead()
567 nr_pages = __readahead_batch(ractl, pages, max_pages); in squashfs_readahead()
574 index = pages[0]->index >> shift; in squashfs_readahead()
576 if ((pages[nr_pages - 1]->index >> shift) != index) in squashfs_readahead()
581 res = squashfs_readahead_fragment(pages, nr_pages, in squashfs_readahead()
592 actor = squashfs_page_actor_init_special(msblk, pages, nr_pages, in squashfs_readahead()
[all …]
Dcache.c206 for (j = 0; j < cache->pages; j++) in squashfs_cache_delete()
245 cache->pages = block_size >> PAGE_SHIFT; in squashfs_cache_init()
246 cache->pages = cache->pages ? cache->pages : 1; in squashfs_cache_init()
258 entry->data = kcalloc(cache->pages, sizeof(void *), GFP_KERNEL); in squashfs_cache_init()
264 for (j = 0; j < cache->pages; j++) { in squashfs_cache_init()
273 cache->pages, 0); in squashfs_cache_init()
408 int pages = (length + PAGE_SIZE - 1) >> PAGE_SHIFT; in squashfs_read_table() local
417 data = kcalloc(pages, sizeof(void *), GFP_KERNEL); in squashfs_read_table()
423 actor = squashfs_page_actor_init(data, pages, length); in squashfs_read_table()
429 for (i = 0; i < pages; i++, buffer += PAGE_SIZE) in squashfs_read_table()
/fs/proc/
Dmeminfo.c38 unsigned long pages[NR_LRU_LISTS]; in meminfo_proc_show() local
52 pages[lru] = global_node_page_state(NR_LRU_BASE + lru); in meminfo_proc_show()
64 show_val_kb(m, "Active: ", pages[LRU_ACTIVE_ANON] + in meminfo_proc_show()
65 pages[LRU_ACTIVE_FILE]); in meminfo_proc_show()
66 show_val_kb(m, "Inactive: ", pages[LRU_INACTIVE_ANON] + in meminfo_proc_show()
67 pages[LRU_INACTIVE_FILE]); in meminfo_proc_show()
68 show_val_kb(m, "Active(anon): ", pages[LRU_ACTIVE_ANON]); in meminfo_proc_show()
69 show_val_kb(m, "Inactive(anon): ", pages[LRU_INACTIVE_ANON]); in meminfo_proc_show()
70 show_val_kb(m, "Active(file): ", pages[LRU_ACTIVE_FILE]); in meminfo_proc_show()
71 show_val_kb(m, "Inactive(file): ", pages[LRU_INACTIVE_FILE]); in meminfo_proc_show()
[all …]
/fs/ntfs/
Dcompress.c502 struct page **pages; in ntfs_read_compressed_block() local
515 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_NOFS); in ntfs_read_compressed_block()
522 if (unlikely(!pages || !bhs || !completed_pages)) { in ntfs_read_compressed_block()
524 kfree(pages); in ntfs_read_compressed_block()
537 pages[xpage] = page; in ntfs_read_compressed_block()
551 kfree(pages); in ntfs_read_compressed_block()
563 pages[i] = grab_cache_page_nowait(mapping, offset); in ntfs_read_compressed_block()
564 page = pages[i]; in ntfs_read_compressed_block()
579 pages[i] = NULL; in ntfs_read_compressed_block()
738 page = pages[cur_page]; in ntfs_read_compressed_block()
[all …]
Dfile.c492 pgoff_t index, const unsigned nr_pages, struct page **pages, in __ntfs_grab_cache_pages() argument
500 pages[nr] = find_get_page_flags(mapping, index, FGP_LOCK | in __ntfs_grab_cache_pages()
502 if (!pages[nr]) { in __ntfs_grab_cache_pages()
518 pages[nr] = *cached_page; in __ntfs_grab_cache_pages()
528 unlock_page(pages[--nr]); in __ntfs_grab_cache_pages()
529 put_page(pages[nr]); in __ntfs_grab_cache_pages()
567 static int ntfs_prepare_pages_for_non_resident_write(struct page **pages, in ntfs_prepare_pages_for_non_resident_write() argument
597 BUG_ON(!pages); in ntfs_prepare_pages_for_non_resident_write()
598 BUG_ON(!*pages); in ntfs_prepare_pages_for_non_resident_write()
599 vi = pages[0]->mapping->host; in ntfs_prepare_pages_for_non_resident_write()
[all …]
/fs/cramfs/
Dinode.c186 struct page *pages[BLKS_PER_BUF]; in cramfs_blkdev_read() local
224 pages[i] = page; in cramfs_blkdev_read()
228 struct page *page = pages[i]; in cramfs_blkdev_read()
235 pages[i] = NULL; in cramfs_blkdev_read()
247 struct page *page = pages[i]; in cramfs_blkdev_read()
297 static u32 cramfs_get_block_range(struct inode *inode, u32 pgoff, u32 *pages) in cramfs_get_block_range() argument
318 pgoff+i, pgoff + *pages - 1, in cramfs_get_block_range()
324 } while (++i < *pages); in cramfs_get_block_range()
326 *pages = i; in cramfs_get_block_range()
359 unsigned int pages, max_pages, offset; in cramfs_physmem_mmap() local
[all …]
/fs/crypto/
Dbio.c125 struct page *pages[16]; /* write up to 16 pages at a time */ in fscrypt_zeroout_range() local
139 BUILD_BUG_ON(ARRAY_SIZE(pages) > BIO_MAX_VECS); in fscrypt_zeroout_range()
140 nr_pages = min_t(u64, ARRAY_SIZE(pages), in fscrypt_zeroout_range()
151 pages[i] = fscrypt_alloc_bounce_page(i == 0 ? GFP_NOFS : in fscrypt_zeroout_range()
153 if (!pages[i]) in fscrypt_zeroout_range()
172 ZERO_PAGE(0), pages[i], in fscrypt_zeroout_range()
182 ret = bio_add_page(bio, pages[i++], offset, 0); in fscrypt_zeroout_range()
200 fscrypt_free_bounce_page(pages[i]); in fscrypt_zeroout_range()
/fs/ceph/
Daddr.c238 ceph_put_page_vector(osd_data->pages, num_pages, false); in finish_netfs_read()
251 struct page **pages; in ceph_netfs_issue_op() local
268 err = iov_iter_get_pages_alloc(&iter, &pages, len, &page_off); in ceph_netfs_issue_op()
278 osd_req_op_extent_osd_data_pages(req, 0, pages, len, 0, false, false); in ceph_netfs_issue_op()
664 page = osd_data->pages[j]; in writepages_finish()
687 release_pages(osd_data->pages, num_pages); in writepages_finish()
697 mempool_free(osd_data->pages, ceph_wb_pagevec_pool); in writepages_finish()
699 kfree(osd_data->pages); in writepages_finish()
788 struct page **pages = NULL, **data_pages; in ceph_writepages_start() local
803 page = pvec.pages[i]; in ceph_writepages_start()
[all …]
/fs/nfs/
Dnfs3acl.c50 struct page *pages[NFSACL_MAXPAGES] = { }; in nfs3_get_acl() local
54 .pages = pages, in nfs3_get_acl()
103 for (count = 0; count < NFSACL_MAXPAGES && args.pages[count]; count++) in nfs3_get_acl()
104 __free_page(args.pages[count]); in nfs3_get_acl()
167 struct page *pages[NFSACL_MAXPAGES]; in __nfs3_proc_setacls() local
172 .pages = pages, in __nfs3_proc_setacls()
206 args.pages[args.npages] = alloc_page(GFP_KERNEL); in __nfs3_proc_setacls()
207 if (args.pages[args.npages] == NULL) in __nfs3_proc_setacls()
243 __free_page(args.pages[args.npages]); in __nfs3_proc_setacls()
Dpnfs_dev.c103 struct page **pages = NULL; in nfs4_get_device_info() local
124 pages = kcalloc(max_pages, sizeof(struct page *), gfp_flags); in nfs4_get_device_info()
125 if (!pages) in nfs4_get_device_info()
129 pages[i] = alloc_page(gfp_flags); in nfs4_get_device_info()
130 if (!pages[i]) in nfs4_get_device_info()
136 pdev->pages = pages; in nfs4_get_device_info()
158 __free_page(pages[i]); in nfs4_get_device_info()
159 kfree(pages); in nfs4_get_device_info()
Dnfs42proc.c1176 struct page *pages[NFS4XATTR_MAXPAGES]; in _nfs42_proc_setxattr() local
1179 .xattr_pages = pages, in _nfs42_proc_setxattr()
1207 put_page(pages[np - 1]); in _nfs42_proc_setxattr()
1216 void *buf, size_t buflen, struct page **pages, in _nfs42_proc_getxattr() argument
1233 arg.xattr_pages = pages; in _nfs42_proc_getxattr()
1250 nfs4_xattr_cache_add(inode, name, NULL, pages, res.xattr_len); in _nfs42_proc_getxattr()
1255 _copy_from_pages(buf, pages, 0, res.xattr_len); in _nfs42_proc_getxattr()
1265 struct page **pages; in _nfs42_proc_listxattrs() local
1294 pages = kcalloc(np, sizeof(struct page *), GFP_KERNEL); in _nfs42_proc_listxattrs()
1295 if (!pages) in _nfs42_proc_listxattrs()
[all …]
/fs/orangefs/
Dinode.c86 struct page **pages; member
93 struct inode *inode = ow->pages[0]->mapping->host; in orangefs_writepages_work()
104 set_page_writeback(ow->pages[i]); in orangefs_writepages_work()
105 ow->bv[i].bv_page = ow->pages[i]; in orangefs_writepages_work()
106 ow->bv[i].bv_len = min(page_offset(ow->pages[i]) + PAGE_SIZE, in orangefs_writepages_work()
108 max(ow->off, page_offset(ow->pages[i])); in orangefs_writepages_work()
111 page_offset(ow->pages[i]); in orangefs_writepages_work()
128 SetPageError(ow->pages[i]); in orangefs_writepages_work()
129 mapping_set_error(ow->pages[i]->mapping, ret); in orangefs_writepages_work()
130 if (PagePrivate(ow->pages[i])) { in orangefs_writepages_work()
[all …]
/fs/btrfs/
Dstruct-funcs.c74 token->kaddr = page_address(token->eb->pages[idx]); \
80 token->kaddr = page_address(token->eb->pages[idx + 1]); \
91 char *kaddr = page_address(eb->pages[idx]); \
101 kaddr = page_address(eb->pages[idx + 1]); \
125 token->kaddr = page_address(token->eb->pages[idx]); \
133 token->kaddr = page_address(token->eb->pages[idx + 1]); \
143 char *kaddr = page_address(eb->pages[idx]); \
156 kaddr = page_address(eb->pages[idx + 1]); \
/fs/cifs/
Dfscache.h107 struct list_head *pages, in cifs_readpages_from_fscache() argument
111 return __cifs_readpages_from_fscache(inode, mapping, pages, in cifs_readpages_from_fscache()
124 struct list_head *pages) in cifs_fscache_readpages_cancel() argument
127 return __cifs_fscache_readpages_cancel(inode, pages); in cifs_fscache_readpages_cancel()
167 struct list_head *pages, in cifs_readpages_from_fscache() argument
177 struct list_head *pages) in cifs_fscache_readpages_cancel() argument
/fs/ntfs3/
Dfile.c841 struct page **pages, u32 pages_per_frame, in ntfs_get_frame_pages() argument
855 page = pages[npages]; in ntfs_get_frame_pages()
866 pages[npages] = page; in ntfs_get_frame_pages()
887 struct page *page, **pages = NULL; in ntfs_compress_write() local
907 pages = kmalloc_array(pages_per_frame, sizeof(struct page *), GFP_NOFS); in ntfs_compress_write()
908 if (!pages) in ntfs_compress_write()
941 pages, pages_per_frame, in ntfs_compress_write()
947 err = ni_read_frame(ni, frame_vbo, pages, in ntfs_compress_write()
951 page = pages[ip]; in ntfs_compress_write()
962 page = pages[ip]; in ntfs_compress_write()
[all …]
/fs/9p/
Dcache.h36 struct list_head *pages,
61 struct list_head *pages, in v9fs_readpages_from_fscache() argument
64 return __v9fs_readpages_from_fscache(inode, mapping, pages, in v9fs_readpages_from_fscache()
117 struct list_head *pages, in v9fs_readpages_from_fscache() argument
/fs/fuse/
Dioctl.c204 ap.pages = fuse_pages_alloc(fm->fc->max_pages, GFP_KERNEL, &ap.descs); in fuse_do_ioctl()
206 if (!ap.pages || !iov_page) in fuse_do_ioctl()
248 ap.pages[ap.num_pages] = alloc_page(GFP_KERNEL | __GFP_HIGHMEM); in fuse_do_ioctl()
249 if (!ap.pages[ap.num_pages]) in fuse_do_ioctl()
269 c = copy_page_from_iter(ap.pages[i], 0, PAGE_SIZE, &ii); in fuse_do_ioctl()
307 vaddr = kmap_atomic(ap.pages[0]); in fuse_do_ioctl()
336 c = copy_page_to_iter(ap.pages[i], 0, PAGE_SIZE, &ii); in fuse_do_ioctl()
344 __free_page(ap.pages[--ap.num_pages]); in fuse_do_ioctl()
345 kfree(ap.pages); in fuse_do_ioctl()
/fs/btrfs/tests/
Dextent-io-tests.c23 struct page *pages[16]; in process_page_range() local
34 ARRAY_SIZE(pages)), pages); in process_page_range()
37 !PageLocked(pages[i])) in process_page_range()
39 if (flags & PROCESS_UNLOCK && PageLocked(pages[i])) in process_page_range()
40 unlock_page(pages[i]); in process_page_range()
41 put_page(pages[i]); in process_page_range()
43 put_page(pages[i]); in process_page_range()

123456