/fs/jfs/ |
D | jfs_metapage.c | 61 unlock_page(mp->page); in __lock_metapage() 63 lock_page(mp->page); in __lock_metapage() 92 #define mp_anchor(page) ((struct meta_anchor *)page_private(page)) argument 94 static inline struct metapage *page_to_mp(struct page *page, int offset) in page_to_mp() argument 96 if (!PagePrivate(page)) in page_to_mp() 98 return mp_anchor(page)->mp[offset >> L2PSIZE]; in page_to_mp() 101 static inline int insert_metapage(struct page *page, struct metapage *mp) in insert_metapage() argument 107 if (PagePrivate(page)) in insert_metapage() 108 a = mp_anchor(page); in insert_metapage() 113 set_page_private(page, (unsigned long)a); in insert_metapage() [all …]
|
/fs/sysv/ |
D | dir.c | 30 static inline void dir_put_page(struct page *page) in dir_put_page() argument 32 kunmap(page); in dir_put_page() 33 page_cache_release(page); in dir_put_page() 36 static int dir_commit_chunk(struct page *page, loff_t pos, unsigned len) in dir_commit_chunk() argument 38 struct address_space *mapping = page->mapping; in dir_commit_chunk() 42 block_write_end(NULL, mapping, pos, len, len, page, NULL); in dir_commit_chunk() 48 err = write_one_page(page, 1); in dir_commit_chunk() 50 unlock_page(page); in dir_commit_chunk() 54 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page() 57 struct page *page = read_mapping_page(mapping, n, NULL); in dir_get_page() local [all …]
|
/fs/9p/ |
D | cache.h | 45 extern int __v9fs_fscache_release_page(struct page *page, gfp_t gfp); 46 extern void __v9fs_fscache_invalidate_page(struct page *page); 48 struct page *page); 53 extern void __v9fs_readpage_to_fscache(struct inode *inode, struct page *page); 55 struct page *page); 57 static inline int v9fs_fscache_release_page(struct page *page, in v9fs_fscache_release_page() argument 60 return __v9fs_fscache_release_page(page, gfp); in v9fs_fscache_release_page() 63 static inline void v9fs_fscache_invalidate_page(struct page *page) in v9fs_fscache_invalidate_page() argument 65 __v9fs_fscache_invalidate_page(page); in v9fs_fscache_invalidate_page() 69 struct page *page) in v9fs_readpage_from_fscache() argument [all …]
|
D | vfs_addr.c | 52 static int v9fs_fid_readpage(void *data, struct page *page) in v9fs_fid_readpage() argument 55 struct inode *inode = page->mapping->host; in v9fs_fid_readpage() 56 struct bio_vec bvec = {.bv_page = page, .bv_len = PAGE_SIZE}; in v9fs_fid_readpage() 62 BUG_ON(!PageLocked(page)); in v9fs_fid_readpage() 64 retval = v9fs_readpage_from_fscache(inode, page); in v9fs_fid_readpage() 70 retval = p9_client_read(fid, page_offset(page), &to, &err); in v9fs_fid_readpage() 72 v9fs_uncache_page(inode, page); in v9fs_fid_readpage() 77 zero_user(page, retval, PAGE_SIZE - retval); in v9fs_fid_readpage() 78 flush_dcache_page(page); in v9fs_fid_readpage() 79 SetPageUptodate(page); in v9fs_fid_readpage() [all …]
|
/fs/ceph/ |
D | addr.c | 58 static inline struct ceph_snap_context *page_snap_context(struct page *page) in page_snap_context() argument 60 if (PagePrivate(page)) in page_snap_context() 61 return (void *)page->private; in page_snap_context() 69 static int ceph_set_page_dirty(struct page *page) in ceph_set_page_dirty() argument 71 struct address_space *mapping = page->mapping; in ceph_set_page_dirty() 76 if (PageDirty(page)) { in ceph_set_page_dirty() 78 mapping->host, page, page->index); in ceph_set_page_dirty() 79 BUG_ON(!PagePrivate(page)); in ceph_set_page_dirty() 106 mapping->host, page, page->index, in ceph_set_page_dirty() 116 BUG_ON(PagePrivate(page)); in ceph_set_page_dirty() [all …]
|
/fs/logfs/ |
D | readwrite.c | 205 static void prelock_page(struct super_block *sb, struct page *page, int lock) in prelock_page() argument 209 BUG_ON(!PageLocked(page)); in prelock_page() 211 BUG_ON(PagePreLocked(page)); in prelock_page() 212 SetPagePreLocked(page); in prelock_page() 215 if (PagePreLocked(page)) in prelock_page() 218 SetPagePreLocked(page); in prelock_page() 222 static void preunlock_page(struct super_block *sb, struct page *page, int lock) in preunlock_page() argument 226 BUG_ON(!PageLocked(page)); in preunlock_page() 228 ClearPagePreLocked(page); in preunlock_page() 231 BUG_ON(!PagePreLocked(page)); in preunlock_page() [all …]
|
D | file.c | 14 struct page **pagep, void **fsdata) in logfs_write_begin() 17 struct page *page; in logfs_write_begin() local 20 page = grab_cache_page_write_begin(mapping, index, flags); in logfs_write_begin() 21 if (!page) in logfs_write_begin() 23 *pagep = page; in logfs_write_begin() 25 if ((len == PAGE_CACHE_SIZE) || PageUptodate(page)) in logfs_write_begin() 32 zero_user_segments(page, 0, start, end, PAGE_CACHE_SIZE); in logfs_write_begin() 35 return logfs_readpage_nolock(page); in logfs_write_begin() 39 loff_t pos, unsigned len, unsigned copied, struct page *page, in logfs_write_end() argument 43 pgoff_t index = page->index; in logfs_write_end() [all …]
|
/fs/afs/ |
D | file.c | 21 static int afs_readpage(struct file *file, struct page *page); 22 static void afs_invalidatepage(struct page *page, unsigned int offset, 24 static int afs_releasepage(struct page *page, gfp_t gfp_flags); 25 static int afs_launder_page(struct page *page); 109 static void afs_file_readpage_read_complete(struct page *page, in afs_file_readpage_read_complete() argument 113 _enter("%p,%p,%d", page, data, error); in afs_file_readpage_read_complete() 118 SetPageUptodate(page); in afs_file_readpage_read_complete() 119 unlock_page(page); in afs_file_readpage_read_complete() 126 int afs_page_filler(void *data, struct page *page) in afs_page_filler() argument 128 struct inode *inode = page->mapping->host; in afs_page_filler() [all …]
|
D | write.c | 20 struct page *page); 25 int afs_set_page_dirty(struct page *page) in afs_set_page_dirty() argument 28 return __set_page_dirty_nobuffers(page); in afs_set_page_dirty() 87 loff_t pos, struct page *page) in afs_fill_page() argument 101 ret = afs_vnode_fetch_data(vnode, key, pos, len, page); in afs_fill_page() 120 struct page **pagep, void **fsdata) in afs_write_begin() 124 struct page *page; in afs_write_begin() local 146 page = grab_cache_page_write_begin(mapping, index, flags); in afs_write_begin() 147 if (!page) { in afs_write_begin() 152 if (!PageUptodate(page) && len != PAGE_CACHE_SIZE) { in afs_write_begin() [all …]
|
/fs/nilfs2/ |
D | dir.c | 58 static inline void nilfs_put_page(struct page *page) in nilfs_put_page() argument 60 kunmap(page); in nilfs_put_page() 61 page_cache_release(page); in nilfs_put_page() 78 static int nilfs_prepare_chunk(struct page *page, unsigned from, unsigned to) in nilfs_prepare_chunk() argument 80 loff_t pos = page_offset(page) + from; in nilfs_prepare_chunk() 81 return __block_write_begin(page, pos, to - from, nilfs_get_block); in nilfs_prepare_chunk() 84 static void nilfs_commit_chunk(struct page *page, in nilfs_commit_chunk() argument 89 loff_t pos = page_offset(page) + from; in nilfs_commit_chunk() 94 nr_dirty = nilfs_page_count_clean_buffers(page, from, to); in nilfs_commit_chunk() 95 copied = block_write_end(NULL, mapping, pos, len, len, page, NULL); in nilfs_commit_chunk() [all …]
|
D | page.c | 43 __nilfs_get_page_block(struct page *page, unsigned long block, pgoff_t index, in __nilfs_get_page_block() argument 50 if (!page_has_buffers(page)) in __nilfs_get_page_block() 51 create_empty_buffers(page, 1 << blkbits, b_state); in __nilfs_get_page_block() 54 bh = nilfs_page_get_nth_block(page, block - first_block); in __nilfs_get_page_block() 68 struct page *page; in nilfs_grab_buffer() local 71 page = grab_cache_page(mapping, index); in nilfs_grab_buffer() 72 if (unlikely(!page)) in nilfs_grab_buffer() 75 bh = __nilfs_get_page_block(page, blkoff, index, blkbits, b_state); in nilfs_grab_buffer() 77 unlock_page(page); in nilfs_grab_buffer() 78 page_cache_release(page); in nilfs_grab_buffer() [all …]
|
/fs/ubifs/ |
D | file.c | 104 static int do_readpage(struct page *page) in do_readpage() argument 110 struct inode *inode = page->mapping->host; in do_readpage() 114 inode->i_ino, page->index, i_size, page->flags); in do_readpage() 115 ubifs_assert(!PageChecked(page)); in do_readpage() 116 ubifs_assert(!PagePrivate(page)); in do_readpage() 118 addr = kmap(page); in do_readpage() 120 block = page->index << UBIFS_BLOCKS_PER_PAGE_SHIFT; in do_readpage() 124 SetPageChecked(page); in do_readpage() 166 SetPageChecked(page); in do_readpage() 171 page->index, inode->i_ino, err); in do_readpage() [all …]
|
/fs/ecryptfs/ |
D | mmap.c | 47 struct page *ecryptfs_get_locked_page(struct inode *inode, loff_t index) in ecryptfs_get_locked_page() 49 struct page *page = read_mapping_page(inode->i_mapping, index, NULL); in ecryptfs_get_locked_page() local 50 if (!IS_ERR(page)) in ecryptfs_get_locked_page() 51 lock_page(page); in ecryptfs_get_locked_page() 52 return page; in ecryptfs_get_locked_page() 65 static int ecryptfs_writepage(struct page *page, struct writeback_control *wbc) in ecryptfs_writepage() argument 69 rc = ecryptfs_encrypt_page(page); in ecryptfs_writepage() 72 "page (upper index [0x%.16lx])\n", page->index); in ecryptfs_writepage() 73 ClearPageUptodate(page); in ecryptfs_writepage() 76 SetPageUptodate(page); in ecryptfs_writepage() [all …]
|
/fs/minix/ |
D | dir.c | 28 static inline void dir_put_page(struct page *page) in dir_put_page() argument 30 kunmap(page); in dir_put_page() 31 page_cache_release(page); in dir_put_page() 48 static int dir_commit_chunk(struct page *page, loff_t pos, unsigned len) in dir_commit_chunk() argument 50 struct address_space *mapping = page->mapping; in dir_commit_chunk() 53 block_write_end(NULL, mapping, pos, len, len, page, NULL); in dir_commit_chunk() 60 err = write_one_page(page, 1); in dir_commit_chunk() 62 unlock_page(page); in dir_commit_chunk() 66 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page() 69 struct page *page = read_mapping_page(mapping, n, NULL); in dir_get_page() local [all …]
|
/fs/exofs/ |
D | dir.c | 41 static inline void exofs_put_page(struct page *page) in exofs_put_page() argument 43 kunmap(page); in exofs_put_page() 44 page_cache_release(page); in exofs_put_page() 57 static int exofs_commit_chunk(struct page *page, loff_t pos, unsigned len) in exofs_commit_chunk() argument 59 struct address_space *mapping = page->mapping; in exofs_commit_chunk() 65 if (!PageUptodate(page)) in exofs_commit_chunk() 66 SetPageUptodate(page); in exofs_commit_chunk() 72 set_page_dirty(page); in exofs_commit_chunk() 75 err = write_one_page(page, 1); in exofs_commit_chunk() 77 unlock_page(page); in exofs_commit_chunk() [all …]
|
/fs/ext2/ |
D | dir.c | 67 static inline void ext2_put_page(struct page *page) in ext2_put_page() argument 69 kunmap(page); in ext2_put_page() 70 page_cache_release(page); in ext2_put_page() 88 static int ext2_commit_chunk(struct page *page, loff_t pos, unsigned len) in ext2_commit_chunk() argument 90 struct address_space *mapping = page->mapping; in ext2_commit_chunk() 95 block_write_end(NULL, mapping, pos, len, len, page, NULL); in ext2_commit_chunk() 103 err = write_one_page(page, 1); in ext2_commit_chunk() 107 unlock_page(page); in ext2_commit_chunk() 113 static void ext2_check_page(struct page *page, int quiet) in ext2_check_page() argument 115 struct inode *dir = page->mapping->host; in ext2_check_page() [all …]
|
/fs/ufs/ |
D | dir.c | 43 static int ufs_commit_chunk(struct page *page, loff_t pos, unsigned len) in ufs_commit_chunk() argument 45 struct address_space *mapping = page->mapping; in ufs_commit_chunk() 50 block_write_end(NULL, mapping, pos, len, len, page, NULL); in ufs_commit_chunk() 56 err = write_one_page(page, 1); in ufs_commit_chunk() 58 unlock_page(page); in ufs_commit_chunk() 62 static inline void ufs_put_page(struct page *page) in ufs_put_page() argument 64 kunmap(page); in ufs_put_page() 65 page_cache_release(page); in ufs_put_page() 72 struct page *page; in ufs_inode_by_name() local 74 de = ufs_find_entry(dir, qstr, &page); in ufs_inode_by_name() [all …]
|
/fs/ |
D | buffer.c | 87 void buffer_check_dirty_writeback(struct page *page, in buffer_check_dirty_writeback() argument 94 BUG_ON(!PageLocked(page)); in buffer_check_dirty_writeback() 96 if (!page_has_buffers(page)) in buffer_check_dirty_writeback() 99 if (PageWriteback(page)) in buffer_check_dirty_writeback() 102 head = page_buffers(page); in buffer_check_dirty_writeback() 128 __clear_page_buffers(struct page *page) in __clear_page_buffers() argument 130 ClearPagePrivate(page); in __clear_page_buffers() 131 set_page_private(page, 0); in __clear_page_buffers() 132 page_cache_release(page); in __clear_page_buffers() 210 struct page *page; in __find_get_block_slow() local [all …]
|
D | mpage.c | 60 struct page *first_page = bio->bi_io_vec[0].bv_page; in mpage_end_io() 69 struct page *page = bv->bv_page; in mpage_end_io() local 70 page_endio(page, bio_data_dir(bio), bio->bi_error); in mpage_end_io() 79 struct page *first_page = bio->bi_io_vec[0].bv_page; in mpage_bio_submit() 134 map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) in map_buffer_to_page() argument 136 struct inode *inode = page->mapping->host; in map_buffer_to_page() 140 if (!page_has_buffers(page)) { in map_buffer_to_page() 147 SetPageUptodate(page); in map_buffer_to_page() 150 create_empty_buffers(page, i_blocksize(inode), 0); in map_buffer_to_page() 152 head = page_buffers(page); in map_buffer_to_page() [all …]
|
/fs/f2fs/ |
D | data.c | 36 static bool __is_cp_guaranteed(struct page *page) in __is_cp_guaranteed() argument 38 struct address_space *mapping = page->mapping; in __is_cp_guaranteed() 51 is_cold_data(page)) in __is_cp_guaranteed() 71 struct page *page; in __read_end_io() local 76 page = bv->bv_page; in __read_end_io() 79 if (bio->bi_error || PageError(page)) { in __read_end_io() 80 ClearPageUptodate(page); in __read_end_io() 81 SetPageError(page); in __read_end_io() 83 SetPageUptodate(page); in __read_end_io() 85 unlock_page(page); in __read_end_io() [all …]
|
D | node.h | 234 static inline nid_t ino_of_node(struct page *node_page) in ino_of_node() 240 static inline nid_t nid_of_node(struct page *node_page) in nid_of_node() 246 static inline unsigned int ofs_of_node(struct page *node_page) in ofs_of_node() 253 static inline __u64 cpver_of_node(struct page *node_page) in cpver_of_node() 259 static inline block_t next_blkaddr_of_node(struct page *node_page) in next_blkaddr_of_node() 265 static inline void fill_node_footer(struct page *page, nid_t nid, in fill_node_footer() argument 268 struct f2fs_node *rn = F2FS_NODE(page); in fill_node_footer() 284 static inline void copy_node_footer(struct page *dst, struct page *src) in copy_node_footer() 291 static inline void fill_node_footer_blkaddr(struct page *page, block_t blkaddr) in fill_node_footer_blkaddr() argument 293 struct f2fs_checkpoint *ckpt = F2FS_CKPT(F2FS_P_SB(page)); in fill_node_footer_blkaddr() [all …]
|
D | inline.c | 46 void read_inline_data(struct page *page, struct page *ipage) in read_inline_data() argument 48 struct inode *inode = page->mapping->host; in read_inline_data() 51 if (PageUptodate(page)) in read_inline_data() 54 f2fs_bug_on(F2FS_P_SB(page), page->index); in read_inline_data() 56 zero_user_segment(page, MAX_INLINE_DATA(inode), PAGE_SIZE); in read_inline_data() 60 dst_addr = kmap_atomic(page); in read_inline_data() 62 flush_dcache_page(page); in read_inline_data() 64 if (!PageUptodate(page)) in read_inline_data() 65 SetPageUptodate(page); in read_inline_data() 68 void truncate_inline_inode(struct inode *inode, struct page *ipage, u64 from) in truncate_inline_inode() [all …]
|
/fs/gfs2/ |
D | aops.c | 41 static void gfs2_page_add_databufs(struct gfs2_inode *ip, struct page *page, in gfs2_page_add_databufs() argument 44 struct buffer_head *head = page_buffers(page); in gfs2_page_add_databufs() 97 static int gfs2_writepage_common(struct page *page, in gfs2_writepage_common() argument 100 struct inode *inode = page->mapping->host; in gfs2_writepage_common() 113 if (page->index > end_index || (page->index == end_index && !offset)) { in gfs2_writepage_common() 114 page->mapping->a_ops->invalidatepage(page, 0, PAGE_CACHE_SIZE); in gfs2_writepage_common() 119 redirty_page_for_writepage(wbc, page); in gfs2_writepage_common() 121 unlock_page(page); in gfs2_writepage_common() 132 static int gfs2_writepage(struct page *page, struct writeback_control *wbc) in gfs2_writepage() argument 136 ret = gfs2_writepage_common(page, wbc); in gfs2_writepage() [all …]
|
/fs/ntfs/ |
D | aops.c | 62 struct page *page; in ntfs_end_buffer_async_read() local 67 page = bh->b_page; in ntfs_end_buffer_async_read() 68 vi = page->mapping->host; in ntfs_end_buffer_async_read() 77 file_ofs = ((s64)page->index << PAGE_CACHE_SHIFT) + in ntfs_end_buffer_async_read() 96 kaddr = kmap_atomic(page); in ntfs_end_buffer_async_read() 99 flush_dcache_page(page); in ntfs_end_buffer_async_read() 105 SetPageError(page); in ntfs_end_buffer_async_read() 109 first = page_buffers(page); in ntfs_end_buffer_async_read() 137 if (likely(page_uptodate && !PageError(page))) in ntfs_end_buffer_async_read() 138 SetPageUptodate(page); in ntfs_end_buffer_async_read() [all …]
|
D | compress.c | 100 static void zero_partial_compressed_page(struct page *page, in zero_partial_compressed_page() argument 103 u8 *kp = page_address(page); in zero_partial_compressed_page() 107 if (((s64)page->index << PAGE_CACHE_SHIFT) >= initialized_size) { in zero_partial_compressed_page() 123 static inline void handle_bounds_compressed_page(struct page *page, in handle_bounds_compressed_page() argument 126 if ((page->index >= (initialized_size >> PAGE_CACHE_SHIFT)) && in handle_bounds_compressed_page() 128 zero_partial_compressed_page(page, initialized_size); in handle_bounds_compressed_page() 169 static int ntfs_decompress(struct page *dest_pages[], int *dest_index, in ntfs_decompress() 185 struct page *dp; /* Current destination page being worked on. */ in ntfs_decompress() 482 int ntfs_read_compressed_block(struct page *page) in ntfs_read_compressed_block() argument 486 struct address_space *mapping = page->mapping; in ntfs_read_compressed_block() [all …]
|