• Home
  • Raw
  • Download

Lines Matching refs:mapping

517 int sync_mapping_buffers(struct address_space *mapping)  in sync_mapping_buffers()  argument
519 struct address_space *buffer_mapping = mapping->private_data; in sync_mapping_buffers()
521 if (buffer_mapping == NULL || list_empty(&mapping->private_list)) in sync_mapping_buffers()
525 &mapping->private_list); in sync_mapping_buffers()
548 struct address_space *mapping = inode->i_mapping; in mark_buffer_dirty_inode() local
549 struct address_space *buffer_mapping = bh->b_page->mapping; in mark_buffer_dirty_inode()
552 if (!mapping->private_data) { in mark_buffer_dirty_inode()
553 mapping->private_data = buffer_mapping; in mark_buffer_dirty_inode()
555 BUG_ON(mapping->private_data != buffer_mapping); in mark_buffer_dirty_inode()
560 &mapping->private_list); in mark_buffer_dirty_inode()
561 bh->b_assoc_map = mapping; in mark_buffer_dirty_inode()
576 void __set_page_dirty(struct page *page, struct address_space *mapping, in __set_page_dirty() argument
581 xa_lock_irqsave(&mapping->i_pages, flags); in __set_page_dirty()
582 if (page->mapping) { /* Race with truncate? */ in __set_page_dirty()
584 account_page_dirtied(page, mapping); in __set_page_dirty()
585 __xa_set_mark(&mapping->i_pages, page_index(page), in __set_page_dirty()
588 xa_unlock_irqrestore(&mapping->i_pages, flags); in __set_page_dirty()
620 struct address_space *mapping = page_mapping(page); in __set_page_dirty_buffers() local
622 if (unlikely(!mapping)) in __set_page_dirty_buffers()
625 spin_lock(&mapping->private_lock); in __set_page_dirty_buffers()
641 spin_unlock(&mapping->private_lock); in __set_page_dirty_buffers()
644 __set_page_dirty(page, mapping, 1); in __set_page_dirty_buffers()
649 __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); in __set_page_dirty_buffers()
678 struct address_space *mapping; in fsync_buffers_list() local
688 mapping = bh->b_assoc_map; in fsync_buffers_list()
695 bh->b_assoc_map = mapping; in fsync_buffers_list()
727 mapping = bh->b_assoc_map; in fsync_buffers_list()
734 &mapping->private_list); in fsync_buffers_list()
735 bh->b_assoc_map = mapping; in fsync_buffers_list()
765 struct address_space *mapping = &inode->i_data; in invalidate_inode_buffers() local
766 struct list_head *list = &mapping->private_list; in invalidate_inode_buffers()
767 struct address_space *buffer_mapping = mapping->private_data; in invalidate_inode_buffers()
788 struct address_space *mapping = &inode->i_data; in remove_inode_buffers() local
789 struct list_head *list = &mapping->private_list; in remove_inode_buffers()
790 struct address_space *buffer_mapping = mapping->private_data; in remove_inode_buffers()
1107 struct address_space *mapping = NULL; in mark_buffer_dirty() local
1111 mapping = page_mapping(page); in mark_buffer_dirty()
1112 if (mapping) in mark_buffer_dirty()
1113 __set_page_dirty(page, mapping, 0); in mark_buffer_dirty()
1116 if (mapping) in mark_buffer_dirty()
1117 __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); in mark_buffer_dirty()
1126 if (bh->b_page && bh->b_page->mapping) in mark_buffer_write_io_error()
1127 mapping_set_error(bh->b_page->mapping, -EIO); in mark_buffer_write_io_error()
1158 struct address_space *buffer_mapping = bh->b_page->mapping; in __bforget()
1539 spin_lock(&page->mapping->private_lock); in create_empty_buffers()
1551 spin_unlock(&page->mapping->private_lock); in create_empty_buffers()
1828 mapping_set_error(page->mapping, err); in __block_write_full_page()
1947 struct inode *inode = page->mapping->host; in __block_write_begin_int()
2078 int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len, in block_write_begin() argument
2085 page = grab_cache_page_write_begin(mapping, index, flags); in block_write_begin()
2101 int block_write_end(struct file *file, struct address_space *mapping, in block_write_end() argument
2105 struct inode *inode = mapping->host; in block_write_end()
2137 int generic_write_end(struct file *file, struct address_space *mapping, in generic_write_end() argument
2141 struct inode *inode = mapping->host; in generic_write_end()
2145 copied = block_write_end(file, mapping, pos, len, copied, page, fsdata); in generic_write_end()
2230 struct inode *inode = page->mapping->host; in block_read_full_page()
2320 struct address_space *mapping = inode->i_mapping; in generic_cont_expand_simple() local
2329 err = pagecache_write_begin(NULL, mapping, size, 0, in generic_cont_expand_simple()
2334 err = pagecache_write_end(NULL, mapping, size, 0, 0, page, fsdata); in generic_cont_expand_simple()
2342 static int cont_expand_zero(struct file *file, struct address_space *mapping, in cont_expand_zero() argument
2345 struct inode *inode = mapping->host; in cont_expand_zero()
2365 err = pagecache_write_begin(file, mapping, curpos, len, 0, in cont_expand_zero()
2370 err = pagecache_write_end(file, mapping, curpos, len, len, in cont_expand_zero()
2377 balance_dirty_pages_ratelimited(mapping); in cont_expand_zero()
2398 err = pagecache_write_begin(file, mapping, curpos, len, 0, in cont_expand_zero()
2403 err = pagecache_write_end(file, mapping, curpos, len, len, in cont_expand_zero()
2418 int cont_write_begin(struct file *file, struct address_space *mapping, in cont_write_begin() argument
2423 struct inode *inode = mapping->host; in cont_write_begin()
2428 err = cont_expand_zero(file, mapping, pos, bytes); in cont_write_begin()
2438 return block_write_begin(mapping, pos, len, flags, pagep, get_block); in cont_write_begin()
2444 struct inode *inode = page->mapping->host; in block_commit_write()
2479 if ((page->mapping != inode->i_mapping) || in block_page_mkwrite()
2528 spin_lock(&page->mapping->private_lock); in attach_nobh_buffers()
2538 spin_unlock(&page->mapping->private_lock); in attach_nobh_buffers()
2546 int nobh_write_begin(struct address_space *mapping, in nobh_write_begin() argument
2551 struct inode *inode = mapping->host; in nobh_write_begin()
2569 page = grab_cache_page_write_begin(mapping, index, flags); in nobh_write_begin()
2687 int nobh_write_end(struct file *file, struct address_space *mapping, in nobh_write_end() argument
2691 struct inode *inode = page->mapping->host; in nobh_write_end()
2699 return generic_write_end(file, mapping, pos, len, in nobh_write_end()
2730 struct inode * const inode = page->mapping->host; in nobh_writepage()
2764 int nobh_truncate_page(struct address_space *mapping, in nobh_truncate_page() argument
2772 struct inode *inode = mapping->host; in nobh_truncate_page()
2787 page = grab_cache_page(mapping, index); in nobh_truncate_page()
2796 return block_truncate_page(mapping, from, get_block); in nobh_truncate_page()
2817 err = mapping->a_ops->readpage(NULL, page); in nobh_truncate_page()
2842 int block_truncate_page(struct address_space *mapping, in block_truncate_page() argument
2850 struct inode *inode = mapping->host; in block_truncate_page()
2865 page = grab_cache_page(mapping, index); in block_truncate_page()
2924 struct inode * const inode = page->mapping->host; in block_write_full_page()
2954 sector_t generic_block_bmap(struct address_space *mapping, sector_t block, in generic_block_bmap() argument
2957 struct inode *inode = mapping->host; in generic_block_bmap()
3243 struct address_space * const mapping = page->mapping; in try_to_free_buffers() local
3251 if (mapping == NULL) { /* can this still happen? */ in try_to_free_buffers()
3256 spin_lock(&mapping->private_lock); in try_to_free_buffers()
3275 spin_unlock(&mapping->private_lock); in try_to_free_buffers()