Home
last modified time | relevance | path

Searched refs:mapping (Results 1 – 25 of 39) sorted by relevance

12

/include/linux/
Dpagemap.h48 static inline void mapping_set_error(struct address_space *mapping, int error) in mapping_set_error() argument
54 filemap_set_wb_err(mapping, error); in mapping_set_error()
58 set_bit(AS_ENOSPC, &mapping->flags); in mapping_set_error()
60 set_bit(AS_EIO, &mapping->flags); in mapping_set_error()
63 static inline void mapping_set_unevictable(struct address_space *mapping) in mapping_set_unevictable() argument
65 set_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_set_unevictable()
68 static inline void mapping_clear_unevictable(struct address_space *mapping) in mapping_clear_unevictable() argument
70 clear_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_clear_unevictable()
73 static inline int mapping_unevictable(struct address_space *mapping) in mapping_unevictable() argument
75 if (mapping) in mapping_unevictable()
[all …]
Dio-mapping.h56 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument
58 iomap_free(mapping->base, mapping->size); in io_mapping_fini()
63 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument
69 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc()
70 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc()
72 return iomap_atomic_prot_pfn(pfn, mapping->prot); in io_mapping_map_atomic_wc()
82 io_mapping_map_wc(struct io_mapping *mapping, in io_mapping_map_wc() argument
88 BUG_ON(offset >= mapping->size); in io_mapping_map_wc()
89 phys_addr = mapping->base + offset; in io_mapping_map_wc()
126 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument
[all …]
Dcleancache.h52 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument
54 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping()
58 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled()
104 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument
108 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page()
109 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page()
112 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument
114 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode()
115 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
Dtpm_eventlog.h164 void *mapping = NULL; in __calc_tpm2_event_size() local
183 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size()
185 if (!mapping) { in __calc_tpm2_event_size()
190 mapping = marker_start; in __calc_tpm2_event_size()
193 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size()
214 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size()
216 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size()
218 if (!mapping) { in __calc_tpm2_event_size()
223 mapping = marker; in __calc_tpm2_event_size()
226 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size()
[all …]
Dpagevec.h29 struct address_space *mapping,
34 struct address_space *mapping,
37 struct address_space *mapping, in pagevec_lookup() argument
40 return pagevec_lookup_range(pvec, mapping, start, (pgoff_t)-1); in pagevec_lookup()
44 struct address_space *mapping, pgoff_t *index, pgoff_t end,
47 struct address_space *mapping, pgoff_t *index, pgoff_t end,
50 struct address_space *mapping, pgoff_t *index, xa_mark_t tag) in pagevec_lookup_tag() argument
52 return pagevec_lookup_range_tag(pvec, mapping, index, (pgoff_t)-1, tag); in pagevec_lookup_tag()
Dshmem_fs.h66 extern bool shmem_mapping(struct address_space *mapping);
68 static inline bool shmem_mapping(struct address_space *mapping) in shmem_mapping() argument
73 extern void shmem_unlock_mapping(struct address_space *mapping);
74 extern struct page *shmem_read_mapping_page_gfp(struct address_space *mapping,
81 extern unsigned long shmem_partial_swap_usage(struct address_space *mapping,
98 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page() argument
100 return shmem_read_mapping_page_gfp(mapping, index, in shmem_read_mapping_page()
101 mapping_gfp_mask(mapping)); in shmem_read_mapping_page()
Ddax.h143 int dax_writeback_mapping_range(struct address_space *mapping,
146 struct page *dax_layout_busy_page(struct address_space *mapping);
177 static inline struct page *dax_layout_busy_page(struct address_space *mapping) in dax_layout_busy_page() argument
182 static inline int dax_writeback_mapping_range(struct address_space *mapping, in dax_writeback_mapping_range() argument
190 if (IS_DAX(page->mapping->host)) in dax_lock_page()
220 int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index);
221 int dax_invalidate_mapping_entry_sync(struct address_space *mapping,
237 static inline bool dax_mapping(struct address_space *mapping) in dax_mapping() argument
239 return mapping->host && IS_DAX(mapping->host); in dax_mapping()
Dmpage.h17 int mpage_readpages(struct address_space *mapping, struct list_head *pages,
20 int mpage_writepages(struct address_space *mapping,
Dmigrate.h65 extern int migrate_page(struct address_space *mapping,
77 extern int migrate_huge_page_move_mapping(struct address_space *mapping,
79 extern int migrate_page_move_mapping(struct address_space *mapping,
101 static inline int migrate_huge_page_move_mapping(struct address_space *mapping, in migrate_huge_page_move_mapping() argument
111 extern void __SetPageMovable(struct page *page, struct address_space *mapping);
116 struct address_space *mapping) in __SetPageMovable() argument
Dwriteback.h388 void balance_dirty_pages_ratelimited(struct address_space *mapping);
394 int generic_writepages(struct address_space *mapping,
396 void tag_pages_for_writeback(struct address_space *mapping,
398 int write_cache_pages(struct address_space *mapping,
401 int do_writepages(struct address_space *mapping, struct writeback_control *wbc);
403 void tag_pages_for_writeback(struct address_space *mapping,
Dfs.h376 int (*readpages)(struct file *filp, struct address_space *mapping,
379 int (*write_begin)(struct file *, struct address_space *mapping,
382 int (*write_end)(struct file *, struct address_space *mapping,
418 int pagecache_write_begin(struct file *, struct address_space *mapping,
422 int pagecache_write_end(struct file *, struct address_space *mapping,
519 static inline bool mapping_tagged(struct address_space *mapping, xa_mark_t tag) in mapping_tagged() argument
521 return xa_marked(&mapping->i_pages, tag); in mapping_tagged()
524 static inline void i_mmap_lock_write(struct address_space *mapping) in i_mmap_lock_write() argument
526 down_write(&mapping->i_mmap_rwsem); in i_mmap_lock_write()
529 static inline void i_mmap_unlock_write(struct address_space *mapping) in i_mmap_unlock_write() argument
[all …]
Diomap.h155 int iomap_readpages(struct address_space *mapping, struct list_head *pages,
164 int iomap_migrate_page(struct address_space *mapping, struct page *newpage,
183 sector_t iomap_bmap(struct address_space *mapping, sector_t bno,
Dvlynq.h143 struct vlynq_mapping *mapping);
145 struct vlynq_mapping *mapping);
Dbacking-dev.h212 static inline bool mapping_cap_writeback_dirty(struct address_space *mapping) in mapping_cap_writeback_dirty() argument
214 return bdi_cap_writeback_dirty(inode_to_bdi(mapping->host)); in mapping_cap_writeback_dirty()
217 static inline bool mapping_cap_account_dirty(struct address_space *mapping) in mapping_cap_account_dirty() argument
219 return bdi_cap_account_dirty(inode_to_bdi(mapping->host)); in mapping_cap_account_dirty()
Dfscache.h111 struct address_space *mapping,
590 struct address_space *mapping, in fscache_read_or_alloc_pages() argument
598 return __fscache_read_or_alloc_pages(cookie, mapping, pages, in fscache_read_or_alloc_pages()
Dswap.h316 #define mapping_set_update(xas, mapping) do { \ argument
317 if (!dax_mapping(mapping) && !shmem_mapping(mapping)) \
365 extern int remove_mapping(struct address_space *mapping, struct page *page);
Dbuffer_head.h174 int sync_mapping_buffers(struct address_space *mapping);
225 int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len,
408 static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } in sync_mapping_buffers() argument
Dballoon_compaction.h86 extern int balloon_page_migrate(struct address_space *mapping,
Dmm.h1359 return page->mapping; in page_file_mapping()
1471 int truncate_inode_page(struct address_space *mapping, struct page *page);
1472 int generic_error_remove_page(struct address_space *mapping, struct page *page);
1481 void unmap_mapping_pages(struct address_space *mapping,
1483 void unmap_mapping_range(struct address_space *mapping,
1501 static inline void unmap_mapping_pages(struct address_space *mapping, in unmap_mapping_pages() argument
1503 static inline void unmap_mapping_range(struct address_space *mapping, in unmap_mapping_range() argument
1507 static inline void unmap_shared_mapping_range(struct address_space *mapping, in unmap_shared_mapping_range() argument
1510 unmap_mapping_range(mapping, holebegin, holelen, 0); in unmap_shared_mapping_range()
1596 void account_page_dirtied(struct page *page, struct address_space *mapping);
[all …]
/include/trace/events/
Dfilemap.h31 __entry->i_ino = page->mapping->host->i_ino;
33 if (page->mapping->host->i_sb)
34 __entry->s_dev = page->mapping->host->i_sb->s_dev;
36 __entry->s_dev = page->mapping->host->i_rdev;
58 TP_PROTO(struct address_space *mapping, errseq_t eseq),
60 TP_ARGS(mapping, eseq),
69 __entry->i_ino = mapping->host->i_ino;
71 if (mapping->host->i_sb)
72 __entry->s_dev = mapping->host->i_sb->s_dev;
74 __entry->s_dev = mapping->host->i_rdev;
Dpage_ref.h24 __field(void *, mapping)
34 __entry->mapping = page->mapping;
43 __entry->mapcount, __entry->mapping, __entry->mt,
72 __field(void *, mapping)
83 __entry->mapping = page->mapping;
93 __entry->mapcount, __entry->mapping, __entry->mt,
Dwriteback.h58 TP_PROTO(struct page *page, struct address_space *mapping),
60 TP_ARGS(page, mapping),
70 mapping ? dev_name(inode_to_bdi(mapping->host)->dev) : "(unknown)",
72 __entry->ino = mapping ? mapping->host->i_ino : 0;
85 TP_PROTO(struct page *page, struct address_space *mapping),
87 TP_ARGS(page, mapping)
92 TP_PROTO(struct page *page, struct address_space *mapping),
94 TP_ARGS(page, mapping)
255 struct address_space *mapping = page_mapping(page);
256 struct inode *inode = mapping ? mapping->host : NULL;
Derofs.h97 __entry->dev = page->mapping->host->i_sb->s_dev;
98 __entry->nid = EROFS_I(page->mapping->host)->nid;
99 __entry->dir = S_ISDIR(page->mapping->host->i_mode);
/include/asm-generic/
Dcacheflush.h55 static inline void flush_dcache_mmap_lock(struct address_space *mapping) in flush_dcache_mmap_lock() argument
61 static inline void flush_dcache_mmap_unlock(struct address_space *mapping) in flush_dcache_mmap_unlock() argument
/include/drm/
Dati_pcigart.h22 struct drm_local_map mapping; member

12