Home
last modified time | relevance | path

Searched refs:mapping (Results 1 – 25 of 45) sorted by relevance

12

/include/linux/
Dpagemap.h21 unsigned long invalidate_mapping_pages(struct address_space *mapping,
30 int invalidate_inode_pages2(struct address_space *mapping);
31 int invalidate_inode_pages2_range(struct address_space *mapping,
35 int filemap_invalidate_pages(struct address_space *mapping,
41 int filemap_fdatawait_keep_errors(struct address_space *mapping);
43 int filemap_fdatawait_range_keep_errors(struct address_space *mapping,
48 static inline int filemap_fdatawait(struct address_space *mapping) in filemap_fdatawait() argument
50 return filemap_fdatawait_range(mapping, 0, LLONG_MAX); in filemap_fdatawait()
54 int filemap_write_and_wait_range(struct address_space *mapping,
56 int __filemap_fdatawrite_range(struct address_space *mapping,
[all …]
Dio-mapping.h58 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument
60 iomap_free(mapping->base, mapping->size); in io_mapping_fini()
65 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument
70 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc()
71 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc()
77 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_atomic_wc()
92 io_mapping_map_local_wc(struct io_mapping *mapping, unsigned long offset) in io_mapping_map_local_wc() argument
96 BUG_ON(offset >= mapping->size); in io_mapping_map_local_wc()
97 phys_addr = mapping->base + offset; in io_mapping_map_local_wc()
98 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_local_wc()
[all …]
Dcleancache.h52 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument
54 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping()
58 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled()
104 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument
108 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page()
109 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page()
112 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument
114 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode()
115 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
Dtpm_eventlog.h166 void *mapping = NULL; in __calc_tpm2_event_size() local
186 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size()
188 if (!mapping) { in __calc_tpm2_event_size()
193 mapping = marker_start; in __calc_tpm2_event_size()
196 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size()
233 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size()
235 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size()
237 if (!mapping) { in __calc_tpm2_event_size()
242 mapping = marker; in __calc_tpm2_event_size()
245 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size()
[all …]
Dshmem_fs.h103 bool shmem_mapping(struct address_space *mapping);
105 static inline bool shmem_mapping(struct address_space *mapping) in shmem_mapping() argument
110 extern void shmem_unlock_mapping(struct address_space *mapping);
111 extern struct page *shmem_read_mapping_page_gfp(struct address_space *mapping,
137 extern unsigned long shmem_partial_swap_usage(struct address_space *mapping,
151 struct folio *shmem_read_folio_gfp(struct address_space *mapping,
154 static inline struct folio *shmem_read_folio(struct address_space *mapping, in shmem_read_folio() argument
157 return shmem_read_folio_gfp(mapping, index, mapping_gfp_mask(mapping)); in shmem_read_folio()
161 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page() argument
163 return shmem_read_mapping_page_gfp(mapping, index, in shmem_read_mapping_page()
[all …]
Dsecretmem.h9 static inline bool secretmem_mapping(struct address_space *mapping) in secretmem_mapping() argument
11 return mapping->a_ops == &secretmem_aops; in secretmem_mapping()
24 static inline bool secretmem_mapping(struct address_space *mapping) in secretmem_mapping() argument
Ddax.h158 int dax_writeback_mapping_range(struct address_space *mapping,
161 struct page *dax_layout_busy_page(struct address_space *mapping);
162 struct page *dax_layout_busy_page_range(struct address_space *mapping, loff_t start, loff_t end);
165 dax_entry_t dax_lock_mapping_entry(struct address_space *mapping,
167 void dax_unlock_mapping_entry(struct address_space *mapping,
170 static inline struct page *dax_layout_busy_page(struct address_space *mapping) in dax_layout_busy_page() argument
175 static inline struct page *dax_layout_busy_page_range(struct address_space *mapping, pgoff_t start,… in dax_layout_busy_page_range() argument
180 static inline int dax_writeback_mapping_range(struct address_space *mapping, in dax_writeback_mapping_range() argument
188 if (IS_DAX(folio->mapping->host)) in dax_lock_folio()
197 static inline dax_entry_t dax_lock_mapping_entry(struct address_space *mapping, in dax_lock_mapping_entry() argument
[all …]
Dmigrate.h66 int migrate_folio(struct address_space *mapping, struct folio *dst,
75 int migrate_huge_page_move_mapping(struct address_space *mapping,
80 int folio_migrate_mapping(struct address_space *mapping,
98 static inline int migrate_huge_page_move_mapping(struct address_space *mapping, in migrate_huge_page_move_mapping() argument
132 ((unsigned long)folio->mapping - PAGE_MAPPING_MOVABLE); in folio_movable_ops()
141 ((unsigned long)page->mapping - PAGE_MAPPING_MOVABLE); in page_movable_ops()
Dwriteback.h369 void balance_dirty_pages_ratelimited(struct address_space *mapping);
370 int balance_dirty_pages_ratelimited_flags(struct address_space *mapping,
375 struct folio *writeback_iter(struct address_space *mapping,
381 int write_cache_pages(struct address_space *mapping,
384 int do_writepages(struct address_space *mapping, struct writeback_control *wbc);
386 void tag_pages_for_writeback(struct address_space *mapping,
389 bool filemap_dirty_folio(struct address_space *mapping, struct folio *folio);
Dmpage.h20 int mpage_writepages(struct address_space *mapping,
Dfs.h411 int (*write_begin)(struct file *, struct address_space *mapping,
414 int (*write_end)(struct file *, struct address_space *mapping,
502 static inline bool mapping_tagged(struct address_space *mapping, xa_mark_t tag) in mapping_tagged() argument
504 return xa_marked(&mapping->i_pages, tag); in mapping_tagged()
507 static inline void i_mmap_lock_write(struct address_space *mapping) in i_mmap_lock_write() argument
509 down_write(&mapping->i_mmap_rwsem); in i_mmap_lock_write()
512 static inline int i_mmap_trylock_write(struct address_space *mapping) in i_mmap_trylock_write() argument
514 return down_write_trylock(&mapping->i_mmap_rwsem); in i_mmap_trylock_write()
517 static inline void i_mmap_unlock_write(struct address_space *mapping) in i_mmap_unlock_write() argument
519 up_write(&mapping->i_mmap_rwsem); in i_mmap_unlock_write()
[all …]
Dfscache.h176 struct address_space *mapping,
586 static inline void fscache_clear_page_bits(struct address_space *mapping, in fscache_clear_page_bits() argument
591 __fscache_clear_page_bits(mapping, start, len); in fscache_clear_page_bits()
620 struct address_space *mapping, in fscache_write_to_cache() argument
627 __fscache_write_to_cache(cookie, mapping, start, len, i_size, in fscache_write_to_cache()
Dnfs_page.h209 return req->wb_page->mapping->host; in nfs_page_to_inode()
210 return folio->mapping->host; in nfs_page_to_inode()
Diomap.h306 bool iomap_dirty_folio(struct address_space *mapping, struct folio *folio);
328 sector_t iomap_bmap(struct address_space *mapping, sector_t bno,
391 int iomap_writepages(struct address_space *mapping,
Duacce.h101 struct address_space *mapping; member
Dbuffer_head.h261 int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len,
514 bool block_dirty_folio(struct address_space *mapping, struct folio *folio);
523 int sync_mapping_buffers(struct address_space *mapping);
536 static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } in sync_mapping_buffers() argument
Dnfs_fs.h411 extern int nfs_sync_mapping(struct address_space *mapping);
412 extern void nfs_zap_mapping(struct inode *inode, struct address_space *mapping);
432 extern int nfs_clear_invalid_mapping(struct address_space *mapping);
434 extern int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping);
/include/trace/events/
Dfilemap.h32 __entry->i_ino = folio->mapping->host->i_ino;
34 if (folio->mapping->host->i_sb)
35 __entry->s_dev = folio->mapping->host->i_sb->s_dev;
37 __entry->s_dev = folio->mapping->host->i_rdev;
62 struct address_space *mapping,
67 TP_ARGS(mapping, index, last_index),
77 __entry->i_ino = mapping->host->i_ino;
78 if (mapping->host->i_sb)
80 mapping->host->i_sb->s_dev;
82 __entry->s_dev = mapping->host->i_rdev;
[all …]
Dpage_ref.h24 __field(void *, mapping)
34 __entry->mapping = page->mapping;
43 __entry->mapcount, __entry->mapping, __entry->mt,
72 __field(void *, mapping)
83 __entry->mapping = page->mapping;
93 __entry->mapcount, __entry->mapping, __entry->mt,
Dwriteback.h65 TP_PROTO(struct folio *folio, struct address_space *mapping),
67 TP_ARGS(folio, mapping),
77 bdi_dev_name(mapping ? inode_to_bdi(mapping->host) :
79 __entry->ino = (mapping && mapping->host) ? mapping->host->i_ino : 0;
92 TP_PROTO(struct folio *folio, struct address_space *mapping),
94 TP_ARGS(folio, mapping)
99 TP_PROTO(struct folio *folio, struct address_space *mapping),
101 TP_ARGS(folio, mapping)
261 struct address_space *mapping = folio_mapping(folio);
262 struct inode *inode = mapping ? mapping->host : NULL;
Derofs.h99 __entry->dev = folio->mapping->host->i_sb->s_dev;
100 __entry->nid = EROFS_I(folio->mapping->host)->nid;
101 __entry->dir = S_ISDIR(folio->mapping->host->i_mode);
/include/asm-generic/
Dcacheflush.h59 static inline void flush_dcache_mmap_lock(struct address_space *mapping) in flush_dcache_mmap_lock() argument
65 static inline void flush_dcache_mmap_unlock(struct address_space *mapping) in flush_dcache_mmap_unlock() argument
/include/trace/hooks/
Dmm.h24 TP_PROTO(struct address_space *mapping, long nr_pages),
25 TP_ARGS(mapping, nr_pages));
27 TP_PROTO(struct address_space *mapping, long nr_pages),
28 TP_ARGS(mapping, nr_pages));
80 TP_PROTO(struct address_space *mapping, unsigned int index,
82 TP_ARGS(mapping, index, nr_page, read, direct));
149 TP_PROTO(struct address_space *mapping, pgoff_t index,
151 TP_ARGS(mapping, index, fgp_flags, gfp_mask, folio));
153 TP_PROTO(struct address_space *mapping, int fgp_flags, gfp_t *gfp_mask),
154 TP_ARGS(mapping, fgp_flags, gfp_mask));
[all …]
Dvmscan.h32 TP_PROTO(struct address_space *mapping, bool *skip),
33 TP_ARGS(mapping, skip));
/include/uapi/drm/
Dtegra_drm.h759 __u32 mapping; member
775 __u32 mapping; member
793 __u32 mapping; member

12