| /include/linux/ |
| D | pagemap.h | 21 unsigned long invalidate_mapping_pages(struct address_space *mapping, 30 int invalidate_inode_pages2(struct address_space *mapping); 31 int invalidate_inode_pages2_range(struct address_space *mapping, 35 int filemap_invalidate_pages(struct address_space *mapping, 41 int filemap_fdatawait_keep_errors(struct address_space *mapping); 43 int filemap_fdatawait_range_keep_errors(struct address_space *mapping, 48 static inline int filemap_fdatawait(struct address_space *mapping) in filemap_fdatawait() argument 50 return filemap_fdatawait_range(mapping, 0, LLONG_MAX); in filemap_fdatawait() 54 int filemap_write_and_wait_range(struct address_space *mapping, 56 int __filemap_fdatawrite_range(struct address_space *mapping, [all …]
|
| D | io-mapping.h | 58 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument 60 iomap_free(mapping->base, mapping->size); in io_mapping_fini() 65 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument 70 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc() 71 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc() 77 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_atomic_wc() 92 io_mapping_map_local_wc(struct io_mapping *mapping, unsigned long offset) in io_mapping_map_local_wc() argument 96 BUG_ON(offset >= mapping->size); in io_mapping_map_local_wc() 97 phys_addr = mapping->base + offset; in io_mapping_map_local_wc() 98 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_local_wc() [all …]
|
| D | cleancache.h | 52 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument 54 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping() 58 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled() 104 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument 108 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page() 109 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page() 112 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument 114 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode() 115 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
|
| D | tpm_eventlog.h | 166 void *mapping = NULL; in __calc_tpm2_event_size() local 186 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size() 188 if (!mapping) { in __calc_tpm2_event_size() 193 mapping = marker_start; in __calc_tpm2_event_size() 196 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size() 233 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size() 235 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size() 237 if (!mapping) { in __calc_tpm2_event_size() 242 mapping = marker; in __calc_tpm2_event_size() 245 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size() [all …]
|
| D | shmem_fs.h | 103 bool shmem_mapping(struct address_space *mapping); 105 static inline bool shmem_mapping(struct address_space *mapping) in shmem_mapping() argument 110 extern void shmem_unlock_mapping(struct address_space *mapping); 111 extern struct page *shmem_read_mapping_page_gfp(struct address_space *mapping, 137 extern unsigned long shmem_partial_swap_usage(struct address_space *mapping, 151 struct folio *shmem_read_folio_gfp(struct address_space *mapping, 154 static inline struct folio *shmem_read_folio(struct address_space *mapping, in shmem_read_folio() argument 157 return shmem_read_folio_gfp(mapping, index, mapping_gfp_mask(mapping)); in shmem_read_folio() 161 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page() argument 163 return shmem_read_mapping_page_gfp(mapping, index, in shmem_read_mapping_page() [all …]
|
| D | secretmem.h | 9 static inline bool secretmem_mapping(struct address_space *mapping) in secretmem_mapping() argument 11 return mapping->a_ops == &secretmem_aops; in secretmem_mapping() 24 static inline bool secretmem_mapping(struct address_space *mapping) in secretmem_mapping() argument
|
| D | dax.h | 158 int dax_writeback_mapping_range(struct address_space *mapping, 161 struct page *dax_layout_busy_page(struct address_space *mapping); 162 struct page *dax_layout_busy_page_range(struct address_space *mapping, loff_t start, loff_t end); 165 dax_entry_t dax_lock_mapping_entry(struct address_space *mapping, 167 void dax_unlock_mapping_entry(struct address_space *mapping, 170 static inline struct page *dax_layout_busy_page(struct address_space *mapping) in dax_layout_busy_page() argument 175 static inline struct page *dax_layout_busy_page_range(struct address_space *mapping, pgoff_t start,… in dax_layout_busy_page_range() argument 180 static inline int dax_writeback_mapping_range(struct address_space *mapping, in dax_writeback_mapping_range() argument 188 if (IS_DAX(folio->mapping->host)) in dax_lock_folio() 197 static inline dax_entry_t dax_lock_mapping_entry(struct address_space *mapping, in dax_lock_mapping_entry() argument [all …]
|
| D | migrate.h | 66 int migrate_folio(struct address_space *mapping, struct folio *dst, 75 int migrate_huge_page_move_mapping(struct address_space *mapping, 80 int folio_migrate_mapping(struct address_space *mapping, 98 static inline int migrate_huge_page_move_mapping(struct address_space *mapping, in migrate_huge_page_move_mapping() argument 132 ((unsigned long)folio->mapping - PAGE_MAPPING_MOVABLE); in folio_movable_ops() 141 ((unsigned long)page->mapping - PAGE_MAPPING_MOVABLE); in page_movable_ops()
|
| D | writeback.h | 369 void balance_dirty_pages_ratelimited(struct address_space *mapping); 370 int balance_dirty_pages_ratelimited_flags(struct address_space *mapping, 375 struct folio *writeback_iter(struct address_space *mapping, 381 int write_cache_pages(struct address_space *mapping, 384 int do_writepages(struct address_space *mapping, struct writeback_control *wbc); 386 void tag_pages_for_writeback(struct address_space *mapping, 389 bool filemap_dirty_folio(struct address_space *mapping, struct folio *folio);
|
| D | mpage.h | 20 int mpage_writepages(struct address_space *mapping,
|
| D | fs.h | 411 int (*write_begin)(struct file *, struct address_space *mapping, 414 int (*write_end)(struct file *, struct address_space *mapping, 502 static inline bool mapping_tagged(struct address_space *mapping, xa_mark_t tag) in mapping_tagged() argument 504 return xa_marked(&mapping->i_pages, tag); in mapping_tagged() 507 static inline void i_mmap_lock_write(struct address_space *mapping) in i_mmap_lock_write() argument 509 down_write(&mapping->i_mmap_rwsem); in i_mmap_lock_write() 512 static inline int i_mmap_trylock_write(struct address_space *mapping) in i_mmap_trylock_write() argument 514 return down_write_trylock(&mapping->i_mmap_rwsem); in i_mmap_trylock_write() 517 static inline void i_mmap_unlock_write(struct address_space *mapping) in i_mmap_unlock_write() argument 519 up_write(&mapping->i_mmap_rwsem); in i_mmap_unlock_write() [all …]
|
| D | fscache.h | 176 struct address_space *mapping, 586 static inline void fscache_clear_page_bits(struct address_space *mapping, in fscache_clear_page_bits() argument 591 __fscache_clear_page_bits(mapping, start, len); in fscache_clear_page_bits() 620 struct address_space *mapping, in fscache_write_to_cache() argument 627 __fscache_write_to_cache(cookie, mapping, start, len, i_size, in fscache_write_to_cache()
|
| D | nfs_page.h | 209 return req->wb_page->mapping->host; in nfs_page_to_inode() 210 return folio->mapping->host; in nfs_page_to_inode()
|
| D | iomap.h | 306 bool iomap_dirty_folio(struct address_space *mapping, struct folio *folio); 328 sector_t iomap_bmap(struct address_space *mapping, sector_t bno, 391 int iomap_writepages(struct address_space *mapping,
|
| D | uacce.h | 101 struct address_space *mapping; member
|
| D | buffer_head.h | 261 int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len, 514 bool block_dirty_folio(struct address_space *mapping, struct folio *folio); 523 int sync_mapping_buffers(struct address_space *mapping); 536 static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } in sync_mapping_buffers() argument
|
| D | nfs_fs.h | 411 extern int nfs_sync_mapping(struct address_space *mapping); 412 extern void nfs_zap_mapping(struct inode *inode, struct address_space *mapping); 432 extern int nfs_clear_invalid_mapping(struct address_space *mapping); 434 extern int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping);
|
| /include/trace/events/ |
| D | filemap.h | 32 __entry->i_ino = folio->mapping->host->i_ino; 34 if (folio->mapping->host->i_sb) 35 __entry->s_dev = folio->mapping->host->i_sb->s_dev; 37 __entry->s_dev = folio->mapping->host->i_rdev; 62 struct address_space *mapping, 67 TP_ARGS(mapping, index, last_index), 77 __entry->i_ino = mapping->host->i_ino; 78 if (mapping->host->i_sb) 80 mapping->host->i_sb->s_dev; 82 __entry->s_dev = mapping->host->i_rdev; [all …]
|
| D | page_ref.h | 24 __field(void *, mapping) 34 __entry->mapping = page->mapping; 43 __entry->mapcount, __entry->mapping, __entry->mt, 72 __field(void *, mapping) 83 __entry->mapping = page->mapping; 93 __entry->mapcount, __entry->mapping, __entry->mt,
|
| D | writeback.h | 65 TP_PROTO(struct folio *folio, struct address_space *mapping), 67 TP_ARGS(folio, mapping), 77 bdi_dev_name(mapping ? inode_to_bdi(mapping->host) : 79 __entry->ino = (mapping && mapping->host) ? mapping->host->i_ino : 0; 92 TP_PROTO(struct folio *folio, struct address_space *mapping), 94 TP_ARGS(folio, mapping) 99 TP_PROTO(struct folio *folio, struct address_space *mapping), 101 TP_ARGS(folio, mapping) 261 struct address_space *mapping = folio_mapping(folio); 262 struct inode *inode = mapping ? mapping->host : NULL;
|
| D | erofs.h | 99 __entry->dev = folio->mapping->host->i_sb->s_dev; 100 __entry->nid = EROFS_I(folio->mapping->host)->nid; 101 __entry->dir = S_ISDIR(folio->mapping->host->i_mode);
|
| /include/asm-generic/ |
| D | cacheflush.h | 59 static inline void flush_dcache_mmap_lock(struct address_space *mapping) in flush_dcache_mmap_lock() argument 65 static inline void flush_dcache_mmap_unlock(struct address_space *mapping) in flush_dcache_mmap_unlock() argument
|
| /include/trace/hooks/ |
| D | mm.h | 24 TP_PROTO(struct address_space *mapping, long nr_pages), 25 TP_ARGS(mapping, nr_pages)); 27 TP_PROTO(struct address_space *mapping, long nr_pages), 28 TP_ARGS(mapping, nr_pages)); 80 TP_PROTO(struct address_space *mapping, unsigned int index, 82 TP_ARGS(mapping, index, nr_page, read, direct)); 149 TP_PROTO(struct address_space *mapping, pgoff_t index, 151 TP_ARGS(mapping, index, fgp_flags, gfp_mask, folio)); 153 TP_PROTO(struct address_space *mapping, int fgp_flags, gfp_t *gfp_mask), 154 TP_ARGS(mapping, fgp_flags, gfp_mask)); [all …]
|
| D | vmscan.h | 32 TP_PROTO(struct address_space *mapping, bool *skip), 33 TP_ARGS(mapping, skip));
|
| /include/uapi/drm/ |
| D | tegra_drm.h | 759 __u32 mapping; member 775 __u32 mapping; member 793 __u32 mapping; member
|