/include/linux/ |
D | pagemap.h | 22 static inline bool mapping_empty(struct address_space *mapping) in mapping_empty() argument 24 return xa_empty(&mapping->i_pages); in mapping_empty() 55 static inline void mapping_set_error(struct address_space *mapping, int error) in mapping_set_error() argument 61 __filemap_set_wb_err(mapping, error); in mapping_set_error() 64 if (mapping->host) in mapping_set_error() 65 errseq_set(&mapping->host->i_sb->s_wb_err, error); in mapping_set_error() 69 set_bit(AS_ENOSPC, &mapping->flags); in mapping_set_error() 71 set_bit(AS_EIO, &mapping->flags); in mapping_set_error() 74 static inline void mapping_set_unevictable(struct address_space *mapping) in mapping_set_unevictable() argument 76 set_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_set_unevictable() [all …]
|
D | io-mapping.h | 58 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument 60 iomap_free(mapping->base, mapping->size); in io_mapping_fini() 65 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument 70 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc() 71 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc() 74 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_atomic_wc() 86 io_mapping_map_local_wc(struct io_mapping *mapping, unsigned long offset) in io_mapping_map_local_wc() argument 90 BUG_ON(offset >= mapping->size); in io_mapping_map_local_wc() 91 phys_addr = mapping->base + offset; in io_mapping_map_local_wc() 92 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_local_wc() [all …]
|
D | secretmem.h | 11 struct address_space *mapping; in page_is_secretmem() local 23 mapping = (struct address_space *) in page_is_secretmem() 24 ((unsigned long)page->mapping & ~PAGE_MAPPING_FLAGS); in page_is_secretmem() 26 if (!mapping || mapping != page->mapping) in page_is_secretmem() 29 return mapping->a_ops == &secretmem_aops; in page_is_secretmem()
|
D | cleancache.h | 54 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument 56 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping() 60 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled() 106 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument 110 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page() 111 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page() 114 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument 116 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode() 117 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
|
D | tpm_eventlog.h | 166 void *mapping = NULL; in __calc_tpm2_event_size() local 186 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size() 188 if (!mapping) { in __calc_tpm2_event_size() 193 mapping = marker_start; in __calc_tpm2_event_size() 196 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size() 233 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size() 235 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size() 237 if (!mapping) { in __calc_tpm2_event_size() 242 mapping = marker; in __calc_tpm2_event_size() 245 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size() [all …]
|
D | pagevec.h | 30 struct address_space *mapping, 33 struct address_space *mapping, in pagevec_lookup() argument 36 return pagevec_lookup_range(pvec, mapping, start, (pgoff_t)-1); in pagevec_lookup() 40 struct address_space *mapping, pgoff_t *index, pgoff_t end, 43 struct address_space *mapping, pgoff_t *index, xa_mark_t tag) in pagevec_lookup_tag() argument 45 return pagevec_lookup_range_tag(pvec, mapping, index, (pgoff_t)-1, tag); in pagevec_lookup_tag()
|
D | shmem_fs.h | 72 static inline bool shmem_mapping(struct address_space *mapping) in shmem_mapping() argument 74 return mapping->a_ops == &shmem_aops; in shmem_mapping() 77 static inline bool shmem_mapping(struct address_space *mapping) in shmem_mapping() argument 82 extern void shmem_unlock_mapping(struct address_space *mapping); 83 extern struct page *shmem_read_mapping_page_gfp(struct address_space *mapping, 96 extern unsigned long shmem_partial_swap_usage(struct address_space *mapping, 112 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page() argument 114 return shmem_read_mapping_page_gfp(mapping, index, in shmem_read_mapping_page() 115 mapping_gfp_mask(mapping)); in shmem_read_mapping_page()
|
D | dax.h | 125 int dax_writeback_mapping_range(struct address_space *mapping, 128 struct page *dax_layout_busy_page(struct address_space *mapping); 129 struct page *dax_layout_busy_page_range(struct address_space *mapping, loff_t start, loff_t end); 151 static inline struct page *dax_layout_busy_page(struct address_space *mapping) in dax_layout_busy_page() argument 156 static inline struct page *dax_layout_busy_page_range(struct address_space *mapping, pgoff_t start,… in dax_layout_busy_page_range() argument 161 static inline int dax_writeback_mapping_range(struct address_space *mapping, in dax_writeback_mapping_range() argument 169 if (IS_DAX(page->mapping->host)) in dax_lock_page() 210 int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index); 211 int dax_invalidate_mapping_entry_sync(struct address_space *mapping, 214 static inline bool dax_mapping(struct address_space *mapping) in dax_mapping() argument [all …]
|
D | migrate.h | 45 extern int migrate_page(struct address_space *mapping, 56 extern int migrate_huge_page_move_mapping(struct address_space *mapping, 58 extern int migrate_page_move_mapping(struct address_space *mapping, 80 static inline int migrate_huge_page_move_mapping(struct address_space *mapping, in migrate_huge_page_move_mapping() argument 89 extern void __SetPageMovable(struct page *page, struct address_space *mapping); 94 struct address_space *mapping) in __SetPageMovable() argument
|
D | fs.h | 392 int (*readpages)(struct file *filp, struct address_space *mapping, 396 int (*write_begin)(struct file *, struct address_space *mapping, 399 int (*write_end)(struct file *, struct address_space *mapping, 440 int pagecache_write_begin(struct file *, struct address_space *mapping, 444 int pagecache_write_end(struct file *, struct address_space *mapping, 510 static inline bool mapping_tagged(struct address_space *mapping, xa_mark_t tag) in mapping_tagged() argument 512 return xa_marked(&mapping->i_pages, tag); in mapping_tagged() 515 static inline void i_mmap_lock_write(struct address_space *mapping) in i_mmap_lock_write() argument 517 down_write(&mapping->i_mmap_rwsem); in i_mmap_lock_write() 520 static inline int i_mmap_trylock_write(struct address_space *mapping) in i_mmap_trylock_write() argument [all …]
|
D | writeback.h | 378 void balance_dirty_pages_ratelimited(struct address_space *mapping); 384 int generic_writepages(struct address_space *mapping, 386 void tag_pages_for_writeback(struct address_space *mapping, 388 int write_cache_pages(struct address_space *mapping, 391 int do_writepages(struct address_space *mapping, struct writeback_control *wbc); 393 void tag_pages_for_writeback(struct address_space *mapping,
|
D | mpage.h | 20 int mpage_writepages(struct address_space *mapping,
|
D | swap.h | 338 #define mapping_set_update(xas, mapping) do { \ argument 339 if (!dax_mapping(mapping) && !shmem_mapping(mapping)) \ 391 extern int remove_mapping(struct address_space *mapping, struct page *page); 459 struct page *find_get_incore_page(struct address_space *mapping, pgoff_t index); 621 struct page *find_get_incore_page(struct address_space *mapping, pgoff_t index) in find_get_incore_page() argument 623 return find_get_page(mapping, index); in find_get_incore_page()
|
D | vlynq.h | 143 struct vlynq_mapping *mapping); 145 struct vlynq_mapping *mapping);
|
D | netfs.h | 136 struct address_space *mapping; /* The mapping being accessed */ member 174 void (*cleanup)(struct address_space *mapping, void *netfs_priv);
|
D | iomap.h | 237 int iomap_migrate_page(struct address_space *mapping, struct page *newpage, 256 sector_t iomap_bmap(struct address_space *mapping, sector_t bno, 309 int iomap_writepages(struct address_space *mapping,
|
D | pagewalk.h | 104 int walk_page_mapping(struct address_space *mapping, pgoff_t first_index,
|
D | fscache.h | 103 struct address_space *mapping, 620 struct address_space *mapping, in fscache_read_or_alloc_pages() argument 628 return __fscache_read_or_alloc_pages(cookie, mapping, pages, in fscache_read_or_alloc_pages()
|
D | nfs_fs.h | 386 extern int nfs_sync_mapping(struct address_space *mapping); 387 extern void nfs_zap_mapping(struct inode *inode, struct address_space *mapping); 407 extern int nfs_clear_invalid_mapping(struct address_space *mapping); 409 extern int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping);
|
/include/trace/events/ |
D | filemap.h | 31 __entry->i_ino = page->mapping->host->i_ino; 33 if (page->mapping->host->i_sb) 34 __entry->s_dev = page->mapping->host->i_sb->s_dev; 36 __entry->s_dev = page->mapping->host->i_rdev; 58 TP_PROTO(struct address_space *mapping, errseq_t eseq), 60 TP_ARGS(mapping, eseq), 69 __entry->i_ino = mapping->host->i_ino; 71 if (mapping->host->i_sb) 72 __entry->s_dev = mapping->host->i_sb->s_dev; 74 __entry->s_dev = mapping->host->i_rdev;
|
D | page_ref.h | 24 __field(void *, mapping) 34 __entry->mapping = page->mapping; 43 __entry->mapcount, __entry->mapping, __entry->mt, 72 __field(void *, mapping) 83 __entry->mapping = page->mapping; 93 __entry->mapcount, __entry->mapping, __entry->mt,
|
D | writeback.h | 57 TP_PROTO(struct page *page, struct address_space *mapping), 59 TP_ARGS(page, mapping), 69 bdi_dev_name(mapping ? inode_to_bdi(mapping->host) : 71 __entry->ino = (mapping && mapping->host) ? mapping->host->i_ino : 0; 84 TP_PROTO(struct page *page, struct address_space *mapping), 86 TP_ARGS(page, mapping) 91 TP_PROTO(struct page *page, struct address_space *mapping), 93 TP_ARGS(page, mapping) 253 struct address_space *mapping = page_mapping(page); 254 struct inode *inode = mapping ? mapping->host : NULL;
|
D | erofs.h | 97 __entry->dev = page->mapping->host->i_sb->s_dev; 98 __entry->nid = EROFS_I(page->mapping->host)->nid; 99 __entry->dir = S_ISDIR(page->mapping->host->i_mode);
|
/include/asm-generic/ |
D | cacheflush.h | 57 static inline void flush_dcache_mmap_lock(struct address_space *mapping) in flush_dcache_mmap_lock() argument 63 static inline void flush_dcache_mmap_unlock(struct address_space *mapping) in flush_dcache_mmap_unlock() argument
|
/include/uapi/drm/ |
D | tegra_drm.h | 759 __u32 mapping; member 775 __u32 mapping; member 793 __u32 mapping; member
|