/include/linux/ |
D | page_ref.h | 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *page, int v, int u); 34 extern void __page_ref_freeze(struct page *page, int v, int ret); 35 extern void __page_ref_unfreeze(struct page *page, int v); 41 static inline void __page_ref_set(struct page *page, int v) in __page_ref_set() argument 44 static inline void __page_ref_mod(struct page *page, int v) in __page_ref_mod() argument 47 static inline void __page_ref_mod_and_test(struct page *page, int v, int ret) in __page_ref_mod_and_test() argument [all …]
|
D | page_pinner.h | 12 extern void __reset_page_pinner(struct page *page, unsigned int order, bool free); 13 extern void __set_page_pinner(struct page *page, unsigned int order); 14 extern void __dump_page_pinner(struct page *page); 15 void __page_pinner_migration_failed(struct page *page); 18 static inline void reset_page_pinner(struct page *page, unsigned int order) in reset_page_pinner() argument 21 __reset_page_pinner(page, order, false); in reset_page_pinner() 24 static inline void free_page_pinner(struct page *page, unsigned int order) in free_page_pinner() argument 27 __reset_page_pinner(page, order, true); in free_page_pinner() 30 static inline void set_page_pinner(struct page *page, unsigned int order) in set_page_pinner() argument 33 __set_page_pinner(page, order); in set_page_pinner() [all …]
|
D | page_idle.h | 12 static inline bool page_is_young(struct page *page) in page_is_young() argument 14 return PageYoung(page); in page_is_young() 17 static inline void set_page_young(struct page *page) in set_page_young() argument 19 SetPageYoung(page); in set_page_young() 22 static inline bool test_and_clear_page_young(struct page *page) in test_and_clear_page_young() argument 24 return TestClearPageYoung(page); in test_and_clear_page_young() 27 static inline bool page_is_idle(struct page *page) in page_is_idle() argument 29 return PageIdle(page); in page_is_idle() 32 static inline void set_page_idle(struct page *page) in set_page_idle() argument 34 SetPageIdle(page); in set_page_idle() [all …]
|
D | balloon_compaction.h | 58 int (*migratepage)(struct balloon_dev_info *, struct page *newpage, 59 struct page *page, enum migrate_mode mode); 63 extern struct page *balloon_page_alloc(void); 65 struct page *page); 66 extern struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info); 83 extern bool balloon_page_isolate(struct page *page, 85 extern void balloon_page_putback(struct page *page); 87 struct page *newpage, 88 struct page *page, enum migrate_mode mode); 100 struct page *page) in balloon_page_insert() argument [all …]
|
D | page_owner.h | 15 extern void __reset_page_owner(struct page *page, unsigned int order); 16 extern void __set_page_owner(struct page *page, 18 extern void __split_page_owner(struct page *page, unsigned int nr); 19 extern void __copy_page_owner(struct page *oldpage, struct page *newpage); 20 extern void __set_page_owner_migrate_reason(struct page *page, int reason); 21 extern void __dump_page_owner(struct page *page); 25 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 28 __reset_page_owner(page, order); in reset_page_owner() 31 static inline void set_page_owner(struct page *page, in set_page_owner() argument 35 __set_page_owner(page, order, gfp_mask); in set_page_owner() [all …]
|
D | migrate.h | 10 typedef struct page *new_page_t(struct page *page, unsigned long private); 11 typedef void free_page_t(struct page *page, unsigned long private); 40 struct page *newpage, struct page *page, 44 extern struct page *alloc_migration_target(struct page *page, unsigned long private); 45 extern int isolate_movable_page(struct page *page, isolate_mode_t mode); 46 extern void putback_movable_page(struct page *page); 48 extern void migrate_page_states(struct page *newpage, struct page *page); 49 extern void migrate_page_copy(struct page *newpage, struct page *page); 51 struct page *newpage, struct page *page); 53 struct page *newpage, struct page *page, int extra_count); [all …]
|
D | page-flags.h | 184 struct page; /* forward declaration */ 186 static inline struct page *compound_head(struct page *page) in compound_head() argument 188 unsigned long head = READ_ONCE(page->compound_head); in compound_head() 191 return (struct page *) (head - 1); in compound_head() 192 return page; in compound_head() 195 static __always_inline int PageTail(struct page *page) in PageTail() argument 197 return READ_ONCE(page->compound_head) & 1; in PageTail() 200 static __always_inline int PageCompound(struct page *page) in PageCompound() argument 202 return test_bit(PG_head, &page->flags) || PageTail(page); in PageCompound() 206 static inline int PagePoisoned(const struct page *page) in PagePoisoned() argument [all …]
|
D | highmem.h | 15 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vma… in flush_anon_page() argument 21 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument 35 extern void *kmap_atomic_high_prot(struct page *page, pgprot_t prot); 47 void *kmap_high(struct page *page); 48 static inline void *kmap(struct page *page) in kmap() argument 53 if (!PageHighMem(page)) in kmap() 54 addr = page_address(page); in kmap() 56 addr = kmap_high(page); in kmap() 61 void kunmap_high(struct page *page); 63 static inline void kunmap(struct page *page) in kunmap() argument [all …]
|
D | hugetlb_cgroup.h | 62 __hugetlb_cgroup_from_page(struct page *page, bool rsvd) in __hugetlb_cgroup_from_page() argument 64 VM_BUG_ON_PAGE(!PageHuge(page), page); in __hugetlb_cgroup_from_page() 66 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in __hugetlb_cgroup_from_page() 69 return (struct hugetlb_cgroup *)page[3].private; in __hugetlb_cgroup_from_page() 71 return (struct hugetlb_cgroup *)page[2].private; in __hugetlb_cgroup_from_page() 74 static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) in hugetlb_cgroup_from_page() argument 76 return __hugetlb_cgroup_from_page(page, false); in hugetlb_cgroup_from_page() 80 hugetlb_cgroup_from_page_rsvd(struct page *page) in hugetlb_cgroup_from_page_rsvd() argument 82 return __hugetlb_cgroup_from_page(page, true); in hugetlb_cgroup_from_page_rsvd() 85 static inline int __set_hugetlb_cgroup(struct page *page, in __set_hugetlb_cgroup() argument [all …]
|
D | pagemap.h | 159 void release_pages(struct page **pages, int nr); 205 static inline int __page_cache_add_speculative(struct page *page, int count) in __page_cache_add_speculative() argument 220 VM_BUG_ON_PAGE(page_count(page) == 0, page); in __page_cache_add_speculative() 221 page_ref_add(page, count); in __page_cache_add_speculative() 224 if (unlikely(!page_ref_add_unless(page, count, 0))) { in __page_cache_add_speculative() 233 VM_BUG_ON_PAGE(PageTail(page), page); in __page_cache_add_speculative() 238 static inline int page_cache_get_speculative(struct page *page) in page_cache_get_speculative() argument 240 return __page_cache_add_speculative(page, 1); in page_cache_get_speculative() 243 static inline int page_cache_add_speculative(struct page *page, int count) in page_cache_add_speculative() argument 245 return __page_cache_add_speculative(page, count); in page_cache_add_speculative() [all …]
|
D | pageblock-flags.h | 55 struct page; 57 unsigned long get_pfnblock_flags_mask(struct page *page, 61 void set_pfnblock_flags_mask(struct page *page, 68 #define get_pageblock_skip(page) \ argument 69 get_pfnblock_flags_mask(page, page_to_pfn(page), \ 71 #define clear_pageblock_skip(page) \ argument 72 set_pfnblock_flags_mask(page, 0, page_to_pfn(page), \ 74 #define set_pageblock_skip(page) \ argument 75 set_pfnblock_flags_mask(page, (1 << PB_migrate_skip), \ 76 page_to_pfn(page), \ [all …]
|
D | mm_inline.h | 26 static inline int page_is_file_lru(struct page *page) in page_is_file_lru() argument 28 return !PageSwapBacked(page); in page_is_file_lru() 52 static __always_inline void add_page_to_lru_list(struct page *page, in add_page_to_lru_list() argument 55 trace_android_vh_add_page_to_lrulist(page, false, lru); in add_page_to_lru_list() 56 update_lru_size(lruvec, lru, page_zonenum(page), thp_nr_pages(page)); in add_page_to_lru_list() 57 list_add(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list() 60 static __always_inline void add_page_to_lru_list_tail(struct page *page, in add_page_to_lru_list_tail() argument 63 trace_android_vh_add_page_to_lrulist(page, false, lru); in add_page_to_lru_list_tail() 64 update_lru_size(lruvec, lru, page_zonenum(page), thp_nr_pages(page)); in add_page_to_lru_list_tail() 65 list_add_tail(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list_tail() [all …]
|
D | swap.h | 317 void *workingset_eviction(struct page *page, struct mem_cgroup *target_memcg); 318 void workingset_refault(struct page *page, void *shadow); 319 void workingset_activation(struct page *page); 339 extern void lru_note_cost_page(struct page *); 340 extern void lru_cache_add(struct page *); 341 extern void lru_add_page_tail(struct page *page, struct page *page_tail, 343 extern void mark_page_accessed(struct page *); 352 extern void rotate_reclaimable_page(struct page *page); 353 extern void deactivate_file_page(struct page *page); 354 extern void deactivate_page(struct page *page); [all …]
|
D | rmap.h | 185 struct anon_vma *page_get_anon_vma(struct page *page); 194 void page_move_anon_rmap(struct page *, struct vm_area_struct *); 195 void page_add_anon_rmap(struct page *, struct vm_area_struct *, 197 void do_page_add_anon_rmap(struct page *, struct vm_area_struct *, 199 void __page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, 201 static inline void page_add_new_anon_rmap(struct page *page, in page_add_new_anon_rmap() argument 206 __page_add_new_anon_rmap(page, vma, address, compound); in page_add_new_anon_rmap() 209 void page_add_file_rmap(struct page *, bool); 210 void page_remove_rmap(struct page *, bool); 212 void hugepage_add_anon_rmap(struct page *, struct vm_area_struct *, [all …]
|
D | mm.h | 162 static inline void __mm_zero_struct_page(struct page *page) in __mm_zero_struct_page() argument 164 unsigned long *_pp = (void *)page; in __mm_zero_struct_page() 167 BUILD_BUG_ON(sizeof(struct page) & 7); in __mm_zero_struct_page() 168 BUILD_BUG_ON(sizeof(struct page) < 56); in __mm_zero_struct_page() 169 BUILD_BUG_ON(sizeof(struct page) > 80); in __mm_zero_struct_page() 171 switch (sizeof(struct page)) { in __mm_zero_struct_page() 192 #define mm_zero_struct_page(pp) ((void)memset((pp), 0, sizeof(struct page))) 230 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) argument 238 #define lru_to_page(head) (list_entry((head)->prev, struct page, lru)) 548 struct page *cow_page; /* Page handler may use for COW fault */ [all …]
|
D | page-isolation.h | 10 static inline bool is_migrate_isolate_page(struct page *page) in is_migrate_isolate_page() argument 12 return get_pageblock_migratetype(page) == MIGRATE_ISOLATE; in is_migrate_isolate_page() 23 static inline bool is_migrate_isolate_page(struct page *page) in is_migrate_isolate_page() argument 36 struct page *has_unmovable_pages(struct zone *zone, struct page *page, 38 void set_pageblock_migratetype(struct page *page, int migratetype); 39 int move_freepages_block(struct zone *zone, struct page *page, 64 struct page *alloc_migrate_target(struct page *page, unsigned long private);
|
D | huge_mm.h | 28 struct page *follow_trans_huge_pmd(struct vm_area_struct *vma, 187 void prep_transhuge_page(struct page *page); 188 void free_transhuge_page(struct page *page); 189 bool is_transparent_hugepage(struct page *page); 191 bool can_split_huge_page(struct page *page, int *pextra_pins); 192 int split_huge_page_to_list(struct page *page, struct list_head *list); 193 static inline int split_huge_page(struct page *page) in split_huge_page() argument 195 return split_huge_page_to_list(page, NULL); in split_huge_page() 197 void deferred_split_huge_page(struct page *page); 200 unsigned long address, bool freeze, struct page *page); [all …]
|
D | cleancache.h | 34 pgoff_t, struct page *); 36 pgoff_t, struct page *); 46 extern int __cleancache_get_page(struct page *); 47 extern void __cleancache_put_page(struct page *); 48 extern void __cleancache_invalidate_page(struct address_space *, struct page *); 58 static inline bool cleancache_fs_enabled(struct page *page) in cleancache_fs_enabled() argument 60 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled() 93 static inline int cleancache_get_page(struct page *page) in cleancache_get_page() argument 95 if (cleancache_enabled && cleancache_fs_enabled(page)) in cleancache_get_page() 96 return __cleancache_get_page(page); in cleancache_get_page() [all …]
|
D | ksm.h | 51 struct page *ksm_might_need_to_copy(struct page *page, 54 void rmap_walk_ksm(struct page *page, struct rmap_walk_control *rwc); 55 void ksm_migrate_page(struct page *newpage, struct page *oldpage); 75 static inline struct page *ksm_might_need_to_copy(struct page *page, in ksm_might_need_to_copy() argument 78 return page; in ksm_might_need_to_copy() 81 static inline void rmap_walk_ksm(struct page *page, in rmap_walk_ksm() argument 86 static inline void ksm_migrate_page(struct page *newpage, struct page *oldpage) in ksm_migrate_page()
|
/include/drm/ttm/ |
D | ttm_set_memory.h | 40 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() 45 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() 50 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc() 55 static inline int ttm_set_pages_wb(struct page *page, int numpages) in ttm_set_pages_wb() argument 57 return set_pages_wb(page, numpages); in ttm_set_pages_wb() 60 static inline int ttm_set_pages_wc(struct page *page, int numpages) in ttm_set_pages_wc() argument 62 unsigned long addr = (unsigned long)page_address(page); in ttm_set_pages_wc() 67 static inline int ttm_set_pages_uc(struct page *page, int numpages) in ttm_set_pages_uc() argument 69 return set_pages_uc(page, numpages); in ttm_set_pages_uc() 78 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() [all …]
|
/include/trace/events/ |
D | pagemap.h | 19 #define trace_pagemap_flags(page) ( \ argument 20 (PageAnon(page) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \ 21 (page_mapped(page) ? PAGEMAP_MAPPED : 0) | \ 22 (PageSwapCache(page) ? PAGEMAP_SWAPCACHE : 0) | \ 23 (PageSwapBacked(page) ? PAGEMAP_SWAPBACKED : 0) | \ 24 (PageMappedToDisk(page) ? PAGEMAP_MAPPEDDISK : 0) | \ 25 (page_has_private(page) ? PAGEMAP_BUFFERS : 0) \ 31 struct page *page, 35 TP_ARGS(page, lru), 38 __field(struct page *, page ) [all …]
|
D | page_ref.h | 15 TP_PROTO(struct page *page, int v), 17 TP_ARGS(page, v), 30 __entry->pfn = page_to_pfn(page); 31 __entry->flags = page->flags; 32 __entry->count = page_ref_count(page); 33 __entry->mapcount = page_mapcount(page); 34 __entry->mapping = page->mapping; 35 __entry->mt = get_pageblock_migratetype(page); 49 TP_PROTO(struct page *page, int v), 51 TP_ARGS(page, v) [all …]
|
D | cma.h | 13 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 16 TP_ARGS(name, pfn, page, count, align), 21 __field(const struct page *, page) 29 __entry->page = page; 37 __entry->page, 44 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 47 TP_ARGS(name, pfn, page, count), 52 __field(const struct page *, page) 59 __entry->page = page; 66 __entry->page, [all …]
|
/include/net/ |
D | page_pool.h | 134 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); 136 static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool) in page_pool_dev_alloc_pages() 157 void page_pool_release_page(struct page_pool *pool, struct page *page); 168 struct page *page) in page_pool_release_page() argument 173 void page_pool_put_page(struct page_pool *pool, struct page *page, 178 struct page *page, bool allow_direct) in page_pool_put_full_page() argument 184 page_pool_put_page(pool, page, -1, allow_direct); in page_pool_put_full_page() 190 struct page *page) in page_pool_recycle_direct() argument 192 page_pool_put_full_page(pool, page, true); in page_pool_recycle_direct() 195 static inline dma_addr_t page_pool_get_dma_addr(struct page *page) in page_pool_get_dma_addr() argument [all …]
|
/include/trace/hooks/ |
D | mm.h | 62 TP_PROTO(struct cma *cma, struct page *page, unsigned long count, 64 TP_ARGS(cma, page, count, align, gfp_mask, ts)); 73 int fgp_flags, gfp_t gfp_mask, struct page *page), 74 TP_ARGS(mapping, index, fgp_flags, gfp_mask, page)); 76 TP_PROTO(struct vm_fault *vmf, struct page **page, bool *retry), 77 TP_ARGS(vmf, page, retry)); 79 TP_PROTO(struct vm_fault *vmf, struct page *page), 80 TP_ARGS(vmf, page)); 159 TP_PROTO(struct vm_area_struct *vma, struct page *page, unsigned long addr, bool ret), 160 TP_ARGS(vma, page, addr, ret)); [all …]
|