Lines Matching refs:page
70 static inline void set_page_refcounted(struct page *page) in set_page_refcounted() argument
72 VM_BUG_ON_PAGE(PageTail(page), page); in set_page_refcounted()
73 VM_BUG_ON_PAGE(page_ref_count(page), page); in set_page_refcounted()
74 set_page_count(page, 1); in set_page_refcounted()
88 extern int isolate_lru_page(struct page *page);
89 extern void putback_lru_page(struct page *page);
147 extern struct page *__pageblock_pfn_to_page(unsigned long start_pfn,
150 static inline struct page *pageblock_pfn_to_page(unsigned long start_pfn, in pageblock_pfn_to_page()
159 extern int __isolate_free_page(struct page *page, unsigned int order);
160 extern void memblock_free_pages(struct page *page, unsigned long pfn,
162 extern void __free_pages_core(struct page *page, unsigned int order);
163 extern void prep_compound_page(struct page *page, unsigned int order);
164 extern void post_alloc_hook(struct page *page, unsigned int order,
214 struct page *page; member
236 static inline unsigned int page_order(struct page *page) in page_order() argument
239 return page_private(page); in page_order()
253 #define page_order_unsafe(page) READ_ONCE(page_private(page)) argument
308 extern void mlock_vma_page(struct page *page);
309 extern unsigned int munlock_vma_page(struct page *page);
320 extern void clear_page_mlock(struct page *page);
327 static inline void mlock_migrate_page(struct page *newpage, struct page *page) in mlock_migrate_page() argument
329 if (TestClearPageMlocked(page)) { in mlock_migrate_page()
330 int nr_pages = hpage_nr_pages(page); in mlock_migrate_page()
333 __mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages); in mlock_migrate_page()
345 __vma_address(struct page *page, struct vm_area_struct *vma) in __vma_address() argument
347 pgoff_t pgoff = page_to_pgoff(page); in __vma_address()
352 vma_address(struct page *page, struct vm_area_struct *vma) in vma_address() argument
356 start = __vma_address(page, vma); in vma_address()
357 end = start + PAGE_SIZE * (hpage_nr_pages(page) - 1); in vma_address()
387 static inline void clear_page_mlock(struct page *page) { } in clear_page_mlock() argument
388 static inline void mlock_vma_page(struct page *page) { } in mlock_vma_page() argument
389 static inline void mlock_migrate_page(struct page *new, struct page *old) { } in mlock_migrate_page()
398 static inline struct page *mem_map_offset(struct page *base, int offset) in mem_map_offset()
409 static inline struct page *mem_map_next(struct page *iter, in mem_map_next()
410 struct page *base, int offset) in mem_map_next()
486 extern int hwpoison_filter(struct page *p);
568 static inline bool is_migrate_highatomic_page(struct page *page) in is_migrate_highatomic_page() argument
570 return get_pageblock_migratetype(page) == MIGRATE_HIGHATOMIC; in is_migrate_highatomic_page()
574 extern struct page *alloc_new_node_page(struct page *page, unsigned long node);