Lines Matching refs:page
37 static inline void *kmap(struct page *page);
46 static inline void kunmap(struct page *page);
54 static inline struct page *kmap_to_page(void *addr);
96 static inline void *kmap_local_page(struct page *page);
179 static inline void *kmap_atomic(struct page *page);
186 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vma… in flush_anon_page() argument
202 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
204 void *addr = kmap_local_page(page); in clear_user_highpage()
205 clear_user_page(addr, vaddr, page); in clear_user_highpage()
231 clear_user_highpage(&folio->page, vaddr); in vma_alloc_zeroed_movable_folio()
237 static inline void clear_highpage(struct page *page) in clear_highpage() argument
239 void *kaddr = kmap_local_page(page); in clear_highpage()
244 static inline void clear_highpage_kasan_tagged(struct page *page) in clear_highpage_kasan_tagged() argument
246 void *kaddr = kmap_local_page(page); in clear_highpage_kasan_tagged()
254 static inline void tag_clear_highpage(struct page *page) in tag_clear_highpage() argument
265 void zero_user_segments(struct page *page, unsigned start1, unsigned end1,
268 static inline void zero_user_segments(struct page *page, in zero_user_segments() argument
272 void *kaddr = kmap_local_page(page); in zero_user_segments()
275 BUG_ON(end1 > page_size(page) || end2 > page_size(page)); in zero_user_segments()
284 for (i = 0; i < compound_nr(page); i++) in zero_user_segments()
285 flush_dcache_page(page + i); in zero_user_segments()
289 static inline void zero_user_segment(struct page *page, in zero_user_segment() argument
292 zero_user_segments(page, start, end, 0, 0); in zero_user_segment()
295 static inline void zero_user(struct page *page, in zero_user() argument
298 zero_user_segments(page, start, start + size, 0, 0); in zero_user()
303 static inline void copy_user_highpage(struct page *to, struct page *from, in copy_user_highpage()
320 static inline void copy_highpage(struct page *to, struct page *from) in copy_highpage()
341 static inline int copy_mc_user_highpage(struct page *to, struct page *from, in copy_mc_user_highpage()
361 static inline int copy_mc_highpage(struct page *to, struct page *from) in copy_mc_highpage()
380 static inline int copy_mc_user_highpage(struct page *to, struct page *from, in copy_mc_user_highpage()
387 static inline int copy_mc_highpage(struct page *to, struct page *from) in copy_mc_highpage()
394 static inline void memcpy_page(struct page *dst_page, size_t dst_off, in memcpy_page()
395 struct page *src_page, size_t src_off, in memcpy_page()
407 static inline void memset_page(struct page *page, size_t offset, int val, in memset_page() argument
410 char *addr = kmap_local_page(page); in memset_page()
417 static inline void memcpy_from_page(char *to, struct page *page, in memcpy_from_page() argument
420 char *from = kmap_local_page(page); in memcpy_from_page()
427 static inline void memcpy_to_page(struct page *page, size_t offset, in memcpy_to_page() argument
430 char *to = kmap_local_page(page); in memcpy_to_page()
434 flush_dcache_page(page); in memcpy_to_page()
438 static inline void memzero_page(struct page *page, size_t offset, size_t len) in memzero_page() argument
440 char *addr = kmap_local_page(page); in memzero_page()
444 flush_dcache_page(page); in memzero_page()
623 zero_user_segments(&folio->page, start1, xend1, start2, xend2); in folio_zero_segments()
635 zero_user_segments(&folio->page, start, xend, 0, 0); in folio_zero_segment()
647 zero_user_segments(&folio->page, start, start + length, 0, 0); in folio_zero_range()
665 static inline void unmap_and_put_page(struct page *page, void *addr) in unmap_and_put_page() argument
667 folio_release_kmap(page_folio(page), addr); in unmap_and_put_page()