• Home
  • Raw
  • Download

Lines Matching refs:page

145 static struct page *kimage_alloc_page(struct kimage *image,
300 static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) in kimage_alloc_pages()
302 struct page *pages; in kimage_alloc_pages()
327 static void kimage_free_pages(struct page *page) in kimage_free_pages() argument
331 order = page_private(page); in kimage_free_pages()
334 arch_kexec_pre_free_pages(page_address(page), count); in kimage_free_pages()
337 ClearPageReserved(page + i); in kimage_free_pages()
338 __free_pages(page, order); in kimage_free_pages()
343 struct page *page, *next; in kimage_free_page_list() local
345 list_for_each_entry_safe(page, next, list, lru) { in kimage_free_page_list()
346 list_del(&page->lru); in kimage_free_page_list()
347 kimage_free_pages(page); in kimage_free_page_list()
351 static struct page *kimage_alloc_normal_control_pages(struct kimage *image, in kimage_alloc_normal_control_pages()
368 struct page *pages; in kimage_alloc_normal_control_pages()
417 static struct page *kimage_alloc_crash_control_pages(struct kimage *image, in kimage_alloc_crash_control_pages()
442 struct page *pages; in kimage_alloc_crash_control_pages()
484 struct page *kimage_alloc_control_pages(struct kimage *image, in kimage_alloc_control_pages()
487 struct page *pages = NULL; in kimage_alloc_control_pages()
503 struct page *vmcoreinfo_page; in kimage_crash_copy_vmcoreinfo()
542 struct page *page; in kimage_add_entry() local
544 page = kimage_alloc_page(image, GFP_KERNEL, KIMAGE_NO_DEST); in kimage_add_entry()
545 if (!page) in kimage_add_entry()
548 ind_page = page_address(page); in kimage_add_entry()
573 static int kimage_add_page(struct kimage *image, unsigned long page) in kimage_add_page() argument
577 page &= PAGE_MASK; in kimage_add_page()
578 result = kimage_add_entry(image, page | IND_SOURCE); in kimage_add_page()
609 struct page *page; in kimage_free_entry() local
611 page = boot_pfn_to_page(entry >> PAGE_SHIFT); in kimage_free_entry()
612 kimage_free_pages(page); in kimage_free_entry()
662 unsigned long page) in kimage_dst_used() argument
671 if (page == destination) in kimage_dst_used()
680 static struct page *kimage_alloc_page(struct kimage *image, in kimage_alloc_page()
702 struct page *page; in kimage_alloc_page() local
709 list_for_each_entry(page, &image->dest_pages, lru) { in kimage_alloc_page()
710 addr = page_to_boot_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
712 list_del(&page->lru); in kimage_alloc_page()
713 return page; in kimage_alloc_page()
716 page = NULL; in kimage_alloc_page()
721 page = kimage_alloc_pages(gfp_mask, 0); in kimage_alloc_page()
722 if (!page) in kimage_alloc_page()
725 if (page_to_boot_pfn(page) > in kimage_alloc_page()
727 list_add(&page->lru, &image->unusable_pages); in kimage_alloc_page()
730 addr = page_to_boot_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
750 struct page *old_page; in kimage_alloc_page()
754 copy_highpage(page, old_page); in kimage_alloc_page()
766 page = old_page; in kimage_alloc_page()
770 list_add(&page->lru, &image->dest_pages); in kimage_alloc_page()
773 return page; in kimage_alloc_page()
798 struct page *page; in kimage_load_normal_segment() local
802 page = kimage_alloc_page(image, GFP_HIGHUSER, maddr); in kimage_load_normal_segment()
803 if (!page) { in kimage_load_normal_segment()
807 result = kimage_add_page(image, page_to_boot_pfn(page) in kimage_load_normal_segment()
812 ptr = kmap_local_page(page); in kimage_load_normal_segment()
866 struct page *page; in kimage_load_crash_segment() local
870 page = boot_pfn_to_page(maddr >> PAGE_SHIFT); in kimage_load_crash_segment()
871 if (!page) { in kimage_load_crash_segment()
875 arch_kexec_post_alloc_pages(page_address(page), 1, 0); in kimage_load_crash_segment()
876 ptr = kmap_local_page(page); in kimage_load_crash_segment()
891 kexec_flush_icache_page(page); in kimage_load_crash_segment()
893 arch_kexec_pre_free_pages(page_address(page), 1); in kimage_load_crash_segment()