• Home
  • Raw
  • Download

Lines Matching refs:page

151 static struct page *kimage_alloc_page(struct kimage *image,
305 static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) in kimage_alloc_pages()
307 struct page *pages; in kimage_alloc_pages()
323 static void kimage_free_pages(struct page *page) in kimage_free_pages() argument
327 order = page_private(page); in kimage_free_pages()
330 ClearPageReserved(page + i); in kimage_free_pages()
331 __free_pages(page, order); in kimage_free_pages()
336 struct page *page, *next; in kimage_free_page_list() local
338 list_for_each_entry_safe(page, next, list, lru) { in kimage_free_page_list()
339 list_del(&page->lru); in kimage_free_page_list()
340 kimage_free_pages(page); in kimage_free_page_list()
344 static struct page *kimage_alloc_normal_control_pages(struct kimage *image, in kimage_alloc_normal_control_pages()
361 struct page *pages; in kimage_alloc_normal_control_pages()
410 static struct page *kimage_alloc_crash_control_pages(struct kimage *image, in kimage_alloc_crash_control_pages()
435 struct page *pages; in kimage_alloc_crash_control_pages()
471 struct page *kimage_alloc_control_pages(struct kimage *image, in kimage_alloc_control_pages()
474 struct page *pages = NULL; in kimage_alloc_control_pages()
495 struct page *page; in kimage_add_entry() local
497 page = kimage_alloc_page(image, GFP_KERNEL, KIMAGE_NO_DEST); in kimage_add_entry()
498 if (!page) in kimage_add_entry()
501 ind_page = page_address(page); in kimage_add_entry()
526 static int kimage_add_page(struct kimage *image, unsigned long page) in kimage_add_page() argument
530 page &= PAGE_MASK; in kimage_add_page()
531 result = kimage_add_entry(image, page | IND_SOURCE); in kimage_add_page()
561 struct page *page; in kimage_free_entry() local
563 page = boot_pfn_to_page(entry >> PAGE_SHIFT); in kimage_free_entry()
564 kimage_free_pages(page); in kimage_free_entry()
609 unsigned long page) in kimage_dst_used() argument
618 if (page == destination) in kimage_dst_used()
627 static struct page *kimage_alloc_page(struct kimage *image, in kimage_alloc_page()
649 struct page *page; in kimage_alloc_page() local
656 list_for_each_entry(page, &image->dest_pages, lru) { in kimage_alloc_page()
657 addr = page_to_boot_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
659 list_del(&page->lru); in kimage_alloc_page()
660 return page; in kimage_alloc_page()
663 page = NULL; in kimage_alloc_page()
668 page = kimage_alloc_pages(gfp_mask, 0); in kimage_alloc_page()
669 if (!page) in kimage_alloc_page()
672 if (page_to_boot_pfn(page) > in kimage_alloc_page()
674 list_add(&page->lru, &image->unusable_pages); in kimage_alloc_page()
677 addr = page_to_boot_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
697 struct page *old_page; in kimage_alloc_page()
701 copy_highpage(page, old_page); in kimage_alloc_page()
714 page = old_page; in kimage_alloc_page()
718 list_add(&page->lru, &image->dest_pages); in kimage_alloc_page()
721 return page; in kimage_alloc_page()
747 struct page *page; in kimage_load_normal_segment() local
751 page = kimage_alloc_page(image, GFP_HIGHUSER, maddr); in kimage_load_normal_segment()
752 if (!page) { in kimage_load_normal_segment()
756 result = kimage_add_page(image, page_to_boot_pfn(page) in kimage_load_normal_segment()
761 ptr = kmap(page); in kimage_load_normal_segment()
774 kunmap(page); in kimage_load_normal_segment()
813 struct page *page; in kimage_load_crash_segment() local
817 page = boot_pfn_to_page(maddr >> PAGE_SHIFT); in kimage_load_crash_segment()
818 if (!page) { in kimage_load_crash_segment()
822 ptr = kmap(page); in kimage_load_crash_segment()
837 kexec_flush_icache_page(page); in kimage_load_crash_segment()
838 kunmap(page); in kimage_load_crash_segment()
1426 VMCOREINFO_STRUCT_SIZE(page); in crash_save_vmcoreinfo_init()
1432 VMCOREINFO_OFFSET(page, flags); in crash_save_vmcoreinfo_init()
1433 VMCOREINFO_OFFSET(page, _refcount); in crash_save_vmcoreinfo_init()
1434 VMCOREINFO_OFFSET(page, mapping); in crash_save_vmcoreinfo_init()
1435 VMCOREINFO_OFFSET(page, lru); in crash_save_vmcoreinfo_init()
1436 VMCOREINFO_OFFSET(page, _mapcount); in crash_save_vmcoreinfo_init()
1437 VMCOREINFO_OFFSET(page, private); in crash_save_vmcoreinfo_init()
1438 VMCOREINFO_OFFSET(page, compound_dtor); in crash_save_vmcoreinfo_init()
1439 VMCOREINFO_OFFSET(page, compound_order); in crash_save_vmcoreinfo_init()
1440 VMCOREINFO_OFFSET(page, compound_head); in crash_save_vmcoreinfo_init()