• Home
  • Raw
  • Download

Lines Matching refs:start_pfn

186 static void register_page_bootmem_info_section(unsigned long start_pfn)  in register_page_bootmem_info_section()  argument
192 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section()
220 static void register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument
226 if (!pfn_valid(start_pfn)) in register_page_bootmem_info_section()
229 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section()
289 static void __meminit grow_zone_span(struct zone *zone, unsigned long start_pfn, in grow_zone_span() argument
297 if (zone_is_empty(zone) || start_pfn < zone->zone_start_pfn) in grow_zone_span()
298 zone->zone_start_pfn = start_pfn; in grow_zone_span()
306 static void resize_zone(struct zone *zone, unsigned long start_pfn, in resize_zone() argument
311 if (end_pfn - start_pfn) { in resize_zone()
312 zone->zone_start_pfn = start_pfn; in resize_zone()
313 zone->spanned_pages = end_pfn - start_pfn; in resize_zone()
326 static void fix_zone_id(struct zone *zone, unsigned long start_pfn, in fix_zone_id() argument
333 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id()
340 unsigned long start_pfn, unsigned long num_pages) in ensure_zone_is_initialized() argument
343 return init_currently_empty_zone(zone, start_pfn, num_pages); in ensure_zone_is_initialized()
349 unsigned long start_pfn, unsigned long end_pfn) in move_pfn_range_left() argument
355 ret = ensure_zone_is_initialized(z1, start_pfn, end_pfn - start_pfn); in move_pfn_range_left()
365 if (start_pfn > z2->zone_start_pfn) in move_pfn_range_left()
375 z1_start_pfn = start_pfn; in move_pfn_range_left()
382 fix_zone_id(z1, start_pfn, end_pfn); in move_pfn_range_left()
391 unsigned long start_pfn, unsigned long end_pfn) in move_pfn_range_right() argument
397 ret = ensure_zone_is_initialized(z2, start_pfn, end_pfn - start_pfn); in move_pfn_range_right()
404 if (z1->zone_start_pfn > start_pfn) in move_pfn_range_right()
410 if (start_pfn >= zone_end_pfn(z1)) in move_pfn_range_right()
419 resize_zone(z1, z1->zone_start_pfn, start_pfn); in move_pfn_range_right()
420 resize_zone(z2, start_pfn, z2_end_pfn); in move_pfn_range_right()
424 fix_zone_id(z2, start_pfn, end_pfn); in move_pfn_range_right()
432 static void __meminit grow_pgdat_span(struct pglist_data *pgdat, unsigned long start_pfn, in grow_pgdat_span() argument
437 if (!pgdat->node_spanned_pages || start_pfn < pgdat->node_start_pfn) in grow_pgdat_span()
438 pgdat->node_start_pfn = start_pfn; in grow_pgdat_span()
534 unsigned long start_pfn, in find_smallest_section_pfn() argument
539 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SECTION) { in find_smallest_section_pfn()
540 ms = __pfn_to_section(start_pfn); in find_smallest_section_pfn()
545 if (unlikely(pfn_to_nid(start_pfn) != nid)) in find_smallest_section_pfn()
548 if (zone && zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn()
551 return start_pfn; in find_smallest_section_pfn()
559 unsigned long start_pfn, in find_biggest_section_pfn() argument
567 for (; pfn >= start_pfn; pfn -= PAGES_PER_SECTION) { in find_biggest_section_pfn()
585 static void shrink_zone_span(struct zone *zone, unsigned long start_pfn, in shrink_zone_span() argument
596 if (zone_start_pfn == start_pfn) { in shrink_zone_span()
617 start_pfn); in shrink_zone_span()
639 if (start_pfn == pfn) in shrink_zone_span()
654 unsigned long start_pfn, unsigned long end_pfn) in shrink_pgdat_span() argument
663 if (pgdat_start_pfn == start_pfn) { in shrink_pgdat_span()
684 start_pfn); in shrink_pgdat_span()
707 if (start_pfn == pfn) in shrink_pgdat_span()
719 static void __remove_zone(struct zone *zone, unsigned long start_pfn) in __remove_zone() argument
729 shrink_zone_span(zone, start_pfn, start_pfn + nr_pages); in __remove_zone()
730 shrink_pgdat_span(pgdat, start_pfn, start_pfn + nr_pages); in __remove_zone()
736 unsigned long start_pfn; in __remove_section() local
748 start_pfn = section_nr_to_pfn(scn_nr); in __remove_section()
749 __remove_zone(zone, start_pfn); in __remove_section()
867 static int online_pages_range(unsigned long start_pfn, unsigned long nr_pages, in online_pages_range() argument
873 if (PageReserved(pfn_to_page(start_pfn))) in online_pages_range()
875 page = pfn_to_page(start_pfn + i); in online_pages_range()
1012 arg.start_pfn = pfn; in online_pages()
1096 unsigned long start_pfn = PFN_DOWN(start); in hotadd_new_pgdat() local
1114 free_area_init_node(nid, zones_size, start_pfn, zholes_size); in hotadd_new_pgdat()
1187 u64 start_pfn = PFN_DOWN(start); in check_hotplug_memory_range() local
1191 if ((start_pfn & ~PAGE_SECTION_MASK) || in check_hotplug_memory_range()
1209 unsigned long start_pfn = start >> PAGE_SHIFT; in should_add_memory_movable() local
1216 if (movable_zone->zone_start_pfn <= start_pfn) in should_add_memory_movable()
1358 int is_mem_section_removable(unsigned long start_pfn, unsigned long nr_pages) in is_mem_section_removable() argument
1360 struct page *page = pfn_to_page(start_pfn); in is_mem_section_removable()
1361 unsigned long end_pfn = min(start_pfn + nr_pages, zone_end_pfn(page_zone(page))); in is_mem_section_removable()
1379 int test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn, in test_pages_in_a_zone() argument
1387 for (pfn = start_pfn, sec_end_pfn = SECTION_ALIGN_UP(start_pfn + 1); in test_pages_in_a_zone()
1453 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) in do_migrate_range() argument
1462 for (pfn = start_pfn; pfn < end_pfn && move_pages > 0; pfn++) { in do_migrate_range()
1555 offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn) in offline_isolated_pages() argument
1557 walk_system_ram_range(start_pfn, end_pfn - start_pfn, NULL, in offline_isolated_pages()
1565 check_pages_isolated_cb(unsigned long start_pfn, unsigned long nr_pages, in check_pages_isolated_cb() argument
1570 ret = test_pages_isolated(start_pfn, start_pfn + nr_pages, true); in check_pages_isolated_cb()
1578 check_pages_isolated(unsigned long start_pfn, unsigned long end_pfn) in check_pages_isolated() argument
1583 ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, &offlined, in check_pages_isolated()
1747 static int __ref __offline_pages(unsigned long start_pfn, in __offline_pages() argument
1759 if (!IS_ALIGNED(start_pfn, pageblock_nr_pages)) in __offline_pages()
1765 if (!test_pages_in_a_zone(start_pfn, end_pfn, &valid_start, &valid_end)) in __offline_pages()
1770 nr_pages = end_pfn - start_pfn; in __offline_pages()
1776 ret = start_isolate_page_range(start_pfn, end_pfn, in __offline_pages()
1781 arg.start_pfn = start_pfn; in __offline_pages()
1790 pfn = start_pfn; in __offline_pages()
1809 pfn = scan_movable_pages(start_pfn, end_pfn); in __offline_pages()
1833 dissolve_free_huge_pages(start_pfn, end_pfn); in __offline_pages()
1835 offlined_pages = check_pages_isolated(start_pfn, end_pfn); in __offline_pages()
1843 offline_isolated_pages(start_pfn, end_pfn); in __offline_pages()
1845 undo_isolate_page_range(start_pfn, end_pfn, MIGRATE_MOVABLE); in __offline_pages()
1847 adjust_managed_page_count(pfn_to_page(start_pfn), -offlined_pages); in __offline_pages()
1876 (unsigned long long) start_pfn << PAGE_SHIFT, in __offline_pages()
1880 undo_isolate_page_range(start_pfn, end_pfn, MIGRATE_MOVABLE); in __offline_pages()
1885 int offline_pages(unsigned long start_pfn, unsigned long nr_pages) in offline_pages() argument
1887 return __offline_pages(start_pfn, start_pfn + nr_pages, 120 * HZ); in offline_pages()
1903 int walk_memory_range(unsigned long start_pfn, unsigned long end_pfn, in walk_memory_range() argument
1911 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in walk_memory_range()
2012 unsigned long start_pfn = pgdat->node_start_pfn; in try_offline_node() local
2013 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in try_offline_node()
2017 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in try_offline_node()