/mm/ |
D | page_isolation.c | 183 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 192 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in start_isolate_page_range() 195 pfn < end_pfn; in start_isolate_page_range() 224 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 231 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in undo_isolate_page_range() 234 pfn < end_pfn; in undo_isolate_page_range() 250 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument 255 while (pfn < end_pfn) { in __test_page_isolated_in_pageblock() 287 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument 299 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in test_pages_isolated() [all …]
|
D | page_idle.c | 124 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 134 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 135 if (end_pfn > max_pfn) in page_idle_bitmap_read() 136 end_pfn = max_pfn; in page_idle_bitmap_read() 138 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 169 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 179 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 180 if (end_pfn > max_pfn) in page_idle_bitmap_write() 181 end_pfn = max_pfn; in page_idle_bitmap_write() 183 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_write()
|
D | memory_hotplug.c | 233 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 244 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 247 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 312 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() local 340 for (; pfn < end_pfn; pfn += cur_nr_pages) { in __add_pages() 342 cur_nr_pages = min(end_pfn - pfn, in __add_pages() 356 unsigned long end_pfn) in find_smallest_section_pfn() argument 358 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 377 unsigned long end_pfn) in find_biggest_section_pfn() argument 382 pfn = end_pfn - 1; in find_biggest_section_pfn() [all …]
|
D | page_alloc.c | 426 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument 434 if (prev_end_pfn != end_pfn) { in defer_init() 435 prev_end_pfn = end_pfn; in defer_init() 440 if (end_pfn < pgdat_end_pfn(NODE_DATA(nid))) in defer_init() 468 static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument 1601 unsigned long end_pfn = PFN_UP(end); in reserve_bootmem_region() local 1603 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region() 1681 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local 1687 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid() 1690 state->last_end = end_pfn; in __early_pfn_to_nid() [all …]
|
D | compaction.c | 557 unsigned long end_pfn, in isolate_freepages_block() argument 576 for (; blockpfn < end_pfn; blockpfn += stride, cursor += stride) { in isolate_freepages_block() 664 if (unlikely(blockpfn > end_pfn)) in isolate_freepages_block() 665 blockpfn = end_pfn; in isolate_freepages_block() 678 if (strict && blockpfn < end_pfn) in isolate_freepages_block() 703 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument 714 for (; pfn < end_pfn; pfn += isolated, in isolate_freepages_range() 720 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range() 730 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range() 758 if (pfn < end_pfn) { in isolate_freepages_range() [all …]
|
D | internal.h | 201 unsigned long end_pfn, struct zone *zone); 204 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument 209 return __pageblock_pfn_to_page(start_pfn, end_pfn, zone); in pageblock_pfn_to_page() 284 unsigned long start_pfn, unsigned long end_pfn); 287 unsigned long low_pfn, unsigned long end_pfn); 554 unsigned long *end_pfn); 557 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
|
D | sparse.c | 155 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument 166 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 169 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 170 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 173 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 175 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 610 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument 614 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 629 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in offline_mem_sections() argument 633 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections()
|
D | page_owner.c | 291 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print() local 304 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print() 312 block_end_pfn = min(block_end_pfn, end_pfn); in pagetypeinfo_showmixedcount_print() 581 unsigned long end_pfn = zone_end_pfn(zone); in init_pages_in_zone() local 589 for (; pfn < end_pfn; ) { in init_pages_in_zone() 598 block_end_pfn = min(block_end_pfn, end_pfn); in init_pages_in_zone()
|
D | page_ext.c | 387 unsigned long start_pfn, end_pfn; in page_ext_init() local 390 end_pfn = node_end_pfn(nid); in page_ext_init() 396 for (pfn = start_pfn; pfn < end_pfn; in page_ext_init()
|
D | shuffle.c | 85 unsigned long end_pfn = zone_end_pfn(z); in __shuffle_zone() local 91 for (i = start_pfn; i < end_pfn; i += order_pages) { in __shuffle_zone()
|
D | memblock.c | 1765 unsigned long *start_pfn, unsigned long *end_pfn) in memblock_search_pfn_nid() argument 1774 *end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size); in memblock_search_pfn_nid() 1930 unsigned long end_pfn = min_t(unsigned long, in __free_memory_core() local 1933 if (start_pfn >= end_pfn) in __free_memory_core() 1936 __free_pages_memory(start_pfn, end_pfn); in __free_memory_core() 1938 return end_pfn - start_pfn; in __free_memory_core()
|
D | ksm.c | 2713 unsigned long end_pfn) in stable_node_dup_remove_range() argument 2716 stable_node->kpfn < end_pfn) { in stable_node_dup_remove_range() 2729 unsigned long end_pfn, in stable_node_chain_remove_range() argument 2738 end_pfn); in stable_node_chain_remove_range() 2744 stable_node_dup_remove_range(dup, start_pfn, end_pfn); in stable_node_chain_remove_range() 2754 unsigned long end_pfn) in ksm_check_stable_tree() argument 2765 start_pfn, end_pfn, in ksm_check_stable_tree() 2776 stable_node->kpfn < end_pfn) in ksm_check_stable_tree()
|
D | kmemleak.c | 1460 unsigned long end_pfn = zone_end_pfn(zone); in kmemleak_scan() local 1463 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in kmemleak_scan()
|
D | vmstat.c | 1502 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print() local 1505 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print()
|
D | hugetlb.c | 1866 int dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) in dissolve_free_huge_pages() argument 1875 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) { in dissolve_free_huge_pages()
|