/mm/ |
D | memory_hotplug.c | 166 static void register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument 173 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section() 201 static void register_page_bootmem_info_section(unsigned long start_pfn) in register_page_bootmem_info_section() argument 208 section_nr = pfn_to_section_nr(start_pfn); in register_page_bootmem_info_section() 330 unsigned long start_pfn, in find_smallest_section_pfn() argument 333 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 334 if (unlikely(!pfn_to_online_page(start_pfn))) in find_smallest_section_pfn() 337 if (unlikely(pfn_to_nid(start_pfn) != nid)) in find_smallest_section_pfn() 340 if (zone && zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn() 343 return start_pfn; in find_smallest_section_pfn() [all …]
|
D | page_isolation.c | 39 arg.start_pfn = pfn; in set_migratetype_isolate() 193 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 201 BUG_ON(!IS_ALIGNED(start_pfn, pageblock_nr_pages)); in start_isolate_page_range() 204 for (pfn = start_pfn; in start_isolate_page_range() 218 for (pfn = start_pfn; in start_isolate_page_range() 233 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 239 BUG_ON(!IS_ALIGNED(start_pfn, pageblock_nr_pages)); in undo_isolate_page_range() 242 for (pfn = start_pfn; in undo_isolate_page_range() 288 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument 300 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in test_pages_isolated() [all …]
|
D | page_ext.c | 286 static int __meminit online_page_ext(unsigned long start_pfn, in online_page_ext() argument 293 start = SECTION_ALIGN_DOWN(start_pfn); in online_page_ext() 294 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext() 302 nid = pfn_to_nid(start_pfn); in online_page_ext() 321 static int __meminit offline_page_ext(unsigned long start_pfn, in offline_page_ext() argument 326 start = SECTION_ALIGN_DOWN(start_pfn); in offline_page_ext() 327 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext() 343 ret = online_page_ext(mn->start_pfn, in page_ext_callback() 347 offline_page_ext(mn->start_pfn, in page_ext_callback() 351 offline_page_ext(mn->start_pfn, in page_ext_callback() [all …]
|
D | page_alloc.c | 578 unsigned long sp, start_pfn; in page_outside_zone_boundaries() local 582 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 591 start_pfn, start_pfn + sp); in page_outside_zone_boundaries() 1404 unsigned long start_pfn = PFN_DOWN(start); in reserve_bootmem_region() local 1407 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region() 1408 if (pfn_valid(start_pfn)) { in reserve_bootmem_region() 1409 struct page *page = pfn_to_page(start_pfn); in reserve_bootmem_region() 1411 init_reserved_page(start_pfn); in reserve_bootmem_region() 1527 struct page *__pageblock_pfn_to_page(unsigned long start_pfn, in __pageblock_pfn_to_page() argument 1536 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in __pageblock_pfn_to_page() [all …]
|
D | sparse.c | 156 void __meminit mminit_validate_memmodel_limits(unsigned long *start_pfn, in mminit_validate_memmodel_limits() argument 165 if (*start_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 168 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 170 *start_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 175 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 609 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument 613 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 628 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in offline_mem_sections() argument 632 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections() 867 int __meminit sparse_add_section(int nid, unsigned long start_pfn, in sparse_add_section() argument [all …]
|
D | compaction.c | 536 unsigned long *start_pfn, in isolate_freepages_block() argument 546 unsigned long blockpfn = *start_pfn; in isolate_freepages_block() 647 trace_mm_compaction_isolate_freepages(*start_pfn, blockpfn, in isolate_freepages_block() 651 *start_pfn = blockpfn; in isolate_freepages_block() 683 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument 688 pfn = start_pfn; in isolate_freepages_range() 791 unsigned long start_pfn = low_pfn; in isolate_migratepages_block() local 1063 trace_mm_compaction_isolate_migratepages(start_pfn, low_pfn, in isolate_migratepages_block() 1085 isolate_migratepages_range(struct compact_control *cc, unsigned long start_pfn, in isolate_migratepages_range() argument 1091 pfn = start_pfn; in isolate_migratepages_range() [all …]
|
D | internal.h | 147 extern struct page *__pageblock_pfn_to_page(unsigned long start_pfn, 150 static inline struct page *pageblock_pfn_to_page(unsigned long start_pfn, in pageblock_pfn_to_page() argument 154 return pfn_to_page(start_pfn); in pageblock_pfn_to_page() 156 return __pageblock_pfn_to_page(start_pfn, end_pfn, zone); in pageblock_pfn_to_page() 219 unsigned long start_pfn, unsigned long end_pfn); 462 extern void mminit_validate_memmodel_limits(unsigned long *start_pfn, 465 static inline void mminit_validate_memmodel_limits(unsigned long *start_pfn, in mminit_validate_memmodel_limits() argument
|
D | shuffle.c | 108 unsigned long start_pfn = z->zone_start_pfn; in __shuffle_zone() local 114 start_pfn = ALIGN(start_pfn, order_pages); in __shuffle_zone() 115 for (i = start_pfn; i < end_pfn; i += order_pages) { in __shuffle_zone()
|
D | memblock.c | 1601 unsigned long start_pfn, end_pfn; in memblock_mem_size() local 1604 start_pfn = memblock_region_memory_base_pfn(r); in memblock_mem_size() 1606 start_pfn = min_t(unsigned long, start_pfn, limit_pfn); in memblock_mem_size() 1608 pages += end_pfn - start_pfn; in memblock_mem_size() 1751 unsigned long *start_pfn, unsigned long *end_pfn) in memblock_search_pfn_nid() argument 1759 *start_pfn = PFN_DOWN(type->regions[mid].base); in memblock_search_pfn_nid() 1911 unsigned long start_pfn = PFN_UP(start); in __free_memory_core() local 1915 if (start_pfn >= end_pfn) in __free_memory_core() 1918 __free_pages_memory(start_pfn, end_pfn); in __free_memory_core() 1920 return end_pfn - start_pfn; in __free_memory_core()
|
D | ksm.c | 2716 unsigned long start_pfn, in stable_node_dup_remove_range() argument 2719 if (stable_node->kpfn >= start_pfn && in stable_node_dup_remove_range() 2732 unsigned long start_pfn, in stable_node_chain_remove_range() argument 2741 return stable_node_dup_remove_range(stable_node, start_pfn, in stable_node_chain_remove_range() 2748 stable_node_dup_remove_range(dup, start_pfn, end_pfn); in stable_node_chain_remove_range() 2757 static void ksm_check_stable_tree(unsigned long start_pfn, in ksm_check_stable_tree() argument 2769 start_pfn, end_pfn, in ksm_check_stable_tree() 2779 if (stable_node->kpfn >= start_pfn && in ksm_check_stable_tree() 2813 ksm_check_stable_tree(mn->start_pfn, in ksm_memory_callback() 2814 mn->start_pfn + mn->nr_pages); in ksm_memory_callback()
|
D | hugetlb.c | 1073 static int __alloc_gigantic_page(unsigned long start_pfn, in __alloc_gigantic_page() argument 1076 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_gigantic_page() 1077 return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE, in __alloc_gigantic_page() 1082 unsigned long start_pfn, unsigned long nr_pages) in pfn_range_valid_gigantic() argument 1084 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_gigantic() 1087 for (i = start_pfn; i < end_pfn; i++) { in pfn_range_valid_gigantic() 1109 unsigned long start_pfn, unsigned long nr_pages) in zone_spans_last_pfn() argument 1111 unsigned long last_pfn = start_pfn + nr_pages - 1; in zone_spans_last_pfn() 1649 int dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) in dissolve_free_huge_pages() argument 1658 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) { in dissolve_free_huge_pages()
|
D | kmemleak.c | 1443 unsigned long start_pfn = node_start_pfn(i); in kmemleak_scan() local 1447 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in kmemleak_scan()
|
D | vmstat.c | 1439 unsigned long start_pfn = zone->zone_start_pfn; in pagetypeinfo_showblockcount_print() local 1443 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print()
|
/mm/kasan/ |
D | common.c | 681 start_kaddr = (unsigned long)pfn_to_kaddr(mem_data->start_pfn); in kasan_mem_notifier() 705 pfn_to_nid(mem_data->start_pfn), in kasan_mem_notifier()
|