Home
last modified time | relevance | path

Searched refs:end_pfn (Results 1 – 25 of 68) sorted by relevance

123

/kernel/linux/linux-5.10/arch/x86/mm/
Dinit.c333 unsigned long start_pfn, unsigned long end_pfn, in save_mr() argument
336 if (start_pfn < end_pfn) { in save_mr()
340 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr()
411 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local
427 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
429 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
431 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
433 if (end_pfn > limit_pfn) in split_mem_range()
434 end_pfn = limit_pfn; in split_mem_range()
435 if (start_pfn < end_pfn) { in split_mem_range()
[all …]
Dinit_32.c265 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local
276 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init()
305 if (pfn >= end_pfn) in kernel_physical_mapping_init()
313 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init()
353 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init()
421 unsigned long start_pfn, unsigned long end_pfn) in add_highpages_with_active_regions() argument
428 start_pfn, end_pfn); in add_highpages_with_active_regions()
430 start_pfn, end_pfn); in add_highpages_with_active_regions()
/kernel/linux/linux-5.10/mm/
Dpage_isolation.c183 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument
191 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in start_isolate_page_range()
194 pfn < end_pfn; in start_isolate_page_range()
221 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument
228 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in undo_isolate_page_range()
231 pfn < end_pfn; in undo_isolate_page_range()
247 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument
252 while (pfn < end_pfn) { in __test_page_isolated_in_pageblock()
284 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument
296 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in test_pages_isolated()
[all …]
Dpage_idle.c124 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local
134 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read()
135 if (end_pfn > max_pfn) in page_idle_bitmap_read()
136 end_pfn = max_pfn; in page_idle_bitmap_read()
138 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read()
169 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local
179 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write()
180 if (end_pfn > max_pfn) in page_idle_bitmap_write()
181 end_pfn = max_pfn; in page_idle_bitmap_write()
183 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_write()
Dmemory_hotplug.c234 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local
245 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
248 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
313 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() local
341 for (; pfn < end_pfn; pfn += cur_nr_pages) { in __add_pages()
343 cur_nr_pages = min(end_pfn - pfn, in __add_pages()
357 unsigned long end_pfn) in find_smallest_section_pfn() argument
359 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn()
378 unsigned long end_pfn) in find_biggest_section_pfn() argument
383 pfn = end_pfn - 1; in find_biggest_section_pfn()
[all …]
Dpage_alloc.c441 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument
449 if (prev_end_pfn != end_pfn) { in defer_init()
450 prev_end_pfn = end_pfn; in defer_init()
455 if (end_pfn < pgdat_end_pfn(NODE_DATA(nid))) in defer_init()
480 static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument
1500 unsigned long end_pfn = PFN_UP(end); in reserve_bootmem_region() local
1502 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region()
1580 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local
1586 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid()
1589 state->last_end = end_pfn; in __early_pfn_to_nid()
[all …]
Dcompaction.c555 unsigned long end_pfn, in isolate_freepages_block() argument
574 for (; blockpfn < end_pfn; blockpfn += stride, cursor += stride) { in isolate_freepages_block()
662 if (unlikely(blockpfn > end_pfn)) in isolate_freepages_block()
663 blockpfn = end_pfn; in isolate_freepages_block()
676 if (strict && blockpfn < end_pfn) in isolate_freepages_block()
701 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument
712 for (; pfn < end_pfn; pfn += isolated, in isolate_freepages_range()
718 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range()
728 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range()
756 if (pfn < end_pfn) { in isolate_freepages_range()
[all …]
Dsparse.c155 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
166 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits()
169 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits()
170 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits()
173 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits()
175 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits()
610 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument
614 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections()
629 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in offline_mem_sections() argument
633 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections()
Dinternal.h313 unsigned long end_pfn, struct zone *zone);
316 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument
321 return __pageblock_pfn_to_page(start_pfn, end_pfn, zone); in pageblock_pfn_to_page()
391 unsigned long start_pfn, unsigned long end_pfn);
394 unsigned long low_pfn, unsigned long end_pfn);
661 unsigned long *end_pfn);
664 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
/kernel/linux/linux-5.10/include/trace/events/
Dpage_isolation.h14 unsigned long end_pfn,
17 TP_ARGS(start_pfn, end_pfn, fin_pfn),
21 __field(unsigned long, end_pfn)
27 __entry->end_pfn = end_pfn;
32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
Dcompaction.h18 unsigned long end_pfn,
22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken),
26 __field(unsigned long, end_pfn)
33 __entry->end_pfn = end_pfn;
40 __entry->end_pfn,
49 unsigned long end_pfn,
53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
60 unsigned long end_pfn,
64 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
/kernel/linux/linux-5.10/arch/mips/loongson64/
Dnuma.c105 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local
121 end_pfn = start_pfn + node_psize; in szmem()
126 start_pfn, end_pfn, num_physpages); in szmem()
133 end_pfn = start_pfn + node_psize; in szmem()
138 start_pfn, end_pfn, num_physpages); in szmem()
155 unsigned long start_pfn, end_pfn; in node_mem_init() local
161 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init()
163 node, start_pfn, end_pfn); in node_mem_init()
168 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init()
175 max_low_pfn = end_pfn; in node_mem_init()
/kernel/linux/linux-5.10/arch/sparc/mm/
Dinit_32.c66 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local
68 if (end_pfn <= max_low_pfn) in calc_highpages()
74 nr += end_pfn - start_pfn; in calc_highpages()
236 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument
241 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region()
244 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region()
285 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local
287 if (end_pfn <= highstart_pfn) in mem_init()
293 map_high_region(start_pfn, end_pfn); in mem_init()
/kernel/linux/linux-5.10/arch/sh/mm/
Dnuma.c28 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local
34 end_pfn = PFN_DOWN(end); in setup_bootmem_node()
41 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node()
52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
Dinit.c210 unsigned long start_pfn, end_pfn; in allocate_pgdat() local
212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat()
224 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat()
229 unsigned long start_pfn, end_pfn; in do_init_bootmem() local
233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) in do_init_bootmem()
234 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem()
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/
Dmemtrace.c115 u64 start_pfn, end_pfn, nr_pages, pfn; in memtrace_alloc_node() local
123 end_pfn = node_end_pfn(nid); in memtrace_alloc_node()
127 end_pfn = round_down(end_pfn - nr_pages, nr_pages); in memtrace_alloc_node()
130 for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { in memtrace_alloc_node()
143 end_pfn = base_pfn + nr_pages; in memtrace_alloc_node()
144 for (pfn = base_pfn; pfn < end_pfn; pfn += bytes>> PAGE_SHIFT) { in memtrace_alloc_node()
/kernel/linux/linux-5.10/include/linux/
Dnode.h103 unsigned long end_pfn,
107 unsigned long end_pfn, in link_mem_sections() argument
126 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node() local
132 link_mem_sections(nid, start_pfn, end_pfn, MEMINIT_EARLY); in register_one_node()
Dpage-isolation.h46 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
54 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
60 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
/kernel/linux/linux-5.10/arch/x86/xen/
Dsetup.c255 unsigned long end_pfn, unsigned long nr_pages) in xen_set_identity_and_release_chunk() argument
260 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk()
263 end = min(end_pfn, nr_pages); in xen_set_identity_and_release_chunk()
282 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk()
388 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_set_identity_and_remap_chunk() argument
393 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk()
435 for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) in xen_set_identity_and_remap_chunk()
444 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_count_remap_pages() argument
450 return remap_pages + min(end_pfn, nr_pages) - start_pfn; in xen_count_remap_pages()
454 unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, in xen_foreach_remap_area() argument
[all …]
/kernel/linux/linux-5.10/arch/sh/kernel/
Dswsusp.c22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local
24 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
Dsetup.c199 unsigned long end_pfn) in __add_active_range() argument
207 end = end_pfn << PAGE_SHIFT; in __add_active_range()
216 start_pfn, end_pfn); in __add_active_range()
240 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
/kernel/linux/linux-5.10/arch/x86/platform/efi/
Defi_32.c38 u64 start_pfn, end_pfn, end; in efi_map_region() local
45 end_pfn = PFN_UP(end); in efi_map_region()
47 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in efi_map_region()
/kernel/linux/linux-5.10/arch/arm64/mm/
Dnuma.c222 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument
229 if (start_pfn >= end_pfn) in setup_node_data()
250 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_node_data()
370 unsigned long start_pfn, end_pfn; in numa_register_nodes() local
372 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes()
373 setup_node_data(nid, start_pfn, end_pfn); in numa_register_nodes()
/kernel/linux/linux-5.10/drivers/hv/
Dhv_balloon.c437 unsigned long end_pfn; member
447 unsigned long end_pfn; member
584 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) in has_pfn_is_backed()
606 (pfn < has->end_pfn) && in hv_page_offline_check()
773 (pfn + (1UL << order) > has->end_pfn)) in hv_online_page()
796 if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) in pfn_covered()
812 gap->end_pfn = start_pfn; in pfn_covered()
822 if ((start_pfn + pfn_cnt) > has->end_pfn) { in pfn_covered()
823 residual = (start_pfn + pfn_cnt - has->end_pfn); in pfn_covered()
831 has->end_pfn += new_inc; in pfn_covered()
[all …]
/kernel/linux/linux-5.10/arch/powerpc/mm/
Dnuma.c93 static int __init fake_numa_create_new_node(unsigned long end_pfn, in fake_numa_create_new_node() argument
124 if ((end_pfn << PAGE_SHIFT) > mem) { in fake_numa_create_new_node()
1040 unsigned long start_pfn, end_pfn; in setup_nonnuma() local
1049 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { in setup_nonnuma()
1050 fake_numa_create_new_node(end_pfn, &nid); in setup_nonnuma()
1052 PFN_PHYS(end_pfn - start_pfn), in setup_nonnuma()
1094 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument
1096 u64 spanned_pages = end_pfn - start_pfn; in setup_node_data()
1227 unsigned long start_pfn, end_pfn; in initmem_init() local
1229 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in initmem_init()
[all …]

123