Searched refs:PAGES_PER_SECTION (Results 1 – 7 of 7) sorted by relevance
/arch/x86/mm/ |
D | numa_32.c | 55 start = round_down(start, PAGES_PER_SECTION); in memory_present() 56 end = round_up(end, PAGES_PER_SECTION); in memory_present() 57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present()
|
D | numa.c | 543 if (pfn_align && pfn_align < PAGES_PER_SECTION) { in numa_register_memblks() 546 PFN_PHYS(PAGES_PER_SECTION) >> 20); in numa_register_memblks()
|
/arch/x86/include/asm/ |
D | mmzone_32.h | 29 #define PAGES_PER_SECTION (MAX_NR_PAGES/MAX_SECTIONS) macro 36 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid()
|
/arch/arm64/mm/ |
D | init.c | 325 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 344 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) in free_unused_memmap() 345 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
/arch/arm/mm/ |
D | init.c | 381 ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 407 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) in free_unused_memmap() 409 ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
/arch/powerpc/mm/ |
D | init_64.c | 180 for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page))) in vmemmap_populated()
|
/arch/powerpc/platforms/pseries/ |
D | hotplug-memory.c | 213 rc &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in lmb_is_removable()
|