Searched refs:PUD_SIZE (Results 1 – 14 of 14) sorted by relevance
113 if (end > PUD_SIZE) in io_remap_pud_range()114 end = PUD_SIZE; in io_remap_pud_range()121 address = (address + PUD_SIZE) & PUD_MASK; in io_remap_pud_range()
35 end_addr = addr + PUD_SIZE; in init_level2_page()63 addr += PUD_SIZE; in init_level3_page()68 addr += PUD_SIZE; in init_level3_page()
86 BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE); in x86_64_start_kernel()
35 #define PUD_SIZE (1UL << PUD_SHIFT) macro36 #define PUD_MASK (~(PUD_SIZE-1))
27 unsigned long s_end = sbase + PUD_SIZE; in page_table_shareable()44 unsigned long end = base + PUD_SIZE; in vma_shareable()137 if (sz == PUD_SIZE) { in huge_pte_alloc()434 } else if (ps == PUD_SIZE && cpu_has_gbpages) { in setup_hugepagesz()
473 for (; i < PTRS_PER_PUD; i++, addr = (addr & PUD_MASK) + PUD_SIZE) { in phys_pud_init()483 !e820_any_mapped(addr, addr+PUD_SIZE, 0)) { in phys_pud_init()519 last_map_addr = (addr & PUD_MASK) + PUD_SIZE; in phys_pud_init()555 puds = (end + PUD_SIZE - 1) >> PUD_SHIFT; in find_early_table_space()727 end_pfn = ((pos + (PUD_SIZE - 1))>>PUD_SHIFT) in init_memory_mapping()738 start_pfn = ((pos + (PUD_SIZE - 1))>>PUD_SHIFT) in init_memory_mapping()
214 i++, pud++, addr += PUD_SIZE) { in pgd_prepopulate_pmd()
828 puds = (end + PUD_SIZE - 1) >> PUD_SHIFT; in find_early_table_space()
143 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro144 #define PUD_MASK (~(PUD_SIZE - 1))
50 paddr = address + i*PUD_SIZE; in res_phys_pud_init()
114 #define PUD_SIZE (1UL << PUD_SHIFT) macro115 #define PUD_MASK (~(PUD_SIZE-1))
73 #define PUD_SIZE (1UL << PUD_SHIFT) macro74 #define PUD_MASK (~(PUD_SIZE-1))
392 end_address += PUD_SIZE; in vmemmap_find_next_valid_pfn()
321 last = ADD_ROUND(addr, PUD_SIZE); in flush_tlb_kernel_range_common()