/arch/sh/include/asm/ |
D | pgtable-3level.h | 22 #define PMD_SIZE (1UL << PMD_SHIFT) macro 23 #define PMD_MASK (~(PMD_SIZE-1)) 25 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
|
/arch/x86/mm/ |
D | init.c | 217 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask() 218 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask() 284 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 286 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 288 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 298 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 300 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 303 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) in split_mem_range() 304 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 325 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() [all …]
|
D | init_64.c | 63 for (; addr < end; addr += PMD_SIZE) { in ident_pmd_init() 336 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping() 382 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in cleanup_highmap() 393 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap() 470 next = (address & PMD_MASK) + PMD_SIZE; in phys_pmd_init() 889 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table() 890 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table() 893 get_order(PMD_SIZE)); in remove_pmd_table() 905 PMD_SIZE)) { in remove_pmd_table() 907 get_order(PMD_SIZE)); in remove_pmd_table() [all …]
|
D | pageattr.c | 102 return __pa_symbol(roundup(_brk_end, PMD_SIZE)) >> PAGE_SHIFT; in highmap_end_pfn() 805 if (start & (PMD_SIZE - 1)) { in unmap_pmd_range() 806 unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; in unmap_pmd_range() 818 while (end - start >= PMD_SIZE) { in unmap_pmd_range() 823 start, start + PMD_SIZE); in unmap_pmd_range() 825 start += PMD_SIZE; in unmap_pmd_range() 969 if (start & (PMD_SIZE - 1)) { in populate_pmd() 971 unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; in populate_pmd() 998 while (end - start >= PMD_SIZE) { in populate_pmd() 1012 start += PMD_SIZE; in populate_pmd() [all …]
|
/arch/x86/include/asm/ |
D | pgtable_32_types.h | 11 # define PMD_SIZE (1UL << PMD_SHIFT) macro 12 # define PMD_MASK (~(PMD_SIZE - 1))
|
D | pgtable_64_types.h | 48 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 49 #define PMD_MASK (~(PMD_SIZE - 1))
|
/arch/powerpc/include/asm/ |
D | pgtable-ppc64-64k.h | 27 #define PMD_SIZE (1UL << PMD_SHIFT) macro 28 #define PMD_MASK (~(PMD_SIZE-1))
|
D | pgtable-ppc64-4k.h | 27 #define PMD_SIZE (1UL << PMD_SHIFT) macro 28 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/tile/mm/ |
D | hugetlbpage.c | 72 if (sz >= PMD_SIZE) { in huge_pte_alloc() 73 BUG_ON(sz != PMD_SIZE && in huge_pte_alloc() 74 sz != (PMD_SIZE << huge_shift[HUGE_SHIFT_PMD])); in huge_pte_alloc() 84 BUG_ON(sz != PMD_SIZE); in huge_pte_alloc() 274 } else if (ps >= PMD_SIZE) { in __setup_hugepagesz() 329 BUILD_BUG_ON(ADDITIONAL_HUGE_SIZE >= PMD_SIZE || in add_default_hugepagesz()
|
/arch/s390/mm/ |
D | vmem.c | 116 !(address & ~PMD_MASK) && (address + PMD_SIZE <= end)) { in vmem_add_mem() 121 address += PMD_SIZE; in vmem_add_mem() 175 address += PMD_SIZE; in vmem_remove_range() 180 address += PMD_SIZE; in vmem_remove_range() 230 new_page = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate() 235 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate() 243 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate()
|
/arch/m68k/include/asm/ |
D | pgtable_mm.h | 37 #define PMD_SIZE (1UL << PMD_SHIFT) macro 38 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/nios2/mm/ |
D | ioremap.c | 33 if (end > PMD_SIZE) in remap_area_pte() 34 end = PMD_SIZE; in remap_area_pte() 70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
|
/arch/parisc/kernel/ |
D | pci-dma.c | 91 if (end > PMD_SIZE) in map_pte_uncached() 92 end = PMD_SIZE; in map_pte_uncached() 126 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached() 127 orig_vaddr += PMD_SIZE; in map_pmd_uncached() 171 if (end > PMD_SIZE) in unmap_uncached_pte() 172 end = PMD_SIZE; in unmap_uncached_pte() 211 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd() 212 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
|
/arch/arm64/mm/ |
D | hugetlbpage.c | 69 *pgsize = PMD_SIZE; in find_num_contig() 128 } else if (sz == PMD_SIZE) { in huge_pte_alloc() 134 } else if (sz == (PMD_SIZE * CONT_PMDS)) { in huge_pte_alloc() 193 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte() 305 if (ps == PMD_SIZE) { in setup_hugepagesz()
|
/arch/arm64/include/asm/ |
D | pgtable-hwdef.h | 59 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 60 #define PMD_MASK (~(PMD_SIZE-1)) 108 #define CONT_PMD_SIZE (CONT_PMDS * PMD_SIZE)
|
/arch/mips/mm/ |
D | ioremap.c | 30 if (end > PMD_SIZE) in remap_area_pte() 31 end = PMD_SIZE; in remap_area_pte() 62 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
|
/arch/arm/include/asm/ |
D | pgtable-2level.h | 88 #define PMD_SIZE (1UL << PMD_SHIFT) macro 89 #define PMD_MASK (~(PMD_SIZE-1))
|
D | highmem.h | 6 #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
|
D | kvm_mmu.h | 148 ({ u64 __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \ 241 unsigned long size = PMD_SIZE; in __kvm_flush_dcache_pmd()
|
/arch/um/include/asm/ |
D | pgtable-3level.h | 27 #define PMD_SIZE (1UL << PMD_SHIFT) macro 28 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/tile/include/asm/ |
D | pgtable_64.h | 32 #define PMD_SIZE HPAGE_SIZE macro 33 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/metag/include/asm/ |
D | highmem.h | 26 #define PKMAP_BASE (FIXADDR_START - PMD_SIZE)
|
/arch/x86/xen/ |
D | p2m.c | 407 (unsigned long)(p2m + pfn) + i * PMD_SIZE); in xen_rebuild_p2m_list() 423 PMD_SIZE * PMDS_PER_MID_PAGE); in xen_vmalloc_p2m_tree() 424 vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE); in xen_vmalloc_p2m_tree() 491 vaddr = addr & ~(PMD_SIZE * PMDS_PER_MID_PAGE - 1); in alloc_p2m_pmd() 520 vaddr += PMD_SIZE; in alloc_p2m_pmd()
|
/arch/arm/mm/ |
D | ioremap.c | 171 addr += PMD_SIZE; in unmap_area_sections() 210 addr += PMD_SIZE; in remap_area_sections() 247 addr += PMD_SIZE; in remap_area_supersections()
|
/arch/mips/include/asm/ |
D | pgtable-64.h | 53 #define PMD_SIZE (1UL << PMD_SHIFT) macro 54 #define PMD_MASK (~(PMD_SIZE-1))
|