/arch/x86/include/asm/ |
D | pgtable_32_types.h | 12 # define PMD_SIZE (1UL << PMD_SHIFT) macro 13 # define PMD_MASK (~(PMD_SIZE - 1)) 56 #define LDT_END_ADDR (LDT_BASE_ADDR + PMD_SIZE)
|
D | pgtable_64_types.h | 98 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 99 #define PMD_MASK (~(PMD_SIZE - 1))
|
/arch/sh/include/asm/ |
D | pgtable-3level.h | 24 #define PMD_SIZE (1UL << PMD_SHIFT) macro 25 #define PMD_MASK (~(PMD_SIZE-1)) 27 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
|
/arch/x86/mm/ |
D | init.c | 289 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask() 290 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask() 356 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 358 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 360 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 370 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 372 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 375 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) in split_mem_range() 376 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 397 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() [all …]
|
D | mem_encrypt.c | 132 vaddr += PMD_SIZE; in __sme_early_map_unmap_mem() 133 paddr += PMD_SIZE; in __sme_early_map_unmap_mem() 134 size = (size <= PMD_SIZE) ? 0 : size - PMD_SIZE; in __sme_early_map_unmap_mem()
|
D | init_64.c | 372 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping() 424 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in cleanup_highmap() 435 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap() 515 paddr_next = (paddr & PMD_MASK) + PMD_SIZE; in phys_pmd_init() 898 vmem_altmap_free(altmap, PMD_SIZE / PAGE_SIZE); in free_hugepage_table() 900 free_pagetable(page, get_order(PMD_SIZE)); in free_hugepage_table() 1045 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table() 1046 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table() 1061 PMD_SIZE)) { in remove_pmd_table() 1324 all_end = roundup((unsigned long)_brk_end, PMD_SIZE); in mark_rodata_ro() [all …]
|
D | kasan_init_64.c | 48 ((end - addr) == PMD_SIZE) && in kasan_populate_pmd() 49 IS_ALIGNED(addr, PMD_SIZE)) { in kasan_populate_pmd() 50 p = early_alloc(PMD_SIZE, nid, false); in kasan_populate_pmd() 54 memblock_free(__pa(p), PMD_SIZE); in kasan_populate_pmd()
|
/arch/parisc/kernel/ |
D | pci-dma.c | 86 if (end > PMD_SIZE) in map_pte_uncached() 87 end = PMD_SIZE; in map_pte_uncached() 121 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached() 122 orig_vaddr += PMD_SIZE; in map_pmd_uncached() 166 if (end > PMD_SIZE) in unmap_uncached_pte() 167 end = PMD_SIZE; in unmap_uncached_pte() 206 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd() 207 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
|
/arch/arm64/mm/ |
D | hugetlbpage.c | 31 case PMD_SIZE: in arch_hugetlb_migration_supported() 77 *pgsize = PMD_SIZE; in find_num_contig() 93 case PMD_SIZE: in num_contig_ptes() 97 *pgsize = PMD_SIZE; in num_contig_ptes() 243 } else if (sz == PMD_SIZE) { in huge_pte_alloc() 283 if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && in huge_pte_offset() 304 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte() 458 add_huge_page_size(PMD_SIZE); in hugetlbpage_init() 474 case PMD_SIZE: in setup_hugepagesz()
|
/arch/riscv/include/asm/ |
D | pgtable-64.h | 18 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 19 #define PMD_MASK (~(PMD_SIZE - 1))
|
/arch/s390/mm/ |
D | vmem.c | 123 !(address & ~PMD_MASK) && (address + PMD_SIZE <= end) && in vmem_add_mem() 126 address += PMD_SIZE; in vmem_add_mem() 190 address += PMD_SIZE; in vmem_remove_range() 195 address += PMD_SIZE; in vmem_remove_range() 267 new_page = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate() 271 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate() 279 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate()
|
D | kasan_init.c | 150 IS_ALIGNED(address, PMD_SIZE) && in kasan_early_vmemmap_populate() 151 end - address >= PMD_SIZE) { in kasan_early_vmemmap_populate() 154 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate() 158 if (has_edat && address && end - address >= PMD_SIZE && in kasan_early_vmemmap_populate() 169 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate() 176 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
|
D | hugetlbpage.c | 141 size = PMD_SIZE; in clear_huge_pte_skeys() 201 else if (sz == PMD_SIZE) in huge_pte_alloc() 257 if (MACHINE_HAS_EDAT1 && size == PMD_SIZE) { in setup_hugepagesz()
|
/arch/powerpc/include/asm/nohash/64/ |
D | pgtable-4k.h | 31 #define PMD_SIZE (1UL << PMD_SHIFT) macro 32 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/nios2/mm/ |
D | ioremap.c | 33 if (end > PMD_SIZE) in remap_area_pte() 34 end = PMD_SIZE; in remap_area_pte() 70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
|
/arch/x86/boot/compressed/ |
D | kaslr_64.c | 138 start = round_down(start, PMD_SIZE); in add_identity_map() 139 end = round_up(end, PMD_SIZE); in add_identity_map()
|
/arch/m68k/include/asm/ |
D | pgtable_mm.h | 38 #define PMD_SIZE (1UL << PMD_SHIFT) macro 39 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/arm/include/asm/ |
D | stage2_pgtable.h | 49 phys_addr_t boundary = (addr + PMD_SIZE) & PMD_MASK; in stage2_pmd_addr_end() 66 #define S2_PMD_SIZE PMD_SIZE
|
D | pgtable-2level.h | 85 #define PMD_SIZE (1UL << PMD_SHIFT) macro 86 #define PMD_MASK (~(PMD_SIZE-1))
|
D | highmem.h | 7 #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
|
/arch/x86/kernel/ |
D | vmlinux.lds.S | 65 #define ALIGN_ENTRY_TEXT_BEGIN . = ALIGN(PMD_SIZE); 66 #define ALIGN_ENTRY_TEXT_END . = ALIGN(PMD_SIZE); 77 . = ALIGN(PMD_SIZE); \ 82 . = ALIGN(PMD_SIZE); \
|
/arch/mips/mm/ |
D | ioremap.c | 31 if (end > PMD_SIZE) in remap_area_pte() 32 end = PMD_SIZE; in remap_area_pte() 63 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
|
/arch/um/include/asm/ |
D | pgtable-3level.h | 28 #define PMD_SIZE (1UL << PMD_SHIFT) macro 29 #define PMD_MASK (~(PMD_SIZE-1))
|
/arch/sparc/mm/ |
D | hugetlbpage.c | 292 if (sz >= PMD_SIZE) in huge_pte_alloc() 332 else if (size >= PMD_SIZE) in set_huge_pte_at() 369 else if (size >= PMD_SIZE) in huge_ptep_get_and_clear() 499 addr += PMD_SIZE; in hugetlb_free_pgd_range() 509 end -= PMD_SIZE; in hugetlb_free_pgd_range()
|
/arch/arm64/include/asm/ |
D | pgtable-hwdef.h | 50 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 51 #define PMD_MASK (~(PMD_SIZE-1)) 99 #define CONT_PMD_SIZE (CONT_PMDS * PMD_SIZE)
|