/kernel/linux/linux-5.10/arch/s390/mm/ |
D | vmem.c | 88 ALIGN(unused_pmd_start, PMD_SIZE) - unused_pmd_start); in vmemmap_flush_unused_pmd() 111 if (likely(IS_ALIGNED(unused_pmd_start, PMD_SIZE))) in vmemmap_use_sub_pmd() 121 void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); in vmemmap_use_new_sub_pmd() 129 if (!IS_ALIGNED(start, PMD_SIZE)) in vmemmap_use_new_sub_pmd() 136 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd() 143 void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); in vmemmap_unuse_sub_pmd() 147 return !memchr_inv(page, PAGE_UNUSED, PMD_SIZE); in vmemmap_unuse_sub_pmd() 227 if (IS_ALIGNED(addr, PMD_SIZE) && in modify_pmd_table() 228 IS_ALIGNED(next, PMD_SIZE)) { in modify_pmd_table() 230 vmem_free_pages(pmd_deref(*pmd), get_order(PMD_SIZE)); in modify_pmd_table() [all …]
|
D | kasan_init.c | 171 if (IS_ALIGNED(address, PMD_SIZE) && in kasan_early_vmemmap_populate() 172 end - address >= PMD_SIZE) { in kasan_early_vmemmap_populate() 175 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate() 187 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate() 194 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
|
/kernel/linux/linux-5.10/arch/m68k/mm/ |
D | kmap.c | 51 #define IO_SIZE PMD_SIZE 86 virtaddr += PMD_SIZE; in __free_io_area() 87 size -= PMD_SIZE; in __free_io_area() 249 if (!(virtaddr & (PMD_SIZE-1))) in __ioremap() 264 physaddr += PMD_SIZE; in __ioremap() 265 virtaddr += PMD_SIZE; in __ioremap() 266 size -= PMD_SIZE; in __ioremap() 371 virtaddr += PMD_SIZE; in kernel_set_cachemode() 372 size -= PMD_SIZE; in kernel_set_cachemode()
|
D | motorola.c | 309 if (!(virtaddr & (PMD_SIZE-1))) in map_node() 344 physaddr += PMD_SIZE; in map_node() 358 size -= PMD_SIZE; in map_node() 359 virtaddr += PMD_SIZE; in map_node()
|
/kernel/linux/linux-5.10/arch/x86/mm/ |
D | init.c | 360 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask() 361 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask() 427 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 429 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 431 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 441 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 443 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 446 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) in split_mem_range() 447 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 468 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() [all …]
|
D | init_64.c | 372 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping() 424 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in cleanup_highmap() 435 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap() 515 paddr_next = (paddr & PMD_MASK) + PMD_SIZE; in phys_pmd_init() 900 vmem_altmap_free(altmap, PMD_SIZE / PAGE_SIZE); in free_hugepage_table() 902 free_pagetable(page, get_order(PMD_SIZE)); in free_hugepage_table() 1047 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table() 1048 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table() 1063 PMD_SIZE)) { in remove_pmd_table() 1354 all_end = roundup((unsigned long)_brk_end, PMD_SIZE); in mark_rodata_ro() [all …]
|
D | mem_encrypt.c | 135 vaddr += PMD_SIZE; in __sme_early_map_unmap_mem() 136 paddr += PMD_SIZE; in __sme_early_map_unmap_mem() 137 size = (size <= PMD_SIZE) ? 0 : size - PMD_SIZE; in __sme_early_map_unmap_mem()
|
/kernel/linux/linux-5.10/arch/riscv/mm/ |
D | init.c | 182 vmlinux_end = (vmlinux_end + PMD_SIZE - 1) & PMD_MASK; in setup_bootmem() 367 if (sz == PMD_SIZE) { in create_pmd_mapping() 431 if ((base & (PMD_SIZE - 1)) || (size & (PMD_SIZE - 1))) in best_map_size() 434 return PMD_SIZE; in best_map_size() 489 map_size = PMD_SIZE; in setup_vm() 508 (uintptr_t)fixmap_pte, PMD_SIZE, PAGE_TABLE); in setup_vm() 513 load_pa, PMD_SIZE, PAGE_KERNEL_EXEC); in setup_vm() 533 pa = dtb_pa & ~(PMD_SIZE - 1); in setup_vm() 535 pa, PMD_SIZE, PAGE_KERNEL); in setup_vm() 536 create_pmd_mapping(early_dtb_pmd, DTB_EARLY_BASE_VA + PMD_SIZE, in setup_vm() [all …]
|
/kernel/linux/linux-5.10/arch/sh/include/asm/ |
D | pgtable-3level.h | 23 #define PMD_SIZE (1UL << PMD_SHIFT) macro 24 #define PMD_MASK (~(PMD_SIZE-1)) 26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
|
/kernel/linux/linux-5.10/arch/parisc/kernel/ |
D | pci-dma.c | 85 if (end > PMD_SIZE) in map_pte_uncached() 86 end = PMD_SIZE; in map_pte_uncached() 120 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached() 121 orig_vaddr += PMD_SIZE; in map_pmd_uncached() 170 if (end > PMD_SIZE) in unmap_uncached_pte() 171 end = PMD_SIZE; in unmap_uncached_pte() 210 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd() 211 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
D | pgtable_32_types.h | 12 # define PMD_SIZE (1UL << PMD_SHIFT) macro 13 # define PMD_MASK (~(PMD_SIZE - 1))
|
D | pgtable_64_types.h | 98 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 99 #define PMD_MASK (~(PMD_SIZE - 1))
|
/kernel/linux/linux-5.10/arch/arm64/mm/ |
D | hugetlbpage.c | 68 case PMD_SIZE: in arch_hugetlb_migration_supported() 116 *pgsize = PMD_SIZE; in find_num_contig() 132 case PMD_SIZE: in num_contig_ptes() 136 *pgsize = PMD_SIZE; in num_contig_ptes() 286 } else if (sz == PMD_SIZE) { in huge_pte_alloc() 331 if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && in huge_pte_offset() 352 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte() 512 case PMD_SIZE: in arch_hugetlb_valid_size()
|
/kernel/linux/linux-5.10/include/asm-generic/ |
D | pgtable-nopmd.h | 22 #define PMD_SIZE (1UL << PMD_SHIFT) macro 23 #define PMD_MASK (~(PMD_SIZE-1))
|
/kernel/linux/linux-5.10/arch/nios2/mm/ |
D | ioremap.c | 33 if (end > PMD_SIZE) in remap_area_pte() 34 end = PMD_SIZE; in remap_area_pte() 70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
|
/kernel/linux/linux-5.10/arch/riscv/include/asm/ |
D | pgtable-64.h | 18 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 19 #define PMD_MASK (~(PMD_SIZE - 1))
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/nohash/64/ |
D | pgtable-4k.h | 31 #define PMD_SIZE (1UL << PMD_SHIFT) macro 32 #define PMD_MASK (~(PMD_SIZE-1))
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
D | vmlinux.lds.S | 69 #define ALIGN_ENTRY_TEXT_BEGIN . = ALIGN(PMD_SIZE); 70 #define ALIGN_ENTRY_TEXT_END . = ALIGN(PMD_SIZE); 81 . = ALIGN(PMD_SIZE); \ 86 . = ALIGN(PMD_SIZE); \
|
/kernel/linux/linux-5.10/arch/m68k/include/asm/ |
D | pgtable_mm.h | 41 #define PMD_SIZE (1UL << PMD_SHIFT) macro 42 #define PMD_MASK (~(PMD_SIZE-1))
|
/kernel/linux/linux-5.10/arch/x86/boot/compressed/ |
D | ident_map_64.c | 98 start = round_down(start, PMD_SIZE); in add_identity_map() 99 end = round_up(end, PMD_SIZE); in add_identity_map() 343 end = address + PMD_SIZE; in do_boot_page_fault()
|
/kernel/linux/linux-5.10/arch/um/include/asm/ |
D | pgtable-3level.h | 27 #define PMD_SIZE (1UL << PMD_SHIFT) macro 28 #define PMD_MASK (~(PMD_SIZE-1))
|
/kernel/linux/linux-5.10/arch/arm/include/asm/ |
D | pgtable-2level.h | 87 #define PMD_SIZE (1UL << PMD_SHIFT) macro 88 #define PMD_MASK (~(PMD_SIZE-1))
|
/kernel/linux/linux-5.10/arch/arm/mm/ |
D | mmu.c | 1106 next = (addr + PMD_SIZE - 1) & PMD_MASK; in fill_pmd_gaps() 1195 if (!IS_ALIGNED(block_start, PMD_SIZE)) { in adjust_lowmem_bounds() 1198 len = round_up(block_start, PMD_SIZE) - block_start; in adjust_lowmem_bounds() 1231 if (!IS_ALIGNED(block_start, PMD_SIZE)) in adjust_lowmem_bounds() 1233 else if (!IS_ALIGNED(block_end, PMD_SIZE)) in adjust_lowmem_bounds() 1252 memblock_limit = round_down(memblock_limit, PMD_SIZE); in adjust_lowmem_bounds() 1282 for (addr = 0; addr < KASAN_SHADOW_START; addr += PMD_SIZE) in prepare_page_table() 1290 for (addr = KASAN_SHADOW_END; addr < MODULES_VADDR; addr += PMD_SIZE) in prepare_page_table() 1293 for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE) in prepare_page_table() 1299 addr = ((unsigned long)_exiprom + PMD_SIZE - 1) & PMD_MASK; in prepare_page_table() [all …]
|
/kernel/linux/linux-5.10/drivers/dax/ |
D | device.c | 113 unsigned int fault_size = PMD_SIZE; in __dev_dax_pmd_fault() 118 if (dev_dax->align > PMD_SIZE) { in __dev_dax_pmd_fault() 131 (pmd_addr + PMD_SIZE) > vmf->vma->vm_end) in __dev_dax_pmd_fault() 135 phys = dax_pgoff_to_phys(dev_dax, pgoff, PMD_SIZE); in __dev_dax_pmd_fault() 216 fault_size = PMD_SIZE; in dev_dax_huge_fault()
|
/kernel/linux/linux-5.10/arch/sparc/mm/ |
D | hugetlbpage.c | 293 if (sz >= PMD_SIZE) in huge_pte_alloc() 337 else if (size >= PMD_SIZE) in set_huge_pte_at() 374 else if (size >= PMD_SIZE) in huge_ptep_get_and_clear() 505 addr += PMD_SIZE; in hugetlb_free_pgd_range() 515 end -= PMD_SIZE; in hugetlb_free_pgd_range()
|