Home
last modified time | relevance | path

Searched refs:PMD_SIZE (Results 1 – 25 of 84) sorted by relevance

1234

/arch/x86/include/asm/
Dpgtable_32_types.h12 # define PMD_SIZE (1UL << PMD_SHIFT) macro
13 # define PMD_MASK (~(PMD_SIZE - 1))
56 #define LDT_END_ADDR (LDT_BASE_ADDR + PMD_SIZE)
Dpgtable_64_types.h98 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
99 #define PMD_MASK (~(PMD_SIZE - 1))
/arch/sh/include/asm/
Dpgtable-3level.h24 #define PMD_SIZE (1UL << PMD_SHIFT) macro
25 #define PMD_MASK (~(PMD_SIZE-1))
27 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
/arch/x86/mm/
Dinit.c289 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask()
290 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask()
356 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
358 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
360 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
370 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
372 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
375 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) in split_mem_range()
376 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
397 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
[all …]
Dmem_encrypt.c132 vaddr += PMD_SIZE; in __sme_early_map_unmap_mem()
133 paddr += PMD_SIZE; in __sme_early_map_unmap_mem()
134 size = (size <= PMD_SIZE) ? 0 : size - PMD_SIZE; in __sme_early_map_unmap_mem()
Dinit_64.c372 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping()
424 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in cleanup_highmap()
435 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap()
515 paddr_next = (paddr & PMD_MASK) + PMD_SIZE; in phys_pmd_init()
898 vmem_altmap_free(altmap, PMD_SIZE / PAGE_SIZE); in free_hugepage_table()
900 free_pagetable(page, get_order(PMD_SIZE)); in free_hugepage_table()
1045 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
1046 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
1061 PMD_SIZE)) { in remove_pmd_table()
1324 all_end = roundup((unsigned long)_brk_end, PMD_SIZE); in mark_rodata_ro()
[all …]
Dkasan_init_64.c48 ((end - addr) == PMD_SIZE) && in kasan_populate_pmd()
49 IS_ALIGNED(addr, PMD_SIZE)) { in kasan_populate_pmd()
50 p = early_alloc(PMD_SIZE, nid, false); in kasan_populate_pmd()
54 memblock_free(__pa(p), PMD_SIZE); in kasan_populate_pmd()
/arch/parisc/kernel/
Dpci-dma.c86 if (end > PMD_SIZE) in map_pte_uncached()
87 end = PMD_SIZE; in map_pte_uncached()
121 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached()
122 orig_vaddr += PMD_SIZE; in map_pmd_uncached()
166 if (end > PMD_SIZE) in unmap_uncached_pte()
167 end = PMD_SIZE; in unmap_uncached_pte()
206 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd()
207 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
/arch/arm64/mm/
Dhugetlbpage.c31 case PMD_SIZE: in arch_hugetlb_migration_supported()
77 *pgsize = PMD_SIZE; in find_num_contig()
93 case PMD_SIZE: in num_contig_ptes()
97 *pgsize = PMD_SIZE; in num_contig_ptes()
243 } else if (sz == PMD_SIZE) { in huge_pte_alloc()
283 if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && in huge_pte_offset()
304 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte()
458 add_huge_page_size(PMD_SIZE); in hugetlbpage_init()
474 case PMD_SIZE: in setup_hugepagesz()
/arch/riscv/include/asm/
Dpgtable-64.h18 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
19 #define PMD_MASK (~(PMD_SIZE - 1))
/arch/s390/mm/
Dvmem.c123 !(address & ~PMD_MASK) && (address + PMD_SIZE <= end) && in vmem_add_mem()
126 address += PMD_SIZE; in vmem_add_mem()
190 address += PMD_SIZE; in vmem_remove_range()
195 address += PMD_SIZE; in vmem_remove_range()
267 new_page = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate()
271 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate()
279 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate()
Dkasan_init.c150 IS_ALIGNED(address, PMD_SIZE) && in kasan_early_vmemmap_populate()
151 end - address >= PMD_SIZE) { in kasan_early_vmemmap_populate()
154 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
158 if (has_edat && address && end - address >= PMD_SIZE && in kasan_early_vmemmap_populate()
169 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
176 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
Dhugetlbpage.c141 size = PMD_SIZE; in clear_huge_pte_skeys()
201 else if (sz == PMD_SIZE) in huge_pte_alloc()
257 if (MACHINE_HAS_EDAT1 && size == PMD_SIZE) { in setup_hugepagesz()
/arch/powerpc/include/asm/nohash/64/
Dpgtable-4k.h31 #define PMD_SIZE (1UL << PMD_SHIFT) macro
32 #define PMD_MASK (~(PMD_SIZE-1))
/arch/nios2/mm/
Dioremap.c33 if (end > PMD_SIZE) in remap_area_pte()
34 end = PMD_SIZE; in remap_area_pte()
70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
/arch/x86/boot/compressed/
Dkaslr_64.c138 start = round_down(start, PMD_SIZE); in add_identity_map()
139 end = round_up(end, PMD_SIZE); in add_identity_map()
/arch/m68k/include/asm/
Dpgtable_mm.h38 #define PMD_SIZE (1UL << PMD_SHIFT) macro
39 #define PMD_MASK (~(PMD_SIZE-1))
/arch/arm/include/asm/
Dstage2_pgtable.h49 phys_addr_t boundary = (addr + PMD_SIZE) & PMD_MASK; in stage2_pmd_addr_end()
66 #define S2_PMD_SIZE PMD_SIZE
Dpgtable-2level.h85 #define PMD_SIZE (1UL << PMD_SHIFT) macro
86 #define PMD_MASK (~(PMD_SIZE-1))
Dhighmem.h7 #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
/arch/x86/kernel/
Dvmlinux.lds.S65 #define ALIGN_ENTRY_TEXT_BEGIN . = ALIGN(PMD_SIZE);
66 #define ALIGN_ENTRY_TEXT_END . = ALIGN(PMD_SIZE);
77 . = ALIGN(PMD_SIZE); \
82 . = ALIGN(PMD_SIZE); \
/arch/mips/mm/
Dioremap.c31 if (end > PMD_SIZE) in remap_area_pte()
32 end = PMD_SIZE; in remap_area_pte()
63 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
/arch/um/include/asm/
Dpgtable-3level.h28 #define PMD_SIZE (1UL << PMD_SHIFT) macro
29 #define PMD_MASK (~(PMD_SIZE-1))
/arch/sparc/mm/
Dhugetlbpage.c292 if (sz >= PMD_SIZE) in huge_pte_alloc()
332 else if (size >= PMD_SIZE) in set_huge_pte_at()
369 else if (size >= PMD_SIZE) in huge_ptep_get_and_clear()
499 addr += PMD_SIZE; in hugetlb_free_pgd_range()
509 end -= PMD_SIZE; in hugetlb_free_pgd_range()
/arch/arm64/include/asm/
Dpgtable-hwdef.h50 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
51 #define PMD_MASK (~(PMD_SIZE-1))
99 #define CONT_PMD_SIZE (CONT_PMDS * PMD_SIZE)

1234