Searched refs:P4D_SIZE (Results 1 – 22 of 22) sorted by relevance
/kernel/linux/linux-5.10/arch/s390/mm/ |
D | kasan_init.c | 131 address = (address + P4D_SIZE) & P4D_MASK; in kasan_early_vmemmap_populate() 138 IS_ALIGNED(address, P4D_SIZE) && in kasan_early_vmemmap_populate() 139 end - address >= P4D_SIZE) { in kasan_early_vmemmap_populate() 142 address = (address + P4D_SIZE) & P4D_MASK; in kasan_early_vmemmap_populate() 308 BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, P4D_SIZE)); in kasan_early_init() 309 BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, P4D_SIZE)); in kasan_early_init()
|
D | vmem.c | 366 const unsigned long end = start + P4D_SIZE; in try_free_pud_table()
|
/kernel/linux/linux-5.10/include/asm-generic/ |
D | pgtable-nop4d.h | 14 #define P4D_SIZE (1UL << P4D_SHIFT) macro 15 #define P4D_MASK (~(P4D_SIZE-1))
|
D | tlb.h | 570 if (_sz >= P4D_SIZE) \
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
D | pgtable_64_types.h | 64 #define P4D_SIZE (_AC(1, UL) << P4D_SHIFT) macro 65 #define P4D_MASK (~(P4D_SIZE - 1))
|
D | pgtable_areas.h | 19 #define CPU_ENTRY_AREA_MAP_SIZE P4D_SIZE
|
/kernel/linux/linux-5.10/mm/ |
D | ioremap.c | 183 if ((end - addr) != P4D_SIZE) in ioremap_try_huge_p4d() 186 if (!IS_ALIGNED(addr, P4D_SIZE)) in ioremap_try_huge_p4d() 189 if (!IS_ALIGNED(phys_addr, P4D_SIZE)) in ioremap_try_huge_p4d()
|
D | page_vma_mapped.c | 200 step_forward(pvmw, P4D_SIZE); in page_vma_mapped_walk()
|
/kernel/linux/linux-5.10/mm/kasan/ |
D | init.c | 191 if (IS_ALIGNED(addr, P4D_SIZE) && end - addr >= P4D_SIZE) { in zero_p4d_populate() 432 if (IS_ALIGNED(addr, P4D_SIZE) && in kasan_remove_p4d_table() 433 IS_ALIGNED(next, P4D_SIZE)) { in kasan_remove_p4d_table()
|
/kernel/linux/linux-5.10/arch/x86/mm/ |
D | ident_map.c | 71 next = (addr & P4D_MASK) + P4D_SIZE; in ident_p4d_init()
|
D | mem_encrypt_identity.c | 272 entries += (DIV_ROUND_UP(len, P4D_SIZE) + 1) * sizeof(pud_t) * PTRS_PER_PUD; in sme_pgtable_calc() 283 tables += DIV_ROUND_UP(entries, P4D_SIZE) * sizeof(pud_t) * PTRS_PER_PUD; in sme_pgtable_calc()
|
D | kasan_init_64.c | 190 for (; start < end; start += P4D_SIZE) in clear_pgds()
|
D | init_64.c | 691 vaddr_next = (vaddr & P4D_MASK) + P4D_SIZE; in phys_p4d_init()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/book3s64/ |
D | radix_pgtable.c | 891 if (!IS_ALIGNED(addr, P4D_SIZE) || in remove_pagetable() 892 !IS_ALIGNED(next, P4D_SIZE)) { in remove_pagetable()
|
/kernel/linux/linux-5.10/arch/arm/mm/ |
D | dump.c | 381 addr = start + i * P4D_SIZE; in walk_p4d()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/ptdump/ |
D | hashpagetable.c | 440 addr = start + i * P4D_SIZE; in walk_p4d()
|
/kernel/linux/linux-5.10/arch/ia64/mm/ |
D | init.c | 401 end_address += P4D_SIZE; in vmemmap_find_next_valid_pfn()
|
/kernel/linux/linux-5.10/arch/um/kernel/ |
D | tlb.c | 396 last = ADD_ROUND(addr, P4D_SIZE); in flush_tlb_kernel_range_common()
|
/kernel/linux/linux-5.10/include/linux/ |
D | pgtable.h | 694 ({ unsigned long __boundary = ((addr) + P4D_SIZE) & P4D_MASK; \
|
/kernel/linux/linux-5.10/arch/x86/xen/ |
D | mmu_pv.c | 1094 xen_free_ro_pages(pa, P4D_SIZE); in xen_cleanmfnmap_p4d() 1871 n_pud = roundup(size, P4D_SIZE) >> P4D_SHIFT; in xen_relocate_p2m()
|
/kernel/linux/linux-5.10/arch/powerpc/kvm/ |
D | book3s_64_mmu_radix.c | 1354 gpa = (gpa & P4D_MASK) + P4D_SIZE; in debugfs_radix_read()
|
/kernel/linux/linux-5.10/arch/s390/include/asm/ |
D | pgtable.h | 332 #define P4D_SIZE _REGION2_SIZE macro
|