/arch/arm64/mm/ |
D | mmu.c | 66 void set_swapper_pgd(pgd_t *pgdp, pgd_t pgd) in set_swapper_pgd() argument 71 fixmap_pgdp = pgd_set_fixmap(__pa_symbol(pgdp)); in set_swapper_pgd() 299 static void alloc_init_pud(pgd_t *pgdp, unsigned long addr, unsigned long end, in alloc_init_pud() argument 306 p4d_t *p4dp = p4d_offset(pgdp, addr); in alloc_init_pud() 365 pgd_t *pgdp = pgd_offset_pgd(pgdir, virt); in __create_pgd_mapping() local 380 alloc_init_pud(pgdp, addr, next, phys, prot, pgtable_alloc, in __create_pgd_mapping() 383 } while (pgdp++, addr = next, addr != end); in __create_pgd_mapping() 464 static void __init __map_memblock(pgd_t *pgdp, phys_addr_t start, in __map_memblock() argument 467 __create_pgd_mapping(pgdp, start, __phys_to_virt(start), end - start, in __map_memblock() 496 static void __init map_mem(pgd_t *pgdp) in map_mem() argument [all …]
|
D | kasan_init.c | 142 static void __init kasan_p4d_populate(pgd_t *pgdp, unsigned long addr, in kasan_p4d_populate() argument 146 p4d_t *p4dp = p4d_offset(pgdp, addr); in kasan_p4d_populate() 158 pgd_t *pgdp; in kasan_pgd_populate() local 160 pgdp = pgd_offset_k(addr); in kasan_pgd_populate() 163 kasan_p4d_populate(pgdp, addr, next, node, early); in kasan_pgd_populate() 164 } while (pgdp++, addr = next, addr != end); in kasan_pgd_populate() 191 pgd_t *pgdp, *pgdp_new, *pgdp_end; in kasan_copy_shadow() local 193 pgdp = pgd_offset_k(KASAN_SHADOW_START); in kasan_copy_shadow() 197 set_pgd(pgdp_new, READ_ONCE(*pgdp)); in kasan_copy_shadow() 198 } while (pgdp++, pgdp_new++, pgdp != pgdp_end); in kasan_copy_shadow()
|
D | hugetlbpage.c | 106 pgd_t *pgdp = pgd_offset(mm, addr); in find_num_contig() local 112 p4dp = p4d_offset(pgdp, addr); in find_num_contig() 258 pgd_t *pgdp; in huge_pte_alloc() local 264 pgdp = pgd_offset(mm, addr); in huge_pte_alloc() 265 p4dp = p4d_offset(pgdp, addr); in huge_pte_alloc() 303 pgd_t *pgdp; in huge_pte_offset() local 308 pgdp = pgd_offset(mm, addr); in huge_pte_offset() 309 if (!pgd_present(READ_ONCE(*pgdp))) in huge_pte_offset() 312 p4dp = p4d_offset(pgdp, addr); in huge_pte_offset()
|
D | pageattr.c | 200 pgd_t *pgdp; in kernel_page_present() local 210 pgdp = pgd_offset_k(addr); in kernel_page_present() 211 if (pgd_none(READ_ONCE(*pgdp))) in kernel_page_present() 214 p4dp = p4d_offset(pgdp, addr); in kernel_page_present()
|
/arch/powerpc/mm/nohash/ |
D | book3e_pgtable.c | 75 pgd_t *pgdp; in map_kernel_page() local 83 pgdp = pgd_offset_k(ea); in map_kernel_page() 84 p4dp = p4d_offset(pgdp, ea); in map_kernel_page() 95 pgdp = pgd_offset_k(ea); in map_kernel_page() 96 p4dp = p4d_offset(pgdp, ea); in map_kernel_page()
|
/arch/s390/mm/ |
D | hugetlbpage.c | 195 pgd_t *pgdp; in huge_pte_alloc() local 200 pgdp = pgd_offset(mm, addr); in huge_pte_alloc() 201 p4dp = p4d_alloc(mm, pgdp, addr); in huge_pte_alloc() 217 pgd_t *pgdp; in huge_pte_offset() local 222 pgdp = pgd_offset(mm, addr); in huge_pte_offset() 223 if (pgd_present(*pgdp)) { in huge_pte_offset() 224 p4dp = p4d_offset(pgdp, addr); in huge_pte_offset()
|
D | pageattr.c | 288 pgd_t *pgdp; in change_page_attr() local 295 pgdp = pgd_offset_k(addr); in change_page_attr() 297 if (pgd_none(*pgdp)) in change_page_attr() 300 rc = walk_p4d_level(pgdp, addr, next, flags); in change_page_attr() 304 } while (pgdp++, addr = next, addr < end && !rc); in change_page_attr()
|
/arch/sparc/mm/ |
D | srmmu.c | 110 static inline void srmmu_ctxd_set(ctxd_t *ctxp, pgd_t *pgdp) in srmmu_ctxd_set() argument 114 pte = __pte((SRMMU_ET_PTD | (__nocache_pa(pgdp) >> 4))); in srmmu_ctxd_set() 496 pgd_t *pgdp; in srmmu_mapioaddr() local 504 pgdp = pgd_offset_k(virt_addr); in srmmu_mapioaddr() 505 p4dp = p4d_offset(pgdp, virt_addr); in srmmu_mapioaddr() 535 pgd_t *pgdp; in srmmu_unmapioaddr() local 542 pgdp = pgd_offset_k(virt_addr); in srmmu_unmapioaddr() 543 p4dp = p4d_offset(pgdp, virt_addr); in srmmu_unmapioaddr() 682 pgd_t *pgdp; in srmmu_early_allocate_ptable_skeleton() local 689 pgdp = pgd_offset_k(start); in srmmu_early_allocate_ptable_skeleton() [all …]
|
D | fault_64.c | 81 pgd_t *pgdp = pgd_offset(current->mm, tpc); in get_user_insn() local 89 if (pgd_none(*pgdp) || unlikely(pgd_bad(*pgdp))) in get_user_insn() 91 p4dp = p4d_offset(pgdp, tpc); in get_user_insn()
|
/arch/arm64/include/asm/ |
D | mmu_context.h | 137 static inline void __nocfi cpu_replace_ttbr1(pgd_t *pgdp) in cpu_replace_ttbr1() argument 144 phys_addr_t ttbr1 = phys_to_ttbr(virt_to_phys(pgdp)); in cpu_replace_ttbr1() 146 if (system_supports_cnp() && !WARN_ON(pgdp != lm_alias(swapper_pg_dir))) { in cpu_replace_ttbr1()
|
/arch/x86/mm/ |
D | pgtable.c | 262 static void mop_up_one_pmd(struct mm_struct *mm, pgd_t *pgdp) in mop_up_one_pmd() argument 264 pgd_t pgd = *pgdp; in mop_up_one_pmd() 269 pgd_clear(pgdp); in mop_up_one_pmd() 277 static void pgd_mop_up_pmds(struct mm_struct *mm, pgd_t *pgdp) in pgd_mop_up_pmds() argument 282 mop_up_one_pmd(mm, &pgdp[i]); in pgd_mop_up_pmds() 289 pgdp = kernel_to_user_pgdp(pgdp); in pgd_mop_up_pmds() 292 mop_up_one_pmd(mm, &pgdp[i + KERNEL_PGD_BOUNDARY]); in pgd_mop_up_pmds()
|
D | pti.c | 124 pgd_t __pti_set_user_pgtbl(pgd_t *pgdp, pgd_t pgd) in __pti_set_user_pgtbl() argument 135 if (!pgdp_maps_userspace(pgdp)) in __pti_set_user_pgtbl() 142 kernel_to_user_pgdp(pgdp)->pgd = pgd.pgd; in __pti_set_user_pgtbl()
|
/arch/x86/include/asm/ |
D | pgtable_64.h | 161 static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd) in native_set_pgd() argument 163 WRITE_ONCE(*pgdp, pti_set_user_pgtbl(pgdp, pgd)); in native_set_pgd()
|
D | paravirt.h | 475 static inline void __set_pgd(pgd_t *pgdp, pgd_t pgd) in __set_pgd() argument 477 PVOP_VCALL2(mmu.set_pgd, pgdp, native_pgd_val(pgd)); in __set_pgd() 480 #define set_pgd(pgdp, pgdval) do { \ argument 482 __set_pgd(pgdp, pgdval); \ 484 set_p4d((p4d_t *)(pgdp), (p4d_t) { (pgdval).pgd }); \ 487 #define pgd_clear(pgdp) do { \ argument 489 set_pgd(pgdp, native_make_pgd(0)); \
|
D | pgtable.h | 73 #define set_pgd(pgdp, pgd) native_set_pgd(pgdp, pgd) argument 719 pgd_t __pti_set_user_pgtbl(pgd_t *pgdp, pgd_t pgd); 726 static inline pgd_t pti_set_user_pgtbl(pgd_t *pgdp, pgd_t pgd) in pti_set_user_pgtbl() argument 730 return __pti_set_user_pgtbl(pgdp, pgd); in pti_set_user_pgtbl() 733 static inline pgd_t pti_set_user_pgtbl(pgd_t *pgdp, pgd_t pgd) in pti_set_user_pgtbl() argument 1223 static inline pgd_t *kernel_to_user_pgdp(pgd_t *pgdp) in kernel_to_user_pgdp() argument 1225 return ptr_set_bit(pgdp, PTI_PGTABLE_SWITCH_BIT); in kernel_to_user_pgdp() 1228 static inline pgd_t *user_to_kernel_pgdp(pgd_t *pgdp) in user_to_kernel_pgdp() argument 1230 return ptr_clear_bit(pgdp, PTI_PGTABLE_SWITCH_BIT); in user_to_kernel_pgdp()
|
/arch/riscv/include/asm/ |
D | mmu.h | 23 void __init create_pgd_mapping(pgd_t *pgdp, uintptr_t va, phys_addr_t pa,
|
/arch/m68k/include/asm/ |
D | pgtable_no.h | 22 #define pgd_clear(pgdp) argument
|
/arch/arm64/kernel/ |
D | hibernate.c | 185 pgd_t *pgdp; in trans_pgd_map_page() local 191 pgdp = pgd_offset_pgd(trans_pgd, dst_addr); in trans_pgd_map_page() 192 if (pgd_none(READ_ONCE(*pgdp))) { in trans_pgd_map_page() 196 pgd_populate(&init_mm, pgdp, pudp); in trans_pgd_map_page() 199 p4dp = p4d_offset(pgdp, dst_addr); in trans_pgd_map_page()
|
/arch/arm/include/asm/ |
D | pgtable-nommu.h | 23 #define pgd_clear(pgdp) argument
|
/arch/mips/kvm/ |
D | mmu.c | 762 pgd_t *pgdp; in kvm_trap_emul_pte_for_gva() local 771 pgdp = vcpu->arch.guest_kernel_mm.pgd; in kvm_trap_emul_pte_for_gva() 773 pgdp = vcpu->arch.guest_user_mm.pgd; in kvm_trap_emul_pte_for_gva() 775 return kvm_mips_walk_pgd(pgdp, memcache, addr); in kvm_trap_emul_pte_for_gva() 781 pgd_t *pgdp; in kvm_trap_emul_invalidate_gva() local 786 pgdp = vcpu->arch.guest_kernel_mm.pgd; in kvm_trap_emul_invalidate_gva() 787 ptep = kvm_mips_walk_pgd(pgdp, NULL, addr); in kvm_trap_emul_invalidate_gva() 794 pgdp = vcpu->arch.guest_user_mm.pgd; in kvm_trap_emul_invalidate_gva() 795 ptep = kvm_mips_walk_pgd(pgdp, NULL, addr); in kvm_trap_emul_invalidate_gva()
|
/arch/powerpc/mm/book3s64/ |
D | radix_pgtable.c | 74 pgd_t *pgdp; in early_map_kernel_page() local 80 pgdp = pgd_offset_k(ea); in early_map_kernel_page() 81 p4dp = p4d_offset(pgdp, ea); in early_map_kernel_page() 126 pgd_t *pgdp; in __map_kernel_page() local 149 pgdp = pgd_offset_k(ea); in __map_kernel_page() 150 p4dp = p4d_offset(pgdp, ea); in __map_kernel_page() 187 pgd_t *pgdp; in radix__change_memory_range() local 200 pgdp = pgd_offset_k(idx); in radix__change_memory_range() 201 p4dp = p4d_offset(pgdp, idx); in radix__change_memory_range()
|
D | hash_pgtable.c | 148 pgd_t *pgdp; in hash__map_kernel_page() local 156 pgdp = pgd_offset_k(ea); in hash__map_kernel_page() 157 p4dp = p4d_offset(pgdp, ea); in hash__map_kernel_page()
|
/arch/riscv/mm/ |
D | init.c | 385 void __init create_pgd_mapping(pgd_t *pgdp, in create_pgd_mapping() argument 394 if (pgd_val(pgdp[pgd_idx]) == 0) in create_pgd_mapping() 395 pgdp[pgd_idx] = pfn_pgd(PFN_DOWN(pa), prot); in create_pgd_mapping() 399 if (pgd_val(pgdp[pgd_idx]) == 0) { in create_pgd_mapping() 401 pgdp[pgd_idx] = pfn_pgd(PFN_DOWN(next_phys), PAGE_TABLE); in create_pgd_mapping() 405 next_phys = PFN_PHYS(_pgd_pfn(pgdp[pgd_idx])); in create_pgd_mapping()
|
/arch/c6x/include/asm/ |
D | pgtable.h | 26 #define pgd_clear(pgdp) argument
|
/arch/powerpc/mm/ |
D | pgtable.c | 349 pgd_t *pgdp; in __find_linux_pte() local 370 pgdp = pgdir + pgd_index(ea); in __find_linux_pte() 371 p4dp = p4d_offset(pgdp, ea); in __find_linux_pte()
|