Home
last modified time | relevance | path

Searched refs:pgd_index (Results 1 – 25 of 54) sorted by relevance

123

/arch/x86/power/
Dhibernate_64.c77 set_pgd(pgd + pgd_index(restore_jump_address), new_pgd); in set_up_temporary_text_mapping()
81 set_pgd(pgd + pgd_index(restore_jump_address), new_pgd); in set_up_temporary_text_mapping()
Dhibernate_32.c89 pgd_idx = pgd_index(PAGE_OFFSET); in resume_physical_mapping_init()
151 pgd = pgd_base + pgd_index(restore_jump_address); in set_up_temporary_text_mapping()
Dhibernate.c166 pgd_index(relocated_restore_code); in relocate_restore_code()
/arch/arm/mm/
Dkasan_init.c226 BUILD_BUG_ON(pgd_index(KASAN_SHADOW_START) != in kasan_init()
227 pgd_index(KASAN_SHADOW_END)); in kasan_init()
231 set_pgd(&tmp_pgd_table[pgd_index(KASAN_SHADOW_START)], in kasan_init()
Dpgd.c57 new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR), in pgd_alloc()
153 pgd = pgd_base + pgd_index(0); in pgd_free()
Didmap.c96 pgd += pgd_index(addr); in identity_mapping_add()
Dioremap.c125 sizeof(pgd_t) * (pgd_index(VMALLOC_END) - in __check_vmalloc_seq()
126 pgd_index(VMALLOC_START))); in __check_vmalloc_seq()
/arch/csky/mm/
Dhighmem.c31 pgd = swapper_pg_dir + pgd_index(vaddr); in kmap_init()
Dinit.c165 i = pgd_index(vaddr); in fixrange_init()
/arch/arc/include/asm/
Dpgtable-levels.h101 #define pgd_index(addr) ((addr) >> PGDIR_SHIFT) macro
102 #define pgd_offset(mm, addr) (((mm)->pgd) + pgd_index(addr))
/arch/nds32/kernel/
Dpm.c23 L1_PPTB_mskBASE)) + pgd_index((unsigned int)cpu_resume); in nds32_suspend2ram()
/arch/mips/kvm/
Dmmu.c111 pgd += pgd_index(addr); in kvm_mips_walk_pgd()
234 int i_min = pgd_index(start_gpa); in kvm_mips_flush_gpa_pgd()
235 int i_max = pgd_index(end_gpa); in kvm_mips_flush_gpa_pgd()
356 int i_min = pgd_index(start); \
357 int i_max = pgd_index(end); \
/arch/x86/mm/
Dmem_encrypt_identity.c109 pgd_p = ppd->pgd + pgd_index(ppd->vaddr); in sme_clear_pgd()
121 pgd = ppd->pgd + pgd_index(ppd->vaddr); in sme_prepare_pgd()
417 decrypted_base = (pgd_index(workarea_end) + 1) & (PTRS_PER_PGD - 1); in sme_encrypt_kernel()
421 check_base = (pgd_index(initrd_end) + 1) & (PTRS_PER_PGD - 1); in sme_encrypt_kernel()
Dpgtable_32.c35 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pte_vaddr()
Dinit_32.c112 int pgd_idx = pgd_index(vaddr); in populate_extra_pmd()
141 pgd_idx = pgd_index(vaddr); in page_table_range_init_count()
223 pgd_idx = pgd_index(vaddr); in page_table_range_init()
300 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init()
464 pgd = base + pgd_index(va); in native_pagetable_init()
Dfault.c175 unsigned index = pgd_index(address); in vmalloc_sync_one()
295 pgd_t *pgd = &base[pgd_index(address)]; in dump_pagetable()
351 pgd_t *pgd = base + pgd_index(address); in dump_pagetable()
518 pgd += pgd_index(address); in show_fault_oops()
1029 pgd = init_mm.pgd + pgd_index(address); in spurious_kernel_fault()
/arch/x86/platform/efi/
Defi_64.c80 pgd = efi_pgd + pgd_index(EFI_VA_END); in efi_alloc_page_tables()
115 pgd_efi = efi_pgd + pgd_index(PAGE_OFFSET); in efi_sync_low_kernel_mappings()
118 num_entries = pgd_index(EFI_VA_END) - pgd_index(PAGE_OFFSET); in efi_sync_low_kernel_mappings()
121 pgd_efi = efi_pgd + pgd_index(EFI_VA_END); in efi_sync_low_kernel_mappings()
/arch/mips/mm/
Dpgtable-32.c83 pgd = swapper_pg_dir + pgd_index(vaddr); in pagetable_init()
/arch/ia64/include/asm/
Dpgtable.h354 pgd_index (unsigned long address) in pgd_index() function
361 #define pgd_index pgd_index macro
/arch/powerpc/mm/
Dpgtable.c314 pgd = mm->pgd + pgd_index(addr); in assert_pte_locked()
379 pgdp = pgdir + pgd_index(ea); in __find_linux_pte()
/arch/sh/mm/
Dfault.c51 pgd += pgd_index(addr); in show_pte()
122 unsigned index = pgd_index(address); in vmalloc_sync_one()
/arch/x86/realmode/
Dinit.c167 for (i = pgd_index(__PAGE_OFFSET); i < PTRS_PER_PGD; i++) in setup_real_mode()
/arch/alpha/mm/
Dfault.c238 long index = pgd_index(address); in do_page_fault()
/arch/um/kernel/
Dmem.c104 i = pgd_index(vaddr); in fixrange_init()
/arch/x86/xen/
Dmmu_pv.c448 if (offset < pgd_index(USER_LIMIT)) { in xen_get_user_pgd()
613 hole_low = pgd_index(GUARD_HOLE_BASE_ADDR); in __xen_pgd_walk()
614 hole_high = pgd_index(GUARD_HOLE_END_ADDR); in __xen_pgd_walk()
616 nr = pgd_index(limit) + 1; in __xen_pgd_walk()
1394 user_pgd[pgd_index(VSYSCALL_ADDR)] = in xen_pgd_alloc()
1715 l3 = m2v(pgd[pgd_index(__START_KERNEL_map)].pgd); in xen_setup_kernel_pagetable()
1744 i = pgd_index(xen_start_info->mfn_list); in xen_setup_kernel_pagetable()
1745 if (i && i < pgd_index(__START_KERNEL_map)) in xen_setup_kernel_pagetable()
1823 pgd = native_make_pgd(xen_read_phys_ulong(pa + pgd_index(vaddr) * in xen_early_virt_to_phys()

123