/arch/powerpc/mm/ |
D | pgtable-book3e.c | 75 pgd_t *pgdp; in map_kernel_page() local 82 pgdp = pgd_offset_k(ea); in map_kernel_page() 83 pudp = pud_alloc(&init_mm, pgdp, ea); in map_kernel_page() 95 pgdp = pgd_offset_k(ea); in map_kernel_page() 97 if (pgd_none(*pgdp)) { in map_kernel_page() 100 pgd_populate(&init_mm, pgdp, pudp); in map_kernel_page() 103 pudp = pud_offset(pgdp, ea); in map_kernel_page()
|
D | pgtable-radix.c | 47 pgd_t *pgdp; in radix__map_kernel_page() local 56 pgdp = pgd_offset_k(ea); in radix__map_kernel_page() 57 pudp = pud_alloc(&init_mm, pgdp, ea); in radix__map_kernel_page() 75 pgdp = pgd_offset_k(ea); in radix__map_kernel_page() 76 if (pgd_none(*pgdp)) { in radix__map_kernel_page() 79 pgd_populate(&init_mm, pgdp, pudp); in radix__map_kernel_page() 81 pudp = pud_offset(pgdp, ea); in radix__map_kernel_page()
|
D | pgtable-hash64.c | 62 pgd_t *pgdp; in hash__map_kernel_page() local 69 pgdp = pgd_offset_k(ea); in hash__map_kernel_page() 70 pudp = pud_alloc(&init_mm, pgdp, ea); in hash__map_kernel_page()
|
/arch/x86/include/asm/ |
D | pgtable_64.h | 110 extern pgd_t kaiser_set_shadow_pgd(pgd_t *pgdp, pgd_t pgd); 112 static inline pgd_t *native_get_shadow_pgd(pgd_t *pgdp) in native_get_shadow_pgd() argument 118 return (pgd_t *)((unsigned long)pgdp | (unsigned long)PAGE_SIZE); in native_get_shadow_pgd() 121 static inline pgd_t kaiser_set_shadow_pgd(pgd_t *pgdp, pgd_t pgd) in kaiser_set_shadow_pgd() argument 125 static inline pgd_t *native_get_shadow_pgd(pgd_t *pgdp) in native_get_shadow_pgd() argument 132 static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd) in native_set_pgd() argument 134 *pgdp = kaiser_set_shadow_pgd(pgdp, pgd); in native_set_pgd()
|
/arch/sh/mm/ |
D | gup.c | 170 pgd_t *pgdp; in __get_user_pages_fast() local 186 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast() 188 pgd_t pgd = *pgdp; in __get_user_pages_fast() 195 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast() 223 pgd_t *pgdp; in get_user_pages_fast() local 235 pgdp = pgd_offset(mm, addr); in get_user_pages_fast() 237 pgd_t pgd = *pgdp; in get_user_pages_fast() 244 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
|
/arch/sparc/mm/ |
D | gup.c | 157 pgd_t *pgdp; in __get_user_pages_fast() local 166 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast() 168 pgd_t pgd = *pgdp; in __get_user_pages_fast() 175 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast() 187 pgd_t *pgdp; in get_user_pages_fast() local 214 pgdp = pgd_offset(mm, addr); in get_user_pages_fast() 216 pgd_t pgd = *pgdp; in get_user_pages_fast() 223 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
|
D | srmmu.c | 110 static inline void srmmu_ctxd_set(ctxd_t *ctxp, pgd_t *pgdp) in srmmu_ctxd_set() argument 114 pte = __pte((SRMMU_ET_PTD | (__nocache_pa(pgdp) >> 4))); in srmmu_ctxd_set() 491 pgd_t *pgdp; in srmmu_mapioaddr() local 497 pgdp = pgd_offset_k(virt_addr); in srmmu_mapioaddr() 498 pmdp = pmd_offset(pgdp, virt_addr); in srmmu_mapioaddr() 526 pgd_t *pgdp; in srmmu_unmapioaddr() local 530 pgdp = pgd_offset_k(virt_addr); in srmmu_unmapioaddr() 531 pmdp = pmd_offset(pgdp, virt_addr); in srmmu_unmapioaddr() 668 pgd_t *pgdp; in srmmu_early_allocate_ptable_skeleton() local 673 pgdp = pgd_offset_k(start); in srmmu_early_allocate_ptable_skeleton() [all …]
|
D | fault_64.c | 94 pgd_t *pgdp = pgd_offset(current->mm, tpc); in get_user_insn() local 101 if (pgd_none(*pgdp) || unlikely(pgd_bad(*pgdp))) in get_user_insn() 103 pudp = pud_offset(pgdp, tpc); in get_user_insn()
|
/arch/powerpc/include/asm/nohash/64/ |
D | pgtable-4k.h | 62 static inline void pgd_clear(pgd_t *pgdp) in pgd_clear() argument 64 *pgdp = __pgd(0); in pgd_clear() 80 #define pud_offset(pgdp, addr) \ argument 81 (((pud_t *) pgd_page_vaddr(*(pgdp))) + \
|
/arch/s390/mm/ |
D | hugetlbpage.c | 156 pgd_t *pgdp; in huge_pte_alloc() local 160 pgdp = pgd_offset(mm, addr); in huge_pte_alloc() 161 pudp = pud_alloc(mm, pgdp, addr); in huge_pte_alloc() 173 pgd_t *pgdp; in huge_pte_offset() local 177 pgdp = pgd_offset(mm, addr); in huge_pte_offset() 178 if (pgd_present(*pgdp)) { in huge_pte_offset() 179 pudp = pud_offset(pgdp, addr); in huge_pte_offset()
|
D | gup.c | 168 static inline int gup_pud_range(pgd_t *pgdp, pgd_t pgd, unsigned long addr, in gup_pud_range() argument 174 pudp = (pud_t *) pgdp; in gup_pud_range() 206 pgd_t *pgdp, pgd; in __get_user_pages_fast() local 223 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast() 225 pgd = *pgdp; in __get_user_pages_fast() 230 if (!gup_pud_range(pgdp, pgd, addr, next, write, pages, &nr)) in __get_user_pages_fast() 232 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast()
|
D | pageattr.c | 253 pgd_t *pgdp; in change_page_attr() local 260 pgdp = pgd_offset_k(addr); in change_page_attr() 262 if (pgd_none(*pgdp)) in change_page_attr() 265 rc = walk_pud_level(pgdp, addr, next, cpa); in change_page_attr() 269 } while (pgdp++, addr = next, addr < end && !rc); in change_page_attr()
|
/arch/mips/mm/ |
D | gup.c | 188 pgd_t *pgdp; in __get_user_pages_fast() local 217 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast() 219 pgd_t pgd = *pgdp; in __get_user_pages_fast() 226 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast() 254 pgd_t *pgdp; in get_user_pages_fast() local 267 pgdp = pgd_offset(mm, addr); in get_user_pages_fast() 269 pgd_t pgd = *pgdp; in get_user_pages_fast() 276 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
|
D | tlb-r8k.c | 186 pgd_t *pgdp; in __update_tlb() local 203 pgdp = pgd_offset(vma->vm_mm, address); in __update_tlb() 204 pmdp = pmd_offset(pgdp, address); in __update_tlb()
|
D | c-r3k.c | 242 pgd_t *pgdp; in r3k_flush_cache_page() local 254 pgdp = pgd_offset(mm, addr); in r3k_flush_cache_page() 255 pudp = pud_offset(pgdp, addr); in r3k_flush_cache_page()
|
D | c-tx39.c | 171 pgd_t *pgdp; in tx39_flush_cache_page() local 184 pgdp = pgd_offset(mm, page); in tx39_flush_cache_page() 185 pudp = pud_offset(pgdp, page); in tx39_flush_cache_page()
|
/arch/x86/mm/ |
D | kaiser.c | 422 static inline bool is_userspace_pgd(pgd_t *pgdp) in is_userspace_pgd() argument 424 return ((unsigned long)pgdp % PAGE_SIZE) < (PAGE_SIZE / 2); in is_userspace_pgd() 427 pgd_t kaiser_set_shadow_pgd(pgd_t *pgdp, pgd_t pgd) in kaiser_set_shadow_pgd() argument 436 if (is_userspace_pgd(pgdp)) { in kaiser_set_shadow_pgd() 437 native_get_shadow_pgd(pgdp)->pgd = pgd.pgd; in kaiser_set_shadow_pgd() 453 if (!WARN_ON_ONCE((unsigned long)pgdp & PAGE_SIZE) && in kaiser_set_shadow_pgd() 454 is_userspace_pgd(pgdp)) in kaiser_set_shadow_pgd() 455 native_get_shadow_pgd(pgdp)->pgd = pgd.pgd; in kaiser_set_shadow_pgd()
|
D | gup.c | 313 pgd_t *pgdp; in __get_user_pages_fast() local 343 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast() 345 pgd_t pgd = *pgdp; in __get_user_pages_fast() 352 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast() 380 pgd_t *pgdp; in get_user_pages_fast() local 415 pgdp = pgd_offset(mm, addr); in get_user_pages_fast() 417 pgd_t pgd = *pgdp; in get_user_pages_fast() 424 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
|
/arch/score/mm/ |
D | cache.c | 177 pgd_t *pgdp; in flush_cache_range() local 185 pgdp = pgd_offset(mm, start); in flush_cache_range() 186 pudp = pud_offset(pgdp, start); in flush_cache_range() 192 pgdp = pgd_offset(mm, start); in flush_cache_range() 193 pudp = pud_offset(pgdp, start); in flush_cache_range()
|
/arch/arm64/include/asm/ |
D | pgalloc.h | 74 static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pud, pgdval_t prot) in __pgd_populate() argument 76 set_pgd(pgdp, __pgd(pud | prot)); in __pgd_populate() 84 static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pud, pgdval_t prot) in __pgd_populate() argument
|
/arch/sparc/include/asm/ |
D | pgalloc_32.h | 30 static inline void pgd_set(pgd_t * pgdp, pmd_t * pmdp) in pgd_set() argument 34 set_pte((pte_t *)pgdp, __pte((SRMMU_ET_PTD | (pa >> 4)))); in pgd_set()
|
/arch/alpha/include/asm/ |
D | pgtable.h | 228 extern inline void pgd_set(pgd_t * pgdp, pmd_t * pmdp) in pgd_set() argument 229 { pgd_val(*pgdp) = _PAGE_TABLE | ((((unsigned long) pmdp) - PAGE_OFFSET) << (32-PAGE_SHIFT)); } in pgd_set() 261 extern inline void pgd_clear(pgd_t * pgdp) { pgd_val(*pgdp) = 0; } in pgd_clear() argument
|
/arch/m68k/include/asm/ |
D | motorola_pgtable.h | 119 static inline void pgd_set(pgd_t *pgdp, pmd_t *pmdp) in pgd_set() argument 121 pgd_val(*pgdp) = _PAGE_TABLE | _PAGE_ACCESSED | __pa(pmdp); in pgd_set() 152 #define pgd_clear(pgdp) ({ pgd_val(*pgdp) = 0; }) argument
|
D | pgtable_no.h | 21 #define pgd_clear(pgdp) argument
|
D | sun3_pgtable.h | 112 static inline void pgd_set(pgd_t *pgdp, pmd_t *pmdp) in pgd_set() argument 114 pgd_val(*pgdp) = virt_to_phys(pmdp); in pgd_set() 150 static inline void pgd_clear (pgd_t *pgdp) {} in pgd_clear() argument
|