Home
last modified time | relevance | path

Searched refs:pgdp (Results 1 – 25 of 50) sorted by relevance

12

/arch/powerpc/mm/
Dpgtable-book3e.c75 pgd_t *pgdp; in map_kernel_page() local
82 pgdp = pgd_offset_k(ea); in map_kernel_page()
83 pudp = pud_alloc(&init_mm, pgdp, ea); in map_kernel_page()
95 pgdp = pgd_offset_k(ea); in map_kernel_page()
97 if (pgd_none(*pgdp)) { in map_kernel_page()
100 pgd_populate(&init_mm, pgdp, pudp); in map_kernel_page()
103 pudp = pud_offset(pgdp, ea); in map_kernel_page()
Dpgtable-radix.c47 pgd_t *pgdp; in radix__map_kernel_page() local
56 pgdp = pgd_offset_k(ea); in radix__map_kernel_page()
57 pudp = pud_alloc(&init_mm, pgdp, ea); in radix__map_kernel_page()
75 pgdp = pgd_offset_k(ea); in radix__map_kernel_page()
76 if (pgd_none(*pgdp)) { in radix__map_kernel_page()
79 pgd_populate(&init_mm, pgdp, pudp); in radix__map_kernel_page()
81 pudp = pud_offset(pgdp, ea); in radix__map_kernel_page()
Dpgtable-hash64.c62 pgd_t *pgdp; in hash__map_kernel_page() local
69 pgdp = pgd_offset_k(ea); in hash__map_kernel_page()
70 pudp = pud_alloc(&init_mm, pgdp, ea); in hash__map_kernel_page()
/arch/x86/include/asm/
Dpgtable_64.h110 extern pgd_t kaiser_set_shadow_pgd(pgd_t *pgdp, pgd_t pgd);
112 static inline pgd_t *native_get_shadow_pgd(pgd_t *pgdp) in native_get_shadow_pgd() argument
118 return (pgd_t *)((unsigned long)pgdp | (unsigned long)PAGE_SIZE); in native_get_shadow_pgd()
121 static inline pgd_t kaiser_set_shadow_pgd(pgd_t *pgdp, pgd_t pgd) in kaiser_set_shadow_pgd() argument
125 static inline pgd_t *native_get_shadow_pgd(pgd_t *pgdp) in native_get_shadow_pgd() argument
132 static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd) in native_set_pgd() argument
134 *pgdp = kaiser_set_shadow_pgd(pgdp, pgd); in native_set_pgd()
/arch/sh/mm/
Dgup.c170 pgd_t *pgdp; in __get_user_pages_fast() local
186 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast()
188 pgd_t pgd = *pgdp; in __get_user_pages_fast()
195 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast()
223 pgd_t *pgdp; in get_user_pages_fast() local
235 pgdp = pgd_offset(mm, addr); in get_user_pages_fast()
237 pgd_t pgd = *pgdp; in get_user_pages_fast()
244 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
/arch/sparc/mm/
Dgup.c157 pgd_t *pgdp; in __get_user_pages_fast() local
166 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast()
168 pgd_t pgd = *pgdp; in __get_user_pages_fast()
175 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast()
187 pgd_t *pgdp; in get_user_pages_fast() local
214 pgdp = pgd_offset(mm, addr); in get_user_pages_fast()
216 pgd_t pgd = *pgdp; in get_user_pages_fast()
223 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
Dsrmmu.c110 static inline void srmmu_ctxd_set(ctxd_t *ctxp, pgd_t *pgdp) in srmmu_ctxd_set() argument
114 pte = __pte((SRMMU_ET_PTD | (__nocache_pa(pgdp) >> 4))); in srmmu_ctxd_set()
491 pgd_t *pgdp; in srmmu_mapioaddr() local
497 pgdp = pgd_offset_k(virt_addr); in srmmu_mapioaddr()
498 pmdp = pmd_offset(pgdp, virt_addr); in srmmu_mapioaddr()
526 pgd_t *pgdp; in srmmu_unmapioaddr() local
530 pgdp = pgd_offset_k(virt_addr); in srmmu_unmapioaddr()
531 pmdp = pmd_offset(pgdp, virt_addr); in srmmu_unmapioaddr()
668 pgd_t *pgdp; in srmmu_early_allocate_ptable_skeleton() local
673 pgdp = pgd_offset_k(start); in srmmu_early_allocate_ptable_skeleton()
[all …]
Dfault_64.c94 pgd_t *pgdp = pgd_offset(current->mm, tpc); in get_user_insn() local
101 if (pgd_none(*pgdp) || unlikely(pgd_bad(*pgdp))) in get_user_insn()
103 pudp = pud_offset(pgdp, tpc); in get_user_insn()
/arch/powerpc/include/asm/nohash/64/
Dpgtable-4k.h62 static inline void pgd_clear(pgd_t *pgdp) in pgd_clear() argument
64 *pgdp = __pgd(0); in pgd_clear()
80 #define pud_offset(pgdp, addr) \ argument
81 (((pud_t *) pgd_page_vaddr(*(pgdp))) + \
/arch/s390/mm/
Dhugetlbpage.c156 pgd_t *pgdp; in huge_pte_alloc() local
160 pgdp = pgd_offset(mm, addr); in huge_pte_alloc()
161 pudp = pud_alloc(mm, pgdp, addr); in huge_pte_alloc()
173 pgd_t *pgdp; in huge_pte_offset() local
177 pgdp = pgd_offset(mm, addr); in huge_pte_offset()
178 if (pgd_present(*pgdp)) { in huge_pte_offset()
179 pudp = pud_offset(pgdp, addr); in huge_pte_offset()
Dgup.c168 static inline int gup_pud_range(pgd_t *pgdp, pgd_t pgd, unsigned long addr, in gup_pud_range() argument
174 pudp = (pud_t *) pgdp; in gup_pud_range()
206 pgd_t *pgdp, pgd; in __get_user_pages_fast() local
223 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast()
225 pgd = *pgdp; in __get_user_pages_fast()
230 if (!gup_pud_range(pgdp, pgd, addr, next, write, pages, &nr)) in __get_user_pages_fast()
232 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast()
Dpageattr.c253 pgd_t *pgdp; in change_page_attr() local
260 pgdp = pgd_offset_k(addr); in change_page_attr()
262 if (pgd_none(*pgdp)) in change_page_attr()
265 rc = walk_pud_level(pgdp, addr, next, cpa); in change_page_attr()
269 } while (pgdp++, addr = next, addr < end && !rc); in change_page_attr()
/arch/mips/mm/
Dgup.c188 pgd_t *pgdp; in __get_user_pages_fast() local
217 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast()
219 pgd_t pgd = *pgdp; in __get_user_pages_fast()
226 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast()
254 pgd_t *pgdp; in get_user_pages_fast() local
267 pgdp = pgd_offset(mm, addr); in get_user_pages_fast()
269 pgd_t pgd = *pgdp; in get_user_pages_fast()
276 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
Dtlb-r8k.c186 pgd_t *pgdp; in __update_tlb() local
203 pgdp = pgd_offset(vma->vm_mm, address); in __update_tlb()
204 pmdp = pmd_offset(pgdp, address); in __update_tlb()
Dc-r3k.c242 pgd_t *pgdp; in r3k_flush_cache_page() local
254 pgdp = pgd_offset(mm, addr); in r3k_flush_cache_page()
255 pudp = pud_offset(pgdp, addr); in r3k_flush_cache_page()
Dc-tx39.c171 pgd_t *pgdp; in tx39_flush_cache_page() local
184 pgdp = pgd_offset(mm, page); in tx39_flush_cache_page()
185 pudp = pud_offset(pgdp, page); in tx39_flush_cache_page()
/arch/x86/mm/
Dkaiser.c422 static inline bool is_userspace_pgd(pgd_t *pgdp) in is_userspace_pgd() argument
424 return ((unsigned long)pgdp % PAGE_SIZE) < (PAGE_SIZE / 2); in is_userspace_pgd()
427 pgd_t kaiser_set_shadow_pgd(pgd_t *pgdp, pgd_t pgd) in kaiser_set_shadow_pgd() argument
436 if (is_userspace_pgd(pgdp)) { in kaiser_set_shadow_pgd()
437 native_get_shadow_pgd(pgdp)->pgd = pgd.pgd; in kaiser_set_shadow_pgd()
453 if (!WARN_ON_ONCE((unsigned long)pgdp & PAGE_SIZE) && in kaiser_set_shadow_pgd()
454 is_userspace_pgd(pgdp)) in kaiser_set_shadow_pgd()
455 native_get_shadow_pgd(pgdp)->pgd = pgd.pgd; in kaiser_set_shadow_pgd()
Dgup.c313 pgd_t *pgdp; in __get_user_pages_fast() local
343 pgdp = pgd_offset(mm, addr); in __get_user_pages_fast()
345 pgd_t pgd = *pgdp; in __get_user_pages_fast()
352 } while (pgdp++, addr = next, addr != end); in __get_user_pages_fast()
380 pgd_t *pgdp; in get_user_pages_fast() local
415 pgdp = pgd_offset(mm, addr); in get_user_pages_fast()
417 pgd_t pgd = *pgdp; in get_user_pages_fast()
424 } while (pgdp++, addr = next, addr != end); in get_user_pages_fast()
/arch/score/mm/
Dcache.c177 pgd_t *pgdp; in flush_cache_range() local
185 pgdp = pgd_offset(mm, start); in flush_cache_range()
186 pudp = pud_offset(pgdp, start); in flush_cache_range()
192 pgdp = pgd_offset(mm, start); in flush_cache_range()
193 pudp = pud_offset(pgdp, start); in flush_cache_range()
/arch/arm64/include/asm/
Dpgalloc.h74 static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pud, pgdval_t prot) in __pgd_populate() argument
76 set_pgd(pgdp, __pgd(pud | prot)); in __pgd_populate()
84 static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pud, pgdval_t prot) in __pgd_populate() argument
/arch/sparc/include/asm/
Dpgalloc_32.h30 static inline void pgd_set(pgd_t * pgdp, pmd_t * pmdp) in pgd_set() argument
34 set_pte((pte_t *)pgdp, __pte((SRMMU_ET_PTD | (pa >> 4)))); in pgd_set()
/arch/alpha/include/asm/
Dpgtable.h228 extern inline void pgd_set(pgd_t * pgdp, pmd_t * pmdp) in pgd_set() argument
229 { pgd_val(*pgdp) = _PAGE_TABLE | ((((unsigned long) pmdp) - PAGE_OFFSET) << (32-PAGE_SHIFT)); } in pgd_set()
261 extern inline void pgd_clear(pgd_t * pgdp) { pgd_val(*pgdp) = 0; } in pgd_clear() argument
/arch/m68k/include/asm/
Dmotorola_pgtable.h119 static inline void pgd_set(pgd_t *pgdp, pmd_t *pmdp) in pgd_set() argument
121 pgd_val(*pgdp) = _PAGE_TABLE | _PAGE_ACCESSED | __pa(pmdp); in pgd_set()
152 #define pgd_clear(pgdp) ({ pgd_val(*pgdp) = 0; }) argument
Dpgtable_no.h21 #define pgd_clear(pgdp) argument
Dsun3_pgtable.h112 static inline void pgd_set(pgd_t *pgdp, pmd_t *pmdp) in pgd_set() argument
114 pgd_val(*pgdp) = virt_to_phys(pmdp); in pgd_set()
150 static inline void pgd_clear (pgd_t *pgdp) {} in pgd_clear() argument

12