/mm/kasan/ |
D | kasan_init.c | 84 static void __init zero_pud_populate(pgd_t *pgd, unsigned long addr, in zero_pud_populate() argument 87 pud_t *pud = pud_offset(pgd, addr); in zero_pud_populate() 120 pgd_t *pgd = pgd_offset_k(addr); in kasan_populate_zero_shadow() local 138 pgd_populate(&init_mm, pgd, kasan_zero_pud); in kasan_populate_zero_shadow() 139 pud = pud_offset(pgd, addr); in kasan_populate_zero_shadow() 146 if (pgd_none(*pgd)) { in kasan_populate_zero_shadow() 147 pgd_populate(&init_mm, pgd, in kasan_populate_zero_shadow() 150 zero_pud_populate(pgd, addr, next); in kasan_populate_zero_shadow() 151 } while (pgd++, addr = next, addr != end); in kasan_populate_zero_shadow()
|
/mm/ |
D | sparse-vmemmap.c | 127 pud_t * __meminit vmemmap_pud_populate(pgd_t *pgd, unsigned long addr, int node) in vmemmap_pud_populate() argument 129 pud_t *pud = pud_offset(pgd, addr); in vmemmap_pud_populate() 141 pgd_t *pgd = pgd_offset_k(addr); in vmemmap_pgd_populate() local 142 if (pgd_none(*pgd)) { in vmemmap_pgd_populate() 146 pgd_populate(&init_mm, pgd, p); in vmemmap_pgd_populate() 148 return pgd; in vmemmap_pgd_populate() 155 pgd_t *pgd; in vmemmap_populate_basepages() local 161 pgd = vmemmap_pgd_populate(addr, node); in vmemmap_populate_basepages() 162 if (!pgd) in vmemmap_populate_basepages() 164 pud = vmemmap_pud_populate(pgd, addr, node); in vmemmap_populate_basepages()
|
D | gup.c | 204 pgd_t *pgd; in follow_page_mask() local 219 pgd = pgd_offset(mm, address); in follow_page_mask() 220 if (pgd_none(*pgd) || unlikely(pgd_bad(*pgd))) in follow_page_mask() 223 pud = pud_offset(pgd, address); in follow_page_mask() 273 pgd_t *pgd; in get_gate_page() local 283 pgd = pgd_offset_k(address); in get_gate_page() 285 pgd = pgd_offset_gate(mm, address); in get_gate_page() 286 BUG_ON(pgd_none(*pgd)); in get_gate_page() 287 pud = pud_offset(pgd, address); in get_gate_page() 1333 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, in gup_pud_range() argument [all …]
|
D | pagewalk.c | 72 static int walk_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, in walk_pud_range() argument 79 pud = pud_offset(pgd, addr); in walk_pud_range() 101 pgd_t *pgd; in walk_pgd_range() local 105 pgd = pgd_offset(walk->mm, addr); in walk_pgd_range() 108 if (pgd_none_or_clear_bad(pgd)) { in walk_pgd_range() 116 err = walk_pud_range(pgd, addr, next, walk); in walk_pgd_range() 119 } while (pgd++, addr = next, addr != end); in walk_pgd_range()
|
D | memory.c | 426 static inline void free_pud_range(struct mmu_gather *tlb, pgd_t *pgd, in free_pud_range() argument 435 pud = pud_offset(pgd, addr); in free_pud_range() 454 pud = pud_offset(pgd, start); in free_pud_range() 455 pgd_clear(pgd); in free_pud_range() 466 pgd_t *pgd; in free_pgd_range() local 511 pgd = pgd_offset(tlb->mm, addr); in free_pgd_range() 514 if (pgd_none_or_clear_bad(pgd)) in free_pgd_range() 516 free_pud_range(tlb, pgd, addr, next, floor, ceiling); in free_pgd_range() 517 } while (pgd++, addr = next, addr != end); in free_pgd_range() 641 pgd_t *pgd = pgd_offset(vma->vm_mm, addr); in print_bad_pte() local [all …]
|
D | mprotect.c | 195 pgd_t *pgd, unsigned long addr, unsigned long end, in change_pud_range() argument 202 pud = pud_offset(pgd, addr); in change_pud_range() 219 pgd_t *pgd; in change_protection_range() local 225 pgd = pgd_offset(mm, addr); in change_protection_range() 230 if (pgd_none_or_clear_bad(pgd)) in change_protection_range() 232 pages += change_pud_range(vma, pgd, addr, next, newprot, in change_protection_range() 234 } while (pgd++, addr = next, addr != end); in change_protection_range()
|
D | mremap.c | 34 pgd_t *pgd; in get_old_pmd() local 38 pgd = pgd_offset(mm, addr); in get_old_pmd() 39 if (pgd_none_or_clear_bad(pgd)) in get_old_pmd() 42 pud = pud_offset(pgd, addr); in get_old_pmd() 56 pgd_t *pgd; in alloc_new_pmd() local 60 pgd = pgd_offset(mm, addr); in alloc_new_pmd() 61 pud = pud_alloc(mm, pgd, addr); in alloc_new_pmd()
|
D | vmalloc.c | 88 static void vunmap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end) in vunmap_pud_range() argument 93 pud = pud_offset(pgd, addr); in vunmap_pud_range() 106 pgd_t *pgd; in vunmap_page_range() local 110 pgd = pgd_offset_k(addr); in vunmap_page_range() 113 if (pgd_none_or_clear_bad(pgd)) in vunmap_page_range() 115 vunmap_pud_range(pgd, addr, next); in vunmap_page_range() 116 } while (pgd++, addr = next, addr != end); in vunmap_page_range() 162 static int vmap_pud_range(pgd_t *pgd, unsigned long addr, in vmap_pud_range() argument 168 pud = pud_alloc(&init_mm, pgd, addr); in vmap_pud_range() 188 pgd_t *pgd; in vmap_page_range_noflush() local [all …]
|
D | pgtable-generic.c | 19 void pgd_clear_bad(pgd_t *pgd) in pgd_clear_bad() argument 21 pgd_ERROR(*pgd); in pgd_clear_bad() 22 pgd_clear(pgd); in pgd_clear_bad()
|
D | userfaultfd.c | 126 pgd_t *pgd; in mm_alloc_pmd() local 130 pgd = pgd_offset(mm, address); in mm_alloc_pmd() 131 pud = pud_alloc(mm, pgd, address); in mm_alloc_pmd()
|
D | init-mm.c | 19 .pgd = swapper_pg_dir,
|
D | huge_memory.c | 1672 pgd_t *pgd; in page_check_address_pmd() local 1679 pgd = pgd_offset(mm, address); in page_check_address_pmd() 1680 if (!pgd_present(*pgd)) in page_check_address_pmd() 1682 pud = pud_offset(pgd, address); in page_check_address_pmd() 3062 pgd_t *pgd; in split_huge_page_address() local 3068 pgd = pgd_offset(mm, address); in split_huge_page_address() 3069 if (!pgd_present(*pgd)) in split_huge_page_address() 3072 pud = pud_offset(pgd, address); in split_huge_page_address()
|
D | rmap.c | 741 pgd_t *pgd; in mm_find_pmd() local 746 pgd = pgd_offset(mm, address); in mm_find_pmd() 747 if (!pgd_present(*pgd)) in mm_find_pmd() 750 pud = pud_offset(pgd, address); in mm_find_pmd()
|
D | swapfile.c | 1245 static inline int unuse_pud_range(struct vm_area_struct *vma, pgd_t *pgd, in unuse_pud_range() argument 1253 pud = pud_offset(pgd, addr); in unuse_pud_range() 1268 pgd_t *pgd; in unuse_vma() local 1283 pgd = pgd_offset(vma->vm_mm, addr); in unuse_vma() 1286 if (pgd_none_or_clear_bad(pgd)) in unuse_vma() 1288 ret = unuse_pud_range(vma, pgd, addr, next, entry, page); in unuse_vma() 1291 } while (pgd++, addr = next, addr != end); in unuse_vma()
|
D | hugetlb.c | 4452 pgd_t *pgd = pgd_offset(mm, *addr); in huge_pmd_unshare() local 4453 pud_t *pud = pud_offset(pgd, *addr); in huge_pmd_unshare() 4488 pgd_t *pgd; in huge_pte_alloc() local 4492 pgd = pgd_offset(mm, addr); in huge_pte_alloc() 4493 pud = pud_alloc(mm, pgd, addr); in huge_pte_alloc() 4512 pgd_t *pgd; in huge_pte_offset() local 4516 pgd = pgd_offset(mm, addr); in huge_pte_offset() 4517 if (pgd_present(*pgd)) { in huge_pte_offset() 4518 pud = pud_offset(pgd, addr); in huge_pte_offset()
|
D | debug.c | 206 mm->pgd, atomic_read(&mm->mm_users), in dump_mm()
|