/mm/kasan/ |
D | kasan_init.c | 62 static void __init zero_pmd_populate(pud_t *pud, unsigned long addr, in zero_pmd_populate() argument 65 pmd_t *pmd = pmd_offset(pud, addr); in zero_pmd_populate() 87 pud_t *pud = pud_offset(pgd, addr); in zero_pud_populate() local 95 pud_populate(&init_mm, pud, kasan_zero_pmd); in zero_pud_populate() 96 pmd = pmd_offset(pud, addr); in zero_pud_populate() 101 if (pud_none(*pud)) { in zero_pud_populate() 102 pud_populate(&init_mm, pud, in zero_pud_populate() 105 zero_pmd_populate(pud, addr, next); in zero_pud_populate() 106 } while (pud++, addr = next, addr != end); in zero_pud_populate() 127 pud_t *pud; in kasan_populate_zero_shadow() local [all …]
|
/mm/ |
D | sparse-vmemmap.c | 187 pmd_t * __meminit vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node) in vmemmap_pmd_populate() argument 189 pmd_t *pmd = pmd_offset(pud, addr); in vmemmap_pmd_populate() 201 pud_t *pud = pud_offset(pgd, addr); in vmemmap_pud_populate() local 202 if (pud_none(*pud)) { in vmemmap_pud_populate() 206 pud_populate(&init_mm, pud, p); in vmemmap_pud_populate() 208 return pud; in vmemmap_pud_populate() 228 pud_t *pud; in vmemmap_populate_basepages() local 236 pud = vmemmap_pud_populate(pgd, addr, node); in vmemmap_populate_basepages() 237 if (!pud) in vmemmap_populate_basepages() 239 pmd = vmemmap_pmd_populate(pud, addr, node); in vmemmap_populate_basepages()
|
D | gup.c | 229 pud_t *pud; in follow_page_mask() local 247 pud = pud_offset(pgd, address); in follow_page_mask() 248 if (pud_none(*pud)) in follow_page_mask() 250 if (pud_huge(*pud) && vma->vm_flags & VM_HUGETLB) { in follow_page_mask() 251 page = follow_huge_pud(mm, address, pud, flags); in follow_page_mask() 256 if (unlikely(pud_bad(*pud))) in follow_page_mask() 259 pmd = pmd_offset(pud, address); in follow_page_mask() 320 pud_t *pud; in get_gate_page() local 333 pud = pud_offset(pgd, address); in get_gate_page() 334 BUG_ON(pud_none(*pud)); in get_gate_page() [all …]
|
D | memory.c | 413 static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud, in free_pmd_range() argument 422 pmd = pmd_offset(pud, addr); in free_pmd_range() 441 pmd = pmd_offset(pud, start); in free_pmd_range() 442 pud_clear(pud); in free_pmd_range() 451 pud_t *pud; in free_pud_range() local 456 pud = pud_offset(pgd, addr); in free_pud_range() 459 if (pud_none_or_clear_bad(pud)) in free_pud_range() 461 free_pmd_range(tlb, pud, addr, next, floor, ceiling); in free_pud_range() 462 } while (pud++, addr = next, addr != end); in free_pud_range() 475 pud = pud_offset(pgd, start); in free_pud_range() [all …]
|
D | pagewalk.c | 27 static int walk_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, in walk_pmd_range() argument 34 pmd = pmd_offset(pud, addr); in walk_pmd_range() 75 pud_t *pud; in walk_pud_range() local 79 pud = pud_offset(pgd, addr); in walk_pud_range() 82 if (pud_none_or_clear_bad(pud)) { in walk_pud_range() 90 err = walk_pmd_range(pud, addr, next, walk); in walk_pud_range() 93 } while (pud++, addr = next, addr != end); in walk_pud_range()
|
D | mremap.c | 34 pud_t *pud; in get_old_pmd() local 41 pud = pud_offset(pgd, addr); in get_old_pmd() 42 if (pud_none_or_clear_bad(pud)) in get_old_pmd() 45 pmd = pmd_offset(pud, addr); in get_old_pmd() 56 pud_t *pud; in alloc_new_pmd() local 60 pud = pud_alloc(mm, pgd, addr); in alloc_new_pmd() 61 if (!pud) in alloc_new_pmd() 64 pmd = pmd_alloc(mm, pud, addr); in alloc_new_pmd()
|
D | mprotect.c | 140 pud_t *pud, unsigned long addr, unsigned long end, in change_pmd_range() argument 150 pmd = pmd_offset(pud, addr); in change_pmd_range() 203 pud_t *pud; in change_pud_range() local 207 pud = pud_offset(pgd, addr); in change_pud_range() 210 if (pud_none_or_clear_bad(pud)) in change_pud_range() 212 pages += change_pmd_range(vma, pud, addr, next, newprot, in change_pud_range() 214 } while (pud++, addr = next, addr != end); in change_pud_range()
|
D | vmalloc.c | 73 static void vunmap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end) in vunmap_pmd_range() argument 78 pmd = pmd_offset(pud, addr); in vunmap_pmd_range() 91 pud_t *pud; in vunmap_pud_range() local 94 pud = pud_offset(pgd, addr); in vunmap_pud_range() 97 if (pud_clear_huge(pud)) in vunmap_pud_range() 99 if (pud_none_or_clear_bad(pud)) in vunmap_pud_range() 101 vunmap_pmd_range(pud, addr, next); in vunmap_pud_range() 102 } while (pud++, addr = next, addr != end); in vunmap_pud_range() 146 static int vmap_pmd_range(pud_t *pud, unsigned long addr, in vmap_pmd_range() argument 152 pmd = pmd_alloc(&init_mm, pud, addr); in vmap_pmd_range() [all …]
|
D | userfaultfd.c | 127 pud_t *pud; in mm_alloc_pmd() local 131 pud = pud_alloc(mm, pgd, address); in mm_alloc_pmd() 132 if (pud) in mm_alloc_pmd() 138 pmd = pmd_alloc(mm, pud, address); in mm_alloc_pmd()
|
D | pgtable-generic.c | 25 void pud_clear_bad(pud_t *pud) in pud_clear_bad() argument 27 pud_ERROR(*pud); in pud_clear_bad() 28 pud_clear(pud); in pud_clear_bad()
|
D | hugetlb.c | 4329 pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) in huge_pmd_share() argument 4342 return (pte_t *)pmd_alloc(mm, pud, addr); in huge_pmd_share() 4364 if (pud_none(*pud)) { in huge_pmd_share() 4365 pud_populate(mm, pud, in huge_pmd_share() 4373 pte = (pte_t *)pmd_alloc(mm, pud, addr); in huge_pmd_share() 4393 pud_t *pud = pud_offset(pgd, *addr); in huge_pmd_unshare() local 4399 pud_clear(pud); in huge_pmd_unshare() 4407 pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) in huge_pmd_share() argument 4424 pud_t *pud; in huge_pte_alloc() local 4428 pud = pud_alloc(mm, pgd, addr); in huge_pte_alloc() [all …]
|
D | rmap.c | 724 pud_t *pud; in mm_find_pmd() local 732 pud = pud_offset(pgd, address); in mm_find_pmd() 733 if (!pud_present(*pud)) in mm_find_pmd() 736 pmd = pmd_offset(pud, address); in mm_find_pmd() 837 pud_t *pud; in page_check_address_transhuge() local 856 pud = pud_offset(pgd, address); in page_check_address_transhuge() 857 if (!pud_present(*pud)) in page_check_address_transhuge() 859 pmd = pmd_offset(pud, address); in page_check_address_transhuge()
|
D | swapfile.c | 1241 static inline int unuse_pmd_range(struct vm_area_struct *vma, pud_t *pud, in unuse_pmd_range() argument 1249 pmd = pmd_offset(pud, addr); in unuse_pmd_range() 1265 pud_t *pud; in unuse_pud_range() local 1269 pud = pud_offset(pgd, addr); in unuse_pud_range() 1272 if (pud_none_or_clear_bad(pud)) in unuse_pud_range() 1274 ret = unuse_pmd_range(vma, pud, addr, next, entry, page); in unuse_pud_range() 1277 } while (pud++, addr = next, addr != end); in unuse_pud_range()
|
D | huge_memory.c | 1784 pud_t *pud; in split_huge_pmd_address() local 1791 pud = pud_offset(pgd, address); in split_huge_pmd_address() 1792 if (!pud_present(*pud)) in split_huge_pmd_address() 1795 pmd = pmd_offset(pud, address); in split_huge_pmd_address()
|