Home
last modified time | relevance | path

Searched refs:pud (Results 1 – 14 of 14) sorted by relevance

/mm/kasan/
Dkasan_init.c62 static void __init zero_pmd_populate(pud_t *pud, unsigned long addr, in zero_pmd_populate() argument
65 pmd_t *pmd = pmd_offset(pud, addr); in zero_pmd_populate()
87 pud_t *pud = pud_offset(pgd, addr); in zero_pud_populate() local
95 pud_populate(&init_mm, pud, kasan_zero_pmd); in zero_pud_populate()
96 pmd = pmd_offset(pud, addr); in zero_pud_populate()
101 if (pud_none(*pud)) { in zero_pud_populate()
102 pud_populate(&init_mm, pud, in zero_pud_populate()
105 zero_pmd_populate(pud, addr, next); in zero_pud_populate()
106 } while (pud++, addr = next, addr != end); in zero_pud_populate()
127 pud_t *pud; in kasan_populate_zero_shadow() local
[all …]
/mm/
Dsparse-vmemmap.c187 pmd_t * __meminit vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node) in vmemmap_pmd_populate() argument
189 pmd_t *pmd = pmd_offset(pud, addr); in vmemmap_pmd_populate()
201 pud_t *pud = pud_offset(pgd, addr); in vmemmap_pud_populate() local
202 if (pud_none(*pud)) { in vmemmap_pud_populate()
206 pud_populate(&init_mm, pud, p); in vmemmap_pud_populate()
208 return pud; in vmemmap_pud_populate()
228 pud_t *pud; in vmemmap_populate_basepages() local
236 pud = vmemmap_pud_populate(pgd, addr, node); in vmemmap_populate_basepages()
237 if (!pud) in vmemmap_populate_basepages()
239 pmd = vmemmap_pmd_populate(pud, addr, node); in vmemmap_populate_basepages()
Dgup.c229 pud_t *pud; in follow_page_mask() local
247 pud = pud_offset(pgd, address); in follow_page_mask()
248 if (pud_none(*pud)) in follow_page_mask()
250 if (pud_huge(*pud) && vma->vm_flags & VM_HUGETLB) { in follow_page_mask()
251 page = follow_huge_pud(mm, address, pud, flags); in follow_page_mask()
256 if (unlikely(pud_bad(*pud))) in follow_page_mask()
259 pmd = pmd_offset(pud, address); in follow_page_mask()
320 pud_t *pud; in get_gate_page() local
333 pud = pud_offset(pgd, address); in get_gate_page()
334 BUG_ON(pud_none(*pud)); in get_gate_page()
[all …]
Dmemory.c413 static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud, in free_pmd_range() argument
422 pmd = pmd_offset(pud, addr); in free_pmd_range()
441 pmd = pmd_offset(pud, start); in free_pmd_range()
442 pud_clear(pud); in free_pmd_range()
451 pud_t *pud; in free_pud_range() local
456 pud = pud_offset(pgd, addr); in free_pud_range()
459 if (pud_none_or_clear_bad(pud)) in free_pud_range()
461 free_pmd_range(tlb, pud, addr, next, floor, ceiling); in free_pud_range()
462 } while (pud++, addr = next, addr != end); in free_pud_range()
475 pud = pud_offset(pgd, start); in free_pud_range()
[all …]
Dpagewalk.c27 static int walk_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, in walk_pmd_range() argument
34 pmd = pmd_offset(pud, addr); in walk_pmd_range()
75 pud_t *pud; in walk_pud_range() local
79 pud = pud_offset(pgd, addr); in walk_pud_range()
82 if (pud_none_or_clear_bad(pud)) { in walk_pud_range()
90 err = walk_pmd_range(pud, addr, next, walk); in walk_pud_range()
93 } while (pud++, addr = next, addr != end); in walk_pud_range()
Dmremap.c34 pud_t *pud; in get_old_pmd() local
41 pud = pud_offset(pgd, addr); in get_old_pmd()
42 if (pud_none_or_clear_bad(pud)) in get_old_pmd()
45 pmd = pmd_offset(pud, addr); in get_old_pmd()
56 pud_t *pud; in alloc_new_pmd() local
60 pud = pud_alloc(mm, pgd, addr); in alloc_new_pmd()
61 if (!pud) in alloc_new_pmd()
64 pmd = pmd_alloc(mm, pud, addr); in alloc_new_pmd()
Dmprotect.c140 pud_t *pud, unsigned long addr, unsigned long end, in change_pmd_range() argument
150 pmd = pmd_offset(pud, addr); in change_pmd_range()
203 pud_t *pud; in change_pud_range() local
207 pud = pud_offset(pgd, addr); in change_pud_range()
210 if (pud_none_or_clear_bad(pud)) in change_pud_range()
212 pages += change_pmd_range(vma, pud, addr, next, newprot, in change_pud_range()
214 } while (pud++, addr = next, addr != end); in change_pud_range()
Dvmalloc.c73 static void vunmap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end) in vunmap_pmd_range() argument
78 pmd = pmd_offset(pud, addr); in vunmap_pmd_range()
91 pud_t *pud; in vunmap_pud_range() local
94 pud = pud_offset(pgd, addr); in vunmap_pud_range()
97 if (pud_clear_huge(pud)) in vunmap_pud_range()
99 if (pud_none_or_clear_bad(pud)) in vunmap_pud_range()
101 vunmap_pmd_range(pud, addr, next); in vunmap_pud_range()
102 } while (pud++, addr = next, addr != end); in vunmap_pud_range()
146 static int vmap_pmd_range(pud_t *pud, unsigned long addr, in vmap_pmd_range() argument
152 pmd = pmd_alloc(&init_mm, pud, addr); in vmap_pmd_range()
[all …]
Duserfaultfd.c127 pud_t *pud; in mm_alloc_pmd() local
131 pud = pud_alloc(mm, pgd, address); in mm_alloc_pmd()
132 if (pud) in mm_alloc_pmd()
138 pmd = pmd_alloc(mm, pud, address); in mm_alloc_pmd()
Dpgtable-generic.c25 void pud_clear_bad(pud_t *pud) in pud_clear_bad() argument
27 pud_ERROR(*pud); in pud_clear_bad()
28 pud_clear(pud); in pud_clear_bad()
Dhugetlb.c4329 pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) in huge_pmd_share() argument
4342 return (pte_t *)pmd_alloc(mm, pud, addr); in huge_pmd_share()
4364 if (pud_none(*pud)) { in huge_pmd_share()
4365 pud_populate(mm, pud, in huge_pmd_share()
4373 pte = (pte_t *)pmd_alloc(mm, pud, addr); in huge_pmd_share()
4393 pud_t *pud = pud_offset(pgd, *addr); in huge_pmd_unshare() local
4399 pud_clear(pud); in huge_pmd_unshare()
4407 pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) in huge_pmd_share() argument
4424 pud_t *pud; in huge_pte_alloc() local
4428 pud = pud_alloc(mm, pgd, addr); in huge_pte_alloc()
[all …]
Drmap.c724 pud_t *pud; in mm_find_pmd() local
732 pud = pud_offset(pgd, address); in mm_find_pmd()
733 if (!pud_present(*pud)) in mm_find_pmd()
736 pmd = pmd_offset(pud, address); in mm_find_pmd()
837 pud_t *pud; in page_check_address_transhuge() local
856 pud = pud_offset(pgd, address); in page_check_address_transhuge()
857 if (!pud_present(*pud)) in page_check_address_transhuge()
859 pmd = pmd_offset(pud, address); in page_check_address_transhuge()
Dswapfile.c1241 static inline int unuse_pmd_range(struct vm_area_struct *vma, pud_t *pud, in unuse_pmd_range() argument
1249 pmd = pmd_offset(pud, addr); in unuse_pmd_range()
1265 pud_t *pud; in unuse_pud_range() local
1269 pud = pud_offset(pgd, addr); in unuse_pud_range()
1272 if (pud_none_or_clear_bad(pud)) in unuse_pud_range()
1274 ret = unuse_pmd_range(vma, pud, addr, next, entry, page); in unuse_pud_range()
1277 } while (pud++, addr = next, addr != end); in unuse_pud_range()
Dhuge_memory.c1784 pud_t *pud; in split_huge_pmd_address() local
1791 pud = pud_offset(pgd, address); in split_huge_pmd_address()
1792 if (!pud_present(*pud)) in split_huge_pmd_address()
1795 pmd = pmd_offset(pud, address); in split_huge_pmd_address()