Home
last modified time | relevance | path

Searched refs:ptl (Results 1 – 23 of 23) sorted by relevance

/mm/
Dpage_vma_mapped.c51 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte()
52 spin_lock(pvmw->ptl); in map_pte()
177 pvmw->ptl = huge_pte_lockptr(page_hstate(page), mm, pvmw->pte); in page_vma_mapped_walk()
178 spin_lock(pvmw->ptl); in page_vma_mapped_walk()
222 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk()
244 spin_unlock(pvmw->ptl); in page_vma_mapped_walk()
245 pvmw->ptl = NULL; in page_vma_mapped_walk()
254 spinlock_t *ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() local
256 spin_unlock(ptl); in page_vma_mapped_walk()
273 if (pvmw->ptl) { in page_vma_mapped_walk()
[all …]
Dhuge_memory.c628 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
640 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page()
657 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page()
664 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page()
752 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page()
757 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page()
760 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page()
768 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page()
771 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page()
792 spinlock_t *ptl; in insert_pfn_pmd() local
[all …]
Dmemory.c280 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); in free_pte_range() local
281 spin_unlock(ptl); in free_pte_range()
499 spinlock_t *ptl; in __pte_alloc() local
519 ptl = pmd_lock(mm, pmd); in __pte_alloc()
525 spin_unlock(ptl); in __pte_alloc()
1384 spinlock_t *ptl; in zap_pte_range() local
1394 start_pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in zap_pte_range()
1499 pte_unmap_unlock(start_pte, ptl); in zap_pte_range()
1541 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); in zap_pmd_range() local
1547 spin_unlock(ptl); in zap_pmd_range()
[all …]
Dmadvise.c210 spinlock_t *ptl; in swapin_walk_pmd_entry() local
212 orig_pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl); in swapin_walk_pmd_entry()
214 pte_unmap_unlock(orig_pte, ptl); in swapin_walk_pmd_entry()
328 spinlock_t *ptl; in madvise_cold_or_pageout_pte_range() local
341 ptl = pmd_trans_huge_lock(pmd, vma); in madvise_cold_or_pageout_pte_range()
342 if (!ptl) in madvise_cold_or_pageout_pte_range()
368 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range()
398 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range()
409 orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in madvise_cold_or_pageout_pte_range()
439 pte_unmap_unlock(orig_pte, ptl); in madvise_cold_or_pageout_pte_range()
[all …]
Dmincore.c102 spinlock_t *ptl; in mincore_pte_range() local
108 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range()
109 if (ptl) { in mincore_pte_range()
111 spin_unlock(ptl); in mincore_pte_range()
120 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in mincore_pte_range()
150 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
Dhmm.c418 spinlock_t *ptl = pud_trans_huge_lock(pudp, walk->vma); in hmm_vma_walk_pud() local
420 if (!ptl) in hmm_vma_walk_pud()
428 spin_unlock(ptl); in hmm_vma_walk_pud()
439 spin_unlock(ptl); in hmm_vma_walk_pud()
451 spin_unlock(ptl); in hmm_vma_walk_pud()
465 spin_unlock(ptl); in hmm_vma_walk_pud()
484 spinlock_t *ptl; in hmm_vma_walk_hugetlb_entry() local
487 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry()
497 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
505 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
Dhugetlb.c4459 spinlock_t *ptl; in __unmap_hugepage_range() local
4490 ptl = huge_pte_lock(h, mm, ptep); in __unmap_hugepage_range()
4492 spin_unlock(ptl); in __unmap_hugepage_range()
4500 spin_unlock(ptl); in __unmap_hugepage_range()
4510 spin_unlock(ptl); in __unmap_hugepage_range()
4522 spin_unlock(ptl); in __unmap_hugepage_range()
4541 spin_unlock(ptl); in __unmap_hugepage_range()
4662 struct page *pagecache_page, spinlock_t *ptl) in hugetlb_cow() argument
4703 spin_unlock(ptl); in hugetlb_cow()
4739 spin_lock(ptl); in hugetlb_cow()
[all …]
Dmigrate.c291 spinlock_t *ptl) in __migration_entry_wait() argument
297 spin_lock(ptl); in __migration_entry_wait()
316 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait()
320 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait()
326 spinlock_t *ptl = pte_lockptr(mm, pmd); in migration_entry_wait() local
328 __migration_entry_wait(mm, ptep, ptl); in migration_entry_wait()
334 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), mm, pte); in migration_entry_wait_huge() local
335 __migration_entry_wait(mm, pte, ptl); in migration_entry_wait_huge()
341 spinlock_t *ptl; in pmd_migration_entry_wait() local
344 ptl = pmd_lock(mm, pmd); in pmd_migration_entry_wait()
[all …]
Dgup.c504 spinlock_t *ptl; in follow_page_pte() local
528 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); in follow_page_pte()
544 pte_unmap_unlock(ptep, ptl); in follow_page_pte()
551 pte_unmap_unlock(ptep, ptl); in follow_page_pte()
639 pte_unmap_unlock(ptep, ptl); in follow_page_pte()
642 pte_unmap_unlock(ptep, ptl); in follow_page_pte()
654 spinlock_t *ptl; in follow_pmd_mask() local
698 ptl = pmd_lock(mm, pmd); in follow_pmd_mask()
700 spin_unlock(ptl); in follow_pmd_mask()
711 ptl = pmd_lock(mm, pmd); in follow_pmd_mask()
[all …]
Duserfaultfd.c67 spinlock_t *ptl; in mfill_atomic_install_pte() local
83 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_install_pte()
119 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_install_pte()
189 spinlock_t *ptl; in mfill_zeropage_pte() local
196 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_zeropage_pte()
214 pte_unmap_unlock(dst_pte, ptl); in mfill_zeropage_pte()
Dkhugepaged.c746 spinlock_t *ptl, in __collapse_huge_page_copy() argument
762 spin_lock(ptl); in __collapse_huge_page_copy()
768 spin_unlock(ptl); in __collapse_huge_page_copy()
780 spin_lock(ptl); in __collapse_huge_page_copy()
787 spin_unlock(ptl); in __collapse_huge_page_copy()
1240 spinlock_t *ptl; in khugepaged_scan_pmd() local
1253 pte = pte_offset_map_lock(mm, pmd, address, &ptl); in khugepaged_scan_pmd()
1373 pte_unmap_unlock(pte, ptl); in khugepaged_scan_pmd()
1447 spinlock_t *ptl; in collapse_pte_mapped_thp() local
1490 start_pte = pte_offset_map_lock(mm, pmd, haddr, &ptl); in collapse_pte_mapped_thp()
[all …]
Dmempolicy.c450 static int queue_pages_pmd(pmd_t *pmd, spinlock_t *ptl, unsigned long addr, in queue_pages_pmd() argument
452 __releases(ptl) in queue_pages_pmd()
465 spin_unlock(ptl); in queue_pages_pmd()
483 spin_unlock(ptl); in queue_pages_pmd()
510 spinlock_t *ptl; in queue_pages_pte_range() local
512 ptl = pmd_trans_huge_lock(pmd, vma); in queue_pages_pte_range()
513 if (ptl) { in queue_pages_pte_range()
514 ret = queue_pages_pmd(pmd, ptl, addr, end, walk); in queue_pages_pte_range()
523 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in queue_pages_pte_range()
555 pte_unmap_unlock(mapped_pte, ptl); in queue_pages_pte_range()
[all …]
Ddebug_vm_pgtable.c1241 spinlock_t *ptl = NULL; in debug_vm_pgtable() local
1305 args.ptep = pte_offset_map_lock(args.mm, args.pmdp, args.vaddr, &ptl); in debug_vm_pgtable()
1308 pte_unmap_unlock(args.ptep, ptl); in debug_vm_pgtable()
1310 ptl = pmd_lock(args.mm, args.pmdp); in debug_vm_pgtable()
1315 spin_unlock(ptl); in debug_vm_pgtable()
1317 ptl = pud_lock(args.mm, args.pudp); in debug_vm_pgtable()
1322 spin_unlock(ptl); in debug_vm_pgtable()
Dpagewalk.c46 spinlock_t *ptl; in walk_pte_range() local
53 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in walk_pte_range()
55 pte_unmap_unlock(pte, ptl); in walk_pte_range()
Dmemory-failure.c635 spinlock_t *ptl; in hwpoison_pte_range() local
637 ptl = pmd_trans_huge_lock(pmdp, walk->vma); in hwpoison_pte_range()
638 if (ptl) { in hwpoison_pte_range()
640 spin_unlock(ptl); in hwpoison_pte_range()
648 addr, &ptl); in hwpoison_pte_range()
655 pte_unmap_unlock(mapped_pte, ptl); in hwpoison_pte_range()
Dmlock.c356 spinlock_t *ptl; in __munlock_pagevec_fill() local
363 pte = get_locked_pte(vma->vm_mm, start, &ptl); in __munlock_pagevec_fill()
400 pte_unmap_unlock(pte, ptl); in __munlock_pagevec_fill()
Dmprotect.c44 spinlock_t *ptl; in change_pte_range() local
66 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in change_pte_range()
193 pte_unmap_unlock(pte - 1, ptl); in change_pte_range()
Dmemcontrol.c5884 spinlock_t *ptl; in mem_cgroup_count_precharge_pte_range() local
5886 ptl = pmd_trans_huge_lock(pmd, vma); in mem_cgroup_count_precharge_pte_range()
5887 if (ptl) { in mem_cgroup_count_precharge_pte_range()
5895 spin_unlock(ptl); in mem_cgroup_count_precharge_pte_range()
5901 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in mem_cgroup_count_precharge_pte_range()
5905 pte_unmap_unlock(pte - 1, ptl); in mem_cgroup_count_precharge_pte_range()
6081 spinlock_t *ptl; in mem_cgroup_move_charge_pte_range() local
6086 ptl = pmd_trans_huge_lock(pmd, vma); in mem_cgroup_move_charge_pte_range()
6087 if (ptl) { in mem_cgroup_move_charge_pte_range()
6089 spin_unlock(ptl); in mem_cgroup_move_charge_pte_range()
[all …]
Dksm.c1121 spinlock_t *ptl; in replace_page() local
1138 ptep = pte_offset_map_lock(mm, pmd, addr, &ptl); in replace_page()
1140 pte_unmap_unlock(ptep, ptl); in replace_page()
1179 pte_unmap_unlock(ptep, ptl); in replace_page()
Dvmscan.c3691 spinlock_t *ptl; in walk_pte_range() local
3702 ptl = pte_lockptr(args->mm, pmd); in walk_pte_range()
3703 if (!spin_trylock(ptl)) in walk_pte_range()
3752 spin_unlock(ptl); in walk_pte_range()
3763 spinlock_t *ptl; in walk_pmd_range_locked() local
3785 ptl = pmd_lockptr(args->mm, pmd); in walk_pmd_range_locked()
3786 if (!spin_trylock(ptl)) in walk_pmd_range_locked()
3829 spin_unlock(ptl); in walk_pmd_range_locked()
4346 lockdep_assert_held(pvmw->ptl); in lru_gen_look_around()
4349 if (spin_is_contended(pvmw->ptl)) in lru_gen_look_around()
Dfilemap.c3280 vmf->ptl = pmd_lock(mm, vmf->pmd); in filemap_map_pmd()
3286 spin_unlock(vmf->ptl); in filemap_map_pmd()
3417 pte_unmap_unlock(vmf->pte, vmf->ptl); in filemap_map_pages()
Dswapfile.c1904 spinlock_t *ptl; in unuse_pte() local
1913 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in unuse_pte()
1932 pte_unmap_unlock(pte, ptl); in unuse_pte()
/mm/damon/
Dvaddr.c374 spinlock_t *ptl; in damon_mkold_pmd_entry() local
377 ptl = pmd_lock(walk->mm, pmd); in damon_mkold_pmd_entry()
379 spin_unlock(ptl); in damon_mkold_pmd_entry()
385 spin_unlock(ptl); in damon_mkold_pmd_entry()
388 spin_unlock(ptl); in damon_mkold_pmd_entry()
393 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in damon_mkold_pmd_entry()
398 pte_unmap_unlock(pte, ptl); in damon_mkold_pmd_entry()
439 spinlock_t *ptl; in damon_mkold_hugetlb_entry() local
442 ptl = huge_pte_lock(h, walk->mm, pte); in damon_mkold_hugetlb_entry()
450 spin_unlock(ptl); in damon_mkold_hugetlb_entry()
[all …]