/mm/ |
D | memory.c | 564 spinlock_t *ptl; in __pte_alloc() local 585 ptl = pmd_lock(mm, pmd); in __pte_alloc() 593 spin_unlock(ptl); in __pte_alloc() 1081 spinlock_t *ptl; in zap_pte_range() local 1087 start_pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in zap_pte_range() 1189 pte_unmap_unlock(start_pte, ptl); in zap_pte_range() 1443 spinlock_t **ptl) in __get_locked_pte() argument 1451 return pte_alloc_map_lock(mm, pmd, addr, ptl); in __get_locked_pte() 1470 spinlock_t *ptl; in insert_page() local 1477 pte = get_locked_pte(mm, addr, &ptl); in insert_page() [all …]
|
D | huge_memory.c | 720 spinlock_t *ptl; in __do_huge_pmd_anonymous_page() local 741 ptl = pmd_lock(mm, pmd); in __do_huge_pmd_anonymous_page() 743 spin_unlock(ptl); in __do_huge_pmd_anonymous_page() 758 spin_unlock(ptl); in __do_huge_pmd_anonymous_page() 810 spinlock_t *ptl; in do_huge_pmd_anonymous_page() local 823 ptl = pmd_lock(mm, pmd); in do_huge_pmd_anonymous_page() 826 spin_unlock(ptl); in do_huge_pmd_anonymous_page() 930 spinlock_t *ptl; in huge_pmd_set_accessed() local 934 ptl = pmd_lock(mm, pmd); in huge_pmd_set_accessed() 944 spin_unlock(ptl); in huge_pmd_set_accessed() [all …]
|
D | rmap.c | 654 spinlock_t *ptl; in __page_check_address() local 662 ptl = huge_pte_lockptr(page_hstate(page), mm, pte); in __page_check_address() 677 ptl = pte_lockptr(mm, pmd); in __page_check_address() 679 spin_lock(ptl); in __page_check_address() 681 *ptlp = ptl; in __page_check_address() 684 pte_unmap_unlock(pte, ptl); in __page_check_address() 701 spinlock_t *ptl; in page_mapped_in_vma() local 706 pte = page_check_address(page, vma->vm_mm, address, &ptl, 1); in page_mapped_in_vma() 709 pte_unmap_unlock(pte, ptl); in page_mapped_in_vma() 727 spinlock_t *ptl; in page_referenced_one() local [all …]
|
D | hugetlb.c | 2659 spinlock_t *ptl; in __unmap_hugepage_range() local 2678 ptl = huge_pte_lock(h, mm, ptep); in __unmap_hugepage_range() 2721 spin_unlock(ptl); in __unmap_hugepage_range() 2726 spin_unlock(ptl); in __unmap_hugepage_range() 2730 spin_unlock(ptl); in __unmap_hugepage_range() 2843 struct page *pagecache_page, spinlock_t *ptl) in hugetlb_cow() argument 2881 spin_unlock(ptl); in hugetlb_cow() 2897 spin_lock(ptl); in hugetlb_cow() 2936 spin_lock(ptl); in hugetlb_cow() 2950 spin_unlock(ptl); in hugetlb_cow() [all …]
|
D | migrate.c | 113 spinlock_t *ptl; in remove_migration_pte() local 119 ptl = huge_pte_lockptr(hstate_vma(vma), mm, ptep); in remove_migration_pte() 132 ptl = pte_lockptr(mm, pmd); in remove_migration_pte() 135 spin_lock(ptl); in remove_migration_pte() 177 pte_unmap_unlock(ptep, ptl); in remove_migration_pte() 234 spinlock_t *ptl) in __migration_entry_wait() argument 240 spin_lock(ptl); in __migration_entry_wait() 260 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait() 265 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait() 271 spinlock_t *ptl = pte_lockptr(mm, pmd); in migration_entry_wait() local [all …]
|
D | gup.c | 50 spinlock_t *ptl; in follow_page_pte() local 57 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); in follow_page_pte() 73 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 80 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 127 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 130 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 134 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 160 spinlock_t *ptl; in follow_page_mask() local 204 ptl = pmd_lock(mm, pmd); in follow_page_mask() 207 spin_unlock(ptl); in follow_page_mask() [all …]
|
D | mprotect.c | 40 unsigned long addr, int prot_numa, spinlock_t **ptl) in lock_pte_protection() argument 47 return pte_offset_map_lock(vma->vm_mm, pmd, addr, ptl); in lock_pte_protection() 55 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, ptl); in lock_pte_protection() 66 spinlock_t *ptl; in change_pte_range() local 69 pte = lock_pte_protection(vma, pmd, addr, prot_numa, &ptl); in change_pte_range() 128 pte_unmap_unlock(pte - 1, ptl); in change_pte_range()
|
D | fremap.c | 77 spinlock_t *ptl; in install_file_pte() local 79 pte = get_locked_pte(mm, addr, &ptl); in install_file_pte() 96 pte_unmap_unlock(pte, ptl); in install_file_pte()
|
D | mincore.c | 121 spinlock_t *ptl; in mincore_pte_range() local 124 ptep = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in mincore_pte_range() 156 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
|
D | madvise.c | 153 spinlock_t *ptl; in swapin_walk_pmd_entry() local 155 orig_pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl); in swapin_walk_pmd_entry() 157 pte_unmap_unlock(orig_pte, ptl); in swapin_walk_pmd_entry()
|
D | filemap_xip.c | 173 spinlock_t *ptl; in __xip_unmap() local 191 pte = page_check_address(page, mm, address, &ptl, 1); in __xip_unmap() 199 pte_unmap_unlock(pte, ptl); in __xip_unmap()
|
D | ksm.c | 862 spinlock_t *ptl; in write_protect_page() local 878 ptep = page_check_address(page, mm, addr, &ptl, 0); in write_protect_page() 914 pte_unmap_unlock(ptep, ptl); in write_protect_page() 936 spinlock_t *ptl; in replace_page() local 954 ptep = pte_offset_map_lock(mm, pmd, addr, &ptl); in replace_page() 956 pte_unmap_unlock(ptep, ptl); in replace_page() 972 pte_unmap_unlock(ptep, ptl); in replace_page()
|
D | mlock.c | 421 spinlock_t *ptl; in __munlock_pagevec_fill() local 428 pte = get_locked_pte(vma->vm_mm, start, &ptl); in __munlock_pagevec_fill() 457 pte_unmap_unlock(pte, ptl); in __munlock_pagevec_fill()
|
D | mempolicy.c | 491 spinlock_t *ptl; in queue_pages_pte_range() local 493 orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in queue_pages_pte_range() 518 pte_unmap_unlock(orig_pte, ptl); in queue_pages_pte_range() 529 spinlock_t *ptl; in queue_pages_hugetlb_pmd_range() local 532 ptl = huge_pte_lock(hstate_vma(vma), vma->vm_mm, (pte_t *)pmd); in queue_pages_hugetlb_pmd_range() 545 spin_unlock(ptl); in queue_pages_hugetlb_pmd_range()
|
D | memcontrol.c | 5841 spinlock_t *ptl; in mem_cgroup_count_precharge_pte_range() local 5843 if (pmd_trans_huge_lock(pmd, vma, &ptl) == 1) { in mem_cgroup_count_precharge_pte_range() 5846 spin_unlock(ptl); in mem_cgroup_count_precharge_pte_range() 5852 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in mem_cgroup_count_precharge_pte_range() 5856 pte_unmap_unlock(pte - 1, ptl); in mem_cgroup_count_precharge_pte_range() 6018 spinlock_t *ptl; in mem_cgroup_move_charge_pte_range() local 6034 if (pmd_trans_huge_lock(pmd, vma, &ptl) == 1) { in mem_cgroup_move_charge_pte_range() 6036 spin_unlock(ptl); in mem_cgroup_move_charge_pte_range() 6053 spin_unlock(ptl); in mem_cgroup_move_charge_pte_range() 6060 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in mem_cgroup_move_charge_pte_range() [all …]
|
D | swapfile.c | 1139 spinlock_t *ptl; in unuse_pte() local 1153 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in unuse_pte() 1180 pte_unmap_unlock(pte, ptl); in unuse_pte()
|