• Home
  • Raw
  • Download

Lines Matching refs:addr

504 static int smaps_pte_hole(unsigned long addr, unsigned long end,  in smaps_pte_hole()  argument
511 linear_page_index(vma, addr), in smaps_pte_hole()
520 static void smaps_pte_hole_lookup(unsigned long addr, struct mm_walk *walk) in smaps_pte_hole_lookup() argument
525 smaps_pte_hole(addr, addr + PAGE_SIZE, 0, walk); in smaps_pte_hole_lookup()
530 static void smaps_pte_entry(pte_t *pte, unsigned long addr, in smaps_pte_entry() argument
540 page = vm_normal_page(vma, addr, *pte); in smaps_pte_entry()
568 smaps_pte_hole_lookup(addr, walk); in smaps_pte_entry()
579 static void smaps_pmd_entry(pmd_t *pmd, unsigned long addr, in smaps_pmd_entry() argument
590 page = follow_trans_huge_pmd(vma, addr, pmd, FOLL_DUMP); in smaps_pmd_entry()
614 static void smaps_pmd_entry(pmd_t *pmd, unsigned long addr, in smaps_pmd_entry() argument
620 static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, in smaps_pte_range() argument
629 smaps_pmd_entry(pmd, addr, walk); in smaps_pte_range()
641 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range()
642 for (; addr != end; pte++, addr += PAGE_SIZE) in smaps_pte_range()
643 smaps_pte_entry(pte, addr, walk); in smaps_pte_range()
731 unsigned long addr, unsigned long end, in smaps_hugetlb_range() argument
739 page = vm_normal_page(vma, addr, *pte); in smaps_hugetlb_range()
1090 static inline bool pte_is_pinned(struct vm_area_struct *vma, unsigned long addr, pte_t pte) in pte_is_pinned() argument
1100 page = vm_normal_page(vma, addr, pte); in pte_is_pinned()
1107 unsigned long addr, pte_t *pte) in clear_soft_dirty() argument
1120 if (pte_is_pinned(vma, addr, ptent)) in clear_soft_dirty()
1122 old_pte = ptep_modify_prot_start(vma, addr, pte); in clear_soft_dirty()
1125 ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent); in clear_soft_dirty()
1128 set_pte_at(vma->vm_mm, addr, pte, ptent); in clear_soft_dirty()
1133 unsigned long addr, pte_t *pte) in clear_soft_dirty() argument
1140 unsigned long addr, pmd_t *pmdp) in clear_soft_dirty_pmd() argument
1146 old = pmdp_invalidate(vma, addr, pmdp); in clear_soft_dirty_pmd()
1155 set_pmd_at(vma->vm_mm, addr, pmdp, pmd); in clear_soft_dirty_pmd()
1158 set_pmd_at(vma->vm_mm, addr, pmdp, pmd); in clear_soft_dirty_pmd()
1163 unsigned long addr, pmd_t *pmdp) in clear_soft_dirty_pmd() argument
1168 static int clear_refs_pte_range(pmd_t *pmd, unsigned long addr, in clear_refs_pte_range() argument
1180 clear_soft_dirty_pmd(vma, addr, pmd); in clear_refs_pte_range()
1190 pmdp_test_and_clear_young(vma, addr, pmd); in clear_refs_pte_range()
1201 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in clear_refs_pte_range()
1202 for (; addr != end; pte++, addr += PAGE_SIZE) { in clear_refs_pte_range()
1206 clear_soft_dirty(vma, addr, pte); in clear_refs_pte_range()
1213 page = vm_normal_page(vma, addr, ptent); in clear_refs_pte_range()
1218 ptep_test_and_clear_young(vma, addr, pte); in clear_refs_pte_range()
1366 static int add_to_pagemap(unsigned long addr, pagemap_entry_t *pme, in add_to_pagemap() argument
1379 unsigned long addr = start; in pagemap_pte_hole() local
1382 while (addr < end) { in pagemap_pte_hole()
1383 struct vm_area_struct *vma = find_vma(walk->mm, addr); in pagemap_pte_hole()
1393 for (; addr < hole_end; addr += PAGE_SIZE) { in pagemap_pte_hole()
1394 err = add_to_pagemap(addr, &pme, pm); in pagemap_pte_hole()
1405 for (; addr < min(end, vma->vm_end); addr += PAGE_SIZE) { in pagemap_pte_hole()
1406 err = add_to_pagemap(addr, &pme, pm); in pagemap_pte_hole()
1416 struct vm_area_struct *vma, unsigned long addr, pte_t pte) in pte_to_pagemap_entry() argument
1426 page = vm_normal_page(vma, addr, pte); in pte_to_pagemap_entry()
1469 static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end, in pagemap_pmd_range() argument
1499 ((addr & ~PMD_MASK) >> PAGE_SHIFT); in pagemap_pmd_range()
1512 ((addr & ~PMD_MASK) >> PAGE_SHIFT); in pagemap_pmd_range()
1530 for (; addr != end; addr += PAGE_SIZE) { in pagemap_pmd_range()
1533 err = add_to_pagemap(addr, &pme, pm); in pagemap_pmd_range()
1555 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range()
1556 for (; addr < end; pte++, addr += PAGE_SIZE) { in pagemap_pmd_range()
1559 pme = pte_to_pagemap_entry(pm, vma, addr, *pte); in pagemap_pmd_range()
1560 err = add_to_pagemap(addr, &pme, pm); in pagemap_pmd_range()
1574 unsigned long addr, unsigned long end, in pagemap_hugetlb_range() argument
1602 ((addr & ~hmask) >> PAGE_SHIFT); in pagemap_hugetlb_range()
1607 for (; addr != end; addr += PAGE_SIZE) { in pagemap_hugetlb_range()
1610 err = add_to_pagemap(addr, &pme, pm); in pagemap_hugetlb_range()
1822 unsigned long addr) in can_gather_numa_stats() argument
1830 page = vm_normal_page(vma, addr, pte); in can_gather_numa_stats()
1847 unsigned long addr) in can_gather_numa_stats_pmd() argument
1855 page = vm_normal_page_pmd(vma, addr, pmd); in can_gather_numa_stats_pmd()
1870 static int gather_pte_stats(pmd_t *pmd, unsigned long addr, in gather_pte_stats() argument
1884 page = can_gather_numa_stats_pmd(*pmd, vma, addr); in gather_pte_stats()
1895 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats()
1897 struct page *page = can_gather_numa_stats(*pte, vma, addr); in gather_pte_stats()
1902 } while (pte++, addr += PAGE_SIZE, addr != end); in gather_pte_stats()
1909 unsigned long addr, unsigned long end, struct mm_walk *walk) in gather_hugetlb_stats() argument
1927 unsigned long addr, unsigned long end, struct mm_walk *walk) in gather_hugetlb_stats() argument