• Home
  • Raw
  • Download

Lines Matching refs:addr

496 static int smaps_pte_hole(unsigned long addr, unsigned long end,  in smaps_pte_hole()  argument
502 walk->vma->vm_file->f_mapping, addr, end); in smaps_pte_hole()
510 static void smaps_pte_entry(pte_t *pte, unsigned long addr, in smaps_pte_entry() argument
520 page = vm_normal_page(vma, addr, *pte); in smaps_pte_entry()
547 linear_page_index(vma, addr)); in smaps_pte_entry()
560 static void smaps_pmd_entry(pmd_t *pmd, unsigned long addr, in smaps_pmd_entry() argument
571 page = follow_trans_huge_pmd(vma, addr, pmd, FOLL_DUMP); in smaps_pmd_entry()
595 static void smaps_pmd_entry(pmd_t *pmd, unsigned long addr, in smaps_pmd_entry() argument
601 static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, in smaps_pte_range() argument
610 smaps_pmd_entry(pmd, addr, walk); in smaps_pte_range()
622 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range()
623 for (; addr != end; pte++, addr += PAGE_SIZE) in smaps_pte_range()
624 smaps_pte_entry(pte, addr, walk); in smaps_pte_range()
712 unsigned long addr, unsigned long end, in smaps_hugetlb_range() argument
720 page = vm_normal_page(vma, addr, *pte); in smaps_hugetlb_range()
1065 static inline bool pte_is_pinned(struct vm_area_struct *vma, unsigned long addr, pte_t pte) in pte_is_pinned() argument
1075 page = vm_normal_page(vma, addr, pte); in pte_is_pinned()
1082 unsigned long addr, pte_t *pte) in clear_soft_dirty() argument
1095 if (pte_is_pinned(vma, addr, ptent)) in clear_soft_dirty()
1097 old_pte = ptep_modify_prot_start(vma, addr, pte); in clear_soft_dirty()
1100 ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent); in clear_soft_dirty()
1103 set_pte_at(vma->vm_mm, addr, pte, ptent); in clear_soft_dirty()
1108 unsigned long addr, pte_t *pte) in clear_soft_dirty() argument
1115 unsigned long addr, pmd_t *pmdp) in clear_soft_dirty_pmd() argument
1121 old = pmdp_invalidate(vma, addr, pmdp); in clear_soft_dirty_pmd()
1130 set_pmd_at(vma->vm_mm, addr, pmdp, pmd); in clear_soft_dirty_pmd()
1133 set_pmd_at(vma->vm_mm, addr, pmdp, pmd); in clear_soft_dirty_pmd()
1138 unsigned long addr, pmd_t *pmdp) in clear_soft_dirty_pmd() argument
1143 static int clear_refs_pte_range(pmd_t *pmd, unsigned long addr, in clear_refs_pte_range() argument
1155 clear_soft_dirty_pmd(vma, addr, pmd); in clear_refs_pte_range()
1165 pmdp_test_and_clear_young(vma, addr, pmd); in clear_refs_pte_range()
1176 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in clear_refs_pte_range()
1177 for (; addr != end; pte++, addr += PAGE_SIZE) { in clear_refs_pte_range()
1181 clear_soft_dirty(vma, addr, pte); in clear_refs_pte_range()
1188 page = vm_normal_page(vma, addr, ptent); in clear_refs_pte_range()
1193 ptep_test_and_clear_young(vma, addr, pte); in clear_refs_pte_range()
1340 static int add_to_pagemap(unsigned long addr, pagemap_entry_t *pme, in add_to_pagemap() argument
1353 unsigned long addr = start; in pagemap_pte_hole() local
1356 while (addr < end) { in pagemap_pte_hole()
1357 struct vm_area_struct *vma = find_vma(walk->mm, addr); in pagemap_pte_hole()
1367 for (; addr < hole_end; addr += PAGE_SIZE) { in pagemap_pte_hole()
1368 err = add_to_pagemap(addr, &pme, pm); in pagemap_pte_hole()
1379 for (; addr < min(end, vma->vm_end); addr += PAGE_SIZE) { in pagemap_pte_hole()
1380 err = add_to_pagemap(addr, &pme, pm); in pagemap_pte_hole()
1390 struct vm_area_struct *vma, unsigned long addr, pte_t pte) in pte_to_pagemap_entry() argument
1400 page = vm_normal_page(vma, addr, pte); in pte_to_pagemap_entry()
1431 static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end, in pagemap_pmd_range() argument
1461 ((addr & ~PMD_MASK) >> PAGE_SHIFT); in pagemap_pmd_range()
1470 ((addr & ~PMD_MASK) >> PAGE_SHIFT); in pagemap_pmd_range()
1488 for (; addr != end; addr += PAGE_SIZE) { in pagemap_pmd_range()
1491 err = add_to_pagemap(addr, &pme, pm); in pagemap_pmd_range()
1513 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range()
1514 for (; addr < end; pte++, addr += PAGE_SIZE) { in pagemap_pmd_range()
1517 pme = pte_to_pagemap_entry(pm, vma, addr, *pte); in pagemap_pmd_range()
1518 err = add_to_pagemap(addr, &pme, pm); in pagemap_pmd_range()
1532 unsigned long addr, unsigned long end, in pagemap_hugetlb_range() argument
1557 ((addr & ~hmask) >> PAGE_SHIFT); in pagemap_hugetlb_range()
1560 for (; addr != end; addr += PAGE_SIZE) { in pagemap_hugetlb_range()
1563 err = add_to_pagemap(addr, &pme, pm); in pagemap_hugetlb_range()
1774 unsigned long addr) in can_gather_numa_stats() argument
1782 page = vm_normal_page(vma, addr, pte); in can_gather_numa_stats()
1799 unsigned long addr) in can_gather_numa_stats_pmd() argument
1807 page = vm_normal_page_pmd(vma, addr, pmd); in can_gather_numa_stats_pmd()
1822 static int gather_pte_stats(pmd_t *pmd, unsigned long addr, in gather_pte_stats() argument
1836 page = can_gather_numa_stats_pmd(*pmd, vma, addr); in gather_pte_stats()
1847 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats()
1849 struct page *page = can_gather_numa_stats(*pte, vma, addr); in gather_pte_stats()
1854 } while (pte++, addr += PAGE_SIZE, addr != end); in gather_pte_stats()
1861 unsigned long addr, unsigned long end, struct mm_walk *walk) in gather_hugetlb_stats() argument
1881 unsigned long addr, unsigned long end, struct mm_walk *walk) in gather_hugetlb_stats() argument