Searched refs:vmas (Results 1 – 5 of 5) sorted by relevance
/mm/damon/ |
D | vaddr-test.h | 17 static void __link_vmas(struct vm_area_struct *vmas, ssize_t nr_vmas) in __link_vmas() argument 26 vmas[i].vm_next = &vmas[i + 1]; in __link_vmas() 28 vmas[i].vm_rb.rb_left = NULL; in __link_vmas() 29 vmas[i].vm_rb.rb_right = &vmas[i + 1].vm_rb; in __link_vmas() 35 gap = vmas[j].vm_start - vmas[j - 1].vm_end; in __link_vmas() 39 vmas[i].rb_subtree_gap = largest_gap; in __link_vmas() 41 vmas[i].vm_next = NULL; in __link_vmas() 42 vmas[i].vm_rb.rb_right = NULL; in __link_vmas() 43 vmas[i].rb_subtree_gap = 0; in __link_vmas() 77 struct vm_area_struct vmas[] = { in damon_test_three_regions_in_vmas() local [all …]
|
/mm/ |
D | gup.c | 1064 struct vm_area_struct **vmas, int *locked) in __get_user_pages() argument 1112 i = follow_hugetlb_page(mm, vma, pages, vmas, in __get_user_pages() 1173 if (vmas) { in __get_user_pages() 1174 vmas[i] = vma; in __get_user_pages() 1295 struct vm_area_struct **vmas, in __get_user_pages_locked() argument 1304 BUG_ON(vmas); in __get_user_pages_locked() 1328 vmas, locked); in __get_user_pages_locked() 1540 struct vm_area_struct **vmas, int *locked, in __get_user_pages_locked() argument 1570 if (vmas) in __get_user_pages_locked() 1571 vmas[i] = vma; in __get_user_pages_locked() [all …]
|
D | vmacache.c | 38 current->vmacache.vmas[VMACACHE_HASH(addr)] = newvma; in vmacache_update() 72 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find() 105 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find_exact()
|
D | hugetlb.c | 4894 struct page **pages, struct vm_area_struct **vmas, in follow_hugetlb_page() argument 5012 if (!pages && !vmas && !pfn_offset && in follow_hugetlb_page() 5043 if (vmas) in follow_hugetlb_page() 5044 vmas[i] = vma; in follow_hugetlb_page()
|
D | nommu.c | 637 if (curr->vmacache.vmas[i] == vma) { in delete_vma_from_mm()
|