/mm/ |
D | vmalloc.c | 775 get_subtree_max_size(struct rb_node *node) in get_subtree_max_size() 779 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size() 790 get_subtree_max_size(va->rb_node.rb_left), in compute_subtree_max_size() 791 get_subtree_max_size(va->rb_node.rb_right)); in compute_subtree_max_size() 795 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size) 812 struct rb_node *n = vmap_area_root.rb_node; in find_vmap_area_exceed_addr() 819 tmp = rb_entry(n, struct vmap_area, rb_node); in find_vmap_area_exceed_addr() 835 struct rb_node *n = vmap_area_root.rb_node; in __find_vmap_area() 842 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area() 862 static __always_inline struct rb_node ** [all …]
|
D | mmap.c | 337 struct rb_node *nd, *pn = NULL; in browse_rb() 383 struct rb_node *nd; in validate_mm_rb() 533 struct rb_node ***rb_link, struct rb_node **rb_parent) in find_vma_links() 535 struct rb_node **__rb_link, *__rb_parent, *rb_prev; in find_vma_links() 538 __rb_link = &mm->mm_rb.rb_node; in find_vma_links() 600 struct vm_area_struct **pprev, struct rb_node ***link, in munmap_vma_range() 601 struct rb_node **parent, struct list_head *uf) in munmap_vma_range() 639 struct rb_node **rb_link, struct rb_node *rb_parent) in __vma_link_rb() 681 struct vm_area_struct *prev, struct rb_node **rb_link, in __vma_link() 682 struct rb_node *rb_parent) in __vma_link() [all …]
|
D | backing-dev.c | 827 static struct rb_node **bdi_lookup_rb_node(u64 id, struct rb_node **parentp) in bdi_lookup_rb_node() 829 struct rb_node **p = &bdi_tree.rb_node; in bdi_lookup_rb_node() 830 struct rb_node *parent = NULL; in bdi_lookup_rb_node() 837 bdi = rb_entry(parent, struct backing_dev_info, rb_node); in bdi_lookup_rb_node() 862 struct rb_node **p; in bdi_get_by_id() 867 bdi = rb_entry(*p, struct backing_dev_info, rb_node); in bdi_get_by_id() 878 struct rb_node *parent, **p; in bdi_register_va() 899 rb_link_node(&bdi->rb_node, parent, p); in bdi_register_va() 900 rb_insert_color(&bdi->rb_node, &bdi_tree); in bdi_register_va() 935 rb_erase(&bdi->rb_node, &bdi_tree); in bdi_remove_from_list()
|
D | kmemleak.c | 144 struct rb_node rb_node; member 383 struct rb_node *rb = object_tree_root.rb_node; in lookup_object() 390 object = rb_entry(rb, struct kmemleak_object, rb_node); in lookup_object() 394 rb = object->rb_node.rb_left; in lookup_object() 396 rb = object->rb_node.rb_right; in lookup_object() 542 rb_erase(&object->rb_node, &object_tree_root); in __remove_object() 582 struct rb_node **link, *rb_parent; in create_object() 633 link = &object_tree_root.rb_node; in create_object() 637 parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); in create_object() 640 link = &parent->rb_node.rb_left; in create_object() [all …]
|
D | ksm.c | 157 struct rb_node node; /* when node of stable tree */ 207 struct rb_node node; /* when node of unstable tree */ 940 while (root_stable_tree[nid].rb_node) { in remove_all_stable_nodes() 941 stable_node = rb_entry(root_stable_tree[nid].rb_node, in remove_all_stable_nodes() 1550 struct rb_node **new; in stable_tree_search() 1551 struct rb_node *parent; in stable_tree_search() 1565 new = &root->rb_node; in stable_tree_search() 1801 struct rb_node **new; in stable_tree_insert() 1802 struct rb_node *parent; in stable_tree_insert() 1811 new = &root->rb_node; in stable_tree_insert() [all …]
|
D | nommu.c | 426 struct rb_node *p, *lastp; in validate_nommu_regions() 459 struct rb_node **p, *parent; in add_nommu_region() 464 p = &nommu_region_tree.rb_node; in add_nommu_region() 489 BUG_ON(!nommu_region_tree.rb_node); in delete_nommu_region() 518 BUG_ON(!nommu_region_tree.rb_node); in __put_nommu_region() 557 struct rb_node **p, *parent, *rb_prev; in add_vma_to_mm() 577 p = &mm->mm_rb.rb_node; in add_vma_to_mm() 1075 struct rb_node *rb; in do_mmap()
|
D | swapfile.c | 156 struct rb_node *rb = rb_first(&sis->swap_extent_root); in first_se() 157 return rb_entry(rb, struct swap_extent, rb_node); in first_se() 162 struct rb_node *rb = rb_next(&se->rb_node); in next_se() 163 return rb ? rb_entry(rb, struct swap_extent, rb_node) : NULL; in next_se() 207 struct rb_node *rb; in offset_to_swap_extent() 209 rb = sis->swap_extent_root.rb_node; in offset_to_swap_extent() 211 se = rb_entry(rb, struct swap_extent, rb_node); in offset_to_swap_extent() 2294 struct rb_node *rb = sis->swap_extent_root.rb_node; in destroy_swap_extents() 2295 struct swap_extent *se = rb_entry(rb, struct swap_extent, rb_node); in destroy_swap_extents() 2321 struct rb_node **link = &sis->swap_extent_root.rb_node, *parent = NULL; in add_swap_extent() [all …]
|
D | mempolicy.c | 2289 struct rb_node *n = sp->root.rb_node; in sp_lookup() 2305 struct rb_node *prev = rb_prev(n); in sp_lookup() 2322 struct rb_node **p = &sp->root.rb_node; in sp_insert() 2323 struct rb_node *parent = NULL; in sp_insert() 2349 if (!sp->root.rb_node) in mpol_shared_policy_lookup() 2524 struct rb_node *next = rb_next(&n->nd); in shared_policy_replace() 2654 struct rb_node *next; in mpol_free_shared_policy() 2656 if (!p->root.rb_node) in mpol_free_shared_policy()
|
D | zswap.c | 172 struct rb_node rbnode; 296 struct rb_node *node = root->rb_node; in zswap_rb_search() 318 struct rb_node **link = &root->rb_node, *parent = NULL; in zswap_rb_insert()
|
D | interval_tree.c | 32 struct rb_node **link; in vma_interval_tree_insert_after()
|
D | memcontrol.c | 120 struct rb_node *rb_rightmost; 484 struct rb_node **p = &mctz->rb_root.rb_node; in __mem_cgroup_insert_exceeded() 485 struct rb_node *parent = NULL; in __mem_cgroup_insert_exceeded()
|
/mm/damon/ |
D | vaddr.c | 132 rbroot.rb_node = &vma->vm_rb; in __damon_va_three_regions()
|