/mm/ |
D | vmalloc.c | 462 get_subtree_max_size(struct rb_node *node) in get_subtree_max_size() 466 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size() 477 get_subtree_max_size(va->rb_node.rb_left), in compute_subtree_max_size() 478 get_subtree_max_size(va->rb_node.rb_right)); in compute_subtree_max_size() 482 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size) 498 struct rb_node *n = vmap_area_root.rb_node; in __find_vmap_area() 503 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area() 523 static __always_inline struct rb_node ** 525 struct rb_root *root, struct rb_node *from, in find_va_links() 526 struct rb_node **parent) in find_va_links() [all …]
|
D | mmap.c | 352 struct rb_node *nd, *pn = NULL; in browse_rb() 398 struct rb_node *nd; in validate_mm_rb() 566 struct rb_node ***rb_link, struct rb_node **rb_parent) in find_vma_links() 568 struct rb_node **__rb_link, *__rb_parent, *rb_prev; in find_vma_links() 570 __rb_link = &mm->mm_rb.rb_node; in find_vma_links() 632 struct vm_area_struct **pprev, struct rb_node ***link, in munmap_vma_range() 633 struct rb_node **parent, struct list_head *uf) in munmap_vma_range() 671 struct rb_node **rb_link, struct rb_node *rb_parent) in __vma_link_rb() 717 struct vm_area_struct *prev, struct rb_node **rb_link, in __vma_link() 718 struct rb_node *rb_parent) in __vma_link() [all …]
|
D | backing-dev.c | 761 static struct rb_node **bdi_lookup_rb_node(u64 id, struct rb_node **parentp) in bdi_lookup_rb_node() 763 struct rb_node **p = &bdi_tree.rb_node; in bdi_lookup_rb_node() 764 struct rb_node *parent = NULL; in bdi_lookup_rb_node() 771 bdi = rb_entry(parent, struct backing_dev_info, rb_node); in bdi_lookup_rb_node() 796 struct rb_node **p; in bdi_get_by_id() 801 bdi = rb_entry(*p, struct backing_dev_info, rb_node); in bdi_get_by_id() 812 struct rb_node *parent, **p; in bdi_register_va() 833 rb_link_node(&bdi->rb_node, parent, p); in bdi_register_va() 834 rb_insert_color(&bdi->rb_node, &bdi_tree); in bdi_register_va() 869 rb_erase(&bdi->rb_node, &bdi_tree); in bdi_remove_from_list()
|
D | kmemleak.c | 143 struct rb_node rb_node; member 382 struct rb_node *rb = object_tree_root.rb_node; in lookup_object() 386 rb_entry(rb, struct kmemleak_object, rb_node); in lookup_object() 388 rb = object->rb_node.rb_left; in lookup_object() 390 rb = object->rb_node.rb_right; in lookup_object() 536 rb_erase(&object->rb_node, &object_tree_root); in __remove_object() 576 struct rb_node **link, *rb_parent; in create_object() 626 link = &object_tree_root.rb_node; in create_object() 630 parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); in create_object() 632 link = &parent->rb_node.rb_left; in create_object() [all …]
|
D | ksm.c | 157 struct rb_node node; /* when node of stable tree */ 207 struct rb_node node; /* when node of unstable tree */ 947 while (root_stable_tree[nid].rb_node) { in remove_all_stable_nodes() 948 stable_node = rb_entry(root_stable_tree[nid].rb_node, in remove_all_stable_nodes() 1557 struct rb_node **new; in stable_tree_search() 1558 struct rb_node *parent; in stable_tree_search() 1572 new = &root->rb_node; in stable_tree_search() 1810 struct rb_node **new; in stable_tree_insert() 1811 struct rb_node *parent; in stable_tree_insert() 1820 new = &root->rb_node; in stable_tree_insert() [all …]
|
D | nommu.c | 436 struct rb_node *p, *lastp; in validate_nommu_regions() 469 struct rb_node **p, *parent; in add_nommu_region() 474 p = &nommu_region_tree.rb_node; in add_nommu_region() 499 BUG_ON(!nommu_region_tree.rb_node); in delete_nommu_region() 528 BUG_ON(!nommu_region_tree.rb_node); in __put_nommu_region() 567 struct rb_node **p, *parent, *rb_prev; in add_vma_to_mm() 587 p = &mm->mm_rb.rb_node; in add_vma_to_mm() 1080 struct rb_node *rb; in do_mmap()
|
D | swapfile.c | 158 struct rb_node *rb = rb_first(&sis->swap_extent_root); in first_se() 159 return rb_entry(rb, struct swap_extent, rb_node); in first_se() 164 struct rb_node *rb = rb_next(&se->rb_node); in next_se() 165 return rb ? rb_entry(rb, struct swap_extent, rb_node) : NULL; in next_se() 209 struct rb_node *rb; in offset_to_swap_extent() 211 rb = sis->swap_extent_root.rb_node; in offset_to_swap_extent() 213 se = rb_entry(rb, struct swap_extent, rb_node); in offset_to_swap_extent() 2363 struct rb_node *rb = sis->swap_extent_root.rb_node; in destroy_swap_extents() 2364 struct swap_extent *se = rb_entry(rb, struct swap_extent, rb_node); in destroy_swap_extents() 2390 struct rb_node **link = &sis->swap_extent_root.rb_node, *parent = NULL; in add_swap_extent() [all …]
|
D | zswap.c | 162 struct rb_node rbnode; 286 struct rb_node *node = root->rb_node; in zswap_rb_search() 308 struct rb_node **link = &root->rb_node, *parent = NULL; in zswap_rb_insert()
|
D | mempolicy.c | 2379 struct rb_node *n = sp->root.rb_node; in sp_lookup() 2395 struct rb_node *prev = rb_prev(n); in sp_lookup() 2412 struct rb_node **p = &sp->root.rb_node; in sp_insert() 2413 struct rb_node *parent = NULL; in sp_insert() 2439 if (!sp->root.rb_node) in mpol_shared_policy_lookup() 2607 struct rb_node *next = rb_next(&n->nd); in shared_policy_replace() 2737 struct rb_node *next; in mpol_free_shared_policy() 2739 if (!p->root.rb_node) in mpol_free_shared_policy()
|
D | interval_tree.c | 32 struct rb_node **link; in vma_interval_tree_insert_after()
|
D | memcontrol.c | 114 struct rb_node *rb_rightmost; 610 struct rb_node **p = &mctz->rb_root.rb_node; in __mem_cgroup_insert_exceeded() 611 struct rb_node *parent = NULL; in __mem_cgroup_insert_exceeded()
|
/mm/damon/ |
D | vaddr.c | 132 rbroot.rb_node = &vma->vm_rb; in __damon_va_three_regions()
|