/mm/ |
D | vmalloc.c | 383 get_subtree_max_size(struct rb_node *node) in get_subtree_max_size() 387 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size() 398 get_subtree_max_size(va->rb_node.rb_left), in compute_subtree_max_size() 399 get_subtree_max_size(va->rb_node.rb_right)); in compute_subtree_max_size() 403 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size) 418 struct rb_node *n = vmap_area_root.rb_node; in __find_vmap_area() 423 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area() 439 static __always_inline struct rb_node ** 441 struct rb_root *root, struct rb_node *from, in find_va_links() 442 struct rb_node **parent) in find_va_links() [all …]
|
D | backing-dev.c | 409 struct rb_node **node, *parent; in wb_congested_get_create() 414 node = &bdi->cgwb_congested_tree.rb_node; in wb_congested_get_create() 420 rb_node); in wb_congested_get_create() 432 rb_link_node(&congested->rb_node, parent, node); in wb_congested_get_create() 433 rb_insert_color(&congested->rb_node, &bdi->cgwb_congested_tree); in wb_congested_get_create() 472 rb_erase(&congested->rb_node, in wb_congested_put() 775 struct rb_node *rbn; in cgwb_bdi_exit() 780 rb_entry(rbn, struct bdi_writeback_congested, rb_node); in cgwb_bdi_exit() 885 static struct rb_node **bdi_lookup_rb_node(u64 id, struct rb_node **parentp) in bdi_lookup_rb_node() 887 struct rb_node **p = &bdi_tree.rb_node; in bdi_lookup_rb_node() [all …]
|
D | mmap.c | 330 struct rb_node *nd, *pn = NULL; in browse_rb() 376 struct rb_node *nd; in validate_mm_rb() 528 struct rb_node ***rb_link, struct rb_node **rb_parent) in find_vma_links() 530 struct rb_node **__rb_link, *__rb_parent, *rb_prev; in find_vma_links() 532 __rb_link = &mm->mm_rb.rb_node; in find_vma_links() 589 struct rb_node **rb_link, struct rb_node *rb_parent) in __vma_link_rb() 633 struct vm_area_struct *prev, struct rb_node **rb_link, in __vma_link() 634 struct rb_node *rb_parent) in __vma_link() 641 struct vm_area_struct *prev, struct rb_node **rb_link, in vma_link() 642 struct rb_node *rb_parent) in vma_link() [all …]
|
D | kmemleak.c | 142 struct rb_node rb_node; member 381 struct rb_node *rb = object_tree_root.rb_node; in lookup_object() 385 rb_entry(rb, struct kmemleak_object, rb_node); in lookup_object() 387 rb = object->rb_node.rb_left; in lookup_object() 389 rb = object->rb_node.rb_right; in lookup_object() 535 rb_erase(&object->rb_node, &object_tree_root); in __remove_object() 575 struct rb_node **link, *rb_parent; in create_object() 625 link = &object_tree_root.rb_node; in create_object() 629 parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); in create_object() 631 link = &parent->rb_node.rb_left; in create_object() [all …]
|
D | ksm.c | 157 struct rb_node node; /* when node of stable tree */ 207 struct rb_node node; /* when node of unstable tree */ 946 while (root_stable_tree[nid].rb_node) { in remove_all_stable_nodes() 947 stable_node = rb_entry(root_stable_tree[nid].rb_node, in remove_all_stable_nodes() 1556 struct rb_node **new; in stable_tree_search() 1557 struct rb_node *parent; in stable_tree_search() 1571 new = &root->rb_node; in stable_tree_search() 1809 struct rb_node **new; in stable_tree_insert() 1810 struct rb_node *parent; in stable_tree_insert() 1819 new = &root->rb_node; in stable_tree_insert() [all …]
|
D | swapfile.c | 157 struct rb_node *rb = rb_first(&sis->swap_extent_root); in first_se() 158 return rb_entry(rb, struct swap_extent, rb_node); in first_se() 163 struct rb_node *rb = rb_next(&se->rb_node); in next_se() 164 return rb ? rb_entry(rb, struct swap_extent, rb_node) : NULL; in next_se() 208 struct rb_node *rb; in offset_to_swap_extent() 210 rb = sis->swap_extent_root.rb_node; in offset_to_swap_extent() 212 se = rb_entry(rb, struct swap_extent, rb_node); in offset_to_swap_extent() 2306 struct rb_node *rb = sis->swap_extent_root.rb_node; in destroy_swap_extents() 2307 struct swap_extent *se = rb_entry(rb, struct swap_extent, rb_node); in destroy_swap_extents() 2333 struct rb_node **link = &sis->swap_extent_root.rb_node, *parent = NULL; in add_swap_extent() [all …]
|
D | nommu.c | 459 struct rb_node *p, *lastp; in validate_nommu_regions() 492 struct rb_node **p, *parent; in add_nommu_region() 497 p = &nommu_region_tree.rb_node; in add_nommu_region() 522 BUG_ON(!nommu_region_tree.rb_node); in delete_nommu_region() 551 BUG_ON(!nommu_region_tree.rb_node); in __put_nommu_region() 590 struct rb_node **p, *parent, *rb_prev; in add_vma_to_mm() 610 p = &mm->mm_rb.rb_node; in add_vma_to_mm() 1110 struct rb_node *rb; in do_mmap()
|
D | zswap.c | 152 struct rb_node rbnode; 269 struct rb_node *node = root->rb_node; in zswap_rb_search() 291 struct rb_node **link = &root->rb_node, *parent = NULL; in zswap_rb_insert()
|
D | mempolicy.c | 2287 struct rb_node *n = sp->root.rb_node; in sp_lookup() 2303 struct rb_node *prev = rb_prev(n); in sp_lookup() 2320 struct rb_node **p = &sp->root.rb_node; in sp_insert() 2321 struct rb_node *parent = NULL; in sp_insert() 2347 if (!sp->root.rb_node) in mpol_shared_policy_lookup() 2515 struct rb_node *next = rb_next(&n->nd); in shared_policy_replace() 2645 struct rb_node *next; in mpol_free_shared_policy() 2647 if (!p->root.rb_node) in mpol_free_shared_policy()
|
D | interval_tree.c | 32 struct rb_node **link; in vma_interval_tree_insert_after()
|
D | internal.h | 293 struct vm_area_struct *prev, struct rb_node *rb_parent);
|
D | util.c | 274 struct vm_area_struct *prev, struct rb_node *rb_parent) in __vma_link_list()
|
D | memcontrol.c | 120 struct rb_node *rb_rightmost; 527 struct rb_node **p = &mctz->rb_root.rb_node; in __mem_cgroup_insert_exceeded() 528 struct rb_node *parent = NULL; in __mem_cgroup_insert_exceeded()
|