/mm/ |
D | vmalloc.c | 380 get_subtree_max_size(struct rb_node *node) in get_subtree_max_size() 384 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size() 395 get_subtree_max_size(va->rb_node.rb_left), in compute_subtree_max_size() 396 get_subtree_max_size(va->rb_node.rb_right)); in compute_subtree_max_size() 400 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size) 415 struct rb_node *n = vmap_area_root.rb_node; in __find_vmap_area() 420 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area() 436 static __always_inline struct rb_node ** 438 struct rb_root *root, struct rb_node *from, in find_va_links() 439 struct rb_node **parent) in find_va_links() [all …]
|
D | backing-dev.c | 408 struct rb_node **node, *parent; in wb_congested_get_create() 413 node = &bdi->cgwb_congested_tree.rb_node; in wb_congested_get_create() 419 rb_node); in wb_congested_get_create() 431 rb_link_node(&congested->rb_node, parent, node); in wb_congested_get_create() 432 rb_insert_color(&congested->rb_node, &bdi->cgwb_congested_tree); in wb_congested_get_create() 471 rb_erase(&congested->rb_node, in wb_congested_put() 774 struct rb_node *rbn; in cgwb_bdi_exit() 779 rb_entry(rbn, struct bdi_writeback_congested, rb_node); in cgwb_bdi_exit() 884 static struct rb_node **bdi_lookup_rb_node(u64 id, struct rb_node **parentp) in bdi_lookup_rb_node() 886 struct rb_node **p = &bdi_tree.rb_node; in bdi_lookup_rb_node() [all …]
|
D | mmap.c | 332 struct rb_node *nd, *pn = NULL; in browse_rb() 378 struct rb_node *nd; in validate_mm_rb() 530 struct rb_node ***rb_link, struct rb_node **rb_parent) in find_vma_links() 532 struct rb_node **__rb_link, *__rb_parent, *rb_prev; in find_vma_links() 534 __rb_link = &mm->mm_rb.rb_node; in find_vma_links() 591 struct rb_node **rb_link, struct rb_node *rb_parent) in __vma_link_rb() 635 struct vm_area_struct *prev, struct rb_node **rb_link, in __vma_link() 636 struct rb_node *rb_parent) in __vma_link() 643 struct vm_area_struct *prev, struct rb_node **rb_link, in vma_link() 644 struct rb_node *rb_parent) in vma_link() [all …]
|
D | kmemleak.c | 142 struct rb_node rb_node; member 381 struct rb_node *rb = object_tree_root.rb_node; in lookup_object() 385 rb_entry(rb, struct kmemleak_object, rb_node); in lookup_object() 387 rb = object->rb_node.rb_left; in lookup_object() 389 rb = object->rb_node.rb_right; in lookup_object() 535 rb_erase(&object->rb_node, &object_tree_root); in __remove_object() 575 struct rb_node **link, *rb_parent; in create_object() 625 link = &object_tree_root.rb_node; in create_object() 629 parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); in create_object() 631 link = &parent->rb_node.rb_left; in create_object() [all …]
|
D | ksm.c | 157 struct rb_node node; /* when node of stable tree */ 207 struct rb_node node; /* when node of unstable tree */ 945 while (root_stable_tree[nid].rb_node) { in remove_all_stable_nodes() 946 stable_node = rb_entry(root_stable_tree[nid].rb_node, in remove_all_stable_nodes() 1555 struct rb_node **new; in stable_tree_search() 1556 struct rb_node *parent; in stable_tree_search() 1570 new = &root->rb_node; in stable_tree_search() 1808 struct rb_node **new; in stable_tree_insert() 1809 struct rb_node *parent; in stable_tree_insert() 1818 new = &root->rb_node; in stable_tree_insert() [all …]
|
D | swapfile.c | 157 struct rb_node *rb = rb_first(&sis->swap_extent_root); in first_se() 158 return rb_entry(rb, struct swap_extent, rb_node); in first_se() 163 struct rb_node *rb = rb_next(&se->rb_node); in next_se() 164 return rb ? rb_entry(rb, struct swap_extent, rb_node) : NULL; in next_se() 208 struct rb_node *rb; in offset_to_swap_extent() 210 rb = sis->swap_extent_root.rb_node; in offset_to_swap_extent() 212 se = rb_entry(rb, struct swap_extent, rb_node); in offset_to_swap_extent() 2287 struct rb_node *rb = sis->swap_extent_root.rb_node; in destroy_swap_extents() 2288 struct swap_extent *se = rb_entry(rb, struct swap_extent, rb_node); in destroy_swap_extents() 2314 struct rb_node **link = &sis->swap_extent_root.rb_node, *parent = NULL; in add_swap_extent() [all …]
|
D | nommu.c | 455 struct rb_node *p, *lastp; in validate_nommu_regions() 488 struct rb_node **p, *parent; in add_nommu_region() 493 p = &nommu_region_tree.rb_node; in add_nommu_region() 518 BUG_ON(!nommu_region_tree.rb_node); in delete_nommu_region() 547 BUG_ON(!nommu_region_tree.rb_node); in __put_nommu_region() 586 struct rb_node **p, *parent, *rb_prev; in add_vma_to_mm() 606 p = &mm->mm_rb.rb_node; in add_vma_to_mm() 1106 struct rb_node *rb; in do_mmap()
|
D | zswap.c | 152 struct rb_node rbnode; 269 struct rb_node *node = root->rb_node; in zswap_rb_search() 291 struct rb_node **link = &root->rb_node, *parent = NULL; in zswap_rb_insert()
|
D | mempolicy.c | 2291 struct rb_node *n = sp->root.rb_node; in sp_lookup() 2307 struct rb_node *prev = rb_prev(n); in sp_lookup() 2324 struct rb_node **p = &sp->root.rb_node; in sp_insert() 2325 struct rb_node *parent = NULL; in sp_insert() 2351 if (!sp->root.rb_node) in mpol_shared_policy_lookup() 2519 struct rb_node *next = rb_next(&n->nd); in shared_policy_replace() 2648 struct rb_node *next; in mpol_free_shared_policy() 2650 if (!p->root.rb_node) in mpol_free_shared_policy()
|
D | interval_tree.c | 32 struct rb_node **link; in vma_interval_tree_insert_after()
|
D | internal.h | 293 struct vm_area_struct *prev, struct rb_node *rb_parent);
|
D | util.c | 274 struct vm_area_struct *prev, struct rb_node *rb_parent) in __vma_link_list()
|
D | memcontrol.c | 120 struct rb_node *rb_rightmost; 525 struct rb_node **p = &mctz->rb_root.rb_node; in __mem_cgroup_insert_exceeded() 526 struct rb_node *parent = NULL; in __mem_cgroup_insert_exceeded()
|