Lines Matching refs:rb_node
380 get_subtree_max_size(struct rb_node *node) in get_subtree_max_size()
384 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size()
395 get_subtree_max_size(va->rb_node.rb_left), in compute_subtree_max_size()
396 get_subtree_max_size(va->rb_node.rb_right)); in compute_subtree_max_size()
400 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size)
415 struct rb_node *n = vmap_area_root.rb_node; in __find_vmap_area()
420 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area()
436 static __always_inline struct rb_node **
438 struct rb_root *root, struct rb_node *from, in find_va_links()
439 struct rb_node **parent) in find_va_links()
442 struct rb_node **link; in find_va_links()
445 link = &root->rb_node; in find_va_links()
460 tmp_va = rb_entry(*link, struct vmap_area, rb_node); in find_va_links()
477 *parent = &tmp_va->rb_node; in find_va_links()
482 get_va_next_sibling(struct rb_node *parent, struct rb_node **link) in get_va_next_sibling()
495 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling()
501 struct rb_node *parent, struct rb_node **link, struct list_head *head) in link_va()
508 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in link_va()
514 rb_link_node(&va->rb_node, parent, link); in link_va()
527 rb_insert_augmented(&va->rb_node, in link_va()
531 rb_insert_color(&va->rb_node, root); in link_va()
541 if (WARN_ON(RB_EMPTY_NODE(&va->rb_node))) in unlink_va()
545 rb_erase_augmented(&va->rb_node, in unlink_va()
548 rb_erase(&va->rb_node, root); in unlink_va()
551 RB_CLEAR_NODE(&va->rb_node); in unlink_va()
556 augment_tree_propagate_check(struct rb_node *n) in augment_tree_propagate_check()
559 struct rb_node *node; in augment_tree_propagate_check()
566 va = rb_entry(n, struct vmap_area, rb_node); in augment_tree_propagate_check()
571 va = rb_entry(node, struct vmap_area, rb_node); in augment_tree_propagate_check()
586 va = rb_entry(n, struct vmap_area, rb_node); in augment_tree_propagate_check()
626 struct rb_node *node = &va->rb_node; in augment_tree_propagate_from()
630 va = rb_entry(node, struct vmap_area, rb_node); in augment_tree_propagate_from()
643 node = rb_parent(&va->rb_node); in augment_tree_propagate_from()
647 augment_tree_propagate_check(free_vmap_area_root.rb_node); in augment_tree_propagate_from()
655 struct rb_node **link; in insert_vmap_area()
656 struct rb_node *parent; in insert_vmap_area()
664 struct rb_node *from, struct rb_root *root, in insert_vmap_area_augment()
667 struct rb_node **link; in insert_vmap_area_augment()
668 struct rb_node *parent; in insert_vmap_area_augment()
691 struct rb_node **link; in merge_or_add_vmap_area()
692 struct rb_node *parent; in merge_or_add_vmap_area()
792 struct rb_node *node; in find_vmap_lowest_match()
796 node = free_vmap_area_root.rb_node; in find_vmap_lowest_match()
802 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
827 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
995 insert_vmap_area_augment(lva, &va->rb_node, in adjust_va_to_fit_type()
1076 kmemleak_scan_area(&va->rb_node, SIZE_MAX, gfp_mask & GFP_RECLAIM_MASK); in alloc_vmap_area()
3122 static struct vmap_area *node_to_va(struct rb_node *n) in node_to_va()
3124 return rb_entry_safe(n, struct vmap_area, rb_node); in node_to_va()
3140 struct rb_node *n; in pvm_find_va_enclose_addr()
3142 n = free_vmap_area_root.rb_node; in pvm_find_va_enclose_addr()
3146 tmp = rb_entry(n, struct vmap_area, rb_node); in pvm_find_va_enclose_addr()
3303 va = node_to_va(rb_prev(&va->rb_node)); in pcpu_get_vm_areas()