/mm/ |
D | interval_tree.c | 33 struct vm_area_struct *parent; in vma_interval_tree_insert_after() local 39 parent = prev; in vma_interval_tree_insert_after() 42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after() 44 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 45 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after() 47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after() 49 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 50 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after() [all …]
|
D | page_counter.c | 23 if (!c->parent) in propagate_protected_usage() 32 atomic_long_add(delta, &c->parent->children_min_usage); in propagate_protected_usage() 41 atomic_long_add(delta, &c->parent->children_low_usage); in propagate_protected_usage() 71 for (c = counter; c; c = c->parent) { in page_counter_charge() 100 for (c = counter; c; c = c->parent) { in page_counter_try_charge() 140 for (c = counter; c != *fail; c = c->parent) in page_counter_try_charge() 155 for (c = counter; c; c = c->parent) in page_counter_uncharge() 214 for (c = counter; c; c = c->parent) in page_counter_set_min() 231 for (c = counter; c; c = c->parent) in page_counter_set_low()
|
D | memcontrol.c | 321 struct mem_cgroup *parent) in memcg_reparent_objcgs() argument 330 xchg(&objcg->memcg, parent); in memcg_reparent_objcgs() 331 css_get(&parent->css); in memcg_reparent_objcgs() 332 list_add(&objcg->list, &parent->objcg_list); in memcg_reparent_objcgs() 336 css_get(&parent->css); in memcg_reparent_objcgs() 337 xchg(&iter->memcg, parent); in memcg_reparent_objcgs() 340 list_splice(&memcg->objcg_list, &parent->objcg_list); in memcg_reparent_objcgs() 611 struct rb_node *parent = NULL; in __mem_cgroup_insert_exceeded() local 622 parent = *p; in __mem_cgroup_insert_exceeded() 623 mz_node = rb_entry(parent, struct mem_cgroup_per_node, in __mem_cgroup_insert_exceeded() [all …]
|
D | vmalloc.c | 526 struct rb_node **parent) in find_va_links() argument 534 *parent = NULL; in find_va_links() 568 *parent = &tmp_va->rb_node; in find_va_links() 573 get_va_next_sibling(struct rb_node *parent, struct rb_node **link) in get_va_next_sibling() argument 577 if (unlikely(!parent)) in get_va_next_sibling() 586 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling() 587 return (&parent->rb_right == link ? list->next : list); in get_va_next_sibling() 592 struct rb_node *parent, struct rb_node **link, struct list_head *head) in link_va() argument 598 if (likely(parent)) { in link_va() 599 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in link_va() [all …]
|
D | backing-dev.c | 557 if (!memcg_css->parent) in wb_get_lookup() 593 if (!memcg_css->parent) in wb_get_create() 764 struct rb_node *parent = NULL; in bdi_lookup_rb_node() local 770 parent = *p; in bdi_lookup_rb_node() 771 bdi = rb_entry(parent, struct backing_dev_info, rb_node); in bdi_lookup_rb_node() 782 *parentp = parent; in bdi_lookup_rb_node() 812 struct rb_node *parent, **p; in bdi_register_va() local 832 p = bdi_lookup_rb_node(bdi->id, &parent); in bdi_register_va() 833 rb_link_node(&bdi->rb_node, parent, p); in bdi_register_va()
|
D | nommu.c | 469 struct rb_node **p, *parent; in add_nommu_region() local 473 parent = NULL; in add_nommu_region() 476 parent = *p; in add_nommu_region() 477 pregion = rb_entry(parent, struct vm_region, vm_rb); in add_nommu_region() 488 rb_link_node(®ion->vm_rb, parent, p); in add_nommu_region() 567 struct rb_node **p, *parent, *rb_prev; in add_vma_to_mm() local 586 parent = rb_prev = NULL; in add_vma_to_mm() 589 parent = *p; in add_vma_to_mm() 590 pvma = rb_entry(parent, struct vm_area_struct, vm_rb); in add_vma_to_mm() 597 rb_prev = parent; in add_vma_to_mm() [all …]
|
D | ksm.c | 1558 struct rb_node *parent; in stable_tree_search() local 1573 parent = NULL; in stable_tree_search() 1636 parent = *new; in stable_tree_search() 1638 new = &parent->rb_left; in stable_tree_search() 1640 new = &parent->rb_right; in stable_tree_search() 1705 rb_link_node(&page_node->node, parent, new); in stable_tree_search() 1811 struct rb_node *parent; in stable_tree_insert() local 1819 parent = NULL; in stable_tree_insert() 1871 parent = *new; in stable_tree_insert() 1873 new = &parent->rb_left; in stable_tree_insert() [all …]
|
D | hugetlb_cgroup.c | 76 return hugetlb_cgroup_from_css(h_cg->css.parent); in parent_hugetlb_cgroup() 167 struct hugetlb_cgroup *parent = parent_hugetlb_cgroup(h_cg); in hugetlb_cgroup_move_parent() local 179 if (!parent) { in hugetlb_cgroup_move_parent() 180 parent = root_h_cgroup; in hugetlb_cgroup_move_parent() 182 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent() 188 set_hugetlb_cgroup(page, parent); in hugetlb_cgroup_move_parent()
|
D | kmemleak.c | 575 struct kmemleak_object *object, *parent; in create_object() local 630 parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); in create_object() 631 if (ptr + size <= parent->pointer) in create_object() 632 link = &parent->rb_node.rb_left; in create_object() 633 else if (parent->pointer + parent->size <= ptr) in create_object() 634 link = &parent->rb_node.rb_right; in create_object() 642 dump_object_info(parent); in create_object()
|
D | slab.c | 227 static void kmem_cache_node_init(struct kmem_cache_node *parent) in kmem_cache_node_init() argument 229 INIT_LIST_HEAD(&parent->slabs_full); in kmem_cache_node_init() 230 INIT_LIST_HEAD(&parent->slabs_partial); in kmem_cache_node_init() 231 INIT_LIST_HEAD(&parent->slabs_free); in kmem_cache_node_init() 232 parent->total_slabs = 0; in kmem_cache_node_init() 233 parent->free_slabs = 0; in kmem_cache_node_init() 234 parent->shared = NULL; in kmem_cache_node_init() 235 parent->alien = NULL; in kmem_cache_node_init() 236 parent->colour_next = 0; in kmem_cache_node_init() 237 spin_lock_init(&parent->list_lock); in kmem_cache_node_init() [all …]
|
D | zswap.c | 308 struct rb_node **link = &root->rb_node, *parent = NULL; in zswap_rb_insert() local 312 parent = *link; in zswap_rb_insert() 313 myentry = rb_entry(parent, struct zswap_entry, rbnode); in zswap_rb_insert() 323 rb_link_node(&entry->rbnode, parent, link); in zswap_rb_insert()
|
D | rmap.c | 96 anon_vma->parent = anon_vma; in anon_vma_alloc() 370 anon_vma->parent = pvma->anon_vma; in anon_vma_fork() 381 anon_vma->parent->num_children++; in anon_vma_fork() 413 anon_vma->parent->num_children--; in unlink_anon_vmas()
|
D | swapfile.c | 2390 struct rb_node **link = &sis->swap_extent_root.rb_node, *parent = NULL; in add_swap_extent() local 2399 parent = *link; in add_swap_extent() 2400 link = &parent->rb_right; in add_swap_extent() 2403 if (parent) { in add_swap_extent() 2404 se = rb_entry(parent, struct swap_extent, rb_node); in add_swap_extent() 2421 rb_link_node(&new_se->rb_node, parent, link); in add_swap_extent()
|
D | mempolicy.c | 2413 struct rb_node *parent = NULL; in sp_insert() local 2417 parent = *p; in sp_insert() 2418 nd = rb_entry(parent, struct sp_node, nd); in sp_insert() 2426 rb_link_node(&new->nd, parent, p); in sp_insert()
|
D | mmap.c | 633 struct rb_node **parent, struct list_head *uf) in munmap_vma_range() argument 636 while (find_vma_links(mm, start, start + len, pprev, link, parent)) in munmap_vma_range()
|
D | hugetlb.c | 2994 static int hugetlb_sysfs_add_hstate(struct hstate *h, struct kobject *parent, in hugetlb_sysfs_add_hstate() argument 3001 hstate_kobjs[hi] = kobject_create_and_add(h->name, parent); in hugetlb_sysfs_add_hstate()
|
D | shmem.c | 3350 struct inode *parent) in shmem_encode_fh() argument
|