Home
last modified time | relevance | path

Searched refs:root (Results 1 – 18 of 18) sorted by relevance

/mm/
Drmap.c92 anon_vma->root = anon_vma; in anon_vma_alloc()
120 if (rwsem_is_locked(&anon_vma->root->rwsem)) { in anon_vma_free()
232 static inline struct anon_vma *lock_anon_vma_root(struct anon_vma *root, struct anon_vma *anon_vma) in lock_anon_vma_root() argument
234 struct anon_vma *new_root = anon_vma->root; in lock_anon_vma_root()
235 if (new_root != root) { in lock_anon_vma_root()
236 if (WARN_ON_ONCE(root)) in lock_anon_vma_root()
237 up_write(&root->rwsem); in lock_anon_vma_root()
238 root = new_root; in lock_anon_vma_root()
239 down_write(&root->rwsem); in lock_anon_vma_root()
241 return root; in lock_anon_vma_root()
[all …]
Dinterval_tree.c30 struct rb_root_cached *root) in vma_interval_tree_insert_after() argument
57 rb_insert_augmented(&node->shared.rb, &root->rb_root, in vma_interval_tree_insert_after()
76 struct rb_root_cached *root) in INTERVAL_TREE_DEFINE()
82 __anon_vma_interval_tree_insert(node, root); in INTERVAL_TREE_DEFINE()
86 struct rb_root_cached *root) in anon_vma_interval_tree_remove() argument
88 __anon_vma_interval_tree_remove(node, root); in anon_vma_interval_tree_remove()
92 anon_vma_interval_tree_iter_first(struct rb_root_cached *root, in anon_vma_interval_tree_iter_first() argument
95 return __anon_vma_interval_tree_iter_first(root, first, last); in anon_vma_interval_tree_iter_first()
Dksm.c591 struct rb_root *root) in alloc_stable_node_chain() argument
609 rb_replace_node(&dup->node, &chain->node, root); in alloc_stable_node_chain()
624 struct rb_root *root) in free_stable_node_chain() argument
626 rb_erase(&chain->node, root); in free_stable_node_chain()
914 struct rb_root *root) in remove_stable_node_chain() argument
934 free_stable_node_chain(stable_node, root); in remove_stable_node_chain()
1356 struct rb_root *root, in stable_node_dup() argument
1429 root); in stable_node_dup()
1474 struct rb_root *root) in stable_node_dup_any() argument
1479 free_stable_node_chain(stable_node, root); in stable_node_dup_any()
[all …]
Dcleancache.c306 struct dentry *root = debugfs_create_dir("cleancache", NULL); in init_cleancache() local
308 debugfs_create_u64("succ_gets", 0444, root, &cleancache_succ_gets); in init_cleancache()
309 debugfs_create_u64("failed_gets", 0444, root, &cleancache_failed_gets); in init_cleancache()
310 debugfs_create_u64("puts", 0444, root, &cleancache_puts); in init_cleancache()
311 debugfs_create_u64("invalidates", 0444, root, &cleancache_invalidates); in init_cleancache()
Dfrontswap.c485 struct dentry *root = debugfs_create_dir("frontswap", NULL); in init_frontswap() local
486 if (root == NULL) in init_frontswap()
488 debugfs_create_u64("loads", 0444, root, &frontswap_loads); in init_frontswap()
489 debugfs_create_u64("succ_stores", 0444, root, &frontswap_succ_stores); in init_frontswap()
490 debugfs_create_u64("failed_stores", 0444, root, in init_frontswap()
492 debugfs_create_u64("invalidates", 0444, root, &frontswap_invalidates); in init_frontswap()
Dsparse.c85 unsigned long root = SECTION_NR_TO_ROOT(section_nr); in sparse_index_init() local
95 if (mem_section[root]) in sparse_index_init()
102 mem_section[root] = section; in sparse_index_init()
117 struct mem_section *root = NULL; in __section_nr() local
120 root = __nr_to_section(root_nr * SECTIONS_PER_ROOT); in __section_nr()
121 if (!root) in __section_nr()
124 if ((ms >= root) && (ms < (root + SECTIONS_PER_ROOT))) in __section_nr()
128 VM_BUG_ON(!root); in __section_nr()
130 return (root_nr * SECTIONS_PER_ROOT) + (ms - root); in __section_nr()
Dmmap.c330 struct rb_root *root = &mm->mm_rb; in browse_rb() local
335 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in browse_rb()
376 static void validate_mm_rb(struct rb_root *root, struct vm_area_struct *ignore) in validate_mm_rb() argument
380 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in validate_mm_rb()
429 #define validate_mm_rb(root, ignore) do { } while (0) argument
452 struct rb_root *root) in vma_rb_insert() argument
455 validate_mm_rb(root, NULL); in vma_rb_insert()
457 rb_insert_augmented(&vma->vm_rb, root, &vma_gap_callbacks); in vma_rb_insert()
460 static void __vma_rb_erase(struct vm_area_struct *vma, struct rb_root *root) in __vma_rb_erase() argument
467 rb_erase_augmented(&vma->vm_rb, root, &vma_gap_callbacks); in __vma_rb_erase()
[all …]
Dvmalloc.c438 struct rb_root *root, struct rb_node *from, in find_va_links() argument
444 if (root) { in find_va_links()
445 link = &root->rb_node; in find_va_links()
500 link_va(struct vmap_area *va, struct rb_root *root, in link_va() argument
515 if (root == &free_vmap_area_root) { in link_va()
528 root, &free_vmap_area_rb_augment_cb); in link_va()
531 rb_insert_color(&va->rb_node, root); in link_va()
539 unlink_va(struct vmap_area *va, struct rb_root *root) in unlink_va() argument
544 if (root == &free_vmap_area_root) in unlink_va()
546 root, &free_vmap_area_rb_augment_cb); in unlink_va()
[all …]
Dzswap.c267 static struct zswap_entry *zswap_rb_search(struct rb_root *root, pgoff_t offset) in zswap_rb_search() argument
269 struct rb_node *node = root->rb_node; in zswap_rb_search()
288 static int zswap_rb_insert(struct rb_root *root, struct zswap_entry *entry, in zswap_rb_insert() argument
291 struct rb_node **link = &root->rb_node, *parent = NULL; in zswap_rb_insert()
307 rb_insert_color(&entry->rbnode, root); in zswap_rb_insert()
311 static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry) in zswap_rb_erase() argument
314 rb_erase(&entry->rbnode, root); in zswap_rb_erase()
358 static struct zswap_entry *zswap_entry_find_get(struct rb_root *root, in zswap_entry_find_get() argument
363 entry = zswap_rb_search(root, offset); in zswap_entry_find_get()
Dslab.h288 #define for_each_memcg_cache(iter, root) \ argument
289 list_for_each_entry(iter, &(root)->memcg_params.children, \
417 #define for_each_memcg_cache(iter, root) \ argument
418 for ((void)(iter), (void)(root); 0; )
Dmemcontrol.c239 #define for_each_mem_cgroup_tree(iter, root) \ argument
240 for (iter = mem_cgroup_iter(root, NULL, NULL); \
242 iter = mem_cgroup_iter(root, iter, NULL))
1025 struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root, in mem_cgroup_iter() argument
1037 if (!root) in mem_cgroup_iter()
1038 root = root_mem_cgroup; in mem_cgroup_iter()
1043 if (!root->use_hierarchy && root != root_mem_cgroup) { in mem_cgroup_iter()
1046 return root; in mem_cgroup_iter()
1054 mz = mem_cgroup_nodeinfo(root, reclaim->pgdat->node_id); in mem_cgroup_iter()
1080 css = css_next_descendant_pre(css, &root->css); in mem_cgroup_iter()
[all …]
Dmempolicy.c2291 struct rb_node *n = sp->root.rb_node; in sp_lookup()
2324 struct rb_node **p = &sp->root.rb_node; in sp_insert()
2339 rb_insert_color(&new->nd, &sp->root); in sp_insert()
2351 if (!sp->root.rb_node) in mpol_shared_policy_lookup()
2472 rb_erase(&n->nd, &sp->root); in sp_delete()
2585 sp->root = RB_ROOT; /* empty tree == default mempolicy */ in mpol_shared_policy_init()
2650 if (!p->root.rb_node) in mpol_free_shared_policy()
2653 next = rb_first(&p->root); in mpol_free_shared_policy()
Dmemblock.c2008 struct dentry *root = debugfs_create_dir("memblock", NULL); in memblock_init_debugfs() local
2010 debugfs_create_file("memory", 0444, root, in memblock_init_debugfs()
2012 debugfs_create_file("reserved", 0444, root, in memblock_init_debugfs()
2015 debugfs_create_file("physmem", 0444, root, in memblock_init_debugfs()
Dvmscan.c2760 struct mem_cgroup *root = sc->target_mem_cgroup; in shrink_node() local
2769 memcg = mem_cgroup_iter(root, NULL, NULL); in shrink_node()
2775 switch (mem_cgroup_protected(root, memcg)) { in shrink_node()
2819 } while ((memcg = mem_cgroup_iter(root, memcg, NULL))); in shrink_node()
2883 set_memcg_congestion(pgdat, root, true); in shrink_node()
2892 current_may_throttle() && pgdat_memcg_congested(pgdat, root)) in shrink_node()
Dslab.c3880 struct kmem_cache *root = memcg_root_cache(cachep); in enable_cpucache() local
3881 limit = root->limit; in enable_cpucache()
3882 shared = root->shared; in enable_cpucache()
3883 batchcount = root->batchcount; in enable_cpucache()
Dshmem.c3541 struct shmem_sb_info *sbinfo = SHMEM_SB(fc->root->d_sb); in shmem_reconfigure()
3593 static int shmem_show_options(struct seq_file *seq, struct dentry *root) in shmem_show_options() argument
3595 struct shmem_sb_info *sbinfo = SHMEM_SB(root->d_sb); in shmem_show_options()
Dmemory.c2673 static inline void unmap_mapping_range_tree(struct rb_root_cached *root, in unmap_mapping_range_tree() argument
2679 vma_interval_tree_foreach(vma, root, in unmap_mapping_range_tree()
DKconfig301 root has set /sys/kernel/mm/ksm/run to 1 (if CONFIG_SYSFS is set).