Searched refs:shared (Results 1 – 5 of 5) sorted by relevance
/mm/ |
D | interval_tree.c | 23 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb, 24 unsigned long, shared.rb_subtree_last, 38 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after() 40 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after() 42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after() 43 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 44 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 45 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after() 47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after() [all …]
|
D | slab.c | 234 parent->shared = NULL; in kmem_cache_node_init() 386 .shared = 1, 708 if (n->shared) in __drain_alien_cache() 709 transfer_objects(n->shared, ac, ac->limit); in __drain_alien_cache() 899 if (cachep->shared) { in setup_kmem_cache_node() 901 cachep->shared * cachep->batchcount, 0xbaadf00d, gfp); in setup_kmem_cache_node() 912 if (n->shared && force_change) { in setup_kmem_cache_node() 913 free_block(cachep, n->shared->entry, in setup_kmem_cache_node() 914 n->shared->avail, node, &list); in setup_kmem_cache_node() 915 n->shared->avail = 0; in setup_kmem_cache_node() [all …]
|
D | slab.h | 204 unsigned int shared; member 576 struct array_cache *shared; /* shared per node */ member
|
D | khugepaged.c | 608 int none_or_zero = 0, shared = 0, result = 0, referenced = 0; in __collapse_huge_page_isolate() local 641 ++shared > khugepaged_max_ptes_shared) { in __collapse_huge_page_isolate() 1237 int none_or_zero = 0, shared = 0; in khugepaged_scan_pmd() local 1306 ++shared > khugepaged_max_ptes_shared) { in khugepaged_scan_pmd()
|
D | slab_common.c | 1094 sinfo.limit, sinfo.batchcount, sinfo.shared); in cache_show()
|