/kernel/cgroup/ |
D | cgroup-v1.c | 59 struct cgroup_root *root; in cgroup_attach_task_all() local 65 for_each_root(root) { in cgroup_attach_task_all() 68 if (root == &cgrp_dfl_root) in cgroup_attach_task_all() 72 from_cgrp = task_cgroup_from_root(from, root); in cgroup_attach_task_all() 558 BUILD_BUG_ON(sizeof(cgrp->root->release_agent_path) < PATH_MAX); in cgroup_release_agent_write() 573 strlcpy(cgrp->root->release_agent_path, strstrip(buf), in cgroup_release_agent_write() 574 sizeof(cgrp->root->release_agent_path)); in cgroup_release_agent_write() 585 seq_puts(seq, cgrp->root->release_agent_path); in cgroup_release_agent_show() 690 ss->legacy_name, ss->root->hierarchy_id, in proc_cgroupstats_show() 691 atomic_read(&ss->root->nr_cgrps), in proc_cgroupstats_show() [all …]
|
D | cgroup.c | 318 return cgrp->root == &cgrp_dfl_root; in cgroup_on_dfl() 439 u16 root_ss_mask = cgrp->root->subsys_mask; in cgroup_control() 470 return cgrp->root->subsys_mask; in cgroup_ss_mask() 1059 BUG_ON(cgrp1->root != cgrp2->root); in compare_css_sets() 1068 if (cgrp1->root == new_cgrp->root) { in compare_css_sets() 1089 struct cgroup_root *root = cgrp->root; in find_existing_css_set() local 1101 if (root->subsys_mask & (1UL << i)) { in find_existing_css_set() 1260 if (c->root == cgrp->root) in find_css_set() 1312 return root_cgrp->root; in cgroup_root_from_kf() 1315 static int cgroup_init_root_id(struct cgroup_root *root) in cgroup_init_root_id() argument [all …]
|
D | cgroup-internal.h | 48 struct cgroup_root *root; member 173 #define for_each_root(root) \ argument 174 list_for_each_entry((root), &cgroup_roots, root_list) 229 struct cgroup_root *root); 235 void cgroup_free_root(struct cgroup_root *root); 237 int cgroup_setup_root(struct cgroup_root *root, u16 ss_mask);
|
D | rstat.c | 88 struct cgroup *root, int cpu) in cgroup_rstat_cpu_pop_updated() argument 92 if (pos == root) in cgroup_rstat_cpu_pop_updated() 100 pos = root; in cgroup_rstat_cpu_pop_updated()
|
D | debug.c | 105 c->root->hierarchy_id, name_buf); in current_css_set_cg_links_read()
|
D | cpuset.c | 414 (cpuset_cgrp_subsys.root->flags & CGRP_ROOT_CPUSET_V2_MODE); in is_in_v2_mode()
|
/kernel/ |
D | resource.c | 98 struct resource *root = PDE_DATA(file_inode(m->file)); in r_show() local 101 int width = root->end < 0x10000 ? 4 : 8; in r_show() 105 if (p->parent == root) in r_show() 159 static struct resource * __request_resource(struct resource *root, struct resource *new) in __request_resource() argument 166 return root; in __request_resource() 167 if (start < root->start) in __request_resource() 168 return root; in __request_resource() 169 if (end > root->end) in __request_resource() 170 return root; in __request_resource() 171 p = &root->child; in __request_resource() [all …]
|
D | audit_tree.c | 16 struct audit_chunk *root; member 105 tree->root = NULL; in alloc_tree() 301 owner->root = new; in replace_chunk() 329 if (owner->root == chunk) { in remove_chunk_node() 331 owner->root = NULL; in remove_chunk_node() 436 if (!tree->root) { in create_chunk() 437 tree->root = chunk; in create_chunk() 508 if (!tree->root) { in tag_chunk() 509 tree->root = chunk; in tag_chunk() 631 if (!tree->root && !tree->goner) { in trim_marked() [all …]
|
D | ucount.c | 30 set_lookup(struct ctl_table_root *root) in set_lookup() argument
|
D | nsproxy.c | 508 set_fs_root(me->fs, &nsset->fs->root); in commit_nsset()
|
D | module.c | 157 struct latch_tree_root root; member 169 latch_tree_insert(&node->node, &mod_tree.root, &mod_tree_ops); in __mod_tree_insert() 174 latch_tree_erase(&node->node, &mod_tree.root, &mod_tree_ops); in __mod_tree_remove() 207 ltn = latch_tree_find((void *)addr, &mod_tree.root, &mod_tree_ops); in mod_find()
|
/kernel/bpf/ |
D | local_storage.c | 24 struct rb_root root; member 69 struct rb_root *root = &map->root; in cgroup_storage_lookup() local 75 node = root->rb_node; in cgroup_storage_lookup() 104 struct rb_root *root = &map->root; in cgroup_storage_insert() local 105 struct rb_node **new = &(root->rb_node), *parent = NULL; in cgroup_storage_insert() 126 rb_insert_color(&storage->node, root); in cgroup_storage_insert() 325 map->root = RB_ROOT; in cgroup_storage_map_alloc() 346 WARN_ON(!RB_EMPTY_ROOT(&map->root)); in cgroup_storage_map_free() 592 struct rb_root *root; in bpf_cgroup_storage_unlink() local 600 root = &map->root; in bpf_cgroup_storage_unlink() [all …]
|
D | lpm_trie.c | 33 struct lpm_trie_node __rcu *root; member 238 for (node = rcu_dereference_check(trie->root, rcu_read_lock_bh_held()); in trie_lookup_elem() 350 slot = &trie->root; in trie_update_elem() 458 trim = &trie->root; in trie_delete_elem() 590 slot = &trie->root; in trie_free() 639 search_root = rcu_dereference(trie->root); in trie_get_next_key()
|
D | inode.c | 604 static int bpf_show_options(struct seq_file *m, struct dentry *root) in bpf_show_options() argument 606 umode_t mode = d_inode(root)->i_mode & S_IALLUGO & ~S_ISVTX; in bpf_show_options()
|
/kernel/trace/ |
D | trace_stat.c | 75 static int insert_stat(struct rb_root *root, void *stat, cmp_func_t cmp) in insert_stat() argument 77 struct rb_node **new = &(root->rb_node), *parent = NULL; in insert_stat() 104 rb_insert_color(&data->node, root); in insert_stat() 126 struct rb_root *root = &session->stat_root; in stat_seq_init() local 141 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init() 155 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init()
|
/kernel/dma/ |
D | pool.c | 38 struct dentry *root; in dma_atomic_pool_debugfs_init() local 40 root = debugfs_create_dir("dma_pools", NULL); in dma_atomic_pool_debugfs_init() 41 debugfs_create_ulong("pool_size_dma", 0400, root, &pool_size_dma); in dma_atomic_pool_debugfs_init() 42 debugfs_create_ulong("pool_size_dma32", 0400, root, &pool_size_dma32); in dma_atomic_pool_debugfs_init() 43 debugfs_create_ulong("pool_size_kernel", 0400, root, &pool_size_kernel); in dma_atomic_pool_debugfs_init()
|
/kernel/locking/ |
D | lockdep.c | 2255 find_usage_forwards(struct lock_list *root, unsigned long usage_mask, in find_usage_forwards() argument 2262 result = __bfs_forwards(root, &usage_mask, usage_match, usage_skip, target_entry); in find_usage_forwards() 2272 find_usage_backwards(struct lock_list *root, unsigned long usage_mask, in find_usage_backwards() argument 2279 result = __bfs_backwards(root, &usage_mask, usage_match, usage_skip, target_entry); in find_usage_backwards() 2364 struct lock_list *root) in print_shortest_lock_dependencies() argument 2378 if (depth == 0 && (entry != root)) { in print_shortest_lock_dependencies() 2410 struct lock_list *root) in print_shortest_lock_dependencies_backwards() argument 2433 if (depth == 0 && (entry != root)) { in print_shortest_lock_dependencies_backwards() 3944 struct lock_list *root, struct lock_list *other, in print_irq_inversion_bug() argument 3975 if (depth == 0 && (entry != root)) { in print_irq_inversion_bug() [all …]
|
/kernel/irq/ |
D | internals.h | 500 void irq_domain_debugfs_init(struct dentry *root); 502 static inline void irq_domain_debugfs_init(struct dentry *root) in irq_domain_debugfs_init() argument
|
D | irqdomain.c | 1994 void __init irq_domain_debugfs_init(struct dentry *root) in irq_domain_debugfs_init() argument 1998 domain_dir = debugfs_create_dir("domains", root); in irq_domain_debugfs_init()
|
/kernel/sched/ |
D | deadline.c | 413 return dl_rq->root.rb_leftmost == &dl_se->rb_node; in is_leftmost() 439 dl_rq->root = RB_ROOT_CACHED; in init_dl_rq() 550 struct rb_root_cached *root = &dl_rq->pushable_dl_tasks_root; in dequeue_pushable_dl_task() local 556 leftmost = rb_erase_cached(&p->pushable_dl_tasks, root); in dequeue_pushable_dl_task() 1425 struct rb_node *leftmost = dl_rq->root.rb_leftmost; in dec_dl_deadline() 1483 rb_add_cached(&dl_se->rb_node, &dl_rq->root, __dl_less); in __enqueue_dl_entity() 1495 rb_erase_cached(&dl_se->rb_node, &dl_rq->root); in __dequeue_dl_entity() 1859 struct rb_node *left = rb_first_cached(&dl_rq->root); in pick_next_dl_entity()
|
D | sched.h | 690 struct rb_root_cached root; member
|
/kernel/rcu/ |
D | Kconfig | 129 root of NR_CPUS, which allows NR_CPUS to be insanely large.
|