/kernel/cgroup/ |
D | cgroup-v1.c | 61 struct cgroup_root *root; in cgroup_attach_task_all() local 66 for_each_root(root) { in cgroup_attach_task_all() 69 if (root == &cgrp_dfl_root) in cgroup_attach_task_all() 73 from_cgrp = task_cgroup_from_root(from, root); in cgroup_attach_task_all() 550 BUILD_BUG_ON(sizeof(cgrp->root->release_agent_path) < PATH_MAX); in cgroup_release_agent_write() 556 strlcpy(cgrp->root->release_agent_path, strstrip(buf), in cgroup_release_agent_write() 557 sizeof(cgrp->root->release_agent_path)); in cgroup_release_agent_write() 568 seq_puts(seq, cgrp->root->release_agent_path); in cgroup_release_agent_show() 673 ss->legacy_name, ss->root->hierarchy_id, in proc_cgroupstats_show() 674 atomic_read(&ss->root->nr_cgrps), in proc_cgroupstats_show() [all …]
|
D | cgroup.c | 300 return cgrp->root == &cgrp_dfl_root; in cgroup_on_dfl() 421 u16 root_ss_mask = cgrp->root->subsys_mask; in cgroup_control() 452 return cgrp->root->subsys_mask; in cgroup_ss_mask() 1034 BUG_ON(cgrp1->root != cgrp2->root); in compare_css_sets() 1043 if (cgrp1->root == new_cgrp->root) { in compare_css_sets() 1064 struct cgroup_root *root = cgrp->root; in find_existing_css_set() local 1076 if (root->subsys_mask & (1UL << i)) { in find_existing_css_set() 1234 if (c->root == cgrp->root) in find_css_set() 1286 return root_cgrp->root; in cgroup_root_from_kf() 1289 static int cgroup_init_root_id(struct cgroup_root *root) in cgroup_init_root_id() argument [all …]
|
D | cgroup-internal.h | 48 struct cgroup_root *root; member 156 #define for_each_root(root) \ argument 157 list_for_each_entry((root), &cgroup_roots, root_list) 212 struct cgroup_root *root); 218 void cgroup_free_root(struct cgroup_root *root); 220 int cgroup_setup_root(struct cgroup_root *root, u16 ss_mask);
|
D | rstat.c | 88 struct cgroup *root, int cpu) in cgroup_rstat_cpu_pop_updated() argument 92 if (pos == root) in cgroup_rstat_cpu_pop_updated() 100 pos = root; in cgroup_rstat_cpu_pop_updated()
|
D | debug.c | 105 c->root->hierarchy_id, name_buf); in current_css_set_cg_links_read()
|
D | cpuset.c | 367 (cpuset_cgrp_subsys.root->flags & CGRP_ROOT_CPUSET_V2_MODE); in is_in_v2_mode()
|
/kernel/ |
D | resource.c | 107 struct resource *root = PDE_DATA(file_inode(m->file)); in r_show() local 110 int width = root->end < 0x10000 ? 4 : 8; in r_show() 114 if (p->parent == root) in r_show() 185 static struct resource * __request_resource(struct resource *root, struct resource *new) in __request_resource() argument 192 return root; in __request_resource() 193 if (start < root->start) in __request_resource() 194 return root; in __request_resource() 195 if (end > root->end) in __request_resource() 196 return root; in __request_resource() 197 p = &root->child; in __request_resource() [all …]
|
D | audit_tree.c | 16 struct audit_chunk *root; member 105 tree->root = NULL; in alloc_tree() 303 owner->root = new; in replace_chunk() 331 if (owner->root == chunk) { in remove_chunk_node() 333 owner->root = NULL; in remove_chunk_node() 438 if (!tree->root) { in create_chunk() 439 tree->root = chunk; in create_chunk() 510 if (!tree->root) { in tag_chunk() 511 tree->root = chunk; in tag_chunk() 633 if (!tree->root && !tree->goner) { in trim_marked() [all …]
|
D | ucount.c | 24 set_lookup(struct ctl_table_root *root) in set_lookup() argument
|
D | module.c | 148 struct latch_tree_root root; member 160 latch_tree_insert(&node->node, &mod_tree.root, &mod_tree_ops); in __mod_tree_insert() 165 latch_tree_erase(&node->node, &mod_tree.root, &mod_tree_ops); in __mod_tree_remove() 198 ltn = latch_tree_find((void *)addr, &mod_tree.root, &mod_tree_ops); in mod_find()
|
/kernel/bpf/ |
D | local_storage.c | 24 struct rb_root root; member 52 struct rb_root *root = &map->root; in cgroup_storage_lookup() local 58 node = root->rb_node; in cgroup_storage_lookup() 87 struct rb_root *root = &map->root; in cgroup_storage_insert() local 88 struct rb_node **new = &(root->rb_node), *parent = NULL; in cgroup_storage_insert() 109 rb_insert_color(&storage->node, root); in cgroup_storage_insert() 312 map->root = RB_ROOT; in cgroup_storage_map_alloc() 322 WARN_ON(!RB_EMPTY_ROOT(&map->root)); in cgroup_storage_map_free() 585 struct rb_root *root; in bpf_cgroup_storage_unlink() local 593 root = &map->root; in bpf_cgroup_storage_unlink() [all …]
|
D | lpm_trie.c | 33 struct lpm_trie_node __rcu *root; member 235 for (node = rcu_dereference(trie->root); node;) { in trie_lookup_elem() 345 slot = &trie->root; in trie_update_elem() 453 trim = &trie->root; in trie_delete_elem() 603 slot = &trie->root; in trie_free() 652 search_root = rcu_dereference(trie->root); in trie_get_next_key()
|
D | inode.c | 558 static int bpf_show_options(struct seq_file *m, struct dentry *root) in bpf_show_options() argument 560 umode_t mode = d_inode(root)->i_mode & S_IALLUGO & ~S_ISVTX; in bpf_show_options()
|
/kernel/trace/ |
D | trace_stat.c | 77 static int insert_stat(struct rb_root *root, void *stat, cmp_stat_t cmp) in insert_stat() argument 79 struct rb_node **new = &(root->rb_node), *parent = NULL; in insert_stat() 106 rb_insert_color(&data->node, root); in insert_stat() 128 struct rb_root *root = &session->stat_root; in stat_seq_init() local 143 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init() 157 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init()
|
/kernel/locking/ |
D | lockdep.c | 1877 find_usage_forwards(struct lock_list *root, unsigned long usage_mask, in find_usage_forwards() argument 1884 result = __bfs_forwards(root, &usage_mask, usage_match, target_entry); in find_usage_forwards() 1900 find_usage_backwards(struct lock_list *root, unsigned long usage_mask, in find_usage_backwards() argument 1907 result = __bfs_backwards(root, &usage_mask, usage_match, target_entry); in find_usage_backwards() 1943 struct lock_list *root) in print_shortest_lock_dependencies() argument 1957 if (depth == 0 && (entry != root)) { in print_shortest_lock_dependencies() 3123 struct lock_list *root, struct lock_list *other, in print_irq_inversion_bug() argument 3154 if (depth == 0 && (entry != root)) { in print_irq_inversion_bug() 3161 } while (entry && entry != root && (depth >= 0)); in print_irq_inversion_bug() 3163 print_irq_lock_scenario(root, other, in print_irq_inversion_bug() [all …]
|
/kernel/dma/ |
D | swiotlb.c | 708 struct dentry *root; in swiotlb_create_debugfs() local 710 root = debugfs_create_dir("swiotlb", NULL); in swiotlb_create_debugfs() 711 debugfs_create_ulong("io_tlb_nslabs", 0400, root, &io_tlb_nslabs); in swiotlb_create_debugfs() 712 debugfs_create_ulong("io_tlb_used", 0400, root, &io_tlb_used); in swiotlb_create_debugfs()
|
/kernel/irq/ |
D | internals.h | 481 void irq_domain_debugfs_init(struct dentry *root); 483 static inline void irq_domain_debugfs_init(struct dentry *root) in irq_domain_debugfs_init() argument
|
D | irqdomain.c | 1786 void __init irq_domain_debugfs_init(struct dentry *root) in irq_domain_debugfs_init() argument 1790 domain_dir = debugfs_create_dir("domains", root); in irq_domain_debugfs_init()
|
/kernel/sched/ |
D | deadline.c | 334 return dl_rq->root.rb_leftmost == &dl_se->rb_node; in is_leftmost() 358 dl_rq->root = RB_ROOT_CACHED; in init_dl_rq() 1358 struct rb_node *leftmost = dl_rq->root.rb_leftmost; in dec_dl_deadline() 1405 struct rb_node **link = &dl_rq->root.rb_root.rb_node; in __enqueue_dl_entity() 1424 rb_insert_color_cached(&dl_se->rb_node, &dl_rq->root, leftmost); in __enqueue_dl_entity() 1436 rb_erase_cached(&dl_se->rb_node, &dl_rq->root); in __dequeue_dl_entity() 1768 struct rb_node *left = rb_first_cached(&dl_rq->root); in pick_next_dl_entity()
|
D | sched.h | 636 struct rb_root_cached root; member
|
/kernel/rcu/ |
D | Kconfig | 102 root of NR_CPUS, which allows NR_CPUS to be insanely large.
|