/kernel/ |
D | ptrace.c | 68 void __ptrace_link(struct task_struct *child, struct task_struct *new_parent, in __ptrace_link() argument 71 BUG_ON(!list_empty(&child->ptrace_entry)); in __ptrace_link() 72 list_add(&child->ptrace_entry, &new_parent->ptraced); in __ptrace_link() 73 child->parent = new_parent; in __ptrace_link() 74 child->ptracer_cred = get_cred(ptracer_cred); in __ptrace_link() 83 static void ptrace_link(struct task_struct *child, struct task_struct *new_parent) in ptrace_link() argument 85 __ptrace_link(child, new_parent, current_cred()); in ptrace_link() 116 void __ptrace_unlink(struct task_struct *child) in __ptrace_unlink() argument 119 BUG_ON(!child->ptrace); in __ptrace_unlink() 121 clear_task_syscall_work(child, SYSCALL_TRACE); in __ptrace_unlink() [all …]
|
D | resource.c | 61 if (p->child) in next_resource() 62 return p->child; in next_resource() 85 for (p = p->child; p && l < *pos; p = r_next(m, p, &l)) in r_start() 171 p = &root->child; in __request_resource() 191 p = &old->parent->child; in __release_resource() 197 if (release_child || !(tmp->child)) { in __release_resource() 200 for (chd = tmp->child;; chd = chd->sibling) { in __release_resource() 205 *p = tmp->child; in __release_resource() 221 p = r->child; in __release_child_resources() 222 r->child = NULL; in __release_child_resources() [all …]
|
D | sysctl.c | 2195 .child = random_table, 2200 .child = usermodehelper_table, 2206 .child = firmware_config_table, 2596 .child = key_sysctls, 3305 .child = inotify_table, 3312 .child = fanotify_table, 3319 .child = epoll_table, 3372 .child = sysctl_mount_point, 3439 .child = kern_table, 3444 .child = vm_table, [all …]
|
D | fork.c | 1414 static int wait_for_vfork_done(struct task_struct *child, in wait_for_vfork_done() argument 1426 task_lock(child); in wait_for_vfork_done() 1427 child->vfork_done = NULL; in wait_for_vfork_done() 1428 task_unlock(child); in wait_for_vfork_done() 1431 put_task_struct(child); in wait_for_vfork_done() 2950 struct task_struct *leader, *parent, *child; in walk_process_tree() local 2957 list_for_each_entry(child, &parent->children, sibling) { in walk_process_tree() 2958 res = visitor(child, data); in walk_process_tree() 2962 leader = child; in walk_process_tree() 2971 child = leader; in walk_process_tree() [all …]
|
D | utsname_sysctl.c | 120 .child = uts_kern_table,
|
D | seccomp.c | 463 struct seccomp_filter *child) in is_ancestor() argument 468 for (; child; child = child->prev) in is_ancestor() 469 if (child == parent) in is_ancestor()
|
D | user_namespace.c | 1296 const struct user_namespace *child) in in_userns() argument 1299 for (ns = child; ns->level > ancestor->level; ns = ns->parent) in in_userns()
|
/kernel/bpf/ |
D | lpm_trie.c | 25 struct lpm_trie_node __rcu *child[2]; member 271 node = rcu_dereference_check(node->child[next_bit], in trie_lookup_elem() 341 RCU_INIT_POINTER(new_node->child[0], NULL); in trie_update_elem() 342 RCU_INIT_POINTER(new_node->child[1], NULL); in trie_update_elem() 362 slot = &node->child[next_bit]; in trie_update_elem() 377 new_node->child[0] = node->child[0]; in trie_update_elem() 378 new_node->child[1] = node->child[1]; in trie_update_elem() 394 rcu_assign_pointer(new_node->child[next_bit], node); in trie_update_elem() 411 rcu_assign_pointer(im_node->child[0], node); in trie_update_elem() 412 rcu_assign_pointer(im_node->child[1], new_node); in trie_update_elem() [all …]
|
/kernel/sched/ |
D | topology.c | 63 if ((meta_flags & SDF_SHARED_CHILD) && sd->child && in sched_domain_debug_one() 64 !(sd->child->flags & flag)) in sched_domain_debug_one() 110 if (group == sd->groups && sd->child && in sched_domain_debug_one() 111 !cpumask_equal(sched_domain_span(sd->child), in sched_domain_debug_one() 695 parent->parent->child = tmp; in cpu_attach_domain() 713 sd->child = NULL; in cpu_attach_domain() 880 if (!sibling->child) in build_balance_mask() 884 if (!cpumask_equal(sg_span, sched_domain_span(sibling->child))) in build_balance_mask() 912 if (sd->child) in build_group_from_child_sched_domain() 913 cpumask_copy(sg_span, sched_domain_span(sd->child)); in build_group_from_child_sched_domain() [all …]
|
D | rt.c | 2705 struct task_group *child; in tg_rt_schedulable() local 2741 list_for_each_entry_rcu(child, &tg->children, siblings) { in tg_rt_schedulable() 2742 period = ktime_to_ns(child->rt_bandwidth.rt_period); in tg_rt_schedulable() 2743 runtime = child->rt_bandwidth.rt_runtime; in tg_rt_schedulable() 2745 if (child == d->tg) { in tg_rt_schedulable()
|
D | fair.c | 6346 sd = sd->child; in find_idlest_cpu() 6352 sd = sd->child; in find_idlest_cpu() 6359 sd = sd->child; in find_idlest_cpu() 8868 struct sched_domain *child = sd->child; in update_group_capacity() local 8877 if (!child) { in update_group_capacity() 8886 if (child->flags & SD_OVERLAP) { in update_group_capacity() 8905 group = child->groups; in update_group_capacity() 8913 } while (group != child->groups); in update_group_capacity() 9692 struct sched_domain *child = env->sd->child; in update_sd_lb_stats() local 9734 sds->prefer_sibling = child && child->flags & SD_PREFER_SIBLING; in update_sd_lb_stats()
|
D | core.c | 1199 struct task_group *parent, *child; in walk_tg_tree_from() local 1208 list_for_each_entry_rcu(child, &parent->children, siblings) { in walk_tg_tree_from() 1209 parent = child; in walk_tg_tree_from() 1219 child = parent; in walk_tg_tree_from()
|
/kernel/livepatch/ |
D | transition.c | 611 void klp_copy_process(struct task_struct *child) in klp_copy_process() argument 625 set_tsk_thread_flag(child, TIF_PATCH_PENDING); in klp_copy_process() 627 clear_tsk_thread_flag(child, TIF_PATCH_PENDING); in klp_copy_process() 629 child->patch_state = current->patch_state; in klp_copy_process()
|
/kernel/cgroup/ |
D | cgroup.c | 730 #define cgroup_for_each_live_child(child, cgrp) \ argument 731 list_for_each_entry((child), &(cgrp)->self.children, self.sibling) \ 733 cgroup_is_dead(child); })) \ 826 struct cgroup *child = NULL; in cgroup_update_populated() local 834 if (!child) { in cgroup_update_populated() 837 if (cgroup_is_threaded(child)) in cgroup_update_populated() 851 child = cgrp; in cgroup_update_populated() 3362 struct cgroup *cgrp, *child; in cgroup_subtree_control_write() local 3417 cgroup_for_each_live_child(child, cgrp) { in cgroup_subtree_control_write() 3418 if (child->subtree_control & (1 << ssid)) { in cgroup_subtree_control_write() [all …]
|
D | legacy_freezer.c | 260 struct freezer *child = css_freezer(pos); in update_if_frozen() local 262 if ((child->state & CGROUP_FREEZER_ONLINE) && in update_if_frozen() 263 !(child->state & CGROUP_FROZEN)) { in update_if_frozen()
|
/kernel/events/ |
D | core.c | 3297 struct perf_event *child; in perf_event_modify_attr() local 3324 list_for_each_entry(child, &event->child_list, child_list) { in perf_event_modify_attr() 3325 perf_event_modify_copy_attr(&child->attr, attr); in perf_event_modify_attr() 3326 err = func(child, attr); in perf_event_modify_attr() 5230 struct perf_event *child, *tmp; in perf_event_release_kernel() local 5266 list_for_each_entry(child, &event->child_list, child_list) { in perf_event_release_kernel() 5272 ctx = READ_ONCE(child->ctx); in perf_event_release_kernel() 5299 if (tmp == child) { in perf_event_release_kernel() 5300 perf_remove_from_context(child, DETACH_GROUP); in perf_event_release_kernel() 5301 list_move(&child->child_list, &free_list); in perf_event_release_kernel() [all …]
|
/kernel/irq/ |
D | irqdomain.c | 1214 struct irq_data *child) in irq_domain_insert_irq_data() argument 1219 irq_data_get_node(child)); in irq_domain_insert_irq_data() 1221 child->parent_data = irq_data; in irq_domain_insert_irq_data() 1222 irq_data->irq = child->irq; in irq_domain_insert_irq_data() 1223 irq_data->common = child->common; in irq_domain_insert_irq_data()
|
/kernel/trace/ |
D | trace_events.c | 975 struct dentry *child; in remove_event_file_dir() local 979 list_for_each_entry(child, &dir->d_subdirs, d_child) { in remove_event_file_dir() 980 if (d_really_is_positive(child)) /* probably unneeded */ in remove_event_file_dir() 981 d_inode(child)->i_private = NULL; in remove_event_file_dir()
|
/kernel/locking/ |
D | lockdep.c | 1489 static inline struct lock_list *get_lock_parent(struct lock_list *child) in get_lock_parent() argument 1491 return child->parent; in get_lock_parent() 1494 static inline int get_lock_depth(struct lock_list *child) in get_lock_depth() argument 1499 while ((parent = get_lock_parent(child))) { in get_lock_depth() 1500 child = parent; in get_lock_depth()
|