Home
last modified time | relevance | path

Searched refs:child (Results 1 – 20 of 20) sorted by relevance

/kernel/
Dptrace.c68 void __ptrace_link(struct task_struct *child, struct task_struct *new_parent, in __ptrace_link() argument
71 BUG_ON(!list_empty(&child->ptrace_entry)); in __ptrace_link()
72 list_add(&child->ptrace_entry, &new_parent->ptraced); in __ptrace_link()
73 child->parent = new_parent; in __ptrace_link()
74 child->ptracer_cred = get_cred(ptracer_cred); in __ptrace_link()
83 static void ptrace_link(struct task_struct *child, struct task_struct *new_parent) in ptrace_link() argument
85 __ptrace_link(child, new_parent, current_cred()); in ptrace_link()
116 void __ptrace_unlink(struct task_struct *child) in __ptrace_unlink() argument
119 BUG_ON(!child->ptrace); in __ptrace_unlink()
121 clear_task_syscall_work(child, SYSCALL_TRACE); in __ptrace_unlink()
[all …]
Dresource.c61 if (p->child) in next_resource()
62 return p->child; in next_resource()
76 for ((_p) = (_root)->child; (_p); \
97 for (p = p->child; p && l < *pos; p = r_next(m, p, &l)) in r_start()
183 p = &root->child; in __request_resource()
203 p = &old->parent->child; in __release_resource()
209 if (release_child || !(tmp->child)) { in __release_resource()
212 for (chd = tmp->child;; chd = chd->sibling) { in __release_resource()
217 *p = tmp->child; in __release_resource()
233 p = r->child; in __release_child_resources()
[all …]
Dfork.c1538 static int wait_for_vfork_done(struct task_struct *child, in wait_for_vfork_done() argument
1549 task_lock(child); in wait_for_vfork_done()
1550 child->vfork_done = NULL; in wait_for_vfork_done()
1551 task_unlock(child); in wait_for_vfork_done()
1554 put_task_struct(child); in wait_for_vfork_done()
3093 struct task_struct *leader, *parent, *child; in walk_process_tree() local
3100 list_for_each_entry(child, &parent->children, sibling) { in walk_process_tree()
3101 res = visitor(child, data); in walk_process_tree()
3105 leader = child; in walk_process_tree()
3114 child = leader; in walk_process_tree()
[all …]
Dutsname_sysctl.c130 .child = uts_kern_table,
Dseccomp.c465 struct seccomp_filter *child) in is_ancestor() argument
470 for (; child; child = child->prev) in is_ancestor()
471 if (child == parent) in is_ancestor()
Duser_namespace.c1301 const struct user_namespace *child) in in_userns() argument
1304 for (ns = child; ns->level > ancestor->level; ns = ns->parent) in in_userns()
Dsysctl.c1824 .child = usermodehelper_table,
2005 .child = key_sysctls,
/kernel/bpf/
Dlpm_trie.c26 struct lpm_trie_node __rcu *child[2]; member
272 node = rcu_dereference_check(node->child[next_bit], in trie_lookup_elem()
342 RCU_INIT_POINTER(new_node->child[0], NULL); in trie_update_elem()
343 RCU_INIT_POINTER(new_node->child[1], NULL); in trie_update_elem()
363 slot = &node->child[next_bit]; in trie_update_elem()
378 new_node->child[0] = node->child[0]; in trie_update_elem()
379 new_node->child[1] = node->child[1]; in trie_update_elem()
395 rcu_assign_pointer(new_node->child[next_bit], node); in trie_update_elem()
412 rcu_assign_pointer(im_node->child[0], node); in trie_update_elem()
413 rcu_assign_pointer(im_node->child[1], new_node); in trie_update_elem()
[all …]
/kernel/sched/
Dtopology.c62 if ((meta_flags & SDF_SHARED_CHILD) && sd->child && in sched_domain_debug_one()
63 !(sd->child->flags & flag)) in sched_domain_debug_one()
109 if (group == sd->groups && sd->child && in sched_domain_debug_one()
110 !cpumask_equal(sched_domain_span(sd->child), in sched_domain_debug_one()
714 parent->parent->child = tmp; in cpu_attach_domain()
743 sd->child = NULL; in cpu_attach_domain()
908 if (!sibling->child) in build_balance_mask()
912 if (!cpumask_equal(sg_span, sched_domain_span(sibling->child))) in build_balance_mask()
940 if (sd->child) { in build_group_from_child_sched_domain()
941 cpumask_copy(sg_span, sched_domain_span(sd->child)); in build_group_from_child_sched_domain()
[all …]
Drt.c2868 struct task_group *child; in tg_rt_schedulable() local
2904 list_for_each_entry_rcu(child, &tg->children, siblings) { in tg_rt_schedulable()
2905 period = ktime_to_ns(child->rt_bandwidth.rt_period); in tg_rt_schedulable()
2906 runtime = child->rt_bandwidth.rt_runtime; in tg_rt_schedulable()
2908 if (child == d->tg) { in tg_rt_schedulable()
Dfair.c6614 sd = sd->child; in find_idlest_cpu()
6620 sd = sd->child; in find_idlest_cpu()
6627 sd = sd->child; in find_idlest_cpu()
9050 struct sched_domain *child = sd->child; in update_group_capacity() local
9059 if (!child) { in update_group_capacity()
9068 if (child->flags & SD_OVERLAP) { in update_group_capacity()
9087 group = child->groups; in update_group_capacity()
9095 } while (group != child->groups); in update_group_capacity()
9996 struct sched_domain *child = env->sd->child; in update_sd_lb_stats() local
10038 sds->prefer_sibling = child && child->flags & SD_PREFER_SIBLING; in update_sd_lb_stats()
Dcore.c1256 struct task_group *parent, *child; in walk_tg_tree_from() local
1265 list_for_each_entry_rcu(child, &parent->children, siblings) { in walk_tg_tree_from()
1266 parent = child; in walk_tg_tree_from()
1276 child = parent; in walk_tg_tree_from()
/kernel/livepatch/
Dtransition.c611 void klp_copy_process(struct task_struct *child) in klp_copy_process() argument
625 set_tsk_thread_flag(child, TIF_PATCH_PENDING); in klp_copy_process()
627 clear_tsk_thread_flag(child, TIF_PATCH_PENDING); in klp_copy_process()
629 child->patch_state = current->patch_state; in klp_copy_process()
/kernel/cgroup/
Dcgroup.c731 #define cgroup_for_each_live_child(child, cgrp) \ argument
732 list_for_each_entry((child), &(cgrp)->self.children, self.sibling) \
734 cgroup_is_dead(child); })) \
827 struct cgroup *child = NULL; in cgroup_update_populated() local
835 if (!child) { in cgroup_update_populated()
838 if (cgroup_is_threaded(child)) in cgroup_update_populated()
852 child = cgrp; in cgroup_update_populated()
3419 struct cgroup *cgrp, *child; in cgroup_subtree_control_write() local
3474 cgroup_for_each_live_child(child, cgrp) { in cgroup_subtree_control_write()
3475 if (child->subtree_control & (1 << ssid)) { in cgroup_subtree_control_write()
[all …]
Dlegacy_freezer.c272 struct freezer *child = css_freezer(pos); in update_if_frozen() local
274 if ((child->state & CGROUP_FREEZER_ONLINE) && in update_if_frozen()
275 !(child->state & CGROUP_FROZEN)) { in update_if_frozen()
Dcpuset.c497 struct cpuset *child; in partition_is_populated() local
505 cpuset_for_each_child(child, css, cs) { in partition_is_populated()
506 if (child == excluded_child) in partition_is_populated()
508 if (is_partition_valid(child)) in partition_is_populated()
510 if (cgroup_is_populated(child->css.cgroup)) { in partition_is_populated()
/kernel/events/
Dcore.c3229 struct perf_event *child; in perf_event_modify_attr() local
3256 list_for_each_entry(child, &event->child_list, child_list) { in perf_event_modify_attr()
3257 perf_event_modify_copy_attr(&child->attr, attr); in perf_event_modify_attr()
3258 err = func(child, attr); in perf_event_modify_attr()
5149 struct perf_event *child, *tmp; in perf_event_release_kernel() local
5185 list_for_each_entry(child, &event->child_list, child_list) { in perf_event_release_kernel()
5191 ctx = READ_ONCE(child->ctx); in perf_event_release_kernel()
5218 if (tmp == child) { in perf_event_release_kernel()
5219 perf_remove_from_context(child, DETACH_GROUP); in perf_event_release_kernel()
5220 list_move(&child->child_list, &free_list); in perf_event_release_kernel()
[all …]
/kernel/irq/
Dirqdomain.c1214 struct irq_data *child) in irq_domain_insert_irq_data() argument
1219 irq_data_get_node(child)); in irq_domain_insert_irq_data()
1221 child->parent_data = irq_data; in irq_domain_insert_irq_data()
1222 irq_data->irq = child->irq; in irq_domain_insert_irq_data()
1223 irq_data->common = child->common; in irq_domain_insert_irq_data()
/kernel/trace/
Dtrace_events.c994 struct dentry *child; in remove_event_file_dir() local
998 list_for_each_entry(child, &dir->d_subdirs, d_child) { in remove_event_file_dir()
999 if (d_really_is_positive(child)) /* probably unneeded */ in remove_event_file_dir()
1000 d_inode(child)->i_private = NULL; in remove_event_file_dir()
/kernel/locking/
Dlockdep.c1526 static inline struct lock_list *get_lock_parent(struct lock_list *child) in get_lock_parent() argument
1528 return child->parent; in get_lock_parent()
1531 static inline int get_lock_depth(struct lock_list *child) in get_lock_depth() argument
1536 while ((parent = get_lock_parent(child))) { in get_lock_depth()
1537 child = parent; in get_lock_depth()