Home
last modified time | relevance | path

Searched refs:child (Results 1 – 19 of 19) sorted by relevance

/kernel/
Dptrace.c68 void __ptrace_link(struct task_struct *child, struct task_struct *new_parent, in __ptrace_link() argument
71 BUG_ON(!list_empty(&child->ptrace_entry)); in __ptrace_link()
72 list_add(&child->ptrace_entry, &new_parent->ptraced); in __ptrace_link()
73 child->parent = new_parent; in __ptrace_link()
74 child->ptracer_cred = get_cred(ptracer_cred); in __ptrace_link()
83 static void ptrace_link(struct task_struct *child, struct task_struct *new_parent) in ptrace_link() argument
85 __ptrace_link(child, new_parent, current_cred()); in ptrace_link()
116 void __ptrace_unlink(struct task_struct *child) in __ptrace_unlink() argument
119 BUG_ON(!child->ptrace); in __ptrace_unlink()
121 clear_task_syscall_work(child, SYSCALL_TRACE); in __ptrace_unlink()
[all …]
Dresource.c61 if (p->child) in next_resource()
62 return p->child; in next_resource()
85 for (p = p->child; p && l < *pos; p = r_next(m, p, &l)) in r_start()
171 p = &root->child; in __request_resource()
191 p = &old->parent->child; in __release_resource()
197 if (release_child || !(tmp->child)) { in __release_resource()
200 for (chd = tmp->child;; chd = chd->sibling) { in __release_resource()
205 *p = tmp->child; in __release_resource()
221 p = r->child; in __release_child_resources()
222 r->child = NULL; in __release_child_resources()
[all …]
Dsysctl.c2195 .child = random_table,
2200 .child = usermodehelper_table,
2206 .child = firmware_config_table,
2596 .child = key_sysctls,
3305 .child = inotify_table,
3312 .child = fanotify_table,
3319 .child = epoll_table,
3372 .child = sysctl_mount_point,
3439 .child = kern_table,
3444 .child = vm_table,
[all …]
Dfork.c1414 static int wait_for_vfork_done(struct task_struct *child, in wait_for_vfork_done() argument
1426 task_lock(child); in wait_for_vfork_done()
1427 child->vfork_done = NULL; in wait_for_vfork_done()
1428 task_unlock(child); in wait_for_vfork_done()
1431 put_task_struct(child); in wait_for_vfork_done()
2950 struct task_struct *leader, *parent, *child; in walk_process_tree() local
2957 list_for_each_entry(child, &parent->children, sibling) { in walk_process_tree()
2958 res = visitor(child, data); in walk_process_tree()
2962 leader = child; in walk_process_tree()
2971 child = leader; in walk_process_tree()
[all …]
Dutsname_sysctl.c120 .child = uts_kern_table,
Dseccomp.c463 struct seccomp_filter *child) in is_ancestor() argument
468 for (; child; child = child->prev) in is_ancestor()
469 if (child == parent) in is_ancestor()
Duser_namespace.c1296 const struct user_namespace *child) in in_userns() argument
1299 for (ns = child; ns->level > ancestor->level; ns = ns->parent) in in_userns()
/kernel/bpf/
Dlpm_trie.c25 struct lpm_trie_node __rcu *child[2]; member
271 node = rcu_dereference_check(node->child[next_bit], in trie_lookup_elem()
341 RCU_INIT_POINTER(new_node->child[0], NULL); in trie_update_elem()
342 RCU_INIT_POINTER(new_node->child[1], NULL); in trie_update_elem()
362 slot = &node->child[next_bit]; in trie_update_elem()
377 new_node->child[0] = node->child[0]; in trie_update_elem()
378 new_node->child[1] = node->child[1]; in trie_update_elem()
394 rcu_assign_pointer(new_node->child[next_bit], node); in trie_update_elem()
411 rcu_assign_pointer(im_node->child[0], node); in trie_update_elem()
412 rcu_assign_pointer(im_node->child[1], new_node); in trie_update_elem()
[all …]
/kernel/sched/
Dtopology.c63 if ((meta_flags & SDF_SHARED_CHILD) && sd->child && in sched_domain_debug_one()
64 !(sd->child->flags & flag)) in sched_domain_debug_one()
110 if (group == sd->groups && sd->child && in sched_domain_debug_one()
111 !cpumask_equal(sched_domain_span(sd->child), in sched_domain_debug_one()
695 parent->parent->child = tmp; in cpu_attach_domain()
713 sd->child = NULL; in cpu_attach_domain()
880 if (!sibling->child) in build_balance_mask()
884 if (!cpumask_equal(sg_span, sched_domain_span(sibling->child))) in build_balance_mask()
912 if (sd->child) in build_group_from_child_sched_domain()
913 cpumask_copy(sg_span, sched_domain_span(sd->child)); in build_group_from_child_sched_domain()
[all …]
Drt.c2705 struct task_group *child; in tg_rt_schedulable() local
2741 list_for_each_entry_rcu(child, &tg->children, siblings) { in tg_rt_schedulable()
2742 period = ktime_to_ns(child->rt_bandwidth.rt_period); in tg_rt_schedulable()
2743 runtime = child->rt_bandwidth.rt_runtime; in tg_rt_schedulable()
2745 if (child == d->tg) { in tg_rt_schedulable()
Dfair.c6346 sd = sd->child; in find_idlest_cpu()
6352 sd = sd->child; in find_idlest_cpu()
6359 sd = sd->child; in find_idlest_cpu()
8868 struct sched_domain *child = sd->child; in update_group_capacity() local
8877 if (!child) { in update_group_capacity()
8886 if (child->flags & SD_OVERLAP) { in update_group_capacity()
8905 group = child->groups; in update_group_capacity()
8913 } while (group != child->groups); in update_group_capacity()
9692 struct sched_domain *child = env->sd->child; in update_sd_lb_stats() local
9734 sds->prefer_sibling = child && child->flags & SD_PREFER_SIBLING; in update_sd_lb_stats()
Dcore.c1199 struct task_group *parent, *child; in walk_tg_tree_from() local
1208 list_for_each_entry_rcu(child, &parent->children, siblings) { in walk_tg_tree_from()
1209 parent = child; in walk_tg_tree_from()
1219 child = parent; in walk_tg_tree_from()
/kernel/livepatch/
Dtransition.c611 void klp_copy_process(struct task_struct *child) in klp_copy_process() argument
625 set_tsk_thread_flag(child, TIF_PATCH_PENDING); in klp_copy_process()
627 clear_tsk_thread_flag(child, TIF_PATCH_PENDING); in klp_copy_process()
629 child->patch_state = current->patch_state; in klp_copy_process()
/kernel/cgroup/
Dcgroup.c730 #define cgroup_for_each_live_child(child, cgrp) \ argument
731 list_for_each_entry((child), &(cgrp)->self.children, self.sibling) \
733 cgroup_is_dead(child); })) \
826 struct cgroup *child = NULL; in cgroup_update_populated() local
834 if (!child) { in cgroup_update_populated()
837 if (cgroup_is_threaded(child)) in cgroup_update_populated()
851 child = cgrp; in cgroup_update_populated()
3362 struct cgroup *cgrp, *child; in cgroup_subtree_control_write() local
3417 cgroup_for_each_live_child(child, cgrp) { in cgroup_subtree_control_write()
3418 if (child->subtree_control & (1 << ssid)) { in cgroup_subtree_control_write()
[all …]
Dlegacy_freezer.c260 struct freezer *child = css_freezer(pos); in update_if_frozen() local
262 if ((child->state & CGROUP_FREEZER_ONLINE) && in update_if_frozen()
263 !(child->state & CGROUP_FROZEN)) { in update_if_frozen()
/kernel/events/
Dcore.c3297 struct perf_event *child; in perf_event_modify_attr() local
3324 list_for_each_entry(child, &event->child_list, child_list) { in perf_event_modify_attr()
3325 perf_event_modify_copy_attr(&child->attr, attr); in perf_event_modify_attr()
3326 err = func(child, attr); in perf_event_modify_attr()
5230 struct perf_event *child, *tmp; in perf_event_release_kernel() local
5266 list_for_each_entry(child, &event->child_list, child_list) { in perf_event_release_kernel()
5272 ctx = READ_ONCE(child->ctx); in perf_event_release_kernel()
5299 if (tmp == child) { in perf_event_release_kernel()
5300 perf_remove_from_context(child, DETACH_GROUP); in perf_event_release_kernel()
5301 list_move(&child->child_list, &free_list); in perf_event_release_kernel()
[all …]
/kernel/irq/
Dirqdomain.c1214 struct irq_data *child) in irq_domain_insert_irq_data() argument
1219 irq_data_get_node(child)); in irq_domain_insert_irq_data()
1221 child->parent_data = irq_data; in irq_domain_insert_irq_data()
1222 irq_data->irq = child->irq; in irq_domain_insert_irq_data()
1223 irq_data->common = child->common; in irq_domain_insert_irq_data()
/kernel/trace/
Dtrace_events.c975 struct dentry *child; in remove_event_file_dir() local
979 list_for_each_entry(child, &dir->d_subdirs, d_child) { in remove_event_file_dir()
980 if (d_really_is_positive(child)) /* probably unneeded */ in remove_event_file_dir()
981 d_inode(child)->i_private = NULL; in remove_event_file_dir()
/kernel/locking/
Dlockdep.c1489 static inline struct lock_list *get_lock_parent(struct lock_list *child) in get_lock_parent() argument
1491 return child->parent; in get_lock_parent()
1494 static inline int get_lock_depth(struct lock_list *child) in get_lock_depth() argument
1499 while ((parent = get_lock_parent(child))) { in get_lock_depth()
1500 child = parent; in get_lock_depth()