/kernel/livepatch/ |
D | transition.c | 73 struct task_struct *g, *task; in klp_complete_transition() local 110 for_each_process_thread(g, task) { in klp_complete_transition() 345 struct task_struct *g, *task; in klp_send_signals() local 351 for_each_process_thread(g, task) { in klp_send_signals() 391 struct task_struct *g, *task; in klp_try_complete_transition() local 407 for_each_process_thread(g, task) in klp_try_complete_transition() 465 struct task_struct *g, *task; in klp_start_transition() local 480 for_each_process_thread(g, task) in klp_start_transition() 506 struct task_struct *g, *task; in klp_init_transition() local 530 for_each_process_thread(g, task) { in klp_init_transition() [all …]
|
/kernel/trace/ |
D | trace_events_filter_test.h | 12 TP_PROTO(int a, int b, int c, int d, int e, int f, int g, int h), 14 TP_ARGS(a, b, c, d, e, f, g, h), 23 __field(int, g) 34 __entry->g = g; 40 __entry->e, __entry->f, __entry->g, __entry->h)
|
D | fgraph.c | 347 struct task_struct *g, *t; in alloc_retstack_tasklist() local 363 do_each_thread(g, t) { in alloc_retstack_tasklist() 378 } while_each_thread(g, t); in alloc_retstack_tasklist()
|
D | ftrace.c | 3676 static int ftrace_match(char *str, struct ftrace_glob *g) in ftrace_match() argument 3681 str = arch_ftrace_match_adjust(str, g->search); in ftrace_match() 3683 switch (g->type) { in ftrace_match() 3685 if (strcmp(str, g->search) == 0) in ftrace_match() 3689 if (strncmp(str, g->search, g->len) == 0) in ftrace_match() 3693 if (strstr(str, g->search)) in ftrace_match() 3698 if (slen >= g->len && in ftrace_match() 3699 memcmp(str + slen - g->len, g->search, g->len) == 0) in ftrace_match() 3703 if (glob_match(g->search, str)) in ftrace_match()
|
D | Kconfig | 304 ftrace interface, e.g.: 451 Tracing also is possible using the ftrace interface, e.g.: 672 However, it should be safe to use on e.g. unused portion of VRAM.
|
D | trace_events_filter.c | 2108 .e = ve, .f = vf, .g = vg, .h = vh }, \
|
/kernel/power/ |
D | process.c | 34 struct task_struct *g, *p; in try_to_freeze_tasks() local 56 for_each_process_thread(g, p) { in try_to_freeze_tasks() 113 for_each_process_thread(g, p) { in try_to_freeze_tasks() 203 struct task_struct *g, *p; in thaw_processes() local 222 for_each_process_thread(g, p) { in thaw_processes() 241 struct task_struct *g, *p; in thaw_kernel_threads() local 249 for_each_process_thread(g, p) { in thaw_kernel_threads()
|
D | main.c | 847 static struct attribute * g[] = { variable 880 .attrs = g,
|
D | hibernate.c | 1145 static struct attribute * g[] = { variable 1156 .attrs = g,
|
D | Kconfig | 9 suspend-to-RAM state (e.g. the ACPI S3 state). 256 notification of APM "events" (e.g. battery status change).
|
/kernel/ |
D | hung_task.c | 152 static bool rcu_lock_break(struct task_struct *g, struct task_struct *t) in rcu_lock_break() argument 156 get_task_struct(g); in rcu_lock_break() 161 can_cont = pid_alive(g) && pid_alive(t); in rcu_lock_break() 163 put_task_struct(g); in rcu_lock_break() 177 struct task_struct *g, *t; in check_hung_uninterruptible_tasks() local 188 for_each_process_thread(g, t) { in check_hung_uninterruptible_tasks() 192 if (!rcu_lock_break(g, t)) in check_hung_uninterruptible_tasks()
|
D | cred.c | 616 int g; in cred_fscmp() local 643 for (g = 0; g < ga->ngroups; g++) { in cred_fscmp() 644 if (gid_lt(ga->gid[g], gb->gid[g])) in cred_fscmp() 646 if (gid_gt(ga->gid[g], gb->gid[g])) in cred_fscmp()
|
D | sys.c | 200 struct task_struct *g, *p; in SYSCALL_DEFINE3() local 247 do_each_thread(g, p) { in SYSCALL_DEFINE3() 250 } while_each_thread(g, p); in SYSCALL_DEFINE3() 270 struct task_struct *g, *p; in SYSCALL_DEFINE2() local 315 do_each_thread(g, p) { in SYSCALL_DEFINE2() 321 } while_each_thread(g, p); in SYSCALL_DEFINE2() 1056 struct task_struct *g; in SYSCALL_DEFINE2() local 1059 g = pid_task(pgrp, PIDTYPE_PGID); in SYSCALL_DEFINE2() 1060 if (!g || task_session(g) != task_session(group_leader)) in SYSCALL_DEFINE2()
|
D | exit.c | 349 struct task_struct *c, *g, *p = current; in mm_update_next_owner() local 388 for_each_process(g) { in mm_update_next_owner() 389 if (g->flags & PF_KTHREAD) in mm_update_next_owner() 391 for_each_thread(g, c) { in mm_update_next_owner()
|
/kernel/debug/kdb/ |
D | kdb_private.h | 233 #define kdb_do_each_thread(g, p) do_each_thread(g, p) argument 234 #define kdb_while_each_thread(g, p) while_each_thread(g, p) argument
|
D | kdb_bt.c | 119 struct task_struct *g, *p; in kdb_bt() local 132 kdb_do_each_thread(g, p) { in kdb_bt() 139 } kdb_while_each_thread(g, p); in kdb_bt()
|
D | kdb_main.c | 2292 const struct task_struct *p, *g; in kdb_ps_suppressed() local 2298 kdb_do_each_thread(g, p) { in kdb_ps_suppressed() 2301 } kdb_while_each_thread(g, p); in kdb_ps_suppressed() 2349 struct task_struct *g, *p; in kdb_ps() local 2368 kdb_do_each_thread(g, p) { in kdb_ps() 2373 } kdb_while_each_thread(g, p); in kdb_ps()
|
/kernel/gcov/ |
D | Kconfig | 10 This option enables gcov-based code profiling (e.g. for code coverage 20 For a single file (e.g. main.o):
|
/kernel/debug/ |
D | gdbstub.c | 700 struct task_struct *g; in gdb_cmd_query() local 728 do_each_thread(g, p) { in gdb_cmd_query() 738 } while_each_thread(g, p); in gdb_cmd_query()
|
/kernel/rcu/ |
D | update.c | 633 struct task_struct *g, *t; in rcu_tasks_kthread() local 693 for_each_process_thread(g, t) { in rcu_tasks_kthread()
|
/kernel/sched/ |
D | debug.c | 464 struct task_struct *g, *p; in print_rq() local 474 for_each_process_thread(g, p) { in print_rq()
|
D | rt.c | 2458 struct task_struct *g, *p; in tg_has_rt_tasks() local 2466 for_each_process_thread(g, p) { in tg_has_rt_tasks()
|
D | core.c | 5974 struct task_struct *g, *p; in show_state_filter() local 5984 for_each_process_thread(g, p) { in show_state_filter() 6829 struct task_struct *g, *p; in normalize_rt_tasks() local 6835 for_each_process_thread(g, p) { in normalize_rt_tasks()
|
/kernel/cgroup/ |
D | cgroup.c | 1896 struct task_struct *p, *g; in cgroup_enable_task_cg_lists() local 1913 do_each_thread(g, p) { in cgroup_enable_task_cg_lists() 1939 } while_each_thread(g, p); in cgroup_enable_task_cg_lists()
|
/kernel/locking/ |
D | lockdep.c | 5332 struct task_struct *g, *p; in debug_show_all_locks() local 5341 for_each_process_thread(g, p) { in debug_show_all_locks()
|