/kernel/livepatch/ |
D | transition.c | 72 struct task_struct *g, *task; in klp_complete_transition() local 109 for_each_process_thread(g, task) { in klp_complete_transition() 343 struct task_struct *g, *task; in klp_send_signals() local 349 for_each_process_thread(g, task) { in klp_send_signals() 387 struct task_struct *g, *task; in klp_try_complete_transition() local 403 for_each_process_thread(g, task) in klp_try_complete_transition() 464 struct task_struct *g, *task; in klp_start_transition() local 479 for_each_process_thread(g, task) in klp_start_transition() 505 struct task_struct *g, *task; in klp_init_transition() local 529 for_each_process_thread(g, task) { in klp_init_transition() [all …]
|
/kernel/trace/ |
D | trace_events_filter_test.h | 12 TP_PROTO(int a, int b, int c, int d, int e, int f, int g, int h), 14 TP_ARGS(a, b, c, d, e, f, g, h), 23 __field(int, g) 34 __entry->g = g; 40 __entry->e, __entry->f, __entry->g, __entry->h)
|
D | fgraph.c | 385 struct task_struct *g, *t; in alloc_retstack_tasklist() local 401 for_each_process_thread(g, t) { in alloc_retstack_tasklist()
|
D | ftrace.c | 4016 static int ftrace_match(char *str, struct ftrace_glob *g) in ftrace_match() argument 4021 str = arch_ftrace_match_adjust(str, g->search); in ftrace_match() 4023 switch (g->type) { in ftrace_match() 4025 if (strcmp(str, g->search) == 0) in ftrace_match() 4029 if (strncmp(str, g->search, g->len) == 0) in ftrace_match() 4033 if (strstr(str, g->search)) in ftrace_match() 4038 if (slen >= g->len && in ftrace_match() 4039 memcmp(str + slen - g->len, g->search, g->len) == 0) in ftrace_match() 4043 if (glob_match(g->search, str)) in ftrace_match()
|
D | Kconfig | 509 ftrace interface, e.g.: 631 Tracing also is possible using the ftrace interface, e.g.: 1046 However, it should be safe to use on e.g. unused portion of VRAM.
|
D | trace_events_filter.c | 2305 .e = ve, .f = vf, .g = vg, .h = vh }, \
|
/kernel/power/ |
D | process.c | 34 struct task_struct *g, *p; in try_to_freeze_tasks() local 55 for_each_process_thread(g, p) { in try_to_freeze_tasks() 104 for_each_process_thread(g, p) { in try_to_freeze_tasks() 189 struct task_struct *g, *p; in thaw_processes() local 208 for_each_process_thread(g, p) { in thaw_processes() 227 struct task_struct *g, *p; in thaw_kernel_threads() local 235 for_each_process_thread(g, p) { in thaw_kernel_threads()
|
D | hibernate.c | 1272 static struct attribute *g[] = { variable 1283 .attrs = g,
|
D | main.c | 867 static struct attribute * g[] = { variable 901 .attrs = g,
|
D | Kconfig | 9 suspend-to-RAM state (e.g. the ACPI S3 state). 286 notification of APM "events" (e.g. battery status change).
|
/kernel/ |
D | hung_task.c | 159 static bool rcu_lock_break(struct task_struct *g, struct task_struct *t) in rcu_lock_break() argument 163 get_task_struct(g); in rcu_lock_break() 168 can_cont = pid_alive(g) && pid_alive(t); in rcu_lock_break() 170 put_task_struct(g); in rcu_lock_break() 184 struct task_struct *g, *t; in check_hung_uninterruptible_tasks() local 196 for_each_process_thread(g, t) { in check_hung_uninterruptible_tasks() 202 if (!rcu_lock_break(g, t)) in check_hung_uninterruptible_tasks()
|
D | cred.c | 635 int g; in cred_fscmp() local 662 for (g = 0; g < ga->ngroups; g++) { in cred_fscmp() 663 if (gid_lt(ga->gid[g], gb->gid[g])) in cred_fscmp() 665 if (gid_gt(ga->gid[g], gb->gid[g])) in cred_fscmp()
|
D | sys.c | 215 struct task_struct *g, *p; in SYSCALL_DEFINE3() local 263 for_each_process_thread(g, p) { in SYSCALL_DEFINE3() 285 struct task_struct *g, *p; in SYSCALL_DEFINE2() local 331 for_each_process_thread(g, p) { in SYSCALL_DEFINE2() 1116 struct task_struct *g; in SYSCALL_DEFINE2() local 1119 g = pid_task(pgrp, PIDTYPE_PGID); in SYSCALL_DEFINE2() 1120 if (!g || task_session(g) != task_session(group_leader)) in SYSCALL_DEFINE2()
|
D | exit.c | 446 struct task_struct *c, *g, *p = current; in mm_update_next_owner() local 485 for_each_process(g) { in mm_update_next_owner() 486 if (g->flags & PF_KTHREAD) in mm_update_next_owner() 488 for_each_thread(g, c) { in mm_update_next_owner()
|
/kernel/debug/kdb/ |
D | kdb_bt.c | 139 struct task_struct *g, *p; in kdb_bt() local 152 for_each_process_thread(g, p) { in kdb_bt()
|
D | kdb_main.c | 2280 const struct task_struct *p, *g; in kdb_ps_suppressed() local 2286 for_each_process_thread(g, p) { in kdb_ps_suppressed() 2340 struct task_struct *g, *p; in kdb_ps() local 2360 for_each_process_thread(g, p) { in kdb_ps()
|
/kernel/gcov/ |
D | Kconfig | 11 This option enables gcov-based code profiling (e.g. for code coverage 21 For a single file (e.g. main.o):
|
/kernel/rcu/ |
D | tasks.h | 664 struct task_struct *g; in rcu_tasks_wait_gp() local 686 for_each_process_thread(g, t) in rcu_tasks_wait_gp() 1644 struct task_struct *g, *t; in check_all_holdout_tasks_trace() local 1649 list_for_each_entry_safe(t, g, hop, trc_holdout_list) { in check_all_holdout_tasks_trace()
|
/kernel/debug/ |
D | gdbstub.c | 697 struct task_struct *g; in gdb_cmd_query() local 725 for_each_process_thread(g, p) { in gdb_cmd_query()
|
/kernel/module/ |
D | Kconfig | 58 list of linked (or loaded) modules e.g. on detection of a bad 96 by developers to identify the SCM version of a given module, e.g.
|
/kernel/sched/ |
D | debug.c | 564 struct task_struct *g, *p; in print_rq() local 574 for_each_process_thread(g, p) { in print_rq()
|
D | core.c | 1791 struct task_struct *g, *p; in uclamp_sync_util_min_rt_default() local 1811 for_each_process_thread(g, p) in uclamp_sync_util_min_rt_default() 9172 struct task_struct *g, *p; in show_state_filter() local 9175 for_each_process_thread(g, p) { in show_state_filter() 9785 struct task_struct *g, *p; in dump_rq_tasks() local 9791 for_each_process_thread(g, p) { in dump_rq_tasks() 10236 struct task_struct *g, *p; in normalize_rt_tasks() local 10242 for_each_process_thread(g, p) { in normalize_rt_tasks()
|
/kernel/locking/ |
D | lockdep.c | 6486 struct task_struct *g, *p; in debug_show_all_locks() local 6495 for_each_process_thread(g, p) { in debug_show_all_locks()
|