/kernel/trace/ |
D | trace_selftest.c | 97 static inline void warn_failed_init_tracer(struct tracer *trace, int init_ret) in warn_failed_init_tracer() argument 100 trace->name, init_ret); in warn_failed_init_tracer() 325 static int trace_selftest_startup_dynamic_tracing(struct tracer *trace, in trace_selftest_startup_dynamic_tracing() argument 355 ret = tracer_init(trace, tr); in trace_selftest_startup_dynamic_tracing() 357 warn_failed_init_tracer(trace, ret); in trace_selftest_startup_dynamic_tracing() 393 trace->reset(tr); in trace_selftest_startup_dynamic_tracing() 401 trace->reset(tr); in trace_selftest_startup_dynamic_tracing() 541 # define trace_selftest_startup_dynamic_tracing(trace, tr, func) ({ 0; }) argument 658 trace_selftest_startup_function(struct tracer *trace, struct trace_array *tr) in trace_selftest_startup_function() argument 677 ret = tracer_init(trace, tr); in trace_selftest_startup_function() [all …]
|
D | fgraph.c | 102 struct ftrace_graph_ent trace; in function_graph_enter() local 104 trace.func = func; in function_graph_enter() 105 trace.depth = ++current->curr_ret_depth; in function_graph_enter() 111 if (!ftrace_graph_entry(&trace)) in function_graph_enter() 124 ftrace_pop_return_trace(struct ftrace_graph_ret *trace, unsigned long *ret, in ftrace_pop_return_trace() argument 168 trace->func = current->ret_stack[index].func; in ftrace_pop_return_trace() 169 trace->calltime = current->ret_stack[index].calltime; in ftrace_pop_return_trace() 170 trace->overrun = atomic_read(¤t->trace_overrun); in ftrace_pop_return_trace() 171 trace->depth = current->curr_ret_depth--; in ftrace_pop_return_trace() 211 struct ftrace_graph_ret trace; in ftrace_return_to_handler() local [all …]
|
D | trace_functions_graph.c | 98 struct ftrace_graph_ent *trace, in __trace_graph_entry() argument 112 entry->graph_ent = *trace; in __trace_graph_entry() 127 int trace_graph_entry(struct ftrace_graph_ent *trace) in trace_graph_entry() argument 147 if (ftrace_graph_notrace_addr(trace->func)) { in trace_graph_entry() 159 if (ftrace_graph_ignore_func(trace)) in trace_graph_entry() 178 ret = __trace_graph_entry(tr, trace, flags, pc); in trace_graph_entry() 218 struct ftrace_graph_ret *trace, in __trace_graph_return() argument 232 entry->ret = *trace; in __trace_graph_return() 237 void trace_graph_return(struct ftrace_graph_ret *trace) in trace_graph_return() argument 246 ftrace_graph_addr_finish(trace); in trace_graph_return() [all …]
|
D | trace_syscalls.c | 128 struct syscall_trace_enter *trace; in print_syscall_enter() local 132 trace = (typeof(trace))ent; in print_syscall_enter() 133 syscall = trace->nr; in print_syscall_enter() 157 trace->args[i], in print_syscall_enter() 174 struct syscall_trace_exit *trace; in print_syscall_exit() local 178 trace = (typeof(trace))ent; in print_syscall_exit() 179 syscall = trace->nr; in print_syscall_exit() 193 trace->ret); in print_syscall_exit() 202 sizeof(type) != sizeof(trace.field) ? \ 204 #type, #name, offsetof(typeof(trace), field), \ [all …]
|
D | trace_stat.c | 310 int register_stat_tracer(struct tracer_stat *trace) in register_stat_tracer() argument 315 if (!trace) in register_stat_tracer() 318 if (!trace->stat_start || !trace->stat_next || !trace->stat_show) in register_stat_tracer() 324 if (node->ts == trace) in register_stat_tracer() 334 session->ts = trace; in register_stat_tracer() 353 void unregister_stat_tracer(struct tracer_stat *trace) in unregister_stat_tracer() argument 359 if (node->ts == trace) { in unregister_stat_tracer()
|
D | trace.h | 432 struct tracer *trace; member 487 int (*selftest)(struct tracer *trace, 739 void trace_graph_return(struct ftrace_graph_ret *trace); 740 int trace_graph_entry(struct ftrace_graph_ent *trace); 821 extern int trace_selftest_startup_function(struct tracer *trace, 823 extern int trace_selftest_startup_function_graph(struct tracer *trace, 825 extern int trace_selftest_startup_irqsoff(struct tracer *trace, 827 extern int trace_selftest_startup_preemptoff(struct tracer *trace, 829 extern int trace_selftest_startup_preemptirqsoff(struct tracer *trace, 831 extern int trace_selftest_startup_wakeup(struct tracer *trace, [all …]
|
D | trace_stat.h | 16 void *(*stat_start)(struct tracer_stat *trace); 31 extern int register_stat_tracer(struct tracer_stat *trace); 32 extern void unregister_stat_tracer(struct tracer_stat *trace);
|
D | Kconfig | 16 See Documentation/trace/ftrace-design.rst 21 See Documentation/trace/ftrace-design.rst 26 See Documentation/trace/ftrace-design.rst 31 See Documentation/trace/ftrace-design.rst 39 See Documentation/trace/ftrace-design.rst 44 See Documentation/trace/ftrace-design.rst 150 Enable the kernel to trace every kernel function. This is done 165 Enable the kernel to trace a function at both its return 167 Its first purpose is to trace the duration of functions and 179 bool "Enable trace events for preempt and irq disable/enable" [all …]
|
D | trace_irqsoff.c | 175 static int irqsoff_graph_entry(struct ftrace_graph_ent *trace) in irqsoff_graph_entry() argument 183 if (ftrace_graph_ignore_func(trace)) in irqsoff_graph_entry() 192 if (ftrace_graph_notrace_addr(trace->func)) in irqsoff_graph_entry() 199 ret = __trace_graph_entry(tr, trace, flags, pc); in irqsoff_graph_entry() 205 static void irqsoff_graph_return(struct ftrace_graph_ret *trace) in irqsoff_graph_return() argument 212 ftrace_graph_addr_finish(trace); in irqsoff_graph_return() 218 __trace_graph_return(tr, trace, flags, pc); in irqsoff_graph_return()
|
D | trace_output.c | 761 if (event->funcs->trace == NULL) in register_trace_event() 762 event->funcs->trace = trace_nop_print; in register_trace_event() 882 .trace = trace_fn_trace, 1022 .trace = trace_ctx_print, 1034 .trace = trace_wake_print, 1074 .trace = trace_stack_print, 1127 .trace = trace_user_stack_print, 1191 .trace = trace_hwlat_print, 1235 .trace = trace_bputs_print, 1279 .trace = trace_bprint_print, [all …]
|
D | trace_sched_wakeup.c | 115 static int wakeup_graph_entry(struct ftrace_graph_ent *trace) in wakeup_graph_entry() argument 122 if (ftrace_graph_ignore_func(trace)) in wakeup_graph_entry() 131 if (ftrace_graph_notrace_addr(trace->func)) in wakeup_graph_entry() 138 ret = __trace_graph_entry(tr, trace, flags, pc); in wakeup_graph_entry() 145 static void wakeup_graph_return(struct ftrace_graph_ret *trace) in wakeup_graph_return() argument 152 ftrace_graph_addr_finish(trace); in wakeup_graph_return() 158 __trace_graph_return(tr, trace, flags, pc); in wakeup_graph_return()
|
D | trace.c | 1859 type->flags->trace = type; in register_tracer() 2592 !event_call->event.funcs->trace) in output_printk() 2600 event_call->event.funcs->trace(iter, 0, event); in output_printk() 3515 if (unlikely(tr->current_trace && iter->trace->name != tr->current_trace->name)) { in s_start() 3517 if (iter->trace->close) in s_start() 3518 iter->trace->close(iter); in s_start() 3519 *iter->trace = *tr->current_trace; in s_start() 3521 if (iter->trace->open) in s_start() 3522 iter->trace->open(iter); in s_start() 3527 if (iter->snapshot && iter->trace->use_max_tr) in s_start() [all …]
|
D | trace_branch.c | 165 .trace = trace_branch_print, 317 static void *annotated_branch_stat_start(struct tracer_stat *trace) in annotated_branch_stat_start() argument 407 static void *all_branch_stat_start(struct tracer_stat *trace) in all_branch_stat_start() argument
|
D | Makefile | 40 obj-$(CONFIG_TRACING) += trace.o
|
D | ftrace.c | 453 static void *function_stat_start(struct tracer_stat *trace) in function_stat_start() argument 456 container_of(trace, struct ftrace_profile_stat, stat); in function_stat_start() 796 static int profile_graph_entry(struct ftrace_graph_ent *trace) in profile_graph_entry() argument 800 function_profile_call(trace->func, 0, NULL, NULL); in profile_graph_entry() 813 static void profile_graph_return(struct ftrace_graph_ret *trace) in profile_graph_return() argument 827 if (!trace->calltime) in profile_graph_return() 830 calltime = trace->rettime - trace->calltime; in profile_graph_return() 846 rec = ftrace_find_profiled_func(stat, trace->func); in profile_graph_return()
|
D | trace_kprobe.c | 1563 .trace = print_kretprobe_event 1567 .trace = print_kprobe_event
|
D | trace_uprobe.c | 1535 .trace = print_uprobe_event
|
D | blktrace.c | 1625 .trace = blk_trace_event_print,
|
D | trace_events_hist.c | 652 struct synth_trace_event trace; in synth_event_define_fields() local 653 int offset = offsetof(typeof(trace), fields); in synth_event_define_fields() 904 .trace = print_synth_event
|
/kernel/ |
D | stacktrace.c | 253 save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace) in save_stack_trace_tsk() argument 259 save_stack_trace_regs(struct pt_regs *regs, struct stack_trace *trace) in save_stack_trace_regs() argument 275 struct stack_trace trace = { in stack_trace_save() local 281 save_stack_trace(&trace); in stack_trace_save() 282 return trace.nr_entries; in stack_trace_save() 299 struct stack_trace trace = { in stack_trace_save_tsk() local 306 save_stack_trace_tsk(task, &trace); in stack_trace_save_tsk() 307 return trace.nr_entries; in stack_trace_save_tsk() 322 struct stack_trace trace = { in stack_trace_save_regs() local 328 save_stack_trace_regs(regs, &trace); in stack_trace_save_regs() [all …]
|
D | Makefile | 99 obj-$(CONFIG_FUNCTION_TRACER) += trace/ 100 obj-$(CONFIG_TRACING) += trace/ 101 obj-$(CONFIG_TRACE_CLOCK) += trace/ 102 obj-$(CONFIG_RING_BUFFER) += trace/ 103 obj-$(CONFIG_TRACEPOINTS) += trace/
|
D | fork.c | 1801 int trace, in copy_process() argument 2207 ptrace_init_task(p, (clone_flags & CLONE_PTRACE) || trace); in copy_process() 2372 int trace = 0; in _do_fork() local 2383 trace = PTRACE_EVENT_VFORK; in _do_fork() 2385 trace = PTRACE_EVENT_CLONE; in _do_fork() 2387 trace = PTRACE_EVENT_FORK; in _do_fork() 2389 if (likely(!ptrace_event_enabled(current, trace))) in _do_fork() 2390 trace = 0; in _do_fork() 2393 p = copy_process(NULL, trace, NUMA_NO_NODE, args); in _do_fork() 2422 if (unlikely(trace)) in _do_fork() [all …]
|
/kernel/locking/ |
D | lockdep.c | 482 struct lock_trace *trace, *t2; in save_trace() local 490 trace = (struct lock_trace *)(stack_trace + nr_stack_trace_entries); in save_trace() 503 trace->nr_entries = stack_trace_save(trace->entries, max_entries, 3); in save_trace() 505 hash = jhash(trace->entries, trace->nr_entries * in save_trace() 506 sizeof(trace->entries[0]), 0); in save_trace() 507 trace->hash = hash; in save_trace() 510 if (traces_identical(trace, t2)) in save_trace() 513 nr_stack_trace_entries += LOCK_TRACE_SIZE_IN_LONGS + trace->nr_entries; in save_trace() 514 hlist_add_head(&trace->hash_entry, hash_head); in save_trace() 516 return trace; in save_trace() [all …]
|
/kernel/bpf/ |
D | stackmap.c | 356 struct perf_callchain_entry *trace; in BPF_CALL_3() local 372 trace = get_perf_callchain(regs, init_nr, kernel, user, in BPF_CALL_3() 375 if (unlikely(!trace)) in BPF_CALL_3() 382 trace_nr = trace->nr - init_nr; in BPF_CALL_3() 390 ips = trace->ip + skip + init_nr; in BPF_CALL_3() 459 struct perf_callchain_entry *trace; in BPF_CALL_4() local 480 trace = get_perf_callchain(regs, init_nr, kernel, user, in BPF_CALL_4() 482 if (unlikely(!trace)) in BPF_CALL_4() 485 trace_nr = trace->nr - init_nr; in BPF_CALL_4() 492 ips = trace->ip + skip + init_nr; in BPF_CALL_4()
|
/kernel/power/ |
D | Kconfig | 222 functions from <linux/resume-trace.h> as well as the 223 <asm/resume-trace.h> header with a TRACE_RESUME() macro.
|