Home
last modified time | relevance | path

Searched refs:trace (Results 1 – 25 of 25) sorted by relevance

/kernel/trace/
Dtrace_selftest.c97 static inline void warn_failed_init_tracer(struct tracer *trace, int init_ret) in warn_failed_init_tracer() argument
100 trace->name, init_ret); in warn_failed_init_tracer()
325 static int trace_selftest_startup_dynamic_tracing(struct tracer *trace, in trace_selftest_startup_dynamic_tracing() argument
355 ret = tracer_init(trace, tr); in trace_selftest_startup_dynamic_tracing()
357 warn_failed_init_tracer(trace, ret); in trace_selftest_startup_dynamic_tracing()
393 trace->reset(tr); in trace_selftest_startup_dynamic_tracing()
401 trace->reset(tr); in trace_selftest_startup_dynamic_tracing()
541 # define trace_selftest_startup_dynamic_tracing(trace, tr, func) ({ 0; }) argument
658 trace_selftest_startup_function(struct tracer *trace, struct trace_array *tr) in trace_selftest_startup_function() argument
677 ret = tracer_init(trace, tr); in trace_selftest_startup_function()
[all …]
Dfgraph.c102 struct ftrace_graph_ent trace; in function_graph_enter() local
104 trace.func = func; in function_graph_enter()
105 trace.depth = ++current->curr_ret_depth; in function_graph_enter()
111 if (!ftrace_graph_entry(&trace)) in function_graph_enter()
124 ftrace_pop_return_trace(struct ftrace_graph_ret *trace, unsigned long *ret, in ftrace_pop_return_trace() argument
168 trace->func = current->ret_stack[index].func; in ftrace_pop_return_trace()
169 trace->calltime = current->ret_stack[index].calltime; in ftrace_pop_return_trace()
170 trace->overrun = atomic_read(&current->trace_overrun); in ftrace_pop_return_trace()
171 trace->depth = current->curr_ret_depth--; in ftrace_pop_return_trace()
211 struct ftrace_graph_ret trace; in ftrace_return_to_handler() local
[all …]
Dtrace_functions_graph.c98 struct ftrace_graph_ent *trace, in __trace_graph_entry() argument
112 entry->graph_ent = *trace; in __trace_graph_entry()
127 int trace_graph_entry(struct ftrace_graph_ent *trace) in trace_graph_entry() argument
147 if (ftrace_graph_notrace_addr(trace->func)) { in trace_graph_entry()
159 if (ftrace_graph_ignore_func(trace)) in trace_graph_entry()
178 ret = __trace_graph_entry(tr, trace, flags, pc); in trace_graph_entry()
218 struct ftrace_graph_ret *trace, in __trace_graph_return() argument
232 entry->ret = *trace; in __trace_graph_return()
237 void trace_graph_return(struct ftrace_graph_ret *trace) in trace_graph_return() argument
246 ftrace_graph_addr_finish(trace); in trace_graph_return()
[all …]
Dtrace_syscalls.c128 struct syscall_trace_enter *trace; in print_syscall_enter() local
132 trace = (typeof(trace))ent; in print_syscall_enter()
133 syscall = trace->nr; in print_syscall_enter()
157 trace->args[i], in print_syscall_enter()
174 struct syscall_trace_exit *trace; in print_syscall_exit() local
178 trace = (typeof(trace))ent; in print_syscall_exit()
179 syscall = trace->nr; in print_syscall_exit()
193 trace->ret); in print_syscall_exit()
202 sizeof(type) != sizeof(trace.field) ? \
204 #type, #name, offsetof(typeof(trace), field), \
[all …]
Dtrace_stat.c310 int register_stat_tracer(struct tracer_stat *trace) in register_stat_tracer() argument
315 if (!trace) in register_stat_tracer()
318 if (!trace->stat_start || !trace->stat_next || !trace->stat_show) in register_stat_tracer()
324 if (node->ts == trace) in register_stat_tracer()
334 session->ts = trace; in register_stat_tracer()
353 void unregister_stat_tracer(struct tracer_stat *trace) in unregister_stat_tracer() argument
359 if (node->ts == trace) { in unregister_stat_tracer()
Dtrace.h432 struct tracer *trace; member
487 int (*selftest)(struct tracer *trace,
739 void trace_graph_return(struct ftrace_graph_ret *trace);
740 int trace_graph_entry(struct ftrace_graph_ent *trace);
821 extern int trace_selftest_startup_function(struct tracer *trace,
823 extern int trace_selftest_startup_function_graph(struct tracer *trace,
825 extern int trace_selftest_startup_irqsoff(struct tracer *trace,
827 extern int trace_selftest_startup_preemptoff(struct tracer *trace,
829 extern int trace_selftest_startup_preemptirqsoff(struct tracer *trace,
831 extern int trace_selftest_startup_wakeup(struct tracer *trace,
[all …]
Dtrace_stat.h16 void *(*stat_start)(struct tracer_stat *trace);
31 extern int register_stat_tracer(struct tracer_stat *trace);
32 extern void unregister_stat_tracer(struct tracer_stat *trace);
DKconfig16 See Documentation/trace/ftrace-design.rst
21 See Documentation/trace/ftrace-design.rst
26 See Documentation/trace/ftrace-design.rst
31 See Documentation/trace/ftrace-design.rst
39 See Documentation/trace/ftrace-design.rst
44 See Documentation/trace/ftrace-design.rst
150 Enable the kernel to trace every kernel function. This is done
165 Enable the kernel to trace a function at both its return
167 Its first purpose is to trace the duration of functions and
179 bool "Enable trace events for preempt and irq disable/enable"
[all …]
Dtrace_irqsoff.c175 static int irqsoff_graph_entry(struct ftrace_graph_ent *trace) in irqsoff_graph_entry() argument
183 if (ftrace_graph_ignore_func(trace)) in irqsoff_graph_entry()
192 if (ftrace_graph_notrace_addr(trace->func)) in irqsoff_graph_entry()
199 ret = __trace_graph_entry(tr, trace, flags, pc); in irqsoff_graph_entry()
205 static void irqsoff_graph_return(struct ftrace_graph_ret *trace) in irqsoff_graph_return() argument
212 ftrace_graph_addr_finish(trace); in irqsoff_graph_return()
218 __trace_graph_return(tr, trace, flags, pc); in irqsoff_graph_return()
Dtrace_output.c761 if (event->funcs->trace == NULL) in register_trace_event()
762 event->funcs->trace = trace_nop_print; in register_trace_event()
882 .trace = trace_fn_trace,
1022 .trace = trace_ctx_print,
1034 .trace = trace_wake_print,
1074 .trace = trace_stack_print,
1127 .trace = trace_user_stack_print,
1191 .trace = trace_hwlat_print,
1235 .trace = trace_bputs_print,
1279 .trace = trace_bprint_print,
[all …]
Dtrace_sched_wakeup.c115 static int wakeup_graph_entry(struct ftrace_graph_ent *trace) in wakeup_graph_entry() argument
122 if (ftrace_graph_ignore_func(trace)) in wakeup_graph_entry()
131 if (ftrace_graph_notrace_addr(trace->func)) in wakeup_graph_entry()
138 ret = __trace_graph_entry(tr, trace, flags, pc); in wakeup_graph_entry()
145 static void wakeup_graph_return(struct ftrace_graph_ret *trace) in wakeup_graph_return() argument
152 ftrace_graph_addr_finish(trace); in wakeup_graph_return()
158 __trace_graph_return(tr, trace, flags, pc); in wakeup_graph_return()
Dtrace.c1859 type->flags->trace = type; in register_tracer()
2592 !event_call->event.funcs->trace) in output_printk()
2600 event_call->event.funcs->trace(iter, 0, event); in output_printk()
3515 if (unlikely(tr->current_trace && iter->trace->name != tr->current_trace->name)) { in s_start()
3517 if (iter->trace->close) in s_start()
3518 iter->trace->close(iter); in s_start()
3519 *iter->trace = *tr->current_trace; in s_start()
3521 if (iter->trace->open) in s_start()
3522 iter->trace->open(iter); in s_start()
3527 if (iter->snapshot && iter->trace->use_max_tr) in s_start()
[all …]
Dtrace_branch.c165 .trace = trace_branch_print,
317 static void *annotated_branch_stat_start(struct tracer_stat *trace) in annotated_branch_stat_start() argument
407 static void *all_branch_stat_start(struct tracer_stat *trace) in all_branch_stat_start() argument
DMakefile40 obj-$(CONFIG_TRACING) += trace.o
Dftrace.c453 static void *function_stat_start(struct tracer_stat *trace) in function_stat_start() argument
456 container_of(trace, struct ftrace_profile_stat, stat); in function_stat_start()
796 static int profile_graph_entry(struct ftrace_graph_ent *trace) in profile_graph_entry() argument
800 function_profile_call(trace->func, 0, NULL, NULL); in profile_graph_entry()
813 static void profile_graph_return(struct ftrace_graph_ret *trace) in profile_graph_return() argument
827 if (!trace->calltime) in profile_graph_return()
830 calltime = trace->rettime - trace->calltime; in profile_graph_return()
846 rec = ftrace_find_profiled_func(stat, trace->func); in profile_graph_return()
Dtrace_kprobe.c1563 .trace = print_kretprobe_event
1567 .trace = print_kprobe_event
Dtrace_uprobe.c1535 .trace = print_uprobe_event
Dblktrace.c1625 .trace = blk_trace_event_print,
Dtrace_events_hist.c652 struct synth_trace_event trace; in synth_event_define_fields() local
653 int offset = offsetof(typeof(trace), fields); in synth_event_define_fields()
904 .trace = print_synth_event
/kernel/
Dstacktrace.c253 save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace) in save_stack_trace_tsk() argument
259 save_stack_trace_regs(struct pt_regs *regs, struct stack_trace *trace) in save_stack_trace_regs() argument
275 struct stack_trace trace = { in stack_trace_save() local
281 save_stack_trace(&trace); in stack_trace_save()
282 return trace.nr_entries; in stack_trace_save()
299 struct stack_trace trace = { in stack_trace_save_tsk() local
306 save_stack_trace_tsk(task, &trace); in stack_trace_save_tsk()
307 return trace.nr_entries; in stack_trace_save_tsk()
322 struct stack_trace trace = { in stack_trace_save_regs() local
328 save_stack_trace_regs(regs, &trace); in stack_trace_save_regs()
[all …]
DMakefile99 obj-$(CONFIG_FUNCTION_TRACER) += trace/
100 obj-$(CONFIG_TRACING) += trace/
101 obj-$(CONFIG_TRACE_CLOCK) += trace/
102 obj-$(CONFIG_RING_BUFFER) += trace/
103 obj-$(CONFIG_TRACEPOINTS) += trace/
Dfork.c1801 int trace, in copy_process() argument
2207 ptrace_init_task(p, (clone_flags & CLONE_PTRACE) || trace); in copy_process()
2372 int trace = 0; in _do_fork() local
2383 trace = PTRACE_EVENT_VFORK; in _do_fork()
2385 trace = PTRACE_EVENT_CLONE; in _do_fork()
2387 trace = PTRACE_EVENT_FORK; in _do_fork()
2389 if (likely(!ptrace_event_enabled(current, trace))) in _do_fork()
2390 trace = 0; in _do_fork()
2393 p = copy_process(NULL, trace, NUMA_NO_NODE, args); in _do_fork()
2422 if (unlikely(trace)) in _do_fork()
[all …]
/kernel/locking/
Dlockdep.c482 struct lock_trace *trace, *t2; in save_trace() local
490 trace = (struct lock_trace *)(stack_trace + nr_stack_trace_entries); in save_trace()
503 trace->nr_entries = stack_trace_save(trace->entries, max_entries, 3); in save_trace()
505 hash = jhash(trace->entries, trace->nr_entries * in save_trace()
506 sizeof(trace->entries[0]), 0); in save_trace()
507 trace->hash = hash; in save_trace()
510 if (traces_identical(trace, t2)) in save_trace()
513 nr_stack_trace_entries += LOCK_TRACE_SIZE_IN_LONGS + trace->nr_entries; in save_trace()
514 hlist_add_head(&trace->hash_entry, hash_head); in save_trace()
516 return trace; in save_trace()
[all …]
/kernel/bpf/
Dstackmap.c356 struct perf_callchain_entry *trace; in BPF_CALL_3() local
372 trace = get_perf_callchain(regs, init_nr, kernel, user, in BPF_CALL_3()
375 if (unlikely(!trace)) in BPF_CALL_3()
382 trace_nr = trace->nr - init_nr; in BPF_CALL_3()
390 ips = trace->ip + skip + init_nr; in BPF_CALL_3()
459 struct perf_callchain_entry *trace; in BPF_CALL_4() local
480 trace = get_perf_callchain(regs, init_nr, kernel, user, in BPF_CALL_4()
482 if (unlikely(!trace)) in BPF_CALL_4()
485 trace_nr = trace->nr - init_nr; in BPF_CALL_4()
492 ips = trace->ip + skip + init_nr; in BPF_CALL_4()
/kernel/power/
DKconfig222 functions from <linux/resume-trace.h> as well as the
223 <asm/resume-trace.h> header with a TRACE_RESUME() macro.