/kernel/trace/ |
D | trace_events.c | 95 trace_find_event_field(struct trace_event_call *call, char *name) in trace_find_event_field() argument 100 head = trace_get_fields(call); in trace_find_event_field() 139 int trace_define_field(struct trace_event_call *call, const char *type, in trace_define_field() argument 145 if (WARN_ON(!call->class)) in trace_define_field() 148 head = trace_get_fields(call); in trace_define_field() 195 static void trace_destroy_fields(struct trace_event_call *call) in trace_destroy_fields() argument 200 head = trace_get_fields(call); in trace_destroy_fields() 207 int trace_event_raw_init(struct trace_event_call *call) in trace_event_raw_init() argument 211 id = register_trace_event(&call->event); in trace_event_raw_init() 301 int trace_event_reg(struct trace_event_call *call, in trace_event_reg() argument [all …]
|
D | trace_syscalls.c | 22 syscall_get_enter_fields(struct trace_event_call *call) in syscall_get_enter_fields() argument 24 struct syscall_metadata *entry = call->data; in syscall_get_enter_fields() 223 static int __init set_syscall_print_fmt(struct trace_event_call *call) in set_syscall_print_fmt() argument 227 struct syscall_metadata *entry = call->data; in set_syscall_print_fmt() 229 if (entry->enter_event != call) { in set_syscall_print_fmt() 230 call->print_fmt = "\"0x%lx\", REC->ret"; in set_syscall_print_fmt() 243 call->print_fmt = print_fmt; in set_syscall_print_fmt() 248 static void __init free_syscall_print_fmt(struct trace_event_call *call) in free_syscall_print_fmt() argument 250 struct syscall_metadata *entry = call->data; in free_syscall_print_fmt() 252 if (entry->enter_event == call) in free_syscall_print_fmt() [all …]
|
D | trace_kprobe.c | 296 tk->tp.call.class = &tk->tp.class; in alloc_trace_kprobe() 297 tk->tp.call.name = kstrdup(event, GFP_KERNEL); in alloc_trace_kprobe() 298 if (!tk->tp.call.name) in alloc_trace_kprobe() 314 kfree(tk->tp.call.name); in alloc_trace_kprobe() 327 kfree(tk->tp.call.class->system); in free_trace_kprobe() 328 kfree(tk->tp.call.name); in free_trace_kprobe() 339 if (strcmp(trace_event_name(&tk->tp.call), event) == 0 && in find_trace_kprobe() 340 strcmp(tk->tp.call.class->system, group) == 0) in find_trace_kprobe() 525 old_tk = find_trace_kprobe(trace_event_name(&tk->tp.call), in register_trace_kprobe() 526 tk->tp.call.class->system); in register_trace_kprobe() [all …]
|
D | trace_export.c | 163 #define FTRACE_ENTRY_REG(call, struct_name, etype, tstruct, print, filter,\ argument 166 struct trace_event_class __refdata event_class_ftrace_##call = { \ 168 .define_fields = ftrace_define_fields_##call, \ 169 .fields = LIST_HEAD_INIT(event_class_ftrace_##call.fields),\ 173 struct trace_event_call __used event_##call = { \ 174 .class = &event_class_ftrace_##call, \ 176 .name = #call, \ 183 __attribute__((section("_ftrace_events"))) *__event_##call = &event_##call; 186 #define FTRACE_ENTRY(call, struct_name, etype, tstruct, print, filter) \ argument 187 FTRACE_ENTRY_REG(call, struct_name, etype, \ [all …]
|
D | trace_uprobe.c | 262 tu->tp.call.class = &tu->tp.class; in alloc_trace_uprobe() 263 tu->tp.call.name = kstrdup(event, GFP_KERNEL); in alloc_trace_uprobe() 264 if (!tu->tp.call.name) in alloc_trace_uprobe() 280 kfree(tu->tp.call.name); in alloc_trace_uprobe() 294 kfree(tu->tp.call.class->system); in free_trace_uprobe() 295 kfree(tu->tp.call.name); in free_trace_uprobe() 305 if (strcmp(trace_event_name(&tu->tp.call), event) == 0 && in find_probe_event() 306 strcmp(tu->tp.call.class->system, group) == 0) in find_probe_event() 335 old_tu = find_probe_event(trace_event_name(&tu->tp.call), in register_trace_uprobe() 336 tu->tp.call.class->system); in register_trace_uprobe() [all …]
|
D | trace_functions_graph.c | 292 struct trace_event_call *call = &event_funcgraph_entry; in __trace_graph_entry() local 303 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_graph_entry() 404 struct trace_event_call *call = &event_funcgraph_exit; in __trace_graph_return() local 415 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_graph_return() 775 struct ftrace_graph_ent *call; in print_graph_entry_leaf() local 781 call = &entry->graph_ent; in print_graph_entry_leaf() 790 if (call->depth < -1) in print_graph_entry_leaf() 791 call->depth += FTRACE_NOTRACE_DEPTH; in print_graph_entry_leaf() 798 cpu_data->depth = call->depth - 1; in print_graph_entry_leaf() 801 if (call->depth < FTRACE_RETFUNC_DEPTH && in print_graph_entry_leaf() [all …]
|
D | trace_events_filter.c | 832 struct trace_event_call *call = file->event_call; in filter_disable() local 834 if (call->flags & TRACE_EVENT_FL_USE_CALL_FILTER) in filter_disable() 835 call->flags &= ~TRACE_EVENT_FL_FILTERED; in filter_disable() 889 struct trace_event_call *call = file->event_call; in __remove_filter() local 892 if (call->flags & TRACE_EVENT_FL_USE_CALL_FILTER) in __remove_filter() 893 remove_filter_string(call->filter); in __remove_filter() 912 struct trace_event_call *call = file->event_call; in __free_subsystem_filter() local 914 if (call->flags & TRACE_EVENT_FL_USE_CALL_FILTER) { in __free_subsystem_filter() 915 __free_filter(call->filter); in __free_subsystem_filter() 916 call->filter = NULL; in __free_subsystem_filter() [all …]
|
D | trace.h | 871 bool ftrace_event_is_function(struct trace_event_call *call); 1127 extern int create_event_filter(struct trace_event_call *call, 1133 trace_find_event_field(struct trace_event_call *call, char *name); 1352 #define FTRACE_ENTRY(call, struct_name, id, tstruct, print, filter) \ argument 1354 __aligned(4) event_##call; 1356 #define FTRACE_ENTRY_DUP(call, struct_name, id, tstruct, print, filter) \ argument 1357 FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print), \ 1362 int perf_ftrace_event_register(struct trace_event_call *call,
|
D | trace_mmiotrace.c | 301 struct trace_event_call *call = &event_mmiotrace_rw; in __trace_mmiotrace_rw() local 316 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_mmiotrace_rw() 331 struct trace_event_call *call = &event_mmiotrace_map; in __trace_mmiotrace_map() local 346 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_mmiotrace_map()
|
D | trace_sched_wakeup.c | 376 struct trace_event_call *call = &event_context_switch; in tracing_sched_switch_trace() local 394 if (!call_filter_check_discard(call, entry, buffer, event)) in tracing_sched_switch_trace() 404 struct trace_event_call *call = &event_wakeup; in tracing_sched_wakeup_trace() local 422 if (!call_filter_check_discard(call, entry, buffer, event)) in tracing_sched_wakeup_trace()
|
D | trace_branch.c | 32 struct trace_event_call *call = &event_branch; in probe_likely_condition() local 83 if (!call_filter_check_discard(call, entry, buffer, event)) in probe_likely_condition()
|
D | Makefile | 3 FRAME_CFLAGS := $(call cc-disable-warning,frame-address)
|
D | Kconfig | 148 sequence is then dynamically patched into a tracer call when 163 draw a call graph for each thread with some information like 375 This tracer works by hooking into every function call that the 527 up since it runs this on every system call defined. 529 TBD - enable a way to actually call the syscalls as we test their
|
D | trace.c | 326 int call_filter_check_discard(struct trace_event_call *call, void *rec, in call_filter_check_discard() argument 330 if (unlikely(call->flags & TRACE_EVENT_FL_FILTERED) && in call_filter_check_discard() 331 !filter_match_preds(call->filter, rec)) { in call_filter_check_discard() 1796 struct trace_event_call *call = &event_function; in trace_function() local 1809 if (!call_filter_check_discard(call, entry, buffer, event)) in trace_function() 1827 struct trace_event_call *call = &event_kernel_stack; in __ftrace_trace_stack() local 1901 if (!call_filter_check_discard(call, entry, buffer, event)) in __ftrace_trace_stack() 1957 struct trace_event_call *call = &event_user_stack; in ftrace_trace_userstack() local 1997 if (!call_filter_check_discard(call, entry, buffer, event)) in ftrace_trace_userstack() 2163 struct trace_event_call *call = &event_bprint; in trace_vbprintk() local [all …]
|
D | trace_event_perf.c | 360 int perf_ftrace_event_register(struct trace_event_call *call, in perf_ftrace_event_register() argument
|
D | trace_probe.h | 276 struct trace_event_call call; member
|
/kernel/rcu/ |
D | sync.c | 34 void (*call)(struct rcu_head *, void (*)(struct rcu_head *)); member 42 .call = call_rcu, 48 .call = call_rcu_sched, 54 .call = call_rcu_bh, 180 gp_ops[rsp->gp_type].call(&rsp->cb_head, rcu_sync_func); in rcu_sync_func() 208 gp_ops[rsp->gp_type].call(&rsp->cb_head, rcu_sync_func); in rcu_sync_exit()
|
D | rcutorture.c | 255 call_rcu_func_t call; member 384 .call = call_rcu, 424 .call = call_rcu_bh, 468 .call = call_rcu_busted, 569 .call = srcu_torture_call, 601 .call = srcu_torture_call, 640 .call = call_rcu_sched, 679 .call = call_rcu_tasks, 829 cur_ops->call && in rcu_torture_cbflood() 846 cur_ops->call(&rhp[i * cbflood_n_per_burst + j], in rcu_torture_cbflood() [all …]
|
D | tree.h | 465 call_rcu_func_t call; /* call_rcu() flavor. */ member
|
/kernel/ |
D | Kconfig.preempt | 30 is in kernel mode executing a system call. This allows 45 even if it is in kernel mode executing a system call and would
|
D | smp.c | 371 goto call; in smp_call_function_any() 378 goto call; in smp_call_function_any() 383 call: in smp_call_function_any()
|
D | Makefile | 119 $(call if_changed,gzip) 124 $(call filechk,ikconfiggz)
|
/kernel/gcov/ |
D | Makefile | 6 obj-$(CONFIG_GCOV_FORMAT_AUTODETECT) += $(call cc-ifversion, -lt, 0407, \
|
/kernel/bpf/ |
D | Makefile | 2 CFLAGS_core.o += $(call cc-disable-warning, override-init)
|
/kernel/debug/kdb/ |
D | Makefile | 25 $(call cmd,gen-kdb)
|