Home
last modified time | relevance | path

Searched refs:call (Results 1 – 25 of 28) sorted by relevance

12

/kernel/trace/
Dtrace_events.c95 trace_find_event_field(struct trace_event_call *call, char *name) in trace_find_event_field() argument
100 head = trace_get_fields(call); in trace_find_event_field()
139 int trace_define_field(struct trace_event_call *call, const char *type, in trace_define_field() argument
145 if (WARN_ON(!call->class)) in trace_define_field()
148 head = trace_get_fields(call); in trace_define_field()
195 static void trace_destroy_fields(struct trace_event_call *call) in trace_destroy_fields() argument
200 head = trace_get_fields(call); in trace_destroy_fields()
207 int trace_event_raw_init(struct trace_event_call *call) in trace_event_raw_init() argument
211 id = register_trace_event(&call->event); in trace_event_raw_init()
301 int trace_event_reg(struct trace_event_call *call, in trace_event_reg() argument
[all …]
Dtrace_syscalls.c22 syscall_get_enter_fields(struct trace_event_call *call) in syscall_get_enter_fields() argument
24 struct syscall_metadata *entry = call->data; in syscall_get_enter_fields()
223 static int __init set_syscall_print_fmt(struct trace_event_call *call) in set_syscall_print_fmt() argument
227 struct syscall_metadata *entry = call->data; in set_syscall_print_fmt()
229 if (entry->enter_event != call) { in set_syscall_print_fmt()
230 call->print_fmt = "\"0x%lx\", REC->ret"; in set_syscall_print_fmt()
243 call->print_fmt = print_fmt; in set_syscall_print_fmt()
248 static void __init free_syscall_print_fmt(struct trace_event_call *call) in free_syscall_print_fmt() argument
250 struct syscall_metadata *entry = call->data; in free_syscall_print_fmt()
252 if (entry->enter_event == call) in free_syscall_print_fmt()
[all …]
Dtrace_kprobe.c296 tk->tp.call.class = &tk->tp.class; in alloc_trace_kprobe()
297 tk->tp.call.name = kstrdup(event, GFP_KERNEL); in alloc_trace_kprobe()
298 if (!tk->tp.call.name) in alloc_trace_kprobe()
314 kfree(tk->tp.call.name); in alloc_trace_kprobe()
327 kfree(tk->tp.call.class->system); in free_trace_kprobe()
328 kfree(tk->tp.call.name); in free_trace_kprobe()
339 if (strcmp(trace_event_name(&tk->tp.call), event) == 0 && in find_trace_kprobe()
340 strcmp(tk->tp.call.class->system, group) == 0) in find_trace_kprobe()
525 old_tk = find_trace_kprobe(trace_event_name(&tk->tp.call), in register_trace_kprobe()
526 tk->tp.call.class->system); in register_trace_kprobe()
[all …]
Dtrace_export.c163 #define FTRACE_ENTRY_REG(call, struct_name, etype, tstruct, print, filter,\ argument
166 struct trace_event_class __refdata event_class_ftrace_##call = { \
168 .define_fields = ftrace_define_fields_##call, \
169 .fields = LIST_HEAD_INIT(event_class_ftrace_##call.fields),\
173 struct trace_event_call __used event_##call = { \
174 .class = &event_class_ftrace_##call, \
176 .name = #call, \
183 __attribute__((section("_ftrace_events"))) *__event_##call = &event_##call;
186 #define FTRACE_ENTRY(call, struct_name, etype, tstruct, print, filter) \ argument
187 FTRACE_ENTRY_REG(call, struct_name, etype, \
[all …]
Dtrace_uprobe.c262 tu->tp.call.class = &tu->tp.class; in alloc_trace_uprobe()
263 tu->tp.call.name = kstrdup(event, GFP_KERNEL); in alloc_trace_uprobe()
264 if (!tu->tp.call.name) in alloc_trace_uprobe()
280 kfree(tu->tp.call.name); in alloc_trace_uprobe()
294 kfree(tu->tp.call.class->system); in free_trace_uprobe()
295 kfree(tu->tp.call.name); in free_trace_uprobe()
305 if (strcmp(trace_event_name(&tu->tp.call), event) == 0 && in find_probe_event()
306 strcmp(tu->tp.call.class->system, group) == 0) in find_probe_event()
335 old_tu = find_probe_event(trace_event_name(&tu->tp.call), in register_trace_uprobe()
336 tu->tp.call.class->system); in register_trace_uprobe()
[all …]
Dtrace_functions_graph.c292 struct trace_event_call *call = &event_funcgraph_entry; in __trace_graph_entry() local
303 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_graph_entry()
404 struct trace_event_call *call = &event_funcgraph_exit; in __trace_graph_return() local
415 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_graph_return()
775 struct ftrace_graph_ent *call; in print_graph_entry_leaf() local
781 call = &entry->graph_ent; in print_graph_entry_leaf()
790 if (call->depth < -1) in print_graph_entry_leaf()
791 call->depth += FTRACE_NOTRACE_DEPTH; in print_graph_entry_leaf()
798 cpu_data->depth = call->depth - 1; in print_graph_entry_leaf()
801 if (call->depth < FTRACE_RETFUNC_DEPTH && in print_graph_entry_leaf()
[all …]
Dtrace_events_filter.c832 struct trace_event_call *call = file->event_call; in filter_disable() local
834 if (call->flags & TRACE_EVENT_FL_USE_CALL_FILTER) in filter_disable()
835 call->flags &= ~TRACE_EVENT_FL_FILTERED; in filter_disable()
889 struct trace_event_call *call = file->event_call; in __remove_filter() local
892 if (call->flags & TRACE_EVENT_FL_USE_CALL_FILTER) in __remove_filter()
893 remove_filter_string(call->filter); in __remove_filter()
912 struct trace_event_call *call = file->event_call; in __free_subsystem_filter() local
914 if (call->flags & TRACE_EVENT_FL_USE_CALL_FILTER) { in __free_subsystem_filter()
915 __free_filter(call->filter); in __free_subsystem_filter()
916 call->filter = NULL; in __free_subsystem_filter()
[all …]
Dtrace.h871 bool ftrace_event_is_function(struct trace_event_call *call);
1127 extern int create_event_filter(struct trace_event_call *call,
1133 trace_find_event_field(struct trace_event_call *call, char *name);
1352 #define FTRACE_ENTRY(call, struct_name, id, tstruct, print, filter) \ argument
1354 __aligned(4) event_##call;
1356 #define FTRACE_ENTRY_DUP(call, struct_name, id, tstruct, print, filter) \ argument
1357 FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print), \
1362 int perf_ftrace_event_register(struct trace_event_call *call,
Dtrace_mmiotrace.c301 struct trace_event_call *call = &event_mmiotrace_rw; in __trace_mmiotrace_rw() local
316 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_mmiotrace_rw()
331 struct trace_event_call *call = &event_mmiotrace_map; in __trace_mmiotrace_map() local
346 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_mmiotrace_map()
Dtrace_sched_wakeup.c376 struct trace_event_call *call = &event_context_switch; in tracing_sched_switch_trace() local
394 if (!call_filter_check_discard(call, entry, buffer, event)) in tracing_sched_switch_trace()
404 struct trace_event_call *call = &event_wakeup; in tracing_sched_wakeup_trace() local
422 if (!call_filter_check_discard(call, entry, buffer, event)) in tracing_sched_wakeup_trace()
Dtrace_branch.c32 struct trace_event_call *call = &event_branch; in probe_likely_condition() local
83 if (!call_filter_check_discard(call, entry, buffer, event)) in probe_likely_condition()
DMakefile3 FRAME_CFLAGS := $(call cc-disable-warning,frame-address)
DKconfig148 sequence is then dynamically patched into a tracer call when
163 draw a call graph for each thread with some information like
375 This tracer works by hooking into every function call that the
527 up since it runs this on every system call defined.
529 TBD - enable a way to actually call the syscalls as we test their
Dtrace.c326 int call_filter_check_discard(struct trace_event_call *call, void *rec, in call_filter_check_discard() argument
330 if (unlikely(call->flags & TRACE_EVENT_FL_FILTERED) && in call_filter_check_discard()
331 !filter_match_preds(call->filter, rec)) { in call_filter_check_discard()
1796 struct trace_event_call *call = &event_function; in trace_function() local
1809 if (!call_filter_check_discard(call, entry, buffer, event)) in trace_function()
1827 struct trace_event_call *call = &event_kernel_stack; in __ftrace_trace_stack() local
1901 if (!call_filter_check_discard(call, entry, buffer, event)) in __ftrace_trace_stack()
1957 struct trace_event_call *call = &event_user_stack; in ftrace_trace_userstack() local
1997 if (!call_filter_check_discard(call, entry, buffer, event)) in ftrace_trace_userstack()
2163 struct trace_event_call *call = &event_bprint; in trace_vbprintk() local
[all …]
Dtrace_event_perf.c360 int perf_ftrace_event_register(struct trace_event_call *call, in perf_ftrace_event_register() argument
Dtrace_probe.h276 struct trace_event_call call; member
/kernel/rcu/
Dsync.c34 void (*call)(struct rcu_head *, void (*)(struct rcu_head *)); member
42 .call = call_rcu,
48 .call = call_rcu_sched,
54 .call = call_rcu_bh,
180 gp_ops[rsp->gp_type].call(&rsp->cb_head, rcu_sync_func); in rcu_sync_func()
208 gp_ops[rsp->gp_type].call(&rsp->cb_head, rcu_sync_func); in rcu_sync_exit()
Drcutorture.c255 call_rcu_func_t call; member
384 .call = call_rcu,
424 .call = call_rcu_bh,
468 .call = call_rcu_busted,
569 .call = srcu_torture_call,
601 .call = srcu_torture_call,
640 .call = call_rcu_sched,
679 .call = call_rcu_tasks,
829 cur_ops->call && in rcu_torture_cbflood()
846 cur_ops->call(&rhp[i * cbflood_n_per_burst + j], in rcu_torture_cbflood()
[all …]
Dtree.h465 call_rcu_func_t call; /* call_rcu() flavor. */ member
/kernel/
DKconfig.preempt30 is in kernel mode executing a system call. This allows
45 even if it is in kernel mode executing a system call and would
Dsmp.c371 goto call; in smp_call_function_any()
378 goto call; in smp_call_function_any()
383 call: in smp_call_function_any()
DMakefile119 $(call if_changed,gzip)
124 $(call filechk,ikconfiggz)
/kernel/gcov/
DMakefile6 obj-$(CONFIG_GCOV_FORMAT_AUTODETECT) += $(call cc-ifversion, -lt, 0407, \
/kernel/bpf/
DMakefile2 CFLAGS_core.o += $(call cc-disable-warning, override-init)
/kernel/debug/kdb/
DMakefile25 $(call cmd,gen-kdb)

12