/kernel/time/ |
D | timer_stats.c | 285 long events = 0; in tstats_show() local 322 events += entry->count; in tstats_show() 329 if (events && period.tv_sec) in tstats_show() 331 events, events * 1000 / ms, in tstats_show() 332 (events * 1000000 / ms) % 1000); in tstats_show() 334 seq_printf(m, "%ld total events\n", events); in tstats_show()
|
D | Kconfig | 32 # The generic clock events infrastructure
|
/kernel/trace/ |
D | Kconfig | 228 bool "Trace process context switches and events" 242 Basic tracer to catch the syscall entry and exit events. 345 This traces the events of likely and unlikely condition 349 events into a running trace buffer to see when and where the 350 events happened, as well as their results. 406 bool "Enable kprobes-based dynamic events" 411 This allows the user to add tracing events (similar to tracepoints) 415 Those events can be inserted wherever kprobes can probe, and record 422 bool "Enable uprobes-based dynamic events" 431 This allows the user to add tracing events on top of userspace [all …]
|
D | trace_events.c | 60 list_for_each_entry(file, &tr->events, list) 65 list_for_each_entry_safe(file, ___n, &tr->events, list) 389 list_for_each_entry(file, &tr->events, list) { in ftrace_clear_events() 491 list_for_each_entry(file, &tr->events, list) { in __ftrace_set_clr_event_nolock() 641 list_for_each_entry_continue(file, &tr->events, list) { in t_next() 663 file = list_entry(&tr->events, struct ftrace_event_file, list); in t_start() 680 list_for_each_entry_continue(file, &tr->events, list) { in s_next() 696 file = list_entry(&tr->events, struct ftrace_event_file, list); in s_start() 803 list_for_each_entry(file, &tr->events, list) { in system_enable_read() 1693 list_add(&file->list, &tr->events); in trace_create_new_event() [all …]
|
D | trace_output.c | 1421 static struct trace_event *events[] __initdata = { variable 1440 for (i = 0; events[i]; i++) { in init_events() 1441 event = events[i]; in init_events()
|
D | trace_events_filter.c | 850 list_for_each_entry(file, &tr->events, list) { in filter_free_subsystem_preds() 875 list_for_each_entry(file, &tr->events, list) { in filter_free_subsystem_filters() 1746 list_for_each_entry(file, &tr->events, list) { in replace_system_preds() 1761 list_for_each_entry(file, &tr->events, list) { in replace_system_preds()
|
D | trace.h | 226 struct list_head events; member
|
D | ring_buffer.c | 4742 unsigned long events; member 4829 data->events++; in rb_write_something() 4970 total_events = data->events + data->events_nested; in test_ringbuffer()
|
D | trace_events_trigger.c | 471 list_for_each_entry(file, &tr->events, list) { in clear_event_triggers()
|
D | trace_kprobe.c | 1368 list_for_each_entry(file, &tr->events, list) in find_trace_probe_file()
|
D | trace.c | 6487 INIT_LIST_HEAD(&tr->events); in instance_mkdir() 6995 INIT_LIST_HEAD(&global_trace.events); in tracer_alloc_buffers()
|
/kernel/ |
D | Kconfig.preempt | 28 This allows reaction to interactive events by allowing a 43 preemptible. This allows reaction to interactive events by
|
D | Kconfig.hz | 46 systems requiring fast interactive responses to events.
|
D | Makefile | 103 obj-$(CONFIG_PERF_EVENTS) += events/
|
/kernel/events/ |
D | internal.h | 25 local_t events; /* event limit */ member
|
D | core.c | 3800 unsigned int events = POLLHUP; in perf_poll() local 3805 return events; in perf_poll() 3814 events = atomic_xchg(&rb->poll, 0); in perf_poll() 3816 return events; in perf_poll() 5067 int events = local_inc_return(&rb->events); in perf_output_sample() local 5069 if (events >= wakeup_events) { in perf_output_sample() 5070 local_sub(wakeup_events, &rb->events); in perf_output_sample() 5777 int events = atomic_read(&event->event_limit); in __perf_event_overflow() local 5821 if (events && atomic_dec_and_test(&event->event_limit)) { in __perf_event_overflow() 7843 LIST_HEAD(events); in perf_pmu_migrate_context() [all …]
|
/kernel/power/ |
D | Kconfig | 153 wake-up events. 248 notification of APM "events" (e.g. battery status change).
|