/kernel/time/ |
D | timer_stats.c | 285 long events = 0; in tstats_show() local 322 events += entry->count; in tstats_show() 329 if (events && period.tv_sec) in tstats_show() 331 events, events * 1000 / ms, in tstats_show() 332 (events * 1000000 / ms) % 1000); in tstats_show() 334 seq_printf(m, "%ld total events\n", events); in tstats_show()
|
D | Kconfig | 32 # The generic clock events infrastructure
|
/kernel/trace/ |
D | Kconfig | 164 bool "Enable trace events for preempt and irq disable/enable" 169 Enable tracing of disable and enable events for preemption and irqs. 170 For tracing preempt disable/enable events, DEBUG_PREEMPT must be 171 enabled. For tracing irq disable/enable events, PROVE_LOCKING must 269 bool "Trace process context switches and events" 283 Basic tracer to catch the syscall entry and exit events. 386 This traces the events of likely and unlikely condition 390 events into a running trace buffer to see when and where the 391 events happened, as well as their results. 447 bool "Enable kprobes-based dynamic events" [all …]
|
D | trace_events.c | 62 list_for_each_entry(file, &tr->events, list) 67 list_for_each_entry_safe(file, ___n, &tr->events, list) 503 list_for_each_entry(file, &tr->events, list) { in ftrace_clear_events() 629 list_for_each_entry(file, &tr->events, list) { in __ftrace_clear_event_pids() 746 list_for_each_entry(file, &tr->events, list) { in __ftrace_set_clr_event_nolock() 903 list_for_each_entry_continue(file, &tr->events, list) { in t_next() 925 file = list_entry(&tr->events, struct trace_event_file, list); in t_start() 942 list_for_each_entry_continue(file, &tr->events, list) { in s_next() 958 file = list_entry(&tr->events, struct trace_event_file, list); in s_start() 1104 list_for_each_entry(file, &tr->events, list) { in system_enable_read() [all …]
|
D | trace_output.c | 1470 static struct trace_event *events[] __initdata = { variable 1490 for (i = 0; events[i]; i++) { in init_events() 1491 event = events[i]; in init_events()
|
D | trace_events_filter.c | 892 list_for_each_entry(file, &tr->events, list) { in filter_free_subsystem_preds() 910 list_for_each_entry(file, &tr->events, list) { in filter_free_subsystem_filters() 1763 list_for_each_entry(file, &tr->events, list) { in replace_system_preds() 1778 list_for_each_entry(file, &tr->events, list) { in replace_system_preds()
|
D | ring_buffer.c | 4729 unsigned long events; member 4816 data->events++; in rb_write_something() 4957 total_events = data->events + data->events_nested; in test_ringbuffer()
|
D | trace_events_trigger.c | 483 list_for_each_entry(file, &tr->events, list) { in clear_event_triggers()
|
D | trace.h | 257 struct list_head events; member
|
D | trace_kprobe.c | 1368 list_for_each_entry(file, &tr->events, list) in find_trace_probe_file()
|
D | trace.c | 7170 INIT_LIST_HEAD(&tr->events); in instance_mkdir() 7806 INIT_LIST_HEAD(&global_trace.events); in tracer_alloc_buffers()
|
/kernel/ |
D | Kconfig.preempt | 28 This allows reaction to interactive events by allowing a 43 preemptible. This allows reaction to interactive events by
|
D | Kconfig.hz | 46 systems requiring fast interactive responses to events.
|
D | Makefile | 109 obj-$(CONFIG_PERF_EVENTS) += events/
|
/kernel/events/ |
D | internal.h | 26 local_t events; /* event limit */ member
|
D | core.c | 4494 unsigned int events = POLLHUP; in perf_poll() local 4499 return events; in perf_poll() 4508 events = atomic_xchg(&rb->poll, 0); in perf_poll() 4510 return events; in perf_poll() 5915 int events = local_inc_return(&rb->events); in perf_output_sample() local 5917 if (events >= wakeup_events) { in perf_output_sample() 5918 local_sub(wakeup_events, &rb->events); in perf_output_sample() 7152 int events = atomic_read(&event->event_limit); in __perf_event_overflow() local 7170 if (events && atomic_dec_and_test(&event->event_limit)) { in __perf_event_overflow() 10122 LIST_HEAD(events); in perf_pmu_migrate_context() [all …]
|
/kernel/power/ |
D | Kconfig | 155 wake-up events. 246 notification of APM "events" (e.g. battery status change).
|