Home
last modified time | relevance | path

Searched refs:disabled (Results 1 – 16 of 16) sorted by relevance

/kernel/trace/
Dtrace_sched_wakeup.c80 long disabled; in func_prolog_preempt_disable() local
94 disabled = atomic_inc_return(&(*data)->disabled); in func_prolog_preempt_disable()
95 if (unlikely(disabled != 1)) in func_prolog_preempt_disable()
101 atomic_dec(&(*data)->disabled); in func_prolog_preempt_disable()
127 atomic_dec(&data->disabled); in wakeup_tracer_call()
238 atomic_dec(&data->disabled); in wakeup_graph_entry()
256 atomic_dec(&data->disabled); in wakeup_graph_return()
374 long disabled; in probe_wakeup_sched_switch() local
399 disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->trace_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch()
400 if (likely(disabled != 1)) in probe_wakeup_sched_switch()
[all …]
Dtrace_irqsoff.c107 long disabled; in func_prolog_dec() local
126 disabled = atomic_inc_return(&(*data)->disabled); in func_prolog_dec()
128 if (likely(disabled == 1)) in func_prolog_dec()
131 atomic_dec(&(*data)->disabled); in func_prolog_dec()
152 atomic_dec(&data->disabled); in irqsoff_tracer_call()
197 atomic_dec(&data->disabled); in irqsoff_graph_entry()
214 atomic_dec(&data->disabled); in irqsoff_graph_return()
386 if (unlikely(!data) || atomic_read(&data->disabled)) in start_critical_timing()
389 atomic_inc(&data->disabled); in start_critical_timing()
401 atomic_dec(&data->disabled); in start_critical_timing()
[all …]
Dtrace_functions.c80 if (!atomic_read(&data->disabled)) { in function_trace_call()
97 long disabled; in function_stack_trace_call() local
111 disabled = atomic_inc_return(&data->disabled); in function_stack_trace_call()
113 if (likely(disabled == 1)) { in function_stack_trace_call()
127 atomic_dec(&data->disabled); in function_stack_trace_call()
Dtrace_functions_graph.c256 long disabled; in trace_graph_entry() local
273 disabled = atomic_inc_return(&data->disabled); in trace_graph_entry()
274 if (likely(disabled == 1)) { in trace_graph_entry()
281 atomic_dec(&data->disabled); in trace_graph_entry()
351 long disabled; in trace_graph_return() local
358 disabled = atomic_inc_return(&data->disabled); in trace_graph_return()
359 if (likely(disabled == 1)) { in trace_graph_return()
363 atomic_dec(&data->disabled); in trace_graph_return()
Dtrace_kdb.c29 atomic_inc(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf()
86 atomic_dec(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf()
Dtrace_sched_switch.c74 if (likely(!atomic_read(&data->disabled))) in probe_sched_switch()
128 if (likely(!atomic_read(&data->disabled))) in probe_sched_wakeup()
DKconfig154 tracing is enabled by the administrator. If it's runtime disabled
189 disabled by default and can be runtime (re-)started
213 disabled by default and can be runtime (re-)started
306 Otherwise keep it disabled.
373 is disabled.
378 The stack tracer can also be enabled or disabled via the
524 implementation and works via page faults. Tracing is disabled by
567 and all ring buffers will be disabled.
Dtrace_branch.c56 if (atomic_inc_return(&data->disabled) != 1) in probe_likely_condition()
85 atomic_dec(&data->disabled); in probe_likely_condition()
Dftrace.c222 *per_cpu_ptr(ops->disabled, cpu) = 1; in control_ops_disable_all()
227 int __percpu *disabled; in control_ops_alloc() local
229 disabled = alloc_percpu(int); in control_ops_alloc()
230 if (!disabled) in control_ops_alloc()
233 ops->disabled = disabled; in control_ops_alloc()
240 free_percpu(ops->disabled); in control_ops_free()
493 atomic_t disabled; member
808 if (atomic_inc_return(&stat->disabled) != 1) in ftrace_profile_alloc()
830 atomic_dec(&stat->disabled); in ftrace_profile_alloc()
Dtrace_events.c2610 long disabled; in function_test_events_call() local
2617 disabled = atomic_inc_return(&per_cpu(ftrace_test_event_disable, cpu)); in function_test_events_call()
2619 if (disabled != 1) in function_test_events_call()
Dtrace.h146 atomic_t disabled; member
Dtrace.c1573 if (likely(!atomic_read(&data->disabled))) in ftrace()
3146 atomic_inc(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); in tracing_cpumask_write()
3151 atomic_dec(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); in tracing_cpumask_write()
6200 atomic_inc(&per_cpu_ptr(iter.tr->trace_buffer.data, cpu)->disabled); in ftrace_dump()
6272 atomic_dec(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump()
/kernel/events/
Dhw_breakpoint.c473 if (attr->disabled) in modify_user_hw_breakpoint()
484 if (!bp->attr.disabled) in modify_user_hw_breakpoint()
491 bp->attr.disabled = attr->disabled; in modify_user_hw_breakpoint()
Dcore.c1023 event->state = event->attr.disabled ? PERF_EVENT_STATE_OFF : in perf_event__state_init()
/kernel/
Dwatchdog.c288 .disabled = 1,
Dcgroup.c1211 if (ss->disabled) in parse_cgroupfs_options()
1236 if (ss->disabled) in parse_cgroupfs_options()
4799 ss->root->number_of_cgroups, !ss->disabled); in proc_cgroupstats_show()
5122 ss->disabled = 1; in cgroup_disable()