Searched refs:disabled (Results 1 – 16 of 16) sorted by relevance
/kernel/trace/ |
D | trace_sched_wakeup.c | 80 long disabled; in func_prolog_preempt_disable() local 94 disabled = atomic_inc_return(&(*data)->disabled); in func_prolog_preempt_disable() 95 if (unlikely(disabled != 1)) in func_prolog_preempt_disable() 101 atomic_dec(&(*data)->disabled); in func_prolog_preempt_disable() 127 atomic_dec(&data->disabled); in wakeup_tracer_call() 238 atomic_dec(&data->disabled); in wakeup_graph_entry() 256 atomic_dec(&data->disabled); in wakeup_graph_return() 374 long disabled; in probe_wakeup_sched_switch() local 399 disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->trace_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 400 if (likely(disabled != 1)) in probe_wakeup_sched_switch() [all …]
|
D | trace_irqsoff.c | 107 long disabled; in func_prolog_dec() local 126 disabled = atomic_inc_return(&(*data)->disabled); in func_prolog_dec() 128 if (likely(disabled == 1)) in func_prolog_dec() 131 atomic_dec(&(*data)->disabled); in func_prolog_dec() 152 atomic_dec(&data->disabled); in irqsoff_tracer_call() 197 atomic_dec(&data->disabled); in irqsoff_graph_entry() 214 atomic_dec(&data->disabled); in irqsoff_graph_return() 386 if (unlikely(!data) || atomic_read(&data->disabled)) in start_critical_timing() 389 atomic_inc(&data->disabled); in start_critical_timing() 401 atomic_dec(&data->disabled); in start_critical_timing() [all …]
|
D | trace_functions.c | 80 if (!atomic_read(&data->disabled)) { in function_trace_call() 97 long disabled; in function_stack_trace_call() local 111 disabled = atomic_inc_return(&data->disabled); in function_stack_trace_call() 113 if (likely(disabled == 1)) { in function_stack_trace_call() 127 atomic_dec(&data->disabled); in function_stack_trace_call()
|
D | trace_functions_graph.c | 256 long disabled; in trace_graph_entry() local 273 disabled = atomic_inc_return(&data->disabled); in trace_graph_entry() 274 if (likely(disabled == 1)) { in trace_graph_entry() 281 atomic_dec(&data->disabled); in trace_graph_entry() 351 long disabled; in trace_graph_return() local 358 disabled = atomic_inc_return(&data->disabled); in trace_graph_return() 359 if (likely(disabled == 1)) { in trace_graph_return() 363 atomic_dec(&data->disabled); in trace_graph_return()
|
D | trace_kdb.c | 29 atomic_inc(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf() 86 atomic_dec(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf()
|
D | trace_sched_switch.c | 74 if (likely(!atomic_read(&data->disabled))) in probe_sched_switch() 128 if (likely(!atomic_read(&data->disabled))) in probe_sched_wakeup()
|
D | Kconfig | 154 tracing is enabled by the administrator. If it's runtime disabled 189 disabled by default and can be runtime (re-)started 213 disabled by default and can be runtime (re-)started 306 Otherwise keep it disabled. 373 is disabled. 378 The stack tracer can also be enabled or disabled via the 524 implementation and works via page faults. Tracing is disabled by 567 and all ring buffers will be disabled.
|
D | trace_branch.c | 56 if (atomic_inc_return(&data->disabled) != 1) in probe_likely_condition() 85 atomic_dec(&data->disabled); in probe_likely_condition()
|
D | ftrace.c | 222 *per_cpu_ptr(ops->disabled, cpu) = 1; in control_ops_disable_all() 227 int __percpu *disabled; in control_ops_alloc() local 229 disabled = alloc_percpu(int); in control_ops_alloc() 230 if (!disabled) in control_ops_alloc() 233 ops->disabled = disabled; in control_ops_alloc() 240 free_percpu(ops->disabled); in control_ops_free() 493 atomic_t disabled; member 808 if (atomic_inc_return(&stat->disabled) != 1) in ftrace_profile_alloc() 830 atomic_dec(&stat->disabled); in ftrace_profile_alloc()
|
D | trace_events.c | 2610 long disabled; in function_test_events_call() local 2617 disabled = atomic_inc_return(&per_cpu(ftrace_test_event_disable, cpu)); in function_test_events_call() 2619 if (disabled != 1) in function_test_events_call()
|
D | trace.h | 146 atomic_t disabled; member
|
D | trace.c | 1573 if (likely(!atomic_read(&data->disabled))) in ftrace() 3146 atomic_inc(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); in tracing_cpumask_write() 3151 atomic_dec(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); in tracing_cpumask_write() 6200 atomic_inc(&per_cpu_ptr(iter.tr->trace_buffer.data, cpu)->disabled); in ftrace_dump() 6272 atomic_dec(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump()
|
/kernel/events/ |
D | hw_breakpoint.c | 473 if (attr->disabled) in modify_user_hw_breakpoint() 484 if (!bp->attr.disabled) in modify_user_hw_breakpoint() 491 bp->attr.disabled = attr->disabled; in modify_user_hw_breakpoint()
|
D | core.c | 1023 event->state = event->attr.disabled ? PERF_EVENT_STATE_OFF : in perf_event__state_init()
|
/kernel/ |
D | watchdog.c | 288 .disabled = 1,
|
D | cgroup.c | 1211 if (ss->disabled) in parse_cgroupfs_options() 1236 if (ss->disabled) in parse_cgroupfs_options() 4799 ss->root->number_of_cgroups, !ss->disabled); in proc_cgroupstats_show() 5122 ss->disabled = 1; in cgroup_disable()
|