Home
last modified time | relevance | path

Searched refs:enabled (Results 1 – 13 of 13) sorted by relevance

/kernel/
Djump_label.c60 if (atomic_inc_not_zero(&key->enabled)) in static_key_slow_inc()
64 if (atomic_read(&key->enabled) == 0) { in static_key_slow_inc()
70 atomic_inc(&key->enabled); in static_key_slow_inc()
78 if (!atomic_dec_and_mutex_lock(&key->enabled, &jump_label_mutex)) { in __static_key_slow_dec()
79 WARN(atomic_read(&key->enabled) < 0, in __static_key_slow_dec()
85 atomic_inc(&key->enabled); in __static_key_slow_dec()
Dstop_machine.c38 bool enabled; /* is this stopper enabled? */ member
74 if (stopper->enabled) { in cpu_stop_queue_work()
318 stopper->enabled = false; in cpu_stop_park()
327 stopper->enabled = true; in cpu_stop_unpark()
Daudit.c662 status_set.enabled = audit_enabled; in audit_receive_msg()
677 err = audit_set_enabled(status_get->enabled); in audit_receive_msg()
827 s.enabled = tsk->signal->audit_tty != 0; in audit_receive_msg()
842 if ((s.enabled != 0 && s.enabled != 1) || in audit_receive_msg()
847 tsk->signal->audit_tty = s.enabled; in audit_receive_msg()
/kernel/trace/
DKconfig95 Adds a very slight overhead to tracing when enabled.
98 # enabled by all tracers (context switch and event tracer) they select TRACING.
154 tracing is enabled by the administrator. If it's runtime disabled
195 enabled. This option and the preempt-off timing option can be
219 enabled. This option and the irqs-off timing option can be
273 When this is enabled, this adds a little more overhead to the
276 of the overall system. This is enabled by default when the preempt
277 or irq latency tracers are enabled, as those need to swap as well
332 This configuration, when enabled, will impose a great overhead
333 on the system. This should only be enabled when the system
[all …]
Dtrace.c1893 static void trace_printk_start_stop_comm(int enabled) in trace_printk_start_stop_comm() argument
1898 if (enabled) in trace_printk_start_stop_comm()
3245 if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set) in trace_keep_overwrite()
3251 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled) in set_tracer_flag() argument
3254 if (!!(trace_flags & mask) == !!enabled) in set_tracer_flag()
3259 if (tr->current_trace->flag_changed(tr->current_trace, mask, !!enabled)) in set_tracer_flag()
3262 if (enabled) in set_tracer_flag()
3268 trace_event_enable_cmd_record(enabled); in set_tracer_flag()
3271 ring_buffer_change_overwrite(tr->trace_buffer.buffer, enabled); in set_tracer_flag()
3273 ring_buffer_change_overwrite(tr->max_buffer.buffer, enabled); in set_tracer_flag()
[all …]
Dtrace.h366 bool enabled; member
1035 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled);
Dtrace_events.c2001 char *glob, char *cmd, char *param, int enabled) in event_enable_func() argument
2014 if (!enabled) in event_enable_func()
/kernel/sched/
Dauto_group.h32 int enabled = ACCESS_ONCE(sysctl_sched_autogroup_enabled); in autogroup_task_group() local
34 if (enabled && task_wants_autogroup(p, tg)) in autogroup_task_group()
Dsched.h724 #define SCHED_FEAT(name, enabled) \ argument
745 #define SCHED_FEAT(name, enabled) \ argument
748 return static_branch__##enabled(key); \
1321 extern void account_cfs_bandwidth_used(int enabled, int was_enabled);
Dcore.c133 #define SCHED_FEAT(name, enabled) \ argument
134 (1UL << __SCHED_FEAT_##name) * enabled |
143 #define SCHED_FEAT(name, enabled) \ argument
171 #define SCHED_FEAT(name, enabled) \ argument
172 jump_label_key__##enabled ,
1648 void set_numabalancing_state(bool enabled) in set_numabalancing_state() argument
1650 if (enabled) in set_numabalancing_state()
1658 void set_numabalancing_state(bool enabled) in set_numabalancing_state() argument
1660 numabalancing_enabled = enabled; in set_numabalancing_state()
Dfair.c2025 void account_cfs_bandwidth_used(int enabled, int was_enabled) in account_cfs_bandwidth_used() argument
2028 if (enabled && !was_enabled) in account_cfs_bandwidth_used()
2030 else if (!enabled && was_enabled) in account_cfs_bandwidth_used()
2039 void account_cfs_bandwidth_used(int enabled, int was_enabled) {} in account_cfs_bandwidth_used() argument
/kernel/power/
Dsuspend_test.c86 alm.enabled = true; in test_wakealarm()
111 alm.enabled = false; in test_wakealarm()
/kernel/events/
Dcore.c2771 int enabled = 0; in perf_event_enable_on_exec() local
2793 enabled = 1; in perf_event_enable_on_exec()
2799 if (enabled) in perf_event_enable_on_exec()
3188 u64 perf_event_read_value(struct perf_event *event, u64 *enabled, u64 *running) in perf_event_read_value() argument
3193 *enabled = 0; in perf_event_read_value()
3198 *enabled += event->total_time_enabled + in perf_event_read_value()
3205 *enabled += child->total_time_enabled; in perf_event_read_value()
3220 u64 count, enabled, running; in perf_event_read_group() local
3225 count = perf_event_read_value(leader, &enabled, &running); in perf_event_read_group()
3229 values[n++] = enabled; in perf_event_read_group()
[all …]