/kernel/ |
D | jump_label.c | 60 if (atomic_inc_not_zero(&key->enabled)) in static_key_slow_inc() 64 if (atomic_read(&key->enabled) == 0) { in static_key_slow_inc() 70 atomic_inc(&key->enabled); in static_key_slow_inc() 78 if (!atomic_dec_and_mutex_lock(&key->enabled, &jump_label_mutex)) { in __static_key_slow_dec() 79 WARN(atomic_read(&key->enabled) < 0, in __static_key_slow_dec() 85 atomic_inc(&key->enabled); in __static_key_slow_dec()
|
D | stop_machine.c | 38 bool enabled; /* is this stopper enabled? */ member 74 if (stopper->enabled) { in cpu_stop_queue_work() 318 stopper->enabled = false; in cpu_stop_park() 327 stopper->enabled = true; in cpu_stop_unpark()
|
D | audit.c | 662 status_set.enabled = audit_enabled; in audit_receive_msg() 677 err = audit_set_enabled(status_get->enabled); in audit_receive_msg() 827 s.enabled = tsk->signal->audit_tty != 0; in audit_receive_msg() 842 if ((s.enabled != 0 && s.enabled != 1) || in audit_receive_msg() 847 tsk->signal->audit_tty = s.enabled; in audit_receive_msg()
|
/kernel/trace/ |
D | Kconfig | 95 Adds a very slight overhead to tracing when enabled. 98 # enabled by all tracers (context switch and event tracer) they select TRACING. 154 tracing is enabled by the administrator. If it's runtime disabled 195 enabled. This option and the preempt-off timing option can be 219 enabled. This option and the irqs-off timing option can be 273 When this is enabled, this adds a little more overhead to the 276 of the overall system. This is enabled by default when the preempt 277 or irq latency tracers are enabled, as those need to swap as well 332 This configuration, when enabled, will impose a great overhead 333 on the system. This should only be enabled when the system [all …]
|
D | trace.c | 1893 static void trace_printk_start_stop_comm(int enabled) in trace_printk_start_stop_comm() argument 1898 if (enabled) in trace_printk_start_stop_comm() 3245 if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set) in trace_keep_overwrite() 3251 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled) in set_tracer_flag() argument 3254 if (!!(trace_flags & mask) == !!enabled) in set_tracer_flag() 3259 if (tr->current_trace->flag_changed(tr->current_trace, mask, !!enabled)) in set_tracer_flag() 3262 if (enabled) in set_tracer_flag() 3268 trace_event_enable_cmd_record(enabled); in set_tracer_flag() 3271 ring_buffer_change_overwrite(tr->trace_buffer.buffer, enabled); in set_tracer_flag() 3273 ring_buffer_change_overwrite(tr->max_buffer.buffer, enabled); in set_tracer_flag() [all …]
|
D | trace.h | 366 bool enabled; member 1035 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled);
|
D | trace_events.c | 2001 char *glob, char *cmd, char *param, int enabled) in event_enable_func() argument 2014 if (!enabled) in event_enable_func()
|
/kernel/sched/ |
D | auto_group.h | 32 int enabled = ACCESS_ONCE(sysctl_sched_autogroup_enabled); in autogroup_task_group() local 34 if (enabled && task_wants_autogroup(p, tg)) in autogroup_task_group()
|
D | sched.h | 724 #define SCHED_FEAT(name, enabled) \ argument 745 #define SCHED_FEAT(name, enabled) \ argument 748 return static_branch__##enabled(key); \ 1321 extern void account_cfs_bandwidth_used(int enabled, int was_enabled);
|
D | core.c | 133 #define SCHED_FEAT(name, enabled) \ argument 134 (1UL << __SCHED_FEAT_##name) * enabled | 143 #define SCHED_FEAT(name, enabled) \ argument 171 #define SCHED_FEAT(name, enabled) \ argument 172 jump_label_key__##enabled , 1648 void set_numabalancing_state(bool enabled) in set_numabalancing_state() argument 1650 if (enabled) in set_numabalancing_state() 1658 void set_numabalancing_state(bool enabled) in set_numabalancing_state() argument 1660 numabalancing_enabled = enabled; in set_numabalancing_state()
|
D | fair.c | 2025 void account_cfs_bandwidth_used(int enabled, int was_enabled) in account_cfs_bandwidth_used() argument 2028 if (enabled && !was_enabled) in account_cfs_bandwidth_used() 2030 else if (!enabled && was_enabled) in account_cfs_bandwidth_used() 2039 void account_cfs_bandwidth_used(int enabled, int was_enabled) {} in account_cfs_bandwidth_used() argument
|
/kernel/power/ |
D | suspend_test.c | 86 alm.enabled = true; in test_wakealarm() 111 alm.enabled = false; in test_wakealarm()
|
/kernel/events/ |
D | core.c | 2771 int enabled = 0; in perf_event_enable_on_exec() local 2793 enabled = 1; in perf_event_enable_on_exec() 2799 if (enabled) in perf_event_enable_on_exec() 3188 u64 perf_event_read_value(struct perf_event *event, u64 *enabled, u64 *running) in perf_event_read_value() argument 3193 *enabled = 0; in perf_event_read_value() 3198 *enabled += event->total_time_enabled + in perf_event_read_value() 3205 *enabled += child->total_time_enabled; in perf_event_read_value() 3220 u64 count, enabled, running; in perf_event_read_group() local 3225 count = perf_event_read_value(leader, &enabled, &running); in perf_event_read_group() 3229 values[n++] = enabled; in perf_event_read_group() [all …]
|