Home
last modified time | relevance | path

Searched refs:enabled (Results 1 – 25 of 28) sorted by relevance

12

/kernel/
Djump_label.c110 int n = atomic_read(&key->enabled); in static_key_count()
135 for (v = atomic_read(&key->enabled); v > 0; v = v1) { in static_key_slow_inc_cpuslocked()
136 v1 = atomic_cmpxchg(&key->enabled, v, v + 1); in static_key_slow_inc_cpuslocked()
142 if (atomic_read(&key->enabled) == 0) { in static_key_slow_inc_cpuslocked()
143 atomic_set(&key->enabled, -1); in static_key_slow_inc_cpuslocked()
149 atomic_set_release(&key->enabled, 1); in static_key_slow_inc_cpuslocked()
151 atomic_inc(&key->enabled); in static_key_slow_inc_cpuslocked()
169 if (atomic_read(&key->enabled) > 0) { in static_key_enable_cpuslocked()
170 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable_cpuslocked()
175 if (atomic_read(&key->enabled) == 0) { in static_key_enable_cpuslocked()
[all …]
Dstop_machine.c41 bool enabled; /* is this stopper enabled? */ member
82 bool enabled; in cpu_stop_queue_work() local
86 enabled = stopper->enabled; in cpu_stop_queue_work()
87 if (enabled) in cpu_stop_queue_work()
96 return enabled; in cpu_stop_queue_work()
264 if (!stopper1->enabled || !stopper2->enabled) { in cpu_stop_queue_two_works()
541 stopper->enabled = false; in stop_machine_park()
563 stopper->enabled = true; in stop_machine_unpark()
Daudit.c1195 s.enabled = audit_enabled; in audit_receive_msg()
1215 err = audit_set_enabled(s.enabled); in audit_receive_msg()
1445 s.enabled = t & AUDIT_TTY_ENABLE; in audit_receive_msg()
1460 if ((s.enabled != 0 && s.enabled != 1) || in audit_receive_msg()
1467 t = s.enabled | (-s.log_passwd & AUDIT_TTY_LOG_PASSWD); in audit_receive_msg()
1470 old.enabled = t & AUDIT_TTY_ENABLE; in audit_receive_msg()
1477 old.enabled, s.enabled, old.log_passwd, in audit_receive_msg()
/kernel/livepatch/
Dcore.c316 bool enabled; in enabled_store() local
318 ret = kstrtobool(buf, &enabled); in enabled_store()
326 if (patch->enabled == enabled) { in enabled_store()
341 else if (!enabled) in enabled_store()
360 return snprintf(buf, PAGE_SIZE-1, "%d\n", patch->enabled); in enabled_show()
402 static struct kobj_attribute enabled_kobj_attr = __ATTR_RW(enabled);
824 patch->enabled = false; in klp_init_patch_early()
876 if (WARN_ON(!patch->enabled)) in __klp_disable_patch()
898 patch->enabled = false; in __klp_disable_patch()
912 if (WARN_ON(patch->enabled)) in __klp_enable_patch()
[all …]
Dtransition.c453 if (!patch->enabled) { in klp_try_complete_transition()
588 klp_transition_patch->enabled = !klp_transition_patch->enabled; in klp_reverse_transition()
/kernel/irq/
Dirq_sim.c18 irq_ctx->enabled = false; in irq_sim_irqmask()
25 irq_ctx->enabled = true; in irq_sim_irqunmask()
98 sim->irqs[i].enabled = false; in irq_sim_init()
176 if (sim->irqs[offset].enabled) { in irq_sim_fire()
Dchip.c945 bool enabled = cpumask_test_cpu(cpu, desc->percpu_enabled); in handle_percpu_devid_irq() local
947 if (enabled) in handle_percpu_devid_irq()
951 enabled ? " and unmasked" : "", irq, cpu); in handle_percpu_devid_irq()
/kernel/trace/
DKconfig89 Adds a very slight overhead to tracing when enabled.
101 # enabled by all tracers (context switch and event tracer) they select TRACING.
154 tracing is enabled by the administrator. If it's runtime disabled
176 and last enabled.
209 enabled. This option and the preempt-off timing option can be
234 enabled. This option and the irqs-off timing option can be
251 This tracer, when enabled will create one or more kernel threads,
260 is enabled:
323 When this is enabled, this adds a little more overhead to the
326 of the overall system. This is enabled by default when the preempt
[all …]
Dtrace_printk.c190 void trace_printk_control(bool enabled) in trace_printk_control() argument
192 trace_printk_enabled = enabled; in trace_printk_control()
Dtrace_kprobe.c356 bool enabled; in enable_trace_kprobe() local
362 enabled = trace_probe_is_enabled(tp); in enable_trace_kprobe()
372 if (enabled) in enable_trace_kprobe()
382 enabled = true; in enable_trace_kprobe()
387 if (enabled) in enable_trace_kprobe()
Dbpf_trace.c348 u64 *value, u64 *enabled, u64 *running) in get_map_perf_counter() argument
366 return perf_event_read_local(ee->event, value, enabled, running); in get_map_perf_counter()
399 err = get_map_perf_counter(map, flags, &buf->counter, &buf->enabled, in BPF_CALL_4()
902 err = perf_event_read_local(ctx->event, &buf->counter, &buf->enabled, in BPF_CALL_3()
Dtrace.c3086 static void trace_printk_start_stop_comm(int enabled) in trace_printk_start_stop_comm() argument
3091 if (enabled) in trace_printk_start_stop_comm()
4591 if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set) in trace_keep_overwrite()
4597 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled) in set_tracer_flag() argument
4604 if (!!(tr->trace_flags & mask) == !!enabled) in set_tracer_flag()
4609 if (tr->current_trace->flag_changed(tr, mask, !!enabled)) in set_tracer_flag()
4612 if (enabled) in set_tracer_flag()
4618 trace_event_enable_cmd_record(enabled); in set_tracer_flag()
4630 trace_event_enable_tgid_record(enabled); in set_tracer_flag()
4634 trace_event_follow_fork(tr, enabled); in set_tracer_flag()
[all …]
Dtrace.h500 int enabled; member
1872 void trace_printk_control(bool enabled);
1876 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled);
Dtrace_uprobe.c1096 bool enabled; in probe_event_enable() local
1102 enabled = trace_probe_is_enabled(tp); in probe_event_enable()
1122 if (enabled) in probe_event_enable()
/kernel/sched/
Dautogroup.h30 int enabled = READ_ONCE(sysctl_sched_autogroup_enabled); in autogroup_task_group() local
32 if (enabled && task_wants_autogroup(p, tg)) in autogroup_task_group()
Dsched.h1563 #define SCHED_FEAT(name, enabled) \ argument
1581 #define SCHED_FEAT(name, enabled) \ argument
1584 return static_key_##enabled(key); \
1600 #define SCHED_FEAT(name, enabled) \ argument
1601 (1UL << __SCHED_FEAT_##name) * enabled |
Ddebug.c50 #define SCHED_FEAT(name, enabled) \ argument
78 #define SCHED_FEAT(name, enabled) \ argument
79 jump_label_key__##enabled ,
Dcore.c48 #define SCHED_FEAT(name, enabled) \ argument
49 (1UL << __SCHED_FEAT_##name) * enabled |
2729 void set_numabalancing_state(bool enabled) in set_numabalancing_state() argument
2731 if (enabled) in set_numabalancing_state()
2765 static void set_schedstats(bool enabled) in set_schedstats() argument
2767 if (enabled) in set_schedstats()
/kernel/power/
Dsuspend_test.c88 alm.enabled = true; in test_wakealarm()
124 alm.enabled = false; in test_wakealarm()
/kernel/events/
Dcore.c620 __perf_update_times(struct perf_event *event, u64 now, u64 *enabled, u64 *running) in __perf_update_times() argument
625 *enabled = event->total_time_enabled; in __perf_update_times()
627 *enabled += delta; in __perf_update_times()
3891 int enabled = 0; in perf_event_enable_on_exec() local
3902 enabled |= event_enable_on_exec(event, ctx); in perf_event_enable_on_exec()
3909 if (enabled) { in perf_event_enable_on_exec()
4021 u64 *enabled, u64 *running) in perf_event_read_local() argument
4070 if (enabled || running) { in perf_event_read_local()
4075 if (enabled) in perf_event_read_local()
4076 *enabled = __enabled; in perf_event_read_local()
[all …]
/kernel/rcu/
Drcu_segcblist.c165 rsclp->enabled = 1; in rcu_segcblist_init()
177 rsclp->enabled = 0; in rcu_segcblist_disable()
Drcu_segcblist.h80 return rsclp->enabled; in rcu_segcblist_is_enabled()
DKconfig42 This option needs to be enabled if you wish to make
/kernel/cgroup/
Dcgroup-v1.c991 u16 enabled = 0; in check_cgroupfs_options() local
1000 enabled |= 1 << i; in check_cgroupfs_options()
1002 ctx->subsys_mask &= enabled; in check_cgroupfs_options()
1016 ctx->subsys_mask = enabled; in check_cgroupfs_options()
/kernel/dma/
DKconfig105 default, but it can be enabled by passing cma=size[MG] to the kernel.
117 enabled by passing cma=size[MG] to the kernel.

12