/kernel/ |
D | jump_label.c | 110 int n = atomic_read(&key->enabled); in static_key_count() 135 for (v = atomic_read(&key->enabled); v > 0; v = v1) { in static_key_slow_inc_cpuslocked() 136 v1 = atomic_cmpxchg(&key->enabled, v, v + 1); in static_key_slow_inc_cpuslocked() 142 if (atomic_read(&key->enabled) == 0) { in static_key_slow_inc_cpuslocked() 143 atomic_set(&key->enabled, -1); in static_key_slow_inc_cpuslocked() 149 atomic_set_release(&key->enabled, 1); in static_key_slow_inc_cpuslocked() 151 atomic_inc(&key->enabled); in static_key_slow_inc_cpuslocked() 169 if (atomic_read(&key->enabled) > 0) { in static_key_enable_cpuslocked() 170 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable_cpuslocked() 175 if (atomic_read(&key->enabled) == 0) { in static_key_enable_cpuslocked() [all …]
|
D | stop_machine.c | 41 bool enabled; /* is this stopper enabled? */ member 82 bool enabled; in cpu_stop_queue_work() local 86 enabled = stopper->enabled; in cpu_stop_queue_work() 87 if (enabled) in cpu_stop_queue_work() 96 return enabled; in cpu_stop_queue_work() 264 if (!stopper1->enabled || !stopper2->enabled) { in cpu_stop_queue_two_works() 541 stopper->enabled = false; in stop_machine_park() 563 stopper->enabled = true; in stop_machine_unpark()
|
D | audit.c | 1195 s.enabled = audit_enabled; in audit_receive_msg() 1215 err = audit_set_enabled(s.enabled); in audit_receive_msg() 1445 s.enabled = t & AUDIT_TTY_ENABLE; in audit_receive_msg() 1460 if ((s.enabled != 0 && s.enabled != 1) || in audit_receive_msg() 1467 t = s.enabled | (-s.log_passwd & AUDIT_TTY_LOG_PASSWD); in audit_receive_msg() 1470 old.enabled = t & AUDIT_TTY_ENABLE; in audit_receive_msg() 1477 old.enabled, s.enabled, old.log_passwd, in audit_receive_msg()
|
/kernel/livepatch/ |
D | core.c | 316 bool enabled; in enabled_store() local 318 ret = kstrtobool(buf, &enabled); in enabled_store() 326 if (patch->enabled == enabled) { in enabled_store() 341 else if (!enabled) in enabled_store() 360 return snprintf(buf, PAGE_SIZE-1, "%d\n", patch->enabled); in enabled_show() 402 static struct kobj_attribute enabled_kobj_attr = __ATTR_RW(enabled); 824 patch->enabled = false; in klp_init_patch_early() 876 if (WARN_ON(!patch->enabled)) in __klp_disable_patch() 898 patch->enabled = false; in __klp_disable_patch() 912 if (WARN_ON(patch->enabled)) in __klp_enable_patch() [all …]
|
D | transition.c | 453 if (!patch->enabled) { in klp_try_complete_transition() 588 klp_transition_patch->enabled = !klp_transition_patch->enabled; in klp_reverse_transition()
|
/kernel/irq/ |
D | irq_sim.c | 18 irq_ctx->enabled = false; in irq_sim_irqmask() 25 irq_ctx->enabled = true; in irq_sim_irqunmask() 98 sim->irqs[i].enabled = false; in irq_sim_init() 176 if (sim->irqs[offset].enabled) { in irq_sim_fire()
|
D | chip.c | 945 bool enabled = cpumask_test_cpu(cpu, desc->percpu_enabled); in handle_percpu_devid_irq() local 947 if (enabled) in handle_percpu_devid_irq() 951 enabled ? " and unmasked" : "", irq, cpu); in handle_percpu_devid_irq()
|
/kernel/trace/ |
D | Kconfig | 89 Adds a very slight overhead to tracing when enabled. 101 # enabled by all tracers (context switch and event tracer) they select TRACING. 154 tracing is enabled by the administrator. If it's runtime disabled 176 and last enabled. 209 enabled. This option and the preempt-off timing option can be 234 enabled. This option and the irqs-off timing option can be 251 This tracer, when enabled will create one or more kernel threads, 260 is enabled: 323 When this is enabled, this adds a little more overhead to the 326 of the overall system. This is enabled by default when the preempt [all …]
|
D | trace_printk.c | 190 void trace_printk_control(bool enabled) in trace_printk_control() argument 192 trace_printk_enabled = enabled; in trace_printk_control()
|
D | trace_kprobe.c | 356 bool enabled; in enable_trace_kprobe() local 362 enabled = trace_probe_is_enabled(tp); in enable_trace_kprobe() 372 if (enabled) in enable_trace_kprobe() 382 enabled = true; in enable_trace_kprobe() 387 if (enabled) in enable_trace_kprobe()
|
D | bpf_trace.c | 348 u64 *value, u64 *enabled, u64 *running) in get_map_perf_counter() argument 366 return perf_event_read_local(ee->event, value, enabled, running); in get_map_perf_counter() 399 err = get_map_perf_counter(map, flags, &buf->counter, &buf->enabled, in BPF_CALL_4() 902 err = perf_event_read_local(ctx->event, &buf->counter, &buf->enabled, in BPF_CALL_3()
|
D | trace.c | 3086 static void trace_printk_start_stop_comm(int enabled) in trace_printk_start_stop_comm() argument 3091 if (enabled) in trace_printk_start_stop_comm() 4591 if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set) in trace_keep_overwrite() 4597 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled) in set_tracer_flag() argument 4604 if (!!(tr->trace_flags & mask) == !!enabled) in set_tracer_flag() 4609 if (tr->current_trace->flag_changed(tr, mask, !!enabled)) in set_tracer_flag() 4612 if (enabled) in set_tracer_flag() 4618 trace_event_enable_cmd_record(enabled); in set_tracer_flag() 4630 trace_event_enable_tgid_record(enabled); in set_tracer_flag() 4634 trace_event_follow_fork(tr, enabled); in set_tracer_flag() [all …]
|
D | trace.h | 500 int enabled; member 1872 void trace_printk_control(bool enabled); 1876 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled);
|
D | trace_uprobe.c | 1096 bool enabled; in probe_event_enable() local 1102 enabled = trace_probe_is_enabled(tp); in probe_event_enable() 1122 if (enabled) in probe_event_enable()
|
/kernel/sched/ |
D | autogroup.h | 30 int enabled = READ_ONCE(sysctl_sched_autogroup_enabled); in autogroup_task_group() local 32 if (enabled && task_wants_autogroup(p, tg)) in autogroup_task_group()
|
D | sched.h | 1563 #define SCHED_FEAT(name, enabled) \ argument 1581 #define SCHED_FEAT(name, enabled) \ argument 1584 return static_key_##enabled(key); \ 1600 #define SCHED_FEAT(name, enabled) \ argument 1601 (1UL << __SCHED_FEAT_##name) * enabled |
|
D | debug.c | 50 #define SCHED_FEAT(name, enabled) \ argument 78 #define SCHED_FEAT(name, enabled) \ argument 79 jump_label_key__##enabled ,
|
D | core.c | 48 #define SCHED_FEAT(name, enabled) \ argument 49 (1UL << __SCHED_FEAT_##name) * enabled | 2729 void set_numabalancing_state(bool enabled) in set_numabalancing_state() argument 2731 if (enabled) in set_numabalancing_state() 2765 static void set_schedstats(bool enabled) in set_schedstats() argument 2767 if (enabled) in set_schedstats()
|
/kernel/power/ |
D | suspend_test.c | 88 alm.enabled = true; in test_wakealarm() 124 alm.enabled = false; in test_wakealarm()
|
/kernel/events/ |
D | core.c | 620 __perf_update_times(struct perf_event *event, u64 now, u64 *enabled, u64 *running) in __perf_update_times() argument 625 *enabled = event->total_time_enabled; in __perf_update_times() 627 *enabled += delta; in __perf_update_times() 3891 int enabled = 0; in perf_event_enable_on_exec() local 3902 enabled |= event_enable_on_exec(event, ctx); in perf_event_enable_on_exec() 3909 if (enabled) { in perf_event_enable_on_exec() 4021 u64 *enabled, u64 *running) in perf_event_read_local() argument 4070 if (enabled || running) { in perf_event_read_local() 4075 if (enabled) in perf_event_read_local() 4076 *enabled = __enabled; in perf_event_read_local() [all …]
|
/kernel/rcu/ |
D | rcu_segcblist.c | 165 rsclp->enabled = 1; in rcu_segcblist_init() 177 rsclp->enabled = 0; in rcu_segcblist_disable()
|
D | rcu_segcblist.h | 80 return rsclp->enabled; in rcu_segcblist_is_enabled()
|
D | Kconfig | 42 This option needs to be enabled if you wish to make
|
/kernel/cgroup/ |
D | cgroup-v1.c | 991 u16 enabled = 0; in check_cgroupfs_options() local 1000 enabled |= 1 << i; in check_cgroupfs_options() 1002 ctx->subsys_mask &= enabled; in check_cgroupfs_options() 1016 ctx->subsys_mask = enabled; in check_cgroupfs_options()
|
/kernel/dma/ |
D | Kconfig | 105 default, but it can be enabled by passing cma=size[MG] to the kernel. 117 enabled by passing cma=size[MG] to the kernel.
|