/kernel/ |
D | jump_label.c | 110 int n = atomic_read(&key->enabled); in static_key_count() 135 for (v = atomic_read(&key->enabled); v > 0; v = v1) { in static_key_slow_inc_cpuslocked() 136 v1 = atomic_cmpxchg(&key->enabled, v, v + 1); in static_key_slow_inc_cpuslocked() 142 if (atomic_read(&key->enabled) == 0) { in static_key_slow_inc_cpuslocked() 143 atomic_set(&key->enabled, -1); in static_key_slow_inc_cpuslocked() 149 atomic_set_release(&key->enabled, 1); in static_key_slow_inc_cpuslocked() 151 atomic_inc(&key->enabled); in static_key_slow_inc_cpuslocked() 169 if (atomic_read(&key->enabled) > 0) { in static_key_enable_cpuslocked() 170 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable_cpuslocked() 175 if (atomic_read(&key->enabled) == 0) { in static_key_enable_cpuslocked() [all …]
|
D | stop_machine.c | 32 bool enabled; /* is this stopper enabled? */ member 73 bool enabled; in cpu_stop_queue_work() local 77 enabled = stopper->enabled; in cpu_stop_queue_work() 78 if (enabled) in cpu_stop_queue_work() 87 return enabled; in cpu_stop_queue_work() 256 if (!stopper1->enabled || !stopper2->enabled) { in cpu_stop_queue_two_works() 552 stopper->enabled = false; in stop_machine_park() 574 stopper->enabled = true; in stop_machine_unpark()
|
D | audit.c | 1242 s.enabled = audit_enabled; in audit_receive_msg() 1263 err = audit_set_enabled(s.enabled); in audit_receive_msg() 1505 s.enabled = t & AUDIT_TTY_ENABLE; in audit_receive_msg() 1520 if ((s.enabled != 0 && s.enabled != 1) || in audit_receive_msg() 1527 t = s.enabled | (-s.log_passwd & AUDIT_TTY_LOG_PASSWD); in audit_receive_msg() 1530 old.enabled = t & AUDIT_TTY_ENABLE; in audit_receive_msg() 1537 old.enabled, s.enabled, old.log_passwd, in audit_receive_msg()
|
/kernel/irq/ |
D | irq_sim.c | 23 bool enabled; member 35 irq_ctx->enabled = false; in irq_sim_irqmask() 42 irq_ctx->enabled = true; in irq_sim_irqunmask() 64 if (irq_ctx->enabled) in irq_sim_get_irqchip_state() 82 if (irq_ctx->enabled) { in irq_sim_set_irqchip_state()
|
D | chip.c | 952 bool enabled = cpumask_test_cpu(cpu, desc->percpu_enabled); in handle_percpu_devid_irq() local 954 if (enabled) in handle_percpu_devid_irq() 958 enabled ? " and unmasked" : "", irq, cpu); in handle_percpu_devid_irq()
|
/kernel/livepatch/ |
D | core.c | 338 bool enabled; in enabled_store() local 340 ret = kstrtobool(buf, &enabled); in enabled_store() 348 if (patch->enabled == enabled) { in enabled_store() 363 else if (!enabled) in enabled_store() 382 return snprintf(buf, PAGE_SIZE-1, "%d\n", patch->enabled); in enabled_show() 424 static struct kobj_attribute enabled_kobj_attr = __ATTR_RW(enabled); 877 patch->enabled = false; in klp_init_patch_early() 929 if (WARN_ON(!patch->enabled)) in __klp_disable_patch() 951 patch->enabled = false; in __klp_disable_patch() 965 if (WARN_ON(patch->enabled)) in __klp_enable_patch() [all …]
|
D | transition.c | 453 if (!patch->enabled) in klp_try_complete_transition() 588 klp_transition_patch->enabled = !klp_transition_patch->enabled; in klp_reverse_transition()
|
/kernel/sched/ |
D | autogroup.h | 30 int enabled = READ_ONCE(sysctl_sched_autogroup_enabled); in autogroup_task_group() local 32 if (enabled && task_wants_autogroup(p, tg)) in autogroup_task_group()
|
D | sched.h | 1675 #define SCHED_FEAT(name, enabled) \ argument 1694 #define SCHED_FEAT(name, enabled) \ argument 1697 return static_key_##enabled(key); \ 1721 #define SCHED_FEAT(name, enabled) \ argument 1722 (1UL << __SCHED_FEAT_##name) * enabled |
|
D | debug.c | 48 #define SCHED_FEAT(name, enabled) \ argument 77 #define SCHED_FEAT(name, enabled) \ argument 78 jump_label_key__##enabled ,
|
/kernel/trace/ |
D | Kconfig | 87 Adds a very slight overhead to tracing when enabled. 112 # enabled by all tracers (context switch and event tracer) they select TRACING. 175 tracing is enabled by the administrator. If it's runtime disabled 204 image. When this is enabled, functions can be individually 205 enabled, and the functions not enabled will not affect 259 The stack tracer can also be enabled or disabled via the 268 and last enabled. 292 enabled. This option and the preempt-off timing option can be 317 enabled. This option and the irqs-off timing option can be 334 This tracer, when enabled will create one or more kernel threads, [all …]
|
D | trace_printk.c | 190 void trace_printk_control(bool enabled) in trace_printk_control() argument 192 trace_printk_enabled = enabled; in trace_printk_control()
|
D | trace_kprobe.c | 357 bool enabled; in enable_trace_kprobe() local 363 enabled = trace_probe_is_enabled(tp); in enable_trace_kprobe() 373 if (enabled) in enable_trace_kprobe() 383 enabled = true; in enable_trace_kprobe() 388 if (enabled) in enable_trace_kprobe()
|
D | trace.c | 3247 static void trace_printk_start_stop_comm(int enabled) in trace_printk_start_stop_comm() argument 3252 if (enabled) in trace_printk_start_stop_comm() 4929 if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set) in trace_keep_overwrite() 4935 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled) in set_tracer_flag() argument 4944 if (!!(tr->trace_flags & mask) == !!enabled) in set_tracer_flag() 4949 if (tr->current_trace->flag_changed(tr, mask, !!enabled)) in set_tracer_flag() 4952 if (enabled) in set_tracer_flag() 4958 trace_event_enable_cmd_record(enabled); in set_tracer_flag() 4979 trace_event_enable_tgid_record(enabled); in set_tracer_flag() 4983 trace_event_follow_fork(tr, enabled); in set_tracer_flag() [all …]
|
D | bpf_trace.c | 805 u64 *value, u64 *enabled, u64 *running) in get_map_perf_counter() argument 823 return perf_event_read_local(ee->event, value, enabled, running); in get_map_perf_counter() 856 err = get_map_perf_counter(map, flags, &buf->counter, &buf->enabled, in BPF_CALL_4() 1516 err = perf_event_read_local(ctx->event, &buf->counter, &buf->enabled, in BPF_CALL_3()
|
D | trace.h | 551 int enabled; member 1989 void trace_printk_control(bool enabled); 1992 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled);
|
D | trace_uprobe.c | 1099 bool enabled; in probe_event_enable() local 1105 enabled = trace_probe_is_enabled(tp); in probe_event_enable() 1125 if (enabled) in probe_event_enable()
|
D | fgraph.c | 108 # error MCOUNT_INSN_SIZE not defined with direct calls enabled
|
/kernel/power/ |
D | suspend_test.c | 88 alm.enabled = true; in test_wakealarm() 124 alm.enabled = false; in test_wakealarm()
|
/kernel/bpf/preload/ |
D | Kconfig | 10 # The dependency on !COMPILE_TEST prevents it from being enabled
|
/kernel/rcu/ |
D | rcu_segcblist.c | 155 rsclp->enabled = 1; in rcu_segcblist_init() 166 rsclp->enabled = 0; in rcu_segcblist_disable()
|
D | rcu_segcblist.h | 59 return rsclp->enabled; in rcu_segcblist_is_enabled()
|
/kernel/events/ |
D | core.c | 628 __perf_update_times(struct perf_event *event, u64 now, u64 *enabled, u64 *running) in __perf_update_times() argument 633 *enabled = event->total_time_enabled; in __perf_update_times() 635 *enabled += delta; in __perf_update_times() 4229 int enabled = 0; in perf_event_enable_on_exec() local 4240 enabled |= event_enable_on_exec(event, ctx); in perf_event_enable_on_exec() 4247 if (enabled) { in perf_event_enable_on_exec() 4359 u64 *enabled, u64 *running) in perf_event_read_local() argument 4408 if (enabled || running) { in perf_event_read_local() 4413 if (enabled) in perf_event_read_local() 4414 *enabled = __enabled; in perf_event_read_local() [all …]
|
/kernel/cgroup/ |
D | cgroup-v1.c | 1019 u16 enabled = 0; in check_cgroupfs_options() local 1028 enabled |= 1 << i; in check_cgroupfs_options() 1030 ctx->subsys_mask &= enabled; in check_cgroupfs_options() 1044 ctx->subsys_mask = enabled; in check_cgroupfs_options()
|
/kernel/dma/ |
D | Kconfig | 146 default, but it can be enabled by passing cma=size[MG] to the kernel. 158 enabled by passing cma=size[MG] to the kernel.
|