/kernel/sched/ |
D | wait.c | 21 unsigned long flags; in add_wait_queue() local 23 wq_entry->flags &= ~WQ_FLAG_EXCLUSIVE; in add_wait_queue() 24 spin_lock_irqsave(&wq_head->lock, flags); in add_wait_queue() 26 spin_unlock_irqrestore(&wq_head->lock, flags); in add_wait_queue() 32 unsigned long flags; in add_wait_queue_exclusive() local 34 wq_entry->flags |= WQ_FLAG_EXCLUSIVE; in add_wait_queue_exclusive() 35 spin_lock_irqsave(&wq_head->lock, flags); in add_wait_queue_exclusive() 37 spin_unlock_irqrestore(&wq_head->lock, flags); in add_wait_queue_exclusive() 43 unsigned long flags; in remove_wait_queue() local 45 spin_lock_irqsave(&wq_head->lock, flags); in remove_wait_queue() [all …]
|
D | isolation.c | 17 bool housekeeping_enabled(enum hk_flags flags) in housekeeping_enabled() argument 19 return !!(housekeeping_flags & flags); in housekeeping_enabled() 23 int housekeeping_any_cpu(enum hk_flags flags) in housekeeping_any_cpu() argument 28 if (housekeeping_flags & flags) { in housekeeping_any_cpu() 40 const struct cpumask *housekeeping_cpumask(enum hk_flags flags) in housekeeping_cpumask() argument 43 if (housekeeping_flags & flags) in housekeeping_cpumask() 49 void housekeeping_affine(struct task_struct *t, enum hk_flags flags) in housekeeping_affine() argument 52 if (housekeeping_flags & flags) in housekeeping_affine() 57 bool housekeeping_test_cpu(int cpu, enum hk_flags flags) in housekeeping_test_cpu() argument 60 if (housekeeping_flags & flags) in housekeeping_test_cpu() [all …]
|
D | membarrier.c | 151 if (p->flags & PF_KTHREAD) in membarrier_global_expedited() 176 static int membarrier_private_expedited(int flags, int cpu_id) in membarrier_private_expedited() argument 182 if (flags == MEMBARRIER_FLAG_SYNC_CORE) { in membarrier_private_expedited() 189 } else if (flags == MEMBARRIER_FLAG_RSEQ) { in membarrier_private_expedited() 197 WARN_ON_ONCE(flags); in membarrier_private_expedited() 203 if (flags != MEMBARRIER_FLAG_SYNC_CORE && in membarrier_private_expedited() 269 if (flags != MEMBARRIER_FLAG_SYNC_CORE) { in membarrier_private_expedited() 372 static int membarrier_register_private_expedited(int flags) in membarrier_register_private_expedited() argument 380 if (flags == MEMBARRIER_FLAG_SYNC_CORE) { in membarrier_register_private_expedited() 385 } else if (flags == MEMBARRIER_FLAG_RSEQ) { in membarrier_register_private_expedited() [all …]
|
/kernel/ |
D | nsproxy.c | 67 static struct nsproxy *create_new_namespaces(unsigned long flags, in create_new_namespaces() argument 78 new_nsp->mnt_ns = copy_mnt_ns(flags, tsk->nsproxy->mnt_ns, user_ns, new_fs); in create_new_namespaces() 84 new_nsp->uts_ns = copy_utsname(flags, user_ns, tsk->nsproxy->uts_ns); in create_new_namespaces() 90 new_nsp->ipc_ns = copy_ipcs(flags, user_ns, tsk->nsproxy->ipc_ns); in create_new_namespaces() 97 copy_pid_ns(flags, user_ns, tsk->nsproxy->pid_ns_for_children); in create_new_namespaces() 103 new_nsp->cgroup_ns = copy_cgroup_ns(flags, user_ns, in create_new_namespaces() 110 new_nsp->net_ns = copy_net_ns(flags, user_ns, tsk->nsproxy->net_ns); in create_new_namespaces() 116 new_nsp->time_ns_for_children = copy_time_ns(flags, user_ns, in create_new_namespaces() 151 int copy_namespaces(unsigned long flags, struct task_struct *tsk) in copy_namespaces() argument 158 if (likely(!(flags & (CLONE_NEWNS | CLONE_NEWUTS | CLONE_NEWIPC | in copy_namespaces() [all …]
|
D | freezer.c | 43 if (p->flags & (PF_NOFREEZE | PF_SUSPEND_TASK)) in freezing_slow_path() 52 if (pm_freezing && !(p->flags & PF_KTHREAD)) in freezing_slow_path() 73 current->flags |= PF_FROZEN; in __refrigerator() 76 current->flags &= ~PF_FROZEN; in __refrigerator() 80 if (!(current->flags & PF_FROZEN)) in __refrigerator() 101 unsigned long flags; in fake_signal_wake_up() local 103 if (lock_task_sighand(p, &flags)) { in fake_signal_wake_up() 105 unlock_task_sighand(p, &flags); in fake_signal_wake_up() 122 unsigned long flags; in freeze_task() local 136 spin_lock_irqsave(&freezer_lock, flags); in freeze_task() [all …]
|
D | up.c | 15 unsigned long flags; in smp_call_function_single() local 20 local_irq_save(flags); in smp_call_function_single() 22 local_irq_restore(flags); in smp_call_function_single() 30 unsigned long flags; in smp_call_function_single_async() local 32 local_irq_save(flags); in smp_call_function_single_async() 34 local_irq_restore(flags); in smp_call_function_single_async() 41 unsigned long flags; in on_each_cpu() local 43 local_irq_save(flags); in on_each_cpu() 45 local_irq_restore(flags); in on_each_cpu() 58 unsigned long flags; in on_each_cpu_mask() local [all …]
|
D | kthread.c | 53 unsigned long flags; member 83 WARN_ON(!(k->flags & PF_KTHREAD)); in to_kthread() 101 if (kthread && !(p->flags & PF_KTHREAD)) in __to_kthread() 130 return test_bit(KTHREAD_SHOULD_STOP, &to_kthread(current)->flags); in kthread_should_stop() 136 return test_bit(KTHREAD_SHOULD_PARK, &to_kthread(k)->flags); in __kthread_should_park() 243 if (!test_bit(KTHREAD_SHOULD_PARK, &self->flags)) in __kthread_parkme() 310 if (!test_bit(KTHREAD_SHOULD_STOP, &self->flags)) { in kthread() 456 unsigned long flags; in __kthread_bind_mask() local 464 raw_spin_lock_irqsave(&p->pi_lock, flags); in __kthread_bind_mask() 466 p->flags |= PF_NO_SETAFFINITY; in __kthread_bind_mask() [all …]
|
D | kexec.c | 42 unsigned long flags) in kimage_alloc_init() argument 46 bool kexec_on_panic = flags & KEXEC_ON_CRASH; in kimage_alloc_init() 107 struct kexec_segment __user *segments, unsigned long flags) in do_kexec_load() argument 121 if (flags & KEXEC_ON_CRASH) { in do_kexec_load() 135 if (flags & KEXEC_ON_CRASH) { in do_kexec_load() 144 ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags); in do_kexec_load() 148 if (flags & KEXEC_PRESERVE_CONTEXT) in do_kexec_load() 179 if ((flags & KEXEC_ON_CRASH) && kexec_crash_image) in do_kexec_load() 210 unsigned long flags) in kexec_load_check() argument 235 if ((flags & KEXEC_FLAGS) != (flags & ~KEXEC_ARCH_MASK)) in kexec_load_check() [all …]
|
D | iomem.c | 24 unsigned long flags) in arch_memremap_can_ram_remap() argument 31 unsigned long flags) in try_ram_remap() argument 37 arch_memremap_can_ram_remap(offset, size, flags)) in try_ram_remap() 71 void *memremap(resource_size_t offset, size_t size, unsigned long flags) in memremap() argument 77 if (!flags) in memremap() 87 if (flags & MEMREMAP_WB) { in memremap() 95 addr = try_ram_remap(offset, size, flags); in memremap() 106 if (!addr && is_ram == REGION_INTERSECTS && flags != MEMREMAP_WB) { in memremap() 112 if (!addr && (flags & MEMREMAP_WT)) in memremap() 115 if (!addr && (flags & MEMREMAP_WC)) in memremap() [all …]
|
/kernel/irq/ |
D | manage.c | 45 unsigned long flags; in __synchronize_hardirq() local 55 raw_spin_lock_irqsave(&desc->lock, flags); in __synchronize_hardirq() 71 raw_spin_unlock_irqrestore(&desc->lock, flags); in __synchronize_hardirq() 393 unsigned long flags; in __irq_set_affinity() local 399 raw_spin_lock_irqsave(&desc->lock, flags); in __irq_set_affinity() 401 raw_spin_unlock_irqrestore(&desc->lock, flags); in __irq_set_affinity() 407 unsigned long flags; in irq_set_affinity_hint() local 408 struct irq_desc *desc = irq_get_desc_lock(irq, &flags, IRQ_GET_DESC_CHECK_GLOBAL); in irq_set_affinity_hint() 413 irq_put_desc_unlock(desc, flags); in irq_set_affinity_hint() 427 unsigned long flags; in irq_affinity_notify() local [all …]
|
D | pm.c | 37 if (action->flags & IRQF_FORCE_RESUME) in irq_pm_install_action() 43 if (action->flags & IRQF_NO_SUSPEND) in irq_pm_install_action() 45 else if (action->flags & IRQF_COND_SUSPEND) in irq_pm_install_action() 61 if (action->flags & IRQF_FORCE_RESUME) in irq_pm_remove_action() 64 if (action->flags & IRQF_NO_SUSPEND) in irq_pm_remove_action() 66 else if (action->flags & IRQF_COND_SUSPEND) in irq_pm_remove_action() 72 unsigned long chipflags = irq_desc_get_chip(desc)->flags; in suspend_device_irq() 137 unsigned long flags; in suspend_device_irqs() local 142 raw_spin_lock_irqsave(&desc->lock, flags); in suspend_device_irqs() 144 raw_spin_unlock_irqrestore(&desc->lock, flags); in suspend_device_irqs() [all …]
|
/kernel/trace/ |
D | trace_functions_graph.c | 95 struct trace_seq *s, u32 flags); 99 unsigned long flags, in __trace_graph_entry() argument 108 sizeof(*entry), flags, pc); in __trace_graph_entry() 131 unsigned long flags; in trace_graph_entry() local 172 local_irq_save(flags); in trace_graph_entry() 178 ret = __trace_graph_entry(tr, trace, flags, pc); in trace_graph_entry() 184 local_irq_restore(flags); in trace_graph_entry() 191 unsigned long ip, unsigned long flags, int pc) in __trace_graph_function() argument 205 __trace_graph_entry(tr, &ent, flags, pc); in __trace_graph_function() 206 __trace_graph_return(tr, &ret, flags, pc); in __trace_graph_function() [all …]
|
D | trace_irqsoff.c | 101 unsigned long *flags) in func_prolog_dec() argument 116 local_save_flags(*flags); in func_prolog_dec() 122 if (!irqs_disabled_flags(*flags) && !preempt_count()) in func_prolog_dec() 145 unsigned long flags; in irqsoff_tracer_call() local 147 if (!func_prolog_dec(tr, &data, &flags)) in irqsoff_tracer_call() 150 trace_function(tr, ip, parent_ip, flags, preempt_count()); in irqsoff_tracer_call() 179 unsigned long flags; in irqsoff_graph_entry() local 195 if (!func_prolog_dec(tr, &data, &flags)) in irqsoff_graph_entry() 199 ret = __trace_graph_entry(tr, trace, flags, pc); in irqsoff_graph_entry() 209 unsigned long flags; in irqsoff_graph_return() local [all …]
|
D | trace_sched_wakeup.c | 119 unsigned long flags; in wakeup_graph_entry() local 137 local_save_flags(flags); in wakeup_graph_entry() 138 ret = __trace_graph_entry(tr, trace, flags, pc); in wakeup_graph_entry() 149 unsigned long flags; in wakeup_graph_return() local 157 local_save_flags(flags); in wakeup_graph_return() 158 __trace_graph_return(tr, trace, flags, pc); in wakeup_graph_return() 221 unsigned long flags; in wakeup_tracer_call() local 227 local_irq_save(flags); in wakeup_tracer_call() 228 trace_function(tr, ip, parent_ip, flags, pc); in wakeup_tracer_call() 229 local_irq_restore(flags); in wakeup_tracer_call() [all …]
|
/kernel/cgroup/ |
D | freezer.c | 27 if (!test_bit(CGRP_FROZEN, &cgrp->flags) && in cgroup_propagate_frozen() 28 test_bit(CGRP_FREEZE, &cgrp->flags) && in cgroup_propagate_frozen() 31 set_bit(CGRP_FROZEN, &cgrp->flags); in cgroup_propagate_frozen() 38 if (test_bit(CGRP_FROZEN, &cgrp->flags)) { in cgroup_propagate_frozen() 39 clear_bit(CGRP_FROZEN, &cgrp->flags); in cgroup_propagate_frozen() 63 frozen = test_bit(CGRP_FREEZE, &cgrp->flags) && in cgroup_update_frozen() 68 if (test_bit(CGRP_FROZEN, &cgrp->flags)) in cgroup_update_frozen() 71 set_bit(CGRP_FROZEN, &cgrp->flags); in cgroup_update_frozen() 74 if (!test_bit(CGRP_FROZEN, &cgrp->flags)) in cgroup_update_frozen() 77 clear_bit(CGRP_FROZEN, &cgrp->flags); in cgroup_update_frozen() [all …]
|
/kernel/time/ |
D | clocksource.c | 108 static inline void clocksource_watchdog_lock(unsigned long *flags) in clocksource_watchdog_lock() argument 110 spin_lock_irqsave(&watchdog_lock, *flags); in clocksource_watchdog_lock() 113 static inline void clocksource_watchdog_unlock(unsigned long *flags) in clocksource_watchdog_unlock() argument 115 spin_unlock_irqrestore(&watchdog_lock, *flags); in clocksource_watchdog_unlock() 154 cs->flags &= ~(CLOCK_SOURCE_VALID_FOR_HRES | CLOCK_SOURCE_WATCHDOG); in __clocksource_unstable() 155 cs->flags |= CLOCK_SOURCE_UNSTABLE; in __clocksource_unstable() 183 unsigned long flags; in clocksource_mark_unstable() local 185 spin_lock_irqsave(&watchdog_lock, flags); in clocksource_mark_unstable() 186 if (!(cs->flags & CLOCK_SOURCE_UNSTABLE)) { in clocksource_mark_unstable() 191 spin_unlock_irqrestore(&watchdog_lock, flags); in clocksource_mark_unstable() [all …]
|
D | posix-timers.c | 105 static struct k_itimer *__lock_timer(timer_t timer_id, unsigned long *flags); 107 #define lock_timer(tid, flags) \ argument 109 __cond_lock(&__timr->it_lock, __timr = __lock_timer(tid, flags)); \ 164 static inline void unlock_timer(struct k_itimer *timr, unsigned long flags) in unlock_timer() argument 166 spin_unlock_irqrestore(&timr->it_lock, flags); in unlock_timer() 316 unsigned long flags; in posixtimer_rearm() local 318 timr = lock_timer(info->si_tid, &flags); in posixtimer_rearm() 333 unlock_timer(timr, flags); in posixtimer_rearm() 369 unsigned long flags; in posix_timer_fn() local 374 spin_lock_irqsave(&timr->it_lock, flags); in posix_timer_fn() [all …]
|
D | alarmtimer.c | 75 unsigned long flags; in alarmtimer_get_rtcdev() local 78 spin_lock_irqsave(&rtcdev_lock, flags); in alarmtimer_get_rtcdev() 80 spin_unlock_irqrestore(&rtcdev_lock, flags); in alarmtimer_get_rtcdev() 89 unsigned long flags; in alarmtimer_rtc_add_device() local 107 spin_lock_irqsave(&rtcdev_lock, flags); in alarmtimer_rtc_add_device() 122 spin_unlock_irqrestore(&rtcdev_lock, flags); in alarmtimer_rtc_add_device() 203 unsigned long flags; in alarmtimer_fired() local 207 spin_lock_irqsave(&base->lock, flags); in alarmtimer_fired() 209 spin_unlock_irqrestore(&base->lock, flags); in alarmtimer_fired() 214 spin_lock_irqsave(&base->lock, flags); in alarmtimer_fired() [all …]
|
/kernel/locking/ |
D | semaphore.c | 55 unsigned long flags; in down() local 57 raw_spin_lock_irqsave(&sem->lock, flags); in down() 62 raw_spin_unlock_irqrestore(&sem->lock, flags); in down() 77 unsigned long flags; in down_interruptible() local 80 raw_spin_lock_irqsave(&sem->lock, flags); in down_interruptible() 85 raw_spin_unlock_irqrestore(&sem->lock, flags); in down_interruptible() 103 unsigned long flags; in down_killable() local 106 raw_spin_lock_irqsave(&sem->lock, flags); in down_killable() 111 raw_spin_unlock_irqrestore(&sem->lock, flags); in down_killable() 132 unsigned long flags; in down_trylock() local [all …]
|
D | spinlock.c | 82 unsigned long flags; \ 86 local_irq_save(flags); \ 89 local_irq_restore(flags); \ 95 return flags; \ 105 unsigned long flags; \ 112 flags = _raw_##op##_lock_irqsave(lock); \ 114 local_irq_restore(flags); \ 189 void __lockfunc _raw_spin_unlock_irqrestore(raw_spinlock_t *lock, unsigned long flags) in _raw_spin_unlock_irqrestore() argument 191 __raw_spin_unlock_irqrestore(lock, flags); in _raw_spin_unlock_irqrestore() 261 void __lockfunc _raw_read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) in _raw_read_unlock_irqrestore() argument [all …]
|
/kernel/power/ |
D | wakeup_reason.c | 153 unsigned long flags; in log_irq_wakeup_reason() local 155 spin_lock_irqsave(&wakeup_reason_lock, flags); in log_irq_wakeup_reason() 157 spin_unlock_irqrestore(&wakeup_reason_lock, flags); in log_irq_wakeup_reason() 162 spin_unlock_irqrestore(&wakeup_reason_lock, flags); in log_irq_wakeup_reason() 170 spin_unlock_irqrestore(&wakeup_reason_lock, flags); in log_irq_wakeup_reason() 176 unsigned long flags; in log_threaded_irq_wakeup_reason() local 188 spin_lock_irqsave(&wakeup_reason_lock, flags); in log_threaded_irq_wakeup_reason() 191 spin_unlock_irqrestore(&wakeup_reason_lock, flags); in log_threaded_irq_wakeup_reason() 196 spin_unlock_irqrestore(&wakeup_reason_lock, flags); in log_threaded_irq_wakeup_reason() 212 spin_unlock_irqrestore(&wakeup_reason_lock, flags); in log_threaded_irq_wakeup_reason() [all …]
|
/kernel/rcu/ |
D | tree_exp.h | 76 unsigned long flags; in sync_exp_reset_tree_hotplug() local 93 raw_spin_lock_irqsave_rcu_node(rnp, flags); in sync_exp_reset_tree_hotplug() 95 raw_spin_unlock_irqrestore_rcu_node(rnp, flags); in sync_exp_reset_tree_hotplug() 102 raw_spin_unlock_irqrestore_rcu_node(rnp, flags); in sync_exp_reset_tree_hotplug() 113 raw_spin_lock_irqsave_rcu_node(rnp_up, flags); in sync_exp_reset_tree_hotplug() 117 raw_spin_unlock_irqrestore_rcu_node(rnp_up, flags); in sync_exp_reset_tree_hotplug() 132 unsigned long flags; in sync_exp_reset_tree() local 137 raw_spin_lock_irqsave_rcu_node(rnp, flags); in sync_exp_reset_tree() 140 raw_spin_unlock_irqrestore_rcu_node(rnp, flags); in sync_exp_reset_tree() 163 unsigned long flags; in sync_rcu_exp_done_unlocked() local [all …]
|
D | tree_plugin.h | 439 rcu_preempt_deferred_qs_irqrestore(struct task_struct *t, unsigned long flags) in rcu_preempt_deferred_qs_irqrestore() argument 458 local_irq_restore(flags); in rcu_preempt_deferred_qs_irqrestore() 529 rcu_report_unblock_qs_rnp(rnp, flags); in rcu_preempt_deferred_qs_irqrestore() 531 raw_spin_unlock_irqrestore_rcu_node(rnp, flags); in rcu_preempt_deferred_qs_irqrestore() 546 local_irq_restore(flags); in rcu_preempt_deferred_qs_irqrestore() 575 unsigned long flags; in rcu_preempt_deferred_qs() local 579 local_irq_save(flags); in rcu_preempt_deferred_qs() 580 rcu_preempt_deferred_qs_irqrestore(t, flags); in rcu_preempt_deferred_qs() 601 unsigned long flags; in rcu_read_unlock_special() local 610 local_irq_save(flags); in rcu_read_unlock_special() [all …]
|
/kernel/printk/ |
D | printk.c | 234 unsigned long flags; in __down_trylock_console_sem() local 241 printk_safe_enter_irqsave(flags); in __down_trylock_console_sem() 243 printk_safe_exit_irqrestore(flags); in __down_trylock_console_sem() 254 unsigned long flags; in __up_console_sem() local 258 printk_safe_enter_irqsave(flags); in __up_console_sem() 260 printk_safe_exit_irqrestore(flags); in __up_console_sem() 386 #define logbuf_lock_irqsave(flags) \ argument 388 printk_safe_enter_irqsave(flags); \ 392 #define logbuf_unlock_irqrestore(flags) \ argument 395 printk_safe_exit_irqrestore(flags); \ [all …]
|
/kernel/dma/ |
D | debug.c | 179 unsigned long flags; in driver_filter() local 203 read_lock_irqsave(&driver_name_lock, flags); in driver_filter() 212 read_unlock_irqrestore(&driver_name_lock, flags); in driver_filter() 249 unsigned long *flags) in get_hash_bucket() argument 256 *flags = __flags; in get_hash_bucket() 264 unsigned long flags) in put_hash_bucket() argument 267 spin_unlock_irqrestore(&bucket->lock, flags); in put_hash_bucket() 350 unsigned long *flags) in bucket_find_contain() argument 366 put_hash_bucket(*bucket, *flags); in bucket_find_contain() 369 *bucket = get_hash_bucket(&index, flags); in bucket_find_contain() [all …]
|