/arch/mips/kernel/ |
D | sync-r4k.c | 75 while (atomic_read(&count_count_start) != nslaves) in synchronise_count_master() 92 while (atomic_read(&count_count_stop) != nslaves) in synchronise_count_master() 133 while (!atomic_read(&count_start_flag)) in synchronise_count_slave() 137 initcount = atomic_read(&count_reference); in synchronise_count_slave() 142 while (atomic_read(&count_count_start) != ncpus) in synchronise_count_slave() 152 while (atomic_read(&count_count_stop) != ncpus) in synchronise_count_slave()
|
D | spinlock_test.c | 66 while (atomic_read(&s->enter_wait)) in multi_other() 73 while (atomic_read(&s->start_wait)) in multi_other() 85 while (atomic_read(&s->exit_wait)) in multi_other()
|
/arch/m68k/include/asm/ |
D | atomic.h | 19 #define atomic_read(v) (*(volatile int *)&(v)->counter) macro 88 : "g" (i), "2" (atomic_read(v))); in atomic_add_return() 102 : "g" (i), "2" (atomic_read(v))); in atomic_sub_return() 117 t = atomic_read(v); in atomic_add_return() 131 t = atomic_read(v); in atomic_sub_return() 145 prev = atomic_read(v); in atomic_cmpxchg() 158 prev = atomic_read(v); in atomic_xchg() 200 c = atomic_read(v); in __atomic_add_unless()
|
/arch/x86/kernel/ |
D | tsc_sync.c | 150 while (atomic_read(&start_count) != cpus-1) in check_tsc_sync_source() 159 while (atomic_read(&stop_count) != cpus-1) in check_tsc_sync_source() 202 while (atomic_read(&start_count) != cpus) in check_tsc_sync_target() 215 while (atomic_read(&stop_count) != cpus) in check_tsc_sync_target()
|
D | ftrace.c | 137 atomic_read(&nmi_update_count)); in ftrace_arch_read_dyn_info() 143 int old = atomic_read(&nmi_running); in clear_mod_flag() 211 if (!atomic_read(&nmi_running)) in wait_for_nmi() 216 } while (atomic_read(&nmi_running)); in wait_for_nmi() 407 if (unlikely(atomic_read(¤t->tracing_graph_pause))) in prepare_ftrace_return()
|
/arch/powerpc/platforms/powermac/ |
D | backlight.c | 102 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_key_worker() 130 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_key() 169 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_set_legacy_worker() 177 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_set_legacy_brightness_pmu()
|
/arch/sh/kernel/ |
D | ftrace.c | 108 atomic_read(&nmi_update_count)); in ftrace_arch_read_dyn_info() 114 int old = atomic_read(&nmi_running); in clear_mod_flag() 174 if (!atomic_read(&nmi_running)) in wait_for_nmi() 179 } while (atomic_read(&nmi_running)); in wait_for_nmi() 350 if (unlikely(atomic_read(¤t->tracing_graph_pause))) in prepare_ftrace_return()
|
/arch/x86/include/asm/ |
D | atomic.h | 23 static inline int atomic_read(const atomic_t *v) in atomic_read() function 187 __i = atomic_read(v); in atomic_add_return() 231 c = atomic_read(v); in __atomic_add_unless() 254 c = atomic_read(v); in atomic_dec_if_positive()
|
/arch/powerpc/kernel/ |
D | crash.c | 125 while ((atomic_read(&cpus_in_crash) < ncpus) && (--msecs > 0)) in crash_kexec_prepare_cpus() 130 if (atomic_read(&cpus_in_crash) >= ncpus) { in crash_kexec_prepare_cpus() 136 ncpus - atomic_read(&cpus_in_crash)); in crash_kexec_prepare_cpus() 167 while (atomic_read(&cpus_in_crash) < ncpus) in crash_kexec_prepare_cpus()
|
D | rtas.c | 721 while (rc == H_MULTI_THREADS_ACTIVE && !atomic_read(&data->done) && in __rtas_suspend_last_cpu() 722 !atomic_read(&data->error)) in __rtas_suspend_last_cpu() 725 if (rc || atomic_read(&data->error)) { in __rtas_suspend_last_cpu() 730 if (atomic_read(&data->error)) in __rtas_suspend_last_cpu() 731 rc = atomic_read(&data->error); in __rtas_suspend_last_cpu() 767 while (rc == H_SUCCESS && !atomic_read(&data->done) && !atomic_read(&data->error)) in __rtas_suspend_cpu() 961 if (atomic_read(&data.error) != 0) in rtas_ibm_suspend_me() 974 return atomic_read(&data.error); in rtas_ibm_suspend_me()
|
/arch/sh/include/asm/ |
D | atomic.h | 16 #define atomic_read(v) (*(volatile int *)&(v)->counter) macro 52 c = atomic_read(v); in __atomic_add_unless()
|
/arch/arm/kernel/ |
D | machine_kexec.c | 89 while ((atomic_read(&waiting_for_crash_ipi) > 0) && msecs) { in machine_crash_shutdown() 93 if (atomic_read(&waiting_for_crash_ipi) > 0) in machine_crash_shutdown()
|
/arch/arm/mach-omap2/ |
D | clockdomain2xxx_3xxx.c | 129 if (atomic_read(&clkdm->usecount) > 0) in omap2_clkdm_allow_idle() 141 if (atomic_read(&clkdm->usecount) > 0) in omap2_clkdm_deny_idle() 228 if (atomic_read(&clkdm->usecount) > 0) in omap3_clkdm_allow_idle() 240 if (atomic_read(&clkdm->usecount) > 0) in omap3_clkdm_deny_idle()
|
/arch/m68k/kernel/ |
D | irq.c | 37 seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); in arch_show_interrupts()
|
/arch/sparc/kernel/ |
D | nmi.c | 57 if (atomic_read(&nmi_active)) { in touch_nmi_watchdog() 179 if (!atomic_read(&nmi_active)) in check_nmi_watchdog() 204 if (!atomic_read(&nmi_active)) { in check_nmi_watchdog()
|
/arch/ia64/include/asm/ |
D | atomic.h | 23 #define atomic_read(v) (*(volatile int *)&(v)->counter) macro 37 old = atomic_read(v); in ia64_atomic_add() 65 old = atomic_read(v); in ia64_atomic_sub() 95 c = atomic_read(v); in __atomic_add_unless()
|
/arch/parisc/kernel/ |
D | ftrace.c | 67 trace->overrun = atomic_read(¤t->trace_overrun); in pop_return_trace() 115 if (unlikely(atomic_read(¤t->tracing_graph_pause))) in prepare_ftrace_return()
|
/arch/avr32/include/asm/ |
D | atomic.h | 22 #define atomic_read(v) (*(volatile int *)&(v)->counter) macro 117 int tmp, old = atomic_read(v); in __atomic_add_unless()
|
/arch/x86/kvm/ |
D | timer.c | 33 if (ktimer->reinject || !atomic_read(&ktimer->pending)) { in kvm_timer_fn()
|
/arch/xtensa/include/asm/ |
D | atomic.h | 49 #define atomic_read(v) (*(volatile int *)&(v)->counter) macro 239 c = atomic_read(v); in __atomic_add_unless()
|
/arch/powerpc/mm/ |
D | icswx.c | 113 if (atomic_read(&mm->mm_users) > 1) in use_cop() 153 if (atomic_read(&mm->mm_users) > 1) in drop_cop()
|
/arch/mn10300/include/asm/ |
D | atomic.h | 37 #define atomic_read(v) (ACCESS_ONCE((v)->counter)) macro 161 c = atomic_read(v); \
|
/arch/blackfin/include/asm/ |
D | atomic.h | 23 #define atomic_read(v) __raw_uncached_fetch_asm(&(v)->counter) macro
|
/arch/tile/include/asm/ |
D | atomic.h | 35 static inline int atomic_read(const atomic_t *v) in atomic_read() function
|
/arch/sparc/oprofile/ |
D | init.c | 61 if (atomic_read(&nmi_active) <= 0) in op_nmi_timer_init()
|