Home
last modified time | relevance | path

Searched refs:dmb (Results 1 – 25 of 29) sorted by relevance

12

/arch/arm64/include/asm/vdso/
Dcompat_barrier.h17 #ifdef dmb
18 #undef dmb
21 #define dmb(option) __asm__ __volatile__ ("dmb " #option : : : "memory") macro
24 #define aarch32_smp_mb() dmb(ish)
25 #define aarch32_smp_rmb() dmb(ishld)
26 #define aarch32_smp_wmb() dmb(ishst)
28 #define aarch32_smp_mb() dmb(ish)
30 #define aarch32_smp_wmb() dmb(ishst)
/arch/arm/include/asm/
Dbarrier.h21 #define dmb(option) __asm__ __volatile__ ("dmb " #option : : : "memory") macro
33 #define dmb(x) __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" \ macro
40 #define dmb(x) __asm__ __volatile__ ("" : : : "memory") macro
45 #define dmb(x) __asm__ __volatile__ ("" : : : "memory") macro
67 #define dma_rmb() dmb(osh)
68 #define dma_wmb() dmb(oshst)
77 #define __smp_mb() dmb(ish)
79 #define __smp_wmb() dmb(ishst)
Dassembler.h310 ALT_SMP(dmb ish)
312 ALT_SMP(W(dmb) ish)
315 ALT_SMP(mcr p15, 0, r0, c7, c10, 5) @ dmb
/arch/arm64/include/asm/
Dbarrier.h22 #define dmb(opt) asm volatile("dmb " #opt : : : "memory") macro
36 #define dma_rmb() dmb(oshld)
37 #define dma_wmb() dmb(oshst)
79 #define __smp_mb() dmb(ish)
80 #define __smp_rmb() dmb(ishld)
81 #define __smp_wmb() dmb(ishst)
Datomic_ll_sc.h86 ATOMIC_OP_RETURN( , dmb ish, , l, "memory", __VA_ARGS__)\
90 ATOMIC_FETCH_OP ( , dmb ish, , l, "memory", __VA_ARGS__)\
101 ATOMIC_FETCH_OP ( , dmb ish, , l, "memory", __VA_ARGS__)\ in ATOMIC_OPS()
182 ATOMIC64_OP_RETURN(, dmb ish, , l, "memory", __VA_ARGS__) \
186 ATOMIC64_FETCH_OP (, dmb ish, , l, "memory", __VA_ARGS__) \
197 ATOMIC64_FETCH_OP (, dmb ish, , l, "memory", __VA_ARGS__) \
290 __CMPXCHG_CASE(w, b, mb_, 8, dmb ish, , l, "memory", K)
291 __CMPXCHG_CASE(w, h, mb_, 16, dmb ish, , l, "memory", K)
292 __CMPXCHG_CASE(w, , mb_, 32, dmb ish, , l, "memory", K)
293 __CMPXCHG_CASE( , , mb_, 64, dmb ish, , l, "memory", L)
[all …]
Dcmpxchg.h57 __XCHG_CASE(w, b, mb_, 8, dmb ish, nop, , a, l, "memory")
58 __XCHG_CASE(w, h, mb_, 16, dmb ish, nop, , a, l, "memory")
59 __XCHG_CASE(w, , mb_, 32, dmb ish, nop, , a, l, "memory")
60 __XCHG_CASE( , , mb_, 64, dmb ish, nop, , a, l, "memory")
/arch/arm/common/
Dvlock.S29 dmb
33 dmb
80 dmb
93 dmb
Dmcpm_head.S121 dmb
136 dmb
148 dmb
152 dmb
173 dmb
182 dmb
196 dmb
Dmcpm_entry.c49 dmb(); in __mcpm_cpu_down()
65 dmb(); in __mcpm_outbound_leave_critical()
/arch/unicore32/include/asm/
Dbarrier.h12 #define dmb() __asm__ __volatile__ ("" : : : "memory") macro
/arch/arm/mm/
Dcache-b15-rac.c66 dmb(); in __b15_rac_disable()
80 dmb(); in __b15_rac_flush()
Dcache-v7.S97 dmb @ ensure ordering with previous memory accesses
125 dmb @ ensure ordering with previous memory accesses
/arch/arm64/mm/
Dflush.c84 dmb(osh); in arch_wb_cache_pmem()
Dproc.S220 dmb sy // lines are written back before
229 dmb sy // that it is visible to all
/arch/arm/mach-omap2/
Dsleep34xx.S97 dmb @ data memory barrier
213 dmb
418 dmb @ data memory barrier
429 dmb @ data memory barrier
444 dmb @ data memory barrier
Domap-smc.S52 dmb
Dsleep33xx.S133 dmb
Dsleep44xx.S350 dmb
Dsleep43xx.S263 dmb
/arch/arm64/kernel/
Dhead.S136 dmb sy // needed before dc ivac with
325 dmb sy
347 dmb sy
401 dmb sy
663 dmb sy
775 dmb sy
/arch/arm/mach-socfpga/
Dself-refresh.S85 dmb
/arch/arm/mach-tegra/
Dsleep.S35 dmb @ ensure ordering
/arch/arm/kernel/
Dsmp_tlb.c153 dmb(); in ipi_flush_tlb_a15_erratum()
/arch/arm/mach-bcm/
Dplatsmp-brcmstb.c75 dmb(); in per_cpu_sw_state_wr()
/arch/arm64/kernel/vdso32/
DMakefile113 dmbinstr := $(call cc32-as-instr,dmb ishld,-DCONFIG_AS_DMB_ISHLD=1)

12