/arch/arc/include/asm/ |
D | cmpxchg.h | 154 #define xchg(ptr, with) \ macro 167 #define xchg(ptr, with) _xchg(ptr, with) macro 201 #define xchg(ptr, with) ({ \ macro 222 #define atomic_xchg(v, new) (xchg(&((v)->counter), new))
|
/arch/x86/include/asm/ |
D | pgtable-2level.h | 52 return __pte(xchg(&xp->pte_low, 0)); in native_ptep_get_and_clear() 61 return __pmd(xchg((pmdval_t *)xp, 0)); in native_pmdp_get_and_clear() 70 return __pud(xchg((pudval_t *)xp, 0)); in native_pudp_get_and_clear()
|
D | pgtable-3level.h | 177 res.pmd_low = xchg(&orig->pmd_low, 0); in native_pmdp_get_and_clear() 207 old.pmd_low = xchg(&ptr->pmd_low, new.pmd_low); in pmdp_establish() 239 res.pud_low = xchg(&orig->pud_low, 0); in native_pudp_get_and_clear()
|
D | pgtable_64.h | 88 return native_make_pte(xchg(&xp->pte, 0)); in native_ptep_get_and_clear() 101 return native_make_pmd(xchg(&xp->pmd, 0)); in native_pmdp_get_and_clear() 124 return native_make_pud(xchg(&xp->pud, 0)); in native_pudp_get_and_clear()
|
D | barrier.h | 64 #define __smp_store_mb(var, value) do { (void)xchg(&var, value); } while (0)
|
/arch/h8300/include/asm/ |
D | cmpxchg.h | 7 #define xchg(ptr, x) \ macro 64 #define atomic_xchg(v, new) (xchg(&((v)->counter), new))
|
/arch/sparc/include/asm/ |
D | atomic_64.h | 57 return xchg(&v->counter, new); in ATOMIC_OPS() 62 #define atomic64_xchg(v, new) (xchg(&((v)->counter), new))
|
D | cmpxchg_32.h | 28 #define xchg(ptr,x) ((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr)))) macro
|
D | parport.h | 52 if (xchg(&sparc_ebus_dmas[dmanr].lock, 1) != 0) in request_dma() 63 if (xchg(&sparc_ebus_dmas[dmanr].lock, 0) == 0) { in free_dma()
|
/arch/arm64/include/asm/xen/ |
D | events.h | 20 #define xchg_xen_ulong(ptr, val) xchg((ptr), (val))
|
/arch/arm/kernel/ |
D | dma.c | 62 if (xchg(&dma->lock, 1) != 0) in request_dma() 74 xchg(&dma->lock, 0); in request_dma() 105 if (xchg(&dma->lock, 0) != 0) { in free_dma()
|
/arch/x86/include/asm/xen/ |
D | events.h | 24 #define xchg_xen_ulong(ptr, val) xchg((ptr), (val))
|
/arch/sh/include/asm/ |
D | atomic.h | 35 #define atomic_xchg(v, new) (xchg(&((v)->counter), new))
|
D | barrier.h | 41 #define __smp_store_mb(var, value) do { (void)xchg(&var, value); } while (0)
|
D | cmpxchg.h | 48 #define xchg(ptr,x) \ macro
|
/arch/microblaze/include/asm/ |
D | cmpxchg.h | 37 #define xchg(ptr, x) \ macro
|
/arch/unicore32/include/asm/ |
D | cmpxchg.h | 41 #define xchg(ptr, x) \ macro
|
/arch/hexagon/include/asm/ |
D | cmpxchg.h | 45 #define xchg(ptr, v) ((__typeof__(*(ptr)))__xchg((unsigned long)(v), (ptr), \ macro
|
/arch/c6x/include/asm/ |
D | cmpxchg.h | 44 #define xchg(ptr, x) \ macro
|
/arch/um/kernel/ |
D | irq.c | 544 mask = xchg(&pending_mask, *mask_out); in to_irq_stack() 558 mask = xchg(&pending_mask, old); in to_irq_stack() 577 mask = xchg(&pending_mask, 0); in to_irq_stack() 596 mask = xchg(&pending_mask, 0); in from_irq_stack()
|
/arch/s390/include/asm/ |
D | atomic.h | 78 #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) in ATOMIC_OPS() 125 #define atomic64_xchg(v, new) (xchg(&((v)->counter), new))
|
D | cmpxchg.h | 26 #define xchg(ptr, x) \ macro
|
/arch/riscv/include/asm/ |
D | cmpxchg.h | 143 #define xchg(ptr, x) \ macro 152 xchg((ptr), (x)); \ 158 xchg((ptr), (x)); \
|
/arch/csky/include/asm/ |
D | cmpxchg.h | 36 #define xchg(ptr, x) (__xchg((x), (ptr), sizeof(*(ptr)))) macro
|
/arch/parisc/include/asm/ |
D | atomic.h | 78 #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) 224 #define atomic64_xchg(v, new) (xchg(&((v)->counter), new))
|