/include/linux/atomic/ |
D | atomic-instrumented.h | 25 atomic_read(const atomic_t *v) in atomic_read() argument 27 instrument_atomic_read(v, sizeof(*v)); in atomic_read() 28 return arch_atomic_read(v); in atomic_read() 32 atomic_read_acquire(const atomic_t *v) in atomic_read_acquire() argument 34 instrument_atomic_read(v, sizeof(*v)); in atomic_read_acquire() 35 return arch_atomic_read_acquire(v); in atomic_read_acquire() 39 atomic_set(atomic_t *v, int i) in atomic_set() argument 41 instrument_atomic_write(v, sizeof(*v)); in atomic_set() 42 arch_atomic_set(v, i); in atomic_set() 46 atomic_set_release(atomic_t *v, int i) in atomic_set_release() argument [all …]
|
D | atomic-long.h | 27 arch_atomic_long_read(const atomic_long_t *v) in arch_atomic_long_read() argument 29 return arch_atomic64_read(v); in arch_atomic_long_read() 33 arch_atomic_long_read_acquire(const atomic_long_t *v) in arch_atomic_long_read_acquire() argument 35 return arch_atomic64_read_acquire(v); in arch_atomic_long_read_acquire() 39 arch_atomic_long_set(atomic_long_t *v, long i) in arch_atomic_long_set() argument 41 arch_atomic64_set(v, i); in arch_atomic_long_set() 45 arch_atomic_long_set_release(atomic_long_t *v, long i) in arch_atomic_long_set_release() argument 47 arch_atomic64_set_release(v, i); in arch_atomic_long_set_release() 51 arch_atomic_long_add(long i, atomic_long_t *v) in arch_atomic_long_add() argument 53 arch_atomic64_add(i, v); in arch_atomic_long_add() [all …]
|
D | atomic-arch-fallback.h | 152 arch_atomic_read_acquire(const atomic_t *v) in arch_atomic_read_acquire() argument 157 ret = smp_load_acquire(&(v)->counter); in arch_atomic_read_acquire() 159 ret = arch_atomic_read(v); in arch_atomic_read_acquire() 170 arch_atomic_set_release(atomic_t *v, int i) in arch_atomic_set_release() argument 173 smp_store_release(&(v)->counter, i); in arch_atomic_set_release() 176 arch_atomic_set(v, i); in arch_atomic_set_release() 190 arch_atomic_add_return_acquire(int i, atomic_t *v) in arch_atomic_add_return_acquire() argument 192 int ret = arch_atomic_add_return_relaxed(i, v); in arch_atomic_add_return_acquire() 201 arch_atomic_add_return_release(int i, atomic_t *v) in arch_atomic_add_return_release() argument 204 return arch_atomic_add_return_relaxed(i, v); in arch_atomic_add_return_release() [all …]
|
/include/linux/ |
D | instrumented.h | 24 static __always_inline void instrument_read(const volatile void *v, size_t size) in instrument_read() argument 26 kasan_check_read(v, size); in instrument_read() 27 kcsan_check_read(v, size); in instrument_read() 39 static __always_inline void instrument_write(const volatile void *v, size_t size) in instrument_write() argument 41 kasan_check_write(v, size); in instrument_write() 42 kcsan_check_write(v, size); in instrument_write() 54 static __always_inline void instrument_read_write(const volatile void *v, size_t size) in instrument_read_write() argument 56 kasan_check_write(v, size); in instrument_read_write() 57 kcsan_check_read_write(v, size); in instrument_read_write() 69 static __always_inline void instrument_atomic_read(const volatile void *v, size_t size) in instrument_atomic_read() argument [all …]
|
D | stdarg.h | 6 #define va_start(v, l) __builtin_va_start(v, l) argument 7 #define va_end(v) __builtin_va_end(v) argument 8 #define va_arg(v, T) __builtin_va_arg(v, T) argument
|
D | atomic.h | 28 #define atomic_cond_read_acquire(v, c) smp_cond_load_acquire(&(v)->counter, (c)) argument 29 #define atomic_cond_read_relaxed(v, c) smp_cond_load_relaxed(&(v)->counter, (c)) argument 31 #define atomic64_cond_read_acquire(v, c) smp_cond_load_acquire(&(v)->counter, (c)) argument 32 #define atomic64_cond_read_relaxed(v, c) smp_cond_load_relaxed(&(v)->counter, (c)) argument
|
D | page_ref.h | 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *page, int v, int u); 34 extern void __page_ref_freeze(struct page *page, int v, int ret); 35 extern void __page_ref_unfreeze(struct page *page, int v); 41 static inline void __page_ref_set(struct page *page, int v) in __page_ref_set() argument 44 static inline void __page_ref_mod(struct page *page, int v) in __page_ref_mod() argument 47 static inline void __page_ref_mod_and_test(struct page *page, int v, int ret) in __page_ref_mod_and_test() argument [all …]
|
D | ssbi.h | 19 u8 v; in ssbi_reg_read() local 21 ret = ssbi_read(context, reg, &v, 1); in ssbi_reg_read() 23 *val = v; in ssbi_reg_read() 31 u8 v = val; in ssbi_reg_write() local 32 return ssbi_write(context, reg, &v, 1); in ssbi_reg_write()
|
D | random.h | 106 static inline bool __must_check arch_get_random_long(unsigned long *v) { return false; } in declare_get_random_var_wait() 107 static inline bool __must_check arch_get_random_int(unsigned int *v) { return false; } 108 static inline bool __must_check arch_get_random_seed_long(unsigned long *v) { return false; } 109 static inline bool __must_check arch_get_random_seed_int(unsigned int *v) { return false; } 117 static inline bool __init arch_get_random_seed_long_early(unsigned long *v) 120 return arch_get_random_seed_long(v); 125 static inline bool __init arch_get_random_long_early(unsigned long *v) in arch_get_random_long_early() argument 128 return arch_get_random_long(v); in arch_get_random_long_early()
|
D | u64_stats_sync.h | 78 local64_t v; member 83 return local64_read(&p->v); in u64_stats_read() 88 local64_add(val, &p->v); in u64_stats_add() 93 local64_inc(&p->v); in u64_stats_inc() 99 u64 v; member 104 return p->v; in u64_stats_read() 109 p->v += val; in u64_stats_add() 114 p->v++; in u64_stats_inc()
|
D | fixp-arith.h | 80 #define fixp_cos32(v) fixp_sin32((v) + 90) argument 88 #define fixp_sin16(v) (fixp_sin32(v) >> 16) argument 89 #define fixp_cos16(v) (fixp_cos32(v) >> 16) argument
|
D | win_minmax.h | 14 u32 v; /* value measured */ member 24 return m->s[0].v; in minmax_get() 29 struct minmax_sample val = { .t = t, .v = meas }; in minmax_reset() 32 return m->s[0].v; in minmax_reset()
|
/include/asm-generic/ |
D | atomic.h | 20 static inline void generic_atomic_##op(int i, atomic_t *v) \ 24 c = v->counter; \ 25 while ((old = arch_cmpxchg(&v->counter, c, c c_op i)) != c) \ 30 static inline int generic_atomic_##op##_return(int i, atomic_t *v) \ 34 c = v->counter; \ 35 while ((old = arch_cmpxchg(&v->counter, c, c c_op i)) != c) \ 42 static inline int generic_atomic_fetch_##op(int i, atomic_t *v) \ 46 c = v->counter; \ 47 while ((old = arch_cmpxchg(&v->counter, c, c c_op i)) != c) \ 58 static inline void generic_atomic_##op(int i, atomic_t *v) \ [all …]
|
D | atomic64.h | 18 extern s64 generic_atomic64_read(const atomic64_t *v); 19 extern void generic_atomic64_set(atomic64_t *v, s64 i); 22 extern void generic_atomic64_##op(s64 a, atomic64_t *v); 25 extern s64 generic_atomic64_##op##_return(s64 a, atomic64_t *v); 28 extern s64 generic_atomic64_fetch_##op(s64 a, atomic64_t *v); 47 extern s64 generic_atomic64_dec_if_positive(atomic64_t *v); 48 extern s64 generic_atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n); 49 extern s64 generic_atomic64_xchg(atomic64_t *v, s64 new); 50 extern s64 generic_atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u);
|
D | barrier.h | 105 #define __smp_store_release(p, v) \ argument 109 WRITE_ONCE(*p, v); \ 138 #define smp_store_release(p, v) __smp_store_release(p, v) argument 160 #define smp_store_release(p, v) \ argument 164 WRITE_ONCE(*p, v); \ 187 #define virt_store_release(p, v) __smp_store_release(p, v) argument
|
/include/net/ |
D | red.h | 160 static inline void red_set_vars(struct red_vars *v) in red_set_vars() argument 166 v->qavg = 0; in red_set_vars() 168 v->qcount = -1; in red_set_vars() 268 static inline int red_is_idling(const struct red_vars *v) in red_is_idling() argument 270 return v->qidlestart != 0; in red_is_idling() 273 static inline void red_start_of_idle_period(struct red_vars *v) in red_start_of_idle_period() argument 275 v->qidlestart = ktime_get(); in red_start_of_idle_period() 278 static inline void red_end_of_idle_period(struct red_vars *v) in red_end_of_idle_period() argument 280 v->qidlestart = 0; in red_end_of_idle_period() 283 static inline void red_restart(struct red_vars *v) in red_restart() argument [all …]
|
/include/linux/platform_data/ |
D | mcs.h | 11 #define MCS_KEY_MAP(v, c) ((((v) & 0xff) << 16) | ((c) & 0xffff)) argument 12 #define MCS_KEY_VAL(v) (((v) >> 16) & 0xff) argument 13 #define MCS_KEY_CODE(v) ((v) & 0xffff) argument
|
/include/trace/events/ |
D | page_ref.h | 15 TP_PROTO(struct page *page, int v), 17 TP_ARGS(page, v), 36 __entry->val = v; 49 TP_PROTO(struct page *page, int v), 51 TP_ARGS(page, v) 56 TP_PROTO(struct page *page, int v), 58 TP_ARGS(page, v) 63 TP_PROTO(struct page *page, int v, int ret), 65 TP_ARGS(page, v, ret), 85 __entry->val = v; [all …]
|
/include/acpi/platform/ |
D | acgcc.h | 20 #define va_start(v, l) __builtin_va_start(v, l) argument 21 #define va_end(v) __builtin_va_end(v) argument 22 #define va_arg(v, l) __builtin_va_arg(v, l) argument
|
/include/linux/dsa/ |
D | brcm.h | 13 #define BRCM_TAG_GET_PORT(v) ((v) >> 8) argument 14 #define BRCM_TAG_GET_QUEUE(v) ((v) & 0xff) argument
|
/include/linux/ceph/ |
D | decode.h | 21 u64 v = get_unaligned_le64(*p); in ceph_decode_64() local 23 return v; in ceph_decode_64() 27 u32 v = get_unaligned_le32(*p); in ceph_decode_32() local 29 return v; in ceph_decode_32() 33 u16 v = get_unaligned_le16(*p); in ceph_decode_16() local 35 return v; in ceph_decode_16() 39 u8 v = *(u8 *)*p; in ceph_decode_8() local 41 return v; in ceph_decode_8() 63 #define ceph_decode_64_safe(p, end, v, bad) \ argument 66 v = ceph_decode_64(p); \ [all …]
|
D | pagelist.h | 42 static inline int ceph_pagelist_encode_64(struct ceph_pagelist *pl, u64 v) in ceph_pagelist_encode_64() argument 44 __le64 ev = cpu_to_le64(v); in ceph_pagelist_encode_64() 47 static inline int ceph_pagelist_encode_32(struct ceph_pagelist *pl, u32 v) in ceph_pagelist_encode_32() argument 49 __le32 ev = cpu_to_le32(v); in ceph_pagelist_encode_32() 52 static inline int ceph_pagelist_encode_16(struct ceph_pagelist *pl, u16 v) in ceph_pagelist_encode_16() argument 54 __le16 ev = cpu_to_le16(v); in ceph_pagelist_encode_16() 57 static inline int ceph_pagelist_encode_8(struct ceph_pagelist *pl, u8 v) in ceph_pagelist_encode_8() argument 59 return ceph_pagelist_append(pl, &v, 1); in ceph_pagelist_encode_8()
|
/include/kvm/ |
D | arm_arch_timer.h | 88 #define vcpu_timer(v) (&(v)->arch.timer_cpu) argument 89 #define vcpu_get_timer(v,t) (&vcpu_timer(v)->timers[(t)]) argument 90 #define vcpu_vtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_VTIMER]) argument 91 #define vcpu_ptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_PTIMER]) argument
|
/include/linux/spi/ |
D | mxs-spi.h | 58 #define BF_SSP_TIMING_CLOCK_DIVIDE(v) \ argument 59 (((v) << 8) & BM_SSP_TIMING_CLOCK_DIVIDE) 62 #define BF_SSP_TIMING_CLOCK_RATE(v) \ argument 63 (((v) << 0) & BM_SSP_TIMING_CLOCK_RATE) 86 #define BF_SSP_CTRL1_WORD_LENGTH(v) \ argument 87 (((v) << 4) & BM_SSP_CTRL1_WORD_LENGTH) 93 #define BF_SSP_CTRL1_SSP_MODE(v) \ argument 94 (((v) << 0) & BM_SSP_CTRL1_SSP_MODE)
|
/include/xen/arm/ |
D | page.h | 79 #define virt_to_gfn(v) \ argument 81 WARN_ON_ONCE(!virt_addr_valid(v)); \ 82 pfn_to_gfn(virt_to_phys(v) >> XEN_PAGE_SHIFT); \ 86 #define percpu_to_gfn(v) \ argument 87 (pfn_to_gfn(per_cpu_ptr_to_phys(v) >> XEN_PAGE_SHIFT))
|