| /include/asm-generic/ |
| D | local64.h | 30 #define local64_read(l) local_read(&(l)->a) argument 31 #define local64_set(l,i) local_set((&(l)->a),(i)) argument 32 #define local64_inc(l) local_inc(&(l)->a) argument 33 #define local64_dec(l) local_dec(&(l)->a) argument 34 #define local64_add(i,l) local_add((i),(&(l)->a)) argument 35 #define local64_sub(i,l) local_sub((i),(&(l)->a)) argument 37 #define local64_sub_and_test(i, l) local_sub_and_test((i), (&(l)->a)) argument 38 #define local64_dec_and_test(l) local_dec_and_test(&(l)->a) argument 39 #define local64_inc_and_test(l) local_inc_and_test(&(l)->a) argument 40 #define local64_add_negative(i, l) local_add_negative((i), (&(l)->a)) argument [all …]
|
| D | local.h | 29 #define local_read(l) atomic_long_read(&(l)->a) argument 30 #define local_set(l,i) atomic_long_set((&(l)->a),(i)) argument 31 #define local_inc(l) atomic_long_inc(&(l)->a) argument 32 #define local_dec(l) atomic_long_dec(&(l)->a) argument 33 #define local_add(i,l) atomic_long_add((i),(&(l)->a)) argument 34 #define local_sub(i,l) atomic_long_sub((i),(&(l)->a)) argument 36 #define local_sub_and_test(i, l) atomic_long_sub_and_test((i), (&(l)->a)) argument 37 #define local_dec_and_test(l) atomic_long_dec_and_test(&(l)->a) argument 38 #define local_inc_and_test(l) atomic_long_inc_and_test(&(l)->a) argument 39 #define local_add_negative(i, l) atomic_long_add_negative((i), (&(l)->a)) argument [all …]
|
| D | qrwlock.h | 139 #define arch_read_lock(l) queued_read_lock(l) argument 140 #define arch_write_lock(l) queued_write_lock(l) argument 141 #define arch_read_trylock(l) queued_read_trylock(l) argument 142 #define arch_write_trylock(l) queued_write_trylock(l) argument 143 #define arch_read_unlock(l) queued_read_unlock(l) argument 144 #define arch_write_unlock(l) queued_write_unlock(l) argument 145 #define arch_rwlock_is_contended(l) queued_rwlock_is_contended(l) argument
|
| D | qspinlock.h | 143 #define arch_spin_is_locked(l) queued_spin_is_locked(l) argument 144 #define arch_spin_is_contended(l) queued_spin_is_contended(l) argument 145 #define arch_spin_value_unlocked(l) queued_spin_value_unlocked(l) argument 146 #define arch_spin_lock(l) queued_spin_lock(l) argument 147 #define arch_spin_trylock(l) queued_spin_trylock(l) argument 148 #define arch_spin_unlock(l) queued_spin_unlock(l) argument
|
| /include/linux/ |
| D | bits.h | 23 #define GENMASK_INPUT_CHECK(h, l) \ argument 25 __is_constexpr((l) > (h)), (l) > (h), 0))) 31 #define GENMASK_INPUT_CHECK(h, l) 0 argument 34 #define GENMASK(h, l) \ argument 35 (GENMASK_INPUT_CHECK(h, l) + __GENMASK(h, l)) 36 #define GENMASK_ULL(h, l) \ argument 37 (GENMASK_INPUT_CHECK(h, l) + __GENMASK_ULL(h, l)) 50 #define GENMASK_U128(h, l) \ argument 51 (GENMASK_INPUT_CHECK(h, l) + __GENMASK_U128(h, l))
|
| D | lockdep.h | 259 #define lock_set_novalidate_class(l, n, i) \ argument 260 lock_set_class(l, n, &__lockdep_no_validate__, 0, i) 284 #define lockdep_assert_held(l) \ argument 285 lockdep_assert(lockdep_is_held(l) != LOCK_STATE_NOT_HELD) 287 #define lockdep_assert_not_held(l) \ argument 288 lockdep_assert(lockdep_is_held(l) != LOCK_STATE_HELD) 290 #define lockdep_assert_held_write(l) \ argument 291 lockdep_assert(lockdep_is_held_type(l, 0)) 293 #define lockdep_assert_held_read(l) \ argument 294 lockdep_assert(lockdep_is_held_type(l, 1)) [all …]
|
| D | local_lock_internal.h | 27 static inline void local_lock_acquire(local_lock_t *l) in local_lock_acquire() argument 29 lock_map_acquire(&l->dep_map); in local_lock_acquire() 30 DEBUG_LOCKS_WARN_ON(l->owner); in local_lock_acquire() 31 l->owner = current; in local_lock_acquire() 34 static inline void local_lock_release(local_lock_t *l) in local_lock_release() argument 36 DEBUG_LOCKS_WARN_ON(l->owner != current); in local_lock_release() 37 l->owner = NULL; in local_lock_release() 38 lock_map_release(&l->dep_map); in local_lock_release() 41 static inline void local_lock_debug_init(local_lock_t *l) in local_lock_debug_init() argument 43 l->owner = NULL; in local_lock_debug_init() [all …]
|
| D | math64.h | 203 } l; in mul_u64_u64_shr() member 210 rl.ll = mul_u32_u32(a0.l.low, b0.l.low); in mul_u64_u64_shr() 211 rm.ll = mul_u32_u32(a0.l.low, b0.l.high); in mul_u64_u64_shr() 212 rn.ll = mul_u32_u32(a0.l.high, b0.l.low); in mul_u64_u64_shr() 213 rh.ll = mul_u32_u32(a0.l.high, b0.l.high); in mul_u64_u64_shr() 220 rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low; in mul_u64_u64_shr() 221 rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low; in mul_u64_u64_shr() 222 rh.l.high = (c >> 32) + rh.l.high; in mul_u64_u64_shr() 267 } l; in mul_u64_u32_div() member 271 rl.ll = mul_u32_u32(u.l.low, mul); in mul_u64_u32_div() [all …]
|
| D | io-pgtable-arm.h | 42 #define ARM_LPAE_LVL_SHIFT(l,d) \ argument 43 (((ARM_LPAE_MAX_LEVELS - (l)) * (d)->bits_per_level) + \ 58 #define ARM_LPAE_PGD_IDX(l,d) \ argument 59 ((l) == (d)->start_level ? (d)->pgd_bits - (d)->bits_per_level : 0) 61 #define ARM_LPAE_LVL_IDX(a,l,d) \ argument 62 (((u64)(a) >> ARM_LPAE_LVL_SHIFT(l,d)) & \ 63 ((1 << ((d)->bits_per_level + ARM_LPAE_PGD_IDX(l,d))) - 1)) 66 #define ARM_LPAE_BLOCK_SIZE(l,d) (1ULL << ARM_LPAE_LVL_SHIFT(l,d)) argument
|
| D | stdarg.h | 6 #define va_start(v, l) __builtin_va_start(v, l) argument
|
| D | lockref.h | 47 static inline bool __lockref_is_dead(const struct lockref *l) in __lockref_is_dead() argument 49 return ((int)l->count < 0); in __lockref_is_dead()
|
| D | time.h | 98 #define time_between32(t, l, h) ((u32)(h) - (u32)(l) >= (u32)(t) - (u32)(l)) argument
|
| D | bitops.h | 204 static inline unsigned int fls_long(unsigned long l) in fls_long() argument 206 if (sizeof(l) == 4) in fls_long() 207 return fls(l); in fls_long() 208 return fls64(l); in fls_long() 225 static inline int get_count_order_long(unsigned long l) in get_count_order_long() argument 227 if (l == 0UL) in get_count_order_long() 229 return (int)fls_long(--l); in get_count_order_long()
|
| D | ww_mutex.h | 30 #define ww_mutex_base_init(l,n,k) __mutex_init(l,n,k) argument 34 #define ww_mutex_base_init(l,n,k) __rt_mutex_init(l,n,k) argument
|
| D | migrate.h | 65 void putback_movable_pages(struct list_head *l); 68 int migrate_pages(struct list_head *l, new_folio_t new, free_folio_t free, 85 static inline void putback_movable_pages(struct list_head *l) {} in putback_movable_pages() argument 86 static inline int migrate_pages(struct list_head *l, new_folio_t new, in migrate_pages() argument
|
| D | fs_context.h | 191 #define __logfc(fc, l, fmt, ...) logfc((fc)->log.log, NULL, \ argument 192 l, fmt, ## __VA_ARGS__) 193 #define __plog(p, l, fmt, ...) logfc((p)->log, (p)->prefix, \ argument 194 l, fmt, ## __VA_ARGS__)
|
| D | badblocks.h | 19 #define BB_MAKE(a, l, ack) (((a)<<9) | ((l)-1) | ((u64)(!!(ack)) << 63)) argument
|
| /include/uapi/linux/ |
| D | bits.h | 7 #define __GENMASK(h, l) \ argument 8 (((~_UL(0)) - (_UL(1) << (l)) + 1) & \ 11 #define __GENMASK_ULL(h, l) \ argument 12 (((~_ULL(0)) - (_ULL(1) << (l)) + 1) & \ 15 #define __GENMASK_U128(h, l) \ argument 16 ((_BIT128((h)) << 1) - (_BIT128(l)))
|
| D | romfs_fs.h | 17 #define __mkw(h,l) (((h)&0x00ff)<< 8|((l)&0x00ff)) argument 18 #define __mkl(h,l) (((h)&0xffff)<<16|((l)&0xffff)) argument
|
| /include/crypto/ |
| D | utils.h | 23 unsigned long l; in crypto_xor() local 26 l = get_unaligned(d) ^ get_unaligned(s++); in crypto_xor() 27 put_unaligned(l, d++); in crypto_xor() 44 unsigned long l; in crypto_xor_cpy() local 47 l = get_unaligned(s1++) ^ get_unaligned(s2++); in crypto_xor_cpy() 48 put_unaligned(l, d++); in crypto_xor_cpy()
|
| /include/linux/ceph/ |
| D | striper.h | 10 void ceph_calc_file_object_mapping(struct ceph_file_layout *l, 36 int ceph_file_to_extents(struct ceph_file_layout *l, u64 off, u64 len, 42 int ceph_iterate_extents(struct ceph_file_layout *l, u64 off, u64 len, 64 int ceph_extent_to_file(struct ceph_file_layout *l, 69 u64 ceph_get_num_objects(struct ceph_file_layout *l, u64 size);
|
| /include/kvm/ |
| D | iodev.h | 44 int l, void *v) in kvm_iodevice_read() argument 46 return dev->ops->read ? dev->ops->read(vcpu, dev, addr, l, v) in kvm_iodevice_read() 52 int l, const void *v) in kvm_iodevice_write() argument 54 return dev->ops->write ? dev->ops->write(vcpu, dev, addr, l, v) in kvm_iodevice_write()
|
| /include/asm-generic/bitops/ |
| D | ext2-atomic-setbit.h | 9 #define ext2_set_bit_atomic(l, nr, addr) test_and_set_bit_le(nr, addr) argument 10 #define ext2_clear_bit_atomic(l, nr, addr) test_and_clear_bit_le(nr, addr) argument
|
| /include/media/i2c/ |
| D | ov772x.h | 35 #define OV772X_AUTO_EDGECTRL(u, l) \ argument 38 .lower = (l & OV772X_EDGE_LOWER_MASK), \
|
| /include/video/ |
| D | pm3fb.h | 591 #define PM3FBDestReadMode_Layout0(l) (((l) & 0x3) << 12) argument 592 #define PM3FBDestReadMode_Layout1(l) (((l) & 0x3) << 14) argument 593 #define PM3FBDestReadMode_Layout2(l) (((l) & 0x3) << 16) argument 594 #define PM3FBDestReadMode_Layout3(l) (((l) & 0x3) << 18) argument 620 #define PM3FBSourceReadMode_Layout(l) (((l) & 0x3) << 8) argument 660 #define PM3FBWriteMode_Layout0(l) (((l) & 0x3) << 16) argument 661 #define PM3FBWriteMode_Layout1(l) (((l) & 0x3) << 18) argument 662 #define PM3FBWriteMode_Layout2(l) (((l) & 0x3) << 20) argument 663 #define PM3FBWriteMode_Layout3(l) (((l) & 0x3) << 22) argument
|