| /include/linux/ |
| D | syscalls.h | 93 #define __MAP1(m,t,a) m(t,a) argument 94 #define __MAP2(m,t,a,...) m(t,a), __MAP1(m,__VA_ARGS__) argument 95 #define __MAP3(m,t,a,...) m(t,a), __MAP2(m,__VA_ARGS__) argument 96 #define __MAP4(m,t,a,...) m(t,a), __MAP3(m,__VA_ARGS__) argument 97 #define __MAP5(m,t,a,...) m(t,a), __MAP4(m,__VA_ARGS__) argument 98 #define __MAP6(m,t,a,...) m(t,a), __MAP5(m,__VA_ARGS__) argument 101 #define __SC_DECL(t, a) t a argument 102 #define __TYPE_IS_L(t) (__same_type((t)0, 0L)) argument 103 #define __TYPE_IS_UL(t) (__same_type((t)0, 0UL)) argument 104 #define __TYPE_IS_LL(t) (__same_type((t)0, 0LL) || __same_type((t)0, 0ULL)) argument [all …]
|
| D | elfcore.h | 30 static inline int elf_core_copy_task_regs(struct task_struct *t, elf_gregset_t* elfregs) in elf_core_copy_task_regs() 42 static inline int elf_core_copy_task_fpregs(struct task_struct *t, struct pt_regs *regs, elf_fpregs… in elf_core_copy_task_fpregs() 52 static inline int elf_core_copy_task_xfpregs(struct task_struct *t, elf_fpxregset_t *xfpu) in elf_core_copy_task_xfpregs()
|
| D | interrupt.h | 538 static inline int tasklet_trylock(struct tasklet_struct *t) in tasklet_trylock() 543 static inline void tasklet_unlock(struct tasklet_struct *t) in tasklet_unlock() 549 static inline void tasklet_unlock_wait(struct tasklet_struct *t) in tasklet_unlock_wait() 554 #define tasklet_trylock(t) 1 argument 555 #define tasklet_unlock_wait(t) do { } while (0) argument 556 #define tasklet_unlock(t) do { } while (0) argument 561 static inline void tasklet_schedule(struct tasklet_struct *t) in tasklet_schedule() 569 static inline void tasklet_hi_schedule(struct tasklet_struct *t) in tasklet_hi_schedule() 583 static inline void tasklet_hi_schedule_first(struct tasklet_struct *t) in tasklet_hi_schedule_first() 590 static inline void tasklet_disable_nosync(struct tasklet_struct *t) in tasklet_disable_nosync() [all …]
|
| D | kcov.h | 29 static inline void kcov_task_init(struct task_struct *t) {} in kcov_task_init() 30 static inline void kcov_task_exit(struct task_struct *t) {} in kcov_task_exit()
|
| D | win_minmax.h | 12 u32 t; /* time measurement was taken */ member 26 static inline u32 minmax_reset(struct minmax *m, u32 t, u32 meas) in minmax_reset()
|
| D | nls.h | 61 static inline unsigned char nls_tolower(struct nls_table *t, unsigned char c) in nls_tolower() 68 static inline unsigned char nls_toupper(struct nls_table *t, unsigned char c) in nls_toupper() 75 static inline int nls_strnicmp(struct nls_table *t, const unsigned char *s1, in nls_strnicmp()
|
| D | lockdep.h | 395 # define lock_acquire(l, s, t, r, c, n, i) do { } while (0) argument 522 #define lock_acquire_exclusive(l, s, t, n, i) lock_acquire(l, s, t, 0, 1, n, i) argument 523 #define lock_acquire_shared(l, s, t, n, i) lock_acquire(l, s, t, 1, 1, n, i) argument 524 #define lock_acquire_shared_recursive(l, s, t, n, i) lock_acquire(l, s, t, 2, 1, n, i) argument 526 #define spin_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 527 #define spin_acquire_nest(l, s, t, n, i) lock_acquire_exclusive(l, s, t, n, i) argument 530 #define rwlock_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 531 #define rwlock_acquire_read(l, s, t, i) lock_acquire_shared_recursive(l, s, t, NULL, i) argument 534 #define seqcount_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 535 #define seqcount_acquire_read(l, s, t, i) lock_acquire_shared_recursive(l, s, t, NULL, i) argument [all …]
|
| D | capability.h | 212 static inline bool has_capability(struct task_struct *t, int cap) in has_capability() 216 static inline bool has_ns_capability(struct task_struct *t, in has_ns_capability() 221 static inline bool has_capability_noaudit(struct task_struct *t, int cap) in has_capability_noaudit() 225 static inline bool has_ns_capability_noaudit(struct task_struct *t, in has_ns_capability_noaudit()
|
| D | profile.h | 122 static inline int profile_event_register(enum profile_type t, struct notifier_block * n) in profile_event_register() 127 static inline int profile_event_unregister(enum profile_type t, struct notifier_block * n) in profile_event_unregister()
|
| D | filter.h | 332 #define __BPF_MAP_1(m, v, t, a, ...) m(t, a) argument 333 #define __BPF_MAP_2(m, v, t, a, ...) m(t, a), __BPF_MAP_1(m, v, __VA_ARGS__) argument 334 #define __BPF_MAP_3(m, v, t, a, ...) m(t, a), __BPF_MAP_2(m, v, __VA_ARGS__) argument 335 #define __BPF_MAP_4(m, v, t, a, ...) m(t, a), __BPF_MAP_3(m, v, __VA_ARGS__) argument 336 #define __BPF_MAP_5(m, v, t, a, ...) m(t, a), __BPF_MAP_4(m, v, __VA_ARGS__) argument 348 #define __BPF_CAST(t, a) \ argument 356 #define __BPF_DECL_ARGS(t, a) t a argument 357 #define __BPF_DECL_REGS(t, a) u64 a argument
|
| D | reciprocal_div.h | 31 u32 t = (u32)(((u64)a * R.m) >> 32); in reciprocal_divide() local
|
| D | page_ref.h | 26 #define page_ref_tracepoint_active(t) static_key_false(&(t).key) argument 38 #define page_ref_tracepoint_active(t) false argument
|
| D | thermal.h | 50 #define DECI_KELVIN_TO_CELSIUS(t) ({ \ argument 54 #define CELSIUS_TO_DECI_KELVIN(t) ((t)*10+2732) argument 55 #define DECI_KELVIN_TO_MILLICELSIUS_WITH_OFFSET(t, off) (((t) - (off)) * 100) argument 56 #define DECI_KELVIN_TO_MILLICELSIUS(t) DECI_KELVIN_TO_MILLICELSIUS_WITH_OFFSET(t, 2732) argument 57 #define MILLICELSIUS_TO_DECI_KELVIN_WITH_OFFSET(t, off) (((t) / 100) + (off)) argument 58 #define MILLICELSIUS_TO_DECI_KELVIN(t) MILLICELSIUS_TO_DECI_KELVIN_WITH_OFFSET(t, 2732) argument
|
| D | watchdog.h | 155 static inline bool watchdog_timeout_invalid(struct watchdog_device *wdd, unsigned int t) in watchdog_timeout_invalid() 175 unsigned int t) in watchdog_pretimeout_invalid()
|
| D | bch.h | 52 unsigned int t; member
|
| /include/crypto/ |
| D | gf128mul.h | 170 be128 t[256]; member 178 static inline void gf128mul_free_4k(struct gf128mul_4k *t) in gf128mul_free_4k() 187 struct gf128mul_4k *t[16]; member
|
| /include/linux/sunrpc/ |
| D | sched.h | 131 #define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC) argument 132 #define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER) argument 133 #define RPC_DO_ROOTOVERRIDE(t) ((t)->tk_flags & RPC_TASK_ROOTCREDS) argument 134 #define RPC_ASSASSINATED(t) ((t)->tk_flags & RPC_TASK_KILLED) argument 135 #define RPC_IS_SOFT(t) ((t)->tk_flags & (RPC_TASK_SOFT|RPC_TASK_TIMEOUT)) argument 136 #define RPC_IS_SOFTCONN(t) ((t)->tk_flags & RPC_TASK_SOFTCONN) argument 137 #define RPC_WAS_SENT(t) ((t)->tk_flags & RPC_TASK_SENT) argument 143 #define RPC_IS_RUNNING(t) test_bit(RPC_TASK_RUNNING, &(t)->tk_runstate) argument 144 #define rpc_set_running(t) set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate) argument 145 #define rpc_test_and_set_running(t) \ argument [all …]
|
| D | timer.h | 28 int *t; in rpc_set_timeo() local
|
| /include/asm-generic/ |
| D | ioctl.h | 7 #define _IOC_TYPECHECK(t) (sizeof(t)) argument 11 #define _IOC_TYPECHECK(t) \ argument
|
| /include/linux/netfilter/ipset/ |
| D | ip_set_timeout.h | 43 ip_set_timeout_expired(unsigned long *t) in ip_set_timeout_expired() 51 unsigned long t; in ip_set_timeout_set() local
|
| /include/net/netfilter/ |
| D | nf_conntrack_tuple.h | 81 static inline void nf_ct_dump_tuple_ip(const struct nf_conntrack_tuple *t) in nf_ct_dump_tuple_ip() 91 static inline void nf_ct_dump_tuple_ipv6(const struct nf_conntrack_tuple *t) in nf_ct_dump_tuple_ipv6() 101 static inline void nf_ct_dump_tuple(const struct nf_conntrack_tuple *t) in nf_ct_dump_tuple() 178 nf_ct_tuple_mask_cmp(const struct nf_conntrack_tuple *t, in nf_ct_tuple_mask_cmp()
|
| /include/scsi/ |
| D | scsi_transport.h | 93 scsi_transport_reserve_target(struct scsi_transport_template * t, int space) in scsi_transport_reserve_target() 100 scsi_transport_reserve_device(struct scsi_transport_template * t, int space) in scsi_transport_reserve_device()
|
| /include/acpi/platform/ |
| D | aclinux.h | 101 #define ACPI_GLOBAL(t,a) argument 102 #define ACPI_INIT_GLOBAL(t,a,b) argument
|
| /include/media/i2c/ |
| D | ov772x.h | 44 #define OV772X_MANUAL_EDGECTRL(s, t) \ argument
|
| /include/trace/ |
| D | perf.h | 27 #define __perf_task(t) (__task = (t)) argument
|