/include/linux/ |
D | units.h | 33 static inline long milli_kelvin_to_millicelsius(long t) in milli_kelvin_to_millicelsius() argument 35 return t + ABSOLUTE_ZERO_MILLICELSIUS; in milli_kelvin_to_millicelsius() 38 static inline long millicelsius_to_milli_kelvin(long t) in millicelsius_to_milli_kelvin() argument 40 return t - ABSOLUTE_ZERO_MILLICELSIUS; in millicelsius_to_milli_kelvin() 46 static inline long kelvin_to_millicelsius(long t) in kelvin_to_millicelsius() argument 48 return milli_kelvin_to_millicelsius(t * MILLIDEGREE_PER_DEGREE); in kelvin_to_millicelsius() 51 static inline long millicelsius_to_kelvin(long t) in millicelsius_to_kelvin() argument 53 t = millicelsius_to_milli_kelvin(t); in millicelsius_to_kelvin() 55 return DIV_ROUND_CLOSEST(t, MILLIDEGREE_PER_DEGREE); in millicelsius_to_kelvin() 58 static inline long deci_kelvin_to_celsius(long t) in deci_kelvin_to_celsius() argument [all …]
|
D | btf.h | 101 int btf_find_spin_lock(const struct btf *btf, const struct btf_type *t); 102 int btf_find_timer(const struct btf *btf, const struct btf_type *t); 103 bool btf_type_is_void(const struct btf_type *t); 114 const char *btf_type_str(const struct btf_type *t); 126 static inline bool btf_type_is_ptr(const struct btf_type *t) in btf_type_is_ptr() argument 128 return BTF_INFO_KIND(t->info) == BTF_KIND_PTR; in btf_type_is_ptr() 131 static inline bool btf_type_is_int(const struct btf_type *t) in btf_type_is_int() argument 133 return BTF_INFO_KIND(t->info) == BTF_KIND_INT; in btf_type_is_int() 136 static inline bool btf_type_is_small_int(const struct btf_type *t) in btf_type_is_small_int() argument 138 return btf_type_is_int(t) && t->size <= sizeof(u64); in btf_type_is_small_int() [all …]
|
D | rcupdate_trace.h | 34 void rcu_read_unlock_trace_special(struct task_struct *t, int nesting); 50 struct task_struct *t = current; in rcu_read_lock_trace() local 52 WRITE_ONCE(t->trc_reader_nesting, READ_ONCE(t->trc_reader_nesting) + 1); in rcu_read_lock_trace() 55 t->trc_reader_special.b.need_mb) in rcu_read_lock_trace() 72 struct task_struct *t = current; in rcu_read_unlock_trace() local 75 nesting = READ_ONCE(t->trc_reader_nesting) - 1; in rcu_read_unlock_trace() 78 WRITE_ONCE(t->trc_reader_nesting, INT_MIN); in rcu_read_unlock_trace() 79 if (likely(!READ_ONCE(t->trc_reader_special.s)) || nesting) { in rcu_read_unlock_trace() 80 WRITE_ONCE(t->trc_reader_nesting, nesting); in rcu_read_unlock_trace() 83 rcu_read_unlock_trace_special(t, nesting); in rcu_read_unlock_trace()
|
D | kcov.h | 28 void kcov_task_init(struct task_struct *t); 29 void kcov_task_exit(struct task_struct *t); 31 #define kcov_prepare_switch(t) \ argument 33 (t)->kcov_mode |= KCOV_IN_CTXSW; \ 36 #define kcov_finish_switch(t) \ argument 38 (t)->kcov_mode &= ~KCOV_IN_CTXSW; \ 77 static inline void kcov_task_init(struct task_struct *t) {} in kcov_task_init() argument 78 static inline void kcov_task_exit(struct task_struct *t) {} in kcov_task_exit() argument 79 static inline void kcov_prepare_switch(struct task_struct *t) {} in kcov_prepare_switch() argument 80 static inline void kcov_finish_switch(struct task_struct *t) {} in kcov_finish_switch() argument
|
D | stackleak.h | 18 static inline void stackleak_task_init(struct task_struct *t) in stackleak_task_init() argument 20 t->lowest_stack = (unsigned long)end_of_stack(t) + sizeof(unsigned long); in stackleak_task_init() 22 t->prev_lowest_stack = t->lowest_stack; in stackleak_task_init() 32 static inline void stackleak_task_init(struct task_struct *t) { } in stackleak_task_init() argument
|
D | damon.h | 311 struct damon_target *t, struct damon_region *r, 314 struct damon_target *t, struct damon_region *r, 428 static inline struct damon_region *damon_last_region(struct damon_target *t) in damon_last_region() argument 430 return list_last_entry(&t->regions_list, struct damon_region, list); in damon_last_region() 433 #define damon_for_each_region(r, t) \ argument 434 list_for_each_entry(r, &t->regions_list, list) 436 #define damon_for_each_region_safe(r, next, t) \ argument 437 list_for_each_entry_safe(r, next, &t->regions_list, list) 439 #define damon_for_each_target(t, ctx) \ argument 440 list_for_each_entry(t, &(ctx)->adaptive_targets, list) [all …]
|
D | cb710.h | 51 #define CB710_PORT_ACCESSORS(t) \ argument 52 static inline void cb710_write_port_##t(struct cb710_slot *slot, \ 53 unsigned port, u##t value) \ 55 iowrite##t(value, slot->iobase + port); \ 58 static inline u##t cb710_read_port_##t(struct cb710_slot *slot, \ 61 return ioread##t(slot->iobase + port); \ 64 static inline void cb710_modify_port_##t(struct cb710_slot *slot, \ 65 unsigned port, u##t set, u##t clear) \ 67 iowrite##t( \ 68 (ioread##t(slot->iobase + port) & ~clear)|set, \
|
D | win_minmax.h | 13 u32 t; /* time measurement was taken */ member 27 static inline u32 minmax_reset(struct minmax *m, u32 t, u32 meas) in minmax_reset() argument 29 struct minmax_sample val = { .t = t, .v = meas }; in minmax_reset() 35 u32 minmax_running_max(struct minmax *m, u32 win, u32 t, u32 meas); 36 u32 minmax_running_min(struct minmax *m, u32 win, u32 t, u32 meas);
|
D | interrupt.h | 602 void (*callback)(struct tasklet_struct *t); 643 static inline int tasklet_trylock(struct tasklet_struct *t) in tasklet_trylock() argument 645 return !test_and_set_bit(TASKLET_STATE_RUN, &(t)->state); in tasklet_trylock() 648 void tasklet_unlock(struct tasklet_struct *t); 649 void tasklet_unlock_wait(struct tasklet_struct *t); 650 void tasklet_unlock_spin_wait(struct tasklet_struct *t); 653 static inline int tasklet_trylock(struct tasklet_struct *t) { return 1; } in tasklet_trylock() argument 654 static inline void tasklet_unlock(struct tasklet_struct *t) { } in tasklet_unlock() argument 655 static inline void tasklet_unlock_wait(struct tasklet_struct *t) { } in tasklet_unlock_wait() argument 656 static inline void tasklet_unlock_spin_wait(struct tasklet_struct *t) { } in tasklet_unlock_spin_wait() argument [all …]
|
D | thread_info.h | 145 #define read_task_thread_flags(t) \ argument 146 read_ti_thread_flags(task_thread_info(t)) 156 #define set_task_syscall_work(t, fl) \ argument 157 set_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work) 158 #define test_task_syscall_work(t, fl) \ argument 159 test_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work) 160 #define clear_task_syscall_work(t, fl) \ argument 161 clear_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work) 172 #define set_task_syscall_work(t, fl) \ argument 173 set_ti_thread_flag(task_thread_info(t), TIF_##fl) [all …]
|
D | nls.h | 62 static inline unsigned char nls_tolower(struct nls_table *t, unsigned char c) in nls_tolower() argument 64 unsigned char nc = t->charset2lower[c]; in nls_tolower() 69 static inline unsigned char nls_toupper(struct nls_table *t, unsigned char c) in nls_toupper() argument 71 unsigned char nc = t->charset2upper[c]; in nls_toupper() 76 static inline int nls_strnicmp(struct nls_table *t, const unsigned char *s1, in nls_strnicmp() argument 80 if (nls_tolower(t, *s1++) != nls_tolower(t, *s2++)) in nls_strnicmp()
|
D | lockdep.h | 361 # define lock_acquire(l, s, t, r, c, n, i) do { } while (0) argument 539 #define lock_acquire_exclusive(l, s, t, n, i) lock_acquire(l, s, t, 0, 1, n, i) argument 540 #define lock_acquire_shared(l, s, t, n, i) lock_acquire(l, s, t, 1, 1, n, i) argument 541 #define lock_acquire_shared_recursive(l, s, t, n, i) lock_acquire(l, s, t, 2, 1, n, i) argument 543 #define spin_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 544 #define spin_acquire_nest(l, s, t, n, i) lock_acquire_exclusive(l, s, t, n, i) argument 547 #define rwlock_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 548 #define rwlock_acquire_read(l, s, t, i) \ argument 551 lock_acquire_shared_recursive(l, s, t, NULL, i); \ 553 lock_acquire_shared(l, s, t, NULL, i); \ [all …]
|
/include/asm-generic/ |
D | ioctl.h | 8 #define _IOC_TYPECHECK(t) (sizeof(t)) argument 12 #define _IOC_TYPECHECK(t) \ argument 13 ((sizeof(t) == sizeof(t[1]) && \ 14 sizeof(t) < (1 << _IOC_SIZEBITS)) ? \ 15 sizeof(t) : __invalid_size_argument_for_IOC)
|
/include/scsi/ |
D | scsi_transport.h | 57 scsi_transport_reserve_target(struct scsi_transport_template * t, int space) in scsi_transport_reserve_target() argument 59 BUG_ON(t->target_private_offset != 0); in scsi_transport_reserve_target() 60 t->target_private_offset = ALIGN(t->target_size, sizeof(void *)); in scsi_transport_reserve_target() 61 t->target_size = t->target_private_offset + space; in scsi_transport_reserve_target() 64 scsi_transport_reserve_device(struct scsi_transport_template * t, int space) in scsi_transport_reserve_device() argument 66 BUG_ON(t->device_private_offset != 0); in scsi_transport_reserve_device() 67 t->device_private_offset = ALIGN(t->device_size, sizeof(void *)); in scsi_transport_reserve_device() 68 t->device_size = t->device_private_offset + space; in scsi_transport_reserve_device()
|
/include/linux/sunrpc/ |
D | sched.h | 137 #define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC) argument 138 #define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER) argument 139 #define RPC_IS_SOFT(t) ((t)->tk_flags & (RPC_TASK_SOFT|RPC_TASK_TIMEOUT)) argument 140 #define RPC_IS_SOFTCONN(t) ((t)->tk_flags & RPC_TASK_SOFTCONN) argument 141 #define RPC_WAS_SENT(t) ((t)->tk_flags & RPC_TASK_SENT) argument 142 #define RPC_IS_MOVEABLE(t) ((t)->tk_flags & RPC_TASK_MOVEABLE) argument 152 #define RPC_IS_RUNNING(t) test_bit(RPC_TASK_RUNNING, &(t)->tk_runstate) argument 153 #define rpc_set_running(t) set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate) argument 154 #define rpc_test_and_set_running(t) \ argument 155 test_and_set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate) [all …]
|
D | timer.h | 29 int *t; in rpc_set_timeo() local 32 t = &rt->ntimeouts[timer-1]; in rpc_set_timeo() 33 if (ntimeo < *t) { in rpc_set_timeo() 34 if (*t > 0) in rpc_set_timeo() 35 (*t)--; in rpc_set_timeo() 39 *t = ntimeo; in rpc_set_timeo()
|
/include/net/netfilter/ |
D | nf_conntrack_tuple.h | 82 static inline void nf_ct_dump_tuple_ip(const struct nf_conntrack_tuple *t) in nf_ct_dump_tuple_ip() argument 86 t, t->dst.protonum, in nf_ct_dump_tuple_ip() 87 &t->src.u3.ip, ntohs(t->src.u.all), in nf_ct_dump_tuple_ip() 88 &t->dst.u3.ip, ntohs(t->dst.u.all)); in nf_ct_dump_tuple_ip() 92 static inline void nf_ct_dump_tuple_ipv6(const struct nf_conntrack_tuple *t) in nf_ct_dump_tuple_ipv6() argument 96 t, t->dst.protonum, in nf_ct_dump_tuple_ipv6() 97 t->src.u3.all, ntohs(t->src.u.all), in nf_ct_dump_tuple_ipv6() 98 t->dst.u3.all, ntohs(t->dst.u.all)); in nf_ct_dump_tuple_ipv6() 102 static inline void nf_ct_dump_tuple(const struct nf_conntrack_tuple *t) in nf_ct_dump_tuple() argument 104 switch (t->src.l3num) { in nf_ct_dump_tuple() [all …]
|
/include/linux/sched/ |
D | task.h | 104 static inline struct task_struct *get_task_struct(struct task_struct *t) in get_task_struct() argument 106 refcount_inc(&t->usage); in get_task_struct() 107 return t; in get_task_struct() 110 extern void __put_task_struct(struct task_struct *t); 113 static inline void put_task_struct(struct task_struct *t) in put_task_struct() argument 115 if (!refcount_dec_and_test(&t->usage)) in put_task_struct() 139 call_rcu(&t->rcu, __put_task_struct_rcu_cb); in put_task_struct() 141 __put_task_struct(t); in put_task_struct() 144 static inline void put_task_struct_many(struct task_struct *t, int nr) in put_task_struct_many() argument 146 if (refcount_sub_and_test(nr, &t->usage)) in put_task_struct_many() [all …]
|
D | cputime.h | 21 extern void task_cputime(struct task_struct *t, 23 extern u64 task_gtime(struct task_struct *t); 25 static inline void task_cputime(struct task_struct *t, in task_cputime() argument 28 *utime = t->utime; in task_cputime() 29 *stime = t->stime; in task_cputime() 32 static inline u64 task_gtime(struct task_struct *t) in task_gtime() argument 34 return t->gtime; in task_gtime() 39 static inline void task_cputime_scaled(struct task_struct *t, in task_cputime_scaled() argument 43 *utimescaled = t->utimescaled; in task_cputime_scaled() 44 *stimescaled = t->stimescaled; in task_cputime_scaled() [all …]
|
/include/trace/hooks/ |
D | binder.h | 23 TP_PROTO(struct binder_transaction *t), 24 TP_ARGS(t)); 29 TP_PROTO(struct binder_transaction *t, struct task_struct *task), 30 TP_ARGS(t, task)); 32 TP_PROTO(struct binder_transaction *t, struct task_struct *task), 33 TP_ARGS(t, task)); 48 TP_PROTO(struct binder_proc *proc, struct binder_transaction *t, 50 TP_ARGS(proc, t, binder_th_task, pending_async, sync)); 52 TP_PROTO(struct binder_proc *proc, struct binder_transaction *t, 55 TP_ARGS(proc, t, thread, node_debug_id, pending_async, sync, skip)); [all …]
|
/include/net/sctp/ |
D | sctp.h | 150 struct sctp_transport *t, __u32 pmtu); 155 struct sctp_transport *t); 160 int sctp_hash_transport(struct sctp_transport *t); 161 void sctp_unhash_transport(struct sctp_transport *t); 566 static inline struct dst_entry *sctp_transport_dst_check(struct sctp_transport *t) in sctp_transport_dst_check() argument 568 if (t->dst && !dst_check(t->dst, t->dst_cookie)) in sctp_transport_dst_check() 569 sctp_transport_dst_release(t); in sctp_transport_dst_check() 571 return t->dst; in sctp_transport_dst_check() 578 const struct sctp_transport *t, in __sctp_mtu_payload() argument 585 if (sp->udp_port && (!t || t->encap_port)) in __sctp_mtu_payload() [all …]
|
/include/media/ |
D | v4l2-dv-timings.h | 22 struct v4l2_fract v4l2_calc_timeperframe(const struct v4l2_dv_timings *t); 37 typedef bool v4l2_check_dv_timings_fnc(const struct v4l2_dv_timings *t, void *handle); 51 bool v4l2_valid_dv_timings(const struct v4l2_dv_timings *t, 72 int v4l2_enum_dv_timings_cap(struct v4l2_enum_dv_timings *t, 95 bool v4l2_find_dv_timings_cap(struct v4l2_dv_timings *t, 109 bool v4l2_find_dv_timings_cea861_vic(struct v4l2_dv_timings *t, u8 vic); 135 const struct v4l2_dv_timings *t, bool detailed); 202 struct v4l2_fract v4l2_dv_timings_aspect_ratio(const struct v4l2_dv_timings *t);
|
/include/crypto/ |
D | gf128mul.h | 223 be128 t[256]; member 228 void gf128mul_4k_lle(be128 *a, const struct gf128mul_4k *t); 229 void gf128mul_4k_bbe(be128 *a, const struct gf128mul_4k *t); 231 static inline void gf128mul_free_4k(struct gf128mul_4k *t) in gf128mul_free_4k() argument 233 kfree_sensitive(t); in gf128mul_free_4k() 240 struct gf128mul_4k *t[16]; member 249 void gf128mul_free_64k(struct gf128mul_64k *t); 250 void gf128mul_64k_bbe(be128 *a, const struct gf128mul_64k *t);
|
/include/net/tc_act/ |
D | tc_tunnel_key.h | 30 struct tcf_tunnel_key *t = to_tunnel_key(a); in is_tcf_tunnel_set() local 33 params = rcu_dereference_protected(t->params, in is_tcf_tunnel_set() 44 struct tcf_tunnel_key *t = to_tunnel_key(a); in is_tcf_tunnel_release() local 47 params = rcu_dereference_protected(t->params, in is_tcf_tunnel_release() 58 struct tcf_tunnel_key *t = to_tunnel_key(a); in tcf_tunnel_info() local 61 params = rcu_dereference_protected(t->params, in tcf_tunnel_info()
|
/include/xen/ |
D | xenbus.h | 156 char **xenbus_directory(struct xenbus_transaction t, 158 void *xenbus_read(struct xenbus_transaction t, 160 int xenbus_write(struct xenbus_transaction t, 162 int xenbus_mkdir(struct xenbus_transaction t, 164 int xenbus_exists(struct xenbus_transaction t, 166 int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node); 167 int xenbus_transaction_start(struct xenbus_transaction *t); 168 int xenbus_transaction_end(struct xenbus_transaction t, int abort); 172 int xenbus_scanf(struct xenbus_transaction t, 181 int xenbus_printf(struct xenbus_transaction t, [all …]
|