| /include/vdso/ |
| D | helpers.h | 12 u32 seq; in vdso_read_begin() local 14 while (unlikely((seq = READ_ONCE(vd->seq)) & 1)) in vdso_read_begin() 18 return seq; in vdso_read_begin() 24 u32 seq; in vdso_read_retry() local 27 seq = READ_ONCE(vd->seq); in vdso_read_retry() 28 return seq != start; in vdso_read_retry() 38 WRITE_ONCE(vd[CS_HRES_COARSE].seq, vd[CS_HRES_COARSE].seq + 1); in vdso_write_begin() 39 WRITE_ONCE(vd[CS_RAW].seq, vd[CS_RAW].seq + 1); in vdso_write_begin() 51 WRITE_ONCE(vd[CS_HRES_COARSE].seq, vd[CS_HRES_COARSE].seq + 1); in vdso_write_end() 52 WRITE_ONCE(vd[CS_RAW].seq, vd[CS_RAW].seq + 1); in vdso_write_end()
|
| /include/trace/events/ |
| D | scmi.h | 38 TP_PROTO(int transfer_id, u8 msg_id, u8 protocol_id, u16 seq, 40 TP_ARGS(transfer_id, msg_id, protocol_id, seq, poll), 46 __field(u16, seq) 54 __entry->seq = seq; 59 __entry->protocol_id, __entry->msg_id, __entry->seq, 64 TP_PROTO(int transfer_id, u8 msg_id, u8 protocol_id, u16 seq, 66 TP_ARGS(transfer_id, msg_id, protocol_id, seq, timeout, poll), 72 __field(u16, seq) 81 __entry->seq = seq; 87 __entry->protocol_id, __entry->msg_id, __entry->seq, [all …]
|
| D | ksm.h | 20 TP_PROTO(int seq, u32 rmap_entries), 22 TP_ARGS(seq, rmap_entries), 25 __field(int, seq) 30 __entry->seq = seq; 35 __entry->seq, __entry->rmap_entries) 48 TP_PROTO(int seq, u32 rmap_entries), 50 TP_ARGS(seq, rmap_entries) 63 TP_PROTO(int seq, u32 rmap_entries), 65 TP_ARGS(seq, rmap_entries)
|
| D | rxrpc.h | 759 __entry->hdr.serial, __entry->hdr.seq, 784 u32 cid, u32 call_id, rxrpc_seq_t seq, int abort_code, int error), 786 TP_ARGS(call_nr, why, cid, call_id, seq, abort_code, error), 793 __field(rxrpc_seq_t, seq) 805 __entry->seq = seq; 810 __entry->cid, __entry->call_id, __entry->seq, 878 TP_PROTO(unsigned int call, rxrpc_seq_t seq, 881 TP_ARGS(call, seq, serial, flags), 885 __field(rxrpc_seq_t, seq) 892 __entry->seq = seq; [all …]
|
| /include/linux/ |
| D | rbtree_latch.h | 45 seqcount_latch_t seq; member 148 raw_write_seqcount_latch(&root->seq); in latch_tree_insert() 150 raw_write_seqcount_latch(&root->seq); in latch_tree_insert() 175 raw_write_seqcount_latch(&root->seq); in latch_tree_erase() 177 raw_write_seqcount_latch(&root->seq); in latch_tree_erase() 204 unsigned int seq; in latch_tree_find() local 207 seq = raw_read_seqcount_latch(&root->seq); in latch_tree_find() 208 node = __lt_find(key, root, seq & 1, ops->comp); in latch_tree_find() 209 } while (raw_read_seqcount_latch_retry(&root->seq, seq)); in latch_tree_find()
|
| D | seq_file_net.h | 18 static inline struct net *seq_file_net(struct seq_file *seq) in seq_file_net() argument 21 return ((struct seq_net_private *)seq->private)->net; in seq_file_net() 31 static inline struct net *seq_file_single_net(struct seq_file *seq) in seq_file_single_net() argument 34 return (struct net *)seq->private; in seq_file_single_net()
|
| D | mroute_base.h | 293 u32 portid, u32 seq, struct mr_mfc *c, 301 u32 portid, u32 seq, struct mr_mfc *c, 352 u32 portid, u32 seq, struct mr_mfc *c, in mr_rtm_dumproute() argument 395 void *mr_vif_seq_next(struct seq_file *seq, void *v, loff_t *pos); 397 static inline void *mr_vif_seq_start(struct seq_file *seq, loff_t *pos) in mr_vif_seq_start() argument 399 return *pos ? mr_vif_seq_idx(seq_file_net(seq), in mr_vif_seq_start() 400 seq->private, *pos - 1) in mr_vif_seq_start() 409 void *mr_mfc_seq_next(struct seq_file *seq, void *v, 412 static inline void *mr_mfc_seq_start(struct seq_file *seq, loff_t *pos, in mr_mfc_seq_start() argument 415 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_start() [all …]
|
| D | u64_stats_sync.h | 66 seqcount_t seq; member 141 seqcount_init(&__s->seq); \ 147 write_seqcount_begin(&syncp->seq); in __u64_stats_update_begin() 152 write_seqcount_end(&syncp->seq); in __u64_stats_update_end() 171 return read_seqcount_begin(&syncp->seq); in __u64_stats_fetch_begin() 177 return read_seqcount_retry(&syncp->seq, start); in __u64_stats_fetch_retry()
|
| D | trace_seq.h | 25 struct seq_buf seq; member 33 seq_buf_init(&s->seq, s->buffer, TRACE_SEQ_BUFFER_SIZE); in trace_seq_init() 53 return seq_buf_used(&s->seq); in trace_seq_used() 68 return s->buffer + seq_buf_used(&s->seq); in trace_seq_buffer_ptr() 80 return s->full || seq_buf_has_overflowed(&s->seq); in trace_seq_has_overflowed()
|
| D | seqlock.h | 160 unsigned seq = smp_load_acquire(&s->seqcount.sequence); \ 163 return seq; \ 165 if (preemptible && unlikely(seq & 1)) { \ 173 seq = smp_load_acquire(&s->seqcount.sequence); \ 176 return seq; \ 1127 static inline void read_seqbegin_or_lock(seqlock_t *lock, int *seq) in read_seqbegin_or_lock() argument 1129 if (!(*seq & 1)) /* Even */ in read_seqbegin_or_lock() 1130 *seq = read_seqbegin(lock); in read_seqbegin_or_lock() 1142 static inline int need_seqretry(seqlock_t *lock, int seq) in need_seqretry() argument 1144 return !(seq & 1) && read_seqretry(lock, seq); in need_seqretry() [all …]
|
| D | sched_clock.h | 36 extern struct clock_read_data *sched_clock_read_begin(unsigned int *seq); 37 extern int sched_clock_read_retry(unsigned int seq);
|
| D | mmap_lock.h | 91 static inline bool mmap_lock_speculate_try_begin(struct mm_struct *mm, unsigned int *seq) in mmap_lock_speculate_try_begin() argument 99 return raw_seqcount_try_begin(&mm->mm_lock_seq, *seq); in mmap_lock_speculate_try_begin() 102 static inline bool mmap_lock_speculate_retry(struct mm_struct *mm, unsigned int seq) in mmap_lock_speculate_retry() argument 104 return read_seqcount_retry(&mm->mm_lock_seq, seq); in mmap_lock_speculate_retry() 124 static inline bool mmap_lock_speculate_try_begin(struct mm_struct *mm, unsigned int *seq) in mmap_lock_speculate_try_begin() argument 129 static inline bool mmap_lock_speculate_retry(struct mm_struct *mm, unsigned int seq) in mmap_lock_speculate_retry() argument
|
| D | connector.h | 45 u32 seq, group; member 51 u32 seq, groups; member
|
| D | cpuset.h | 153 static inline bool read_mems_allowed_retry(unsigned int seq) in read_mems_allowed_retry() argument 158 return read_seqcount_retry(¤t->mems_allowed_seq, seq); in read_mems_allowed_retry() 280 static inline bool read_mems_allowed_retry(unsigned int seq) in read_mems_allowed_retry() argument
|
| /include/net/sctp/ |
| D | command.h | 199 static inline int sctp_init_cmd_seq(struct sctp_cmd_seq *seq) in sctp_init_cmd_seq() argument 202 seq->last_used_slot = seq->cmds + SCTP_MAX_NUM_COMMANDS; in sctp_init_cmd_seq() 203 seq->next_cmd = seq->last_used_slot; in sctp_init_cmd_seq() 213 static inline void sctp_add_cmd_sf(struct sctp_cmd_seq *seq, in sctp_add_cmd_sf() argument 216 struct sctp_cmd *cmd = seq->last_used_slot - 1; in sctp_add_cmd_sf() 218 BUG_ON(cmd < seq->cmds); in sctp_add_cmd_sf() 222 seq->last_used_slot = cmd; in sctp_add_cmd_sf() 228 static inline struct sctp_cmd *sctp_next_cmd(struct sctp_cmd_seq *seq) in sctp_next_cmd() argument 230 if (seq->next_cmd <= seq->last_used_slot) in sctp_next_cmd() 233 return --seq->next_cmd; in sctp_next_cmd()
|
| /include/net/ |
| D | raw.h | 66 static inline struct raw_iter_state *raw_seq_private(struct seq_file *seq) in raw_seq_private() argument 68 return seq->private; in raw_seq_private() 70 void *raw_seq_start(struct seq_file *seq, loff_t *pos); 71 void *raw_seq_next(struct seq_file *seq, void *v, loff_t *pos); 72 void raw_seq_stop(struct seq_file *seq, void *v);
|
| D | transp_v6.h | 47 void __ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, 50 ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, __u16 srcp, in ip6_dgram_sock_seq_show() argument 53 __ip6_dgram_sock_seq_show(seq, sp, srcp, destp, sk_rmem_alloc_get(sp), in ip6_dgram_sock_seq_show()
|
| D | ping.h | 75 void *ping_seq_start(struct seq_file *seq, loff_t *pos, sa_family_t family); 76 void *ping_seq_next(struct seq_file *seq, void *v, loff_t *pos); 77 void ping_seq_stop(struct seq_file *seq, void *v);
|
| D | tls.h | 283 struct sock *sk, u32 seq, u8 *rcd_sn, 342 u32 seq, u64 *p_record_sn); 452 static inline void tls_offload_rx_resync_request(struct sock *sk, __be32 seq) in tls_offload_rx_resync_request() argument 457 atomic64_set(&rx_ctx->resync_req, ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_request() 462 tls_offload_rx_resync_async_request_start(struct sock *sk, __be32 seq, u16 len) in tls_offload_rx_resync_async_request_start() argument 467 atomic64_set(&rx_ctx->resync_async->req, ((u64)ntohl(seq) << 32) | in tls_offload_rx_resync_async_request_start() 474 tls_offload_rx_resync_async_request_end(struct sock *sk, __be32 seq) in tls_offload_rx_resync_async_request_end() argument 480 ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_async_request_end()
|
| D | gtp.h | 25 __be16 seq; member 44 __be16 seq; member
|
| D | lib80211.h | 71 int (*set_key) (void *key, int len, u8 * seq, void *priv); 72 int (*get_key) (void *key, int len, u8 * seq, void *priv);
|
| /include/net/netfilter/ |
| D | nf_conntrack_seqadj.h | 21 struct nf_ct_seqadj seq[IP_CT_DIR_MAX]; member 37 __be32 seq, s32 off); 43 s32 nf_ct_seq_offset(const struct nf_conn *ct, enum ip_conntrack_dir, u32 seq);
|
| /include/drm/ |
| D | drm_mipi_dsi.h | 393 #define mipi_dsi_generic_write_seq(dsi, seq...) \ argument 395 static const u8 d[] = { seq }; \ 411 #define mipi_dsi_generic_write_seq_multi(ctx, seq...) \ argument 413 static const u8 d[] = { seq }; \ 430 #define mipi_dsi_dcs_write_seq(dsi, cmd, seq...) \ argument 432 static const u8 d[] = { cmd, seq }; \ 449 #define mipi_dsi_dcs_write_seq_multi(ctx, cmd, seq...) \ argument 451 static const u8 d[] = { cmd, seq }; \
|
| /include/net/netns/ |
| D | nexthop.h | 16 unsigned int seq; /* protected by rtnl_mutex */ member
|
| /include/uapi/asm-generic/ |
| D | ipcbuf.h | 31 unsigned short seq; member
|