Home
last modified time | relevance | path

Searched refs:seq (Results 1 – 25 of 115) sorted by relevance

12345

/include/vdso/
Dhelpers.h12 u32 seq; in vdso_read_begin() local
14 while (unlikely((seq = READ_ONCE(vd->seq)) & 1)) in vdso_read_begin()
18 return seq; in vdso_read_begin()
24 u32 seq; in vdso_read_retry() local
27 seq = READ_ONCE(vd->seq); in vdso_read_retry()
28 return seq != start; in vdso_read_retry()
38 WRITE_ONCE(vd[CS_HRES_COARSE].seq, vd[CS_HRES_COARSE].seq + 1); in vdso_write_begin()
39 WRITE_ONCE(vd[CS_RAW].seq, vd[CS_RAW].seq + 1); in vdso_write_begin()
51 WRITE_ONCE(vd[CS_HRES_COARSE].seq, vd[CS_HRES_COARSE].seq + 1); in vdso_write_end()
52 WRITE_ONCE(vd[CS_RAW].seq, vd[CS_RAW].seq + 1); in vdso_write_end()
/include/trace/events/
Dscmi.h38 TP_PROTO(int transfer_id, u8 msg_id, u8 protocol_id, u16 seq,
40 TP_ARGS(transfer_id, msg_id, protocol_id, seq, poll),
46 __field(u16, seq)
54 __entry->seq = seq;
59 __entry->protocol_id, __entry->msg_id, __entry->seq,
64 TP_PROTO(int transfer_id, u8 msg_id, u8 protocol_id, u16 seq,
66 TP_ARGS(transfer_id, msg_id, protocol_id, seq, timeout, poll),
72 __field(u16, seq)
81 __entry->seq = seq;
87 __entry->protocol_id, __entry->msg_id, __entry->seq,
[all …]
Dksm.h20 TP_PROTO(int seq, u32 rmap_entries),
22 TP_ARGS(seq, rmap_entries),
25 __field(int, seq)
30 __entry->seq = seq;
35 __entry->seq, __entry->rmap_entries)
48 TP_PROTO(int seq, u32 rmap_entries),
50 TP_ARGS(seq, rmap_entries)
63 TP_PROTO(int seq, u32 rmap_entries),
65 TP_ARGS(seq, rmap_entries)
Drxrpc.h759 __entry->hdr.serial, __entry->hdr.seq,
784 u32 cid, u32 call_id, rxrpc_seq_t seq, int abort_code, int error),
786 TP_ARGS(call_nr, why, cid, call_id, seq, abort_code, error),
793 __field(rxrpc_seq_t, seq)
805 __entry->seq = seq;
810 __entry->cid, __entry->call_id, __entry->seq,
878 TP_PROTO(unsigned int call, rxrpc_seq_t seq,
881 TP_ARGS(call, seq, serial, flags),
885 __field(rxrpc_seq_t, seq)
892 __entry->seq = seq;
[all …]
/include/linux/
Drbtree_latch.h45 seqcount_latch_t seq; member
148 raw_write_seqcount_latch(&root->seq); in latch_tree_insert()
150 raw_write_seqcount_latch(&root->seq); in latch_tree_insert()
175 raw_write_seqcount_latch(&root->seq); in latch_tree_erase()
177 raw_write_seqcount_latch(&root->seq); in latch_tree_erase()
204 unsigned int seq; in latch_tree_find() local
207 seq = raw_read_seqcount_latch(&root->seq); in latch_tree_find()
208 node = __lt_find(key, root, seq & 1, ops->comp); in latch_tree_find()
209 } while (raw_read_seqcount_latch_retry(&root->seq, seq)); in latch_tree_find()
Dseq_file_net.h18 static inline struct net *seq_file_net(struct seq_file *seq) in seq_file_net() argument
21 return ((struct seq_net_private *)seq->private)->net; in seq_file_net()
31 static inline struct net *seq_file_single_net(struct seq_file *seq) in seq_file_single_net() argument
34 return (struct net *)seq->private; in seq_file_single_net()
Dmroute_base.h293 u32 portid, u32 seq, struct mr_mfc *c,
301 u32 portid, u32 seq, struct mr_mfc *c,
352 u32 portid, u32 seq, struct mr_mfc *c, in mr_rtm_dumproute() argument
395 void *mr_vif_seq_next(struct seq_file *seq, void *v, loff_t *pos);
397 static inline void *mr_vif_seq_start(struct seq_file *seq, loff_t *pos) in mr_vif_seq_start() argument
399 return *pos ? mr_vif_seq_idx(seq_file_net(seq), in mr_vif_seq_start()
400 seq->private, *pos - 1) in mr_vif_seq_start()
409 void *mr_mfc_seq_next(struct seq_file *seq, void *v,
412 static inline void *mr_mfc_seq_start(struct seq_file *seq, loff_t *pos, in mr_mfc_seq_start() argument
415 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_start()
[all …]
Du64_stats_sync.h66 seqcount_t seq; member
141 seqcount_init(&__s->seq); \
147 write_seqcount_begin(&syncp->seq); in __u64_stats_update_begin()
152 write_seqcount_end(&syncp->seq); in __u64_stats_update_end()
171 return read_seqcount_begin(&syncp->seq); in __u64_stats_fetch_begin()
177 return read_seqcount_retry(&syncp->seq, start); in __u64_stats_fetch_retry()
Dtrace_seq.h25 struct seq_buf seq; member
33 seq_buf_init(&s->seq, s->buffer, TRACE_SEQ_BUFFER_SIZE); in trace_seq_init()
53 return seq_buf_used(&s->seq); in trace_seq_used()
68 return s->buffer + seq_buf_used(&s->seq); in trace_seq_buffer_ptr()
80 return s->full || seq_buf_has_overflowed(&s->seq); in trace_seq_has_overflowed()
Dseqlock.h160 unsigned seq = smp_load_acquire(&s->seqcount.sequence); \
163 return seq; \
165 if (preemptible && unlikely(seq & 1)) { \
173 seq = smp_load_acquire(&s->seqcount.sequence); \
176 return seq; \
1127 static inline void read_seqbegin_or_lock(seqlock_t *lock, int *seq) in read_seqbegin_or_lock() argument
1129 if (!(*seq & 1)) /* Even */ in read_seqbegin_or_lock()
1130 *seq = read_seqbegin(lock); in read_seqbegin_or_lock()
1142 static inline int need_seqretry(seqlock_t *lock, int seq) in need_seqretry() argument
1144 return !(seq & 1) && read_seqretry(lock, seq); in need_seqretry()
[all …]
Dsched_clock.h36 extern struct clock_read_data *sched_clock_read_begin(unsigned int *seq);
37 extern int sched_clock_read_retry(unsigned int seq);
Dmmap_lock.h91 static inline bool mmap_lock_speculate_try_begin(struct mm_struct *mm, unsigned int *seq) in mmap_lock_speculate_try_begin() argument
99 return raw_seqcount_try_begin(&mm->mm_lock_seq, *seq); in mmap_lock_speculate_try_begin()
102 static inline bool mmap_lock_speculate_retry(struct mm_struct *mm, unsigned int seq) in mmap_lock_speculate_retry() argument
104 return read_seqcount_retry(&mm->mm_lock_seq, seq); in mmap_lock_speculate_retry()
124 static inline bool mmap_lock_speculate_try_begin(struct mm_struct *mm, unsigned int *seq) in mmap_lock_speculate_try_begin() argument
129 static inline bool mmap_lock_speculate_retry(struct mm_struct *mm, unsigned int seq) in mmap_lock_speculate_retry() argument
Dconnector.h45 u32 seq, group; member
51 u32 seq, groups; member
Dcpuset.h153 static inline bool read_mems_allowed_retry(unsigned int seq) in read_mems_allowed_retry() argument
158 return read_seqcount_retry(&current->mems_allowed_seq, seq); in read_mems_allowed_retry()
280 static inline bool read_mems_allowed_retry(unsigned int seq) in read_mems_allowed_retry() argument
/include/net/sctp/
Dcommand.h199 static inline int sctp_init_cmd_seq(struct sctp_cmd_seq *seq) in sctp_init_cmd_seq() argument
202 seq->last_used_slot = seq->cmds + SCTP_MAX_NUM_COMMANDS; in sctp_init_cmd_seq()
203 seq->next_cmd = seq->last_used_slot; in sctp_init_cmd_seq()
213 static inline void sctp_add_cmd_sf(struct sctp_cmd_seq *seq, in sctp_add_cmd_sf() argument
216 struct sctp_cmd *cmd = seq->last_used_slot - 1; in sctp_add_cmd_sf()
218 BUG_ON(cmd < seq->cmds); in sctp_add_cmd_sf()
222 seq->last_used_slot = cmd; in sctp_add_cmd_sf()
228 static inline struct sctp_cmd *sctp_next_cmd(struct sctp_cmd_seq *seq) in sctp_next_cmd() argument
230 if (seq->next_cmd <= seq->last_used_slot) in sctp_next_cmd()
233 return --seq->next_cmd; in sctp_next_cmd()
/include/net/
Draw.h66 static inline struct raw_iter_state *raw_seq_private(struct seq_file *seq) in raw_seq_private() argument
68 return seq->private; in raw_seq_private()
70 void *raw_seq_start(struct seq_file *seq, loff_t *pos);
71 void *raw_seq_next(struct seq_file *seq, void *v, loff_t *pos);
72 void raw_seq_stop(struct seq_file *seq, void *v);
Dtransp_v6.h47 void __ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp,
50 ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, __u16 srcp, in ip6_dgram_sock_seq_show() argument
53 __ip6_dgram_sock_seq_show(seq, sp, srcp, destp, sk_rmem_alloc_get(sp), in ip6_dgram_sock_seq_show()
Dping.h75 void *ping_seq_start(struct seq_file *seq, loff_t *pos, sa_family_t family);
76 void *ping_seq_next(struct seq_file *seq, void *v, loff_t *pos);
77 void ping_seq_stop(struct seq_file *seq, void *v);
Dtls.h283 struct sock *sk, u32 seq, u8 *rcd_sn,
342 u32 seq, u64 *p_record_sn);
452 static inline void tls_offload_rx_resync_request(struct sock *sk, __be32 seq) in tls_offload_rx_resync_request() argument
457 atomic64_set(&rx_ctx->resync_req, ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_request()
462 tls_offload_rx_resync_async_request_start(struct sock *sk, __be32 seq, u16 len) in tls_offload_rx_resync_async_request_start() argument
467 atomic64_set(&rx_ctx->resync_async->req, ((u64)ntohl(seq) << 32) | in tls_offload_rx_resync_async_request_start()
474 tls_offload_rx_resync_async_request_end(struct sock *sk, __be32 seq) in tls_offload_rx_resync_async_request_end() argument
480 ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_async_request_end()
Dgtp.h25 __be16 seq; member
44 __be16 seq; member
Dlib80211.h71 int (*set_key) (void *key, int len, u8 * seq, void *priv);
72 int (*get_key) (void *key, int len, u8 * seq, void *priv);
/include/net/netfilter/
Dnf_conntrack_seqadj.h21 struct nf_ct_seqadj seq[IP_CT_DIR_MAX]; member
37 __be32 seq, s32 off);
43 s32 nf_ct_seq_offset(const struct nf_conn *ct, enum ip_conntrack_dir, u32 seq);
/include/drm/
Ddrm_mipi_dsi.h393 #define mipi_dsi_generic_write_seq(dsi, seq...) \ argument
395 static const u8 d[] = { seq }; \
411 #define mipi_dsi_generic_write_seq_multi(ctx, seq...) \ argument
413 static const u8 d[] = { seq }; \
430 #define mipi_dsi_dcs_write_seq(dsi, cmd, seq...) \ argument
432 static const u8 d[] = { cmd, seq }; \
449 #define mipi_dsi_dcs_write_seq_multi(ctx, cmd, seq...) \ argument
451 static const u8 d[] = { cmd, seq }; \
/include/net/netns/
Dnexthop.h16 unsigned int seq; /* protected by rtnl_mutex */ member
/include/uapi/asm-generic/
Dipcbuf.h31 unsigned short seq; member

12345