Home
last modified time | relevance | path

Searched refs:seq (Results 1 – 25 of 117) sorted by relevance

12345

/include/vdso/
A Dhelpers.h12 u32 seq; in vdso_read_begin() local
14 while (unlikely((seq = READ_ONCE(vc->seq)) & 1)) in vdso_read_begin()
18 return seq; in vdso_read_begin()
24 u32 seq; in vdso_read_retry() local
27 seq = READ_ONCE(vc->seq); in vdso_read_retry()
28 return seq != start; in vdso_read_retry()
38 WRITE_ONCE(vc->seq, vc->seq + 1); in vdso_write_seq_begin()
48 WRITE_ONCE(vc->seq, vc->seq + 1); in vdso_write_seq_end()
/include/trace/events/
A Dscmi.h46 __field(u16, seq)
55 __entry->seq = seq;
74 __field(u16, seq)
83 __entry->seq = seq;
102 __field(u16, seq)
111 __entry->seq = seq;
130 __field(u16, seq)
138 __entry->seq = seq;
149 unsigned char *tag, u16 seq, int status,
160 __field(u16, seq)
[all …]
A Dksm.h20 TP_PROTO(int seq, u32 rmap_entries),
22 TP_ARGS(seq, rmap_entries),
25 __field(int, seq)
30 __entry->seq = seq;
35 __entry->seq, __entry->rmap_entries)
48 TP_PROTO(int seq, u32 rmap_entries),
50 TP_ARGS(seq, rmap_entries)
63 TP_PROTO(int seq, u32 rmap_entries),
65 TP_ARGS(seq, rmap_entries)
A Damdxdna.h33 TP_PROTO(struct drm_sched_job *sched_job, const char *name, const char *str, u64 seq),
35 TP_ARGS(sched_job, name, str, seq),
41 __field(u64, seq)),
47 __entry->seq = seq;),
51 __get_str(name), __entry->seq,
A Drxrpc.h957 __entry->seq = seq;
1084 __entry->seq = seq;
1111 __entry->seq = seq;
1430 __entry->seq = seq;
1506 __entry->seq = seq;
1563 __entry->seq = seq;
2150 __entry->seq = seq;
2318 __entry->seq = seq;
2346 __entry->seq = seq;
2375 __entry->seq = seq;
[all …]
/include/linux/
A Drbtree_latch.h45 seqcount_latch_t seq; member
148 write_seqcount_latch_begin(&root->seq); in latch_tree_insert()
150 write_seqcount_latch(&root->seq); in latch_tree_insert()
152 write_seqcount_latch_end(&root->seq); in latch_tree_insert()
176 write_seqcount_latch_begin(&root->seq); in latch_tree_erase()
178 write_seqcount_latch(&root->seq); in latch_tree_erase()
180 write_seqcount_latch_end(&root->seq); in latch_tree_erase()
206 unsigned int seq; in latch_tree_find() local
209 seq = read_seqcount_latch(&root->seq); in latch_tree_find()
210 node = __lt_find(key, root, seq & 1, ops->comp); in latch_tree_find()
[all …]
A Dseq_file_net.h18 static inline struct net *seq_file_net(struct seq_file *seq) in seq_file_net() argument
21 return ((struct seq_net_private *)seq->private)->net; in seq_file_net()
31 static inline struct net *seq_file_single_net(struct seq_file *seq) in seq_file_single_net() argument
34 return (struct net *)seq->private; in seq_file_single_net()
A Dfs_struct.h11 seqlock_t seq; member
28 read_seqlock_excl(&fs->seq); in get_fs_root()
31 read_sequnlock_excl(&fs->seq); in get_fs_root()
36 read_seqlock_excl(&fs->seq); in get_fs_pwd()
39 read_sequnlock_excl(&fs->seq); in get_fs_pwd()
A Dmroute_base.h298 u32 portid, u32 seq, struct mr_mfc *c,
306 u32 portid, u32 seq, struct mr_mfc *c,
357 u32 portid, u32 seq, struct mr_mfc *c, in mr_rtm_dumproute() argument
404 return *pos ? mr_vif_seq_idx(seq_file_net(seq), in mr_vif_seq_start()
405 seq->private, *pos - 1) in mr_vif_seq_start()
414 void *mr_mfc_seq_next(struct seq_file *seq, void *v,
420 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_start()
426 return *pos ? mr_mfc_seq_idx(seq_file_net(seq), in mr_mfc_seq_start()
427 seq->private, *pos - 1) in mr_mfc_seq_start()
433 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_stop()
[all …]
A Du64_stats_sync.h66 seqcount_t seq; member
141 seqcount_init(&__s->seq); \
147 write_seqcount_begin(&syncp->seq); in __u64_stats_update_begin()
152 write_seqcount_end(&syncp->seq); in __u64_stats_update_end()
171 return read_seqcount_begin(&syncp->seq); in __u64_stats_fetch_begin()
177 return read_seqcount_retry(&syncp->seq, start); in __u64_stats_fetch_retry()
A Dtrace_seq.h25 struct seq_buf seq; member
33 seq_buf_init(&s->seq, s->buffer, TRACE_SEQ_BUFFER_SIZE); in trace_seq_init()
53 return seq_buf_used(&s->seq); in trace_seq_used()
68 return s->buffer + seq_buf_used(&s->seq); in trace_seq_buffer_ptr()
80 return s->full || seq_buf_has_overflowed(&s->seq); in trace_seq_has_overflowed()
A Dseqlock.h163 return seq; \
165 if (preemptible && unlikely(seq & 1)) { \
173 seq = smp_load_acquire(&s->seqcount.sequence); \
176 return seq; \
1129 if (!(*seq & 1)) /* Even */ in read_seqbegin_or_lock()
1130 *seq = read_seqbegin(lock); in read_seqbegin_or_lock()
1144 return !(seq & 1) && read_seqretry(lock, seq); in need_seqretry()
1157 if (seq & 1) in done_seqretry()
1186 if (!(*seq & 1)) /* Even */ in read_seqbegin_or_lock_irqsave()
1187 *seq = read_seqbegin(lock); in read_seqbegin_or_lock_irqsave()
[all …]
A Dsched_clock.h36 extern struct clock_read_data *sched_clock_read_begin(unsigned int *seq);
37 extern int sched_clock_read_retry(unsigned int seq);
A Dconnector.h45 u32 seq, group; member
51 u32 seq, groups; member
A Dipc.h22 unsigned long seq; member
/include/net/sctp/
A Dcommand.h199 static inline int sctp_init_cmd_seq(struct sctp_cmd_seq *seq) in sctp_init_cmd_seq() argument
202 seq->last_used_slot = seq->cmds + SCTP_MAX_NUM_COMMANDS; in sctp_init_cmd_seq()
203 seq->next_cmd = seq->last_used_slot; in sctp_init_cmd_seq()
213 static inline void sctp_add_cmd_sf(struct sctp_cmd_seq *seq, in sctp_add_cmd_sf() argument
216 struct sctp_cmd *cmd = seq->last_used_slot - 1; in sctp_add_cmd_sf()
218 BUG_ON(cmd < seq->cmds); in sctp_add_cmd_sf()
222 seq->last_used_slot = cmd; in sctp_add_cmd_sf()
228 static inline struct sctp_cmd *sctp_next_cmd(struct sctp_cmd_seq *seq) in sctp_next_cmd() argument
230 if (seq->next_cmd <= seq->last_used_slot) in sctp_next_cmd()
233 return --seq->next_cmd; in sctp_next_cmd()
/include/net/
A Draw.h66 static inline struct raw_iter_state *raw_seq_private(struct seq_file *seq) in raw_seq_private() argument
68 return seq->private; in raw_seq_private()
70 void *raw_seq_start(struct seq_file *seq, loff_t *pos);
71 void *raw_seq_next(struct seq_file *seq, void *v, loff_t *pos);
72 void raw_seq_stop(struct seq_file *seq, void *v);
A Dtransp_v6.h47 void __ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp,
50 ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, __u16 srcp, in ip6_dgram_sock_seq_show() argument
53 __ip6_dgram_sock_seq_show(seq, sp, srcp, destp, sk_rmem_alloc_get(sp), in ip6_dgram_sock_seq_show()
A Dping.h75 void *ping_seq_start(struct seq_file *seq, loff_t *pos, sa_family_t family);
76 void *ping_seq_next(struct seq_file *seq, void *v, loff_t *pos);
77 void ping_seq_stop(struct seq_file *seq, void *v);
A Dtls.h281 struct sock *sk, u32 seq, u8 *rcd_sn,
334 u32 seq, u64 *p_record_sn);
444 static inline void tls_offload_rx_resync_request(struct sock *sk, __be32 seq) in tls_offload_rx_resync_request() argument
449 atomic64_set(&rx_ctx->resync_req, ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_request()
454 tls_offload_rx_resync_async_request_start(struct sock *sk, __be32 seq, u16 len) in tls_offload_rx_resync_async_request_start() argument
459 atomic64_set(&rx_ctx->resync_async->req, ((u64)ntohl(seq) << 32) | in tls_offload_rx_resync_async_request_start()
466 tls_offload_rx_resync_async_request_end(struct sock *sk, __be32 seq) in tls_offload_rx_resync_async_request_end() argument
472 ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_async_request_end()
A Dgtp.h25 __be16 seq; member
44 __be16 seq; member
A Dpptp.h22 __be32 seq; member
/include/net/netfilter/
A Dnf_conntrack_seqadj.h21 struct nf_ct_seqadj seq[IP_CT_DIR_MAX]; member
37 __be32 seq, s32 off);
43 s32 nf_ct_seq_offset(const struct nf_conn *ct, enum ip_conntrack_dir, u32 seq);
/include/net/netns/
A Dnexthop.h16 unsigned int seq; /* protected by rtnl_mutex */ member
/include/uapi/asm-generic/
A Dipcbuf.h31 unsigned short seq; member

Completed in 48 milliseconds

12345