/include/vdso/ |
A D | helpers.h | 12 u32 seq; in vdso_read_begin() local 14 while (unlikely((seq = READ_ONCE(vc->seq)) & 1)) in vdso_read_begin() 18 return seq; in vdso_read_begin() 24 u32 seq; in vdso_read_retry() local 27 seq = READ_ONCE(vc->seq); in vdso_read_retry() 28 return seq != start; in vdso_read_retry() 38 WRITE_ONCE(vc->seq, vc->seq + 1); in vdso_write_seq_begin() 48 WRITE_ONCE(vc->seq, vc->seq + 1); in vdso_write_seq_end()
|
/include/trace/events/ |
A D | scmi.h | 46 __field(u16, seq) 55 __entry->seq = seq; 74 __field(u16, seq) 83 __entry->seq = seq; 102 __field(u16, seq) 111 __entry->seq = seq; 130 __field(u16, seq) 138 __entry->seq = seq; 149 unsigned char *tag, u16 seq, int status, 160 __field(u16, seq) [all …]
|
A D | ksm.h | 20 TP_PROTO(int seq, u32 rmap_entries), 22 TP_ARGS(seq, rmap_entries), 25 __field(int, seq) 30 __entry->seq = seq; 35 __entry->seq, __entry->rmap_entries) 48 TP_PROTO(int seq, u32 rmap_entries), 50 TP_ARGS(seq, rmap_entries) 63 TP_PROTO(int seq, u32 rmap_entries), 65 TP_ARGS(seq, rmap_entries)
|
A D | amdxdna.h | 33 TP_PROTO(struct drm_sched_job *sched_job, const char *name, const char *str, u64 seq), 35 TP_ARGS(sched_job, name, str, seq), 41 __field(u64, seq)), 47 __entry->seq = seq;), 51 __get_str(name), __entry->seq,
|
A D | rxrpc.h | 957 __entry->seq = seq; 1084 __entry->seq = seq; 1111 __entry->seq = seq; 1430 __entry->seq = seq; 1506 __entry->seq = seq; 1563 __entry->seq = seq; 2150 __entry->seq = seq; 2318 __entry->seq = seq; 2346 __entry->seq = seq; 2375 __entry->seq = seq; [all …]
|
/include/linux/ |
A D | rbtree_latch.h | 45 seqcount_latch_t seq; member 148 write_seqcount_latch_begin(&root->seq); in latch_tree_insert() 150 write_seqcount_latch(&root->seq); in latch_tree_insert() 152 write_seqcount_latch_end(&root->seq); in latch_tree_insert() 176 write_seqcount_latch_begin(&root->seq); in latch_tree_erase() 178 write_seqcount_latch(&root->seq); in latch_tree_erase() 180 write_seqcount_latch_end(&root->seq); in latch_tree_erase() 206 unsigned int seq; in latch_tree_find() local 209 seq = read_seqcount_latch(&root->seq); in latch_tree_find() 210 node = __lt_find(key, root, seq & 1, ops->comp); in latch_tree_find() [all …]
|
A D | seq_file_net.h | 18 static inline struct net *seq_file_net(struct seq_file *seq) in seq_file_net() argument 21 return ((struct seq_net_private *)seq->private)->net; in seq_file_net() 31 static inline struct net *seq_file_single_net(struct seq_file *seq) in seq_file_single_net() argument 34 return (struct net *)seq->private; in seq_file_single_net()
|
A D | fs_struct.h | 11 seqlock_t seq; member 28 read_seqlock_excl(&fs->seq); in get_fs_root() 31 read_sequnlock_excl(&fs->seq); in get_fs_root() 36 read_seqlock_excl(&fs->seq); in get_fs_pwd() 39 read_sequnlock_excl(&fs->seq); in get_fs_pwd()
|
A D | mroute_base.h | 298 u32 portid, u32 seq, struct mr_mfc *c, 306 u32 portid, u32 seq, struct mr_mfc *c, 357 u32 portid, u32 seq, struct mr_mfc *c, in mr_rtm_dumproute() argument 404 return *pos ? mr_vif_seq_idx(seq_file_net(seq), in mr_vif_seq_start() 405 seq->private, *pos - 1) in mr_vif_seq_start() 414 void *mr_mfc_seq_next(struct seq_file *seq, void *v, 420 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_start() 426 return *pos ? mr_mfc_seq_idx(seq_file_net(seq), in mr_mfc_seq_start() 427 seq->private, *pos - 1) in mr_mfc_seq_start() 433 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_stop() [all …]
|
A D | u64_stats_sync.h | 66 seqcount_t seq; member 141 seqcount_init(&__s->seq); \ 147 write_seqcount_begin(&syncp->seq); in __u64_stats_update_begin() 152 write_seqcount_end(&syncp->seq); in __u64_stats_update_end() 171 return read_seqcount_begin(&syncp->seq); in __u64_stats_fetch_begin() 177 return read_seqcount_retry(&syncp->seq, start); in __u64_stats_fetch_retry()
|
A D | trace_seq.h | 25 struct seq_buf seq; member 33 seq_buf_init(&s->seq, s->buffer, TRACE_SEQ_BUFFER_SIZE); in trace_seq_init() 53 return seq_buf_used(&s->seq); in trace_seq_used() 68 return s->buffer + seq_buf_used(&s->seq); in trace_seq_buffer_ptr() 80 return s->full || seq_buf_has_overflowed(&s->seq); in trace_seq_has_overflowed()
|
A D | seqlock.h | 163 return seq; \ 165 if (preemptible && unlikely(seq & 1)) { \ 173 seq = smp_load_acquire(&s->seqcount.sequence); \ 176 return seq; \ 1129 if (!(*seq & 1)) /* Even */ in read_seqbegin_or_lock() 1130 *seq = read_seqbegin(lock); in read_seqbegin_or_lock() 1144 return !(seq & 1) && read_seqretry(lock, seq); in need_seqretry() 1157 if (seq & 1) in done_seqretry() 1186 if (!(*seq & 1)) /* Even */ in read_seqbegin_or_lock_irqsave() 1187 *seq = read_seqbegin(lock); in read_seqbegin_or_lock_irqsave() [all …]
|
A D | sched_clock.h | 36 extern struct clock_read_data *sched_clock_read_begin(unsigned int *seq); 37 extern int sched_clock_read_retry(unsigned int seq);
|
A D | connector.h | 45 u32 seq, group; member 51 u32 seq, groups; member
|
A D | ipc.h | 22 unsigned long seq; member
|
/include/net/sctp/ |
A D | command.h | 199 static inline int sctp_init_cmd_seq(struct sctp_cmd_seq *seq) in sctp_init_cmd_seq() argument 202 seq->last_used_slot = seq->cmds + SCTP_MAX_NUM_COMMANDS; in sctp_init_cmd_seq() 203 seq->next_cmd = seq->last_used_slot; in sctp_init_cmd_seq() 213 static inline void sctp_add_cmd_sf(struct sctp_cmd_seq *seq, in sctp_add_cmd_sf() argument 216 struct sctp_cmd *cmd = seq->last_used_slot - 1; in sctp_add_cmd_sf() 218 BUG_ON(cmd < seq->cmds); in sctp_add_cmd_sf() 222 seq->last_used_slot = cmd; in sctp_add_cmd_sf() 228 static inline struct sctp_cmd *sctp_next_cmd(struct sctp_cmd_seq *seq) in sctp_next_cmd() argument 230 if (seq->next_cmd <= seq->last_used_slot) in sctp_next_cmd() 233 return --seq->next_cmd; in sctp_next_cmd()
|
/include/net/ |
A D | raw.h | 66 static inline struct raw_iter_state *raw_seq_private(struct seq_file *seq) in raw_seq_private() argument 68 return seq->private; in raw_seq_private() 70 void *raw_seq_start(struct seq_file *seq, loff_t *pos); 71 void *raw_seq_next(struct seq_file *seq, void *v, loff_t *pos); 72 void raw_seq_stop(struct seq_file *seq, void *v);
|
A D | transp_v6.h | 47 void __ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, 50 ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, __u16 srcp, in ip6_dgram_sock_seq_show() argument 53 __ip6_dgram_sock_seq_show(seq, sp, srcp, destp, sk_rmem_alloc_get(sp), in ip6_dgram_sock_seq_show()
|
A D | ping.h | 75 void *ping_seq_start(struct seq_file *seq, loff_t *pos, sa_family_t family); 76 void *ping_seq_next(struct seq_file *seq, void *v, loff_t *pos); 77 void ping_seq_stop(struct seq_file *seq, void *v);
|
A D | tls.h | 281 struct sock *sk, u32 seq, u8 *rcd_sn, 334 u32 seq, u64 *p_record_sn); 444 static inline void tls_offload_rx_resync_request(struct sock *sk, __be32 seq) in tls_offload_rx_resync_request() argument 449 atomic64_set(&rx_ctx->resync_req, ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_request() 454 tls_offload_rx_resync_async_request_start(struct sock *sk, __be32 seq, u16 len) in tls_offload_rx_resync_async_request_start() argument 459 atomic64_set(&rx_ctx->resync_async->req, ((u64)ntohl(seq) << 32) | in tls_offload_rx_resync_async_request_start() 466 tls_offload_rx_resync_async_request_end(struct sock *sk, __be32 seq) in tls_offload_rx_resync_async_request_end() argument 472 ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_async_request_end()
|
A D | gtp.h | 25 __be16 seq; member 44 __be16 seq; member
|
A D | pptp.h | 22 __be32 seq; member
|
/include/net/netfilter/ |
A D | nf_conntrack_seqadj.h | 21 struct nf_ct_seqadj seq[IP_CT_DIR_MAX]; member 37 __be32 seq, s32 off); 43 s32 nf_ct_seq_offset(const struct nf_conn *ct, enum ip_conntrack_dir, u32 seq);
|
/include/net/netns/ |
A D | nexthop.h | 16 unsigned int seq; /* protected by rtnl_mutex */ member
|
/include/uapi/asm-generic/ |
A D | ipcbuf.h | 31 unsigned short seq; member
|