| /net/batman-adv/ |
| A D | fragmentation.c | 107 u16 seqno) in batadv_frag_init_chain() argument 111 if (chain->seqno == seqno) in batadv_frag_init_chain() 118 chain->seqno = seqno; in batadv_frag_init_chain() 145 u16 seqno, hdr_size = sizeof(struct batadv_frag_packet); in batadv_frag_insert_packet() local 156 seqno = ntohs(frag_packet->seqno); in batadv_frag_insert_packet() 157 bucket = seqno % BATADV_FRAG_BUFFER_COUNT; in batadv_frag_insert_packet() 172 if (batadv_frag_init_chain(chain, seqno)) { in batadv_frag_insert_packet() 487 frag_header.seqno = htons(atomic_inc_return(&bat_priv->frag_seqno)); in batadv_frag_send_packet()
|
| A D | tp_meter.c | 606 icmp->seqno = htonl(seqno); in batadv_tp_send_msg() 654 if (batadv_seq_before(ntohl(icmp->seqno), in batadv_tp_recv_ack() 674 recv_ack = ntohl(icmp->seqno); in batadv_tp_recv_ack() 1194 icmp->seqno = htonl(seq); in batadv_tp_send_ack() 1237 new->seqno = ntohl(icmp->seqno); in batadv_tp_handle_out_of_order() 1257 if (new->seqno == un->seqno) { in batadv_tp_handle_out_of_order() 1266 if (batadv_seq_before(new->seqno, un->seqno)) in batadv_tp_handle_out_of_order() 1387 u32 seqno; in batadv_tp_recv_msg() local 1391 seqno = ntohl(icmp->seqno); in batadv_tp_recv_msg() 1395 if (seqno == BATADV_TP_FIRST_SEQ) { in batadv_tp_recv_msg() [all …]
|
| A D | bat_v_ogm.c | 297 ogm_packet->seqno = htonl(atomic_read(&bat_priv->bat_v.ogm_seqno)); in batadv_v_ogm_send_meshif() 334 ogm_packet->orig, ntohl(ogm_packet->seqno), in batadv_v_ogm_send_meshif() 546 if (orig_ifinfo->last_seqno_forwarded == ntohl(ogm_received->seqno)) in batadv_v_ogm_forward() 549 orig_ifinfo->last_seqno_forwarded = ntohl(ogm_received->seqno); in batadv_v_ogm_forward() 621 seq_diff = ntohl(ogm2->seqno) - orig_ifinfo->last_real_seqno; in batadv_v_ogm_metric_update() 647 orig_ifinfo->last_real_seqno = ntohl(ogm2->seqno); in batadv_v_ogm_metric_update() 658 neigh_ifinfo->bat_v.last_seqno = ntohl(ogm2->seqno); in batadv_v_ogm_metric_update() 870 ntohl(ogm_packet->seqno), ogm_throughput, ogm_packet->ttl, in batadv_v_ogm_process()
|
| A D | bat_iv_ogm.c | 373 ntohl(batadv_ogm_packet->seqno), in batadv_iv_ogm_send_to_if() 795 u32 seqno; in batadv_iv_ogm_schedule_buff() local 830 seqno = (u32)atomic_read(&hard_iface->bat_iv.ogm_seqno); in batadv_iv_ogm_schedule_buff() 831 batadv_ogm_packet->seqno = htonl(seqno); in batadv_iv_ogm_schedule_buff() 1226 u32 seqno = ntohl(batadv_ogm_packet->seqno); in batadv_iv_ogm_update_seqnos() local 1242 seq_diff = seqno - orig_ifinfo->last_real_seqno; in batadv_iv_ogm_update_seqnos() 1263 seqno); in batadv_iv_ogm_update_seqnos() 1292 orig_ifinfo->last_real_seqno, seqno); in batadv_iv_ogm_update_seqnos() 1293 orig_ifinfo->last_real_seqno = seqno; in batadv_iv_ogm_update_seqnos() 1539 bit_pos -= ntohl(ogm_packet->seqno); in batadv_iv_ogm_process_reply() [all …]
|
| A D | bat_v_elp.c | 316 elp_packet->seqno = htonl(atomic_read(&hard_iface->bat_v.elp_seqno)); in batadv_v_elp_periodic_work() 519 seqno_diff = ntohl(elp_packet->seqno) - elp_latest_seqno; in batadv_v_elp_neigh_update() 529 hardif_neigh->bat_v.elp_latest_seqno = ntohl(elp_packet->seqno); in batadv_v_elp_neigh_update() 576 ethhdr->h_source, ntohl(elp_packet->seqno), in batadv_v_elp_packet_recv()
|
| A D | routing.c | 1174 u32 seqno; in batadv_recv_bcast_packet() local 1211 seqno = ntohl(bcast_packet->seqno); in batadv_recv_bcast_packet() 1214 seqno)) in batadv_recv_bcast_packet() 1217 seq_diff = seqno - orig_node->last_bcast_seqno; in batadv_recv_bcast_packet() 1229 orig_node->last_bcast_seqno = seqno; in batadv_recv_bcast_packet()
|
| A D | mesh-interface.c | 186 u32 seqno; in batadv_interface_tx() local 341 seqno = atomic_inc_return(&bat_priv->bcast_seqno); in batadv_interface_tx() 342 bcast_packet->seqno = htonl(seqno); in batadv_interface_tx()
|
| A D | types.h | 300 u16 seqno; member 1401 u32 seqno; member
|
| A D | network-coding.c | 773 if (last_real_seqno != ntohl(ogm_packet->seqno)) in batadv_can_nc_with_orig()
|
| /net/tipc/ |
| A D | name_distr.c | 151 u32 dnode, struct list_head *pls, u16 seqno) in named_distribute() argument 197 msg_set_named_seqno(hdr, seqno); in named_distribute() 211 u16 seqno; in tipc_named_node_up() local 217 seqno = nt->snd_nxt; in tipc_named_node_up() 221 named_distribute(net, &head, dnode, &nt->cluster_scope, seqno); in tipc_named_node_up() 318 u16 seqno; in tipc_named_dequeue() local 328 seqno = msg_named_seqno(hdr); in tipc_named_dequeue() 330 *rcv_nxt = seqno; in tipc_named_dequeue() 340 if (*open && (*rcv_nxt == seqno)) { in tipc_named_dequeue() 347 if (less(seqno, *rcv_nxt)) { in tipc_named_dequeue()
|
| A D | link.c | 1022 u16 seqno = l->snd_nxt; in tipc_link_xmit() local 1079 seqno++; in tipc_link_xmit() 1103 l->snd_nxt = seqno; in tipc_link_xmit() 1156 u16 seqno = l->snd_nxt; in tipc_link_advance_backlog() local 1184 seqno++; in tipc_link_advance_backlog() 1186 l->snd_nxt = seqno; in tipc_link_advance_backlog() 1353 u16 seqno; in tipc_link_tnl_rcv() local 1453 u16 expect, seqno = 0; in __tipc_build_gap_ack_blks() local 1475 expect = seqno + 1; in __tipc_build_gap_ack_blks() 1549 u16 seqno, n = 0; in tipc_link_advance_transmq() local [all …]
|
| A D | msg.c | 823 bool __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno, in __tipc_skb_queue_sorted() argument 828 if (skb_queue_empty(list) || less(seqno, buf_seqno(skb_peek(list)))) { in __tipc_skb_queue_sorted() 833 if (more(seqno, buf_seqno(skb_peek_tail(list)))) { in __tipc_skb_queue_sorted() 839 if (more(seqno, buf_seqno(_skb))) in __tipc_skb_queue_sorted() 841 if (seqno == buf_seqno(_skb)) in __tipc_skb_queue_sorted()
|
| A D | crypto.c | 164 atomic64_t seqno ____cacheline_aligned; 606 atomic64_set(&tmp->seqno, 0); in tipc_aead_init() 660 atomic64_set(&aead->seqno, 0); in tipc_aead_clone() 798 memcpy(iv + 4, (u8 *)&ehdr->seqno, 8); in tipc_aead_encrypt() 932 memcpy(iv + 4, (u8 *)&ehdr->seqno, 8); in tipc_aead_decrypt() 1046 u64 seqno; in tipc_ehdr_build() local 1059 seqno = atomic64_inc_return(&aead->seqno); in tipc_ehdr_build() 1061 seqno = atomic64_inc_return(&__rx->sndnxt); in tipc_ehdr_build() 1064 if (unlikely(!seqno)) in tipc_ehdr_build() 1068 ehdr->seqno = cpu_to_be64(seqno); in tipc_ehdr_build()
|
| A D | crypto.h | 147 __be64 seqno; member
|
| A D | msg.h | 1180 bool __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno, 1299 u16 seqno) in __tipc_skb_dequeue() argument 1303 if (skb && less_eq(buf_seqno(skb), seqno)) { in __tipc_skb_dequeue()
|
| A D | group.c | 670 u32 event, u16 seqno, in tipc_group_create_event() argument 697 msg_set_grp_bc_seqno(hdr, seqno); in tipc_group_create_event()
|
| /net/smc/ |
| A D | smc_cdc.h | 44 __be16 seqno; member 214 peer->seqno = htons(local->seqno); in smc_host_msg_to_cdc() 247 local->seqno = ntohs(peer->seqno); in smcr_cdc_msg_to_host()
|
| A D | smc_cdc.c | 123 conn->local_tx_ctrl.seqno = conn->tx_cdc_seq; in smc_cdc_msg_send() 135 conn->local_tx_ctrl.seqno = conn->tx_cdc_seq; in smc_cdc_msg_send() 155 peer->seqno = htons(conn->tx_cdc_seq_fin); /* seqno last compl. tx */ in smcr_cdc_msg_send_validation() 312 u16 recv_seq = ntohs(cdc->seqno); in smc_cdc_msg_validate() 316 diff = conn->local_rx_ctrl.seqno - recv_seq; in smc_cdc_msg_validate() 495 if (smc_cdc_before(ntohs(cdc->seqno), in smc_cdc_rx_handler() 496 conn->local_rx_ctrl.seqno)) in smc_cdc_rx_handler()
|
| A D | smc.h | 172 u16 seqno; /* connection seq # */ member
|
| /net/sunrpc/auth_gss/ |
| A D | auth_gss.c | 1539 u32 seqno; in gss_marshal() local 1551 seqno = (ctx->gc_seq < MAXSEQ) ? ctx->gc_seq++ : MAXSEQ; in gss_marshal() 1552 xprt_rqst_add_seqno(req, seqno); in gss_marshal() 1675 gss_validate_seqno_mic(struct gss_cl_ctx *ctx, u32 seqno, __be32 *seq, __be32 *p, u32 len) in gss_validate_seqno_mic() argument 1681 *seq = cpu_to_be32(seqno); in gss_validate_seqno_mic() 1993 u32 len, offset, seqno, maj_stat; in gss_unwrap_resp_integ() local 2006 if (xdr_stream_decode_u32(xdr, &seqno)) in gss_unwrap_resp_integ() 2008 if (seqno != *rqstp->rq_seqnos) in gss_unwrap_resp_integ() 2052 trace_rpcgss_bad_seqno(task, *rqstp->rq_seqnos, seqno); in gss_unwrap_resp_integ()
|
| /net/ipv6/ |
| A D | esp6_offload.c | 365 esp.seqno = cpu_to_be64(xo->seq.low + ((u64)xo->seq.hi << 32)); in esp6_xmit()
|
| A D | esp6.c | 613 memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8), in esp6_output_tail() 688 esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low + in esp6_output()
|
| /net/ipv4/ |
| A D | esp4_offload.c | 335 esp.seqno = cpu_to_be64(seq + ((u64)xo->seq.hi << 32)); in esp_xmit()
|
| A D | esp4.c | 582 memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8), in esp_output_tail() 656 esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low + in esp_output()
|
| /net/9p/ |
| A D | client.c | 980 static atomic_t seqno = ATOMIC_INIT(0); in p9_client_create() local 1041 "9p-fcall-cache-%u", atomic_inc_return(&seqno)); in p9_client_create()
|