| /net/rxrpc/ |
| A D | peer_object.c | 143 if (peer) in rxrpc_lookup_peer_rcu() 164 if (peer->max_data < peer->if_mtu - peer->hdrsize) { in rxrpc_assess_MTU_size() 165 trace_rxrpc_pmtud_reduce(peer, 0, peer->if_mtu - peer->hdrsize, in rxrpc_assess_MTU_size() 167 peer->max_data = peer->if_mtu - peer->hdrsize; in rxrpc_assess_MTU_size() 210 peer->max_data = umin(RXRPC_JUMBO(1), peer->if_mtu - peer->hdrsize); in rxrpc_assess_MTU_size() 212 peer->pmtud_bad = peer->if_mtu - peer->hdrsize + 1; in rxrpc_assess_MTU_size() 213 peer->pmtud_trial = umin(peer->max_data, peer->pmtud_bad - 1); in rxrpc_assess_MTU_size() 230 if (peer) { in rxrpc_alloc_peer() 278 peer->max_data = peer->if_mtu - peer->hdrsize; in rxrpc_init_peer() 375 if (peer) in rxrpc_lookup_peer() [all …]
|
| A D | peer_event.c | 108 if (mtu > 0 && peer->if_mtu == 65535 && mtu < peer->if_mtu) in rxrpc_adjust_mtu() 109 peer->if_mtu = mtu; in rxrpc_adjust_mtu() 113 mtu = peer->if_mtu; in rxrpc_adjust_mtu() 120 if (mtu < peer->hdrsize) in rxrpc_adjust_mtu() 155 if (peer && !rxrpc_get_peer_maybe(peer, rxrpc_peer_get_input_error)) in rxrpc_input_error() 156 peer = NULL; in rxrpc_input_error() 158 if (!peer) in rxrpc_input_error() 248 struct rxrpc_peer *peer; in rxrpc_peer_keepalive_dispatch() local 271 cursor, peer->debug_id, slot, &peer->srx.transport); in rxrpc_peer_keepalive_dispatch() 370 struct rxrpc_peer *peer = conn->peer; in rxrpc_input_probe_for_pmtud() local [all …]
|
| A D | conn_service.c | 41 p = rcu_dereference_raw(peer->service_conns.rb_node); in rxrpc_find_service_conn_rcu() 55 done_seqretry(&peer->service_conn_lock, seq); in rxrpc_find_service_conn_rcu() 71 write_seqlock(&peer->service_conn_lock); in rxrpc_publish_service_conn() 73 pp = &peer->service_conns.rb_node; in rxrpc_publish_service_conn() 92 write_sequnlock(&peer->service_conn_lock); in rxrpc_publish_service_conn() 99 write_sequnlock(&peer->service_conn_lock); in rxrpc_publish_service_conn() 111 &peer->service_conns); in rxrpc_publish_service_conn() 180 rxrpc_publish_service_conn(conn->peer, conn); in rxrpc_new_incoming_connection() 189 struct rxrpc_peer *peer = conn->peer; in rxrpc_unpublish_service_conn() local 191 write_seqlock(&peer->service_conn_lock); in rxrpc_unpublish_service_conn() [all …]
|
| A D | call_accept.c | 71 struct rxrpc_peer *peer; in rxrpc_service_prealloc_one() local 74 if (!peer) in rxrpc_service_prealloc_one() 76 b->peer_backlog[head] = peer; in rxrpc_service_prealloc_one() 200 kfree(peer); in rxrpc_discard_prealloc() 279 if (peer && !rxrpc_get_peer_maybe(peer, rxrpc_peer_get_service_conn)) in rxrpc_alloc_incoming_call() 280 peer = NULL; in rxrpc_alloc_incoming_call() 281 if (!peer) { in rxrpc_alloc_incoming_call() 283 peer->srx = *peer_srx; in rxrpc_alloc_incoming_call() 298 conn->peer = peer; in rxrpc_alloc_incoming_call() 317 call->peer = rxrpc_get_peer(conn->peer, rxrpc_peer_get_accept); in rxrpc_alloc_incoming_call() [all …]
|
| A D | output.c | 221 if_mtu = call->peer->if_mtu - call->peer->hdrsize; in rxrpc_fill_out_ack() 222 if (call->peer->ackr_adv_pmtud) { in rxrpc_fill_out_ack() 261 msg.msg_name = &call->peer->srx.transport; in rxrpc_send_ack_packet() 289 call->peer->pmtud_pending = false; in rxrpc_send_ack_packet() 290 call->peer->pmtud_probing = true; in rxrpc_send_ack_packet() 388 msg.msg_name = &call->peer->srx.transport; in rxrpc_send_abort_packet() 669 msg.msg_name = &call->peer->srx.transport; in rxrpc_send_data_packet() 760 msg.msg_name = &conn->peer->srx.transport; in rxrpc_send_conn_abort() 886 msg.msg_name = &peer->srx.transport; in rxrpc_send_keepalive() 887 msg.msg_namelen = peer->srx.transport_len; in rxrpc_send_keepalive() [all …]
|
| A D | conn_object.c | 103 struct rxrpc_peer *peer; in rxrpc_find_client_connection_rcu() local 120 peer = conn->peer; in rxrpc_find_client_connection_rcu() 123 if (peer->srx.transport.sin.sin_port != in rxrpc_find_client_connection_rcu() 129 if (peer->srx.transport.sin6.sin6_port != in rxrpc_find_client_connection_rcu() 198 call->peer->cong_ssthresh = call->cong_ssthresh; in rxrpc_disconnect_call() 201 spin_lock_irq(&call->peer->lock); in rxrpc_disconnect_call() 203 spin_unlock_irq(&call->peer->lock); in rxrpc_disconnect_call() 328 conn->peer->pmtud_probing = false; in rxrpc_clean_up_connection() 329 conn->peer->pmtud_pending = true; in rxrpc_clean_up_connection() 340 rxrpc_put_peer(conn->peer, rxrpc_peer_put_conn); in rxrpc_clean_up_connection()
|
| A D | proc.c | 165 sprintf(rbuff, "%pISpc", &conn->peer->srx.transport); in rxrpc_connection_seq_show() 280 struct rxrpc_peer *peer; in rxrpc_peer_seq_show() local 291 peer = list_entry(v, struct rxrpc_peer, hash_link); in rxrpc_peer_seq_show() 293 sprintf(lbuff, "%pISpc", &peer->local->srx.transport); in rxrpc_peer_seq_show() 295 sprintf(rbuff, "%pISpc", &peer->srx.transport); in rxrpc_peer_seq_show() 302 refcount_read(&peer->ref), in rxrpc_peer_seq_show() 303 peer->cong_ssthresh, in rxrpc_peer_seq_show() 304 peer->max_data, in rxrpc_peer_seq_show() 305 now - peer->last_tx_at, in rxrpc_peer_seq_show() 306 READ_ONCE(peer->recent_srtt_us), in rxrpc_peer_seq_show() [all …]
|
| A D | conn_client.c | 82 bundle->peer = rxrpc_get_peer(call->peer, rxrpc_peer_get_bundle); in rxrpc_alloc_bundle() 119 rxrpc_put_peer(bundle->peer, rxrpc_peer_put_bundle); in rxrpc_free_bundle() 178 conn->peer = rxrpc_get_peer(bundle->peer, rxrpc_peer_get_client_conn); in rxrpc_alloc_client_connection() 259 call->peer, key_serial(call->key), call->security_level, in rxrpc_look_up_bundle() 275 diff = (cmp(bundle->peer, call->peer) ?: in rxrpc_look_up_bundle() 304 diff = (cmp(bundle->peer, call->peer) ?: in rxrpc_look_up_bundle() 438 call->cong_ssthresh = call->peer->cong_ssthresh; in rxrpc_activate_one_channel() 550 spin_lock_irq(&call->peer->lock); in rxrpc_expose_client_call() 551 hlist_add_head(&call->error_link, &call->peer->error_targets); in rxrpc_expose_client_call() 552 spin_unlock_irq(&call->peer->lock); in rxrpc_expose_client_call()
|
| A D | io_thread.c | 200 struct rxrpc_peer *peer = NULL; in rxrpc_input_packet() local 321 peer = rxrpc_lookup_peer_rcu(local, &peer_srx); in rxrpc_input_packet() 322 if (!peer) { in rxrpc_input_packet() 327 conn = rxrpc_find_service_conn_rcu(peer, skb); in rxrpc_input_packet() 336 peer = rxrpc_get_peer_maybe(peer, rxrpc_peer_get_input); in rxrpc_input_packet() 339 ret = rxrpc_new_incoming_call(local, peer, NULL, &peer_srx, skb); in rxrpc_input_packet() 340 rxrpc_put_peer(peer, rxrpc_peer_put_input); in rxrpc_input_packet() 437 return rxrpc_new_incoming_call(conn->local, conn->peer, conn, in rxrpc_input_packet_on_conn()
|
| /net/tipc/ |
| A D | monitor.c | 179 peer = peer_prev(peer); in peer_head() 180 return peer; in peer_head() 289 peer = peer_nxt(peer); in mon_update_local_domain() 316 peer = peer_prev(peer); in mon_update_neighbors() 330 for (; peer != self; peer = peer_nxt(peer)) { in mon_assign_roles() 364 if (!peer) in tipc_mon_remove_peer() 397 *peer = p; in tipc_mon_add_peer() 430 if (!peer && !tipc_mon_add_peer(mon, addr, &peer)) in tipc_mon_peer_up() 524 if (!peer || !peer->is_up) in tipc_mon_rcv() 620 if (peer) { in tipc_mon_get_state() [all …]
|
| /net/sctp/ |
| A D | associola.c | 495 if (asoc->peer.retran_path == peer) in sctp_assoc_rm_peer() 508 if (asoc->peer.primary_path == peer) in sctp_assoc_rm_peer() 510 if (asoc->peer.active_path == peer) in sctp_assoc_rm_peer() 512 if (asoc->peer.retran_path == peer) in sctp_assoc_rm_peer() 514 if (asoc->peer.last_data_from == peer) in sctp_assoc_rm_peer() 728 asoc->peer.retran_path = peer; in sctp_assoc_add_peer() 733 asoc->peer.retran_path = peer; in sctp_assoc_add_peer() 1093 asoc->peer.rwnd = new->peer.rwnd; in sctp_assoc_update() 1094 asoc->peer.sack_needed = new->peer.sack_needed; in sctp_assoc_update() 1096 asoc->peer.i = new->peer.i; in sctp_assoc_update() [all …]
|
| A D | sm_make_chunk.c | 2399 if (asoc->peer.auth_capable && (!asoc->peer.peer_random || in sctp_process_init() 2410 (asoc->peer.asconf_capable && !asoc->peer.auth_capable)) { in sctp_process_init() 2429 asoc->peer.i.init_tag = in sctp_process_init() 2431 asoc->peer.i.a_rwnd = in sctp_process_init() 2461 asoc->peer.rwnd = asoc->peer.i.a_rwnd; in sctp_process_init() 2505 asoc->peer.addip_serial = asoc->peer.i.initial_tsn - 1; in sctp_process_init() 2622 asoc->peer.cookie_len = in sctp_process_param() 2625 asoc->peer.cookie = kmemdup(param.cookie->body, asoc->peer.cookie_len, gfp); in sctp_process_param() 3112 if (!peer) in sctp_process_asconf_param() 3156 if (!peer) in sctp_process_asconf_param() [all …]
|
| A D | transport.c | 41 struct sctp_transport *peer, in sctp_transport_init() argument 47 memcpy(&peer->ipaddr, addr, peer->af_specific->sockaddr_len); in sctp_transport_init() 50 peer->sack_generation = 0; in sctp_transport_init() 60 peer->last_time_heard = 0; in sctp_transport_init() 71 INIT_LIST_HEAD(&peer->transmitted); in sctp_transport_init() 72 INIT_LIST_HEAD(&peer->send_ready); in sctp_transport_init() 73 INIT_LIST_HEAD(&peer->transports); in sctp_transport_init() 83 get_random_bytes(&peer->hb_nonce, sizeof(peer->hb_nonce)); in sctp_transport_init() 85 refcount_set(&peer->refcnt, 1); in sctp_transport_init() 87 return peer; in sctp_transport_init() [all …]
|
| A D | sm_sideeffect.c | 149 asoc->peer.sack_needed = 1; in sctp_gen_sack() 165 asoc->peer.sack_needed = 1; in sctp_gen_sack() 175 if (!asoc->peer.sack_needed) { in sctp_gen_sack() 176 asoc->peer.sack_cnt++; in sctp_gen_sack() 186 asoc->peer.sack_needed = 1; in sctp_gen_sack() 193 asoc->peer.sack_needed = 1; in sctp_gen_sack() 212 asoc->peer.sack_needed = 0; in sctp_gen_sack() 213 asoc->peer.sack_cnt = 0; in sctp_gen_sack() 908 kfree(asoc->peer.cookie); in sctp_cmd_new_state() 909 asoc->peer.cookie = NULL; in sctp_cmd_new_state() [all …]
|
| A D | outqueue.c | 300 if (chunk->asoc->peer.prsctp_capable && in sctp_outq_tail() 772 const __u16 dport = asoc->peer.port; in sctp_packet_singleton() 773 const __u32 vtag = asoc->peer.i.init_tag; in sctp_packet_singleton() 870 ctx->asoc->peer.i.init_tag, in sctp_outq_select_transport() 871 ctx->asoc->peer.ecn_capable); in sctp_outq_select_transport() 1009 ctx->asoc->peer.ecn_capable); in sctp_outq_flush_rtx() 1378 if (asoc->peer.prsctp_capable && in sctp_outq_sack() 1399 asoc->peer.rwnd = sack_a_rwnd; in sctp_outq_sack() 1521 q->asoc->peer.primary_path->cacc. in sctp_check_transmitted() 1671 if (!q->asoc->peer.rwnd && in sctp_check_transmitted() [all …]
|
| A D | proc.c | 81 struct sctp_transport *peer; in sctp_seq_dump_local_addrs() local 88 peer = asoc->peer.primary_path; in sctp_seq_dump_local_addrs() 89 if (unlikely(peer == NULL)) { in sctp_seq_dump_local_addrs() 94 primary = &peer->saddr; in sctp_seq_dump_local_addrs() 119 primary = &assoc->peer.primary_addr; in sctp_seq_dump_remote_addrs() 120 list_for_each_entry_rcu(transport, &assoc->peer.transport_addr_list, in sctp_seq_dump_remote_addrs() 273 assoc->peer.port); in sctp_assocs_seq_show() 314 list_for_each_entry_rcu(tsp, &assoc->peer.transport_addr_list, in sctp_remaddr_seq_show()
|
| A D | auth.c | 236 return sctp_auth_make_key_vector(asoc->peer.peer_random, in sctp_auth_make_peer_vector() 237 asoc->peer.peer_chunks, in sctp_auth_make_peer_vector() 238 asoc->peer.peer_hmacs, in sctp_auth_make_peer_vector() 392 if (!asoc->peer.auth_capable) in sctp_auth_asoc_init_active_key() 538 hmacs = asoc->peer.peer_hmacs; in sctp_auth_asoc_get_hmac() 679 if (!asoc->peer.auth_capable) in sctp_auth_send_cid() 691 if (!asoc->peer.auth_capable) in sctp_auth_recv_cid() 830 if (!asoc->peer.auth_capable) in sctp_auth_set_key() 890 if (!asoc->peer.auth_capable) in sctp_auth_set_active_key() 935 if (!asoc->peer.auth_capable) in sctp_auth_del_key_id() [all …]
|
| /net/bluetooth/ |
| A D | 6lowpan.c | 114 return peer; in __peer_lookup_chan() 127 return peer; in __peer_lookup_conn() 172 &peer->chan->dst, peer->chan->dst_type, in peer_lookup_dst() 208 if (peer) in lookup_peer() 214 return peer; in lookup_peer() 352 if (!peer) in chan_recv_cb() 640 peer = kzalloc(sizeof(*peer), GFP_ATOMIC); in add_peer_chan() 641 if (!peer) in add_peer_chan() 917 if (!peer) in bt_6lowpan_disconnect() 920 BT_DBG("peer %p chan %p", peer, peer->chan); in bt_6lowpan_disconnect() [all …]
|
| /net/smc/ |
| A D | smc_cdc.h | 201 peer->count = htonl(save->count); in smc_host_cursor_to_cdc() 202 peer->wrap = htons(save->wrap); in smc_host_cursor_to_cdc() 213 peer->len = local->len; in smc_host_msg_to_cdc() 223 union smc_cdc_cursor *peer, in smc_cdc_cursor_to_host() argument 242 struct smc_cdc_msg *peer, in smcr_cdc_msg_to_host() argument 246 local->len = peer->len; in smcr_cdc_msg_to_host() 256 struct smcd_cdc_msg *peer, in smcd_cdc_msg_to_host() argument 261 temp.wrap = peer->prod.wrap; in smcd_cdc_msg_to_host() 262 temp.count = peer->prod.count; in smcd_cdc_msg_to_host() 265 temp.wrap = peer->cons.wrap; in smcd_cdc_msg_to_host() [all …]
|
| /net/core/ |
| A D | net_namespace.c | 340 put_net(peer); in peernet2id_alloc() 373 struct net *peer; in get_net_ns_by_id() local 380 if (peer) in get_net_ns_by_id() 381 peer = maybe_get_net(peer); in get_net_ns_by_id() 384 return peer; in get_net_ns_by_id() 856 struct net *peer; in rtnl_net_newid() local 879 if (IS_ERR(peer)) { in rtnl_net_newid() 907 put_net(peer); in rtnl_net_newid() 1020 if (!peer) in rtnl_net_getid() 1028 if (IS_ERR(peer)) { in rtnl_net_getid() [all …]
|
| /net/mac80211/ |
| A D | tdls.c | 213 rsp_addr = peer; in ieee80211_tdls_add_link_ie() 215 init_addr = peer; in ieee80211_tdls_add_link_ie() 437 sta = sta_info_get(sdata, peer); in ieee80211_tdls_add_setup_start_ies() 608 sta = sta_info_get(sdata, peer); in ieee80211_tdls_add_setup_cfm_ies() 754 skb, peer, in ieee80211_tdls_add_ies() 772 peer, initiator); in ieee80211_tdls_add_ies() 801 memcpy(tf->da, peer, ETH_ALEN); in ieee80211_prep_tdls_encap_data() 972 peer, link, action_code, in ieee80211_tdls_build_mgmt_packet_data() 999 const u8 *peer, int link_id, in ieee80211_tdls_prep_mgmt_packet() argument 1256 const u8 *peer, int link_id, in ieee80211_tdls_mgmt() argument [all …]
|
| /net/ipv4/ |
| A D | inetpeer.c | 247 bool inet_peer_xrlim_allow(struct inet_peer *peer, int timeout) in inet_peer_xrlim_allow() argument 252 if (!peer) in inet_peer_xrlim_allow() 255 token = otoken = READ_ONCE(peer->rate_tokens); in inet_peer_xrlim_allow() 257 delta = now - READ_ONCE(peer->rate_last); in inet_peer_xrlim_allow() 259 WRITE_ONCE(peer->rate_last, now); in inet_peer_xrlim_allow() 269 WRITE_ONCE(peer->rate_tokens, token); in inet_peer_xrlim_allow() 279 struct inet_peer *peer = rb_entry(p, struct inet_peer, rb_node); in inetpeer_invalidate_tree() local 282 rb_erase(&peer->rb_node, &base->rb_root); in inetpeer_invalidate_tree() 283 inet_putpeer(peer); in inetpeer_invalidate_tree()
|
| /net/wireless/ |
| A D | pmsr.c | 193 struct nlattr *peer, in pmsr_parse_peer() argument 209 NL_SET_ERR_MSG_ATTR(info->extack, peer, in pmsr_parse_peer() 274 struct nlattr *peers, *peer; in nl80211_pmsr_start() local 289 nla_for_each_nested(peer, peers, rem) { in nl80211_pmsr_start() 293 NL_SET_ERR_MSG_ATTR(info->extack, peer, in nl80211_pmsr_start() 326 nla_for_each_nested(peer, peers, rem) { in nl80211_pmsr_start() 328 err = pmsr_parse_peer(rdev, peer, &req->peers[idx], info); in nl80211_pmsr_start() 493 struct nlattr *pmsr, *peers, *peer, *resp, *data, *typedata; in nl80211_pmsr_send_result() local 503 peer = nla_nest_start_noflag(msg, 1); in nl80211_pmsr_send_result() 504 if (!peer) in nl80211_pmsr_send_result() [all …]
|
| A D | trace.h | 1833 MAC_ASSIGN(peer, peer); 1972 MAC_ASSIGN(peer, peer); 2062 MAC_ASSIGN(peer, peer); 2099 MAC_ASSIGN(peer, peer); 2543 MAC_ASSIGN(peer, peer); 2566 MAC_ASSIGN(peer, peer); 2916 MAC_ASSIGN(peer, tid_conf->peer); 2935 MAC_ASSIGN(peer, peer); 3648 MAC_ASSIGN(peer, peer); 3713 MAC_ASSIGN(peer, peer); [all …]
|
| /net/x25/ |
| A D | x25_forward.c | 93 struct net_device *peer = NULL; in x25_forward_data() local 103 peer = frwd->dev2; in x25_forward_data() 105 peer = frwd->dev1; in x25_forward_data() 112 if ( (nb = x25_get_neigh(peer)) == NULL) in x25_forward_data()
|