/linux-6.3-rc2/net/mptcp/ |
A D | sockopt.c | 87 ssk->sk_prot->keepalive(ssk, !!val); in mptcp_sol_socket_sync_intval() 606 lock_sock(ssk); in mptcp_setsockopt_sol_tcp_congestion() 640 lock_sock(ssk); in mptcp_setsockopt_sol_tcp_cork() 670 lock_sock(ssk); in mptcp_setsockopt_sol_tcp_nodelay() 825 struct sock *ssk; in mptcp_setsockopt() local 844 if (ssk) in mptcp_setsockopt() 865 struct sock *ssk; in mptcp_getsockopt_first_sf_only() local 869 if (ssk) { in mptcp_getsockopt_first_sf_only() 1238 if (ssk) in mptcp_getsockopt() 1257 ssk->sk_prot->keepalive(ssk, 1); in sync_socket_options() [all …]
|
A D | protocol.c | 1452 send_info[subflow->backup].ssk = ssk; in mptcp_subflow_get_send() 1473 ssk = send_info[SSK_MODE_ACTIVE].ssk; in mptcp_subflow_get_send() 1474 if (!ssk || !sk_stream_memory_free(ssk)) in mptcp_subflow_get_send() 1481 return ssk; in mptcp_subflow_get_send() 1490 return ssk; in mptcp_subflow_get_send() 1556 if (!ssk) in __mptcp_push_pending() 1584 if (ssk) in __mptcp_push_pending() 1644 tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle, in __mptcp_subflow_push_pending() 2521 if (!ssk) in __mptcp_retrans() 2540 tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle, in __mptcp_retrans() [all …]
|
A D | subflow.c | 405 tcp_done(ssk); in mptcp_subflow_reset() 1072 sk_eat_skb(ssk, skb); in get_mapping_status() 1202 sk_eat_skb(ssk, skb); in mptcp_subflow_discard_data() 1257 tcp_send_ack(ssk); in mptcp_subflow_fail() 1338 ssk->sk_err = EBADMSG; in subflow_check_data_avail() 1341 sk_eat_skb(ssk, skb); in subflow_check_data_avail() 1552 struct sock *ssk; in __mptcp_subflow_connect() local 1565 ssk = sf->sk; in __mptcp_subflow_connect() 1603 sock_hold(ssk); in __mptcp_subflow_connect() 1819 if (!sk_is_mptcp(ssk)) in mptcp_subflow_queue_clean() [all …]
|
A D | protocol.h | 629 void __mptcp_subflow_send_ack(struct sock *ssk); 630 void mptcp_subflow_reset(struct sock *ssk); 732 WRITE_ONCE(sk->sk_sndbuf, ssk->sk_sndbuf); in mptcp_propagate_sndbuf() 758 int mptcp_token_new_connect(struct sock *ssk); 843 const struct sock *ssk, gfp_t gfp); 846 void mptcp_event_pm_listener(const struct sock *ssk, 983 gfp_t saved_allocation = ssk->sk_allocation; in mptcp_do_fallback() 988 ssk->sk_allocation = GFP_ATOMIC; in mptcp_do_fallback() 989 ssk->sk_shutdown |= SEND_SHUTDOWN; in mptcp_do_fallback() 990 tcp_shutdown(ssk, SEND_SHUTDOWN); in mptcp_do_fallback() [all …]
|
A D | pm.c | 81 mptcp_event(MPTCP_EVENT_CREATED, msk, ssk, GFP_ATOMIC); in mptcp_pm_new_connection() 153 mptcp_event(MPTCP_EVENT_ESTABLISHED, msk, ssk, gfp); in mptcp_pm_fully_established() 202 void mptcp_pm_add_addr_received(const struct sock *ssk, in mptcp_pm_add_addr_received() argument 205 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk); in mptcp_pm_add_addr_received() 212 mptcp_event_addr_announced(ssk, addr); in mptcp_pm_add_addr_received() 277 void mptcp_pm_mp_prio_received(struct sock *ssk, u8 bkup) in mptcp_pm_mp_prio_received() argument 279 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk); in mptcp_pm_mp_prio_received() 295 mptcp_event(MPTCP_EVENT_SUB_PRIORITY, msk, ssk, GFP_ATOMIC); in mptcp_pm_mp_prio_received() 406 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk); in mptcp_pm_subflow_chk_stale() 407 u32 rcv_tstamp = READ_ONCE(tcp_sk(ssk)->rcv_tstamp); in mptcp_pm_subflow_chk_stale() [all …]
|
A D | pm_userspace.c | 327 struct sock *ssk; in mptcp_nl_find_ssk() local 329 ssk = mptcp_subflow_tcp_sock(subflow); in mptcp_nl_find_ssk() 331 if (local->family != ssk->sk_family) in mptcp_nl_find_ssk() 334 issk = inet_sk(ssk); in mptcp_nl_find_ssk() 336 switch (ssk->sk_family) { in mptcp_nl_find_ssk() 344 const struct ipv6_pinfo *pinfo = inet6_sk(ssk); in mptcp_nl_find_ssk() 358 return ssk; in mptcp_nl_find_ssk() 372 struct sock *sk, *ssk; in mptcp_nl_cmd_sf_destroy() local 420 ssk = mptcp_nl_find_ssk(msk, &addr_l, &addr_r); in mptcp_nl_cmd_sf_destroy() 421 if (ssk) { in mptcp_nl_cmd_sf_destroy() [all …]
|
A D | pm_netlink.c | 473 slow = lock_sock_fast(ssk); in __mptcp_pm_send_ack() 483 __mptcp_subflow_send_ack(ssk); in __mptcp_pm_send_ack() 484 unlock_sock_fast(ssk, slow); in __mptcp_pm_send_ack() 1956 switch (ssk->sk_family) { in mptcp_event_add_subflow() 1984 sf = mptcp_subflow_ctx(ssk); in mptcp_event_add_subflow() 1999 const struct sock *ssk) in mptcp_event_put_token_and_ssk() argument 2011 sf = mptcp_subflow_ctx(ssk); in mptcp_event_put_token_and_ssk() 2018 if (ssk->sk_bound_dev_if && in mptcp_event_put_token_and_ssk() 2022 sk_err = ssk->sk_err; in mptcp_event_put_token_and_ssk() 2039 const struct sock *ssk) in mptcp_event_sub_closed() argument [all …]
|
A D | fastopen.c | 12 struct sock *ssk = subflow->tcp_sock; in mptcp_fastopen_subflow_synack_set_params() local 17 tp = tcp_sk(ssk); in mptcp_fastopen_subflow_synack_set_params() 21 skb = skb_peek(&ssk->sk_receive_queue); in mptcp_fastopen_subflow_synack_set_params() 26 __skb_unlink(skb, &ssk->sk_receive_queue); in mptcp_fastopen_subflow_synack_set_params()
|
A D | options.c | 941 tcp_send_ack(ssk); in check_fully_established() 978 mptcp_do_fallback(ssk); in check_fully_established() 1001 clear_3rdack_retransmission(ssk); in check_fully_established() 1009 mptcp_subflow_reset(ssk); in check_fully_established() 1030 struct sock *ssk, in ack_update_msk() argument 1050 new_wnd_end = new_snd_una + tcp_sk(ssk)->snd_wnd; in ack_update_msk() 1057 __mptcp_check_push(sk, ssk); in ack_update_msk() 1239 const struct sock *ssk = (const struct sock *)tp; in mptcp_set_rwin() local 1246 subflow = mptcp_subflow_ctx(ssk); in mptcp_set_rwin() 1568 subflow = mptcp_subflow_ctx(ssk); in mptcp_write_options() [all …]
|
A D | token.c | 151 int mptcp_token_new_connect(struct sock *ssk) in mptcp_token_new_connect() argument 153 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk); in mptcp_token_new_connect() 173 ssk, subflow->local_key, subflow->token, subflow->idsn); in mptcp_token_new_connect()
|
/linux-6.3-rc2/include/trace/events/ |
A D | mptcp.h | 34 struct sock *ssk; 44 ssk = mptcp_subflow_tcp_sock(subflow); 45 if (ssk && sk_fullsock(ssk)) { 46 __entry->snd_wnd = tcp_sk(ssk)->snd_wnd; 47 __entry->pace = ssk->sk_pacing_rate; 53 if (ssk && sk_fullsock(ssk) && __entry->pace) 54 __entry->ratio = div_u64((u64)ssk->sk_wmem_queued << 32, __entry->pace);
|
/linux-6.3-rc2/include/linux/ |
A D | netlink.h | 206 int netlink_unicast(struct sock *ssk, struct sk_buff *skb, __u32 portid, int nonblock); 207 int netlink_broadcast(struct sock *ssk, struct sk_buff *skb, __u32 portid, 209 int netlink_set_err(struct sock *ssk, __u32 portid, __u32 group, int code); 216 long *timeo, struct sock *ssk); 298 int __netlink_dump_start(struct sock *ssk, struct sk_buff *skb, 301 static inline int netlink_dump_start(struct sock *ssk, struct sk_buff *skb, in netlink_dump_start() argument 308 return __netlink_dump_start(ssk, skb, nlh, control); in netlink_dump_start()
|
/linux-6.3-rc2/net/netlink/ |
A D | af_netlink.c | 1173 sock = netlink_lookup(sock_net(ssk), ssk->sk_protocol, portid); in netlink_getsockbyportid() 1240 long *timeo, struct sock *ssk) in netlink_attachskb() argument 1250 if (!ssk || netlink_is_kernel(ssk)) in netlink_attachskb() 1328 struct sock *ssk) in netlink_unicast_kernel() argument 1337 NETLINK_CB(skb).sk = ssk; in netlink_unicast_kernel() 1338 netlink_deliver_tap_kernel(sk, ssk, skb); in netlink_unicast_kernel() 1357 timeo = sock_sndtimeo(ssk, nonblock); in netlink_unicast() 1513 struct net *net = sock_net(ssk); in netlink_broadcast() 1519 info.exclude_sk = ssk; in netlink_broadcast() 1606 info.exclude_sk = ssk; in netlink_set_err() [all …]
|
/linux-6.3-rc2/include/net/ |
A D | mptcp.h | 117 void mptcp_space(const struct sock *ssk, int *space, int *full_space); 268 static inline void mptcp_space(const struct sock *ssk, int *s, int *fs) { } in mptcp_space() argument
|
/linux-6.3-rc2/tools/testing/selftests/bpf/progs/ |
A D | mptcp_sock.c | 82 const struct sock *ssk, int server_side) in BPF_PROG() argument
|