| /net/ipv4/ |
| A D | tcp_timer.c | 36 return icsk->icsk_rto; in tcp_clamp_rto_to_user_timeout() 250 if (icsk->icsk_retransmits) in tcp_write_timeout() 264 tcp_mtu_probing(icsk, sk); in tcp_write_timeout() 335 icsk->icsk_ack.ato = min_t(u32, icsk->icsk_ack.ato << 1, icsk->icsk_rto); in tcp_delack_timer_handler() 395 icsk->icsk_probes_out = 0; in tcp_probe_timer() 447 icsk->icsk_retransmits++; in tcp_update_rto_stats() 667 icsk->icsk_backoff = 0; in tcp_retransmit_timer() 677 icsk->icsk_backoff++; in tcp_retransmit_timer() 678 icsk->icsk_rto = min(icsk->icsk_rto << 1, tcp_rto_max(sk)); in tcp_retransmit_timer() 697 !icsk->icsk_pending) in tcp_write_timer_handler() [all …]
|
| A D | tcp_cong.c | 43 if (icsk->icsk_ca_ops->set_state) in tcp_set_ca_state() 45 icsk->icsk_ca_state = ca_state; in tcp_set_ca_state() 225 icsk->icsk_ca_ops = ca; in tcp_assign_congestion_control() 228 memset(icsk->icsk_ca_priv, 0, sizeof(icsk->icsk_ca_priv)); in tcp_assign_congestion_control() 240 if (icsk->icsk_ca_ops->init) in tcp_init_congestion_control() 241 icsk->icsk_ca_ops->init(sk); in tcp_init_congestion_control() 255 icsk->icsk_ca_ops = ca; in tcp_reinit_congestion_control() 256 icsk->icsk_ca_setsockopt = 1; in tcp_reinit_congestion_control() 257 memset(icsk->icsk_ca_priv, 0, sizeof(icsk->icsk_ca_priv)); in tcp_reinit_congestion_control() 273 if (icsk->icsk_ca_initialized && icsk->icsk_ca_ops->release) in tcp_cleanup_congestion_control() [all …]
|
| A D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() local 108 if (icsk->icsk_ulp_ops->update) in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() local 120 if (!icsk->icsk_ulp_ops) in tcp_cleanup_ulp() 123 if (icsk->icsk_ulp_ops->release) in tcp_cleanup_ulp() 124 icsk->icsk_ulp_ops->release(sk); in tcp_cleanup_ulp() 125 module_put(icsk->icsk_ulp_ops->owner); in tcp_cleanup_ulp() 127 icsk->icsk_ulp_ops = NULL; in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp() local 136 if (icsk->icsk_ulp_ops) in __tcp_set_ulp() [all …]
|
| A D | inet_connection_sock.c | 765 icsk->icsk_pending = icsk->icsk_ack.pending = 0; in inet_csk_init_xmit_timers() 772 smp_store_release(&icsk->icsk_pending, 0); in inet_csk_clear_xmit_timers() 773 smp_store_release(&icsk->icsk_ack.pending, 0); in inet_csk_clear_xmit_timers() 776 sk_stop_timer(sk, &icsk->icsk_delack_timer); in inet_csk_clear_xmit_timers() 787 smp_store_release(&icsk->icsk_pending, 0); in inet_csk_clear_xmit_timers_sync() 1059 struct inet_connection_sock *icsk; in reqsk_timer_handler() local 1088 icsk = inet_csk(sk_listener); in reqsk_timer_handler() 1109 queue = &icsk->icsk_accept_queue; in reqsk_timer_handler() 1202 if (!icsk->icsk_ulp_ops) in inet_clone_ulp() 1324 if (icsk->icsk_ulp_ops && !icsk->icsk_ulp_ops->clone) in inet_ulp_can_listen() [all …]
|
| A D | tcp_output.c | 177 if ((u32)(now - icsk->icsk_ack.lrcvtime) < icsk->icsk_ack.ato) in tcp_event_data_sent() 2468 interval = icsk->icsk_mtup.search_high - icsk->icsk_mtup.search_low; in tcp_mtu_probe() 3121 icsk->icsk_ack.quick = 0; in __tcp_select_window() 3180 icsk->icsk_ack.quick = 0; in __tcp_select_window() 3881 bpf_module_put(icsk->icsk_ca_ops, icsk->icsk_ca_ops->owner); in tcp_ca_dst_init() 3883 icsk->icsk_ca_ops = ca; in tcp_ca_dst_init() 4267 icsk->icsk_ack.retry++; in __tcp_send_ack() 4396 icsk->icsk_probes_out = 0; in tcp_send_probe0() 4397 icsk->icsk_backoff = 0; in tcp_send_probe0() 4402 icsk->icsk_probes_out++; in tcp_send_probe0() [all …]
|
| A D | tcp_input.c | 620 icsk->icsk_ack.quick = 0; in tcp_clamp_window() 842 if (!icsk->icsk_ack.ato) { in tcp_event_data_recv() 853 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + TCP_ATO_MIN / 2; in tcp_event_data_recv() 855 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + m; in tcp_event_data_recv() 856 if (icsk->icsk_ack.ato > icsk->icsk_rto) in tcp_event_data_recv() 857 icsk->icsk_ack.ato = icsk->icsk_rto; in tcp_event_data_recv() 2189 (icsk->icsk_ca_state == TCP_CA_Loss && !icsk->icsk_retransmits)) { in tcp_enter_loss() 2700 icsk->icsk_mtup.search_high = icsk->icsk_mtup.probe_size - 1; in tcp_mtup_probe_failed() 2722 icsk->icsk_mtup.search_low = icsk->icsk_mtup.probe_size; in tcp_mtup_probe_success() 3403 icsk = inet_csk(sk); in tcp_clean_rtx_queue() [all …]
|
| A D | tcp.c | 432 icsk->icsk_rto = TCP_TIMEOUT_INIT; in tcp_init_sock() 470 icsk->icsk_sync_mss = tcp_sync_mss; in tcp_init_sock() 3378 icsk->icsk_backoff = 0; in tcp_disconnect() 3379 icsk->icsk_probes_out = 0; in tcp_disconnect() 3380 icsk->icsk_probes_tstamp = 0; in tcp_disconnect() 3392 if (icsk->icsk_ca_initialized && icsk->icsk_ca_ops->release) in tcp_disconnect() 3393 icsk->icsk_ca_ops->release(sk); in tcp_disconnect() 3394 memset(icsk->icsk_ca_priv, 0, sizeof(icsk->icsk_ca_priv)); in tcp_disconnect() 3395 icsk->icsk_ca_initialized = 0; in tcp_disconnect() 4461 ca_ops = icsk->icsk_ca_ops; in do_tcp_getsockopt() [all …]
|
| A D | inet_timewait_sock.c | 113 const struct inet_connection_sock *icsk = inet_csk(sk); in inet_twsk_hashdance_schedule() local 130 tw->tw_tb = icsk->icsk_bind_hash; in inet_twsk_hashdance_schedule() 131 WARN_ON(!icsk->icsk_bind_hash); in inet_twsk_hashdance_schedule() 133 tw->tw_tb2 = icsk->icsk_bind2_hash; in inet_twsk_hashdance_schedule() 134 WARN_ON(!icsk->icsk_bind2_hash); in inet_twsk_hashdance_schedule()
|
| A D | tcp_htcp.c | 84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt() local 92 if (icsk->icsk_ca_state == TCP_CA_Open) { in measure_rtt() 104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput() local 109 if (icsk->icsk_ca_state == TCP_CA_Open) in measure_achieved_throughput() 119 if (!((1 << icsk->icsk_ca_state) & (TCPF_CA_Open | TCPF_CA_Disorder))) { in measure_achieved_throughput()
|
| A D | tcp_diag.c | 115 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux() local 133 ulp_ops = icsk->icsk_ulp_ops; in tcp_diag_get_aux() 145 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux_size() local 169 ulp_ops = icsk->icsk_ulp_ops; in tcp_diag_get_aux_size()
|
| A D | tcp_minisocks.c | 310 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() local 319 const int rto = (icsk->icsk_rto << 2) - (icsk->icsk_rto >> 1); in tcp_time_wait() 474 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() local 484 icsk->icsk_ca_dst_locked = tcp_ca_dst_locked(dst); in tcp_ca_openreq_child() 485 icsk->icsk_ca_ops = ca; in tcp_ca_openreq_child() 493 (!icsk->icsk_ca_setsockopt || in tcp_ca_openreq_child() 494 !bpf_try_module_get(icsk->icsk_ca_ops, icsk->icsk_ca_ops->owner))) in tcp_ca_openreq_child()
|
| A D | inet_diag.c | 237 int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk, in inet_sk_diag_fill() argument 306 if (!icsk) { in inet_sk_diag_fill() 311 icsk_pending = smp_load_acquire(&icsk->icsk_pending); in inet_sk_diag_fill() 316 r->idiag_retrans = icsk->icsk_retransmits; in inet_sk_diag_fill() 318 jiffies_delta_to_msecs(icsk_timeout(icsk) - jiffies); in inet_sk_diag_fill() 321 r->idiag_retrans = icsk->icsk_probes_out; in inet_sk_diag_fill() 323 jiffies_delta_to_msecs(icsk_timeout(icsk) - jiffies); in inet_sk_diag_fill() 326 r->idiag_retrans = icsk->icsk_probes_out; in inet_sk_diag_fill() 345 ca_ops = READ_ONCE(icsk->icsk_ca_ops); in inet_sk_diag_fill() 365 ca_ops = READ_ONCE(icsk->icsk_ca_ops); in inet_sk_diag_fill()
|
| A D | tcp_nv.c | 241 const struct inet_connection_sock *icsk = inet_csk(sk); in tcpnv_acked() local 255 if (icsk->icsk_ca_state != TCP_CA_Open && in tcpnv_acked() 256 icsk->icsk_ca_state != TCP_CA_Disorder) in tcpnv_acked()
|
| A D | tcp_ipv4.c | 444 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ld_RTO_revert() local 454 !icsk->icsk_backoff) in tcp_ld_RTO_revert() 461 icsk->icsk_backoff--; in tcp_ld_RTO_revert() 463 icsk->icsk_rto = inet_csk_rto_backoff(icsk, tcp_rto_max(sk)); in tcp_ld_RTO_revert() 2515 icsk->icsk_af_ops = &ipv4_specific; 2932 timer_expires = icsk_timeout(icsk); 2935 timer_expires = icsk_timeout(icsk); 2961 icsk->icsk_retransmits, 2963 icsk->icsk_probes_out, 2966 jiffies_to_clock_t(icsk->icsk_rto), [all …]
|
| A D | tcp_bic.c | 192 const struct inet_connection_sock *icsk = inet_csk(sk); in bictcp_acked() local 194 if (icsk->icsk_ca_state == TCP_CA_Open) { in bictcp_acked()
|
| A D | tcp_fastopen.c | 96 int tcp_fastopen_get_cipher(struct net *net, struct inet_connection_sock *icsk, in tcp_fastopen_get_cipher() argument 103 if (icsk) in tcp_fastopen_get_cipher() 104 ctx = rcu_dereference(icsk->icsk_accept_queue.fastopenq.ctx); in tcp_fastopen_get_cipher()
|
| A D | syncookies.c | 200 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_get_cookie_sock() local 204 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst, in tcp_get_cookie_sock()
|
| A D | ip_sockglue.c | 1093 struct inet_connection_sock *icsk = inet_csk(sk); in do_ip_setsockopt() local 1101 icsk->icsk_ext_hdr_len -= old->opt.optlen; in do_ip_setsockopt() 1103 icsk->icsk_ext_hdr_len += opt->opt.optlen; in do_ip_setsockopt() 1104 icsk->icsk_sync_mss(sk, icsk->icsk_pmtu_cookie); in do_ip_setsockopt()
|
| /net/mptcp/ |
| A D | token_test.c | 34 struct inet_connection_sock *icsk; in build_icsk() local 36 icsk = kunit_kzalloc(test, sizeof(struct inet_connection_sock), in build_icsk() 38 KUNIT_EXPECT_NOT_ERR_OR_NULL(test, icsk); in build_icsk() 39 return icsk; in build_icsk() 73 struct inet_connection_sock *icsk = build_icsk(test); in mptcp_token_test_msk_basic() local 79 rcu_assign_pointer(icsk->icsk_ulp_data, ctx); in mptcp_token_test_msk_basic() 84 mptcp_token_new_connect((struct sock *)icsk)); in mptcp_token_test_msk_basic()
|
| A D | subflow.c | 769 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_ulp_fallback() local 772 icsk->icsk_ulp_ops = NULL; in subflow_ulp_fallback() 773 rcu_assign_pointer(icsk->icsk_ulp_data, NULL); in subflow_ulp_fallback() 1554 struct inet_connection_sock *icsk = inet_csk(sk); in mptcpv6_handle_mapped() local 1562 if (likely(icsk->icsk_af_ops == target)) in mptcpv6_handle_mapped() 1565 subflow->icsk_af_ops = icsk->icsk_af_ops; in mptcpv6_handle_mapped() 1566 icsk->icsk_af_ops = target; in mptcpv6_handle_mapped() 1823 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_create_ctx() local 1830 rcu_assign_pointer(icsk->icsk_ulp_data, ctx); in subflow_create_ctx() 1973 ctx->icsk_af_ops = icsk->icsk_af_ops; in subflow_ulp_init() [all …]
|
| A D | protocol.c | 355 struct inet_connection_sock *icsk = inet_csk(sk); in mptcp_stop_rtx_timer() local 357 sk_stop_timer(sk, &icsk->icsk_retransmit_timer); in mptcp_stop_rtx_timer() 429 struct inet_connection_sock *icsk = inet_csk(sk); in mptcp_set_datafin_timeout() local 432 retransmits = min_t(u32, icsk->icsk_retransmits, in mptcp_set_datafin_timeout() 508 READ_ONCE(icsk->icsk_ack.rcv_mss)) || in mptcp_subflow_could_cleanup() 867 struct inet_connection_sock *icsk = inet_csk(sk); in mptcp_reset_rtx_timer() local 2233 struct inet_connection_sock *icsk = timer_container_of(icsk, t, in mptcp_retransmit_timer() local 2235 struct sock *sk = &icsk->icsk_inet.sk; in mptcp_retransmit_timer() 2590 icsk->icsk_retransmits++; in __mptcp_retrans() 2799 icsk->icsk_ca_ops = NULL; in mptcp_ca_reset() [all …]
|
| /net/ipv6/ |
| A D | tcp_ipv6.c | 222 u32 exthdrlen = icsk->icsk_ext_hdr_len; in tcp_v6_connect() 233 WRITE_ONCE(icsk->icsk_af_ops, &ipv6_mapped); in tcp_v6_connect() 244 icsk->icsk_ext_hdr_len = exthdrlen; in tcp_v6_connect() 304 icsk->icsk_ext_hdr_len = 0; in tcp_v6_connect() 306 icsk->icsk_ext_hdr_len = opt->opt_flen + in tcp_v6_connect() 2131 icsk->icsk_af_ops = &ipv6_specific; 2198 timer_expires = icsk_timeout(icsk); 2201 timer_expires = icsk_timeout(icsk); 2233 icsk->icsk_retransmits, 2235 icsk->icsk_probes_out, [all …]
|
| A D | ipv6_sockglue.c | 109 struct inet_connection_sock *icsk = inet_csk(sk); in ipv6_update_options() local 110 icsk->icsk_ext_hdr_len = opt->opt_flen + opt->opt_nflen; in ipv6_update_options() 111 icsk->icsk_sync_mss(sk, icsk->icsk_pmtu_cookie); in ipv6_update_options() 593 struct inet_connection_sock *icsk = inet_csk(sk); in do_ipv6_setsockopt() local 601 WRITE_ONCE(icsk->icsk_af_ops, &ipv4_specific); in do_ipv6_setsockopt() 604 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in do_ipv6_setsockopt()
|
| /net/tls/ |
| A D | tls_toe.c | 48 struct inet_connection_sock *icsk = inet_csk(sk); in tls_toe_sk_destruct() local 53 rcu_assign_pointer(icsk->icsk_ulp_data, NULL); in tls_toe_sk_destruct()
|
| A D | tls_main.c | 370 struct inet_connection_sock *icsk = inet_csk(sk); in tls_sk_proto_close() local 386 rcu_assign_pointer(icsk->icsk_ulp_data, NULL); in tls_sk_proto_close() 862 struct inet_connection_sock *icsk = inet_csk(sk); in tls_ctx_create() local 880 rcu_assign_pointer(icsk->icsk_ulp_data, ctx); in tls_ctx_create()
|