| /net/ipv4/ |
| A D | bpf_tcp_ca.c | 88 case offsetof(struct tcp_sock, snd_cwnd): in bpf_tcp_ca_btf_struct_access() 89 end = offsetofend(struct tcp_sock, snd_cwnd); in bpf_tcp_ca_btf_struct_access() 91 case offsetof(struct tcp_sock, snd_cwnd_cnt): in bpf_tcp_ca_btf_struct_access() 92 end = offsetofend(struct tcp_sock, snd_cwnd_cnt); in bpf_tcp_ca_btf_struct_access() 94 case offsetof(struct tcp_sock, snd_cwnd_stamp): in bpf_tcp_ca_btf_struct_access() 97 case offsetof(struct tcp_sock, snd_ssthresh): in bpf_tcp_ca_btf_struct_access() 98 end = offsetofend(struct tcp_sock, snd_ssthresh); in bpf_tcp_ca_btf_struct_access() 100 case offsetof(struct tcp_sock, ecn_flags): in bpf_tcp_ca_btf_struct_access() 101 end = offsetofend(struct tcp_sock, ecn_flags); in bpf_tcp_ca_btf_struct_access() 103 case offsetof(struct tcp_sock, app_limited): in bpf_tcp_ca_btf_struct_access() [all …]
|
| A D | tcp.c | 423 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock() 626 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl() 684 struct tcp_sock *tp = tcp_sk(sk); in tcp_skb_entail() 728 struct tcp_sock *tp = tcp_sk(sk); in tcp_push() 920 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_size_goal() 1014 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_fastopen() 1061 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_locked() 1403 struct tcp_sock *tp = tcp_sk(sk); in tcp_splice_eof() 1423 struct tcp_sock *tp = tcp_sk(sk); in tcp_recv_urg() 1496 struct tcp_sock *tp = tcp_sk(sk); in __tcp_cleanup_rbuf() [all …]
|
| A D | tcp_input.c | 369 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_ecn_check() 538 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window() 578 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space() 615 struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_window() 732 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_rtt_measure_ts() 750 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcvbuf_grow() 779 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_space_adjust() 830 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_data_recv() 885 struct tcp_sock *tp = tcp_sk(sk); in tcp_rtt_estimator() 1090 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reordering() [all …]
|
| A D | tcp_recovery.c | 7 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd() 32 s32 tcp_rack_skb_timeout(struct tcp_sock *tp, struct sk_buff *skb, u32 reo_wnd) in tcp_rack_skb_timeout() 60 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss() 97 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost() 118 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance() 151 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout() 189 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd() 220 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost()
|
| A D | tcp_output.c | 70 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_new_data_sent() 125 struct tcp_sock *tp = tcp_sk(sk); in tcp_advertise_mss() 146 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_restart() 184 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_ack_sent() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_select_window() 338 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_syn() 382 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send() 463 struct tcp_sock *tp, in mptcp_options_write() 1118 struct tcp_sock *tp; in tcp_tsq_workfn() 1250 struct tcp_sock *tp = container_of(timer, struct tcp_sock, pacing_timer); in tcp_pace_kick() [all …]
|
| A D | tcp_dctcp.c | 81 static void dctcp_reset(const struct tcp_sock *tp, struct dctcp *ca) in dctcp_reset() 91 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_init() 121 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ssthresh() 129 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_update_alpha() 177 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss() 219 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_get_info() 246 struct tcp_sock *tp = tcp_sk(sk); in dctcp_cwnd_undo()
|
| A D | tcp_bbr.c | 268 struct tcp_sock *tp = tcp_sk(sk); in bbr_init_pacing_rate_from_rtt() 288 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_pacing_rate() 306 struct tcp_sock *tp = tcp_sk(sk); in bbr_tso_segs_goal() 323 struct tcp_sock *tp = tcp_sk(sk); in bbr_save_cwnd() 334 struct tcp_sock *tp = tcp_sk(sk); in bbr_cwnd_event() 439 struct tcp_sock *tp = tcp_sk(sk); in bbr_packets_in_net_at_edt() 483 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd_to_recover_or_restore() 522 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd() 557 struct tcp_sock *tp = tcp_sk(sk); in bbr_is_next_cycle_phase() 593 struct tcp_sock *tp = tcp_sk(sk); in bbr_advance_cycle_phase() [all …]
|
| A D | tcp_timer.c | 30 const struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_rto_to_user_timeout() 107 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources() 217 struct tcp_sock *tp = tcp_sk(sk); in retransmits_timed_out() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout() 310 struct tcp_sock *tp = tcp_sk(sk); in tcp_delack_timer_handler() 391 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer() 441 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rto_stats() 458 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer() 533 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer() 783 struct tcp_sock *tp = tcp_sk(sk); in tcp_keepalive_timer() [all …]
|
| A D | tcp_cdg.c | 143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff() 265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid() 302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked() 331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh() 348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event() 376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
|
| A D | tcp_rate.c | 42 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_sent() 83 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_delivered() 120 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_gen() 196 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited()
|
| A D | tcp_highspeed.c | 102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init() 114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid() 153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
|
| A D | tcp_westwood.c | 165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw() 182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count() 219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin() 242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
|
| A D | tcp_illinois.c | 59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset() 224 struct tcp_sock *tp = tcp_sk(sk); in update_params() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid() 297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
|
| A D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid() 36 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
|
| A D | tcp_yeah.c | 43 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_init() 60 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_cong_avoid() 189 const struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_ssthresh()
|
| A D | tcp_cong.c | 455 __bpf_kfunc u32 tcp_slow_start(struct tcp_sock *tp, u32 acked) in tcp_slow_start() 469 __bpf_kfunc void tcp_cong_avoid_ai(struct tcp_sock *tp, u32 w, u32 acked) in tcp_cong_avoid_ai() 497 struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_cong_avoid() 516 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_ssthresh() 524 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_undo_cwnd()
|
| A D | tcp_cubic.c | 120 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 326 struct tcp_sock *tp = tcp_sk(sk); in cubictcp_cong_avoid() 343 const struct tcp_sock *tp = tcp_sk(sk); in cubictcp_recalc_ssthresh() 388 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 454 const struct tcp_sock *tp = tcp_sk(sk); in cubictcp_acked()
|
| A D | tcp_vegas.c | 73 const struct tcp_sock *tp = tcp_sk(sk); in vegas_enable() 160 static inline u32 tcp_vegas_ssthresh(struct tcp_sock *tp) in tcp_vegas_ssthresh() 167 struct tcp_sock *tp = tcp_sk(sk); in tcp_vegas_cong_avoid()
|
| A D | tcp_minisocks.c | 275 const struct tcp_sock *tp = tcp_sk(sk); in tcp_time_wait_init() 311 struct tcp_sock *tp = tcp_sk(sk); in tcp_time_wait() 430 const struct tcp_sock *tp = tcp_sk(sk_listener); in tcp_openreq_init_rwin() 464 static void tcp_ecn_openreq_child(struct tcp_sock *tp, in tcp_ecn_openreq_child() 501 static void smc_check_reset_syn_req(const struct tcp_sock *oldtp, in smc_check_reset_syn_req() 503 struct tcp_sock *newtp) in smc_check_reset_syn_req() 530 const struct tcp_sock *oldtp; in tcp_create_openreq_child() 531 struct tcp_sock *newtp; in tcp_create_openreq_child()
|
| A D | tcp_nv.c | 125 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_reset() 182 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_cong_avoid() 210 const struct tcp_sock *tp = tcp_sk(sk); in tcpnv_recalc_ssthresh() 242 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_acked()
|
| A D | tcp_fastopen.c | 173 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_add_skb() 244 struct tcp_sock *tp; in tcp_fastopen_create_child() 454 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_defer_connect() 561 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_disable_ofo_check() 590 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_detect_blackhole()
|
| A D | tcp_lp.c | 139 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_remote_hz_estimator() 193 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_owd_calculator() 273 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_pkts_acked()
|
| A D | tcp_htcp.c | 105 const struct tcp_sock *tp = tcp_sk(sk); in measure_achieved_throughput() 224 const struct tcp_sock *tp = tcp_sk(sk); in htcp_recalc_ssthresh() 233 struct tcp_sock *tp = tcp_sk(sk); in htcp_cong_avoid()
|
| A D | tcp_hybla.c | 48 struct tcp_sock *tp = tcp_sk(sk); in hybla_init() 92 struct tcp_sock *tp = tcp_sk(sk); in hybla_cong_avoid()
|
| /net/mptcp/ |
| A D | fastopen.c | 14 struct tcp_sock *tp; in mptcp_fastopen_subflow_synack_set_params() 22 ssk = subflow->tcp_sock; in mptcp_fastopen_subflow_synack_set_params()
|