Lines Matching refs:flag

1033 	int	flag;  member
1059 state->flag |= FLAG_DSACK_TLP; in tcp_dsack_seen()
1073 if (tp->reord_seen && !(state->flag & FLAG_DSACK_TLP)) in tcp_dsack_seen()
1076 state->flag |= FLAG_DSACKING_ACK; in tcp_dsack_seen()
1437 state->flag |= FLAG_ORIG_SACK_ACKED; in tcp_sacktag_one()
1450 state->flag |= FLAG_DATA_SACKED; in tcp_sacktag_one()
1859 state->flag = 0; in tcp_sacktag_write_queue()
2035 return state->flag; in tcp_sacktag_write_queue()
2607 void tcp_cwnd_reduction(struct sock *sk, int newly_acked_sacked, int newly_lost, int flag) in tcp_cwnd_reduction() argument
2616 trace_tcp_cwnd_reduction_tp(sk, newly_acked_sacked, newly_lost, flag); in tcp_cwnd_reduction()
2626 if (flag & FLAG_SND_UNA_ADVANCED && !newly_lost) in tcp_cwnd_reduction()
2679 static void tcp_try_to_open(struct sock *sk, int flag) in tcp_try_to_open() argument
2688 if (flag & FLAG_ECE) in tcp_try_to_open()
2837 static void tcp_process_loss(struct sock *sk, int flag, int num_dupack, in tcp_process_loss() argument
2843 if ((flag & FLAG_SND_UNA_ADVANCED || rcu_access_pointer(tp->fastopen_rsk)) && in tcp_process_loss()
2851 if ((flag & FLAG_ORIG_SACK_ACKED) && in tcp_process_loss()
2856 if (flag & FLAG_DATA_SACKED || num_dupack) in tcp_process_loss()
2858 } else if (flag & FLAG_SND_UNA_ADVANCED && !recovered) { in tcp_process_loss()
2883 tcp_add_reno_sack(sk, num_dupack, flag & FLAG_ECE); in tcp_process_loss()
2884 else if (flag & FLAG_SND_UNA_ADVANCED) in tcp_process_loss()
2956 int flag = *ack_flag; in tcp_fastretrans_alert() local
2957 bool ece_ack = flag & FLAG_ECE; in tcp_fastretrans_alert()
3003 if (!(flag & FLAG_SND_UNA_ADVANCED)) { in tcp_fastretrans_alert()
3010 tcp_try_to_open(sk, flag); in tcp_fastretrans_alert()
3023 tcp_process_loss(sk, flag, num_dupack, rexmit); in tcp_fastretrans_alert()
3034 if (flag & FLAG_SND_UNA_ADVANCED) in tcp_fastretrans_alert()
3044 tcp_try_to_open(sk, flag); in tcp_fastretrans_alert()
3066 static void tcp_update_rtt_min(struct sock *sk, u32 rtt_us, const int flag) in tcp_update_rtt_min() argument
3071 if ((flag & FLAG_ACK_MAYBE_DELAYED) && rtt_us > tcp_min_rtt(tp)) { in tcp_update_rtt_min()
3082 static bool tcp_ack_update_rtt(struct sock *sk, const int flag, in tcp_ack_update_rtt() argument
3103 tp->rx_opt.rcv_tsecr && flag & FLAG_ACKED) in tcp_ack_update_rtt()
3114 tcp_update_rtt_min(sk, ca_rtt_us, flag); in tcp_ack_update_rtt()
3241 int flag = 0; in tcp_clean_rtx_queue() local
3268 flag |= FLAG_RETRANS_DATA_ACKED; in tcp_clean_rtx_queue()
3278 flag |= FLAG_ORIG_SACK_ACKED; in tcp_clean_rtx_queue()
3304 flag |= FLAG_DATA_ACKED; in tcp_clean_rtx_queue()
3306 flag |= FLAG_SYN_ACKED; in tcp_clean_rtx_queue()
3331 flag |= FLAG_SACK_RENEGING; in tcp_clean_rtx_queue()
3334 if (likely(first_ackt) && !(flag & FLAG_RETRANS_DATA_ACKED)) { in tcp_clean_rtx_queue()
3341 !(flag & (FLAG_CA_ALERT | FLAG_SYN_ACKED))) { in tcp_clean_rtx_queue()
3346 flag |= FLAG_ACK_MAYBE_DELAYED; in tcp_clean_rtx_queue()
3353 rtt_update = tcp_ack_update_rtt(sk, flag, seq_rtt_us, sack_rtt_us, in tcp_clean_rtx_queue()
3356 if (flag & FLAG_ACKED) { in tcp_clean_rtx_queue()
3357 flag |= FLAG_SET_XMIT_TIMER; /* set TLP or RTO timer */ in tcp_clean_rtx_queue()
3372 if (flag & FLAG_RETRANS_DATA_ACKED) in tcp_clean_rtx_queue()
3373 flag &= ~FLAG_ORIG_SACK_ACKED; in tcp_clean_rtx_queue()
3386 flag |= FLAG_SET_XMIT_TIMER; /* set TLP or RTO timer */ in tcp_clean_rtx_queue()
3421 return flag; in tcp_clean_rtx_queue()
3448 static inline bool tcp_ack_is_dubious(const struct sock *sk, const int flag) in tcp_ack_is_dubious() argument
3450 return !(flag & FLAG_NOT_DUP) || (flag & FLAG_CA_ALERT) || in tcp_ack_is_dubious()
3455 static inline bool tcp_may_raise_cwnd(const struct sock *sk, const int flag) in tcp_may_raise_cwnd() argument
3465 return flag & FLAG_FORWARD_PROGRESS; in tcp_may_raise_cwnd()
3467 return flag & FLAG_DATA_ACKED; in tcp_may_raise_cwnd()
3476 int flag, const struct rate_sample *rs) in tcp_cong_control() argument
3481 icsk->icsk_ca_ops->cong_control(sk, ack, flag, rs); in tcp_cong_control()
3487 tcp_cwnd_reduction(sk, acked_sacked, rs->losses, flag); in tcp_cong_control()
3488 } else if (tcp_may_raise_cwnd(sk, flag)) { in tcp_cong_control()
3572 int flag = 0; in tcp_ack_update_window() local
3579 flag |= FLAG_WIN_UPDATE; in tcp_ack_update_window()
3603 return flag; in tcp_ack_update_window()
3719 static void tcp_process_tlp_ack(struct sock *sk, u32 ack, int flag) in tcp_process_tlp_ack() argument
3729 } else if (flag & FLAG_DSACK_TLP) { in tcp_process_tlp_ack()
3742 } else if (!(flag & (FLAG_SND_UNA_ADVANCED | in tcp_process_tlp_ack()
3749 static void tcp_in_ack_event(struct sock *sk, int flag) in tcp_in_ack_event() argument
3756 if (flag & FLAG_WIN_UPDATE) in tcp_in_ack_event()
3758 if (flag & FLAG_SLOWPATH) { in tcp_in_ack_event()
3760 if (flag & FLAG_ECE) in tcp_in_ack_event()
3790 static u32 tcp_newly_delivered(struct sock *sk, u32 prior_delivered, int flag) in tcp_newly_delivered() argument
3798 if (flag & FLAG_ECE) in tcp_newly_delivered()
3805 static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag) in tcp_ack() argument
3839 if (!(flag & FLAG_NO_CHALLENGE_ACK)) in tcp_ack()
3853 flag |= FLAG_SND_UNA_ADVANCED; in tcp_ack()
3869 if (flag & FLAG_UPDATE_TS_RECENT) in tcp_ack()
3870 flag |= tcp_replace_ts_recent(tp, TCP_SKB_CB(skb)->seq); in tcp_ack()
3872 if ((flag & (FLAG_SLOWPATH | FLAG_SND_UNA_ADVANCED)) == in tcp_ack()
3880 flag |= FLAG_WIN_UPDATE; in tcp_ack()
3885 flag |= FLAG_DATA; in tcp_ack()
3889 flag |= tcp_ack_update_window(sk, skb, ack, ack_seq); in tcp_ack()
3892 flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una, in tcp_ack()
3896 flag |= FLAG_ECE; in tcp_ack()
3900 flag & FLAG_ECE); in tcp_ack()
3922 flag |= tcp_clean_rtx_queue(sk, skb, prior_fack, prior_snd_una, in tcp_ack()
3923 &sack_state, flag & FLAG_ECE); in tcp_ack()
3927 tcp_in_ack_event(sk, flag); in tcp_ack()
3930 tcp_process_tlp_ack(sk, ack, flag); in tcp_ack()
3932 if (tcp_ack_is_dubious(sk, flag)) { in tcp_ack()
3933 if (!(flag & (FLAG_SND_UNA_ADVANCED | in tcp_ack()
3937 if (!(flag & FLAG_DATA)) in tcp_ack()
3940 tcp_fastretrans_alert(sk, prior_snd_una, num_dupack, &flag, in tcp_ack()
3945 if (flag & FLAG_SET_XMIT_TIMER) in tcp_ack()
3948 if ((flag & FLAG_FORWARD_PROGRESS) || !(flag & FLAG_NOT_DUP)) in tcp_ack()
3951 delivered = tcp_newly_delivered(sk, delivered, flag); in tcp_ack()
3953 rs.is_ack_delayed = !!(flag & FLAG_ACK_MAYBE_DELAYED); in tcp_ack()
3955 tcp_cong_control(sk, ack, delivered, flag, sack_state.rate); in tcp_ack()
3960 tcp_in_ack_event(sk, flag); in tcp_ack()
3962 if (flag & FLAG_DSACKING_ACK) { in tcp_ack()
3963 tcp_fastretrans_alert(sk, prior_snd_una, num_dupack, &flag, in tcp_ack()
3965 tcp_newly_delivered(sk, delivered, flag); in tcp_ack()
3974 tcp_process_tlp_ack(sk, ack, flag); in tcp_ack()
3982 flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una, in tcp_ack()
3984 tcp_fastretrans_alert(sk, prior_snd_una, num_dupack, &flag, in tcp_ack()
3986 tcp_newly_delivered(sk, delivered, flag); in tcp_ack()
6078 int flag = 0; in tcp_rcv_established() local
6114 flag |= __tcp_replace_ts_recent(tp, in tcp_rcv_established()
6120 tcp_ack(sk, skb, flag); in tcp_rcv_established()
6155 flag |= __tcp_replace_ts_recent(tp, in tcp_rcv_established()
6171 tcp_ack(sk, skb, flag | FLAG_DATA); in tcp_rcv_established()