| /net/netfilter/ |
| A D | xt_rateest.c | 18 struct gnet_stats_rate_est64 sample = {0}; in xt_rateest_mt() local 22 gen_estimator_read(&info->est1->rate_est, &sample); in xt_rateest_mt() 25 bps1 = info->bps1 >= sample.bps ? info->bps1 - sample.bps : 0; in xt_rateest_mt() 26 pps1 = info->pps1 >= sample.pps ? info->pps1 - sample.pps : 0; in xt_rateest_mt() 28 bps1 = sample.bps; in xt_rateest_mt() 29 pps1 = sample.pps; in xt_rateest_mt() 36 gen_estimator_read(&info->est2->rate_est, &sample); in xt_rateest_mt() 39 bps2 = info->bps2 >= sample.bps ? info->bps2 - sample.bps : 0; in xt_rateest_mt() 40 pps2 = info->pps2 >= sample.pps ? info->pps2 - sample.pps : 0; in xt_rateest_mt() 42 bps2 = sample.bps; in xt_rateest_mt() [all …]
|
| /net/ipv4/ |
| A D | tcp_nv.c | 239 static void tcpnv_acked(struct sock *sk, const struct ack_sample *sample) in tcpnv_acked() argument 251 if (sample->rtt_us < 0) in tcpnv_acked() 268 if (sample->in_flight == 0) in tcpnv_acked() 274 avg_rtt = (((u64)sample->rtt_us) * nv_rtt_factor + in tcpnv_acked() 278 avg_rtt = sample->rtt_us; in tcpnv_acked() 283 avg_rtt = sample->rtt_us; in tcpnv_acked() 287 rate64 = ((u64)sample->in_flight) * 80000; in tcpnv_acked()
|
| A D | tcp_cdg.c | 299 static void tcp_cdg_acked(struct sock *sk, const struct ack_sample *sample) in tcp_cdg_acked() argument 304 if (sample->rtt_us <= 0) in tcp_cdg_acked() 312 if (sample->pkts_acked == 1 && ca->delack) { in tcp_cdg_acked() 316 ca->rtt.min = min(ca->rtt.min, sample->rtt_us); in tcp_cdg_acked() 319 } else if (sample->pkts_acked > 1 && ca->delack < 5) { in tcp_cdg_acked() 324 ca->rtt.min = min_not_zero(ca->rtt.min, sample->rtt_us); in tcp_cdg_acked() 325 ca->rtt.max = max(ca->rtt.max, sample->rtt_us); in tcp_cdg_acked()
|
| A D | tcp_htcp.c | 102 const struct ack_sample *sample) in measure_achieved_throughput() argument 110 ca->pkts_acked = sample->pkts_acked; in measure_achieved_throughput() 112 if (sample->rtt_us > 0) in measure_achieved_throughput() 113 measure_rtt(sk, usecs_to_jiffies(sample->rtt_us)); in measure_achieved_throughput() 125 ca->packetcount += sample->pkts_acked; in measure_achieved_throughput()
|
| A D | tcp_veno.c | 74 const struct ack_sample *sample) in tcp_veno_pkts_acked() argument 79 if (sample->rtt_us < 0) in tcp_veno_pkts_acked() 83 vrtt = sample->rtt_us + 1; in tcp_veno_pkts_acked()
|
| A D | tcp_vegas.c | 111 void tcp_vegas_pkts_acked(struct sock *sk, const struct ack_sample *sample) in tcp_vegas_pkts_acked() argument 116 if (sample->rtt_us < 0) in tcp_vegas_pkts_acked() 120 vrtt = sample->rtt_us + 1; in tcp_vegas_pkts_acked()
|
| A D | tcp_lp.c | 271 static void tcp_lp_pkts_acked(struct sock *sk, const struct ack_sample *sample) in tcp_lp_pkts_acked() argument 278 if (sample->rtt_us > 0) in tcp_lp_pkts_acked() 279 tcp_lp_rtt_sample(sk, sample->rtt_us); in tcp_lp_pkts_acked()
|
| A D | tcp_westwood.c | 104 const struct ack_sample *sample) in tcp_westwood_pkts_acked() argument 108 if (sample->rtt_us > 0) in tcp_westwood_pkts_acked() 109 w->rtt = usecs_to_jiffies(sample->rtt_us); in tcp_westwood_pkts_acked()
|
| A D | tcp_illinois.c | 86 static void tcp_illinois_acked(struct sock *sk, const struct ack_sample *sample) in tcp_illinois_acked() argument 89 s32 rtt_us = sample->rtt_us; in tcp_illinois_acked() 91 ca->acked = sample->pkts_acked; in tcp_illinois_acked()
|
| A D | tcp_bic.c | 190 static void bictcp_acked(struct sock *sk, const struct ack_sample *sample) in bictcp_acked() argument 197 ca->delayed_ack += sample->pkts_acked - in bictcp_acked()
|
| A D | tcp_cubic.c | 452 __bpf_kfunc static void cubictcp_acked(struct sock *sk, const struct ack_sample *sample) in cubictcp_acked() argument 459 if (sample->rtt_us < 0) in cubictcp_acked() 466 delay = sample->rtt_us; in cubictcp_acked()
|
| A D | tcp_vegas.h | 21 void tcp_vegas_pkts_acked(struct sock *sk, const struct ack_sample *sample);
|
| A D | bpf_tcp_ca.c | 279 static void bpf_tcp_ca_pkts_acked(struct sock *sk, const struct ack_sample *sample) in bpf_tcp_ca_pkts_acked() argument
|
| A D | tcp_input.c | 665 static void tcp_rcv_rtt_update(struct tcp_sock *tp, u32 sample, int win_dep) in tcp_rcv_rtt_update() argument 668 long m = sample << 3; in tcp_rcv_rtt_update() 688 new_sample = old_sample - (old_sample >> 3) + sample; in tcp_rcv_rtt_update() 3390 struct ack_sample sample = { .pkts_acked = pkts_acked, in tcp_clean_rtx_queue() local 3393 sample.in_flight = tp->mss_cache * in tcp_clean_rtx_queue() 3395 icsk->icsk_ca_ops->pkts_acked(sk, &sample); in tcp_clean_rtx_queue()
|
| /net/core/ |
| A D | gen_stats.c | 311 struct gnet_stats_rate_est64 sample; in gnet_stats_copy_rate_est() local 315 if (!gen_estimator_read(rate_est, &sample)) in gnet_stats_copy_rate_est() 317 est.bps = min_t(u64, UINT_MAX, sample.bps); in gnet_stats_copy_rate_est() 319 est.pps = sample.pps; in gnet_stats_copy_rate_est() 329 if (res < 0 || est.bps == sample.bps) in gnet_stats_copy_rate_est() 332 return gnet_stats_copy(d, TCA_STATS_RATE_EST64, &sample, in gnet_stats_copy_rate_est() 333 sizeof(sample), TCA_STATS_PAD); in gnet_stats_copy_rate_est()
|
| A D | gen_estimator.c | 257 struct gnet_stats_rate_est64 *sample) in gen_estimator_read() argument 271 sample->bps = est->avbps >> 8; in gen_estimator_read() 272 sample->pps = est->avpps >> 8; in gen_estimator_read()
|
| /net/sched/ |
| A D | act_sample.c | 291 entry->sample.psample_group = in tcf_offload_sample_get_group() 293 entry->destructor_priv = entry->sample.psample_group; in tcf_offload_sample_get_group() 304 entry->sample.trunc_size = tcf_sample_trunc_size(act); in tcf_sample_offload_act_setup() 305 entry->sample.truncate = tcf_sample_truncate(act); in tcf_sample_offload_act_setup() 306 entry->sample.rate = tcf_sample_rate(act); in tcf_sample_offload_act_setup()
|
| A D | act_police.c | 262 struct gnet_stats_rate_est64 sample; in tcf_police_act() local 264 if (!gen_estimator_read(&police->tcf_rate_est, &sample) || in tcf_police_act() 265 sample.bps >= p->tcfp_ewma_rate) in tcf_police_act()
|
| A D | sch_fq.c | 609 unsigned long sample; in fq_check_throttled() local 618 sample = (unsigned long)(now - q->time_next_delayed_flow); in fq_check_throttled() 619 if ((long)sample > 0) { in fq_check_throttled() 621 q->unthrottle_latency_ns += sample >> 3; in fq_check_throttled()
|
| A D | sch_cake.c | 1354 static u64 cake_ewma(u64 avg, u64 sample, u32 shift) in cake_ewma() argument 1357 avg += sample >> shift; in cake_ewma()
|
| A D | Kconfig | 757 Say Y here to allow packet sampling tc action. The packet sample
|
| /net/mac80211/ |
| A D | rc80211_minstrel_ht.c | 704 u16 *rates = mi->sample[type].sample_rates; in __minstrel_ht_get_sample_rate() 791 u16 cur = mi->sample[type].sample_rates[i]; in minstrel_ht_find_sample_rate() 807 u16 *rates = mi->sample[type].sample_rates; in minstrel_ht_move_sample_rates() 881 group = mi->sample[type].sample_group; in minstrel_ht_next_inc_rate() 897 mi->sample[type].sample_group = group; in minstrel_ht_next_inc_rate() 952 group = mi->sample[MINSTREL_SAMPLE_TYPE_JUMP].sample_group; in minstrel_ht_next_jump_rate() 1002 mi->sample[MINSTREL_SAMPLE_TYPE_JUMP].sample_group = group; in minstrel_ht_next_jump_rate() 1018 rates = mi->sample[MINSTREL_SAMPLE_TYPE_INC].sample_rates; in minstrel_ht_refill_sample_rates() 1043 for (i = 0; i < ARRAY_SIZE(mi->sample); i++) in minstrel_ht_refill_sample_rates() 1044 memcpy(mi->sample[i].cur_sample_rates, mi->sample[i].sample_rates, in minstrel_ht_refill_sample_rates() [all …]
|
| A D | rc80211_minstrel_ht_debugfs.c | 40 for (type = 0; type < ARRAY_SIZE(mi->sample); type++) in minstrel_ht_is_sample_rate() 42 if (mi->sample[type].cur_sample_rates[i] == idx) in minstrel_ht_is_sample_rate()
|
| A D | rc80211_minstrel_ht.h | 189 struct minstrel_sample_category sample[__MINSTREL_SAMPLE_TYPE_MAX]; member
|
| /net/openvswitch/ |
| A D | actions.c | 1003 static int sample(struct datapath *dp, struct sk_buff *skb, in sample() function 1434 err = sample(dp, skb, key, a, last); in do_execute_actions()
|