| /net/ethtool/ |
| A D | pse-pd.c | 87 if (st->pw_d_id) in pse_reply_size() 97 if (st->c33_pw_class > 0) in pse_reply_size() 115 if (st->prio_max) in pse_reply_size() 157 if (st->pw_d_id && in pse_fill_reply() 159 st->pw_d_id)) in pse_fill_reply() 169 st->podl_pw_status)) in pse_fill_reply() 174 st->c33_admin_state)) in pse_fill_reply() 179 st->c33_pw_status)) in pse_fill_reply() 184 st->c33_pw_class)) in pse_fill_reply() 189 st->c33_actual_pw)) in pse_fill_reply() [all …]
|
| /net/netfilter/ |
| A D | nf_conntrack_standalone.c | 109 for (i = st->bucket; i < st->htable_size; i++) { in ct_get_next() 134 st->bucket = i; in ct_get_next() 145 st->bucket = i; in ct_get_next() 158 nf_conntrack_get_ht(&st->hash, &st->htable_size); in ct_seq_start() 162 st->bucket = 0; in ct_seq_start() 165 st->skip_elems--; in ct_seq_start() 453 st->found, in ct_cpu_seq_show() 455 st->invalid, in ct_cpu_seq_show() 459 st->insert, in ct_cpu_seq_show() 461 st->drop, in ct_cpu_seq_show() [all …]
|
| A D | xt_recent.c | 483 const struct recent_table *t = st->table; in recent_seq_start() 489 for (st->bucket = 0; st->bucket < ip_list_hash_size; st->bucket++) in recent_seq_start() 499 const struct recent_table *t = st->table; in recent_seq_next() 504 while (head == &t->iphash[st->bucket]) { in recent_seq_next() 505 if (++st->bucket >= ip_list_hash_size) in recent_seq_next() 507 head = t->iphash[st->bucket].next; in recent_seq_next() 522 const struct recent_table *t = st->table; in recent_seq_show() 548 struct recent_iter_state *st; in recent_seq_open() local 550 st = __seq_open_private(file, &recent_seq_ops, sizeof(*st)); in recent_seq_open() 551 if (st == NULL) in recent_seq_open() [all …]
|
| A D | nf_flow_table_procfs.c | 45 const struct nf_flow_table_stat *st = v; in nf_flow_table_cpu_seq_show() local 53 st->count_wq_add, in nf_flow_table_cpu_seq_show() 54 st->count_wq_del, in nf_flow_table_cpu_seq_show() 55 st->count_wq_stats in nf_flow_table_cpu_seq_show()
|
| A D | nfnetlink_log.c | 1029 static struct hlist_node *get_first(struct net *net, struct iter_state *st) in get_first() argument 1032 if (!st) in get_first() 1037 for (st->bucket = 0; st->bucket < INSTANCE_BUCKETS; st->bucket++) { in get_first() 1038 struct hlist_head *head = &log->instance_table[st->bucket]; in get_first() 1046 static struct hlist_node *get_next(struct net *net, struct iter_state *st, in get_next() argument 1054 if (++st->bucket >= INSTANCE_BUCKETS) in get_next() 1058 head = &log->instance_table[st->bucket]; in get_next() 1064 static struct hlist_node *get_idx(struct net *net, struct iter_state *st, in get_idx() argument 1068 head = get_first(net, st); in get_idx() 1071 while (pos && (head = get_next(net, st, head))) in get_idx()
|
| A D | nf_conntrack_expect.c | 572 struct ct_expect_iter_state *st = seq->private; in ct_expect_get_first() local 575 for (st->bucket = 0; st->bucket < nf_ct_expect_hsize; st->bucket++) { in ct_expect_get_first() 576 n = rcu_dereference(hlist_first_rcu(&nf_ct_expect_hash[st->bucket])); in ct_expect_get_first() 586 struct ct_expect_iter_state *st = seq->private; in ct_expect_get_next() local 590 if (++st->bucket >= nf_ct_expect_hsize) in ct_expect_get_next() 592 head = rcu_dereference(hlist_first_rcu(&nf_ct_expect_hash[st->bucket])); in ct_expect_get_next()
|
| A D | nf_conntrack_netlink.c | 2532 htonl(st->insert_failed)) || in ctnetlink_ct_stat_cpu_fill_info() 2537 htonl(st->search_restart)) || in ctnetlink_ct_stat_cpu_fill_info() 2539 htonl(st->clash_resolve)) || in ctnetlink_ct_stat_cpu_fill_info() 2541 htonl(st->chaintoolong))) in ctnetlink_ct_stat_cpu_fill_info() 2563 const struct ip_conntrack_stat *st; in ctnetlink_ct_stat_cpu_dump() local 2568 st = per_cpu_ptr(net->ct.stat, cpu); in ctnetlink_ct_stat_cpu_dump() 2572 cpu, st) < 0) in ctnetlink_ct_stat_cpu_dump() 3709 const struct ip_conntrack_stat *st) in ctnetlink_exp_stat_fill_info() argument 3745 const struct ip_conntrack_stat *st; in ctnetlink_exp_stat_cpu_dump() local 3750 st = per_cpu_ptr(net->ct.stat, cpu); in ctnetlink_exp_stat_cpu_dump() [all …]
|
| A D | nfnetlink_queue.c | 1658 struct iter_state *st = seq->private; in get_first() local 1662 if (!st) in get_first() 1667 for (st->bucket = 0; st->bucket < INSTANCE_BUCKETS; st->bucket++) { in get_first() 1668 if (!hlist_empty(&q->instance_table[st->bucket])) in get_first() 1669 return q->instance_table[st->bucket].first; in get_first() 1676 struct iter_state *st = seq->private; in get_next() local 1683 if (++st->bucket >= INSTANCE_BUCKETS) in get_next() 1687 h = q->instance_table[st->bucket].first; in get_next()
|
| /net/ipv4/ |
| A D | tcp_ipv4.c | 2623 for (; st->bucket <= hinfo->lhash2_mask; st->bucket++) { 2656 ++st->num; 2657 ++st->offset; 2668 ++st->bucket; 2704 for (; st->bucket <= hinfo->ehash_mask; ++st->bucket) { 2733 ++st->num; 2734 ++st->offset; 2744 ++st->bucket; 2825 st->num = 0; 3117 ++st->bucket; [all …]
|
| A D | ip_input.c | 385 struct ip_rt_acct *st = this_cpu_ptr(ip_rt_acct); in ip_rcv_finish_core() local 387 st[idx&0xFF].o_packets++; in ip_rcv_finish_core() 388 st[idx&0xFF].o_bytes += skb->len; in ip_rcv_finish_core() 389 st[(idx>>16)&0xFF].i_packets++; in ip_rcv_finish_core() 390 st[(idx>>16)&0xFF].i_bytes += skb->len; in ip_rcv_finish_core()
|
| /net/ipv6/ |
| A D | addrconf_core.c | 40 __be32 st; in __ipv6_addr_type() local 42 st = addr->s6_addr32[0]; in __ipv6_addr_type() 47 if ((st & htonl(0xE0000000)) != htonl(0x00000000) && in __ipv6_addr_type() 48 (st & htonl(0xE0000000)) != htonl(0xE0000000)) in __ipv6_addr_type() 52 if ((st & htonl(0xFF000000)) == htonl(0xFF000000)) { in __ipv6_addr_type() 59 if ((st & htonl(0xFFC00000)) == htonl(0xFE800000)) in __ipv6_addr_type() 62 if ((st & htonl(0xFFC00000)) == htonl(0xFEC00000)) in __ipv6_addr_type() 65 if ((st & htonl(0xFE000000)) == htonl(0xFC000000)) in __ipv6_addr_type()
|
| /net/sched/ |
| A D | sch_fq.c | 1284 struct tc_fq_qd_stats st; in fq_dump_stats() local 1287 st.pad = 0; in fq_dump_stats() 1291 st.gc_flows = q->stat_gc_flows; in fq_dump_stats() 1292 st.highprio_packets = 0; in fq_dump_stats() 1294 st.tcp_retrans = 0; in fq_dump_stats() 1295 st.throttled = q->stat_throttled; in fq_dump_stats() 1296 st.flows_plimit = q->stat_flows_plimit; in fq_dump_stats() 1301 st.flows = q->flows; in fq_dump_stats() 1302 st.inactive_flows = q->inactive_flows; in fq_dump_stats() 1306 st.ce_mark = q->stat_ce_mark; in fq_dump_stats() [all …]
|
| A D | sch_fq_codel.c | 580 struct tc_fq_codel_xstats st = { in fq_codel_dump_stats() local 585 st.qdisc_stats.maxpacket = q->cstats.maxpacket; in fq_codel_dump_stats() 586 st.qdisc_stats.drop_overlimit = q->drop_overlimit; in fq_codel_dump_stats() 587 st.qdisc_stats.ecn_mark = q->cstats.ecn_mark; in fq_codel_dump_stats() 588 st.qdisc_stats.new_flow_count = q->new_flow_count; in fq_codel_dump_stats() 589 st.qdisc_stats.ce_mark = q->cstats.ce_mark; in fq_codel_dump_stats() 590 st.qdisc_stats.memory_usage = q->memory_usage; in fq_codel_dump_stats() 591 st.qdisc_stats.drop_overmemory = q->drop_overmemory; in fq_codel_dump_stats() 595 st.qdisc_stats.new_flows_len++; in fq_codel_dump_stats() 598 st.qdisc_stats.old_flows_len++; in fq_codel_dump_stats() [all …]
|
| A D | sch_codel.c | 220 struct tc_codel_xstats st = { in codel_dump_stats() local 235 st.drop_next = codel_time_to_us(delta); in codel_dump_stats() 237 st.drop_next = -codel_time_to_us(-delta); in codel_dump_stats() 240 return gnet_stats_copy_app(d, &st, sizeof(st)); in codel_dump_stats()
|
| A D | sch_red.c | 452 struct tc_red_xstats st = {0}; in red_dump_stats() local 466 st.early = q->stats.prob_drop + q->stats.forced_drop; in red_dump_stats() 467 st.pdrop = q->stats.pdrop; in red_dump_stats() 468 st.marked = q->stats.prob_mark + q->stats.forced_mark; in red_dump_stats() 470 return gnet_stats_copy_app(d, &st, sizeof(st)); in red_dump_stats()
|
| A D | sch_pie.c | 501 struct tc_pie_xstats st = { in pie_dump_stats() local 513 st.dq_rate_estimating = q->params.dq_rate_estimator; in pie_dump_stats() 517 st.avg_dq_rate = q->vars.avg_dq_rate * in pie_dump_stats() 520 return gnet_stats_copy_app(d, &st, sizeof(st)); in pie_dump_stats()
|
| A D | sch_sfb.c | 601 struct tc_sfb_xstats st = { in sfb_dump_stats() local 610 st.maxqlen = sfb_compute_qlen(&st.maxprob, &st.avgprob, q); in sfb_dump_stats() 612 return gnet_stats_copy_app(d, &st, sizeof(st)); in sfb_dump_stats()
|
| A D | sch_fq_pie.c | 511 struct tc_fq_pie_xstats st = { in fq_pie_dump_stats() local 524 st.new_flows_len++; in fq_pie_dump_stats() 527 st.old_flows_len++; in fq_pie_dump_stats() 530 return gnet_stats_copy_app(d, &st, sizeof(st)); in fq_pie_dump_stats()
|
| /net/bridge/netfilter/ |
| A D | ebt_stp.c | 141 const struct stp_config_pdu *st; in ebt_stp_mt() local 144 st = skb_header_pointer(skb, sizeof(_stph), in ebt_stp_mt() 146 if (st == NULL) in ebt_stp_mt() 148 return ebt_filter_config(info, st); in ebt_stp_mt()
|
| /net/core/ |
| A D | skbuff.c | 4325 st->root_skb = st->cur_skb = skb; in skb_prepare_seq_read() 4326 st->frag_idx = st->stepped_offset = 0; in skb_prepare_seq_read() 4372 block_limit = skb_headlen(st->cur_skb) + st->stepped_offset; in skb_seq_read() 4375 *data = st->cur_skb->data + (abs_offset - st->stepped_offset); in skb_seq_read() 4382 if (st->frag_idx == 0 && !st->frag_data) in skb_seq_read() 4383 st->stepped_offset += skb_headlen(st->cur_skb); in skb_seq_read() 4385 while (st->frag_idx < skb_shinfo(st->cur_skb)->nr_frags) { in skb_seq_read() 4388 frag = &skb_shinfo(st->cur_skb)->frags[st->frag_idx]; in skb_seq_read() 4430 if (st->root_skb == st->cur_skb && skb_has_frag_list(st->root_skb)) { in skb_seq_read() 4431 st->cur_skb = skb_shinfo(st->root_skb)->frag_list; in skb_seq_read() [all …]
|
| A D | neighbour.c | 3506 st->allocs, in neigh_stat_seq_show() 3507 st->destroys, in neigh_stat_seq_show() 3508 st->hash_grows, in neigh_stat_seq_show() 3510 st->lookups, in neigh_stat_seq_show() 3511 st->hits, in neigh_stat_seq_show() 3513 st->res_failed, in neigh_stat_seq_show() 3515 st->rcv_probes_mcast, in neigh_stat_seq_show() 3516 st->rcv_probes_ucast, in neigh_stat_seq_show() 3519 st->forced_gc_runs, in neigh_stat_seq_show() 3520 st->unres_discards, in neigh_stat_seq_show() [all …]
|
| A D | gen_stats.c | 426 gnet_stats_copy_app(struct gnet_dump *d, void *st, int len) in gnet_stats_copy_app() argument 429 d->xstats = kmemdup(st, len, GFP_ATOMIC); in gnet_stats_copy_app() 436 return gnet_stats_copy(d, TCA_STATS_APP, st, len, in gnet_stats_copy_app()
|
| /net/xfrm/ |
| A D | xfrm_iptfs.c | 533 u32 off, u32 len, struct skb_seq_state *st, u32 copy_len) in iptfs_pskb_add_frags() argument 549 skb_copy_seq_read(st, off, skb_put(skb, copy_len), copy_len)) { in iptfs_pskb_add_frags() 550 XFRM_INC_STATS(dev_net(st->root_skb->dev), in iptfs_pskb_add_frags() 579 struct sk_buff *skb = iptfs_alloc_skb(st->root_skb, skblen, false); in iptfs_pskb_extract_seq() 583 if (skb_copy_seq_read(st, off, skb_put(skb, len), len)) { in iptfs_pskb_extract_seq() 584 XFRM_INC_STATS(dev_net(st->root_skb->dev), LINUX_MIB_XFRMINERROR); in iptfs_pskb_extract_seq() 742 struct skb_seq_state *st, struct sk_buff *skb, in iptfs_reassem_cont() argument 811 if (skb_copy_seq_read(st, data, &xtfs->ra_runt[runtlen], in iptfs_reassem_cont() 870 if (skb_copy_seq_read(st, data, skb_put(newskb, copylen), in iptfs_reassem_cont() 905 if (skb_copy_seq_read(st, data, skb_put(newskb, copylen), in iptfs_reassem_cont() [all …]
|
| /net/mac80211/ |
| A D | rate.c | 91 struct ieee80211_tx_status *st) in rate_control_tx_status() argument 94 struct sta_info *sta = container_of(st->sta, struct sta_info, sta); in rate_control_tx_status() 101 sband = local->hw.wiphy->bands[st->info->band]; in rate_control_tx_status() 105 ref->ops->tx_status_ext(ref->priv, sband, priv_sta, st); in rate_control_tx_status() 106 else if (st->skb) in rate_control_tx_status() 107 ref->ops->tx_status(ref->priv, sband, st->sta, priv_sta, st->skb); in rate_control_tx_status()
|
| /net/batman-adv/ |
| A D | main.c | 597 struct skb_seq_state st; in batadv_skb_crc32() local 604 skb_prepare_seq_read(skb, from, to, &st); in batadv_skb_crc32() 605 while ((len = skb_seq_read(consumed, &data, &st)) != 0) { in batadv_skb_crc32()
|