| /net/netfilter/ |
| A D | xt_connbytes.c | 28 u_int64_t pkts = 0; in connbytes_mt() local 74 pkts = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].packets); in connbytes_mt() 78 pkts = atomic64_read(&counters[IP_CT_DIR_REPLY].packets); in connbytes_mt() 83 pkts = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].packets) + in connbytes_mt() 87 if (pkts != 0) in connbytes_mt() 88 what = div64_u64(bytes, pkts); in connbytes_mt()
|
| A D | nfnetlink_acct.c | 30 atomic64_t pkts; member 89 atomic64_set(&matching->pkts, 0); in nfnl_acct_new() 133 atomic64_set(&nfacct->pkts, in nfnl_acct_new() 147 u64 pkts, bytes; in nfnl_acct_fill_info() local 161 pkts = atomic64_xchg(&acct->pkts, 0); in nfnl_acct_fill_info() 167 pkts = atomic64_read(&acct->pkts); in nfnl_acct_fill_info() 170 if (nla_put_be64(skb, NFACCT_PKTS, cpu_to_be64(pkts), in nfnl_acct_fill_info() 455 atomic64_inc(&nfacct->pkts); in nfnl_acct_update() 490 atomic64_read(&nfacct->pkts) : atomic64_read(&nfacct->bytes); in nfnl_acct_overquota()
|
| A D | nft_limit.c | 59 const struct nlattr * const tb[], bool pkts) in nft_limit_init() argument 79 if (pkts && priv->burst == 0) in nft_limit_init() 85 if (pkts) { in nft_limit_init()
|
| A D | nf_flow_table_offload.c | 960 if (stats[0].pkts) in flow_offload_work_stats() 963 stats[0].pkts, stats[0].bytes); in flow_offload_work_stats() 964 if (stats[1].pkts) in flow_offload_work_stats() 967 stats[1].pkts, stats[1].bytes); in flow_offload_work_stats()
|
| A D | nft_counter.c | 280 u64_stats_add(&this_cpu->packets, stats->pkts); in nft_counter_offload_stats()
|
| A D | nf_tables_core.c | 192 stats->pkts++; in nft_update_chain_stats()
|
| A D | nf_conntrack_netlink.c | 264 u64 pkts, bytes; in dump_counters() local 267 pkts = atomic64_xchg(&counter[dir].packets, 0); in dump_counters() 270 pkts = atomic64_read(&counter[dir].packets); in dump_counters() 278 if (nla_put_be64(skb, CTA_COUNTERS_PACKETS, cpu_to_be64(pkts), in dump_counters()
|
| A D | nf_tables_api.c | 1928 u64 pkts, bytes; in nft_dump_stats() local 1939 pkts = cpu_stats->pkts; in nft_dump_stats() 1942 total.pkts += pkts; in nft_dump_stats() 1949 if (nla_put_be64(skb, NFTA_COUNTER_PACKETS, cpu_to_be64(total.pkts), in nft_dump_stats() 2240 stats->pkts = be64_to_cpu(nla_get_be64(tb[NFTA_COUNTER_PACKETS])); in nft_stats_alloc()
|
| /net/tipc/ |
| A D | bcast.c | 259 int tipc_bcast_xmit(struct net *net, struct sk_buff_head *pkts, in tipc_bcast_xmit() argument 269 rc = tipc_link_xmit(l, pkts, &xmitq); in tipc_bcast_xmit() 272 __skb_queue_purge(pkts); in tipc_bcast_xmit() 295 selector = msg_link_selector(buf_msg(skb_peek(pkts))); in tipc_rcast_xmit() 300 if (!tipc_msg_pskb_copy(dnode, pkts, &_pkts)) in tipc_rcast_xmit() 378 int tipc_mcast_xmit(struct net *net, struct sk_buff_head *pkts, in tipc_mcast_xmit() argument 392 if (dests->local && !tipc_msg_reassemble(pkts, &localq)) { in tipc_mcast_xmit() 400 skb = skb_peek(pkts); in tipc_mcast_xmit() 417 rc = tipc_rcast_xmit(net, pkts, dests, cong_link_cnt); in tipc_mcast_xmit() 419 rc = tipc_bcast_xmit(net, pkts, cong_link_cnt); in tipc_mcast_xmit() [all …]
|
| A D | bearer.h | 246 void tipc_clone_to_loopback(struct net *net, struct sk_buff_head *pkts); 251 struct sk_buff_head *pkts) in tipc_loopback_trace() argument 254 tipc_clone_to_loopback(net, pkts); in tipc_loopback_trace()
|
| A D | bcast.h | 90 int tipc_mcast_xmit(struct net *net, struct sk_buff_head *pkts, 93 int tipc_bcast_xmit(struct net *net, struct sk_buff_head *pkts,
|
| A D | socket.c | 849 struct sk_buff_head pkts; in tipc_sendmcast() local 878 __skb_queue_head_init(&pkts); in tipc_sendmcast() 883 trace_tipc_sk_sendmcast(sk, skb_peek(&pkts), in tipc_sendmcast() 912 struct sk_buff_head pkts; in tipc_send_group_msg() local 923 __skb_queue_head_init(&pkts); in tipc_send_group_msg() 925 rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); in tipc_send_group_msg() 1093 struct sk_buff_head pkts; in tipc_send_group_bcast() local 1124 __skb_queue_head_init(&pkts); in tipc_send_group_bcast() 1417 struct sk_buff_head pkts; in __tipc_sendmsg() local 1499 __skb_queue_head_init(&pkts); in __tipc_sendmsg() [all …]
|
| A D | crypto.c | 2241 struct sk_buff_head pkts; in tipc_crypto_key_xmit() local 2266 __skb_queue_head_init(&pkts); in tipc_crypto_key_xmit() 2267 __skb_queue_tail(&pkts, skb); in tipc_crypto_key_xmit() 2269 rc = tipc_node_xmit(net, &pkts, dnode, 0); in tipc_crypto_key_xmit() 2271 rc = tipc_bcast_xmit(net, &pkts, &cong_link_cnt); in tipc_crypto_key_xmit()
|
| A D | bearer.c | 763 void tipc_clone_to_loopback(struct net *net, struct sk_buff_head *pkts) in tipc_clone_to_loopback() argument 769 skb_queue_walk(pkts, _skb) { in tipc_clone_to_loopback()
|
| /net/vmw_vsock/ |
| A D | vsock_loopback.c | 118 struct sk_buff_head pkts; in vsock_loopback_work() local 121 skb_queue_head_init(&pkts); in vsock_loopback_work() 124 skb_queue_splice_init(&vsock->pkt_queue, &pkts); in vsock_loopback_work() 127 while ((skb = __skb_dequeue(&pkts))) { in vsock_loopback_work()
|
| /net/netfilter/ipvs/ |
| A D | ip_vs_sync.c | 461 struct ip_vs_conn *cp, int pkts) in ip_vs_sync_conn_needed() argument 527 pkts != sysctl_sync_threshold(ipvs)) in ip_vs_sync_conn_needed() 541 int pkts) in ip_vs_sync_conn_v0() argument 556 if (!ip_vs_sync_conn_needed(ipvs, cp, pkts)) in ip_vs_sync_conn_v0() 618 pkts = atomic_inc_return(&cp->in_pkts); in ip_vs_sync_conn_v0() 620 pkts = sysctl_sync_threshold(ipvs); in ip_vs_sync_conn_v0() 621 ip_vs_sync_conn(ipvs, cp, pkts); in ip_vs_sync_conn_v0() 642 ip_vs_sync_conn_v0(ipvs, cp, pkts); in ip_vs_sync_conn() 649 if (!ip_vs_sync_conn_needed(ipvs, cp, pkts)) in ip_vs_sync_conn() 779 pkts = atomic_inc_return(&cp->in_pkts); in ip_vs_sync_conn() [all …]
|
| A D | ip_vs_core.c | 1919 int ret, pkts; in ip_vs_in_hook() local 2080 pkts = sysctl_sync_threshold(ipvs); in ip_vs_in_hook() 2082 pkts = atomic_inc_return(&cp->in_pkts); in ip_vs_in_hook() 2085 ip_vs_sync_conn(ipvs, cp, pkts); in ip_vs_in_hook()
|
| /net/mctp/test/ |
| A D | utils.c | 91 skb_queue_head_init(&tpq->pkts); in mctp_test_pktqueue_init() 101 skb_queue_tail(&tpq->pkts, skb); in mctp_test_dst_output() 194 skb_queue_purge(&tpq->pkts); in mctp_test_dst_release()
|
| A D | route-test.c | 51 n = tpq.pkts.qlen; in mctp_test_fragment() 64 skb2 = skb_dequeue(&tpq.pkts); in mctp_test_fragment() 154 KUNIT_EXPECT_EQ(test, !!tpq.pkts.qlen, params->input); in mctp_test_rx_input() 156 skb_queue_purge(&tpq.pkts); in mctp_test_rx_input() 991 n = tpq.pkts.qlen; in mctp_test_packet_flow() 994 skb2 = skb_dequeue(&tpq.pkts); in mctp_test_packet_flow() 1026 n = tpq.pkts.qlen; in mctp_test_fragment_flow() 1030 tx_skbs[0] = skb_dequeue(&tpq.pkts); in mctp_test_fragment_flow() 1031 tx_skbs[1] = skb_dequeue(&tpq.pkts); in mctp_test_fragment_flow()
|
| A D | utils.h | 31 struct sk_buff_head pkts; member
|
| /net/ethtool/ |
| A D | coalesce.c | 182 profile[i].pkts); in coalesce_put_profile() 479 ret = ethnl_update_irq_moder(irq_moder, &new_profile[i].pkts, in ethnl_update_profile()
|
| A D | tsinfo.c | 194 if (tsinfo_put_stat(skb, stats->tx_stats.pkts, in tsinfo_put_stats()
|
| /net/wireless/ |
| A D | rdev-ops.h | 840 struct net_device *dev, u32 rate, u32 pkts, u32 intvl) in rdev_set_cqm_txe_config() argument 843 trace_rdev_set_cqm_txe_config(&rdev->wiphy, dev, rate, pkts, intvl); in rdev_set_cqm_txe_config() 844 ret = rdev->ops->set_cqm_txe_config(&rdev->wiphy, dev, rate, pkts, in rdev_set_cqm_txe_config()
|
| A D | trace.h | 1615 u32 pkts, u32 intvl), 1616 TP_ARGS(wiphy, netdev, rate, pkts, intvl), 1621 __field(u32, pkts) 1628 __entry->pkts = pkts; 1632 WIPHY_PR_ARG, NETDEV_PR_ARG, __entry->rate, __entry->pkts,
|
| /net/bluetooth/ |
| A D | hci_core.c | 3616 unsigned int pkts = hdev->sco_pkts; in hci_sched_sco() local 3620 if (!hci_conn_num(hdev, type) || !pkts) in hci_sched_sco() 3627 cnt = &pkts; in hci_sched_sco() 3648 if (!pkts && !hci_dev_test_flag(hdev, HCI_SCO_FLOWCTL)) in hci_sched_sco()
|