| /net/netfilter/ |
| A D | nft_numgen.c | 20 atomic_t *counter; member 29 oval = atomic_read(priv->counter); in nft_ng_inc_gen() 31 } while (atomic_cmpxchg(priv->counter, oval, nval) != oval); in nft_ng_inc_gen() 69 priv->counter = kmalloc(sizeof(*priv->counter), GFP_KERNEL_ACCOUNT); in nft_ng_inc_init() 70 if (!priv->counter) in nft_ng_inc_init() 73 atomic_set(priv->counter, priv->modulus - 1); in nft_ng_inc_init() 82 kfree(priv->counter); in nft_ng_inc_init() 129 kfree(priv->counter); in nft_ng_inc_destroy()
|
| A D | nft_counter.c | 30 struct nft_counter __percpu *counter; member 43 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_do_eval() 83 priv->counter = cpu_stats; in nft_counter_do_init() 98 free_percpu(priv->counter); in nft_counter_do_destroy() 116 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_reset() 139 this_cpu = per_cpu_ptr(priv->counter, cpu); in nft_counter_fetch() 256 priv_clone->counter = cpu_stats; in nft_counter_clone() 276 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_offload_stats()
|
| A D | nf_conntrack_standalone.c | 291 struct nf_conn_counter *counter; in seq_print_acct() local 297 counter = acct->counter; in seq_print_acct() 299 (unsigned long long)atomic64_read(&counter[dir].packets), in seq_print_acct() 300 (unsigned long long)atomic64_read(&counter[dir].bytes)); in seq_print_acct()
|
| A D | xt_connbytes.c | 40 counters = acct->counter; in connbytes_mt()
|
| A D | nf_conntrack_core.c | 955 struct nf_conn_counter *counter = acct->counter; in nf_ct_acct_add() local 957 atomic64_add(packets, &counter[dir].packets); in nf_ct_acct_add() 958 atomic64_add(bytes, &counter[dir].bytes); in nf_ct_acct_add() 970 struct nf_conn_counter *counter = acct->counter; in nf_ct_acct_merge() local 974 bytes = atomic64_read(&counter[CTINFO2DIR(ctinfo)].bytes); in nf_ct_acct_merge()
|
| A D | x_tables.c | 1926 struct xt_counters *counter) in xt_percpu_counter_alloc() argument 1928 BUILD_BUG_ON(XT_PCPU_BLOCK_SIZE < (sizeof(*counter) * 2)); in xt_percpu_counter_alloc() 1939 counter->pcnt = (__force unsigned long)(state->mem + state->off); in xt_percpu_counter_alloc() 1940 state->off += sizeof(*counter); in xt_percpu_counter_alloc() 1941 if (state->off > (XT_PCPU_BLOCK_SIZE - sizeof(*counter))) { in xt_percpu_counter_alloc()
|
| A D | nft_ct.c | 130 count = nft_ct_get_eval_counter(acct->counter, in nft_ct_get_eval() 140 pcnt = nft_ct_get_eval_counter(acct->counter, in nft_ct_get_eval() 142 bcnt = nft_ct_get_eval_counter(acct->counter, in nft_ct_get_eval()
|
| A D | nf_conntrack_netlink.c | 262 struct nf_conn_counter *counter = acct->counter; in dump_counters() local 267 pkts = atomic64_xchg(&counter[dir].packets, 0); in dump_counters() 268 bytes = atomic64_xchg(&counter[dir].bytes, 0); in dump_counters() 270 pkts = atomic64_read(&counter[dir].packets); in dump_counters() 271 bytes = atomic64_read(&counter[dir].bytes); in dump_counters()
|
| /net/openvswitch/ |
| A D | flow_table.c | 205 u64 counter; in tbl_mask_array_reset_counters() local 210 counter = stats->usage_cntrs[i]; in tbl_mask_array_reset_counters() 213 ma->masks_usage_zero_cntr[i] += counter; in tbl_mask_array_reset_counters() 1101 return (s64)mc_b->counter - (s64)mc_a->counter; in compare_mask_and_count() 1128 masks_and_count[i].counter = 0; in ovs_flow_masks_rebalance() 1133 u64 counter; in ovs_flow_masks_rebalance() local 1138 counter = stats->usage_cntrs[i]; in ovs_flow_masks_rebalance() 1141 masks_and_count[i].counter += counter; in ovs_flow_masks_rebalance() 1145 masks_and_count[i].counter -= ma->masks_usage_zero_cntr[i]; in ovs_flow_masks_rebalance() 1150 ma->masks_usage_zero_cntr[i] += masks_and_count[i].counter; in ovs_flow_masks_rebalance()
|
| A D | flow_table.h | 38 u64 counter; member
|
| /net/netfilter/ipset/ |
| A D | ip_set_core.c | 526 return (u64)atomic64_read(&(counter)->bytes); in ip_set_get_bytes() 532 return (u64)atomic64_read(&(counter)->packets); in ip_set_get_packets() 539 cpu_to_be64(ip_set_get_bytes(counter)), in ip_set_put_counter() 542 cpu_to_be64(ip_set_get_packets(counter)), in ip_set_put_counter() 595 return counter == match; in ip_set_match_counter() 597 return counter != match; in ip_set_match_counter() 599 return counter < match; in ip_set_match_counter() 601 return counter > match; in ip_set_match_counter() 624 ip_set_add_bytes(ext->bytes, counter); in ip_set_update_counter() 625 ip_set_add_packets(ext->packets, counter); in ip_set_update_counter() [all …]
|
| /net/netfilter/ipvs/ |
| A D | ip_vs_lblc.c | 109 int counter; /* counter for no expire */ member 304 if ((tbl->counter % COUNT_FOR_FULL_EXPIRATION) == 0) { in ip_vs_lblc_check_expire() 307 tbl->counter = 1; in ip_vs_lblc_check_expire() 312 tbl->counter++; in ip_vs_lblc_check_expire() 367 tbl->counter = 1; in ip_vs_lblc_init_svc()
|
| A D | ip_vs_lblcr.c | 279 int counter; /* counter for no expire */ member 468 if ((tbl->counter % COUNT_FOR_FULL_EXPIRATION) == 0) { in ip_vs_lblcr_check_expire() 471 tbl->counter = 1; in ip_vs_lblcr_check_expire() 476 tbl->counter++; in ip_vs_lblcr_check_expire() 530 tbl->counter = 1; in ip_vs_lblcr_init_svc()
|
| /net/ipv4/netfilter/ |
| A D | nf_nat_snmp_basic.asn1 | 51 counter-value 60 big-counter-value
|
| A D | arp_tables.c | 223 struct xt_counters *counter; in arpt_do_table() local 230 counter = xt_get_this_cpu_counter(&e->counters); in arpt_do_table() 231 ADD_COUNTER(*counter, arp_hdr_len(skb->dev), 1); in arpt_do_table()
|
| A D | ip_tables.c | 280 struct xt_counters *counter; in ipt_do_table() local 297 counter = xt_get_this_cpu_counter(&e->counters); in ipt_do_table() 298 ADD_COUNTER(*counter, skb->len, 1); in ipt_do_table()
|
| /net/ipv6/ |
| A D | tcp_ao.c | 23 u8 counter; in tcp_v6_ao_calc_key() member 36 tmp->counter = 1; in tcp_v6_ao_calc_key()
|
| /net/sctp/ |
| A D | objcnt.c | 69 atomic_read(sctp_dbg_objcnt[i].counter)); in sctp_objcnt_seq_show()
|
| /net/sched/ |
| A D | sch_netem.c | 95 u32 counter; member 553 q->counter < q->gap - 1 || /* inside last reordering gap */ in netem_enqueue() 602 ++q->counter; in netem_enqueue() 610 q->counter = 0; in netem_enqueue() 1067 q->counter = 0; in netem_change()
|
| /net/smc/ |
| A D | smc_cdc.c | 246 curs.acurs.counter = atomic64_read(&conn->local_tx_ctrl.prod.acurs); in smcd_cdc_msg_send() 249 curs.acurs.counter = atomic64_read(&conn->local_tx_ctrl.cons.acurs); in smcd_cdc_msg_send()
|
| /net/rxrpc/ |
| A D | rxkad.c | 1218 u32 counter = READ_ONCE(conn->channels[i].call_counter); in rxkad_verify_response() local 1226 if (call_id < counter) { in rxkad_verify_response() 1232 if (call_id > counter) { in rxkad_verify_response()
|
| /net/ipv6/netfilter/ |
| A D | ip6_tables.c | 302 struct xt_counters *counter; in ip6t_do_table() local 320 counter = xt_get_this_cpu_counter(&e->counters); in ip6t_do_table() 321 ADD_COUNTER(*counter, skb->len, 1); in ip6t_do_table()
|
| /net/batman-adv/ |
| A D | types.h | 2008 atomic_t counter; member
|
| /net/ceph/ |
| A D | messenger_v2.c | 832 u64 counter; in gcm_inc_nonce() local 834 counter = le64_to_cpu(nonce->counter); in gcm_inc_nonce() 835 nonce->counter = cpu_to_le64(counter + 1); in gcm_inc_nonce()
|
| /net/devlink/ |
| A D | dpipe.c | 409 entry->counter)) in devlink_dpipe_entry_put()
|