| /net/netfilter/ |
| A D | nft_limit.c | 24 struct nft_limit *limit; member 39 tokens = priv->limit->tokens + now - priv->limit->last; in nft_limit_eval() 43 priv->limit->last = now; in nft_limit_eval() 113 priv->limit = kmalloc(sizeof(*priv->limit), GFP_KERNEL_ACCOUNT); in nft_limit_init() 114 if (!priv->limit) in nft_limit_init() 149 kfree(priv->limit); in nft_limit_destroy() 161 priv_dst->limit = kmalloc(sizeof(*priv_dst->limit), gfp); in nft_limit_clone() 162 if (!priv_dst->limit) in nft_limit_clone() 206 priv->cost = div64_u64(priv->limit.nsecs, priv->limit.rate); in nft_limit_pkts_init() 234 return nft_limit_clone(&priv_dst->limit, &priv_src->limit, gfp); in nft_limit_pkts_clone() [all …]
|
| A D | nf_conntrack_sip.c | 252 if (++dptr >= limit) in ct_sip_parse_request() 315 if (++dptr >= limit) in sip_follow_continuation() 353 for (limit -= len; dptr < limit; dptr++) { in ct_sip_header_search() 547 limit = ct_sip_header_search(dptr + dataoff, limit, ",", strlen(",")); in ct_sip_parse_param() 548 if (!limit) in ct_sip_parse_param() 558 end = limit; in ct_sip_parse_param() 575 limit = ct_sip_header_search(dptr + dataoff, limit, ",", strlen(",")); in ct_sip_parse_address_param() 576 if (!limit) in ct_sip_parse_address_param() 603 limit = ct_sip_header_search(dptr + dataoff, limit, ",", strlen(",")); in ct_sip_parse_numerical_param() 604 if (!limit) in ct_sip_parse_numerical_param() [all …]
|
| A D | nft_connlimit.c | 18 u32 limit; member 53 if ((count > priv->limit) ^ priv->invert) { in nft_connlimit_do_eval() 64 u32 flags, limit; in nft_connlimit_do_init() local 70 limit = ntohl(nla_get_be32(tb[NFTA_CONNLIMIT_COUNT])); in nft_connlimit_do_init() 85 priv->limit = limit; in nft_connlimit_do_init() 110 if (nla_put_be32(skb, NFTA_CONNLIMIT_COUNT, htonl(priv->limit))) in nft_connlimit_do_dump() 223 priv_dst->limit = priv_src->limit; in nft_connlimit_clone()
|
| /net/sched/ |
| A D | sch_fifo.c | 23 READ_ONCE(sch->limit))) in bfifo_enqueue() 43 if (unlikely(READ_ONCE(sch->limit) == 0)) in pfifo_tail_enqueue() 107 u32 limit = qdisc_dev(sch)->tx_queue_len; in __fifo_init() local 110 limit *= psched_mtu(qdisc_dev(sch)); in __fifo_init() 112 WRITE_ONCE(sch->limit, limit); in __fifo_init() 119 WRITE_ONCE(sch->limit, ctl->limit); in __fifo_init() 125 bypass = sch->limit >= 1; in __fifo_init() 161 struct tc_fifo_qopt opt = { .limit = READ_ONCE(sch->limit) }; in __fifo_dump() 247 ((struct tc_fifo_qopt *)nla_data(nla))->limit = limit; in fifo_set_limit() 257 unsigned int limit, in fifo_create_dflt() argument [all …]
|
| A D | sch_plug.c | 66 u32 limit; member 92 if (likely(sch->qstats.backlog + skb->len <= q->limit)) { in plug_enqueue() 133 q->limit = qdisc_dev(sch)->tx_queue_len in plug_init() 141 q->limit = ctl->limit; in plug_init() 196 q->limit = msg->limit; in plug_change()
|
| A D | sch_sfb.c | 62 u32 limit; /* HARD maximal queue length */ member 296 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue() 305 if (unlikely(time_after(jiffies, limit))) { in sfb_enqueue() 309 time_after(jiffies, limit - q->warmup_time))) { in sfb_enqueue() 483 .limit = 0, 499 u32 limit; in sfb_change() local 514 limit = ctl->limit; in sfb_change() 515 if (limit == 0) in sfb_change() 516 limit = qdisc_dev(sch)->tx_queue_len; in sfb_change() 533 q->limit = limit; in sfb_change() [all …]
|
| A D | sch_gred.c | 34 u32 limit; /* HARD maximal queue length */ member 183 sch->limit)) in gred_enqueue() 339 opt->set.tab[i].limit = q->limit; in gred_offload() 505 if (ctl->limit > sch->limit) in gred_change_vq() 506 q->limit = sch->limit; in gred_change_vq() 508 q->limit = ctl->limit; in gred_change_vq() 660 sch->limit = nla_get_u32(tb[TCA_GRED_LIMIT]); in gred_change() 754 sch->limit = nla_get_u32(tb[TCA_GRED_LIMIT]); in gred_init() 756 sch->limit = qdisc_dev(sch)->tx_queue_len in gred_init() 798 if (nla_put_u32(skb, TCA_GRED_LIMIT, sch->limit)) in gred_dump() [all …]
|
| A D | sch_sfq.c | 457 if (++sch->q.qlen <= q->limit) in sfq_enqueue() 643 int limit; in sfq_change() local 678 limit = q->limit; in sfq_change() 709 if (ctl->limit) { in sfq_change() 710 limit = min_t(u32, ctl->limit, maxdepth * maxflows); in sfq_change() 713 if (limit == 1) { in sfq_change() 721 q->limit = limit; in sfq_change() 733 while (sch->q.qlen > q->limit) { in sfq_change() 793 q->limit = SFQ_MAX_DEPTH; in sfq_init() 823 if (q->limit >= 1) in sfq_init() [all …]
|
| A D | sch_pie.c | 92 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in pie_qdisc_enqueue() 172 u32 limit = nla_get_u32(tb[TCA_PIE_LIMIT]); in pie_change() local 174 WRITE_ONCE(q->params.limit, limit); in pie_change() 175 WRITE_ONCE(sch->limit, limit); in pie_change() 197 while (sch->q.qlen > sch->limit) { in pie_change() 450 sch->limit = q->params.limit; in pie_init() 479 nla_put_u32(skb, TCA_PIE_LIMIT, READ_ONCE(sch->limit)) || in pie_dump()
|
| A D | sch_fq_pie.c | 158 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in fq_pie_qdisc_enqueue() 302 u32 limit = nla_get_u32(tb[TCA_FQ_PIE_LIMIT]); in fq_pie_change() local 304 WRITE_ONCE(q->p_params.limit, limit); in fq_pie_change() 305 WRITE_ONCE(sch->limit, limit); in fq_pie_change() 368 while (sch->q.qlen > sch->limit) { in fq_pie_change() 426 sch->limit = 10 * 1024; in fq_pie_init() 427 q->p_params.limit = sch->limit; in fq_pie_init() 482 if (nla_put_u32(skb, TCA_FQ_PIE_LIMIT, READ_ONCE(sch->limit)) || in fq_pie_dump()
|
| A D | sch_codel.c | 83 if (likely(qdisc_qlen(sch) < sch->limit)) { in codel_qdisc_enqueue() 138 WRITE_ONCE(sch->limit, in codel_change() 146 while (sch->q.qlen > sch->limit) { in codel_change() 164 sch->limit = DEFAULT_CODEL_LIMIT; in codel_init() 178 if (sch->limit >= 1) in codel_init() 199 READ_ONCE(sch->limit)) || in codel_dump()
|
| A D | sch_skbprio.c | 84 if (sch->q.qlen < READ_ONCE(sch->limit)) { in skbprio_enqueue() 174 WRITE_ONCE(sch->limit, ctl->limit); in skbprio_change() 191 sch->limit = 64; in skbprio_init() 202 opt.limit = READ_ONCE(sch->limit); in skbprio_dump()
|
| A D | sch_choke.c | 50 u32 limit; member 267 if (sch->q.qlen < q->limit) { in choke_enqueue() 366 if (ctl->limit > CHOKE_MAX_QUEUE) in choke_change() 369 mask = roundup_pow_of_two(ctl->limit + 1) - 1; in choke_change() 409 WRITE_ONCE(q->limit, ctl->limit); in choke_change() 437 .limit = READ_ONCE(q->limit), in choke_dump()
|
| A D | sch_red.c | 37 u32 limit; /* HARD maximal queue length */ member 203 opt.set.limit = q->limit; in red_offload() 267 if (ctl->limit > 0) { in __red_change() 268 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, ctl->limit, in __red_change() 286 q->limit = ctl->limit; in __red_change() 416 .limit = q->limit, in red_dump()
|
| A D | sch_tbf.c | 100 u32 limit; /* Maximal length of backlog: bytes */ member 438 err = fifo_set_limit(q->qdisc, qopt->limit); in tbf_change() 441 } else if (qopt->limit > 0) { in tbf_change() 442 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, qopt->limit, in tbf_change() 459 q->limit = qopt->limit; in tbf_change() 524 opt.limit = q->limit; in tbf_dump()
|
| A D | sch_dualpi2.c | 391 if (unlikely(qdisc_qlen(sch) >= sch->limit) || in dualpi2_enqueue_skb() 677 static u32 get_memory_limit(struct Qdisc *sch, u32 limit) in get_memory_limit() argument 682 u64 memlim = mul_u32_u32(limit, 2 * psched_mtu(qdisc_dev(sch))); in get_memory_limit() 784 u32 limit = nla_get_u32(tb[TCA_DUALPI2_LIMIT]); in dualpi2_change() local 786 WRITE_ONCE(sch->limit, limit); in dualpi2_change() 787 WRITE_ONCE(q->memory_limit, get_memory_limit(sch, limit)); in dualpi2_change() 872 while (qdisc_qlen(sch) > sch->limit || in dualpi2_change() 892 q->sch->limit = 10000; /* Max 125ms at 1Gbps */ in dualpi2_reset_default() 893 q->memory_limit = get_memory_limit(sch, q->sch->limit); in dualpi2_reset_default() 959 (nla_put_u32(skb, TCA_DUALPI2_LIMIT, READ_ONCE(sch->limit)) || in dualpi2_dump() [all …]
|
| A D | sch_fq_codel.c | 218 if (++sch->q.qlen <= sch->limit && !memory_limited) in fq_codel_enqueue() 424 WRITE_ONCE(sch->limit, in fq_codel_change() 442 while (sch->q.qlen > sch->limit || in fq_codel_change() 474 sch->limit = 10*1024; in fq_codel_init() 516 if (sch->limit >= 1) in fq_codel_init() 543 READ_ONCE(sch->limit)) || in fq_codel_dump()
|
| A D | sch_hhf.c | 402 if (++sch->q.qlen <= sch->limit) in hhf_enqueue() 537 WRITE_ONCE(sch->limit, nla_get_u32(tb[TCA_HHF_BACKLOG_LIMIT])); in hhf_change() 566 while (sch->q.qlen > sch->limit) { in hhf_change() 584 sch->limit = 1000; in hhf_init() 664 if (nla_put_u32(skb, TCA_HHF_BACKLOG_LIMIT, READ_ONCE(sch->limit)) || in hhf_dump()
|
| /net/wireless/ |
| A D | of.c | 30 struct ieee80211_freq_range *limit = &freq_limits[i]; in wiphy_freq_limits_valid_chan() local 32 if (cfg80211_does_bw_fit_range(limit, in wiphy_freq_limits_valid_chan() 109 struct ieee80211_freq_range *limit = &freq_limits[i]; in wiphy_read_of_freq_limits() local 111 p = of_prop_next_u32(prop, p, &limit->start_freq_khz); in wiphy_read_of_freq_limits() 117 p = of_prop_next_u32(prop, p, &limit->end_freq_khz); in wiphy_read_of_freq_limits() 123 if (!limit->start_freq_khz || in wiphy_read_of_freq_limits() 124 !limit->end_freq_khz || in wiphy_read_of_freq_limits() 125 limit->start_freq_khz >= limit->end_freq_khz) { in wiphy_read_of_freq_limits()
|
| /net/devlink/ |
| A D | dev.c | 22 enum devlink_reload_limit limit; member 29 .limit = DEVLINK_RELOAD_LIMIT_NO_RESET, 35 enum devlink_reload_limit limit) in devlink_reload_combination_is_invalid() argument 41 devlink_reload_invalid_combinations[i].limit == limit) in devlink_reload_combination_is_invalid() 370 limit > DEVLINK_RELOAD_LIMIT_MAX)) in devlink_remote_reload_actions_performed() 444 enum devlink_reload_limit limit, in devlink_reload() argument 523 enum devlink_reload_limit limit; in devlink_nl_reload_doit() local 542 limit = DEVLINK_RELOAD_LIMIT_UNSPEC; in devlink_nl_reload_doit() 553 for (limit = 0 ; limit <= DEVLINK_RELOAD_LIMIT_MAX ; limit++) in devlink_nl_reload_doit() 554 if (limits_selected & BIT(limit)) in devlink_nl_reload_doit() [all …]
|
| /net/bridge/netfilter/ |
| A D | ebt_among.c | 28 int start, limit, i; in ebt_mac_wormhash_contains() local 34 limit = wh->table[key + 1]; in ebt_mac_wormhash_contains() 36 for (i = start; i < limit; i++) { in ebt_mac_wormhash_contains() 43 for (i = start; i < limit; i++) { in ebt_mac_wormhash_contains()
|
| /net/rxrpc/ |
| A D | call_event.c | 194 static void rxrpc_transmit_fresh_data(struct rxrpc_call *call, unsigned int limit, in rxrpc_transmit_fresh_data() argument 215 int limit = min(space, max(call->peer->pmtud_jumbo, 1)); in rxrpc_transmit_fresh_data() local 245 } while (req.n < limit && before(seq, send_top)); in rxrpc_transmit_fresh_data() 259 void rxrpc_transmit_some_data(struct rxrpc_call *call, unsigned int limit, in rxrpc_transmit_some_data() argument 277 rxrpc_transmit_fresh_data(call, limit, trace); in rxrpc_transmit_some_data()
|
| /net/batman-adv/ |
| A D | fragmentation.c | 85 int limit = BATADV_FRAG_MAX_FRAG_SIZE; in batadv_frag_size_limit() local 87 limit -= sizeof(struct batadv_frag_packet); in batadv_frag_size_limit() 88 limit *= BATADV_FRAG_MAX_FRAGMENTS; in batadv_frag_size_limit() 90 return limit; in batadv_frag_size_limit()
|
| /net/openvswitch/ |
| A D | conntrack.c | 91 u32 limit; member 924 return ct_limit->limit; in ct_limit_get() 1683 info->default_limit = zone_limit->limit; in ovs_ct_limit_set_zone_limit() 1697 ct_limit->limit = zone_limit->limit; in ovs_ct_limit_set_zone_limit() 1754 .limit = info->default_limit, in ovs_ct_limit_get_default_limit() 1762 u16 zone_id, u32 limit, in __ovs_ct_limit_get_zone_limit() argument 1770 zone_limit.limit = limit; in __ovs_ct_limit_get_zone_limit() 1785 u32 limit; in ovs_ct_limit_get_zone_limit() local 1802 limit = ct_limit_get(info, zone); in ovs_ct_limit_get_zone_limit() 1806 net, info->data, zone, limit, reply); in ovs_ct_limit_get_zone_limit() [all …]
|
| /net/9p/ |
| A D | trans_virtio.c | 171 int limit, char *data, int count) in pack_sg_list() argument 180 BUG_ON(index >= limit); in pack_sg_list() 217 pack_sg_list_p(struct scatterlist *sg, int start, int limit, in pack_sg_list_p() argument 224 BUG_ON(nr_pages > (limit - start)); in pack_sg_list_p() 233 BUG_ON(index >= limit); in pack_sg_list_p()
|