| /net/ipv6/netfilter/ |
| A D | nf_conntrack_reasm.c | 137 struct frag_queue *fq; in nf_ct_frag6_expire() local 141 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in nf_ct_frag6_expire() 210 if (end < fq->q.len || in nf_ct_frag6_queue() 211 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq->q.len)) { in nf_ct_frag6_queue() 216 fq->q.len = end; in nf_ct_frag6_queue() 229 if (end > fq->q.len) { in nf_ct_frag6_queue() 235 fq->q.len = end; in nf_ct_frag6_queue() 274 fq->ecn |= ecn; in nf_ct_frag6_queue() 283 fq->nhoffset = nhoff; in nf_ct_frag6_queue() 288 fq->q.meat == fq->q.len) { in nf_ct_frag6_queue() [all …]
|
| /net/ieee802154/6lowpan/ |
| A D | reassembly.c | 48 struct frag_queue *fq; in lowpan_frag_expire() local 53 spin_lock(&fq->q.lock); in lowpan_frag_expire() 60 spin_unlock(&fq->q.lock); in lowpan_frag_expire() 111 if (end < fq->q.len || in lowpan_frag_queue() 112 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq->q.len)) in lowpan_frag_queue() 115 fq->q.len = end; in lowpan_frag_queue() 117 if (end > fq->q.len) { in lowpan_frag_queue() 121 fq->q.len = end; in lowpan_frag_queue() 140 fq->q.meat += skb->len; in lowpan_frag_queue() 144 fq->q.meat == fq->q.len) { in lowpan_frag_queue() [all …]
|
| /net/ipv6/ |
| A D | reassembly.c | 77 struct frag_queue *fq; in ip6_frag_expire() local 81 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in ip6_frag_expire() 153 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq->q.len)) in ip6_frag_queue() 156 fq->q.len = end; in ip6_frag_queue() 172 fq->q.len = end; in ip6_frag_queue() 204 fq->ecn |= ecn; in ip6_frag_queue() 215 fq->nhoffset = nhoff; in ip6_frag_queue() 220 fq->q.meat == fq->q.len) { in ip6_frag_queue() 285 nhoff = fq->nhoffset; in ip6_frag_reasm() 380 if (fq) { in ipv6_frag_rcv() [all …]
|
| /net/ipv4/ |
| A D | inet_fragment.c | 133 struct inet_frag_queue *fq = ptr; in inet_frags_free_cb() local 138 spin_lock_bh(&fq->lock); in inet_frags_free_cb() 139 fq->flags |= INET_FRAG_DROP; in inet_frags_free_cb() 141 fq->flags |= INET_FRAG_COMPLETE; in inet_frags_free_cb() 146 spin_unlock_bh(&fq->lock); in inet_frags_free_cb() 148 inet_frag_putn(fq, count); in inet_frags_free_cb() 230 if (timer_delete(&fq->timer)) in inet_frag_kill() 234 struct fqdir *fqdir = fq->fqdir; in inet_frag_kill() 236 fq->flags |= INET_FRAG_COMPLETE; in inet_frag_kill() 380 fq = prev; in inet_frag_find() [all …]
|
| A D | ip_fragment.c | 719 const struct inet_frag_queue *fq = data; in ip4_obj_hashfn() local 721 return jhash2((const u32 *)&fq->key.v4, in ip4_obj_hashfn() 728 const struct inet_frag_queue *fq = ptr; in ip4_obj_cmpfn() local 730 return !!memcmp(&fq->key, key, sizeof(*key)); in ip4_obj_cmpfn()
|
| A D | Kconfig | 679 AQM schemes that do not provide a delay signal. It requires the fq
|
| /net/xdp/ |
| A D | xsk_buff_pool.c | 100 pool->fq = xs->fq_tmp; in xp_create_and_assign_umem() 258 if (!pool->fq || !pool->cq) in xp_assign_dev_shared() 292 if (pool->fq) { in xp_release_deferred() 293 xskq_destroy(pool->fq); in xp_release_deferred() 294 pool->fq = NULL; in xp_release_deferred() 547 pool->fq->invalid_descs++; in __xp_alloc() 548 xskq_cons_release(pool->fq); in __xp_alloc() 556 xskq_cons_release(pool->fq); in __xp_alloc() 607 pool->fq->invalid_descs++; in xp_alloc_new_from_fq() 678 pool->fq->queue_empty_descs++; in xp_alloc_batch() [all …]
|
| A D | xsk_diag.c | 70 if (!err && pool && pool->fq) in xsk_diag_put_umem() 71 err = xsk_diag_put_ring(pool->fq, in xsk_diag_put_umem() 86 du.n_fill_ring_empty = xs->pool ? xskq_nb_queue_empty_descs(xs->pool->fq) : 0; in xsk_diag_put_stats()
|
| A D | xsk.c | 44 pool->fq->ring->flags |= XDP_RING_NEED_WAKEUP; in xsk_set_rx_need_wakeup() 71 pool->fq->ring->flags &= ~XDP_RING_NEED_WAKEUP; in xsk_clear_rx_need_wakeup() 339 __xskq_cons_release(xs->pool->fq); in xsk_flush() 1522 xs->pool ? xskq_nb_queue_empty_descs(xs->pool->fq) : 0; in xsk_getsockopt() 1638 READ_ONCE(xs->pool->fq); in xsk_mmap()
|
| /net/mac80211/ |
| A D | debugfs.c | 80 struct fq *fq = &local->fq; in aqm_read() local 84 spin_lock_bh(&local->fq.lock); in aqm_read() 98 fq->flows_cnt, in aqm_read() 99 fq->backlog, in aqm_read() 100 fq->overmemory, in aqm_read() 101 fq->overlimit, in aqm_read() 102 fq->collisions, in aqm_read() 103 fq->memory_usage, in aqm_read() 104 fq->memory_limit, in aqm_read() 105 fq->limit, in aqm_read() [all …]
|
| A D | tx.c | 1361 struct fq *fq; in codel_dequeue_func() local 1366 fq = &local->fq; in codel_dequeue_func() 1400 local = container_of(fq, struct ieee80211_local, fq); in fq_tin_dequeue_func() 1421 static void fq_skb_free_func(struct fq *fq, in fq_skb_free_func() argument 1436 struct fq *fq = &local->fq; in ieee80211_txq_enqueue() local 1473 struct fq *fq = &local->fq; in ieee80211_txq_remove_vlan() local 1539 struct fq *fq = &local->fq; in ieee80211_txq_purge() local 1572 struct fq *fq = &local->fq; in ieee80211_txq_setup_flows() local 1623 struct fq *fq = &local->fq; in ieee80211_txq_teardown_flows() local 3396 struct fq *fq = &local->fq; in ieee80211_amsdu_aggregate() local [all …]
|
| A D | ht.c | 397 struct fq *fq = &sdata->local->fq; in ieee80211_ba_session_work() local 399 spin_lock_bh(&fq->lock); in ieee80211_ba_session_work() 409 spin_unlock_bh(&fq->lock); in ieee80211_ba_session_work() 421 spin_unlock_bh(&fq->lock); in ieee80211_ba_session_work()
|
| A D | agg-tx.c | 185 struct fq *fq; in ieee80211_agg_stop_txq() local 193 fq = &sdata->local->fq; in ieee80211_agg_stop_txq() 196 spin_lock_bh(&fq->lock); in ieee80211_agg_stop_txq() 198 spin_unlock_bh(&fq->lock); in ieee80211_agg_stop_txq()
|
| A D | util.c | 327 struct fq *fq = &local->fq; in __ieee80211_wake_txqs() local 334 spin_lock(&fq->lock); in __ieee80211_wake_txqs() 361 spin_unlock(&fq->lock); in __ieee80211_wake_txqs() 363 spin_lock(&fq->lock); in __ieee80211_wake_txqs() 376 spin_unlock(&fq->lock); in __ieee80211_wake_txqs() 382 spin_unlock(&fq->lock); in __ieee80211_wake_txqs()
|
| A D | cfg.c | 4827 spin_lock_bh(&local->fq.lock); in ieee80211_get_txq_stats() 4845 txqstats->backlog_packets = local->fq.backlog; in ieee80211_get_txq_stats() 4846 txqstats->backlog_bytes = local->fq.memory_usage; in ieee80211_get_txq_stats() 4847 txqstats->overlimit = local->fq.overlimit; in ieee80211_get_txq_stats() 4848 txqstats->overmemory = local->fq.overmemory; in ieee80211_get_txq_stats() 4849 txqstats->collisions = local->fq.collisions; in ieee80211_get_txq_stats() 4850 txqstats->max_flows = local->fq.flows_cnt; in ieee80211_get_txq_stats() 4855 spin_unlock_bh(&local->fq.lock); in ieee80211_get_txq_stats()
|
| A D | debugfs_netdev.c | 627 spin_lock_bh(&local->fq.lock); in ieee80211_if_fmt_aqm() 646 spin_unlock_bh(&local->fq.lock); in ieee80211_if_fmt_aqm()
|
| A D | debugfs_sta.c | 150 spin_lock_bh(&local->fq.lock); in sta_aqm_read() 182 spin_unlock_bh(&local->fq.lock); in sta_aqm_read()
|
| A D | ieee80211_i.h | 1405 struct fq fq; member
|
| A D | sta_info.c | 2639 spin_lock_bh(&local->fq.lock); in sta_set_tidstats() 2647 spin_unlock_bh(&local->fq.lock); in sta_set_tidstats()
|
| /net/sched/ |
| A D | Kconfig | 472 default "fq" if DEFAULT_FQ
|