Lines Matching refs:tbl
57 static int pneigh_ifdown_and_unlock(struct neigh_table *tbl,
123 atomic_dec(&n->tbl->gc_entries); in neigh_mark_dead()
133 write_lock_bh(&n->tbl->lock); in neigh_update_gc_list()
147 atomic_dec(&n->tbl->gc_entries); in neigh_update_gc_list()
150 list_add_tail(&n->gc_list, &n->tbl->gc_list); in neigh_update_gc_list()
151 atomic_inc(&n->tbl->gc_entries); in neigh_update_gc_list()
155 write_unlock_bh(&n->tbl->lock); in neigh_update_gc_list()
162 write_lock_bh(&n->tbl->lock); in neigh_update_managed_list()
173 list_add_tail(&n->managed_list, &n->tbl->managed_list); in neigh_update_managed_list()
176 write_unlock_bh(&n->tbl->lock); in neigh_update_managed_list()
209 struct neigh_table *tbl) in neigh_del() argument
218 lockdep_is_held(&tbl->lock)); in neigh_del()
229 bool neigh_remove_one(struct neighbour *ndel, struct neigh_table *tbl) in neigh_remove_one() argument
237 nht = rcu_dereference_protected(tbl->nht, in neigh_remove_one()
238 lockdep_is_held(&tbl->lock)); in neigh_remove_one()
239 hash_val = tbl->hash(pkey, ndel->dev, nht->hash_rnd); in neigh_remove_one()
244 lockdep_is_held(&tbl->lock)))) { in neigh_remove_one()
246 return neigh_del(n, np, tbl); in neigh_remove_one()
252 static int neigh_forced_gc(struct neigh_table *tbl) in neigh_forced_gc() argument
254 int max_clean = atomic_read(&tbl->gc_entries) - tbl->gc_thresh2; in neigh_forced_gc()
259 NEIGH_CACHE_STAT_INC(tbl, forced_gc_runs); in neigh_forced_gc()
261 write_lock_bh(&tbl->lock); in neigh_forced_gc()
263 list_for_each_entry_safe(n, tmp, &tbl->gc_list, gc_list) { in neigh_forced_gc()
270 (tbl->is_multicast && in neigh_forced_gc()
271 tbl->is_multicast(n->primary_key)) || in neigh_forced_gc()
276 if (remove && neigh_remove_one(n, tbl)) in neigh_forced_gc()
283 tbl->last_flush = jiffies; in neigh_forced_gc()
285 write_unlock_bh(&tbl->lock); in neigh_forced_gc()
320 static void neigh_flush_dev(struct neigh_table *tbl, struct net_device *dev, in neigh_flush_dev() argument
326 nht = rcu_dereference_protected(tbl->nht, in neigh_flush_dev()
327 lockdep_is_held(&tbl->lock)); in neigh_flush_dev()
334 lockdep_is_held(&tbl->lock))) != NULL) { in neigh_flush_dev()
345 lockdep_is_held(&tbl->lock))); in neigh_flush_dev()
374 void neigh_changeaddr(struct neigh_table *tbl, struct net_device *dev) in neigh_changeaddr() argument
376 write_lock_bh(&tbl->lock); in neigh_changeaddr()
377 neigh_flush_dev(tbl, dev, false); in neigh_changeaddr()
378 write_unlock_bh(&tbl->lock); in neigh_changeaddr()
382 static int __neigh_ifdown(struct neigh_table *tbl, struct net_device *dev, in __neigh_ifdown() argument
385 write_lock_bh(&tbl->lock); in __neigh_ifdown()
386 neigh_flush_dev(tbl, dev, skip_perm); in __neigh_ifdown()
387 pneigh_ifdown_and_unlock(tbl, dev); in __neigh_ifdown()
389 del_timer_sync(&tbl->proxy_timer); in __neigh_ifdown()
390 pneigh_queue_purge(&tbl->proxy_queue); in __neigh_ifdown()
394 int neigh_carrier_down(struct neigh_table *tbl, struct net_device *dev) in neigh_carrier_down() argument
396 __neigh_ifdown(tbl, dev, true); in neigh_carrier_down()
401 int neigh_ifdown(struct neigh_table *tbl, struct net_device *dev) in neigh_ifdown() argument
403 __neigh_ifdown(tbl, dev, false); in neigh_ifdown()
408 static struct neighbour *neigh_alloc(struct neigh_table *tbl, in neigh_alloc() argument
419 entries = atomic_inc_return(&tbl->gc_entries) - 1; in neigh_alloc()
420 if (entries >= tbl->gc_thresh3 || in neigh_alloc()
421 (entries >= tbl->gc_thresh2 && in neigh_alloc()
422 time_after(now, tbl->last_flush + 5 * HZ))) { in neigh_alloc()
423 if (!neigh_forced_gc(tbl) && in neigh_alloc()
424 entries >= tbl->gc_thresh3) { in neigh_alloc()
426 tbl->id); in neigh_alloc()
427 NEIGH_CACHE_STAT_INC(tbl, table_fulls); in neigh_alloc()
433 n = kzalloc(tbl->entry_size + dev->neigh_priv_len, GFP_ATOMIC); in neigh_alloc()
445 n->parms = neigh_parms_clone(&tbl->parms); in neigh_alloc()
448 NEIGH_CACHE_STAT_INC(tbl, allocs); in neigh_alloc()
449 n->tbl = tbl; in neigh_alloc()
455 atomic_inc(&tbl->entries); in neigh_alloc()
461 atomic_dec(&tbl->gc_entries); in neigh_alloc()
516 static struct neigh_hash_table *neigh_hash_grow(struct neigh_table *tbl, in neigh_hash_grow() argument
522 NEIGH_CACHE_STAT_INC(tbl, hash_grows); in neigh_hash_grow()
524 old_nht = rcu_dereference_protected(tbl->nht, in neigh_hash_grow()
525 lockdep_is_held(&tbl->lock)); in neigh_hash_grow()
534 lockdep_is_held(&tbl->lock)); in neigh_hash_grow()
537 hash = tbl->hash(n->primary_key, n->dev, in neigh_hash_grow()
542 lockdep_is_held(&tbl->lock)); in neigh_hash_grow()
547 lockdep_is_held(&tbl->lock))); in neigh_hash_grow()
552 rcu_assign_pointer(tbl->nht, new_nht); in neigh_hash_grow()
557 struct neighbour *neigh_lookup(struct neigh_table *tbl, const void *pkey, in neigh_lookup() argument
562 NEIGH_CACHE_STAT_INC(tbl, lookups); in neigh_lookup()
565 n = __neigh_lookup_noref(tbl, pkey, dev); in neigh_lookup()
569 NEIGH_CACHE_STAT_INC(tbl, hits); in neigh_lookup()
577 struct neighbour *neigh_lookup_nodev(struct neigh_table *tbl, struct net *net, in neigh_lookup_nodev() argument
581 unsigned int key_len = tbl->key_len; in neigh_lookup_nodev()
585 NEIGH_CACHE_STAT_INC(tbl, lookups); in neigh_lookup_nodev()
588 nht = rcu_dereference_bh(tbl->nht); in neigh_lookup_nodev()
589 hash_val = tbl->hash(pkey, NULL, nht->hash_rnd) >> (32 - nht->hash_shift); in neigh_lookup_nodev()
598 NEIGH_CACHE_STAT_INC(tbl, hits); in neigh_lookup_nodev()
609 ___neigh_create(struct neigh_table *tbl, const void *pkey, in ___neigh_create() argument
613 u32 hash_val, key_len = tbl->key_len; in ___neigh_create()
618 n = neigh_alloc(tbl, dev, flags, exempt_from_gc); in ___neigh_create()
619 trace_neigh_create(tbl, dev, pkey, n, exempt_from_gc); in ___neigh_create()
630 if (tbl->constructor && (error = tbl->constructor(n)) < 0) { in ___neigh_create()
652 write_lock_bh(&tbl->lock); in ___neigh_create()
653 nht = rcu_dereference_protected(tbl->nht, in ___neigh_create()
654 lockdep_is_held(&tbl->lock)); in ___neigh_create()
656 if (atomic_read(&tbl->entries) > (1 << nht->hash_shift)) in ___neigh_create()
657 nht = neigh_hash_grow(tbl, nht->hash_shift + 1); in ___neigh_create()
659 hash_val = tbl->hash(n->primary_key, dev, nht->hash_rnd) >> (32 - nht->hash_shift); in ___neigh_create()
667 lockdep_is_held(&tbl->lock)); in ___neigh_create()
670 lockdep_is_held(&tbl->lock))) { in ___neigh_create()
681 list_add_tail(&n->gc_list, &n->tbl->gc_list); in ___neigh_create()
683 list_add_tail(&n->managed_list, &n->tbl->managed_list); in ___neigh_create()
688 lockdep_is_held(&tbl->lock))); in ___neigh_create()
690 write_unlock_bh(&tbl->lock); in ___neigh_create()
696 write_unlock_bh(&tbl->lock); in ___neigh_create()
699 atomic_dec(&tbl->gc_entries); in ___neigh_create()
704 struct neighbour *__neigh_create(struct neigh_table *tbl, const void *pkey, in __neigh_create() argument
707 return ___neigh_create(tbl, pkey, dev, 0, false, want_ref); in __neigh_create()
737 struct pneigh_entry *__pneigh_lookup(struct neigh_table *tbl, in __pneigh_lookup() argument
740 unsigned int key_len = tbl->key_len; in __pneigh_lookup()
743 return __pneigh_lookup_1(tbl->phash_buckets[hash_val], in __pneigh_lookup()
748 struct pneigh_entry * pneigh_lookup(struct neigh_table *tbl, in pneigh_lookup() argument
753 unsigned int key_len = tbl->key_len; in pneigh_lookup()
756 read_lock_bh(&tbl->lock); in pneigh_lookup()
757 n = __pneigh_lookup_1(tbl->phash_buckets[hash_val], in pneigh_lookup()
759 read_unlock_bh(&tbl->lock); in pneigh_lookup()
775 if (tbl->pconstructor && tbl->pconstructor(n)) { in pneigh_lookup()
782 write_lock_bh(&tbl->lock); in pneigh_lookup()
783 n->next = tbl->phash_buckets[hash_val]; in pneigh_lookup()
784 tbl->phash_buckets[hash_val] = n; in pneigh_lookup()
785 write_unlock_bh(&tbl->lock); in pneigh_lookup()
792 int pneigh_delete(struct neigh_table *tbl, struct net *net, const void *pkey, in pneigh_delete() argument
796 unsigned int key_len = tbl->key_len; in pneigh_delete()
799 write_lock_bh(&tbl->lock); in pneigh_delete()
800 for (np = &tbl->phash_buckets[hash_val]; (n = *np) != NULL; in pneigh_delete()
805 write_unlock_bh(&tbl->lock); in pneigh_delete()
806 if (tbl->pdestructor) in pneigh_delete()
807 tbl->pdestructor(n); in pneigh_delete()
813 write_unlock_bh(&tbl->lock); in pneigh_delete()
817 static int pneigh_ifdown_and_unlock(struct neigh_table *tbl, in pneigh_ifdown_and_unlock() argument
824 np = &tbl->phash_buckets[h]; in pneigh_ifdown_and_unlock()
835 write_unlock_bh(&tbl->lock); in pneigh_ifdown_and_unlock()
839 if (tbl->pdestructor) in pneigh_ifdown_and_unlock()
840 tbl->pdestructor(n); in pneigh_ifdown_and_unlock()
863 NEIGH_CACHE_STAT_INC(neigh->tbl, destroys); in neigh_destroy()
887 atomic_dec(&neigh->tbl->entries); in neigh_destroy()
918 struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); in neigh_periodic_work() local
924 NEIGH_CACHE_STAT_INC(tbl, periodic_gc_runs); in neigh_periodic_work()
926 write_lock_bh(&tbl->lock); in neigh_periodic_work()
927 nht = rcu_dereference_protected(tbl->nht, in neigh_periodic_work()
928 lockdep_is_held(&tbl->lock)); in neigh_periodic_work()
934 if (time_after(jiffies, tbl->last_rand + 300 * HZ)) { in neigh_periodic_work()
936 tbl->last_rand = jiffies; in neigh_periodic_work()
937 list_for_each_entry(p, &tbl->parms_list, list) in neigh_periodic_work()
942 if (atomic_read(&tbl->entries) < tbl->gc_thresh1) in neigh_periodic_work()
949 lockdep_is_held(&tbl->lock))) != NULL) { in neigh_periodic_work()
982 write_unlock_bh(&tbl->lock); in neigh_periodic_work()
984 write_lock_bh(&tbl->lock); in neigh_periodic_work()
985 nht = rcu_dereference_protected(tbl->nht, in neigh_periodic_work()
986 lockdep_is_held(&tbl->lock)); in neigh_periodic_work()
993 queue_delayed_work(system_power_efficient_wq, &tbl->gc_work, in neigh_periodic_work()
994 NEIGH_VAR(&tbl->parms, BASE_REACHABLE_TIME) >> 1); in neigh_periodic_work()
995 write_unlock_bh(&tbl->lock); in neigh_periodic_work()
1012 NEIGH_CACHE_STAT_INC(neigh->tbl, res_failed); in neigh_invalidate()
1191 NEIGH_CACHE_STAT_INC(neigh->tbl, unres_discards); in __neigh_event_send()
1469 struct neighbour *neigh_event_ns(struct neigh_table *tbl, in neigh_event_ns() argument
1473 struct neighbour *neigh = __neigh_lookup(tbl, saddr, dev, in neigh_event_ns()
1486 __be16 prot = n->tbl->protocol; in neigh_hh_init()
1568 struct neigh_table *tbl = container_of(work, struct neigh_table, in neigh_managed_work() local
1572 write_lock_bh(&tbl->lock); in neigh_managed_work()
1573 list_for_each_entry(neigh, &tbl->managed_list, managed_list) in neigh_managed_work()
1575 queue_delayed_work(system_power_efficient_wq, &tbl->managed_work, in neigh_managed_work()
1576 NEIGH_VAR(&tbl->parms, DELAY_PROBE_TIME)); in neigh_managed_work()
1577 write_unlock_bh(&tbl->lock); in neigh_managed_work()
1582 struct neigh_table *tbl = from_timer(tbl, t, proxy_timer); in neigh_proxy_process() local
1587 spin_lock(&tbl->proxy_queue.lock); in neigh_proxy_process()
1589 skb_queue_walk_safe(&tbl->proxy_queue, skb, n) { in neigh_proxy_process()
1595 __skb_unlink(skb, &tbl->proxy_queue); in neigh_proxy_process()
1596 if (tbl->proxy_redo && netif_running(dev)) { in neigh_proxy_process()
1598 tbl->proxy_redo(skb); in neigh_proxy_process()
1608 del_timer(&tbl->proxy_timer); in neigh_proxy_process()
1610 mod_timer(&tbl->proxy_timer, jiffies + sched_next); in neigh_proxy_process()
1611 spin_unlock(&tbl->proxy_queue.lock); in neigh_proxy_process()
1614 void pneigh_enqueue(struct neigh_table *tbl, struct neigh_parms *p, in pneigh_enqueue() argument
1620 if (tbl->proxy_queue.qlen > NEIGH_VAR(p, PROXY_QLEN)) { in pneigh_enqueue()
1628 spin_lock(&tbl->proxy_queue.lock); in pneigh_enqueue()
1629 if (del_timer(&tbl->proxy_timer)) { in pneigh_enqueue()
1630 if (time_before(tbl->proxy_timer.expires, sched_next)) in pneigh_enqueue()
1631 sched_next = tbl->proxy_timer.expires; in pneigh_enqueue()
1635 __skb_queue_tail(&tbl->proxy_queue, skb); in pneigh_enqueue()
1636 mod_timer(&tbl->proxy_timer, sched_next); in pneigh_enqueue()
1637 spin_unlock(&tbl->proxy_queue.lock); in pneigh_enqueue()
1641 static inline struct neigh_parms *lookup_neigh_parms(struct neigh_table *tbl, in lookup_neigh_parms() argument
1646 list_for_each_entry(p, &tbl->parms_list, list) { in lookup_neigh_parms()
1656 struct neigh_table *tbl) in neigh_parms_alloc() argument
1662 p = kmemdup(&tbl->parms, sizeof(*p), GFP_KERNEL); in neigh_parms_alloc()
1664 p->tbl = tbl; in neigh_parms_alloc()
1679 write_lock_bh(&tbl->lock); in neigh_parms_alloc()
1680 list_add(&p->list, &tbl->parms.list); in neigh_parms_alloc()
1681 write_unlock_bh(&tbl->lock); in neigh_parms_alloc()
1697 void neigh_parms_release(struct neigh_table *tbl, struct neigh_parms *parms) in neigh_parms_release() argument
1699 if (!parms || parms == &tbl->parms) in neigh_parms_release()
1701 write_lock_bh(&tbl->lock); in neigh_parms_release()
1704 write_unlock_bh(&tbl->lock); in neigh_parms_release()
1719 void neigh_table_init(int index, struct neigh_table *tbl) in neigh_table_init() argument
1724 INIT_LIST_HEAD(&tbl->parms_list); in neigh_table_init()
1725 INIT_LIST_HEAD(&tbl->gc_list); in neigh_table_init()
1726 INIT_LIST_HEAD(&tbl->managed_list); in neigh_table_init()
1728 list_add(&tbl->parms.list, &tbl->parms_list); in neigh_table_init()
1729 write_pnet(&tbl->parms.net, &init_net); in neigh_table_init()
1730 refcount_set(&tbl->parms.refcnt, 1); in neigh_table_init()
1731 tbl->parms.reachable_time = in neigh_table_init()
1732 neigh_rand_reach_time(NEIGH_VAR(&tbl->parms, BASE_REACHABLE_TIME)); in neigh_table_init()
1734 tbl->stats = alloc_percpu(struct neigh_statistics); in neigh_table_init()
1735 if (!tbl->stats) in neigh_table_init()
1739 if (!proc_create_seq_data(tbl->id, 0, init_net.proc_net_stat, in neigh_table_init()
1740 &neigh_stat_seq_ops, tbl)) in neigh_table_init()
1744 RCU_INIT_POINTER(tbl->nht, neigh_hash_alloc(3)); in neigh_table_init()
1747 tbl->phash_buckets = kzalloc(phsize, GFP_KERNEL); in neigh_table_init()
1749 if (!tbl->nht || !tbl->phash_buckets) in neigh_table_init()
1752 if (!tbl->entry_size) in neigh_table_init()
1753 tbl->entry_size = ALIGN(offsetof(struct neighbour, primary_key) + in neigh_table_init()
1754 tbl->key_len, NEIGH_PRIV_ALIGN); in neigh_table_init()
1756 WARN_ON(tbl->entry_size % NEIGH_PRIV_ALIGN); in neigh_table_init()
1758 rwlock_init(&tbl->lock); in neigh_table_init()
1760 INIT_DEFERRABLE_WORK(&tbl->gc_work, neigh_periodic_work); in neigh_table_init()
1761 queue_delayed_work(system_power_efficient_wq, &tbl->gc_work, in neigh_table_init()
1762 tbl->parms.reachable_time); in neigh_table_init()
1763 INIT_DEFERRABLE_WORK(&tbl->managed_work, neigh_managed_work); in neigh_table_init()
1764 queue_delayed_work(system_power_efficient_wq, &tbl->managed_work, 0); in neigh_table_init()
1766 timer_setup(&tbl->proxy_timer, neigh_proxy_process, 0); in neigh_table_init()
1767 skb_queue_head_init_class(&tbl->proxy_queue, in neigh_table_init()
1770 tbl->last_flush = now; in neigh_table_init()
1771 tbl->last_rand = now + tbl->parms.reachable_time * 20; in neigh_table_init()
1773 neigh_tables[index] = tbl; in neigh_table_init()
1777 int neigh_table_clear(int index, struct neigh_table *tbl) in neigh_table_clear() argument
1781 cancel_delayed_work_sync(&tbl->managed_work); in neigh_table_clear()
1782 cancel_delayed_work_sync(&tbl->gc_work); in neigh_table_clear()
1783 del_timer_sync(&tbl->proxy_timer); in neigh_table_clear()
1784 pneigh_queue_purge(&tbl->proxy_queue); in neigh_table_clear()
1785 neigh_ifdown(tbl, NULL); in neigh_table_clear()
1786 if (atomic_read(&tbl->entries)) in neigh_table_clear()
1789 call_rcu(&rcu_dereference_protected(tbl->nht, 1)->rcu, in neigh_table_clear()
1791 tbl->nht = NULL; in neigh_table_clear()
1793 kfree(tbl->phash_buckets); in neigh_table_clear()
1794 tbl->phash_buckets = NULL; in neigh_table_clear()
1796 remove_proc_entry(tbl->id, init_net.proc_net_stat); in neigh_table_clear()
1798 free_percpu(tbl->stats); in neigh_table_clear()
1799 tbl->stats = NULL; in neigh_table_clear()
1807 struct neigh_table *tbl = NULL; in neigh_find_table() local
1811 tbl = neigh_tables[NEIGH_ARP_TABLE]; in neigh_find_table()
1814 tbl = neigh_tables[NEIGH_ND_TABLE]; in neigh_find_table()
1817 tbl = neigh_tables[NEIGH_DN_TABLE]; in neigh_find_table()
1821 return tbl; in neigh_find_table()
1847 struct neigh_table *tbl; in neigh_delete() local
1871 tbl = neigh_find_table(ndm->ndm_family); in neigh_delete()
1872 if (tbl == NULL) in neigh_delete()
1875 if (nla_len(dst_attr) < (int)tbl->key_len) { in neigh_delete()
1881 err = pneigh_delete(tbl, net, nla_data(dst_attr), dev); in neigh_delete()
1888 neigh = neigh_lookup(tbl, nla_data(dst_attr), dev); in neigh_delete()
1897 write_lock_bh(&tbl->lock); in neigh_delete()
1899 neigh_remove_one(neigh, tbl); in neigh_delete()
1900 write_unlock_bh(&tbl->lock); in neigh_delete()
1914 struct neigh_table *tbl; in neigh_add() local
1957 tbl = neigh_find_table(ndm->ndm_family); in neigh_add()
1958 if (tbl == NULL) in neigh_add()
1961 if (nla_len(tb[NDA_DST]) < (int)tbl->key_len) { in neigh_add()
1980 pn = pneigh_lookup(tbl, net, dst, dev, 1); in neigh_add()
1995 if (tbl->allow_add && !tbl->allow_add(dev, extack)) { in neigh_add()
2000 neigh = neigh_lookup(tbl, dst, dev); in neigh_add()
2016 neigh = ___neigh_create(tbl, dst, dev, in neigh_add()
2106 static int neightbl_fill_info(struct sk_buff *skb, struct neigh_table *tbl, in neightbl_fill_info() argument
2118 read_lock_bh(&tbl->lock); in neightbl_fill_info()
2119 ndtmsg->ndtm_family = tbl->family; in neightbl_fill_info()
2123 if (nla_put_string(skb, NDTA_NAME, tbl->id) || in neightbl_fill_info()
2124 nla_put_msecs(skb, NDTA_GC_INTERVAL, tbl->gc_interval, NDTA_PAD) || in neightbl_fill_info()
2125 nla_put_u32(skb, NDTA_THRESH1, tbl->gc_thresh1) || in neightbl_fill_info()
2126 nla_put_u32(skb, NDTA_THRESH2, tbl->gc_thresh2) || in neightbl_fill_info()
2127 nla_put_u32(skb, NDTA_THRESH3, tbl->gc_thresh3)) in neightbl_fill_info()
2131 long flush_delta = now - tbl->last_flush; in neightbl_fill_info()
2132 long rand_delta = now - tbl->last_rand; in neightbl_fill_info()
2135 .ndtc_key_len = tbl->key_len, in neightbl_fill_info()
2136 .ndtc_entry_size = tbl->entry_size, in neightbl_fill_info()
2137 .ndtc_entries = atomic_read(&tbl->entries), in neightbl_fill_info()
2140 .ndtc_proxy_qlen = tbl->proxy_queue.qlen, in neightbl_fill_info()
2144 nht = rcu_dereference_bh(tbl->nht); in neightbl_fill_info()
2162 st = per_cpu_ptr(tbl->stats, cpu); in neightbl_fill_info()
2181 BUG_ON(tbl->parms.dev); in neightbl_fill_info()
2182 if (neightbl_fill_parms(skb, &tbl->parms) < 0) in neightbl_fill_info()
2185 read_unlock_bh(&tbl->lock); in neightbl_fill_info()
2190 read_unlock_bh(&tbl->lock); in neightbl_fill_info()
2196 struct neigh_table *tbl, in neightbl_fill_param_info() argument
2210 read_lock_bh(&tbl->lock); in neightbl_fill_param_info()
2211 ndtmsg->ndtm_family = tbl->family; in neightbl_fill_param_info()
2215 if (nla_put_string(skb, NDTA_NAME, tbl->id) < 0 || in neightbl_fill_param_info()
2219 read_unlock_bh(&tbl->lock); in neightbl_fill_param_info()
2223 read_unlock_bh(&tbl->lock); in neightbl_fill_param_info()
2258 struct neigh_table *tbl; in neightbl_set() local
2277 tbl = neigh_tables[tidx]; in neightbl_set()
2278 if (!tbl) in neightbl_set()
2280 if (ndtmsg->ndtm_family && tbl->family != ndtmsg->ndtm_family) in neightbl_set()
2282 if (nla_strcmp(tb[NDTA_NAME], tbl->id) == 0) { in neightbl_set()
2295 write_lock_bh(&tbl->lock); in neightbl_set()
2311 p = lookup_neigh_parms(tbl, net, ifindex); in neightbl_set()
2397 tbl->gc_thresh1 = nla_get_u32(tb[NDTA_THRESH1]); in neightbl_set()
2400 tbl->gc_thresh2 = nla_get_u32(tb[NDTA_THRESH2]); in neightbl_set()
2403 tbl->gc_thresh3 = nla_get_u32(tb[NDTA_THRESH3]); in neightbl_set()
2406 tbl->gc_interval = nla_get_msecs(tb[NDTA_GC_INTERVAL]); in neightbl_set()
2411 write_unlock_bh(&tbl->lock); in neightbl_set()
2447 struct neigh_table *tbl; in neightbl_dump_info() local
2461 tbl = neigh_tables[tidx]; in neightbl_dump_info()
2462 if (!tbl) in neightbl_dump_info()
2465 if (tidx < tbl_skip || (family && tbl->family != family)) in neightbl_dump_info()
2468 if (neightbl_fill_info(skb, tbl, NETLINK_CB(cb->skb).portid, in neightbl_dump_info()
2474 p = list_next_entry(&tbl->parms, list); in neightbl_dump_info()
2475 list_for_each_entry_from(p, &tbl->parms_list, list) { in neightbl_dump_info()
2482 if (neightbl_fill_param_info(skb, tbl, p, in neightbl_dump_info()
2525 if (nla_put(skb, NDA_DST, neigh->tbl->key_len, neigh->primary_key)) in neigh_fill_info()
2565 struct neigh_table *tbl) in pneigh_fill_info() argument
2579 ndm->ndm_family = tbl->family; in pneigh_fill_info()
2587 if (nla_put(skb, NDA_DST, tbl->key_len, pn->key)) in pneigh_fill_info()
2643 static int neigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, in neigh_dump_table() argument
2658 nht = rcu_dereference_bh(tbl->nht); in neigh_dump_table()
2690 static int pneigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, in pneigh_dump_table() argument
2703 read_lock_bh(&tbl->lock); in pneigh_dump_table()
2708 for (n = tbl->phash_buckets[h], idx = 0; n; n = n->next) { in pneigh_dump_table()
2716 RTM_NEWNEIGH, flags, tbl) < 0) { in pneigh_dump_table()
2717 read_unlock_bh(&tbl->lock); in pneigh_dump_table()
2726 read_unlock_bh(&tbl->lock); in pneigh_dump_table()
2800 struct neigh_table *tbl; in neigh_dump_info() local
2821 tbl = neigh_tables[t]; in neigh_dump_info()
2823 if (!tbl) in neigh_dump_info()
2825 if (t < s_t || (family && tbl->family != family)) in neigh_dump_info()
2831 err = pneigh_dump_table(tbl, skb, cb, &filter); in neigh_dump_info()
2833 err = neigh_dump_table(tbl, skb, cb, &filter); in neigh_dump_info()
2843 struct neigh_table **tbl, in neigh_valid_get_req() argument
2875 *tbl = neigh_find_table(ndm->ndm_family); in neigh_valid_get_req()
2876 if (*tbl == NULL) { in neigh_valid_get_req()
2887 if (nla_len(tb[i]) != (int)(*tbl)->key_len) { in neigh_valid_get_req()
2943 u32 pid, u32 seq, struct neigh_table *tbl) in pneigh_get_reply() argument
2952 err = pneigh_fill_info(skb, neigh, pid, seq, RTM_NEWNEIGH, 0, tbl); in pneigh_get_reply()
2968 struct neigh_table *tbl = NULL; in neigh_get() local
2975 err = neigh_valid_get_req(nlh, &tbl, &dst, &dev_idx, &ndm_flags, in neigh_get()
2996 pn = pneigh_lookup(tbl, net, dst, dev, 0); in neigh_get()
3002 nlh->nlmsg_seq, tbl); in neigh_get()
3010 neigh = neigh_lookup(tbl, dst, dev); in neigh_get()
3024 void neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie) in neigh_for_each() argument
3030 nht = rcu_dereference_bh(tbl->nht); in neigh_for_each()
3032 read_lock(&tbl->lock); /* avoid resizes */ in neigh_for_each()
3041 read_unlock(&tbl->lock); in neigh_for_each()
3047 void __neigh_for_each_release(struct neigh_table *tbl, in __neigh_for_each_release() argument
3053 nht = rcu_dereference_protected(tbl->nht, in __neigh_for_each_release()
3054 lockdep_is_held(&tbl->lock)); in __neigh_for_each_release()
3061 lockdep_is_held(&tbl->lock))) != NULL) { in __neigh_for_each_release()
3069 lockdep_is_held(&tbl->lock))); in __neigh_for_each_release()
3086 struct neigh_table *tbl; in neigh_xmit() local
3089 tbl = neigh_tables[index]; in neigh_xmit()
3090 if (!tbl) in neigh_xmit()
3098 neigh = __neigh_lookup_noref(tbl, addr, dev); in neigh_xmit()
3101 neigh = __neigh_create(tbl, addr, dev, false); in neigh_xmit()
3233 struct neigh_table *tbl = state->tbl; in pneigh_get_first() local
3239 pn = tbl->phash_buckets[bucket]; in pneigh_get_first()
3256 struct neigh_table *tbl = state->tbl; in pneigh_get_next() local
3265 pn = tbl->phash_buckets[state->bucket]; in pneigh_get_next()
3306 void *neigh_seq_start(struct seq_file *seq, loff_t *pos, struct neigh_table *tbl, unsigned int neig… in neigh_seq_start() argument
3307 __acquires(tbl->lock) in neigh_seq_start()
3312 state->tbl = tbl; in neigh_seq_start()
3317 state->nht = rcu_dereference_bh(tbl->nht); in neigh_seq_start()
3318 read_lock(&tbl->lock); in neigh_seq_start()
3352 __releases(tbl->lock) in neigh_seq_stop()
3356 struct neigh_table *tbl = state->tbl; in neigh_seq_stop() local
3358 read_unlock(&tbl->lock); in neigh_seq_stop()
3367 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); in neigh_stat_seq_start() local
3377 return per_cpu_ptr(tbl->stats, cpu); in neigh_stat_seq_start()
3384 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); in neigh_stat_seq_next() local
3391 return per_cpu_ptr(tbl->stats, cpu); in neigh_stat_seq_next()
3404 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); in neigh_stat_seq_show() local
3415 atomic_read(&tbl->entries), in neigh_stat_seq_show()
3741 struct neigh_table *tbl = p->tbl; in neigh_sysctl_register() local
3743 t->neigh_vars[NEIGH_VAR_GC_INTERVAL].data = &tbl->gc_interval; in neigh_sysctl_register()
3744 t->neigh_vars[NEIGH_VAR_GC_THRESH1].data = &tbl->gc_thresh1; in neigh_sysctl_register()
3745 t->neigh_vars[NEIGH_VAR_GC_THRESH2].data = &tbl->gc_thresh2; in neigh_sysctl_register()
3746 t->neigh_vars[NEIGH_VAR_GC_THRESH3].data = &tbl->gc_thresh3; in neigh_sysctl_register()