Home
last modified time | relevance | path

Searched refs:buckets (Results 1 – 25 of 99) sorted by relevance

1234

/linux-6.3-rc2/drivers/s390/scsi/
A Dzfcp_reqlist.h26 struct list_head buckets[ZFCP_REQ_LIST_BUCKETS]; member
52 INIT_LIST_HEAD(&rl->buckets[i]); in zfcp_reqlist_alloc()
68 if (!list_empty(&rl->buckets[i])) in zfcp_reqlist_isempty()
92 list_for_each_entry(req, &rl->buckets[i], list) in _zfcp_reqlist_find()
165 list_add_tail(&req->list, &rl->buckets[i]); in zfcp_reqlist_add()
182 list_splice_init(&rl->buckets[i], list); in zfcp_reqlist_move()
209 list_for_each_entry(req, &rl->buckets[i], list) in zfcp_reqlist_apply_for_all()
/linux-6.3-rc2/Documentation/networking/
A Dnexthop-group-resilient.rst56 the buckets that held it are simply reassigned to other next hops::
70 choose a subset of buckets that are currently not used for forwarding
80 certain number of buckets, according to its weight and the number of
86 Next hops that have fewer buckets than their wants count, are called
98 buckets:
112 of balance until idle buckets do appear, possibly never. If set to a
117 item. Thus buckets:
133 created and buckets populated for the first time.
247 buckets 8 idle_timer 60 unbalanced_timer 300
259 This can be confirmed by looking at individual buckets::
[all …]
/linux-6.3-rc2/tools/perf/util/
A Dhashmap.h77 struct hashmap_entry **buckets; member
87 .buckets = NULL, \
180 for (cur = map->buckets[bkt]; cur; cur = cur->next)
192 for (cur = map->buckets[bkt]; \
203 for (cur = map->buckets \
204 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \
211 for (cur = map->buckets \
212 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \
A Dhashmap.c45 map->buckets = NULL; in hashmap__init()
71 free(map->buckets); in hashmap__clear()
72 map->buckets = NULL; in hashmap__clear()
124 free(map->buckets); in hashmap_grow()
125 map->buckets = new_buckets; in hashmap_grow()
137 if (!map->buckets) in hashmap_find_entry()
140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry()
200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
A Dftrace.h44 int buckets[]);
69 int buckets[] __maybe_unused) in perf_ftrace__latency_read_bpf()
A Dbpf_ftrace.c122 int buckets[]) in perf_ftrace__latency_read_bpf() argument
138 buckets[idx] = 0; in perf_ftrace__latency_read_bpf()
143 buckets[idx] += hist[i]; in perf_ftrace__latency_read_bpf()
/linux-6.3-rc2/tools/lib/bpf/
A Dhashmap.h77 struct hashmap_entry **buckets; member
87 .buckets = NULL, \
180 for (cur = map->buckets[bkt]; cur; cur = cur->next)
192 for (cur = map->buckets[bkt]; \
203 for (cur = map->buckets \
204 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \
211 for (cur = map->buckets \
212 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \
A Dhashmap.c45 map->buckets = NULL; in hashmap__init()
71 free(map->buckets); in hashmap__clear()
72 map->buckets = NULL; in hashmap__clear()
124 free(map->buckets); in hashmap_grow()
125 map->buckets = new_buckets; in hashmap_grow()
137 if (!map->buckets) in hashmap_find_entry()
140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry()
200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
/linux-6.3-rc2/block/
A Dblk-stat.c87 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
94 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
106 unsigned int buckets, void *data) in blk_stat_alloc_callback() argument
114 cb->stat = kmalloc_array(buckets, sizeof(struct blk_rq_stat), in blk_stat_alloc_callback()
120 cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat), in blk_stat_alloc_callback()
131 cb->buckets = buckets; in blk_stat_alloc_callback()
148 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_add_callback()
A Dblk-stat.h45 unsigned int buckets; member
89 unsigned int buckets, void *data);
A Dkyber-iosched.c136 atomic_t buckets[KYBER_OTHER][2][KYBER_LATENCY_BUCKETS]; member
216 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in flush_latency_buckets() local
217 atomic_t *cpu_buckets = cpu_latency->buckets[sched_domain][type]; in flush_latency_buckets()
221 buckets[bucket] += atomic_xchg(&cpu_buckets[bucket], 0); in flush_latency_buckets()
232 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in calculate_percentile() local
236 samples += buckets[bucket]; in calculate_percentile()
255 if (buckets[bucket] >= percentile_samples) in calculate_percentile()
257 percentile_samples -= buckets[bucket]; in calculate_percentile()
259 memset(buckets, 0, sizeof(kqd->latency_buckets[sched_domain][type])); in calculate_percentile()
637 atomic_inc(&cpu_latency->buckets[sched_domain][type][bucket]); in add_latency_sample()
/linux-6.3-rc2/tools/perf/
A Dbuiltin-ftrace.c740 buckets[i]++; in make_histogram()
760 total += buckets[i]; in display_histogram()
770 bar_len = buckets[0] * bar_total / total; in display_histogram()
784 bar_len = buckets[i] * bar_total / total; in display_histogram()
786 start, stop, unit, buckets[i], bar_len, bar, in display_histogram()
790 bar_len = buckets[NUM_BUCKET - 1] * bar_total / total; in display_histogram()
862 return perf_ftrace__latency_read_bpf(ftrace, buckets); in read_func_latency()
884 int buckets[NUM_BUCKET] = { }; in __cmd_latency() local
937 make_histogram(buckets, buf, n, line, ftrace->use_nsec); in __cmd_latency()
940 read_func_latency(ftrace, buckets); in __cmd_latency()
[all …]
/linux-6.3-rc2/tools/testing/selftests/drivers/net/netdevsim/
A Dnexthop.sh213 $IP nexthop add id 10 group 1/2 type resilient buckets 4
325 $IP nexthop add id 10 group 1/2 type resilient buckets 6
353 $IP nexthop add id 10 group 1/2 type resilient buckets 6
434 type resilient buckets 8 idle_timer 6
535 type resilient buckets 8 $timer 4
581 buckets 8 idle_timer 6 unbalanced_timer 10
613 buckets 8 idle_timer 6 unbalanced_timer 10
648 $IP nexthop add id 10 group 1/2 type resilient buckets 8
672 buckets 8 idle_timer 120 unbalanced_timer 4
707 buckets 8 idle_timer 120
[all …]
/linux-6.3-rc2/lib/
A Dhashtable_test.c189 int buckets[2]; in hashtable_test_hash_for_each_possible() local
223 buckets[y->key] = bkt; in hashtable_test_hash_for_each_possible()
230 if (buckets[0] == buckets[1]) { in hashtable_test_hash_for_each_possible()
244 int buckets[2]; in hashtable_test_hash_for_each_possible_safe() local
281 buckets[y->key] = bkt; in hashtable_test_hash_for_each_possible_safe()
288 if (buckets[0] == buckets[1]) { in hashtable_test_hash_for_each_possible_safe()
/linux-6.3-rc2/net/ceph/crush/
A Dcrush.c111 if (map->buckets) { in crush_destroy()
114 if (map->buckets[b] == NULL) in crush_destroy()
116 crush_destroy_bucket(map->buckets[b]); in crush_destroy()
118 kfree(map->buckets); in crush_destroy()
A Dmapper.c527 itemtype = map->buckets[-1-item]->type; in crush_choose_firstn()
540 in = map->buckets[-1-item]; in crush_choose_firstn()
564 map->buckets[-1-item], in crush_choose_firstn()
741 itemtype = map->buckets[-1-item]->type; in crush_choose_indep()
758 in = map->buckets[-1-item]; in crush_choose_indep()
778 map->buckets[-1-item], in crush_choose_indep()
865 if (!map->buckets[b]) in crush_init_workspace()
869 switch (map->buckets[b]->alg) { in crush_init_workspace()
948 map->buckets[-1-curstep->arg1])) { in crush_do_rule()
1032 map->buckets[bno], in crush_do_rule()
[all …]
/linux-6.3-rc2/net/netfilter/ipvs/
A Dip_vs_sh.c70 struct ip_vs_sh_bucket buckets[IP_VS_SH_TAB_SIZE]; member
108 struct ip_vs_dest *dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get()
130 dest = rcu_dereference(s->buckets[ihash].dest); in ip_vs_sh_get_fallback()
145 dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get_fallback()
172 b = &s->buckets[0]; in ip_vs_sh_reassign()
216 b = &s->buckets[0]; in ip_vs_sh_flush()
A Dip_vs_dh.c64 struct ip_vs_dh_bucket buckets[IP_VS_DH_TAB_SIZE]; member
90 return rcu_dereference(s->buckets[ip_vs_dh_hashkey(af, addr)].dest); in ip_vs_dh_get()
106 b = &s->buckets[0]; in ip_vs_dh_reassign()
140 b = &s->buckets[0]; in ip_vs_dh_flush()
/linux-6.3-rc2/fs/nfs/
A Dpnfs_nfs.c102 p = kmalloc(struct_size(p, buckets, n), gfp_flags); in pnfs_alloc_commit_array()
109 for (b = &p->buckets[0]; n != 0; b++, n--) { in pnfs_alloc_commit_array()
259 struct pnfs_commit_bucket *buckets, in pnfs_bucket_scan_array() argument
288 cnt = pnfs_bucket_scan_array(cinfo, array->buckets, in pnfs_generic_scan_commit_lists()
304 struct pnfs_commit_bucket *buckets, in pnfs_bucket_recover_commit_reqs() argument
314 for (i = 0, b = buckets; i < nbuckets; i++, b++) { in pnfs_bucket_recover_commit_reqs()
343 array->buckets, in pnfs_generic_recover_commit_reqs()
364 for (i = 0, b = buckets; i < nbuckets; i++, b++) { in pnfs_bucket_search_commit_reqs()
392 req = pnfs_bucket_search_commit_reqs(array->buckets, in pnfs_generic_search_commit_reqs()
452 struct pnfs_commit_bucket *buckets, in pnfs_bucket_alloc_ds_commits() argument
[all …]
A Dnfs42xattr.c70 struct nfs4_xattr_bucket buckets[NFS4_XATTR_HASH_SIZE]; member
111 INIT_HLIST_HEAD(&cache->buckets[i].hlist); in nfs4_xattr_hash_init()
112 spin_lock_init(&cache->buckets[i].lock); in nfs4_xattr_hash_init()
113 cache->buckets[i].cache = cache; in nfs4_xattr_hash_init()
114 cache->buckets[i].draining = false; in nfs4_xattr_hash_init()
276 if (WARN_ON(!hlist_empty(&cache->buckets[i].hlist))) in nfs4_xattr_free_cache_cb()
278 cache->buckets[i].draining = false; in nfs4_xattr_free_cache_cb()
394 bucket = &cache->buckets[i]; in nfs4_xattr_discard_cache()
506 return &cache->buckets[jhash(name, strlen(name), 0) & in nfs4_xattr_hash_bucket()
507 (ARRAY_SIZE(cache->buckets) - 1)]; in nfs4_xattr_hash_bucket()
[all …]
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlx5/core/ipoib/
A Dipoib_vlan.c45 struct hlist_head buckets[1 << MLX5I_MAX_LOG_PKEY_SUP]; member
71 static struct qpn_to_netdev *mlx5i_find_qpn_to_netdev_node(struct hlist_head *buckets, in mlx5i_find_qpn_to_netdev_node() argument
74 struct hlist_head *h = &buckets[hash_32(qpn, MLX5I_MAX_LOG_PKEY_SUP)]; in mlx5i_find_qpn_to_netdev_node()
99 hlist_add_head(&new_node->hlist, &ht->buckets[key]); in mlx5i_pkey_add_qpn()
112 node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); in mlx5i_pkey_del_qpn()
131 node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn); in mlx5i_pkey_get_netdev()
/linux-6.3-rc2/drivers/md/
A Ddm-region-hash.c73 struct list_head *buckets; member
209 rh->buckets = vmalloc(array_size(nr_buckets, sizeof(*rh->buckets))); in dm_region_hash_create()
210 if (!rh->buckets) { in dm_region_hash_create()
217 INIT_LIST_HEAD(rh->buckets + i); in dm_region_hash_create()
231 vfree(rh->buckets); in dm_region_hash_create()
247 list_for_each_entry_safe(reg, nreg, rh->buckets + h, in dm_region_hash_destroy()
258 vfree(rh->buckets); in dm_region_hash_destroy()
277 struct list_head *bucket = rh->buckets + rh_hash(rh, region); in __rh_lookup()
288 list_add(&reg->hash_list, rh->buckets + rh_hash(rh, reg->key)); in __rh_insert()
/linux-6.3-rc2/kernel/bpf/
A Dbpf_local_storage.c24 return &smap->buckets[hash_ptr(selem, smap->bucket_log)]; in select_bucket()
571 smap->buckets = bpf_map_kvcalloc(&smap->map, sizeof(*smap->buckets), in __bpf_local_storage_map_alloc()
573 if (!smap->buckets) { in __bpf_local_storage_map_alloc()
579 INIT_HLIST_HEAD(&smap->buckets[i].list); in __bpf_local_storage_map_alloc()
580 raw_spin_lock_init(&smap->buckets[i].lock); in __bpf_local_storage_map_alloc()
681 b = &smap->buckets[i]; in bpf_local_storage_map_free()
716 kvfree(smap->buckets); in bpf_local_storage_map_free()
A Dstackmap.c31 struct stack_map_bucket *buckets[]; member
233 bucket = READ_ONCE(smap->buckets[id]); in __bpf_get_stackid()
277 old_bucket = xchg(&smap->buckets[id], new_bucket); in __bpf_get_stackid()
578 bucket = xchg(&smap->buckets[id], NULL); in bpf_stackmap_copy()
586 old_bucket = xchg(&smap->buckets[id], bucket); in bpf_stackmap_copy()
605 if (id >= smap->n_buckets || !smap->buckets[id]) in stack_map_get_next_key()
611 while (id < smap->n_buckets && !smap->buckets[id]) in stack_map_get_next_key()
637 old_bucket = xchg(&smap->buckets[id], NULL); in stack_map_delete_elem()
/linux-6.3-rc2/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
A Dpno.c298 struct brcmf_gscan_bucket_config **buckets, in brcmf_pno_prep_fwconfig() argument
323 *buckets = NULL; in brcmf_pno_prep_fwconfig()
355 *buckets = fw_buckets; in brcmf_pno_prep_fwconfig()
396 struct brcmf_gscan_bucket_config *buckets; in brcmf_pno_config_sched_scans() local
403 n_buckets = brcmf_pno_prep_fwconfig(pi, &pno_cfg, &buckets, in brcmf_pno_config_sched_scans()
437 memcpy(gscan_cfg->bucket, buckets, in brcmf_pno_config_sched_scans()
438 array_size(n_buckets, sizeof(*buckets))); in brcmf_pno_config_sched_scans()
463 kfree(buckets); in brcmf_pno_config_sched_scans()

Completed in 62 milliseconds

1234