| /tools/testing/selftests/bpf/ |
| A D | bpf_arena_htab.h | 14 htab_bucket_t *buckets; member 21 htab_bucket_t *b = htab->buckets; in __select_bucket() 95 void __arena *buckets = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in htab_init() local 97 cast_user(buckets); in htab_init() 98 htab->buckets = buckets; in htab_init()
|
| /tools/perf/util/ |
| A D | hashmap.h | 77 struct hashmap_entry **buckets; member 170 for (cur = (map)->buckets[bkt]; cur; cur = cur->next) 182 for (cur = (map)->buckets[bkt]; \ 193 for (cur = (map)->buckets \ 194 ? (map)->buckets[hash_bits((map)->hash_fn((_key), (map)->ctx), (map)->cap_bits)] \ 201 for (cur = (map)->buckets \ 202 ? (map)->buckets[hash_bits((map)->hash_fn((_key), (map)->ctx), (map)->cap_bits)] \
|
| A D | hashmap.c | 45 map->buckets = NULL; in hashmap__init() 71 free(map->buckets); in hashmap__clear() 72 map->buckets = NULL; in hashmap__clear() 124 free(map->buckets); in hashmap_grow() 125 map->buckets = new_buckets; in hashmap_grow() 137 if (!map->buckets) in hashmap_find_entry() 140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry() 200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
|
| A D | ftrace.h | 57 int buckets[], struct stats *stats); 82 int buckets[] __maybe_unused, in perf_ftrace__latency_read_bpf()
|
| A D | bpf_ftrace.c | 167 int buckets[], struct stats *stats) in perf_ftrace__latency_read_bpf() argument 183 buckets[idx] = 0; in perf_ftrace__latency_read_bpf() 188 buckets[idx] += hist[i]; in perf_ftrace__latency_read_bpf()
|
| /tools/lib/bpf/ |
| A D | hashmap.h | 77 struct hashmap_entry **buckets; member 170 for (cur = (map)->buckets[bkt]; cur; cur = cur->next) 182 for (cur = (map)->buckets[bkt]; \ 193 for (cur = (map)->buckets \ 194 ? (map)->buckets[hash_bits((map)->hash_fn((_key), (map)->ctx), (map)->cap_bits)] \ 201 for (cur = (map)->buckets \ 202 ? (map)->buckets[hash_bits((map)->hash_fn((_key), (map)->ctx), (map)->cap_bits)] \
|
| A D | hashmap.c | 45 map->buckets = NULL; in hashmap__init() 71 free(map->buckets); in hashmap__clear() 72 map->buckets = NULL; in hashmap__clear() 124 free(map->buckets); in hashmap_grow() 125 map->buckets = new_buckets; in hashmap_grow() 137 if (!map->buckets) in hashmap_find_entry() 140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry() 200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
|
| /tools/testing/selftests/bpf/prog_tests/ |
| A D | arena_htab.c | 20 printf("htab %p buckets %p n_buckets %d\n", htab, htab->buckets, htab->n_buckets); in test_arena_htab_common() 21 ASSERT_OK_PTR(htab->buckets, "htab->buckets shouldn't be NULL"); in test_arena_htab_common() 22 for (i = 0; htab->buckets && i < 16; i += 4) { in test_arena_htab_common()
|
| /tools/testing/selftests/drivers/net/netdevsim/ |
| A D | nexthop.sh | 213 $IP nexthop add id 10 group 1/2 type resilient buckets 4 325 $IP nexthop add id 10 group 1/2 type resilient buckets 6 353 $IP nexthop add id 10 group 1/2 type resilient buckets 6 434 type resilient buckets 8 idle_timer 6 535 type resilient buckets 8 $timer 4 581 buckets 8 idle_timer 6 unbalanced_timer 10 613 buckets 8 idle_timer 6 unbalanced_timer 10 648 $IP nexthop add id 10 group 1/2 type resilient buckets 8 672 buckets 8 idle_timer 120 unbalanced_timer 4 707 buckets 8 idle_timer 120 [all …]
|
| /tools/perf/ |
| A D | builtin-ftrace.c | 929 buckets[i]++; in make_histogram() 953 total += buckets[i]; in display_histogram() 963 bar_len = buckets[0] * bar_total / total; in display_histogram() 965 if (!ftrace->hide_empty || buckets[0]) in display_histogram() 974 if (ftrace->hide_empty && !buckets[i]) in display_histogram() 1132 int *buckets; in __cmd_latency() local 1146 buckets = calloc(ftrace->bucket_num, sizeof(*buckets)); in __cmd_latency() 1147 if (buckets == NULL) { in __cmd_latency() 1182 read_func_latency(ftrace, buckets); in __cmd_latency() 1184 display_histogram(ftrace, buckets); in __cmd_latency() [all …]
|
| /tools/testing/selftests/net/forwarding/ |
| A D | router_mpath_nh_res.sh | 207 ip nexthop add id 103 group 101/102 type resilient buckets 512 \ 213 ip nexthop add id 106 group 104/105 type resilient buckets 512 \ 219 ip nexthop add id 203 group 201/202 type resilient buckets 512 \ 225 ip nexthop add id 206 group 204/205 type resilient buckets 512 \
|
| A D | gre_multipath_nh_res.sh | 103 ip nexthop add id 103 group 101/102 type resilient buckets 512 \ 150 ip nexthop add id 203 group 201/202 type resilient buckets 512 \
|
| /tools/testing/selftests/drivers/net/mlxsw/ |
| A D | rtnetlink.sh | 624 ip nexthop add id 10 group 3 type resilient buckets 7 627 ip nexthop add id 10 group 3 type resilient buckets 129 630 ip nexthop add id 10 group 1/2 type resilient buckets 32 633 ip nexthop add id 10 group 3 type resilient buckets 32 767 ip nexthop add id 10 group 1/2 type resilient buckets 32 idle_timer 0
|
| A D | devlink_trap_l3_exceptions.sh | 464 ip nexthop add id 10 group 1 type resilient buckets 32
|
| /tools/memory-model/Documentation/ |
| A D | simple.txt | 99 in each hash bucket allows operations on different buckets to proceed 100 concurrently. This works because the buckets do not overlap with each 104 As the number of buckets increases, data locking scales naturally. 106 increasing the number of buckets as the number of CPUs increase results
|
| /tools/testing/selftests/net/ |
| A D | fib_nexthops.sh | 362 local buckets=$2 373 run_cmd "$IP nexthop add id 1000 group 100 type resilient buckets $buckets" 375 log_test $? 0 "Dump large (x$buckets) nexthop buckets"
|
| /tools/perf/Documentation/ |
| A D | perf-ftrace.txt | 171 22 buckets.
|