| /tools/testing/selftests/bpf/prog_tests/ |
| A D | btf.c | 69 __u32 value_size; member 137 .value_size = 180, 192 .value_size = 68, 217 .value_size = 16, 258 .value_size = 48, 303 .value_size = 48, 325 .value_size = 4, 347 .value_size = 4, 369 .value_size = 4, 394 .value_size = 4, [all …]
|
| /tools/testing/selftests/bpf/benchs/ |
| A D | bench_bloom_filter_map.c | 41 __u8 value_size; member 45 .value_size = 8, 93 args.value_size = ret; in parse_arg() 136 val_size = args.value_size; in map_prepare_thread() 243 if (args.value_size < 8) { in check_args() 244 __u64 nr_unique_entries = 1ULL << (args.value_size * 8); in check_args() 279 bpf_map__set_value_size(skel->maps.array_map, args.value_size); in setup_skeleton() 281 bpf_map__set_value_size(skel->maps.bloom_map, args.value_size); in setup_skeleton() 283 bpf_map__set_value_size(skel->maps.hashmap, args.value_size); in setup_skeleton() 286 bpf_map__set_key_size(skel->maps.hashmap, args.value_size); in setup_skeleton() [all …]
|
| A D | bench_htab_mem.c | 40 u32 value_size; member 44 .value_size = 8, 68 args.value_size = strtoul(arg, NULL, 10); in htab_mem_parse_arg() 69 if (args.value_size > 4096) { in htab_mem_parse_arg() 70 fprintf(stderr, "too big value size %u\n", args.value_size); in htab_mem_parse_arg() 181 bpf_map__set_value_size(map, args.value_size); in htab_mem_setup()
|
| /tools/testing/selftests/bpf/map_tests/ |
| A D | array_map_batch_ops.c | 76 int err, step, value_size; in __test_map_lookup_and_update_batch() local 88 value_size = sizeof(__s64); in __test_map_lookup_and_update_batch() 90 value_size *= nr_cpus; in __test_map_lookup_and_update_batch() 93 values = calloc(max_entries, value_size); in __test_map_lookup_and_update_batch() 104 memset(values, 0, max_entries * value_size); in __test_map_lookup_and_update_batch() 115 values + total * value_size, in __test_map_lookup_and_update_batch()
|
| A D | htab_map_batch_ops.c | 84 int err, step, value_size; in __test_map_lookup_and_delete_batch() local 97 value_size = is_pcpu ? sizeof(value) : sizeof(int); in __test_map_lookup_and_delete_batch() 125 memset(values, 0, max_entries * value_size); in __test_map_lookup_and_delete_batch() 144 memset(values, 0, max_entries * value_size); in __test_map_lookup_and_delete_batch() 156 total * value_size, in __test_map_lookup_and_delete_batch() 210 memset(values, 0, max_entries * value_size); in __test_map_lookup_and_delete_batch() 219 total * value_size, in __test_map_lookup_and_delete_batch()
|
| A D | map_in_map_batch_ops.c | 129 __u32 value_size = sizeof(__u32); in fetch_and_validate() local 132 fetched_keys = calloc(max_entries, value_size); in fetch_and_validate() 133 fetched_values = calloc(max_entries, value_size); in fetch_and_validate() 194 __u32 value_size = sizeof(__u32); in _map_in_map_batch_ops() local 199 outer_map_keys = calloc(OUTER_MAP_ENTRIES, value_size); in _map_in_map_batch_ops() 200 inner_map_fds = calloc(OUTER_MAP_ENTRIES, value_size); in _map_in_map_batch_ops()
|
| /tools/perf/util/bpf_skel/ |
| A D | bperf_leader.bpf.c | 10 __uint(value_size, sizeof(int)); 17 __uint(value_size, sizeof(struct bpf_perf_event_value)); 24 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
| A D | bpf_prog_profiler.bpf.c | 11 __uint(value_size, sizeof(int)); 18 __uint(value_size, sizeof(struct bpf_perf_event_value)); 26 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
| A D | func_latency.bpf.c | 13 __uint(value_size, sizeof(__u64)); 20 __uint(value_size, sizeof(__u8)); 27 __uint(value_size, sizeof(__u8)); 34 __uint(value_size, sizeof(__u64));
|
| A D | lock_contention.bpf.c | 32 __uint(value_size, sizeof(__u64)); 40 __uint(value_size, sizeof(__u64)); 80 __uint(value_size, sizeof(struct tstamp_data)); 102 __uint(value_size, sizeof(__u32)); 109 __uint(value_size, sizeof(__u8)); 116 __uint(value_size, sizeof(__u8)); 123 __uint(value_size, sizeof(__u8)); 130 __uint(value_size, sizeof(__u8)); 137 __uint(value_size, sizeof(__u8)); 144 __uint(value_size, sizeof(__u8)); [all …]
|
| A D | off_cpu.bpf.c | 47 __uint(value_size, MAX_STACKS * sizeof(__u64)); 58 __uint(value_size, sizeof(int)); 65 __uint(value_size, sizeof(struct offcpu_data)); 79 __uint(value_size, sizeof(__u64)); 86 __uint(value_size, sizeof(__u8)); 93 __uint(value_size, sizeof(__u8)); 100 __uint(value_size, sizeof(__u8));
|
| A D | bperf_follower.bpf.c | 13 __uint(value_size, sizeof(struct bpf_perf_event_value)); 20 __uint(value_size, sizeof(struct bpf_perf_event_value)); 27 __uint(value_size, sizeof(struct bperf_filter_value));
|
| A D | bperf_cgroup.bpf.c | 19 __uint(value_size, sizeof(int)); 27 __uint(value_size, sizeof(__u32)); 35 __uint(value_size, sizeof(struct bpf_perf_event_value)); 43 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
| A D | kwork_trace.bpf.c | 38 __uint(value_size, MAX_KWORKNAME); 45 __uint(value_size, sizeof(__u64)); 52 __uint(value_size, sizeof(struct report_data)); 59 __uint(value_size, sizeof(__u8)); 66 __uint(value_size, MAX_KWORKNAME);
|
| /tools/bpf/bpftool/ |
| A D | map.c | 68 return malloc(info->value_size); in alloc_value() 159 step = round_up(info->value_size, 8); in print_entry_json() 173 info->value_size); in print_entry_json() 264 if (info->value_size) { in print_entry_plain() 286 if (info->value_size) { in print_entry_plain() 291 info->value_size, " "); in print_entry_plain() 370 if (value_size) in parse_elem() 382 if (value_size != 4) { in parse_elem() 400 if (value_size != 4) { in parse_elem() 861 info->value_size != 8) { in map_dump() [all …]
|
| /tools/lib/bpf/ |
| A D | libbpf_probes.c | 285 int key_size, value_size, max_entries; in probe_map_create() local 290 value_size = sizeof(__u32); in probe_map_create() 295 value_size = sizeof(__u64); in probe_map_create() 299 value_size = sizeof(__u64); in probe_map_create() 305 value_size = sizeof(__u64); in probe_map_create() 318 value_size = 8; in probe_map_create() 328 value_size = 0; in probe_map_create() 343 value_size = 0; in probe_map_create() 388 fd = bpf_map_create(map_type, NULL, key_size, value_size, max_entries, &opts); in probe_map_create()
|
| A D | bpf_gen_internal.h | 61 __u32 key_size, __u32 value_size, __u32 max_entries, 67 void bpf_gen__map_update_elem(struct bpf_gen *gen, int map_idx, void *value, __u32 value_size);
|
| /tools/bpf/bpftool/skeleton/ |
| A D | profiler.bpf.c | 17 __uint(value_size, sizeof(int)); 24 __uint(value_size, sizeof(struct bpf_perf_event_value___local)); 31 __uint(value_size, sizeof(struct bpf_perf_event_value___local)); 38 __uint(value_size, sizeof(u64));
|
| /tools/testing/selftests/bpf/progs/ |
| A D | test_queue_stack_map.h | 16 __uint(value_size, sizeof(__u32)); 24 __uint(value_size, sizeof(__u32));
|
| A D | tailcall_bpf2bpf4.c | 9 __uint(value_size, sizeof(__u32)); 16 __uint(value_size, sizeof(__u32));
|
| A D | kmem_cache_iter.c | 20 __uint(value_size, SLAB_NAME_MAX); 27 __uint(value_size, sizeof(struct kmem_cache_result));
|
| A D | bloom_filter_bench.c | 56 __u8 value_size; variable 126 for (i = 0; i < 1024; i++, index += value_size) { in bloom_hashmap_lookup()
|
| A D | xdp_redirect_multi_kern.c | 18 __uint(value_size, sizeof(int)); 25 __uint(value_size, sizeof(struct bpf_devmap_val));
|
| A D | tailcall_freplace.c | 10 __uint(value_size, sizeof(__u32));
|
| A D | test_xdp_with_devmap_frags_helpers.c | 8 __uint(value_size, sizeof(struct bpf_devmap_val));
|