| /kernel/bpf/ |
| A D | hashtab.c | 662 u32 hash, key_size; in __htab_map_lookup_elem() local 667 key_size = map->key_size; in __htab_map_lookup_elem() 828 key_size = map->key_size; in htab_map_get_next_key() 1096 key_size = map->key_size; in htab_map_update_elem() 1204 key_size = map->key_size; in htab_lru_map_update_elem() 1273 key_size = map->key_size; in htab_map_update_elem_in_place() 1336 key_size = map->key_size; in __htab_lru_percpu_map_update_elem() 1414 key_size = map->key_size; in htab_map_delete_elem() 1450 key_size = map->key_size; in htab_lru_map_delete_elem() 1594 key_size = map->key_size; in __htab_map_lookup_and_delete_elem() [all …]
|
| A D | map_in_map.c | 37 inner_map_meta->key_size = inner_map->key_size; in bpf_map_meta_alloc() 88 meta0->key_size == meta1->key_size && in bpf_map_meta_equal()
|
| A D | map_iter.c | 104 u32 key_acc_size, value_acc_size, key_size, value_size; in bpf_iter_attach_map() local 127 key_size = map->key_size; in bpf_iter_attach_map() 133 if (key_acc_size > key_size || value_acc_size > value_size) { in bpf_iter_attach_map()
|
| A D | syscall.c | 434 map->key_size = attr->key_size; in bpf_map_init_from_attr() 1003 map->key_size, in bpf_map_show_fdinfo() 1220 u32 key_size, value_size; in map_check_btf() local 1226 if (!key_type || key_size != map->key_size) in map_check_btf() 1637 if (key_size) in __bpf_copy_key() 1648 if (key_size) in ___bpf_copy_key() 1920 map->key_size)) in generic_map_delete_batch() 1985 map->key_size) || in generic_map_update_batch() 2049 value = key + map->key_size; in generic_map_lookup_batch() 2069 map->key_size)) { in generic_map_lookup_batch() [all …]
|
| A D | local_storage.c | 36 return map->key_size == sizeof(struct bpf_cgroup_storage_key); in attach_type_isolated() 298 if (attr->key_size != sizeof(struct bpf_cgroup_storage_key) && in cgroup_storage_map_alloc() 299 attr->key_size != sizeof(__u64)) in cgroup_storage_map_alloc()
|
| A D | lpm_trie.c | 582 attr->key_size < LPM_KEY_SIZE_MIN || in trie_alloc() 583 attr->key_size > LPM_KEY_SIZE_MAX || in trie_alloc() 594 trie->data_size = attr->key_size - in trie_alloc()
|
| A D | bloom_filter.c | 101 if (attr->key_size != 0 || attr->value_size == 0 || in bloom_map_alloc()
|
| A D | queue_stack_maps.c | 50 if (attr->max_entries == 0 || attr->key_size != 0 || in queue_stack_map_alloc_check()
|
| A D | arena.c | 106 if (attr->key_size || attr->value_size || attr->max_entries == 0 || in arena_map_alloc()
|
| A D | bpf_local_storage.c | 709 attr->key_size != sizeof(int) || !attr->value_size || in bpf_local_storage_map_alloc_check()
|
| A D | cpumap.c | 89 if (attr->max_entries == 0 || attr->key_size != 4 || in cpu_map_alloc()
|
| A D | ringbuf.c | 197 if (attr->key_size || attr->value_size || in ringbuf_map_alloc()
|
| A D | stackmap.c | 81 if (attr->max_entries == 0 || attr->key_size != 4 || in stack_map_alloc()
|
| A D | log.c | 722 reg->map_ptr->key_size, in print_reg_state()
|
| A D | inode.c | 203 iter->key = kzalloc(map->key_size, GFP_KERNEL | __GFP_NOWARN); in map_iter_alloc()
|
| A D | devmap.c | 117 if (attr->max_entries == 0 || attr->key_size != 4 || in dev_map_alloc_check()
|
| A D | bpf_struct_ops.c | 1007 if (attr->key_size != sizeof(unsigned int) || attr->max_entries != 1 || in bpf_struct_ops_map_alloc_check()
|
| A D | arraymap.c | 58 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check()
|
| A D | helpers.c | 1176 key = value - round_up(map->key_size, 8); in bpf_timer_cb() 1210 key = value - round_up(map->key_size, 8); in bpf_wq_work()
|
| A D | verifier.c | 7562 reg->map_ptr->key_size, false); in check_mem_access() 8192 reg->map_ptr->key_size, false); in check_helper_mem_access() 9585 u32 key_size, in get_constant_map_key() argument 9613 if (zero_size >= key_size) { in get_constant_map_key() 9623 if (spill_size != key_size) in get_constant_map_key() 9655 u32 key_size; in check_func_arg() local 9789 key_size = meta->map_ptr->key_size; in check_func_arg() 9790 err = check_helper_mem_access(env, regno, key_size, BPF_READ, false, NULL); in check_func_arg() 9794 err = get_constant_map_key(env, reg, key_size, &meta->const_map_key); in check_func_arg()
|
| /kernel/trace/ |
| A D | tracing_map.c | 414 elt->key = kzalloc(map->key_size, GFP_KERNEL); in tracing_map_elt_alloc() 509 if (memcmp(key, test_key, key_size)) in keys_match() 523 key_hash = jhash(key, map->key_size, 0); in __tracing_map_insert() 536 keys_match(key, val->key, map->key_size)) { in __tracing_map_insert() 576 memcpy(elt->key, key, map->key_size); in __tracing_map_insert() 769 unsigned int key_size, in tracing_map_create() argument 798 map->key_size = key_size; in tracing_map_create() 852 return memcmp(a->key, b->key, a->elt->map->key_size); in cmp_entries_dup() 963 int n_entries, unsigned int key_size) in detect_dups() argument 977 if (!memcmp(sort_entries[i]->key, key, key_size)) { in detect_dups() [all …]
|
| A D | tracing_map.h | 184 unsigned int key_size; member 242 unsigned int key_size,
|
| A D | trace_events_hist.c | 558 unsigned int key_size; member 4505 unsigned int key_size; in create_key_field() local 4515 key_size = sizeof(unsigned long) * HIST_STACKTRACE_DEPTH; in create_key_field() 4532 key_size = hist_field->size; in create_key_field() 4537 key_size = ALIGN(key_size, sizeof(u64)); in create_key_field() 4538 hist_data->fields[key_idx]->size = key_size; in create_key_field() 4541 hist_data->key_size += key_size; in create_key_field() 4543 if (hist_data->key_size > HIST_KEY_SIZE_MAX) { in create_key_field() 4554 ret = key_size; in create_key_field() 5144 hist_data->map = tracing_map_create(map_bits, hist_data->key_size, in create_hist_data() [all …]
|
| /kernel/ |
| A D | crash_dump_dm_crypt.c | 17 unsigned int key_size; member 71 dm_key->data, dm_key->key_size, in add_key_to_keyring() 135 kexec_dprintk("Get key (size=%u)\n", key->key_size); in restore_dm_crypt_keys_to_thread_keyring() 165 dm_key->key_size = ukp->datalen; in read_key_from_user_keying() 166 kexec_dprintk("Get dm crypt key (size=%u) %s: %8ph\n", dm_key->key_size, in read_key_from_user_keying()
|