| /linux/kernel/bpf/ |
| A D | bpf_local_storage.c | 24 return &smap->buckets[hash_ptr(selem, smap->bucket_log)]; in select_bucket() 97 selem = bpf_map_kzalloc(&smap->map, smap->elem_size, in bpf_selem_alloc() 109 mem_uncharge(smap, owner, smap->elem_size); in bpf_selem_alloc() 174 if (smap) { in bpf_local_storage_free() 269 mem_uncharge(smap, owner, smap->elem_size); in bpf_selem_unlink_storage_nolock() 301 if (rcu_access_pointer(local_storage->smap) == smap) in bpf_selem_unlink_storage_nolock() 402 RCU_INIT_POINTER(SDATA(selem)->smap, smap); in bpf_selem_link_map() 475 RCU_INIT_POINTER(storage->smap, smap); in bpf_local_storage_alloc() 560 mem_uncharge(smap, owner, smap->elem_size); in bpf_local_storage_update() 633 mem_uncharge(smap, owner, smap->elem_size); in bpf_local_storage_update() [all …]
|
| A D | stackmap.c | 48 (u64)smap->map.value_size; in prealloc_elems_and_freelist() 51 smap->elems = bpf_map_area_alloc(elem_size * smap->map.max_entries, in prealloc_elems_and_freelist() 52 smap->map.numa_node); in prealloc_elems_and_freelist() 53 if (!smap->elems) in prealloc_elems_and_freelist() 60 pcpu_freelist_populate(&smap->freelist, smap->elems, elem_size, in prealloc_elems_and_freelist() 73 struct bpf_stack_map *smap; in stack_map_alloc() local 104 if (!smap) in stack_map_alloc() 118 return &smap->map; in stack_map_alloc() 123 bpf_map_area_free(smap); in stack_map_alloc() 685 if (id >= smap->n_buckets || !smap->buckets[id]) in stack_map_get_next_key() [all …]
|
| A D | bpf_inode_storage.c | 41 struct bpf_local_storage_map *smap; in inode_storage_lookup() local 53 smap = (struct bpf_local_storage_map *)map; in inode_storage_lookup() 54 return bpf_local_storage_lookup(inode_storage, smap, cacheit_lockit); in inode_storage_lookup()
|
| A D | bpf_cgrp_storage.c | 67 struct bpf_local_storage_map *smap; in cgroup_storage_lookup() local 74 smap = (struct bpf_local_storage_map *)map; in cgroup_storage_lookup() 75 return bpf_local_storage_lookup(cgroup_storage, smap, cacheit_lockit); in cgroup_storage_lookup()
|
| A D | bpf_task_storage.c | 61 struct bpf_local_storage_map *smap; in task_storage_lookup() local 68 smap = (struct bpf_local_storage_map *)map; in task_storage_lookup() 69 return bpf_local_storage_lookup(task_storage, smap, cacheit_lockit); in task_storage_lookup()
|
| /linux/tools/testing/selftests/mm/ |
| A D | map_populate.c | 40 *smap = 0x22222BAD; in parent_f() 41 ret = msync(smap, MMAP_SZ, MS_SYNC); in parent_f() 58 smap = mmap(0, MMAP_SZ, PROT_READ | PROT_WRITE, in child_f() 60 BUG_ON(smap == MAP_FAILED, "mmap()"); in child_f() 81 unsigned long *smap; in main() local 92 smap = mmap(0, MMAP_SZ, PROT_READ | PROT_WRITE, in main() 94 BUG_ON(smap == MAP_FAILED, "mmap()"); in main() 96 *smap = 0xdeadbabe; in main() 98 ret = msync(smap, MMAP_SZ, MS_SYNC); in main() 111 parent_f(sock[1], smap, child); in main() [all …]
|
| /linux/include/linux/ |
| A D | bpf_local_storage.h | 71 struct bpf_local_storage_map __rcu *smap; member 90 struct bpf_local_storage_map __rcu *smap; member 133 struct bpf_local_storage_map *smap, 138 struct bpf_local_storage_map *smap, in bpf_local_storage_lookup() argument 145 sdata = rcu_dereference_check(local_storage->cache[smap->cache_idx], in bpf_local_storage_lookup() 147 if (sdata && rcu_access_pointer(sdata->smap) == smap) in bpf_local_storage_lookup() 153 if (rcu_access_pointer(SDATA(selem)->smap) == smap) in bpf_local_storage_lookup() 159 __bpf_local_storage_insert_cache(local_storage, smap, selem); in bpf_local_storage_lookup() 179 void bpf_selem_link_map(struct bpf_local_storage_map *smap, 187 struct bpf_local_storage_map *smap, [all …]
|
| A D | bpf.h | 148 int (*map_local_storage_charge)(struct bpf_local_storage_map *smap, 150 void (*map_local_storage_uncharge)(struct bpf_local_storage_map *smap,
|
| /linux/net/core/ |
| A D | bpf_sk_storage.c | 24 struct bpf_local_storage_map *smap; in bpf_sk_storage_lookup() local 171 struct bpf_local_storage_map *smap; in bpf_sk_storage_clone() local 174 smap = rcu_dereference(SDATA(selem)->smap); in bpf_sk_storage_clone() 201 atomic_sub(smap->elem_size, in bpf_sk_storage_clone() 543 struct bpf_local_storage_map *smap; in diag_get() local 552 smap = rcu_dereference(sdata->smap); in diag_get() 557 smap->map.value_size, in diag_get() 584 struct bpf_local_storage_map *smap; in bpf_sk_storage_diag_put_all() local 604 smap = rcu_dereference(SDATA(selem)->smap); in bpf_sk_storage_diag_put_all() 729 b = &smap->buckets[bucket_id++]; in bpf_sk_storage_map_seq_find_next() [all …]
|
| /linux/fs/btrfs/ |
| A D | bio.c | 491 bio->bi_private = smap->dev; in btrfs_submit_bio() 493 btrfs_submit_dev_bio(smap->dev, bio); in btrfs_submit_bio() 526 struct btrfs_io_stripe smap; member 636 async->smap = *smap; in btrfs_wq_submit_bio() 667 struct btrfs_io_stripe smap; in btrfs_submit_chunk() local 672 smap.rst_search_commit_root = true; in btrfs_submit_chunk() 674 smap.rst_search_commit_root = false; in btrfs_submit_chunk() 678 &bioc, &smap, &mirror_num); in btrfs_submit_chunk() 791 struct btrfs_io_stripe smap = { 0 }; in btrfs_repair_io_failure() local 812 if (!smap.dev->bdev || in btrfs_repair_io_failure() [all …]
|
| A D | volumes.h | 683 struct btrfs_io_stripe *smap, int *mirror_num_ret); 685 struct btrfs_io_stripe *smap, u64 logical,
|
| A D | volumes.c | 6263 const struct btrfs_io_stripe *smap, in is_single_device_io() argument 6268 if (!smap) in is_single_device_io() 6451 struct btrfs_io_stripe *smap, int *mirror_num_ret) in btrfs_map_block() argument 8135 struct btrfs_io_stripe *smap, in map_raid56_repair_block() argument 8150 smap->dev = bioc->stripes[i].dev; in map_raid56_repair_block() 8151 smap->physical = bioc->stripes[i].physical + in map_raid56_repair_block() 8170 struct btrfs_io_stripe *smap, u64 logical, in btrfs_map_repair_block() argument 8181 &bioc, smap, &mirror_ret); in btrfs_map_repair_block() 8194 map_raid56_repair_block(bioc, smap, logical); in btrfs_map_repair_block() 8199 smap->dev = bioc->stripes[mirror_num - 1].dev; in btrfs_map_repair_block() [all …]
|
| /linux/fs/nilfs2/ |
| A D | page.c | 245 struct address_space *smap) in nilfs_copy_dirty_pages() argument 254 if (!filemap_get_folios_tag(smap, &index, (pgoff_t)-1, in nilfs_copy_dirty_pages() 300 struct address_space *smap) in nilfs_copy_back_pages() argument 308 n = filemap_get_folios(smap, &start, ~0UL, &fbatch); in nilfs_copy_back_pages() 329 xa_lock_irq(&smap->i_pages); in nilfs_copy_back_pages() 330 f = __xa_erase(&smap->i_pages, index); in nilfs_copy_back_pages() 332 smap->nrpages--; in nilfs_copy_back_pages() 333 xa_unlock_irq(&smap->i_pages); in nilfs_copy_back_pages()
|
| /linux/fs/xfs/ |
| A D | xfs_reflink.c | 1130 struct xfs_bmbt_irec smap; in xfs_reflink_remap_extent() local 1182 &smap, &nimaps, 0); in xfs_reflink_remap_extent() 1186 smap_real = xfs_bmap_is_real_extent(&smap); in xfs_reflink_remap_extent() 1195 trace_xfs_reflink_remap_extent_dest(ip, &smap); in xfs_reflink_remap_extent() 1202 if (dmap->br_startblock == smap.br_startblock) { in xfs_reflink_remap_extent() 1203 if (dmap->br_state != smap.br_state) { in xfs_reflink_remap_extent() 1212 smap.br_state == XFS_EXT_UNWRITTEN) in xfs_reflink_remap_extent() 1268 xfs_refcount_decrease_extent(tp, &smap); in xfs_reflink_remap_extent() 1269 qdelta -= smap.br_blockcount; in xfs_reflink_remap_extent() 1279 error = xfs_bunmapi(NULL, ip, smap.br_startoff, in xfs_reflink_remap_extent() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | local_storage.c | 99 if (!local_storage || local_storage->smap) in BPF_PROG() 176 if (!sk->sk_bpf_storage || sk->sk_bpf_storage->smap) in BPF_PROG()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | hwmon.c | 209 DECLARE_BITMAP(smap, BITS_PER_TYPE(sensor_map)); in mlx5_hwmon_init_sensors_indexes() 214 bitmap_from_u64(smap, sensor_map); in mlx5_hwmon_init_sensors_indexes() 216 for_each_set_bit(bit_pos, smap, BITS_PER_TYPE(sensor_map)) { in mlx5_hwmon_init_sensors_indexes()
|
| /linux/lib/zstd/common/ |
| A D | cpu.h | 175 B(smap, 20)
|
| /linux/fs/ |
| A D | dax.c | 2005 const struct iomap *smap = &it_src->iomap; in dax_range_compare_iter() local 2011 len = min(len, min(smap->length, dmap->length)); in dax_range_compare_iter() 2013 if (smap->type == IOMAP_HOLE && dmap->type == IOMAP_HOLE) { in dax_range_compare_iter() 2018 if (smap->type == IOMAP_HOLE || dmap->type == IOMAP_HOLE) { in dax_range_compare_iter() 2024 ret = dax_iomap_direct_access(smap, pos1, ALIGN(pos1 + len, PAGE_SIZE), in dax_range_compare_iter()
|
| /linux/fs/ocfs2/dlm/ |
| A D | dlmdomain.c | 49 static inline void byte_copymap(u8 dmap[], unsigned long smap[], in byte_copymap() argument 59 if (test_bit(nn, smap)) in byte_copymap()
|
| /linux/arch/x86/kvm/svm/ |
| A D | svm.c | 4800 bool smep, smap, is_user; in svm_check_emulate_instruction() local 4904 smap = kvm_is_cr4_bit_set(vcpu, X86_CR4_SMAP); in svm_check_emulate_instruction() 4906 if (smap && (!smep || is_user)) { in svm_check_emulate_instruction()
|
| /linux/Documentation/virt/kvm/x86/ |
| A D | mmu.rst | 192 Contains the value of cr4.smap && !cr0.wp for which the page is valid
|
| /linux/arch/x86/kvm/mmu/ |
| A D | mmu.c | 211 BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, smap, X86_CR4_SMAP); 231 BUILD_MMU_ROLE_ACCESSOR(ext, cr4, smap);
|
| /linux/tools/arch/x86/kcpuid/ |
| A D | cpuid.csv | 209 7, 0, ebx, 20, smap , Supervisor mode access prevention
|