| /tools/testing/selftests/damon/ |
| A D | access_memory_even.c | 16 char **regions; in main() local 29 regions = malloc(sizeof(*regions) * nr_regions); in main() 31 regions[i] = malloc(sz_region); in main() 36 memset(regions[i], i, sz_region); in main()
|
| A D | access_memory.c | 13 char **regions; in main() local 30 regions = malloc(sizeof(*regions) * nr_regions); in main() 32 regions[i] = malloc(sz_region); in main() 38 memset(regions[i], i, sz_region); in main()
|
| A D | drgn_dump_damon_status.py | 66 def regions_to_list(regions): argument 69 'struct damon_region', regions.address_of_(), 'list')]
|
| /tools/testing/memblock/tests/ |
| A D | alloc_exact_nid_api.c | 30 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_simple_check() 31 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_simple_check() 82 struct memblock_region *new_rgn = &memblock.reserved.regions[1]; in alloc_exact_nid_top_down_numa_part_reserved_check() 143 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_split_range_low_check() 200 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_no_overlap_split_check() 202 struct memblock_region *node2 = &memblock.memory.regions[6]; in alloc_exact_nid_top_down_numa_no_overlap_split_check() 258 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_no_overlap_low_check() 260 struct memblock_region *min_node = &memblock.memory.regions[2]; in alloc_exact_nid_top_down_numa_no_overlap_low_check() 261 struct memblock_region *max_node = &memblock.memory.regions[5]; in alloc_exact_nid_top_down_numa_no_overlap_low_check() 299 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_bottom_up_numa_simple_check() [all …]
|
| A D | alloc_nid_api.c | 66 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_simple_check() 118 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_end_misaligned_check() 169 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_exact_address_generic_check() 221 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_narrow_range_check() 311 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_min_reserved_generic_check() 363 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_max_reserved_generic_check() 416 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_nid_top_down_reserved_with_space_check() 417 struct memblock_region *rgn2 = &memblock.reserved.regions[0]; in alloc_nid_top_down_reserved_with_space_check() 481 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_reserved_full_merge_generic_check() 652 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_cap_max_check() [all …]
|
| A D | basic_api.c | 17 ASSERT_NE(memblock.memory.regions, NULL); in memblock_initialization_check() 45 rgn = &memblock.memory.regions[0]; in memblock_add_simple_check() 78 rgn = &memblock.memory.regions[0]; in memblock_add_node_simple_check() 121 rgn1 = &memblock.memory.regions[0]; in memblock_add_disjoint_check() 122 rgn2 = &memblock.memory.regions[1]; in memblock_add_disjoint_check() 175 rgn = &memblock.memory.regions[0]; in memblock_add_overlap_top_check() 227 rgn = &memblock.memory.regions[0]; in memblock_add_overlap_bottom_check() 276 rgn = &memblock.memory.regions[0]; in memblock_add_within_check() 347 rgn = &memblock.memory.regions[0]; in memblock_add_between_check() 401 rgn = &memblock.memory.regions[0]; in memblock_add_near_max_check() [all …]
|
| A D | alloc_helpers_api.c | 20 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_simple_generic_check() 63 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_misaligned_generic_check() 110 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_high_addr_check() 153 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_no_space_above_check() 190 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_min_addr_cap_check() 236 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_high_addr_check() 278 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_no_space_above_check() 314 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_min_addr_cap_check()
|
| A D | alloc_api.c | 26 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_simple_check() 73 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_top_down_disjoint_check() 74 struct memblock_region *rgn2 = &memblock.reserved.regions[0]; in alloc_top_down_disjoint_check() 125 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_before_check() 168 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_after_check() 217 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_second_fit_check() 266 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_in_between_generic_check() 416 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_limited_space_generic_check() 450 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_no_memory_generic_check() 484 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_too_large_generic_check() [all …]
|
| A D | common.c | 41 memset(memblock.memory.regions, 0, in reset_memblock_regions() 47 memset(memblock.reserved.regions, 0, in reset_memblock_regions()
|
| /tools/virtio/ |
| A D | virtio_test.c | 141 dev->mem = malloc(offsetof(struct vhost_memory, regions) + in vdev_info_init() 142 sizeof dev->mem->regions[0]); in vdev_info_init() 144 memset(dev->mem, 0, offsetof(struct vhost_memory, regions) + in vdev_info_init() 145 sizeof dev->mem->regions[0]); in vdev_info_init() 147 dev->mem->regions[0].guest_phys_addr = (long)dev->buf; in vdev_info_init() 148 dev->mem->regions[0].userspace_addr = (long)dev->buf; in vdev_info_init() 149 dev->mem->regions[0].memory_size = dev->buf_size; in vdev_info_init()
|
| A D | vhost_net_test.c | 238 dev->mem = malloc(offsetof(struct vhost_memory, regions) + in vdev_info_init() 239 sizeof(dev->mem->regions[0])); in vdev_info_init() 241 memset(dev->mem, 0, offsetof(struct vhost_memory, regions) + in vdev_info_init() 242 sizeof(dev->mem->regions[0])); in vdev_info_init() 244 dev->mem->regions[0].guest_phys_addr = (long)dev->buf; in vdev_info_init() 245 dev->mem->regions[0].userspace_addr = (long)dev->buf; in vdev_info_init() 246 dev->mem->regions[0].memory_size = dev->buf_size; in vdev_info_init()
|
| /tools/testing/selftests/mm/ |
| A D | .gitignore | 59 guard-regions
|
| A D | Makefile | 102 TEST_GEN_FILES += guard-regions
|
| A D | run_vmtests.sh | 425 CATEGORY="madv_guard" run_test ./guard-regions
|
| /tools/testing/nvdimm/test/ |
| A D | ndtest.h | 72 struct ndtest_region *regions; member
|
| A D | ndtest.c | 185 .regions = bus0_regions, 193 .regions = bus1_regions, 474 ret = ndtest_create_region(p, &p->config->regions[i]); in ndtest_init_regions()
|
| /tools/testing/selftests/kvm/lib/ |
| A D | kvm_util.c | 299 vm->regions.gpa_tree = RB_ROOT; in ____vm_create() 300 vm->regions.hva_tree = RB_ROOT; in ____vm_create() 301 hash_init(vm->regions.slot_hash); in ____vm_create() 596 hash_for_each(vmp->regions.slot_hash, ctr, region, slot_node) { in kvm_vm_restart() 716 for (node = vm->regions.gpa_tree.rb_node; node; ) { in userspace_mem_region_find() 815 rb_erase(®ion->gpa_node, &vm->regions.gpa_tree); in __vm_mem_region_delete() 816 rb_erase(®ion->hva_node, &vm->regions.hva_tree); in __vm_mem_region_delete() 1036 hash_for_each_possible(vm->regions.slot_hash, region, slot_node, in vm_mem_add() 1155 hash_add(vm->regions.slot_hash, ®ion->slot_node, slot); in vm_mem_add() 1675 for (node = vm->regions.hva_tree.rb_node; node; ) { in addr_hva2gpa() [all …]
|
| /tools/testing/selftests/kvm/lib/x86/ |
| A D | sev.c | 97 hash_for_each(vm->regions.slot_hash, ctr, region, slot_node) in sev_vm_launch() 148 hash_for_each(vm->regions.slot_hash, ctr, region, slot_node) in snp_vm_launch_update()
|
| /tools/testing/memblock/ |
| A D | README | 8 Memblock is a boot time memory allocator[1] that manages memory regions before 10 memory regions, mark them as available or reserved, allocate a block of memory
|
| /tools/include/uapi/drm/ |
| A D | i915_drm.h | 3620 struct drm_i915_memory_region_info regions[]; member 3828 __u64 regions; member
|
| /tools/perf/Documentation/ |
| A D | topdown.txt | 142 It's not a good idea to do this with too short code regions, 271 short regions over time because the number of cycles covered by each
|
| A D | tips.txt | 63 To compare hot regions in two workloads use perf record -b -o file ... ; perf diff --stream file1 f…
|
| /tools/testing/selftests/kvm/include/ |
| A D | kvm_util.h | 99 struct userspace_mem_regions regions; member
|
| /tools/memory-model/Documentation/ |
| A D | access-marking.txt | 92 to x allows KCSAN to detect reads from x from within regions of code
|