| /kernel/lib/unittest/ |
| A D | user_memory.cpp | 37 fbl::RefPtr<VmMapping> mapping; in Create() local 41 vmo, 0, arch_mmu_flags, "unittest", &mapping); in Create() 47 if (mapping) { in Create() 48 zx_status_t status = mapping->Unmap(mapping->base(), mapping->size()); in Create() 54 ktl::unique_ptr<UserMemory> mem(new (&ac) UserMemory(mapping)); in Create()
|
| /kernel/lib/hypervisor/ |
| A D | guest_physical_address_space.cpp | 74 fbl::RefPtr<VmMapping> mapping; in MapInterruptController() local 78 &mapping); in MapInterruptController() 84 status = mapping->MapRange(0, vmo->size(), true); in MapInterruptController() 86 mapping->Destroy(); in MapInterruptController() 108 fbl::RefPtr<VmMapping> mapping = FindMapping(RootVmar(), guest_paddr); in GetPage() local 109 if (!mapping) { in GetPage() 114 zx_gpaddr_t offset = guest_paddr - mapping->base(); in GetPage() 120 if (!mapping) { in PageFault() 129 if (mapping->arch_mmu_flags() & ARCH_MMU_FLAG_PERM_WRITE) { in PageFault() 132 if (mapping->arch_mmu_flags() & ARCH_MMU_FLAG_PERM_EXECUTE) { in PageFault() [all …]
|
| A D | hypervisor_unittest.cpp | 61 fbl::RefPtr<VmMapping> mapping; in create_mapping() local 63 0 /* vmo_offset */, mmu_flags, "vmo", &mapping); in create_mapping()
|
| /kernel/dev/udisplay/ |
| A D | udisplay.cpp | 60 fbl::RefPtr<VmMapping> mapping; in udisplay_set_framebuffer() local 63 ktl::move(vmo), 0, kFramebufferArchMmuFlags, "framebuffer_vmo", &mapping); in udisplay_set_framebuffer() 67 status = mapping->MapRange(0, size, true); in udisplay_set_framebuffer() 69 mapping->Destroy(); in udisplay_set_framebuffer() 73 g_udisplay.framebuffer_virt = reinterpret_cast<void*>(mapping->base()); in udisplay_set_framebuffer() 75 g_udisplay.framebuffer_vmo_mapping = mapping; in udisplay_set_framebuffer()
|
| /kernel/vm/ |
| A D | kstack.cpp | 115 fbl::RefPtr<VmMapping> mapping; in vm_allocate_kstack() local 117 zx_status_t status = allocate_vmar(false, &mapping, &vmar); in vm_allocate_kstack() 121 stack->size = mapping->size(); in vm_allocate_kstack() 122 stack->base = mapping->base(); in vm_allocate_kstack() 123 stack->top = mapping->base() + DEFAULT_STACK_SIZE; in vm_allocate_kstack() 129 status = allocate_vmar(true, &mapping, &vmar); in vm_allocate_kstack() 134 stack->size = mapping->size(); in vm_allocate_kstack() 135 stack->unsafe_base = mapping->base(); in vm_allocate_kstack()
|
| A D | vm_mapping.cpp | 151 fbl::RefPtr<VmMapping> mapping(fbl::AdoptRef( in ProtectLocked() local 163 mapping->ActivateLocked(); in ProtectLocked() 172 fbl::RefPtr<VmMapping> mapping(fbl::AdoptRef( in ProtectLocked() local 184 mapping->ActivateLocked(); in ProtectLocked() 300 fbl::RefPtr<VmMapping> mapping(fbl::AdoptRef( in UnmapLocked() local 316 mapping->ActivateLocked(); in UnmapLocked() 395 VmMappingCoalescer(VmMapping* mapping, vaddr_t base); 435 VmMappingCoalescer::VmMappingCoalescer(VmMapping* mapping, vaddr_t base) in VmMappingCoalescer() argument 436 : mapping_(mapping), base_(base), count_(0), aborted_(false) {} in VmMappingCoalescer()
|
| A D | vm_address_region.cpp | 545 VmMapping* mapping = curr->as_vm_mapping().get(); in EnumerateChildrenLocked() local 546 DEBUG_ASSERT(mapping != nullptr); in EnumerateChildrenLocked() 547 if (!ve->OnVmMapping(mapping, this, depth)) { in EnumerateChildrenLocked()
|
| /kernel/lib/vdso/ |
| A D | rodso.cpp | 68 fbl::RefPtr<VmMapping> mapping; in MapSegment() local 70 start_offset, len, flags, &mapping); in MapSegment() 80 DEBUG_ASSERT(mapping->base() == vmar->vmar()->base() + vmar_offset); in MapSegment() 83 mapping->base(), mapping->base() + len); in MapSegment()
|
| /kernel/lib/unittest/include/lib/unittest/ |
| A D | user_memory.h | 34 UserMemory(fbl::RefPtr<VmMapping> mapping) in UserMemory() argument 35 : mapping_(ktl::move(mapping)) {} in UserMemory()
|
| /kernel/platform/pc/ |
| A D | smbios.cpp | 66 fbl::RefPtr<VmMapping>* mapping, uintptr_t* struct_table_virt) { in MapStructs2_1() argument 87 *mapping = ktl::move(m); in MapStructs2_1() 108 fbl::RefPtr<VmMapping> mapping; in pc_init_smbios() local 109 auto cleanup_mapping = fbl::MakeAutoCall([&mapping] { in pc_init_smbios() 110 if (mapping) { in pc_init_smbios() 111 mapping->Destroy(); in pc_init_smbios() 132 status = MapStructs2_1(ep, &mapping, &struct_table_virt); in pc_init_smbios()
|
| /kernel/arch/x86/ |
| A D | descriptor.cpp | 158 fbl::RefPtr<VmMapping> mapping; in gdt_setup() local 161 /*vmo_offset*/0u, mmu_flags, "gdt", &mapping); in gdt_setup() 164 status = mapping->MapRange(0, gdt_real_size, /*commit*/true); in gdt_setup() 167 memcpy((void*)mapping->base(), _temp_gdt, gdt_real_size); in gdt_setup() 168 gdt = mapping->base(); in gdt_setup()
|
| A D | perf_mon.cpp | 264 fbl::RefPtr<VmMapping> mapping; member 1435 if (state->mchbar_data.mapping) { in x86_perfmon_unmap_buffers_locked() 1436 state->mchbar_data.mapping->Destroy(); in x86_perfmon_unmap_buffers_locked() 1438 state->mchbar_data.mapping.reset(); in x86_perfmon_unmap_buffers_locked() 1469 fbl::RefPtr<VmMapping> mapping; in x86_map_mchbar_stat_registers() local 1473 &mapping); in x86_map_mchbar_stat_registers() 1477 status = mapping->MapRange(0, PAGE_SIZE, false); in x86_map_mchbar_stat_registers() 1481 state->mchbar_data.mapping = mapping; in x86_map_mchbar_stat_registers() 1483 reinterpret_cast<void*>(mapping->base() + begin_offset); in x86_map_mchbar_stat_registers() 1506 mapping->base(), num_bytes_to_map); in x86_map_mchbar_stat_registers()
|
| /kernel/lib/hypervisor/include/hypervisor/ |
| A D | guest_physical_address_space.h | 23 GuestPtr(fbl::RefPtr<VmMapping> mapping, zx_vaddr_t offset) in GuestPtr() argument 24 : mapping_(ktl::move(mapping)), offset_(offset) {} in GuestPtr()
|
| /kernel/lib/fbl/ |
| A D | arena_tests.cpp | 175 const auto mapping = region->as_vm_mapping(); in count_committed_pages() local 176 if (mapping == nullptr) { in count_committed_pages() 183 auto start_off = ROUNDDOWN(start, PAGE_SIZE) - mapping->base(); in count_committed_pages() 184 auto end_off = ROUNDUP(end, PAGE_SIZE) - mapping->base(); in count_committed_pages() 185 *committed = mapping->vmo()->AllocatedPagesInRange( in count_committed_pages() 186 start_off + mapping->object_offset(), end_off - start_off); in count_committed_pages()
|
| A D | arena.cpp | 143 void Arena::Pool::Init(const char* name, fbl::RefPtr<VmMapping> mapping, in Init() argument 145 DEBUG_ASSERT(mapping != nullptr); in Init() 150 mapping_ = mapping; in Init()
|
| /kernel/arch/x86/page_tables/include/arch/x86/page_tables/ |
| A D | page_tables.h | 170 volatile pt_entry_t** mapping) TA_REQ(lock_); 173 volatile pt_entry_t** mapping) TA_REQ(lock_);
|
| /kernel/arch/x86/page_tables/ |
| A D | page_tables.cpp | 376 volatile pt_entry_t** mapping) { in GetMapping() argument 379 DEBUG_ASSERT(mapping); in GetMapping() 382 return GetMappingL0(table, vaddr, ret_level, mapping); in GetMapping() 395 *mapping = e; in GetMapping() 401 return GetMapping(next_table, vaddr, lower_level(level), ret_level, mapping); in GetMapping() 406 volatile pt_entry_t** mapping) { in GetMappingL0() argument 413 *mapping = e; in GetMappingL0()
|
| /kernel/lib/fbl/include/fbl/ |
| A D | arena.h | 74 void Init(const char* name, fbl::RefPtr<VmMapping> mapping,
|
| /kernel/object/ |
| A D | diagnostics.cpp | 496 zx_info_maps_mapping_t* u = &entry.u.mapping; in OnVmMapping()
|