| /hyp/core/cspace_twolevel/src/ |
| A D | cspace_twolevel.c | 308 object_header_t *header; in cspace_destroy_cap_table() local 329 spinlock_acquire(&header->cap_list_lock); in cspace_destroy_cap_table() 332 spinlock_release(&header->cap_list_lock); in cspace_destroy_cap_table() 638 object_header_t *header; in cspace_copy_cap() local 675 spinlock_acquire(&header->cap_list_lock); in cspace_copy_cap() 686 list_insert_after_node(&header->cap_list, in cspace_copy_cap() 691 spinlock_release(&header->cap_list_lock); in cspace_copy_cap() 772 object_header_t *header; in cspace_revoke_caps() local 793 spinlock_acquire(&header->cap_list_lock); in cspace_revoke_caps() 806 list_t *list = &header->cap_list; in cspace_revoke_caps() [all …]
|
| A D | hypercalls.c | 126 spinlock_acquire(&target_cspace->header.lock); in hypercall_cspace_configure() 128 if (atomic_load_relaxed(&target_cspace->header.state) == in hypercall_cspace_configure() 135 spinlock_release(&target_cspace->header.lock); in hypercall_cspace_configure() 172 spinlock_acquire(&thread->header.lock); in hypercall_cspace_attach_thread() 174 if (atomic_load_relaxed(&thread->header.state) == OBJECT_STATE_INIT) { in hypercall_cspace_attach_thread() 180 spinlock_release(&thread->header.lock); in hypercall_cspace_attach_thread()
|
| A D | cspace_tests.c | 48 spinlock_acquire(&test_cspace->header.lock); in tests_cspace_init() 50 spinlock_release(&test_cspace->header.lock); in tests_cspace_init()
|
| /hyp/vm/vrtc_pl031/src/ |
| A D | hypercalls.c | 43 spinlock_acquire(&vrtc->header.lock); in hypercall_vrtc_configure() 44 if (atomic_load_relaxed(&vrtc->header.state) == OBJECT_STATE_INIT) { in hypercall_vrtc_configure() 49 spinlock_release(&vrtc->header.lock); in hypercall_vrtc_configure() 124 spinlock_acquire(&addrspace->header.lock); in hypercall_vrtc_attach_addrspace() 125 if (atomic_load_relaxed(&addrspace->header.state) != in hypercall_vrtc_attach_addrspace() 144 spinlock_release(&addrspace->header.lock); in hypercall_vrtc_attach_addrspace()
|
| /hyp/vm/vcpu/aarch64/src/ |
| A D | hypercalls.c | 65 spinlock_acquire(&vcpu->header.lock); in hypercall_vcpu_configure() 66 object_state_t state = atomic_load_relaxed(&vcpu->header.state); in hypercall_vcpu_configure() 72 spinlock_release(&vcpu->header.lock); in hypercall_vcpu_configure() 198 spinlock_acquire(&vcpu->header.lock); 199 object_state_t state = atomic_load_relaxed(&vcpu->header.state); 211 spinlock_release(&vcpu->header.lock); 342 spinlock_acquire(&vcpu->header.lock); 343 object_state_t state = atomic_load_relaxed(&vcpu->header.state); 351 spinlock_release(&vcpu->header.lock); 379 spinlock_acquire(&vcpu->header.lock); [all …]
|
| /hyp/vm/vpm_base/src/ |
| A D | hypercalls.c | 45 spinlock_acquire(&vpm_group->header.lock); in hypercall_vpm_group_configure() 47 if (atomic_load_relaxed(&vpm_group->header.state) == in hypercall_vpm_group_configure() 54 spinlock_release(&vpm_group->header.lock); in hypercall_vpm_group_configure() 89 spinlock_acquire(&thread->header.lock); in hypercall_vpm_group_attach_vcpu() 90 if (atomic_load_relaxed(&thread->header.state) == OBJECT_STATE_INIT) { in hypercall_vpm_group_attach_vcpu() 95 spinlock_release(&thread->header.lock); in hypercall_vpm_group_attach_vcpu()
|
| /hyp/vm/vic_base/src/ |
| A D | hypercalls.c | 103 spinlock_acquire(&vic->header.lock); in hypercall_vic_configure() 104 if (atomic_load_relaxed(&vic->header.state) == OBJECT_STATE_INIT) { in hypercall_vic_configure() 111 spinlock_release(&vic->header.lock); in hypercall_vic_configure() 146 spinlock_acquire(&thread->header.lock); in hypercall_vic_attach_vcpu() 147 if (atomic_load_relaxed(&thread->header.state) == OBJECT_STATE_INIT) { in hypercall_vic_attach_vcpu() 152 spinlock_release(&thread->header.lock); in hypercall_vic_attach_vcpu()
|
| /hyp/mem/memextent/src/ |
| A D | memextent_tests.c | 157 spinlock_acquire(&me->header.lock); in create_memextent() 165 spinlock_release(&me->header.lock); in create_memextent() 370 (void)partition_destroy_memextent(&me_dd2->header.rcu_entry); in tests_memextent_test1() 373 (void)partition_destroy_memextent(&me_dd->header.rcu_entry); in tests_memextent_test1() 376 (void)partition_destroy_memextent(&me_d2->header.rcu_entry); in tests_memextent_test1() 379 (void)partition_destroy_memextent(&me_d->header.rcu_entry); in tests_memextent_test1() 382 (void)partition_destroy_memextent(&me2->header.rcu_entry); in tests_memextent_test1() 385 (void)partition_destroy_memextent(&me->header.rcu_entry); in tests_memextent_test1() 504 (void)partition_destroy_memextent(&me_d->header.rcu_entry); in tests_memextent_test2() 511 (void)partition_destroy_memextent(&me->header.rcu_entry); in tests_memextent_test2() [all …]
|
| A D | hypercalls.c | 94 spinlock_acquire(&target_me->header.lock); in hypercall_memextent_configure() 96 if (atomic_load_relaxed(&target_me->header.state) == in hypercall_memextent_configure() 104 spinlock_release(&target_me->header.lock); in hypercall_memextent_configure() 144 spinlock_acquire(&target_me->header.lock); in hypercall_memextent_configure_derive() 146 if (atomic_load_relaxed(&target_me->header.state) == in hypercall_memextent_configure_derive() 155 spinlock_release(&target_me->header.lock); in hypercall_memextent_configure_derive() 202 if (child.r->header.partition != p.r) { in hypercall_memextent_donate_child()
|
| A D | memextent.c | 806 me_ret = partition_allocate_memextent(parent->header.partition, in memextent_derive() 818 spinlock_acquire(&me->header.lock); in memextent_derive() 822 spinlock_release(&me->header.lock); in memextent_derive() 827 spinlock_release(&me->header.lock); in memextent_derive() 872 if (owner != me->header.partition) { in memextent_attach() 918 assert(owner == me->header.partition); in memextent_detach()
|
| /hyp/mem/addrspace/src/ |
| A D | addrspace.c | 93 assert(atomic_load_relaxed(&addrspace->header.state) == in addrspace_attach_thread() 187 root_thread->header.partition, as_params); in addrspace_handle_rootvm_init() 195 spinlock_acquire(&root_addrspace->header.lock); in addrspace_handle_rootvm_init() 199 spinlock_release(&root_addrspace->header.lock); in addrspace_handle_rootvm_init() 202 spinlock_release(&root_addrspace->header.lock); in addrspace_handle_rootvm_init() 289 hyp_aspace_deallocate(addrspace->header.partition, in addrspace_handle_object_cleanup_addrspace() 377 partition_t *partition = addrspace->header.partition; in addrspace_handle_object_activate_addrspace() 425 memextent_detach(addrspace->header.partition, in addrspace_handle_object_deactivate_addrspace() 433 pgtable_vm_destroy(addrspace->header.partition, in addrspace_handle_object_deactivate_addrspace() 468 err = pgtable_vm_map(addrspace->header.partition, in addrspace_map() [all …]
|
| A D | hypercalls.c | 55 spinlock_acquire(&thread->header.lock); in hypercall_addrspace_attach_thread() 57 if (atomic_load_relaxed(&thread->header.state) == OBJECT_STATE_INIT) { in hypercall_addrspace_attach_thread() 63 spinlock_release(&thread->header.lock); in hypercall_addrspace_attach_thread() 296 spinlock_acquire(&target_as->header.lock); in hypercall_addrspace_configure() 298 if (atomic_load_relaxed(&target_as->header.state) == in hypercall_addrspace_configure() 305 spinlock_release(&target_as->header.lock); in hypercall_addrspace_configure() 406 spinlock_acquire(&target_as->header.lock); in hypercall_addrspace_configure_info_area() 407 if (atomic_load_relaxed(&target_as->header.state) == in hypercall_addrspace_configure_info_area() 414 spinlock_release(&target_as->header.lock); in hypercall_addrspace_configure_info_area() 440 object_state_t state = atomic_load_relaxed(&target_as->header.state); in hypercall_addrspace_configure_vmmio()
|
| /hyp/core/partition_standard/src/ |
| A D | init.c | 58 refcount_init(&hyp_partition.header.refcount); in partition_standard_handle_boot_cold_init() 59 hyp_partition.header.type = OBJECT_TYPE_PARTITION; in partition_standard_handle_boot_cold_init() 60 atomic_store_release(&hyp_partition.header.state, OBJECT_STATE_ACTIVE); in partition_standard_handle_boot_cold_init()
|
| A D | partition.c | 129 assert(partition->header.partition != NULL); in partition_standard_handle_object_activate_partition() 130 assert(partition->header.partition != partition); in partition_standard_handle_object_activate_partition() 134 &partition->header.partition->options)) { in partition_standard_handle_object_activate_partition() 333 spinlock_acquire(&partition->header.lock); in partition_map_and_add_heap() 338 spinlock_release(&partition->header.lock); in partition_map_and_add_heap()
|
| /hyp/core/idle/src/ |
| A D | idle.c | 86 thread->header.partition = in idle_thread_init_boot() 88 thread->header.type = OBJECT_TYPE_THREAD; in idle_thread_init_boot() 89 atomic_init(&thread->header.state, OBJECT_STATE_INIT); in idle_thread_init_boot()
|
| /hyp/core/thread_standard/src/ |
| A D | init.c | 50 refcount_init(&idle_thread->header.refcount); in thread_standard_handle_boot_runtime_first_init()
|
| A D | thread.c | 61 thread->header.partition, stack_size, thread_stack_alloc_align); in thread_standard_handle_object_create_thread() 92 (void)partition_free(thread->header.partition, in thread_standard_unwind_object_create_thread() 155 (void)partition_free(thread->header.partition, in thread_standard_handle_object_deactivate_thread()
|
| /hyp/platform/soc_qemu/src/ |
| A D | boot.c | 117 spinlock_acquire(&me->header.lock); in create_memextent() 122 spinlock_release(&me->header.lock); in create_memextent()
|
| /hyp/ipc/msgqueue/src/ |
| A D | hypercalls.c | 273 spinlock_acquire(&target_msgqueue->header.lock); in hypercall_msgqueue_configure() 280 if (atomic_load_relaxed(&target_msgqueue->header.state) == in hypercall_msgqueue_configure() 288 spinlock_release(&target_msgqueue->header.lock); in hypercall_msgqueue_configure()
|
| A D | msgqueue.c | 196 partition_t *partition = msgqueue->header.partition; in msgqueue_handle_object_activate_msgqueue() 224 partition_t *partition = msgqueue->header.partition; in msgqueue_handle_object_deactivate_msgqueue()
|
| /hyp/vm/rootvm/src/ |
| A D | rootvm_init.c | 194 spinlock_acquire_nopreempt(&root_cspace->header.lock); in rootvm_init() 196 spinlock_release_nopreempt(&root_cspace->header.lock); in rootvm_init() 199 spinlock_release_nopreempt(&root_cspace->header.lock); in rootvm_init()
|
| /hyp/vm/vgic/src/ |
| A D | distrib.c | 62 partition_t *partition = vic->header.partition; in vgic_handle_object_create_vic() 142 partition_t *partition = vic->header.partition; in vgic_handle_object_activate_vic() 338 partition_t *partition = vic->header.partition; in vgic_handle_object_cleanup_vic() 358 (void)partition_free(vic->header.partition, in vgic_handle_object_cleanup_vic() 575 vcpu->header.partition, vlpi_pendbase_size, in vgic_handle_object_activate_thread() 857 spinlock_acquire(&vic_r.r->header.lock); in vgic_handle_rootvm_init() 883 spinlock_release(&vic_r.r->header.lock); in vgic_handle_rootvm_init() 886 spinlock_release(&vic_r.r->header.lock); in vgic_handle_rootvm_init() 2045 gicr_vcpu->header.partition, in vgic_gicr_enable_lpis() 2893 spinlock_acquire(&vic->header.lock); in hypercall_vgic_set_mpidr_mapping() [all …]
|
| /hyp/core/thread_standard/aarch64/src/ |
| A D | thread_arch.c | 277 partition_t *partition = thread->header.partition; in thread_arch_map_stack() 295 pgtable_hyp_unmap(thread->header.partition, thread->stack_base, in thread_arch_unmap_stack()
|
| /hyp/vm/virtio_mmio/src/ |
| A D | virtio_mmio.c | 90 partition_t *partition = virtio_mmio->header.partition; in virtio_mmio_handle_object_activate_virtio_mmio() 178 partition_t *partition = virtio_mmio->header.partition; in virtio_mmio_handle_object_cleanup_virtio_mmio()
|
| /hyp/vm/vcpu/src/ |
| A D | vcpu.c | 131 &thread->header.partition->options)) { in vcpu_handle_vcpu_activate_thread()
|