/xen-4.10.0-shim-comet/xen/include/xen/ |
A D | rwlock.h | 11 spinlock_t lock; member 46 cnts = atomic_read(&lock->cnts); in _read_trylock() 70 queue_read_lock_slowpath(lock); in _read_lock() 77 _read_lock(lock); in _read_lock_irq() 84 _read_lock(lock); in _read_lock_irqsave() 102 _read_unlock(lock); in _read_unlock_irq() 108 _read_unlock(lock); in _read_unlock_irqrestore() 134 _write_lock(lock); in _write_lock_irq() 142 _write_lock(lock); in _write_lock_irqsave() 173 _write_unlock(lock); in _write_unlock_irq() [all …]
|
A D | spinlock.h | 155 void _spin_lock(spinlock_t *lock); 157 void _spin_lock_irq(spinlock_t *lock); 160 void _spin_unlock(spinlock_t *lock); 161 void _spin_unlock_irq(spinlock_t *lock); 164 int _spin_is_locked(spinlock_t *lock); 165 int _spin_trylock(spinlock_t *lock); 166 void _spin_barrier(spinlock_t *lock); 168 int _spin_trylock_recursive(spinlock_t *lock); 169 void _spin_lock_recursive(spinlock_t *lock); 170 void _spin_unlock_recursive(spinlock_t *lock); [all …]
|
A D | sched-if.h | 75 spin_lock##irq(lock, ## arg); \ 77 return lock; \ 78 spin_unlock##irq(lock, ## arg); \ 83 static inline void kind##_schedule_unlock##irq(spinlock_t *lock \ 86 ASSERT(lock == per_cpu(schedule_data, cpu).schedule_lock); \ 87 spin_unlock##irq(lock, ## arg); \ 115 spinlock_t *lock = per_cpu(schedule_data, cpu).schedule_lock; in sched_lock() local 117 if ( !spin_trylock(lock) ) in sched_lock() 119 if ( lock == per_cpu(schedule_data, cpu).schedule_lock ) in sched_lock() 120 return lock; in sched_lock() [all …]
|
/xen-4.10.0-shim-comet/xen/common/ |
A D | spinlock.c | 98 lock->profile->time_hold += NOW() - lock->profile->time_locked; \ 109 lock->profile->time_block += lock->profile->time_locked - block; \ 167 _spin_lock(lock); in _spin_lock_irq() 175 _spin_lock(lock); in _spin_lock_irqsave() 190 _spin_unlock(lock); in _spin_unlock_irq() 196 _spin_unlock(lock); in _spin_unlock_irqrestore() 210 ? lock->tickets.head != lock->tickets.tail in _spin_is_locked() 228 if (lock->profile) in _spin_trylock() 282 lock->recurse_cnt++; in _spin_trylock_recursive() 293 _spin_lock(lock); in _spin_lock_recursive() [all …]
|
A D | rwlock.c | 18 cnts = atomic_read(&lock->cnts); in rspin_until_writer_unlock() 26 void queue_read_lock_slowpath(rwlock_t *lock) in queue_read_lock_slowpath() argument 33 atomic_sub(_QR_BIAS, &lock->cnts); in queue_read_lock_slowpath() 38 spin_lock(&lock->lock); in queue_read_lock_slowpath() 51 rspin_until_writer_unlock(lock, cnts); in queue_read_lock_slowpath() 56 spin_unlock(&lock->lock); in queue_read_lock_slowpath() 68 spin_lock(&lock->lock); in queue_write_lock_slowpath() 71 if ( !atomic_read(&lock->cnts) && in queue_write_lock_slowpath() 81 cnts = atomic_read(&lock->cnts); in queue_write_lock_slowpath() 93 cnts = atomic_read(&lock->cnts); in queue_write_lock_slowpath() [all …]
|
A D | rangeset.c | 31 rwlock_t lock; member 123 write_lock(&r->lock); in rangeset_add_range() 191 write_lock(&r->lock); in rangeset_remove_range() 259 read_lock(&r->lock); in rangeset_contains_range() 262 read_unlock(&r->lock); in rangeset_contains_range() 275 read_lock(&r->lock); in rangeset_overlaps_range() 278 read_unlock(&r->lock); in rangeset_overlaps_range() 290 read_lock(&r->lock); in rangeset_report_ranges() 296 read_unlock(&r->lock); in rangeset_report_ranges() 307 write_lock(&r->lock); in rangeset_claim_range() [all …]
|
A D | timer.c | 32 spinlock_t lock; member 437 spin_unlock_irq(&ts->lock); in execute_timer() 439 spin_lock_irq(&ts->lock); in execute_timer() 462 spin_lock_irq(&ts->lock); in timer_softirq_action() 473 spin_lock_irq(&ts->lock); in timer_softirq_action() 515 spin_unlock_irq(&ts->lock); in timer_softirq_action() 571 spin_lock(&new_ts->lock); in migrate_timers_from_cpu() 576 spin_lock(&old_ts->lock); in migrate_timers_from_cpu() 595 spin_unlock(&old_ts->lock); in migrate_timers_from_cpu() 596 spin_unlock_irq(&new_ts->lock); in migrate_timers_from_cpu() [all …]
|
A D | wait.c | 85 spin_lock_init(&wq->lock); in init_waitqueue_head() 98 spin_lock(&wq->lock); in wake_up_nr() 108 spin_unlock(&wq->lock); in wake_up_nr() 216 spin_lock(&wq->lock); in prepare_to_wait() 220 spin_unlock(&wq->lock); in prepare_to_wait() 233 spin_lock(&wq->lock); in finish_wait() 240 spin_unlock(&wq->lock); in finish_wait()
|
A D | sched_null.c | 150 spin_lock_init(&prv->lock); in null_init() 443 spinlock_t *lock; in null_vcpu_insert() local 447 lock = vcpu_schedule_lock_irq(v); in null_vcpu_insert() 452 spin_unlock(lock); in null_vcpu_insert() 454 lock = vcpu_schedule_lock(v); in null_vcpu_insert() 489 spin_unlock_irq(lock); in null_vcpu_insert() 536 spinlock_t *lock; in null_vcpu_remove() local 540 lock = vcpu_schedule_lock_irq(v); in null_vcpu_remove() 558 vcpu_schedule_unlock_irq(lock, v); in null_vcpu_remove() 819 spinlock_t *lock; in null_dump_pcpu() local [all …]
|
A D | sched_credit.c | 207 spinlock_t lock; member 633 spin_lock(&prv->lock); in csched_switch_sched() 635 spin_unlock(&prv->lock); in csched_switch_sched() 1029 spinlock_t *lock; in csched_vcpu_insert() local 1038 spin_unlock_irq(lock); in csched_vcpu_insert() 1077 spin_lock_irq(&prv->lock); in csched_vcpu_remove() 1338 spinlock_t *lock; in csched_runq_sort() local 1787 if ( !lock ) in csched_load_balance() 2047 spinlock_t *lock; in csched_dump_pcpu() local 2146 spinlock_t *lock; in csched_dump() local [all …]
|
A D | schedule.c | 209 if ( unlikely(lock != NULL) ) in vcpu_runstate_get() 361 spinlock_t *lock; in sched_move_domain() local 379 spin_unlock_irq(lock); in sched_move_domain() 442 spinlock_t *lock; in vcpu_sleep_nosync() local 472 spinlock_t *lock; in vcpu_wake() local 553 spinlock_t *lock, *new_lock; in vcpu_move_nosched() local 675 spinlock_t *lock; in restore_vcpu_affinity() local 703 spin_unlock_irq(lock); in restore_vcpu_affinity() 707 spin_unlock_irq(lock); in restore_vcpu_affinity() 837 spinlock_t *lock; in vcpu_set_affinity() local [all …]
|
A D | sched_rt.c | 364 spin_lock_irqsave(&prv->lock, flags); in rt_dump_pcpu() 683 spin_lock_init(&prv->lock); in rt_init() 931 spinlock_t *lock; in rt_vcpu_insert() local 938 lock = vcpu_schedule_lock_irq(vc); in rt_vcpu_insert() 951 vcpu_schedule_unlock_irq(lock, vc); in rt_vcpu_insert() 964 spinlock_t *lock; in rt_vcpu_remove() local 970 lock = vcpu_schedule_lock_irq(vc); in rt_vcpu_remove() 977 vcpu_schedule_unlock_irq(lock, vc); in rt_vcpu_remove() 1385 vcpu_schedule_unlock_irq(lock, vc); in rt_context_saved() 1513 spin_lock_irq(&prv->lock); in repl_timer_handler() [all …]
|
A D | sched_credit2.c | 1863 spinlock_t *lock; in unpark_parked_vcpus() local 2377 read_unlock(&prv->lock); in csched2_cpu_pick() 2579 read_unlock(&prv->lock); in balance_load() 2828 spinlock_t *lock; in csched2_dom_cntl() local 3094 spinlock_t *lock; in csched2_vcpu_insert() local 3104 spin_unlock_irq(lock); in csched2_vcpu_insert() 3132 spinlock_t *lock; in csched2_vcpu_remove() local 3854 write_lock(&prv->lock); in csched2_switch_sched() 3904 spin_lock(&rqd->lock); in csched2_deinit_pdata() 3922 spin_unlock(&rqd->lock); in csched2_deinit_pdata() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/ |
A D | mm-locks.h | 38 spin_lock_init(&l->lock); in mm_lock_init() 46 return (l->lock.recurse_cpu == current->processor); in mm_locked_by_me() 72 spin_lock_recursive(&l->lock); in _mm_lock() 73 if ( l->lock.recurse_cnt == 1 ) in _mm_lock() 144 percpu_read_lock(p2m_percpu_rwlock, &l->lock); in _mm_read_lock() 151 percpu_read_unlock(p2m_percpu_rwlock, &l->lock); in mm_read_unlock() 181 if ( l->lock.recurse_cnt == 1 ) in mm_unlock() 186 spin_unlock_recursive(&l->lock); in mm_unlock() 293 mm_write_lock(altp2m, &(p)->lock); \ 295 mm_write_lock(p2m, &(p)->lock); \ [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/hvm/ |
A D | pmtimer.c | 76 spin_lock(&s->lock); in hvm_acpi_power_button() 79 spin_unlock(&s->lock); in hvm_acpi_power_button() 89 spin_lock(&s->lock); in hvm_acpi_sleep_button() 92 spin_unlock(&s->lock); in hvm_acpi_sleep_button() 133 spin_lock(&s->lock); in pmt_timer_callback() 151 spin_unlock(&s->lock); in pmt_timer_callback() 169 spin_lock(&s->lock); in handle_evt_io() 212 spin_unlock(&s->lock); in handle_evt_io() 262 spin_lock(&s->lock); in acpi_save() 278 spin_unlock(&s->lock); in acpi_save() [all …]
|
A D | i8254.c | 248 spin_lock(&pit->lock); in pit_ioport_write() 314 spin_unlock(&pit->lock); in pit_ioport_write() 325 spin_lock(&pit->lock); in pit_ioport_read() 377 spin_unlock(&pit->lock); in pit_ioport_read() 388 spin_lock(&pit->lock); in pit_stop_channel0_irq() 390 spin_unlock(&pit->lock); in pit_stop_channel0_irq() 401 spin_lock(&pit->lock); in pit_save() 405 spin_unlock(&pit->lock); in pit_save() 418 spin_lock(&pit->lock); in pit_load() 455 spin_lock(&pit->lock); in pit_reset() [all …]
|
A D | rtc.c | 87 spin_lock(&s->lock); in rtc_pf_callback() 101 spin_unlock(&s->lock); in rtc_pf_callback() 223 spin_lock(&s->lock); in rtc_update_timer() 236 spin_lock(&s->lock); in rtc_update_timer2() 427 spin_lock(&s->lock); in rtc_alarm_cb() 443 spin_lock(&s->lock); in rtc_ioport_write() 654 spin_lock(&s->lock); in rtc_ioport_read() 748 spin_lock(&s->lock); in rtc_save() 762 spin_lock(&s->lock); in rtc_load() 818 spin_lock(&s->lock); in rtc_init() [all …]
|
A D | ioreq.c | 316 spin_lock(&s->lock); in hvm_ioreq_server_add_vcpu() 346 spin_unlock(&s->lock); in hvm_ioreq_server_add_vcpu() 353 spin_unlock(&s->lock); in hvm_ioreq_server_add_vcpu() 365 spin_lock(&s->lock); in hvm_ioreq_server_remove_vcpu() 385 spin_unlock(&s->lock); in hvm_ioreq_server_remove_vcpu() 392 spin_lock(&s->lock); in hvm_ioreq_server_remove_all_vcpus() 411 spin_unlock(&s->lock); in hvm_ioreq_server_remove_all_vcpus() 552 spin_lock(&s->lock); in hvm_ioreq_server_enable() 573 spin_unlock(&s->lock); in hvm_ioreq_server_enable() 582 spin_lock(&s->lock); in hvm_ioreq_server_disable() [all …]
|
/xen-4.10.0-shim-comet/unmodified_drivers/linux-2.6/platform-pci/ |
A D | evtchn.c | 49 spinlock_t lock; member 129 spin_lock_irq(&irq_evtchn[irq].lock); in bind_listening_port_to_irqhandler() 150 spin_unlock_irq(&irq_evtchn[irq].lock); in bind_listening_port_to_irqhandler() 169 spin_lock_irq(&irq_evtchn[irq].lock); in bind_caller_port_to_irqhandler() 180 spin_unlock_irq(&irq_evtchn[irq].lock); in bind_caller_port_to_irqhandler() 190 spin_lock_irq(&irq_evtchn[irq].lock); in unbind_from_irqhandler() 295 spin_lock(&irq_evtchn[irq].lock); in evtchn_interrupt() 301 spin_unlock(&irq_evtchn[irq].lock); in evtchn_interrupt() 305 spin_unlock(&irq_evtchn[irq].lock); in evtchn_interrupt() 311 spin_lock(&irq_evtchn[irq].lock); in evtchn_interrupt() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/arm/ |
A D | irq.c | 133 ASSERT(spin_is_locked(&desc->lock)); in irq_get_guest_info() 203 spin_lock(&desc->lock); in do_IRQ() 249 spin_unlock_irq(&desc->lock); in do_IRQ() 257 spin_lock_irq(&desc->lock); in do_IRQ() 265 spin_unlock(&desc->lock); in do_IRQ() 277 spin_lock_irqsave(&desc->lock,flags); in release_irq() 306 spin_unlock_irqrestore(&desc->lock,flags); in release_irq() 351 spin_lock_irqsave(&desc->lock, flags); in setup_irq() 460 spin_lock_irqsave(&desc->lock, flags); in route_irq_to_guest() 548 spin_lock_irqsave(&desc->lock, flags); in release_guest_irq() [all …]
|
A D | vgic.c | 88 spin_lock_init(&rank->lock); in vgic_rank_init() 132 spin_lock_init(&d->arch.vgic.lock); in domain_vgic_init() 216 spin_lock_init(&v->arch.vgic.lock); in vcpu_vgic_init() 254 spin_lock_irqsave(&old->arch.vgic.lock, flags); in vgic_migrate_irq() 354 spin_lock_irqsave(&desc->lock, flags); in vgic_disable_irqs() 370 ASSERT(spin_is_locked(&r->lock)); in vgic_get_virq_type() 403 spin_lock_irqsave(&p->desc->lock, flags); in vgic_enable_irqs() 504 spin_lock_irqsave(&v->arch.vgic.lock, flags); in vgic_clear_pending_irqs() 508 spin_unlock_irqrestore(&v->arch.vgic.lock, flags); in vgic_clear_pending_irqs() 518 spin_lock_irqsave(&v->arch.vgic.lock, flags); in vgic_vcpu_inject_irq() [all …]
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/amd/ |
A D | iommu_intr.c | 169 spinlock_t *lock; in update_intremap_entry_from_ioapic() local 180 spin_lock_irqsave(lock, flags); in update_intremap_entry_from_ioapic() 213 spin_unlock_irqrestore(lock, flags); in update_intremap_entry_from_ioapic() 236 spinlock_t *lock; in amd_iommu_setup_ioapic_remapping() local 433 spinlock_t *lock; in update_intremap_entry_from_msi_msg() local 442 spin_lock_irqsave(lock, flags); in update_intremap_entry_from_msi_msg() 451 spin_lock_irqsave(lock, flags); in update_intremap_entry_from_msi_msg() 471 spin_unlock_irqrestore(lock, flags); in update_intremap_entry_from_msi_msg() 635 spinlock_t *lock; in amd_setup_hpet_msi() local 653 spin_lock_irqsave(lock, flags); in amd_setup_hpet_msi() [all …]
|
/xen-4.10.0-shim-comet/tools/libxl/ |
A D | libxl_internal.c | 369 if (pthread_mutex_init(lock, &attr) != 0) { in libxl__init_recursive_mutex() 406 libxl__domain_userdata_lock *lock = NULL; in libxl__lock_domain_userdata() local 415 lock->path = libxl__strdup(NOGC, lockfile); in libxl__lock_domain_userdata() 423 lock->carefd = libxl__carefd_opened(CTX, fd); in libxl__lock_domain_userdata() 458 libxl__carefd_close(lock->carefd); in libxl__lock_domain_userdata() 467 return lock; in libxl__lock_domain_userdata() 470 if (lock) libxl__unlock_domain_userdata(lock); in libxl__lock_domain_userdata() 490 if (lock->path) unlink(lock->path); in libxl__unlock_domain_userdata() 491 if (lock->carefd) libxl__carefd_close(lock->carefd); in libxl__unlock_domain_userdata() 492 free(lock->path); in libxl__unlock_domain_userdata() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/ |
A D | irq.c | 663 spin_lock(&desc->lock); in irq_move_cleanup_interrupt() 704 spin_unlock(&desc->lock); in irq_move_cleanup_interrupt() 807 spin_lock(&lock); in alloc_direct_apic_vector() 812 spin_unlock(&lock); in alloc_direct_apic_vector() 864 spin_lock(&desc->lock); in do_IRQ() 930 spin_unlock(&desc->lock); in do_IRQ() 945 spin_lock(&desc->lock); in irq_ratelimit_timer_fn() 1382 spin_lock(&desc->lock); in flush_ready_eoi() 1421 spin_lock(&desc->lock); in set_eoi_ready() 1423 spin_unlock(&desc->lock); in set_eoi_ready() [all …]
|
/xen-4.10.0-shim-comet/tools/libs/toolcore/ |
A D | handlereg.c | 36 static void lock(void) { in lock() function 47 lock(); in xentoolcore__register_active_handle() 53 lock(); in xentoolcore__deregister_active_handle() 62 lock(); in xentoolcore_restrict_all()
|