Searched refs:arch_spin_lock (Results 1 – 25 of 30) sorted by relevance
12
30 static inline void arch_spin_lock(arch_spinlock_t *x) in arch_spin_lock() function85 arch_spin_lock(&(rw->lock_mutex)); in arch_read_trylock()109 arch_spin_lock(&(rw->lock_mutex)); in arch_write_trylock()144 arch_spin_lock(&(rw->lock_mutex)); in arch_read_unlock()155 arch_spin_lock(&(rw->lock_mutex)); in arch_write_unlock()
24 arch_spin_lock(s); in _futex_spin_lock_irqsave()
39 arch_spin_lock(s); \
17 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function222 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function304 arch_spin_lock(&(rw->lock_mutex)); in arch_read_trylock()328 arch_spin_lock(&(rw->lock_mutex)); in arch_write_trylock()363 arch_spin_lock(&(rw->lock_mutex)); in arch_read_unlock()374 arch_spin_lock(&(rw->lock_mutex)); in arch_write_unlock()
113 arch_spin_lock(&smp_atomic_ops_lock); \
62 arch_spin_lock(&sync_lock); in check_counter_warp()85 arch_spin_lock(&sync_lock); in check_counter_warp()
210 arch_spin_lock(&mcpm_lock); in mcpm_cpu_power_up()256 arch_spin_lock(&mcpm_lock); in mcpm_cpu_power_down()333 arch_spin_lock(&mcpm_lock); in mcpm_cpu_suspend()353 arch_spin_lock(&mcpm_lock); in mcpm_cpu_powered_up()
353 arch_spin_lock(&kmmio_lock); in post_kmmio_handler()447 arch_spin_lock(&kmmio_lock); in register_kmmio_probe()504 arch_spin_lock(&kmmio_lock); in remove_kmmio_fault_pages()552 arch_spin_lock(&kmmio_lock); in unregister_kmmio_probe()
115 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
30 static inline void arch_spin_lock(arch_spinlock_t * lock) in arch_spin_lock() function
28 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
26 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
164 #define arch_spin_lock(l) queued_spin_lock(l) macro
114 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
18 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
283 arch_spin_lock(&sync_lock); in check_tsc_warp()306 arch_spin_lock(&sync_lock); in check_tsc_warp()
344 arch_spin_lock(&die_lock); in oops_begin()
265 arch_spin_lock(&sb->lock); in xive_native_esb_fault()370 arch_spin_lock(&sb->lock); in kvmppc_xive_native_set_source()426 arch_spin_lock(&sb->lock); in kvmppc_xive_native_update_source_config()529 arch_spin_lock(&sb->lock); in kvmppc_xive_native_sync_source()845 arch_spin_lock(&sb->lock); in kvmppc_xive_reset()927 arch_spin_lock(&sb->lock); in kvmppc_xive_native_eq_sync()1248 arch_spin_lock(&sb->lock); in xive_native_debug_show()
148 arch_spin_lock(&ics->lock); in write_xive()213 arch_spin_lock(&ics->lock); in kvmppc_xics_get_xive()412 arch_spin_lock(&ics->lock); in icp_deliver_irq()998 arch_spin_lock(&ics->lock); in xics_debug_show()1192 arch_spin_lock(&ics->lock); in xics_get_source()1257 arch_spin_lock(&ics->lock); in xics_set_source()
573 arch_spin_lock(&sb->lock); in xive_vm_h_eoi()1098 arch_spin_lock(&sb->lock); in xive_lock_and_mask()1136 arch_spin_lock(&sb->lock); in xive_lock_for_unmask()1384 arch_spin_lock(&sb->lock); in kvmppc_xive_get_xive()1752 arch_spin_lock(&sb->lock); in kvmppc_xive_disable_vcpu_interrupts()2945 arch_spin_lock(&sb->lock); in xive_debug_show()
319 arch_spin_lock(&rw->wait); in arch_read_lock_wait()337 arch_spin_lock(&rw->wait); in arch_write_lock_wait()
82 static inline void arch_spin_lock(arch_spinlock_t *lp) in arch_spin_lock() function
110 arch_spin_lock(&timebase_lock); in pas_give_timebase()127 arch_spin_lock(&timebase_lock); in pas_take_timebase()
56 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
184 arch_spin_lock(&nmi_lock); in nmi_dump()
Completed in 31 milliseconds