Home
last modified time | relevance | path

Searched defs:lock (Results 1 – 25 of 147) sorted by relevance

123456

/arch/hexagon/include/asm/
A Dspinlock.h28 static inline void arch_read_lock(arch_rwlock_t *lock) in arch_read_lock()
43 static inline void arch_read_unlock(arch_rwlock_t *lock) in arch_read_unlock()
58 static inline int arch_read_trylock(arch_rwlock_t *lock) in arch_read_trylock()
76 static inline void arch_write_lock(arch_rwlock_t *lock) in arch_write_lock()
91 static inline int arch_write_trylock(arch_rwlock_t *lock) in arch_write_trylock()
109 static inline void arch_write_unlock(arch_rwlock_t *lock) in arch_write_unlock()
115 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
130 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
136 static inline unsigned int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
A Dspinlock_types.h16 volatile unsigned int lock; member
22 volatile unsigned int lock; member
/arch/alpha/include/asm/
A Dspinlock.h19 static inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked()
24 static inline void arch_spin_unlock(arch_spinlock_t * lock) in arch_spin_unlock()
30 static inline void arch_spin_lock(arch_spinlock_t * lock) in arch_spin_lock()
50 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
57 static inline void arch_read_lock(arch_rwlock_t *lock) in arch_read_lock()
77 static inline void arch_write_lock(arch_rwlock_t *lock) in arch_write_lock()
97 static inline int arch_read_trylock(arch_rwlock_t * lock) in arch_read_trylock()
119 static inline int arch_write_trylock(arch_rwlock_t * lock) in arch_write_trylock()
141 static inline void arch_read_unlock(arch_rwlock_t * lock) in arch_read_unlock()
157 static inline void arch_write_unlock(arch_rwlock_t * lock) in arch_write_unlock()
A Dspinlock_types.h10 volatile unsigned int lock; member
16 volatile unsigned int lock; member
/arch/powerpc/include/asm/
A Dsimple_spinlock.h40 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked()
70 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
94 static inline void splpar_spin_yield(arch_spinlock_t *lock) {} in splpar_spin_yield()
95 static inline void splpar_rw_yield(arch_rwlock_t *lock) {} in splpar_rw_yield()
98 static inline void spin_yield(arch_spinlock_t *lock) in spin_yield()
106 static inline void rw_yield(arch_rwlock_t *lock) in rw_yield()
114 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
128 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
264 #define arch_spin_relax(lock) spin_yield(lock) argument
265 #define arch_read_relax(lock) rw_yield(lock) argument
[all …]
A Dqspinlock.h71 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked()
76 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked()
81 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended()
92 static __always_inline int __queued_spin_trylock_nosteal(struct qspinlock *lock) in __queued_spin_trylock_nosteal()
114 static __always_inline int __queued_spin_trylock_steal(struct qspinlock *lock) in __queued_spin_trylock_steal()
138 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock()
148 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock()
154 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
/arch/sparc/include/asm/
A Dspinlock_32.h16 #define arch_spin_is_locked(lock) (*((volatile unsigned char *)(lock)) != 0) argument
18 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
38 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
48 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
92 #define arch_read_lock(lock) \ argument
112 #define arch_read_unlock(lock) \ argument
133 static inline void arch_write_unlock(arch_rwlock_t *lock) in arch_write_unlock()
177 #define arch_read_trylock(lock) \ argument
A Dspinlock_types.h10 volatile unsigned char lock; member
20 volatile unsigned int lock; member
/arch/x86/include/asm/
A Dqspinlock.h14 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire()
44 static inline void native_queued_spin_unlock(struct qspinlock *lock) in native_queued_spin_unlock()
49 static inline void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath()
54 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
88 static inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
A Dcmpxchg.h42 #define __xchg_op(ptr, arg, op, lock) \ argument
85 #define __raw_cmpxchg(ptr, old, new, size, lock) \ argument
158 #define __raw_try_cmpxchg(_ptr, _pold, _new, size, lock) \ argument
245 #define __xadd(ptr, inc, lock) __xchg_op((ptr), (inc), xadd, lock) argument
A Drqspinlock.h24 static inline int resilient_virt_spin_lock(rqspinlock_t *lock) in resilient_virt_spin_lock()
/arch/mips/include/asm/octeon/
A Dcvmx-spinlock.h63 static inline void cvmx_spinlock_init(cvmx_spinlock_t *lock) in cvmx_spinlock_init()
74 static inline int cvmx_spinlock_locked(cvmx_spinlock_t *lock) in cvmx_spinlock_locked()
84 static inline void cvmx_spinlock_unlock(cvmx_spinlock_t *lock) in cvmx_spinlock_unlock()
103 static inline unsigned int cvmx_spinlock_trylock(cvmx_spinlock_t *lock) in cvmx_spinlock_trylock()
128 static inline void cvmx_spinlock_lock(cvmx_spinlock_t *lock) in cvmx_spinlock_lock()
/arch/arm64/kvm/hyp/include/nvhe/
A Dspinlock.h44 static inline void hyp_spin_lock(hyp_spinlock_t *lock) in hyp_spin_lock()
82 static inline void hyp_spin_unlock(hyp_spinlock_t *lock) in hyp_spin_unlock()
101 static inline bool hyp_spin_is_locked(hyp_spinlock_t *lock) in hyp_spin_is_locked()
109 static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) in hyp_assert_lock_held()
122 static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) { } in hyp_assert_lock_held()
/arch/arm/include/asm/
A Dspinlock.h56 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
81 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
107 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
114 static inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked()
119 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked()
124 static inline int arch_spin_is_contended(arch_spinlock_t *lock) in arch_spin_is_contended()
A Dmcs_spinlock.h9 #define arch_mcs_spin_lock_contended(lock) \ argument
17 #define arch_mcs_spin_unlock_contended(lock) \ argument
/arch/arc/include/asm/
A Dspinlock.h17 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
44 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
67 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
222 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
244 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
261 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
/arch/sh/include/asm/
A Dspinlock_types.h10 volatile unsigned int lock; member
16 volatile unsigned int lock; member
A Dspinlock-cas.h28 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
33 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
38 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
A Dspinlock-llsc.h26 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
46 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
61 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
/arch/powerpc/lib/
A Dqspinlock.c17 struct qspinlock *lock; member
122 static __always_inline u32 trylock_clean_tail(struct qspinlock *lock, u32 tail) in trylock_clean_tail()
161 static __always_inline u32 publish_tail_cpu(struct qspinlock *lock, u32 tail) in publish_tail_cpu()
181 static __always_inline u32 set_mustq(struct qspinlock *lock) in set_mustq()
197 static __always_inline u32 clear_mustq(struct qspinlock *lock) in clear_mustq()
213 static __always_inline bool try_set_sleepy(struct qspinlock *lock, u32 old) in try_set_sleepy()
235 static __always_inline void seen_sleepy_owner(struct qspinlock *lock, u32 val) in seen_sleepy_owner()
260 static struct qnode *get_tail_qnode(struct qspinlock *lock, int prev_cpu) in get_tail_qnode()
456 static __always_inline bool try_to_steal_lock(struct qspinlock *lock, bool paravirt) in try_to_steal_lock()
526 static __always_inline void queued_spin_lock_mcs_queue(struct qspinlock *lock, bool paravirt) in queued_spin_lock_mcs_queue()
[all …]
/arch/s390/lib/
A Dspinlock.c93 static inline int arch_load_niai4(int *lock) in arch_load_niai4()
106 static inline int arch_try_cmpxchg_niai8(int *lock, int old, int new) in arch_try_cmpxchg_niai8()
121 static inline int arch_try_cmpxchg_niai8(int *lock, int old, int new) in arch_try_cmpxchg_niai8()
136 static inline struct spin_wait *arch_spin_decode_tail(int lock) in arch_spin_decode_tail()
145 static inline int arch_spin_yield_target(int lock, struct spin_wait *node) in arch_spin_yield_target()
/arch/arm/mach-omap1/
A Dsram.S49 lock: ldrh r4, [r2], #0 @ read back dpll value label
/arch/mips/include/asm/
A Dspinlock.h21 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
/arch/loongarch/include/asm/
A Dqspinlock.h13 static inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
A Dkvm_ipi.h14 spinlock_t lock; member
20 spinlock_t lock; member

Completed in 32 milliseconds

123456