Lines Matching refs:op

49 #define ATOMIC_OP(op, asm_op, suffix, sign, ...)			\  argument
50 static __inline__ void arch_atomic_##op(int a, atomic_t *v) \
55 "1: lwarx %0,0,%3 # atomic_" #op "\n" \
64 #define ATOMIC_OP_RETURN_RELAXED(op, asm_op, suffix, sign, ...) \ argument
65 static inline int arch_atomic_##op##_return_relaxed(int a, atomic_t *v) \
70 "1: lwarx %0,0,%3 # atomic_" #op "_return_relaxed\n" \
81 #define ATOMIC_FETCH_OP_RELAXED(op, asm_op, suffix, sign, ...) \ argument
82 static inline int arch_atomic_fetch_##op##_relaxed(int a, atomic_t *v) \
87 "1: lwarx %0,0,%4 # atomic_fetch_" #op "_relaxed\n" \
98 #define ATOMIC_OPS(op, asm_op, suffix, sign, ...) \ argument
99 ATOMIC_OP(op, asm_op, suffix, sign, ##__VA_ARGS__) \
100 ATOMIC_OP_RETURN_RELAXED(op, asm_op, suffix, sign, ##__VA_ARGS__)\
101 ATOMIC_FETCH_OP_RELAXED(op, asm_op, suffix, sign, ##__VA_ARGS__)
113 #define ATOMIC_OPS(op, asm_op, suffix, sign) \ argument
114 ATOMIC_OP(op, asm_op, suffix, sign) \
115 ATOMIC_FETCH_OP_RELAXED(op, asm_op, suffix, sign)
215 #define ATOMIC64_OP(op, asm_op) \ argument
216 static __inline__ void arch_atomic64_##op(s64 a, atomic64_t *v) \
221 "1: ldarx %0,0,%3 # atomic64_" #op "\n" \
230 #define ATOMIC64_OP_RETURN_RELAXED(op, asm_op) \ argument
232 arch_atomic64_##op##_return_relaxed(s64 a, atomic64_t *v) \
237 "1: ldarx %0,0,%3 # atomic64_" #op "_return_relaxed\n" \
248 #define ATOMIC64_FETCH_OP_RELAXED(op, asm_op) \ argument
250 arch_atomic64_fetch_##op##_relaxed(s64 a, atomic64_t *v) \
255 "1: ldarx %0,0,%4 # atomic64_fetch_" #op "_relaxed\n" \
266 #define ATOMIC64_OPS(op, asm_op) \ argument
267 ATOMIC64_OP(op, asm_op) \
268 ATOMIC64_OP_RETURN_RELAXED(op, asm_op) \
269 ATOMIC64_FETCH_OP_RELAXED(op, asm_op)
281 #define ATOMIC64_OPS(op, asm_op) \ in ATOMIC64_OPS() argument
282 ATOMIC64_OP(op, asm_op) \ in ATOMIC64_OPS()
283 ATOMIC64_FETCH_OP_RELAXED(op, asm_op) in ATOMIC64_OPS()