| /include/asm-generic/ |
| A D | local64.h | 30 #define local64_read(l) local_read(&(l)->a) argument 32 #define local64_inc(l) local_inc(&(l)->a) argument 33 #define local64_dec(l) local_dec(&(l)->a) argument 61 #define __local64_inc(l) local64_set((l), local64_read(l) + 1) argument 62 #define __local64_dec(l) local64_set((l), local64_read(l) - 1) argument 63 #define __local64_add(i,l) local64_set((l), local64_read(l) + (i)) argument 64 #define __local64_sub(i,l) local64_set((l), local64_read(l) - (i)) argument 100 #define __local64_inc(l) local64_set((l), local64_read(l) + 1) argument 101 #define __local64_dec(l) local64_set((l), local64_read(l) - 1) argument 102 #define __local64_add(i,l) local64_set((l), local64_read(l) + (i)) argument [all …]
|
| A D | local.h | 29 #define local_read(l) atomic_long_read(&(l)->a) argument 30 #define local_set(l,i) atomic_long_set((&(l)->a),(i)) argument 31 #define local_inc(l) atomic_long_inc(&(l)->a) argument 32 #define local_dec(l) atomic_long_dec(&(l)->a) argument 33 #define local_add(i,l) atomic_long_add((i),(&(l)->a)) argument 34 #define local_sub(i,l) atomic_long_sub((i),(&(l)->a)) argument 46 #define local_xchg(l, n) atomic_long_xchg((&(l)->a), (n)) argument 52 #define __local_inc(l) local_set((l), local_read(l) + 1) argument 53 #define __local_dec(l) local_set((l), local_read(l) - 1) argument 54 #define __local_add(i,l) local_set((l), local_read(l) + (i)) argument [all …]
|
| A D | qrwlock.h | 139 #define arch_read_lock(l) queued_read_lock(l) argument 140 #define arch_write_lock(l) queued_write_lock(l) argument 141 #define arch_read_trylock(l) queued_read_trylock(l) argument 142 #define arch_write_trylock(l) queued_write_trylock(l) argument 143 #define arch_read_unlock(l) queued_read_unlock(l) argument 144 #define arch_write_unlock(l) queued_write_unlock(l) argument 145 #define arch_rwlock_is_contended(l) queued_rwlock_is_contended(l) argument
|
| A D | qspinlock.h | 144 #define arch_spin_is_locked(l) queued_spin_is_locked(l) argument 145 #define arch_spin_is_contended(l) queued_spin_is_contended(l) argument 146 #define arch_spin_value_unlocked(l) queued_spin_value_unlocked(l) argument 147 #define arch_spin_lock(l) queued_spin_lock(l) argument 148 #define arch_spin_trylock(l) queued_spin_trylock(l) argument 149 #define arch_spin_unlock(l) queued_spin_unlock(l) argument
|
| A D | ticket_spinlock.h | 97 #define arch_spin_is_locked(l) ticket_spin_is_locked(l) argument 98 #define arch_spin_is_contended(l) ticket_spin_is_contended(l) argument 99 #define arch_spin_value_unlocked(l) ticket_spin_value_unlocked(l) argument 100 #define arch_spin_lock(l) ticket_spin_lock(l) argument 101 #define arch_spin_trylock(l) ticket_spin_trylock(l) argument 102 #define arch_spin_unlock(l) ticket_spin_unlock(l) argument
|
| /include/linux/ |
| A D | bits.h | 34 #define GENMASK_INPUT_CHECK(h, l) BUILD_BUG_ON_ZERO(const_true((l) > (h))) argument 48 (type_max(t) << (l) & \ 51 #define GENMASK(h, l) GENMASK_TYPE(unsigned long, h, l) argument 52 #define GENMASK_ULL(h, l) GENMASK_TYPE(unsigned long long, h, l) argument 54 #define GENMASK_U8(h, l) GENMASK_TYPE(u8, h, l) argument 55 #define GENMASK_U16(h, l) GENMASK_TYPE(u16, h, l) argument 56 #define GENMASK_U32(h, l) GENMASK_TYPE(u32, h, l) argument 57 #define GENMASK_U64(h, l) GENMASK_TYPE(u64, h, l) argument 58 #define GENMASK_U128(h, l) GENMASK_TYPE(u128, h, l) argument 84 #define GENMASK(h, l) __GENMASK(h, l) argument [all …]
|
| A D | lockdep.h | 304 #define lockdep_pin_lock(l) lock_pin_lock(&(l)->dep_map) argument 392 #define lockdep_assert_held(l) do { (void)(l); } while (0) argument 393 #define lockdep_assert_not_held(l) do { (void)(l); } while (0) argument 395 #define lockdep_assert_held_read(l) do { (void)(l); } while (0) argument 396 #define lockdep_assert_held_once(l) do { (void)(l); } while (0) argument 515 #define spin_release(l, i) lock_release(l, i) argument 526 #define rwlock_release(l, i) lock_release(l, i) argument 530 #define seqcount_release(l, i) lock_release(l, i) argument 534 #define mutex_release(l, i) lock_release(l, i) argument 539 #define rwsem_release(l, i) lock_release(l, i) argument [all …]
|
| A D | local_lock_internal.h | 38 lock_map_acquire(&l->dep_map); in local_lock_acquire() 39 DEBUG_LOCKS_WARN_ON(l->owner); in local_lock_acquire() 40 l->owner = current; in local_lock_acquire() 46 DEBUG_LOCKS_WARN_ON(l->owner); in local_trylock_acquire() 47 l->owner = current; in local_trylock_acquire() 53 l->owner = NULL; in local_lock_release() 54 lock_map_release(&l->dep_map); in local_lock_release() 59 l->owner = NULL; in local_lock_debug_init() 100 local_lock_t *l; \ 168 local_lock_t *l; \ [all …]
|
| A D | math64.h | 203 } l; in mul_u64_u64_shr() member 210 rl.ll = mul_u32_u32(a0.l.low, b0.l.low); in mul_u64_u64_shr() 211 rm.ll = mul_u32_u32(a0.l.low, b0.l.high); in mul_u64_u64_shr() 212 rn.ll = mul_u32_u32(a0.l.high, b0.l.low); in mul_u64_u64_shr() 213 rh.ll = mul_u32_u32(a0.l.high, b0.l.high); in mul_u64_u64_shr() 220 rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low; in mul_u64_u64_shr() 221 rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low; in mul_u64_u64_shr() 222 rh.l.high = (c >> 32) + rh.l.high; in mul_u64_u64_shr() 267 } l; in mul_u64_u32_div() member 272 rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high; in mul_u64_u32_div() [all …]
|
| A D | sort.h | 15 #define cmp_int(l, r) (((l) > (r)) - ((l) < (r))) argument
|
| A D | stdarg.h | 6 #define va_start(v, l) __builtin_va_start(v, l) argument
|
| A D | time.h | 98 #define time_between32(t, l, h) ((u32)(h) - (u32)(l) >= (u32)(t) - (u32)(l)) argument
|
| A D | lockref.h | 58 static inline bool __lockref_is_dead(const struct lockref *l) in __lockref_is_dead() argument 60 return ((int)l->count < 0); in __lockref_is_dead()
|
| A D | bitops.h | 203 static inline unsigned int fls_long(unsigned long l) in fls_long() argument 205 if (sizeof(l) == 4) in fls_long() 206 return fls(l); in fls_long() 207 return fls64(l); in fls_long() 224 static inline int get_count_order_long(unsigned long l) in get_count_order_long() argument 226 if (l == 0UL) in get_count_order_long() 228 return (int)fls_long(--l); in get_count_order_long()
|
| A D | migrate.h | 65 void putback_movable_pages(struct list_head *l); 68 int migrate_pages(struct list_head *l, new_folio_t new, free_folio_t free, 85 static inline void putback_movable_pages(struct list_head *l) {} in putback_movable_pages() argument 86 static inline int migrate_pages(struct list_head *l, new_folio_t new, in migrate_pages() argument
|
| A D | ww_mutex.h | 30 #define ww_mutex_base_init(l,n,k) __mutex_init(l,n,k) argument 34 #define ww_mutex_base_init(l,n,k) __rt_mutex_init(l,n,k) argument
|
| A D | fs_context.h | 189 #define __logfc(fc, l, fmt, ...) logfc((fc)->log.log, NULL, \ argument 190 l, fmt, ## __VA_ARGS__) 191 #define __plog(p, l, fmt, ...) logfc((p)->log, (p)->prefix, \ argument 192 l, fmt, ## __VA_ARGS__)
|
| /include/uapi/linux/ |
| A D | bits.h | 7 #define __GENMASK(h, l) (((~_UL(0)) << (l)) & (~_UL(0) >> (__BITS_PER_LONG - 1 - (h)))) argument 9 #define __GENMASK_ULL(h, l) (((~_ULL(0)) << (l)) & (~_ULL(0) >> (__BITS_PER_LONG_LONG - 1 - (h)))) argument 11 #define __GENMASK_U128(h, l) \ argument 12 ((_BIT128((h)) << 1) - (_BIT128(l)))
|
| A D | romfs_fs.h | 17 #define __mkw(h,l) (((h)&0x00ff)<< 8|((l)&0x00ff)) argument 18 #define __mkl(h,l) (((h)&0xffff)<<16|((l)&0xffff)) argument
|
| /include/crypto/ |
| A D | utils.h | 23 unsigned long l; in crypto_xor() local 26 l = get_unaligned(d) ^ get_unaligned(s++); in crypto_xor() 27 put_unaligned(l, d++); in crypto_xor() 44 unsigned long l; in crypto_xor_cpy() local 47 l = get_unaligned(s1++) ^ get_unaligned(s2++); in crypto_xor_cpy() 48 put_unaligned(l, d++); in crypto_xor_cpy()
|
| /include/linux/ceph/ |
| A D | striper.h | 10 void ceph_calc_file_object_mapping(struct ceph_file_layout *l, 36 int ceph_file_to_extents(struct ceph_file_layout *l, u64 off, u64 len, 42 int ceph_iterate_extents(struct ceph_file_layout *l, u64 off, u64 len, 64 int ceph_extent_to_file(struct ceph_file_layout *l, 69 u64 ceph_get_num_objects(struct ceph_file_layout *l, u64 size);
|
| /include/kvm/ |
| A D | iodev.h | 44 int l, void *v) in kvm_iodevice_read() argument 46 return dev->ops->read ? dev->ops->read(vcpu, dev, addr, l, v) in kvm_iodevice_read() 52 int l, const void *v) in kvm_iodevice_write() argument 54 return dev->ops->write ? dev->ops->write(vcpu, dev, addr, l, v) in kvm_iodevice_write()
|
| /include/asm-generic/bitops/ |
| A D | ext2-atomic-setbit.h | 9 #define ext2_set_bit_atomic(l, nr, addr) test_and_set_bit_le(nr, addr) argument 10 #define ext2_clear_bit_atomic(l, nr, addr) test_and_clear_bit_le(nr, addr) argument
|
| /include/media/i2c/ |
| A D | ov772x.h | 35 #define OV772X_AUTO_EDGECTRL(u, l) \ argument 38 .lower = (l & OV772X_EDGE_LOWER_MASK), \
|
| /include/video/ |
| A D | pm3fb.h | 591 #define PM3FBDestReadMode_Layout0(l) (((l) & 0x3) << 12) argument 592 #define PM3FBDestReadMode_Layout1(l) (((l) & 0x3) << 14) argument 593 #define PM3FBDestReadMode_Layout2(l) (((l) & 0x3) << 16) argument 594 #define PM3FBDestReadMode_Layout3(l) (((l) & 0x3) << 18) argument 620 #define PM3FBSourceReadMode_Layout(l) (((l) & 0x3) << 8) argument 660 #define PM3FBWriteMode_Layout0(l) (((l) & 0x3) << 16) argument 661 #define PM3FBWriteMode_Layout1(l) (((l) & 0x3) << 18) argument 662 #define PM3FBWriteMode_Layout2(l) (((l) & 0x3) << 20) argument 663 #define PM3FBWriteMode_Layout3(l) (((l) & 0x3) << 22) argument
|