/include/linux/ |
A D | kmemleak.h | 24 extern void kmemleak_free(const void *ptr) __ref; 28 extern void kmemleak_not_leak(const void *ptr) __ref; 30 extern void kmemleak_ignore(const void *ptr) __ref; 33 extern void kmemleak_no_scan(const void *ptr) __ref; 44 kmemleak_alloc(ptr, size, min_count, gfp); in kmemleak_alloc_recursive() 50 kmemleak_free(ptr); in kmemleak_free_recursive() 53 static inline void kmemleak_erase(void **ptr) in kmemleak_erase() argument 55 *ptr = NULL; in kmemleak_erase() 80 static inline void kmemleak_free(const void *ptr) in kmemleak_free() argument 104 static inline void kmemleak_ignore(const void *ptr) in kmemleak_ignore() argument [all …]
|
A D | err.h | 55 static inline long __must_check PTR_ERR(__force const void *ptr) in PTR_ERR() argument 57 return (long) ptr; in PTR_ERR() 61 #define PTR_ERR_PCPU(ptr) (PTR_ERR((const void *)(__force const unsigned long)(ptr))) argument 68 static inline bool __must_check IS_ERR(__force const void *ptr) in IS_ERR() argument 70 return IS_ERR_VALUE((unsigned long)ptr); in IS_ERR() 74 #define IS_ERR_PCPU(ptr) (IS_ERR((const void *)(__force const unsigned long)(ptr))) argument 84 return unlikely(!ptr) || IS_ERR_VALUE((unsigned long)ptr); in IS_ERR_OR_NULL() 94 static inline void * __must_check ERR_CAST(__force const void *ptr) in ERR_CAST() argument 97 return (void *) ptr; in ERR_CAST() 119 if (IS_ERR(ptr)) in PTR_ERR_OR_ZERO() [all …]
|
A D | container_of.h | 19 #define container_of(ptr, type, member) ({ \ argument 20 void *__mptr = (void *)(ptr); \ 21 static_assert(__same_type(*(ptr), ((type *)0)->member) || \ 22 __same_type(*(ptr), void), \ 35 #define container_of_const(ptr, type, member) \ argument 36 _Generic(ptr, \ 37 const typeof(*(ptr)) *: ((const type *)container_of(ptr, type, member)),\ 38 default: ((type *)container_of(ptr, type, member)) \
|
A D | if_tun.h | 19 void *ptr; member 26 static inline bool tun_is_xdp_frame(void *ptr) in tun_is_xdp_frame() argument 28 return (unsigned long)ptr & TUN_XDP_FLAG; in tun_is_xdp_frame() 36 static inline struct xdp_frame *tun_ptr_to_xdp(void *ptr) in tun_ptr_to_xdp() argument 38 return (void *)((unsigned long)ptr & ~TUN_XDP_FLAG); in tun_ptr_to_xdp() 41 void tun_ptr_free(void *ptr); 58 static inline bool tun_is_xdp_frame(void *ptr) in tun_is_xdp_frame() argument 68 static inline struct xdp_frame *tun_ptr_to_xdp(void *ptr) in tun_ptr_to_xdp() argument 73 static inline void tun_ptr_free(void *ptr) in tun_ptr_free() argument
|
A D | kcsan-checks.h | 37 void __kcsan_check_access(const volatile void *ptr, size_t size, int type); 142 const volatile void *ptr; member 173 kcsan_begin_scoped_access(const volatile void *ptr, size_t size, int type, 209 kcsan_begin_scoped_access(const volatile void *ptr, size_t size, int type, in kcsan_begin_scoped_access() argument 276 #define __kcsan_check_read(ptr, size) __kcsan_check_access(ptr, size, 0) argument 285 __kcsan_check_access(ptr, size, KCSAN_ACCESS_WRITE) 302 #define kcsan_check_read(ptr, size) kcsan_check_access(ptr, size, 0) argument 311 kcsan_check_access(ptr, size, KCSAN_ACCESS_WRITE) 320 kcsan_check_access(ptr, size, KCSAN_ACCESS_COMPOUND | KCSAN_ACCESS_WRITE) 332 kcsan_check_access(ptr, size, KCSAN_ACCESS_ATOMIC) [all …]
|
A D | list_nulls.h | 29 #define INIT_HLIST_NULLS_HEAD(ptr, nulls) \ argument 30 ((ptr)->first = (struct hlist_nulls_node *) NULLS_MARKER(nulls)) 33 #define hlist_nulls_entry(ptr, type, member) container_of(ptr,type,member) argument 35 #define hlist_nulls_entry_safe(ptr, type, member) \ argument 36 ({ typeof(ptr) ____ptr = (ptr); \ 44 static inline int is_a_nulls(const struct hlist_nulls_node *ptr) in is_a_nulls() argument 46 return ((unsigned long)ptr & 1); in is_a_nulls() 55 static inline unsigned long get_nulls_value(const struct hlist_nulls_node *ptr) in get_nulls_value() argument 57 return ((unsigned long)ptr) >> 1; in get_nulls_value()
|
A D | adfs_fs.h | 12 static inline int adfs_checkbblk(unsigned char *ptr) in adfs_checkbblk() argument 15 unsigned char *p = ptr + 511; in adfs_checkbblk() 20 } while (p != ptr); in adfs_checkbblk() 22 return (result & 0xff) != ptr[511]; in adfs_checkbblk()
|
A D | ptr_ring.h | 293 void *ptr; in __ptr_ring_consume() local 300 if (ptr) in __ptr_ring_consume() 303 return ptr; in __ptr_ring_consume() 309 void *ptr; in __ptr_ring_consume_batched() local 314 if (!ptr) in __ptr_ring_consume_batched() 329 void *ptr; in ptr_ring_consume() local 340 void *ptr; in ptr_ring_consume_irq() local 352 void *ptr; in ptr_ring_consume_any() local 363 void *ptr; in ptr_ring_consume_bh() local 565 void *ptr; in __ptr_ring_swap_queue() local [all …]
|
A D | kallsyms.h | 49 static inline void *dereference_symbol_descriptor(void *ptr) in dereference_symbol_descriptor() argument 54 ptr = dereference_kernel_function_descriptor(ptr); in dereference_symbol_descriptor() 55 if (is_ksym_addr((unsigned long)ptr)) in dereference_symbol_descriptor() 56 return ptr; in dereference_symbol_descriptor() 59 mod = __module_address((unsigned long)ptr); in dereference_symbol_descriptor() 62 ptr = dereference_module_function_descriptor(mod, ptr); in dereference_symbol_descriptor() 64 return ptr; in dereference_symbol_descriptor()
|
A D | bpf_mem_alloc.h | 41 void bpf_mem_free(struct bpf_mem_alloc *ma, void *ptr); 42 void bpf_mem_free_rcu(struct bpf_mem_alloc *ma, void *ptr); 46 void bpf_mem_cache_free(struct bpf_mem_alloc *ma, void *ptr); 47 void bpf_mem_cache_free_rcu(struct bpf_mem_alloc *ma, void *ptr); 48 void bpf_mem_cache_raw_free(void *ptr);
|
/include/asm-generic/ |
A D | cmpxchg.h | 31 return __xchg_u8(x, ptr); in __generic_xchg() 34 ret = *(volatile u8 *)ptr; in __generic_xchg() 42 return __xchg_u16(x, ptr); in __generic_xchg() 45 ret = *(volatile u16 *)ptr; in __generic_xchg() 53 return __xchg_u32(x, ptr); in __generic_xchg() 56 ret = *(volatile u32 *)ptr; in __generic_xchg() 65 return __xchg_u64(x, ptr); in __generic_xchg() 68 ret = *(volatile u64 *)ptr; in __generic_xchg() 69 *(volatile u64 *)ptr = x; in __generic_xchg() 83 __generic_xchg((unsigned long)(x), (ptr), sizeof(*(ptr)))); \ [all …]
|
A D | cmpxchg-local.h | 24 wrong_size_cmpxchg(ptr); in __generic_cmpxchg_local() 28 case 1: prev = *(u8 *)ptr; in __generic_cmpxchg_local() 30 *(u8 *)ptr = (new & 0xffu); in __generic_cmpxchg_local() 32 case 2: prev = *(u16 *)ptr; in __generic_cmpxchg_local() 34 *(u16 *)ptr = (new & 0xffffu); in __generic_cmpxchg_local() 36 case 4: prev = *(u32 *)ptr; in __generic_cmpxchg_local() 40 case 8: prev = *(u64 *)ptr; in __generic_cmpxchg_local() 42 *(u64 *)ptr = (u64)new; in __generic_cmpxchg_local() 45 wrong_size_cmpxchg(ptr); in __generic_cmpxchg_local() 61 prev = *(u64 *)ptr; in __generic_cmpxchg64_local() [all …]
|
A D | uaccess.h | 104 #define __put_user(x, ptr) \ argument 115 ptr, &__x); \ 124 #define put_user(x, ptr) \ argument 126 void __user *__p = (ptr); \ 149 __chk_user_ptr(ptr); \ 150 switch (sizeof(*(ptr))) { \ 154 ptr, &__x); \ 161 ptr, &__x); \ 168 ptr, &__x); \ 175 ptr, &__x); \ [all …]
|
/include/rdma/ |
A D | iba.h | 14 return *ptr; in _iba_get8() 19 *ptr = (*ptr & ~mask) | prep_value; in _iba_set8() 24 return be16_to_cpu(*ptr); in _iba_get16() 29 *ptr = cpu_to_be16((be16_to_cpu(*ptr) & ~mask) | prep_value); in _iba_set16() 34 return be32_to_cpu(*ptr); in _iba_get32() 39 *ptr = cpu_to_be32((be32_to_cpu(*ptr) & ~mask) | prep_value); in _iba_set32() 53 put_unaligned(cpu_to_be64((_iba_get64(ptr) & ~mask) | prep_value), ptr); in _iba_set64() 62 #define IBA_SET(field, ptr, value) _IBA_SET(field, ptr, value) argument 69 #define IBA_GET_MEM_PTR(field, ptr) _IBA_GET_MEM_PTR(field, ptr) argument 82 #define IBA_SET_MEM(field, ptr, in, bytes) _IBA_SET_MEM(field, ptr, in, bytes) argument [all …]
|
/include/linux/unaligned/ |
A D | packed_struct.h | 12 const struct __una_u16 *ptr = (const struct __una_u16 *)p; in __get_unaligned_cpu16() local 13 return ptr->x; in __get_unaligned_cpu16() 18 const struct __una_u32 *ptr = (const struct __una_u32 *)p; in __get_unaligned_cpu32() local 19 return ptr->x; in __get_unaligned_cpu32() 25 return ptr->x; in __get_unaligned_cpu64() 30 struct __una_u16 *ptr = (struct __una_u16 *)p; in __put_unaligned_cpu16() local 31 ptr->x = val; in __put_unaligned_cpu16() 36 struct __una_u32 *ptr = (struct __una_u32 *)p; in __put_unaligned_cpu32() local 37 ptr->x = val; in __put_unaligned_cpu32() 42 struct __una_u64 *ptr = (struct __una_u64 *)p; in __put_unaligned_cpu64() local [all …]
|
/include/linux/sunrpc/xdrgen/ |
A D | _builtins.h | 34 *ptr = (*p != xdr_zero); in xdrgen_decode_bool() 56 *ptr = be32_to_cpup(p); in xdrgen_decode_int() 78 *ptr = be32_to_cpup(p); in xdrgen_decode_unsigned_int() 100 *ptr = be32_to_cpup(p); in xdrgen_decode_long() 122 *ptr = be32_to_cpup(p); in xdrgen_decode_unsigned_long() 144 *ptr = get_unaligned_be64(p); in xdrgen_decode_hyper() 166 *ptr = get_unaligned_be64(p); in xdrgen_decode_unsigned_hyper() 195 ptr->data = (unsigned char *)p; in xdrgen_decode_string() 197 ptr->len = len; in xdrgen_decode_string() 226 ptr->data = (u8 *)p; in xdrgen_decode_opaque() [all …]
|
/include/net/netns/ |
A D | generic.h | 36 DECLARE_FLEX_ARRAY(void *, ptr); 43 void *ptr; in net_generic() local 47 ptr = ng->ptr[id]; in net_generic() 50 return ptr; in net_generic()
|
/include/linux/mlx4/ |
A D | doorbell.h | 52 #define MLX4_INIT_DOORBELL_LOCK(ptr) do { } while (0) argument 53 #define MLX4_GET_DOORBELL_LOCK(ptr) (NULL) argument 70 #define MLX4_INIT_DOORBELL_LOCK(ptr) spin_lock_init(ptr) argument 71 #define MLX4_GET_DOORBELL_LOCK(ptr) (ptr) argument
|
/include/net/ |
A D | snmp.h | 146 __typeof__((mib->mibs) + 0) ptr = mib->mibs; \ 147 this_cpu_inc(ptr[basefield##PKTS]); \ 152 __typeof__((mib->mibs) + 0) ptr = mib->mibs; \ 153 __this_cpu_inc(ptr[basefield##PKTS]); \ 163 u64_stats_update_begin(&ptr->syncp); \ 164 ptr->mibs[field] += addend; \ 165 u64_stats_update_end(&ptr->syncp); \ 180 u64_stats_update_begin(&ptr->syncp); \ 181 ptr->mibs[basefield##PKTS]++; \ 182 ptr->mibs[basefield##OCTETS] += addend; \ [all …]
|
/include/trace/events/ |
A D | percpu.h | 16 void __percpu *ptr, size_t bytes_alloc, gfp_t gfp_flags), 19 ptr, bytes_alloc, gfp_flags), 29 __field( void __percpu *, ptr ) 41 __entry->ptr = ptr; 50 __entry->base_addr, __entry->off, __entry->ptr, 56 TP_PROTO(void *base_addr, int off, void __percpu *ptr), 58 TP_ARGS(base_addr, off, ptr), 63 __field( void __percpu *, ptr ) 69 __entry->ptr = ptr; 73 __entry->base_addr, __entry->off, __entry->ptr)
|
A D | tegra_apb_dma.h | 28 TP_PROTO(struct dma_chan *dc, int count, void *ptr), 29 TP_ARGS(dc, count, ptr), 33 __field(void *, ptr) 38 __entry->ptr = ptr; 41 __get_str(chan), __entry->count, __entry->ptr)
|
A D | kmem.h | 15 const void *ptr, 24 __field( const void *, ptr ) 34 __entry->ptr = ptr; 46 __entry->ptr, 57 const void *ptr, 76 __entry->ptr = ptr; 85 __entry->ptr, 98 TP_ARGS(call_site, ptr), 107 __entry->ptr = ptr; 118 TP_ARGS(call_site, ptr, s), [all …]
|
/include/vdso/ |
A D | unaligned.h | 5 #define __get_unaligned_t(type, ptr) ({ \ argument 6 const struct { type x; } __packed * __get_pptr = (typeof(__get_pptr))(ptr); \ 10 #define __put_unaligned_t(type, val, ptr) do { \ argument 11 struct { type x; } __packed * __put_pptr = (typeof(__put_pptr))(ptr); \
|
/include/linux/soc/marvell/octeontx2/ |
A D | asm.h | 39 static inline u64 otx2_atomic64_fetch_add(u64 incr, u64 *ptr) in otx2_atomic64_fetch_add() argument 45 : [r] "=r" (result), "+m" (*ptr) in otx2_atomic64_fetch_add() 46 : [i] "r" (incr), [b] "r" (ptr) in otx2_atomic64_fetch_add() 54 #define otx2_atomic64_fetch_add(incr, ptr) ({ incr; }) argument
|
/include/linux/atomic/ |
A D | atomic-instrumented.h | 4755 typeof(ptr) __ai_ptr = (ptr); \ 4763 typeof(ptr) __ai_ptr = (ptr); \ 4770 typeof(ptr) __ai_ptr = (ptr); \ 4778 typeof(ptr) __ai_ptr = (ptr); \ 4785 typeof(ptr) __ai_ptr = (ptr); \ 4793 typeof(ptr) __ai_ptr = (ptr); \ 4800 typeof(ptr) __ai_ptr = (ptr); \ 4808 typeof(ptr) __ai_ptr = (ptr); \ 4815 typeof(ptr) __ai_ptr = (ptr); \ 4823 typeof(ptr) __ai_ptr = (ptr); \ [all …]
|