Searched refs:__force (Results 1 – 11 of 11) sorted by relevance
16 #define __constant_htonl(x) ((__force __be32)(__u32)(x))17 #define __constant_ntohl(x) ((__force __u32)(__be32)(x))47 return (__force __le64)__swab64p(p); in __cpu_to_le64p()55 return (__force __le32)__swab32p(p); in __cpu_to_le32p()63 return (__force __le16)__swab16p(p); in __cpu_to_le16p()71 return (__force __be64)*p; in __cpu_to_be64p()75 return (__force __u64)*p; in __be64_to_cpup()79 return (__force __be32)*p; in __cpu_to_be32p()83 return (__force __u32)*p; in __be32_to_cpup()87 return (__force __be16)*p; in __cpu_to_be16p()[all …]
32 #define __cpu_to_le64(x) ((__force __le64)(__u64)(x))33 #define __le64_to_cpu(x) ((__force __u64)(__le64)(x))47 return (__force __le64)*p; in __cpu_to_le64p()51 return (__force __u64)*p; in __le64_to_cpup()55 return (__force __le32)*p; in __cpu_to_le32p()59 return (__force __u32)*p; in __le32_to_cpup()63 return (__force __le16)*p; in __cpu_to_le16p()67 return (__force __u16)*p; in __le16_to_cpup()71 return (__force __be64)__swab64p(p); in __cpu_to_be64p()79 return (__force __be32)__swab32p(p); in __cpu_to_be32p()[all …]
10 #define __get_unaligned_le(ptr) ((__force typeof(*(ptr)))({ \18 #define __get_unaligned_be(ptr) ((__force typeof(*(ptr)))({ \30 *(u8 *)__gu_p = (__force u8)(val); \33 put_unaligned_le16((__force u16)(val), __gu_p); \36 put_unaligned_le32((__force u32)(val), __gu_p); \39 put_unaligned_le64((__force u64)(val), __gu_p); \51 *(u8 *)__gu_p = (__force u8)(val); \54 put_unaligned_be16((__force u16)(val), __gu_p); \57 put_unaligned_be32((__force u32)(val), __gu_p); \60 put_unaligned_be64((__force u64)(val), __gu_p); \
378 return le16_to_cpu((__force __le16)val); in __virtio16_to_cpu()380 return be16_to_cpu((__force __be16)val); in __virtio16_to_cpu()386 return (__force __virtio16)cpu_to_le16(val); in __cpu_to_virtio16()388 return (__force __virtio16)cpu_to_be16(val); in __cpu_to_virtio16()394 return le32_to_cpu((__force __le32)val); in __virtio32_to_cpu()396 return be32_to_cpu((__force __be32)val); in __virtio32_to_cpu()402 return (__force __virtio32)cpu_to_le32(val); in __cpu_to_virtio32()404 return (__force __virtio32)cpu_to_be32(val); in __cpu_to_virtio32()410 return le64_to_cpu((__force __le64)val); in __virtio64_to_cpu()412 return be64_to_cpu((__force __be64)val); in __virtio64_to_cpu()[all …]
120 return *(const volatile u8 __force *)addr; in __raw_readb()128 return *(const volatile u16 __force *)addr; in __raw_readw()136 return *(const volatile u32 __force *)addr; in __raw_readl()145 return *(const volatile u64 __force *)addr; in __raw_readq()154 *(volatile u8 __force *)addr = value; in __raw_writeb()162 *(volatile u16 __force *)addr = value; in __raw_writew()170 *(volatile u32 __force *)addr = value; in __raw_writel()179 *(volatile u64 __force *)addr = value; in __raw_writeq()
17 writel((u32 __force)cpu_to_le32(val), addr); in _write_litex_subregister()22 return le32_to_cpu((__le32 __force)readl(addr)); in _read_litex_subregister()
50 static inline void * __must_check ERR_CAST(__force const void *ptr) in ERR_CAST()
29 #define BUILD_BUG_ON_INVALID(e) ((void)(sizeof((__force long)(e))))
11 # define __force __attribute__((force)) macro25 # define ACCESS_PRIVATE(p, member) (*((typeof((p)->member) __force *) &(p)->member))34 # define __force macro
283 { .__val = (__force typeof(x)) (val) }; \
34 #define __GFP_ZERO ((__force gfp_t)0x8000u) /* Return zeroed page on success */
Completed in 27 milliseconds