/include/scsi/ |
D | scsi_devinfo.h | 9 #define BLIST_NOLUN ((__force blist_flags_t)(1ULL << 0)) 12 #define BLIST_FORCELUN ((__force blist_flags_t)(1ULL << 1)) 14 #define BLIST_BORKEN ((__force blist_flags_t)(1ULL << 2)) 16 #define BLIST_KEY ((__force blist_flags_t)(1ULL << 3)) 18 #define BLIST_SINGLELUN ((__force blist_flags_t)(1ULL << 4)) 20 #define BLIST_NOTQ ((__force blist_flags_t)(1ULL << 5)) 22 #define BLIST_SPARSELUN ((__force blist_flags_t)(1ULL << 6)) 24 #define BLIST_MAX5LUN ((__force blist_flags_t)(1ULL << 7)) 26 #define BLIST_ISROM ((__force blist_flags_t)(1ULL << 8)) 28 #define BLIST_LARGELUN ((__force blist_flags_t)(1ULL << 9)) [all …]
|
/include/uapi/linux/ |
D | eventpoll.h | 31 #define EPOLLIN (__force __poll_t)0x00000001 32 #define EPOLLPRI (__force __poll_t)0x00000002 33 #define EPOLLOUT (__force __poll_t)0x00000004 34 #define EPOLLERR (__force __poll_t)0x00000008 35 #define EPOLLHUP (__force __poll_t)0x00000010 36 #define EPOLLNVAL (__force __poll_t)0x00000020 37 #define EPOLLRDNORM (__force __poll_t)0x00000040 38 #define EPOLLRDBAND (__force __poll_t)0x00000080 39 #define EPOLLWRNORM (__force __poll_t)0x00000100 40 #define EPOLLWRBAND (__force __poll_t)0x00000200 [all …]
|
/include/linux/ |
D | gfp_types.h | 83 #define __GFP_DMA ((__force gfp_t)___GFP_DMA) 84 #define __GFP_HIGHMEM ((__force gfp_t)___GFP_HIGHMEM) 85 #define __GFP_DMA32 ((__force gfp_t)___GFP_DMA32) 86 #define __GFP_MOVABLE ((__force gfp_t)___GFP_MOVABLE) /* ZONE_MOVABLE allowed */ 87 #define __GFP_CMA ((__force gfp_t)___GFP_CMA) 117 #define __GFP_RECLAIMABLE ((__force gfp_t)___GFP_RECLAIMABLE) 118 #define __GFP_WRITE ((__force gfp_t)___GFP_WRITE) 119 #define __GFP_HARDWALL ((__force gfp_t)___GFP_HARDWALL) 120 #define __GFP_THISNODE ((__force gfp_t)___GFP_THISNODE) 121 #define __GFP_ACCOUNT ((__force gfp_t)___GFP_ACCOUNT) [all …]
|
D | blk_types.h | 103 #define BLK_STS_NOTSUPP ((__force blk_status_t)1) 104 #define BLK_STS_TIMEOUT ((__force blk_status_t)2) 105 #define BLK_STS_NOSPC ((__force blk_status_t)3) 106 #define BLK_STS_TRANSPORT ((__force blk_status_t)4) 107 #define BLK_STS_TARGET ((__force blk_status_t)5) 108 #define BLK_STS_RESV_CONFLICT ((__force blk_status_t)6) 109 #define BLK_STS_MEDIUM ((__force blk_status_t)7) 110 #define BLK_STS_PROTECTION ((__force blk_status_t)8) 111 #define BLK_STS_RESOURCE ((__force blk_status_t)9) 112 #define BLK_STS_IOERR ((__force blk_status_t)10) [all …]
|
D | virtio_byteorder.h | 19 return le16_to_cpu((__force __le16)val); in __virtio16_to_cpu() 21 return be16_to_cpu((__force __be16)val); in __virtio16_to_cpu() 27 return (__force __virtio16)cpu_to_le16(val); in __cpu_to_virtio16() 29 return (__force __virtio16)cpu_to_be16(val); in __cpu_to_virtio16() 35 return le32_to_cpu((__force __le32)val); in __virtio32_to_cpu() 37 return be32_to_cpu((__force __be32)val); in __virtio32_to_cpu() 43 return (__force __virtio32)cpu_to_le32(val); in __cpu_to_virtio32() 45 return (__force __virtio32)cpu_to_be32(val); in __cpu_to_virtio32() 51 return le64_to_cpu((__force __le64)val); in __virtio64_to_cpu() 53 return be64_to_cpu((__force __be64)val); in __virtio64_to_cpu() [all …]
|
D | err.h | 49 static inline long __must_check PTR_ERR(__force const void *ptr) in PTR_ERR() 59 static inline bool __must_check IS_ERR(__force const void *ptr) in IS_ERR() 70 static inline bool __must_check IS_ERR_OR_NULL(__force const void *ptr) in IS_ERR_OR_NULL() 82 static inline void * __must_check ERR_CAST(__force const void *ptr) in ERR_CAST() 105 static inline int __must_check PTR_ERR_OR_ZERO(__force const void *ptr) in PTR_ERR_OR_ZERO()
|
D | serial_core.h | 509 #define UPF_FOURPORT ((__force upf_t) ASYNC_FOURPORT /* 1 */ ) 510 #define UPF_SAK ((__force upf_t) ASYNC_SAK /* 2 */ ) 511 #define UPF_SPD_HI ((__force upf_t) ASYNC_SPD_HI /* 4 */ ) 512 #define UPF_SPD_VHI ((__force upf_t) ASYNC_SPD_VHI /* 5 */ ) 513 #define UPF_SPD_CUST ((__force upf_t) ASYNC_SPD_CUST /* 0x0030 */ ) 514 #define UPF_SPD_WARP ((__force upf_t) ASYNC_SPD_WARP /* 0x1010 */ ) 515 #define UPF_SPD_MASK ((__force upf_t) ASYNC_SPD_MASK /* 0x1030 */ ) 516 #define UPF_SKIP_TEST ((__force upf_t) ASYNC_SKIP_TEST /* 6 */ ) 517 #define UPF_AUTO_IRQ ((__force upf_t) ASYNC_AUTO_IRQ /* 7 */ ) 518 #define UPF_HARDPPS_CD ((__force upf_t) ASYNC_HARDPPS_CD /* 11 */ ) [all …]
|
D | slab.h | 30 #define SLAB_CONSISTENCY_CHECKS ((slab_flags_t __force)0x00000100U) 32 #define SLAB_RED_ZONE ((slab_flags_t __force)0x00000400U) 34 #define SLAB_POISON ((slab_flags_t __force)0x00000800U) 36 #define SLAB_KMALLOC ((slab_flags_t __force)0x00001000U) 38 #define SLAB_HWCACHE_ALIGN ((slab_flags_t __force)0x00002000U) 40 #define SLAB_CACHE_DMA ((slab_flags_t __force)0x00004000U) 42 #define SLAB_CACHE_DMA32 ((slab_flags_t __force)0x00008000U) 44 #define SLAB_STORE_USER ((slab_flags_t __force)0x00010000U) 46 #define SLAB_PANIC ((slab_flags_t __force)0x00040000U) 98 #define SLAB_TYPESAFE_BY_RCU ((slab_flags_t __force)0x00080000U) [all …]
|
D | omap-mailbox.h | 14 #define IRQ_TX ((__force omap_mbox_irq_t) 1) 15 #define IRQ_RX ((__force omap_mbox_irq_t) 2)
|
D | mm_types.h | 1192 VM_FAULT_OOM = (__force vm_fault_t)0x000001, 1193 VM_FAULT_SIGBUS = (__force vm_fault_t)0x000002, 1194 VM_FAULT_MAJOR = (__force vm_fault_t)0x000004, 1195 VM_FAULT_HWPOISON = (__force vm_fault_t)0x000010, 1196 VM_FAULT_HWPOISON_LARGE = (__force vm_fault_t)0x000020, 1197 VM_FAULT_SIGSEGV = (__force vm_fault_t)0x000040, 1198 VM_FAULT_NOPAGE = (__force vm_fault_t)0x000100, 1199 VM_FAULT_LOCKED = (__force vm_fault_t)0x000200, 1200 VM_FAULT_RETRY = (__force vm_fault_t)0x000400, 1201 VM_FAULT_FALLBACK = (__force vm_fault_t)0x000800, [all …]
|
D | blk-mq.h | 32 #define RQF_STARTED ((__force req_flags_t)(1 << 1)) 34 #define RQF_FLUSH_SEQ ((__force req_flags_t)(1 << 4)) 36 #define RQF_MIXED_MERGE ((__force req_flags_t)(1 << 5)) 38 #define RQF_MQ_INFLIGHT ((__force req_flags_t)(1 << 6)) 40 #define RQF_DONTPREP ((__force req_flags_t)(1 << 7)) 42 #define RQF_SCHED_TAGS ((__force req_flags_t)(1 << 8)) 44 #define RQF_USE_SCHED ((__force req_flags_t)(1 << 9)) 46 #define RQF_FAILED ((__force req_flags_t)(1 << 10)) 48 #define RQF_QUIET ((__force req_flags_t)(1 << 11)) 50 #define RQF_IO_STAT ((__force req_flags_t)(1 << 13)) [all …]
|
/include/uapi/linux/byteorder/ |
D | little_endian.h | 16 #define __constant_htonl(x) ((__force __be32)___constant_swab32((x))) 17 #define __constant_ntohl(x) ___constant_swab32((__force __be32)(x)) 18 #define __constant_htons(x) ((__force __be16)___constant_swab16((x))) 19 #define __constant_ntohs(x) ___constant_swab16((__force __be16)(x)) 20 #define __constant_cpu_to_le64(x) ((__force __le64)(__u64)(x)) 21 #define __constant_le64_to_cpu(x) ((__force __u64)(__le64)(x)) 22 #define __constant_cpu_to_le32(x) ((__force __le32)(__u32)(x)) 23 #define __constant_le32_to_cpu(x) ((__force __u32)(__le32)(x)) 24 #define __constant_cpu_to_le16(x) ((__force __le16)(__u16)(x)) 25 #define __constant_le16_to_cpu(x) ((__force __u16)(__le16)(x)) [all …]
|
D | big_endian.h | 16 #define __constant_htonl(x) ((__force __be32)(__u32)(x)) 17 #define __constant_ntohl(x) ((__force __u32)(__be32)(x)) 18 #define __constant_htons(x) ((__force __be16)(__u16)(x)) 19 #define __constant_ntohs(x) ((__force __u16)(__be16)(x)) 20 #define __constant_cpu_to_le64(x) ((__force __le64)___constant_swab64((x))) 21 #define __constant_le64_to_cpu(x) ___constant_swab64((__force __u64)(__le64)(x)) 22 #define __constant_cpu_to_le32(x) ((__force __le32)___constant_swab32((x))) 23 #define __constant_le32_to_cpu(x) ___constant_swab32((__force __u32)(__le32)(x)) 24 #define __constant_cpu_to_le16(x) ((__force __le16)___constant_swab16((x))) 25 #define __constant_le16_to_cpu(x) ___constant_swab16((__force __u16)(__le16)(x)) [all …]
|
/include/linux/rpmsg/ |
D | byteorder.h | 22 return le16_to_cpu((__force __le16)val); in __rpmsg16_to_cpu() 24 return be16_to_cpu((__force __be16)val); in __rpmsg16_to_cpu() 30 return (__force __rpmsg16)cpu_to_le16(val); in __cpu_to_rpmsg16() 32 return (__force __rpmsg16)cpu_to_be16(val); in __cpu_to_rpmsg16() 38 return le32_to_cpu((__force __le32)val); in __rpmsg32_to_cpu() 40 return be32_to_cpu((__force __be32)val); in __rpmsg32_to_cpu() 46 return (__force __rpmsg32)cpu_to_le32(val); in __cpu_to_rpmsg32() 48 return (__force __rpmsg32)cpu_to_be32(val); in __cpu_to_rpmsg32() 54 return le64_to_cpu((__force __le64)val); in __rpmsg64_to_cpu() 56 return be64_to_cpu((__force __be64)val); in __rpmsg64_to_cpu() [all …]
|
/include/uapi/sound/ |
D | asound.h | 173 #define SNDRV_PCM_ACCESS_MMAP_INTERLEAVED ((__force snd_pcm_access_t) 0) /* interleaved mmap */ 174 #define SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED ((__force snd_pcm_access_t) 1) /* noninterleaved mmap … 175 #define SNDRV_PCM_ACCESS_MMAP_COMPLEX ((__force snd_pcm_access_t) 2) /* complex mmap */ 176 #define SNDRV_PCM_ACCESS_RW_INTERLEAVED ((__force snd_pcm_access_t) 3) /* readi/writei */ 177 #define SNDRV_PCM_ACCESS_RW_NONINTERLEAVED ((__force snd_pcm_access_t) 4) /* readn/writen */ 181 #define SNDRV_PCM_FORMAT_S8 ((__force snd_pcm_format_t) 0) 182 #define SNDRV_PCM_FORMAT_U8 ((__force snd_pcm_format_t) 1) 183 #define SNDRV_PCM_FORMAT_S16_LE ((__force snd_pcm_format_t) 2) 184 #define SNDRV_PCM_FORMAT_S16_BE ((__force snd_pcm_format_t) 3) 185 #define SNDRV_PCM_FORMAT_U16_LE ((__force snd_pcm_format_t) 4) [all …]
|
/include/net/netfilter/ |
D | nf_queue.h | 52 if ((__force u32)iph->saddr < (__force u32)iph->daddr) in hash_v4() 53 return jhash_3words((__force u32)iph->saddr, in hash_v4() 54 (__force u32)iph->daddr, iph->protocol, initval); in hash_v4() 56 return jhash_3words((__force u32)iph->daddr, in hash_v4() 57 (__force u32)iph->saddr, iph->protocol, initval); in hash_v4() 64 if ((__force u32)ip6h->saddr.s6_addr32[3] < in hash_v6() 65 (__force u32)ip6h->daddr.s6_addr32[3]) { in hash_v6() 66 a = (__force u32) ip6h->saddr.s6_addr32[3]; in hash_v6() 67 b = (__force u32) ip6h->daddr.s6_addr32[3]; in hash_v6() 69 b = (__force u32) ip6h->saddr.s6_addr32[3]; in hash_v6() [all …]
|
/include/net/ |
D | checksum.h | 61 u32 res = (__force u32)csum; in csum_add() 62 res += (__force u32)addend; in csum_add() 63 return (__force __wsum)(res + (res < (__force u32)addend)); in csum_add() 74 u16 res = (__force u16)csum; in csum16_add() 76 res += (__force u16)addend; in csum16_add() 77 return (__force __sum16)(res + (res < (__force u16)addend)); in csum16_add() 90 return (__force __wsum)ror32((__force u32)sum, 8); in csum_shift() 115 return (__force __wsum)n; in csum_unfold() 124 #define CSUM_MANGLED_0 ((__force __sum16)0xffff) 133 __wsum tmp = csum_sub(~csum_unfold(*sum), (__force __wsum)from); in csum_replace4() [all …]
|
D | dsfield.h | 24 return ntohs(*(__force const __be16 *)ipv6h) >> 4; in ipv6_get_dsfield() 31 __u32 check = ntohs((__force __be16)iph->check); in ipv4_change_dsfield() 39 iph->check = (__force __sum16)htons(check); in ipv4_change_dsfield() 47 __be16 *p = (__force __be16 *)ipv6h; in ipv6_change_dsfield()
|
D | inet_dscp.h | 44 return (__force dscp_t)(dsfield & INET_DSCP_MASK); in inet_dsfield_to_dscp() 49 return (__force __u8)dscp; in inet_dscp_to_dsfield()
|
/include/net/sctp/ |
D | checksum.h | 36 return (__force __wsum)crc32c((__force __u32)sum, buff, len); in sctp_csum_update() 42 return (__force __wsum)__crc32c_le_combine((__force __u32)csum, in sctp_csum_combine() 43 (__force __u32)csum2, len); in sctp_csum_combine() 63 return cpu_to_le32((__force __u32)new); in sctp_compute_cksum()
|
/include/asm-generic/ |
D | uaccess.h | 23 *(u8 *)to = *((u8 __force *)from); in __get_user_fn() 26 *(u16 *)to = get_unaligned((u16 __force *)from); in __get_user_fn() 29 *(u32 *)to = get_unaligned((u32 __force *)from); in __get_user_fn() 32 *(u64 *)to = get_unaligned((u64 __force *)from); in __get_user_fn() 49 *(u8 __force *)to = *(u8 *)from; in __put_user_fn() 52 put_unaligned(*(u16 *)from, (u16 __force *)to); in __put_user_fn() 55 put_unaligned(*(u32 *)from, (u32 __force *)to); in __put_user_fn() 58 put_unaligned(*(u64 *)from, (u64 __force *)to); in __put_user_fn() 84 memcpy(to, (const void __force *)from, n); in raw_copy_from_user() 91 memcpy((void __force *)to, from, n); in raw_copy_to_user() [all …]
|
D | io.h | 116 return *(const volatile u8 __force *)addr; in __raw_readb() 124 return *(const volatile u16 __force *)addr; in __raw_readw() 132 return *(const volatile u32 __force *)addr; in __raw_readl() 141 return *(const volatile u64 __force *)addr; in __raw_readq() 150 *(volatile u8 __force *)addr = value; in __raw_writeb() 158 *(volatile u16 __force *)addr = value; in __raw_writew() 166 *(volatile u32 __force *)addr = value; in __raw_writel() 175 *(volatile u64 __force *)addr = value; in __raw_writeq() 208 val = __le16_to_cpu((__le16 __force)__raw_readw(addr)); in readw() 223 val = __le32_to_cpu((__le32 __force)__raw_readl(addr)); in readl() [all …]
|
/include/uapi/asm-generic/ |
D | signal-defs.h | 88 #define SIG_DFL ((__force __sighandler_t)0) /* default signal handling */ 89 #define SIG_IGN ((__force __sighandler_t)1) /* ignore signal */ 90 #define SIG_ERR ((__force __sighandler_t)-1) /* error return from signal */
|
D | poll.h | 32 #define POLLFREE (__force __poll_t)0x4000 34 #define POLL_BUSY_LOOP (__force __poll_t)0x8000
|
/include/linux/mlx5/ |
D | doorbell.h | 55 __raw_writel((__force u32) val[0], dest); in mlx5_write64() 56 __raw_writel((__force u32) val[1], dest + 4); in mlx5_write64()
|