Home
last modified time | relevance | path

Searched refs:__force (Results 1 – 25 of 100) sorted by relevance

1234

/include/scsi/
Dscsi_devinfo.h9 #define BLIST_NOLUN ((__force blist_flags_t)(1ULL << 0))
12 #define BLIST_FORCELUN ((__force blist_flags_t)(1ULL << 1))
14 #define BLIST_BORKEN ((__force blist_flags_t)(1ULL << 2))
16 #define BLIST_KEY ((__force blist_flags_t)(1ULL << 3))
18 #define BLIST_SINGLELUN ((__force blist_flags_t)(1ULL << 4))
20 #define BLIST_NOTQ ((__force blist_flags_t)(1ULL << 5))
22 #define BLIST_SPARSELUN ((__force blist_flags_t)(1ULL << 6))
24 #define BLIST_MAX5LUN ((__force blist_flags_t)(1ULL << 7))
26 #define BLIST_ISROM ((__force blist_flags_t)(1ULL << 8))
28 #define BLIST_LARGELUN ((__force blist_flags_t)(1ULL << 9))
[all …]
/include/uapi/linux/
Deventpoll.h31 #define EPOLLIN (__force __poll_t)0x00000001
32 #define EPOLLPRI (__force __poll_t)0x00000002
33 #define EPOLLOUT (__force __poll_t)0x00000004
34 #define EPOLLERR (__force __poll_t)0x00000008
35 #define EPOLLHUP (__force __poll_t)0x00000010
36 #define EPOLLNVAL (__force __poll_t)0x00000020
37 #define EPOLLRDNORM (__force __poll_t)0x00000040
38 #define EPOLLRDBAND (__force __poll_t)0x00000080
39 #define EPOLLWRNORM (__force __poll_t)0x00000100
40 #define EPOLLWRBAND (__force __poll_t)0x00000200
[all …]
/include/linux/
Dgfp_types.h83 #define __GFP_DMA ((__force gfp_t)___GFP_DMA)
84 #define __GFP_HIGHMEM ((__force gfp_t)___GFP_HIGHMEM)
85 #define __GFP_DMA32 ((__force gfp_t)___GFP_DMA32)
86 #define __GFP_MOVABLE ((__force gfp_t)___GFP_MOVABLE) /* ZONE_MOVABLE allowed */
87 #define __GFP_CMA ((__force gfp_t)___GFP_CMA)
117 #define __GFP_RECLAIMABLE ((__force gfp_t)___GFP_RECLAIMABLE)
118 #define __GFP_WRITE ((__force gfp_t)___GFP_WRITE)
119 #define __GFP_HARDWALL ((__force gfp_t)___GFP_HARDWALL)
120 #define __GFP_THISNODE ((__force gfp_t)___GFP_THISNODE)
121 #define __GFP_ACCOUNT ((__force gfp_t)___GFP_ACCOUNT)
[all …]
Dblk_types.h103 #define BLK_STS_NOTSUPP ((__force blk_status_t)1)
104 #define BLK_STS_TIMEOUT ((__force blk_status_t)2)
105 #define BLK_STS_NOSPC ((__force blk_status_t)3)
106 #define BLK_STS_TRANSPORT ((__force blk_status_t)4)
107 #define BLK_STS_TARGET ((__force blk_status_t)5)
108 #define BLK_STS_RESV_CONFLICT ((__force blk_status_t)6)
109 #define BLK_STS_MEDIUM ((__force blk_status_t)7)
110 #define BLK_STS_PROTECTION ((__force blk_status_t)8)
111 #define BLK_STS_RESOURCE ((__force blk_status_t)9)
112 #define BLK_STS_IOERR ((__force blk_status_t)10)
[all …]
Dvirtio_byteorder.h19 return le16_to_cpu((__force __le16)val); in __virtio16_to_cpu()
21 return be16_to_cpu((__force __be16)val); in __virtio16_to_cpu()
27 return (__force __virtio16)cpu_to_le16(val); in __cpu_to_virtio16()
29 return (__force __virtio16)cpu_to_be16(val); in __cpu_to_virtio16()
35 return le32_to_cpu((__force __le32)val); in __virtio32_to_cpu()
37 return be32_to_cpu((__force __be32)val); in __virtio32_to_cpu()
43 return (__force __virtio32)cpu_to_le32(val); in __cpu_to_virtio32()
45 return (__force __virtio32)cpu_to_be32(val); in __cpu_to_virtio32()
51 return le64_to_cpu((__force __le64)val); in __virtio64_to_cpu()
53 return be64_to_cpu((__force __be64)val); in __virtio64_to_cpu()
[all …]
Derr.h49 static inline long __must_check PTR_ERR(__force const void *ptr) in PTR_ERR()
59 static inline bool __must_check IS_ERR(__force const void *ptr) in IS_ERR()
70 static inline bool __must_check IS_ERR_OR_NULL(__force const void *ptr) in IS_ERR_OR_NULL()
82 static inline void * __must_check ERR_CAST(__force const void *ptr) in ERR_CAST()
105 static inline int __must_check PTR_ERR_OR_ZERO(__force const void *ptr) in PTR_ERR_OR_ZERO()
Dserial_core.h509 #define UPF_FOURPORT ((__force upf_t) ASYNC_FOURPORT /* 1 */ )
510 #define UPF_SAK ((__force upf_t) ASYNC_SAK /* 2 */ )
511 #define UPF_SPD_HI ((__force upf_t) ASYNC_SPD_HI /* 4 */ )
512 #define UPF_SPD_VHI ((__force upf_t) ASYNC_SPD_VHI /* 5 */ )
513 #define UPF_SPD_CUST ((__force upf_t) ASYNC_SPD_CUST /* 0x0030 */ )
514 #define UPF_SPD_WARP ((__force upf_t) ASYNC_SPD_WARP /* 0x1010 */ )
515 #define UPF_SPD_MASK ((__force upf_t) ASYNC_SPD_MASK /* 0x1030 */ )
516 #define UPF_SKIP_TEST ((__force upf_t) ASYNC_SKIP_TEST /* 6 */ )
517 #define UPF_AUTO_IRQ ((__force upf_t) ASYNC_AUTO_IRQ /* 7 */ )
518 #define UPF_HARDPPS_CD ((__force upf_t) ASYNC_HARDPPS_CD /* 11 */ )
[all …]
Dslab.h30 #define SLAB_CONSISTENCY_CHECKS ((slab_flags_t __force)0x00000100U)
32 #define SLAB_RED_ZONE ((slab_flags_t __force)0x00000400U)
34 #define SLAB_POISON ((slab_flags_t __force)0x00000800U)
36 #define SLAB_KMALLOC ((slab_flags_t __force)0x00001000U)
38 #define SLAB_HWCACHE_ALIGN ((slab_flags_t __force)0x00002000U)
40 #define SLAB_CACHE_DMA ((slab_flags_t __force)0x00004000U)
42 #define SLAB_CACHE_DMA32 ((slab_flags_t __force)0x00008000U)
44 #define SLAB_STORE_USER ((slab_flags_t __force)0x00010000U)
46 #define SLAB_PANIC ((slab_flags_t __force)0x00040000U)
98 #define SLAB_TYPESAFE_BY_RCU ((slab_flags_t __force)0x00080000U)
[all …]
Domap-mailbox.h14 #define IRQ_TX ((__force omap_mbox_irq_t) 1)
15 #define IRQ_RX ((__force omap_mbox_irq_t) 2)
Dmm_types.h1192 VM_FAULT_OOM = (__force vm_fault_t)0x000001,
1193 VM_FAULT_SIGBUS = (__force vm_fault_t)0x000002,
1194 VM_FAULT_MAJOR = (__force vm_fault_t)0x000004,
1195 VM_FAULT_HWPOISON = (__force vm_fault_t)0x000010,
1196 VM_FAULT_HWPOISON_LARGE = (__force vm_fault_t)0x000020,
1197 VM_FAULT_SIGSEGV = (__force vm_fault_t)0x000040,
1198 VM_FAULT_NOPAGE = (__force vm_fault_t)0x000100,
1199 VM_FAULT_LOCKED = (__force vm_fault_t)0x000200,
1200 VM_FAULT_RETRY = (__force vm_fault_t)0x000400,
1201 VM_FAULT_FALLBACK = (__force vm_fault_t)0x000800,
[all …]
Dblk-mq.h32 #define RQF_STARTED ((__force req_flags_t)(1 << 1))
34 #define RQF_FLUSH_SEQ ((__force req_flags_t)(1 << 4))
36 #define RQF_MIXED_MERGE ((__force req_flags_t)(1 << 5))
38 #define RQF_MQ_INFLIGHT ((__force req_flags_t)(1 << 6))
40 #define RQF_DONTPREP ((__force req_flags_t)(1 << 7))
42 #define RQF_SCHED_TAGS ((__force req_flags_t)(1 << 8))
44 #define RQF_USE_SCHED ((__force req_flags_t)(1 << 9))
46 #define RQF_FAILED ((__force req_flags_t)(1 << 10))
48 #define RQF_QUIET ((__force req_flags_t)(1 << 11))
50 #define RQF_IO_STAT ((__force req_flags_t)(1 << 13))
[all …]
/include/uapi/linux/byteorder/
Dlittle_endian.h16 #define __constant_htonl(x) ((__force __be32)___constant_swab32((x)))
17 #define __constant_ntohl(x) ___constant_swab32((__force __be32)(x))
18 #define __constant_htons(x) ((__force __be16)___constant_swab16((x)))
19 #define __constant_ntohs(x) ___constant_swab16((__force __be16)(x))
20 #define __constant_cpu_to_le64(x) ((__force __le64)(__u64)(x))
21 #define __constant_le64_to_cpu(x) ((__force __u64)(__le64)(x))
22 #define __constant_cpu_to_le32(x) ((__force __le32)(__u32)(x))
23 #define __constant_le32_to_cpu(x) ((__force __u32)(__le32)(x))
24 #define __constant_cpu_to_le16(x) ((__force __le16)(__u16)(x))
25 #define __constant_le16_to_cpu(x) ((__force __u16)(__le16)(x))
[all …]
Dbig_endian.h16 #define __constant_htonl(x) ((__force __be32)(__u32)(x))
17 #define __constant_ntohl(x) ((__force __u32)(__be32)(x))
18 #define __constant_htons(x) ((__force __be16)(__u16)(x))
19 #define __constant_ntohs(x) ((__force __u16)(__be16)(x))
20 #define __constant_cpu_to_le64(x) ((__force __le64)___constant_swab64((x)))
21 #define __constant_le64_to_cpu(x) ___constant_swab64((__force __u64)(__le64)(x))
22 #define __constant_cpu_to_le32(x) ((__force __le32)___constant_swab32((x)))
23 #define __constant_le32_to_cpu(x) ___constant_swab32((__force __u32)(__le32)(x))
24 #define __constant_cpu_to_le16(x) ((__force __le16)___constant_swab16((x)))
25 #define __constant_le16_to_cpu(x) ___constant_swab16((__force __u16)(__le16)(x))
[all …]
/include/linux/rpmsg/
Dbyteorder.h22 return le16_to_cpu((__force __le16)val); in __rpmsg16_to_cpu()
24 return be16_to_cpu((__force __be16)val); in __rpmsg16_to_cpu()
30 return (__force __rpmsg16)cpu_to_le16(val); in __cpu_to_rpmsg16()
32 return (__force __rpmsg16)cpu_to_be16(val); in __cpu_to_rpmsg16()
38 return le32_to_cpu((__force __le32)val); in __rpmsg32_to_cpu()
40 return be32_to_cpu((__force __be32)val); in __rpmsg32_to_cpu()
46 return (__force __rpmsg32)cpu_to_le32(val); in __cpu_to_rpmsg32()
48 return (__force __rpmsg32)cpu_to_be32(val); in __cpu_to_rpmsg32()
54 return le64_to_cpu((__force __le64)val); in __rpmsg64_to_cpu()
56 return be64_to_cpu((__force __be64)val); in __rpmsg64_to_cpu()
[all …]
/include/uapi/sound/
Dasound.h173 #define SNDRV_PCM_ACCESS_MMAP_INTERLEAVED ((__force snd_pcm_access_t) 0) /* interleaved mmap */
174 #define SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED ((__force snd_pcm_access_t) 1) /* noninterleaved mmap …
175 #define SNDRV_PCM_ACCESS_MMAP_COMPLEX ((__force snd_pcm_access_t) 2) /* complex mmap */
176 #define SNDRV_PCM_ACCESS_RW_INTERLEAVED ((__force snd_pcm_access_t) 3) /* readi/writei */
177 #define SNDRV_PCM_ACCESS_RW_NONINTERLEAVED ((__force snd_pcm_access_t) 4) /* readn/writen */
181 #define SNDRV_PCM_FORMAT_S8 ((__force snd_pcm_format_t) 0)
182 #define SNDRV_PCM_FORMAT_U8 ((__force snd_pcm_format_t) 1)
183 #define SNDRV_PCM_FORMAT_S16_LE ((__force snd_pcm_format_t) 2)
184 #define SNDRV_PCM_FORMAT_S16_BE ((__force snd_pcm_format_t) 3)
185 #define SNDRV_PCM_FORMAT_U16_LE ((__force snd_pcm_format_t) 4)
[all …]
/include/net/netfilter/
Dnf_queue.h52 if ((__force u32)iph->saddr < (__force u32)iph->daddr) in hash_v4()
53 return jhash_3words((__force u32)iph->saddr, in hash_v4()
54 (__force u32)iph->daddr, iph->protocol, initval); in hash_v4()
56 return jhash_3words((__force u32)iph->daddr, in hash_v4()
57 (__force u32)iph->saddr, iph->protocol, initval); in hash_v4()
64 if ((__force u32)ip6h->saddr.s6_addr32[3] < in hash_v6()
65 (__force u32)ip6h->daddr.s6_addr32[3]) { in hash_v6()
66 a = (__force u32) ip6h->saddr.s6_addr32[3]; in hash_v6()
67 b = (__force u32) ip6h->daddr.s6_addr32[3]; in hash_v6()
69 b = (__force u32) ip6h->saddr.s6_addr32[3]; in hash_v6()
[all …]
/include/net/
Dchecksum.h61 u32 res = (__force u32)csum; in csum_add()
62 res += (__force u32)addend; in csum_add()
63 return (__force __wsum)(res + (res < (__force u32)addend)); in csum_add()
74 u16 res = (__force u16)csum; in csum16_add()
76 res += (__force u16)addend; in csum16_add()
77 return (__force __sum16)(res + (res < (__force u16)addend)); in csum16_add()
90 return (__force __wsum)ror32((__force u32)sum, 8); in csum_shift()
115 return (__force __wsum)n; in csum_unfold()
124 #define CSUM_MANGLED_0 ((__force __sum16)0xffff)
133 __wsum tmp = csum_sub(~csum_unfold(*sum), (__force __wsum)from); in csum_replace4()
[all …]
Ddsfield.h24 return ntohs(*(__force const __be16 *)ipv6h) >> 4; in ipv6_get_dsfield()
31 __u32 check = ntohs((__force __be16)iph->check); in ipv4_change_dsfield()
39 iph->check = (__force __sum16)htons(check); in ipv4_change_dsfield()
47 __be16 *p = (__force __be16 *)ipv6h; in ipv6_change_dsfield()
Dinet_dscp.h44 return (__force dscp_t)(dsfield & INET_DSCP_MASK); in inet_dsfield_to_dscp()
49 return (__force __u8)dscp; in inet_dscp_to_dsfield()
/include/net/sctp/
Dchecksum.h36 return (__force __wsum)crc32c((__force __u32)sum, buff, len); in sctp_csum_update()
42 return (__force __wsum)__crc32c_le_combine((__force __u32)csum, in sctp_csum_combine()
43 (__force __u32)csum2, len); in sctp_csum_combine()
63 return cpu_to_le32((__force __u32)new); in sctp_compute_cksum()
/include/asm-generic/
Duaccess.h23 *(u8 *)to = *((u8 __force *)from); in __get_user_fn()
26 *(u16 *)to = get_unaligned((u16 __force *)from); in __get_user_fn()
29 *(u32 *)to = get_unaligned((u32 __force *)from); in __get_user_fn()
32 *(u64 *)to = get_unaligned((u64 __force *)from); in __get_user_fn()
49 *(u8 __force *)to = *(u8 *)from; in __put_user_fn()
52 put_unaligned(*(u16 *)from, (u16 __force *)to); in __put_user_fn()
55 put_unaligned(*(u32 *)from, (u32 __force *)to); in __put_user_fn()
58 put_unaligned(*(u64 *)from, (u64 __force *)to); in __put_user_fn()
84 memcpy(to, (const void __force *)from, n); in raw_copy_from_user()
91 memcpy((void __force *)to, from, n); in raw_copy_to_user()
[all …]
Dio.h116 return *(const volatile u8 __force *)addr; in __raw_readb()
124 return *(const volatile u16 __force *)addr; in __raw_readw()
132 return *(const volatile u32 __force *)addr; in __raw_readl()
141 return *(const volatile u64 __force *)addr; in __raw_readq()
150 *(volatile u8 __force *)addr = value; in __raw_writeb()
158 *(volatile u16 __force *)addr = value; in __raw_writew()
166 *(volatile u32 __force *)addr = value; in __raw_writel()
175 *(volatile u64 __force *)addr = value; in __raw_writeq()
208 val = __le16_to_cpu((__le16 __force)__raw_readw(addr)); in readw()
223 val = __le32_to_cpu((__le32 __force)__raw_readl(addr)); in readl()
[all …]
/include/uapi/asm-generic/
Dsignal-defs.h88 #define SIG_DFL ((__force __sighandler_t)0) /* default signal handling */
89 #define SIG_IGN ((__force __sighandler_t)1) /* ignore signal */
90 #define SIG_ERR ((__force __sighandler_t)-1) /* error return from signal */
Dpoll.h32 #define POLLFREE (__force __poll_t)0x4000
34 #define POLL_BUSY_LOOP (__force __poll_t)0x8000
/include/linux/mlx5/
Ddoorbell.h55 __raw_writel((__force u32) val[0], dest); in mlx5_write64()
56 __raw_writel((__force u32) val[1], dest + 4); in mlx5_write64()

1234