Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 58) sorted by relevance

123

/include/linux/
Dlwq.h33 static inline void lwq_init(struct lwq *q) in lwq_init() argument
35 spin_lock_init(&q->lock); in lwq_init()
36 q->ready = NULL; in lwq_init()
37 init_llist_head(&q->new); in lwq_init()
48 static inline bool lwq_empty(struct lwq *q) in lwq_empty() argument
51 return smp_load_acquire(&q->ready) == NULL && llist_empty(&q->new); in lwq_empty()
54 struct llist_node *__lwq_dequeue(struct lwq *q);
65 #define lwq_dequeue(q, type, member) \ argument
66 ({ struct llist_node *_n = __lwq_dequeue(q); \
69 struct llist_node *lwq_dequeue_all(struct lwq *q);
[all …]
Dblktrace_api.h49 #define blk_add_cgroup_trace_msg(q, css, fmt, ...) \ argument
54 bt = rcu_dereference((q)->blk_trace); \
59 #define blk_add_trace_msg(q, fmt, ...) \ argument
60 blk_add_cgroup_trace_msg(q, NULL, fmt, ##__VA_ARGS__)
63 static inline bool blk_trace_note_message_enabled(struct request_queue *q) in blk_trace_note_message_enabled() argument
69 bt = rcu_dereference(q->blk_trace); in blk_trace_note_message_enabled()
76 extern int blk_trace_setup(struct request_queue *q, char *name, dev_t dev,
79 extern int blk_trace_startstop(struct request_queue *q, int start);
80 extern int blk_trace_remove(struct request_queue *q);
84 # define blk_trace_shutdown(q) do { } while (0) argument
[all …]
Dblk-pm.h13 extern void blk_pm_runtime_init(struct request_queue *q, struct device *dev);
14 extern int blk_pre_runtime_suspend(struct request_queue *q);
15 extern void blk_post_runtime_suspend(struct request_queue *q, int err);
16 extern void blk_pre_runtime_resume(struct request_queue *q);
17 extern void blk_post_runtime_resume(struct request_queue *q);
19 static inline void blk_pm_runtime_init(struct request_queue *q, in blk_pm_runtime_init() argument
Dblkdev.h646 void blk_queue_flag_set(unsigned int flag, struct request_queue *q);
647 void blk_queue_flag_clear(unsigned int flag, struct request_queue *q);
649 #define blk_queue_dying(q) test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags) argument
650 #define blk_queue_init_done(q) test_bit(QUEUE_FLAG_INIT_DONE, &(q)->queue_flags) argument
651 #define blk_queue_nomerges(q) test_bit(QUEUE_FLAG_NOMERGES, &(q)->queue_flags) argument
652 #define blk_queue_noxmerges(q) \ argument
653 test_bit(QUEUE_FLAG_NOXMERGES, &(q)->queue_flags)
654 #define blk_queue_nonrot(q) (!((q)->limits.features & BLK_FEAT_ROTATIONAL)) argument
655 #define blk_queue_io_stat(q) ((q)->limits.features & BLK_FEAT_IO_STAT) argument
656 #define blk_queue_dax(q) ((q)->limits.features & BLK_FEAT_DAX) argument
[all …]
Dfortify-string.h82 extern void *__underlying_memmove(void *p, const void *q, __kernel_size_t size) __RENAME(memmove);
83 extern void *__underlying_memcpy(void *p, const void *q, __kernel_size_t size) __RENAME(memcpy);
86 extern void *__underlying_memmove(void *p, const void *q, __kernel_size_t size) __RENAME(__asan_mem…
87 extern void *__underlying_memcpy(void *p, const void *q, __kernel_size_t size) __RENAME(__asan_memc…
90 extern void *__underlying_memmove(void *p, const void *q, __kernel_size_t size) __RENAME(__hwasan_m…
91 extern void *__underlying_memcpy(void *p, const void *q, __kernel_size_t size) __RENAME(__hwasan_me…
95 extern int __underlying_memcmp(const void *p, const void *q, __kernel_size_t size) __RENAME(memcmp);
96 extern char *__underlying_strcat(char *p, const char *q) __RENAME(strcat);
97 extern char *__underlying_strcpy(char *p, const char *q) __RENAME(strcpy);
99 extern char *__underlying_strncat(char *p, const char *q, __kernel_size_t count) __RENAME(strncat);
[all …]
Dt10-pi.h42 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in t10_pi_ref_tag()
45 rq->q->limits.integrity.interval_exp) in t10_pi_ref_tag()
46 shift = rq->q->limits.integrity.interval_exp; in t10_pi_ref_tag()
67 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in ext_pi_ref_tag()
70 rq->q->limits.integrity.interval_exp) in ext_pi_ref_tag()
71 shift = rq->q->limits.integrity.interval_exp; in ext_pi_ref_tag()
Dblk-integrity.h34 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity() argument
36 return q->limits.integrity.tuple_size; in blk_integrity_queue_supports_integrity()
53 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments() argument
55 return q->limits.max_integrity_segments; in queue_max_integrity_segments()
95 static inline int blk_rq_count_integrity_sg(struct request_queue *q, in blk_rq_count_integrity_sg() argument
100 static inline int blk_rq_map_integrity_sg(struct request *q, in blk_rq_map_integrity_sg() argument
121 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity() argument
126 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments() argument
Dswait.h69 extern void __init_swait_queue_head(struct swait_queue_head *q, const char *name,
72 #define init_swait_queue_head(q) \ argument
75 __init_swait_queue_head((q), #q, &__key); \
147 extern void swake_up_one(struct swait_queue_head *q);
148 extern void swake_up_all(struct swait_queue_head *q);
149 extern void swake_up_locked(struct swait_queue_head *q, int wake_flags);
151 extern void prepare_to_swait_exclusive(struct swait_queue_head *q, struct swait_queue *wait, int st…
152 extern long prepare_to_swait_event(struct swait_queue_head *q, struct swait_queue *wait, int state);
154 extern void __finish_swait(struct swait_queue_head *q, struct swait_queue *wait);
155 extern void finish_swait(struct swait_queue_head *q, struct swait_queue *wait);
Duacce.h41 struct uacce_queue *q);
42 void (*put_queue)(struct uacce_queue *q);
43 int (*start_queue)(struct uacce_queue *q);
44 void (*stop_queue)(struct uacce_queue *q);
45 int (*is_q_updated)(struct uacce_queue *q);
46 int (*mmap)(struct uacce_queue *q, struct vm_area_struct *vma,
48 long (*ioctl)(struct uacce_queue *q, unsigned int cmd,
Dblk-mq.h104 struct request_queue *q; member
731 struct gendisk *blk_mq_alloc_disk_for_queue(struct request_queue *q,
736 struct request_queue *q);
749 bool blk_mq_queue_inflight(struct request_queue *q);
760 struct request *blk_mq_alloc_request(struct request_queue *q, blk_opf_t opf,
762 struct request *blk_mq_alloc_request_hctx(struct request_queue *q,
928 void blk_mq_kick_requeue_list(struct request_queue *q);
929 void blk_mq_delay_kick_requeue_list(struct request_queue *q, unsigned long msecs);
934 void blk_mq_stop_hw_queues(struct request_queue *q);
935 void blk_mq_start_hw_queues(struct request_queue *q);
[all …]
/include/media/
Dvideobuf2-core.h435 int (*queue_setup)(struct vb2_queue *q,
439 void (*wait_prepare)(struct vb2_queue *q);
440 void (*wait_finish)(struct vb2_queue *q);
448 int (*prepare_streaming)(struct vb2_queue *q);
449 int (*start_streaming)(struct vb2_queue *q, unsigned int count);
450 void (*stop_streaming)(struct vb2_queue *q);
451 void (*unprepare_streaming)(struct vb2_queue *q);
698 static inline bool vb2_queue_allows_cache_hints(struct vb2_queue *q) in vb2_queue_allows_cache_hints() argument
700 return q->allow_cache_hints && q->memory == VB2_MEMORY_MMAP; in vb2_queue_allows_cache_hints()
762 void vb2_discard_done(struct vb2_queue *q);
[all …]
Dvideobuf2-v4l2.h76 struct vb2_buffer *vb2_find_buffer(struct vb2_queue *q, u64 timestamp);
78 int vb2_querybuf(struct vb2_queue *q, struct v4l2_buffer *b);
88 int vb2_reqbufs(struct vb2_queue *q, struct v4l2_requestbuffers *req);
98 int vb2_create_bufs(struct vb2_queue *q, struct v4l2_create_buffers *create);
122 int vb2_prepare_buf(struct vb2_queue *q, struct media_device *mdev,
148 int vb2_qbuf(struct vb2_queue *q, struct media_device *mdev,
160 int vb2_expbuf(struct vb2_queue *q, struct v4l2_exportbuffer *eb);
186 int vb2_dqbuf(struct vb2_queue *q, struct v4l2_buffer *b, bool nonblocking);
204 int vb2_streamon(struct vb2_queue *q, enum v4l2_buf_type type);
223 int vb2_streamoff(struct vb2_queue *q, enum v4l2_buf_type type);
[all …]
/include/trace/events/
Dvb2.h12 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
13 TP_ARGS(q, vb),
26 __entry->owner = q->owner;
27 __entry->queued_count = q->queued_count;
29 atomic_read(&q->owned_by_drv_count);
47 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
48 TP_ARGS(q, vb)
52 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
53 TP_ARGS(q, vb)
57 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
[all …]
Dqdisc.h79 TP_PROTO(struct Qdisc *q),
81 TP_ARGS(q),
84 __string( dev, qdisc_dev(q) ? qdisc_dev(q)->name : "(null)" )
85 __string( kind, q->ops->id )
93 __entry->parent = q->parent;
94 __entry->handle = q->handle;
104 TP_PROTO(struct Qdisc *q),
106 TP_ARGS(q),
109 __string( dev, qdisc_dev(q)->name )
110 __string( kind, q->ops->id )
[all …]
Dblock.h99 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0;
135 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0;
208 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0;
309 TP_PROTO(struct request_queue *q, struct bio *bio),
311 TP_ARGS(q, bio),
443 TP_PROTO(struct request_queue *q),
445 TP_ARGS(q),
460 TP_PROTO(struct request_queue *q, unsigned int depth, bool explicit),
462 TP_ARGS(q, depth, explicit),
488 TP_PROTO(struct request_queue *q, unsigned int depth, bool explicit),
[all …]
/include/net/
Dipv6_frag.h24 struct inet_frag_queue q; member
32 static inline void ip6frag_init(struct inet_frag_queue *q, const void *a) in ip6frag_init() argument
34 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init()
37 q->key.v6 = *key; in ip6frag_init()
72 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue()
74 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue()
76 if (fq->q.flags & INET_FRAG_COMPLETE) in ip6frag_expire_frag_queue()
79 fq->q.flags |= INET_FRAG_DROP; in ip6frag_expire_frag_queue()
80 inet_frag_kill(&fq->q); in ip6frag_expire_frag_queue()
90 if (!(fq->q.flags & INET_FRAG_FIRST_IN)) in ip6frag_expire_frag_queue()
[all …]
Dinet_frag.h110 void (*constructor)(struct inet_frag_queue *q,
140 void inet_frag_kill(struct inet_frag_queue *q);
141 void inet_frag_destroy(struct inet_frag_queue *q);
148 static inline void inet_frag_put(struct inet_frag_queue *q) in inet_frag_put() argument
150 if (refcount_dec_and_test(&q->refcnt)) in inet_frag_put()
151 inet_frag_destroy(q); in inet_frag_put()
186 int inet_frag_queue_insert(struct inet_frag_queue *q, struct sk_buff *skb,
188 void *inet_frag_reasm_prepare(struct inet_frag_queue *q, struct sk_buff *skb,
190 void inet_frag_reasm_finish(struct inet_frag_queue *q, struct sk_buff *head,
192 struct sk_buff *inet_frag_pull_head(struct inet_frag_queue *q);
Dpkt_sched.h23 #define qdisc_priv(q) \ argument
24 _Generic(q, \
25 const struct Qdisc * : (const void *)&q->privdata, \
26 struct Qdisc * : (void *)&q->privdata)
96 int fifo_set_limit(struct Qdisc *q, unsigned int limit);
108 void qdisc_hash_add(struct Qdisc *q, bool invisible);
109 void qdisc_hash_del(struct Qdisc *q);
118 bool sch_direct_xmit(struct sk_buff *skb, struct Qdisc *q,
122 void __qdisc_run(struct Qdisc *q);
124 static inline void qdisc_run(struct Qdisc *q) in qdisc_run() argument
[all …]
Dsch_generic.h118 struct qdisc_skb_head q; member
183 static inline bool qdisc_is_percpu_stats(const struct Qdisc *q) in qdisc_is_percpu_stats() argument
185 return q->flags & TCQ_F_CPUSTATS; in qdisc_is_percpu_stats()
192 return !READ_ONCE(qdisc->q.qlen); in qdisc_is_empty()
481 struct Qdisc *q; member
525 static inline int qdisc_qlen(const struct Qdisc *q) in qdisc_qlen() argument
527 return q->q.qlen; in qdisc_qlen()
530 static inline int qdisc_qlen_sum(const struct Qdisc *q) in qdisc_qlen_sum() argument
532 __u32 qlen = q->qstats.qlen; in qdisc_qlen_sum()
535 if (qdisc_is_percpu_stats(q)) { in qdisc_qlen_sum()
[all …]
Dpkt_cls.h56 struct tcf_proto __rcu **p_filter_chain, struct Qdisc *q,
58 int tcf_block_get_ext(struct tcf_block **p_block, struct Qdisc *q,
62 void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q,
89 return block->q; in tcf_block_q()
123 struct tcf_proto __rcu **p_filter_chain, struct Qdisc *q, in tcf_block_get() argument
130 int tcf_block_get_ext(struct tcf_block **p_block, struct Qdisc *q, in tcf_block_get_ext() argument
142 void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q, in tcf_block_put_ext() argument
169 __tcf_bind_filter(struct Qdisc *q, struct tcf_result *r, unsigned long base) in __tcf_bind_filter() argument
173 cl = q->ops->cl_ops->bind_tcf(q, base, r->classid); in __tcf_bind_filter()
176 q->ops->cl_ops->unbind_tcf(q, cl); in __tcf_bind_filter()
[all …]
/include/crypto/
Db128ops.h60 static inline void be128_xor(be128 *r, const be128 *p, const be128 *q) in be128_xor() argument
62 r->a = p->a ^ q->a; in be128_xor()
63 r->b = p->b ^ q->b; in be128_xor()
66 static inline void le128_xor(le128 *r, const le128 *p, const le128 *q) in le128_xor() argument
68 r->a = p->a ^ q->a; in le128_xor()
69 r->b = p->b ^ q->b; in le128_xor()
Daria.h436 int q = 4 - (n / 32); in aria_gsrk() local
440 ((y[q % 4]) >> r) ^ in aria_gsrk()
441 ((y[(q + 3) % 4]) << (32 - r)); in aria_gsrk()
443 ((y[(q + 1) % 4]) >> r) ^ in aria_gsrk()
444 ((y[q % 4]) << (32 - r)); in aria_gsrk()
446 ((y[(q + 2) % 4]) >> r) ^ in aria_gsrk()
447 ((y[(q + 1) % 4]) << (32 - r)); in aria_gsrk()
449 ((y[(q + 3) % 4]) >> r) ^ in aria_gsrk()
450 ((y[(q + 2) % 4]) << (32 - r)); in aria_gsrk()
/include/net/sctp/
Dstream_sched.h33 void (*enqueue)(struct sctp_outq *q, struct sctp_datamsg *msg);
35 struct sctp_chunk *(*dequeue)(struct sctp_outq *q);
37 void (*dequeue_done)(struct sctp_outq *q, struct sctp_chunk *chunk);
51 void sctp_sched_dequeue_done(struct sctp_outq *q, struct sctp_chunk *ch);
53 void sctp_sched_dequeue_common(struct sctp_outq *q, struct sctp_chunk *ch);
/include/linux/dsa/
Dbrcm.h12 #define BRCM_TAG_SET_PORT_QUEUE(p, q) ((p) << 8 | q) argument
/include/linux/sunrpc/
Dsched.h213 #define RPC_IS_PRIORITY(q) ((q)->maxpriority > 0) argument
286 static inline const char * rpc_qname(const struct rpc_wait_queue *q) in rpc_qname() argument
288 return ((q && q->name) ? q->name : "unknown"); in rpc_qname()
291 static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q, in rpc_assign_waitqueue_name() argument
294 q->name = name; in rpc_assign_waitqueue_name()
297 static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q, in rpc_assign_waitqueue_name() argument

123