Searched refs:sq_entries (Results 1 – 4 of 4) sorted by relevance
/io_uring/ |
D | fdinfo.c | 56 unsigned int sq_mask = ctx->sq_entries - 1, cq_mask = ctx->cq_entries - 1; in __io_uring_show_fdinfo() 63 unsigned int sq_entries, cq_entries; in __io_uring_show_fdinfo() local 88 sq_entries = min(sq_tail - sq_head, ctx->sq_entries); in __io_uring_show_fdinfo() 89 for (i = 0; i < sq_entries; i++) { in __io_uring_show_fdinfo()
|
D | io_uring.c | 2230 unsigned head, mask = ctx->sq_entries - 1; in io_get_sqe() 2242 if (likely(head < ctx->sq_entries)) { in io_get_sqe() 2268 ret = left = min3(nr, ctx->sq_entries, entries); in io_submit_sqes() 2502 static unsigned long rings_size(struct io_ring_ctx *ctx, unsigned int sq_entries, in rings_size() argument 2525 sq_array_size = array_size(sizeof(u32), sq_entries); in rings_size() 3396 ctx->sq_entries = p->sq_entries; in io_allocate_scq_urings() 3399 size = rings_size(ctx, p->sq_entries, p->cq_entries, &sq_array_offset); in io_allocate_scq_urings() 3409 rings->sq_ring_mask = p->sq_entries - 1; in io_allocate_scq_urings() 3411 rings->sq_ring_entries = p->sq_entries; in io_allocate_scq_urings() 3415 size = array_size(2 * sizeof(struct io_uring_sqe), p->sq_entries); in io_allocate_scq_urings() [all …]
|
D | tctx.c | 42 concurrency = min(ctx->sq_entries, 4 * num_online_cpus()); in io_init_wq_offload()
|
D | io_uring.h | 239 return READ_ONCE(r->sq.tail) - ctx->cached_sq_head == ctx->sq_entries; in io_sqring_full()
|