Home
last modified time | relevance | path

Searched refs:entry (Results 1 – 11 of 11) sorted by relevance

/io_uring/
Dalloc_cache.h14 struct io_cache_entry *entry) in io_alloc_cache_put() argument
18 wq_stack_add_head(&entry->node, &cache->list); in io_alloc_cache_put()
20 kasan_slab_free_mempool(entry); in io_alloc_cache_put()
34 struct io_cache_entry *entry; in io_alloc_cache_get() local
36 entry = container_of(cache->list.next, struct io_cache_entry, node); in io_alloc_cache_get()
37 kasan_unpoison_range(entry, cache->elem_size); in io_alloc_cache_get()
40 return entry; in io_alloc_cache_get()
59 struct io_cache_entry *entry = io_alloc_cache_get(cache); in io_alloc_cache_free() local
61 if (!entry) in io_alloc_cache_free()
63 free(entry); in io_alloc_cache_free()
Dfdinfo.c92 unsigned int entry = i + sq_head; in io_uring_show_fdinfo() local
98 sq_idx = READ_ONCE(ctx->sq_array[entry & sq_mask]); in io_uring_show_fdinfo()
125 unsigned int entry = i + cq_head; in io_uring_show_fdinfo() local
126 struct io_uring_cqe *cqe = &r->cqes[(entry & cq_mask) << cq_shift]; in io_uring_show_fdinfo()
129 entry & cq_mask, cqe->user_data, cqe->res, in io_uring_show_fdinfo()
Dnet.h67 void io_netmsg_cache_free(struct io_cache_entry *entry);
69 static inline void io_netmsg_cache_free(struct io_cache_entry *entry) in io_netmsg_cache_free() argument
Dpoll.c176 INIT_LIST_HEAD(&poll->wait.entry); in io_init_poll_iocb()
186 list_del_init(&poll->wait.entry); in io_poll_remove_entry()
411 list_del_init(&poll->wait.entry); in io_pollfree_wake()
448 list_del_init(&poll->wait.entry); in io_poll_wake()
674 struct io_cache_entry *entry; in io_req_alloc_apoll() local
681 entry = io_alloc_cache_get(&ctx->apoll_cache); in io_req_alloc_apoll()
682 if (entry == NULL) in io_req_alloc_apoll()
684 apoll = container_of(entry, struct async_poll, cache); in io_req_alloc_apoll()
1045 void io_apoll_cache_free(struct io_cache_entry *entry) in io_apoll_cache_free() argument
1047 kfree(container_of(entry, struct async_poll, cache)); in io_apoll_cache_free()
Dpoll.h40 void io_apoll_cache_free(struct io_cache_entry *entry);
Dtimeout.c576 struct list_head *entry; in io_timeout() local
587 entry = ctx->timeout_list.prev; in io_timeout()
604 list_for_each_prev(entry, &ctx->timeout_list) { in io_timeout()
605 struct io_timeout *nextt = list_entry(entry, struct io_timeout, list); in io_timeout()
615 list_add(&timeout->list, entry); in io_timeout()
Dio-wq.c465 if (list_empty(&wq->wait.entry)) { in io_wait_on_hash()
469 list_del_init(&wq->wait.entry); in io_wait_on_hash()
1136 list_del_init(&wait->entry); in io_wq_hash_wake()
1176 INIT_LIST_HEAD(&wq->wait.entry); in io_wq_create()
1253 list_del_init(&wq->wait.entry); in io_wq_exit_workers()
Drsrc.c200 struct io_cache_entry *entry; in io_rsrc_node_alloc() local
202 entry = io_alloc_cache_get(&ctx->rsrc_node_cache); in io_rsrc_node_alloc()
203 if (entry) { in io_rsrc_node_alloc()
204 ref_node = container_of(entry, struct io_rsrc_node, cache); in io_rsrc_node_alloc()
Dnet.c149 struct io_cache_entry *entry; in io_msg_alloc_async() local
153 entry = io_alloc_cache_get(&ctx->netmsg_cache); in io_msg_alloc_async()
154 if (entry) { in io_msg_alloc_async()
155 hdr = container_of(entry, struct io_async_msghdr, cache); in io_msg_alloc_async()
1575 void io_netmsg_cache_free(struct io_cache_entry *entry) in io_netmsg_cache_free() argument
1577 kfree(container_of(entry, struct io_async_msghdr, cache)); in io_netmsg_cache_free()
Drw.c602 list_del_init(&wait->entry); in io_async_buf_func()
644 INIT_LIST_HEAD(&wait->wait.entry); in io_rw_should_retry()
Dio_uring.c2599 INIT_LIST_HEAD(&iowq.wq.entry); in io_cqring_wait()
2929 static void io_rsrc_node_cache_free(struct io_cache_entry *entry) in io_rsrc_node_cache_free() argument
2931 kfree(container_of(entry, struct io_rsrc_node, cache)); in io_rsrc_node_cache_free()