Searched refs:entry (Results 1 – 11 of 11) sorted by relevance
/io_uring/ |
D | alloc_cache.h | 14 struct io_cache_entry *entry) in io_alloc_cache_put() argument 18 wq_stack_add_head(&entry->node, &cache->list); in io_alloc_cache_put() 20 kasan_slab_free_mempool(entry); in io_alloc_cache_put() 34 struct io_cache_entry *entry; in io_alloc_cache_get() local 36 entry = container_of(cache->list.next, struct io_cache_entry, node); in io_alloc_cache_get() 37 kasan_unpoison_range(entry, cache->elem_size); in io_alloc_cache_get() 40 return entry; in io_alloc_cache_get() 59 struct io_cache_entry *entry = io_alloc_cache_get(cache); in io_alloc_cache_free() local 61 if (!entry) in io_alloc_cache_free() 63 free(entry); in io_alloc_cache_free()
|
D | fdinfo.c | 92 unsigned int entry = i + sq_head; in io_uring_show_fdinfo() local 98 sq_idx = READ_ONCE(ctx->sq_array[entry & sq_mask]); in io_uring_show_fdinfo() 125 unsigned int entry = i + cq_head; in io_uring_show_fdinfo() local 126 struct io_uring_cqe *cqe = &r->cqes[(entry & cq_mask) << cq_shift]; in io_uring_show_fdinfo() 129 entry & cq_mask, cqe->user_data, cqe->res, in io_uring_show_fdinfo()
|
D | net.h | 67 void io_netmsg_cache_free(struct io_cache_entry *entry); 69 static inline void io_netmsg_cache_free(struct io_cache_entry *entry) in io_netmsg_cache_free() argument
|
D | poll.c | 176 INIT_LIST_HEAD(&poll->wait.entry); in io_init_poll_iocb() 186 list_del_init(&poll->wait.entry); in io_poll_remove_entry() 411 list_del_init(&poll->wait.entry); in io_pollfree_wake() 448 list_del_init(&poll->wait.entry); in io_poll_wake() 674 struct io_cache_entry *entry; in io_req_alloc_apoll() local 681 entry = io_alloc_cache_get(&ctx->apoll_cache); in io_req_alloc_apoll() 682 if (entry == NULL) in io_req_alloc_apoll() 684 apoll = container_of(entry, struct async_poll, cache); in io_req_alloc_apoll() 1045 void io_apoll_cache_free(struct io_cache_entry *entry) in io_apoll_cache_free() argument 1047 kfree(container_of(entry, struct async_poll, cache)); in io_apoll_cache_free()
|
D | poll.h | 40 void io_apoll_cache_free(struct io_cache_entry *entry);
|
D | timeout.c | 576 struct list_head *entry; in io_timeout() local 587 entry = ctx->timeout_list.prev; in io_timeout() 604 list_for_each_prev(entry, &ctx->timeout_list) { in io_timeout() 605 struct io_timeout *nextt = list_entry(entry, struct io_timeout, list); in io_timeout() 615 list_add(&timeout->list, entry); in io_timeout()
|
D | io-wq.c | 465 if (list_empty(&wq->wait.entry)) { in io_wait_on_hash() 469 list_del_init(&wq->wait.entry); in io_wait_on_hash() 1136 list_del_init(&wait->entry); in io_wq_hash_wake() 1176 INIT_LIST_HEAD(&wq->wait.entry); in io_wq_create() 1253 list_del_init(&wq->wait.entry); in io_wq_exit_workers()
|
D | rsrc.c | 200 struct io_cache_entry *entry; in io_rsrc_node_alloc() local 202 entry = io_alloc_cache_get(&ctx->rsrc_node_cache); in io_rsrc_node_alloc() 203 if (entry) { in io_rsrc_node_alloc() 204 ref_node = container_of(entry, struct io_rsrc_node, cache); in io_rsrc_node_alloc()
|
D | net.c | 149 struct io_cache_entry *entry; in io_msg_alloc_async() local 153 entry = io_alloc_cache_get(&ctx->netmsg_cache); in io_msg_alloc_async() 154 if (entry) { in io_msg_alloc_async() 155 hdr = container_of(entry, struct io_async_msghdr, cache); in io_msg_alloc_async() 1575 void io_netmsg_cache_free(struct io_cache_entry *entry) in io_netmsg_cache_free() argument 1577 kfree(container_of(entry, struct io_async_msghdr, cache)); in io_netmsg_cache_free()
|
D | rw.c | 602 list_del_init(&wait->entry); in io_async_buf_func() 644 INIT_LIST_HEAD(&wait->wait.entry); in io_rw_should_retry()
|
D | io_uring.c | 2599 INIT_LIST_HEAD(&iowq.wq.entry); in io_cqring_wait() 2929 static void io_rsrc_node_cache_free(struct io_cache_entry *entry) in io_rsrc_node_cache_free() argument 2931 kfree(container_of(entry, struct io_rsrc_node, cache)); in io_rsrc_node_cache_free()
|