Home
last modified time | relevance | path

Searched refs:ret (Results 1 – 25 of 26) sorted by relevance

12

/io_uring/
Dxattr.c36 static void io_xattr_finish(struct io_kiocb *req, int ret) in io_xattr_finish() argument
41 io_req_set_res(req, ret, 0); in io_xattr_finish()
49 int ret; in __io_getxattr_prep() local
68 ret = strncpy_from_user(ix->ctx.kname->name, name, in __io_getxattr_prep()
70 if (!ret || ret == sizeof(ix->ctx.kname->name)) in __io_getxattr_prep()
71 ret = -ERANGE; in __io_getxattr_prep()
72 if (ret < 0) { in __io_getxattr_prep()
74 return ret; in __io_getxattr_prep()
90 int ret; in io_getxattr_prep() local
92 ret = __io_getxattr_prep(req, sqe); in io_getxattr_prep()
[all …]
Dfiletable.c20 int ret; in io_file_bitmap_get() local
26 ret = find_next_zero_bit(table->bitmap, nr, table->alloc_hint); in io_file_bitmap_get()
27 if (ret != nr) in io_file_bitmap_get()
28 return ret; in io_file_bitmap_get()
69 int ret; in io_install_fixed_file() local
84 ret = io_rsrc_node_switch_start(ctx); in io_install_fixed_file()
85 if (ret) in io_install_fixed_file()
89 ret = io_queue_rsrc_removal(ctx->file_data, slot_index, in io_install_fixed_file()
91 if (ret) in io_install_fixed_file()
98 ret = io_scm_file_account(ctx, file); in io_install_fixed_file()
[all …]
Dnet.c89 int ret; in io_shutdown() local
98 ret = __sys_shutdown_sock(sock, shutdown->how); in io_shutdown()
99 io_req_set_res(req, ret, 0); in io_shutdown()
190 int ret; in io_sendmsg_copy_hdr() local
194 ret = sendmsg_copy_msghdr(&iomsg->msg, sr->umsg, sr->msg_flags, in io_sendmsg_copy_hdr()
198 return ret; in io_sendmsg_copy_hdr()
205 int ret; in io_send_prep_async() local
212 ret = move_addr_to_kernel(zc->addr, zc->addr_len, &io->addr); in io_send_prep_async()
213 return ret; in io_send_prep_async()
234 int ret; in io_sendmsg_prep_async() local
[all …]
Dcancel.c58 int ret = 0; in io_async_cancel_one() local
68 ret = 0; in io_async_cancel_one()
71 ret = -EALREADY; in io_async_cancel_one()
74 ret = -ENOENT; in io_async_cancel_one()
78 return ret; in io_async_cancel_one()
85 int ret; in io_try_cancel() local
89 ret = io_async_cancel_one(tctx, cd); in io_try_cancel()
94 if (!ret) in io_try_cancel()
97 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel()
98 if (ret != -ENOENT) in io_try_cancel()
[all …]
Dtctx.c64 int ret; in io_uring_alloc_task_context() local
70 ret = percpu_counter_init(&tctx->inflight, 0, GFP_KERNEL); in io_uring_alloc_task_context()
71 if (unlikely(ret)) { in io_uring_alloc_task_context()
73 return ret; in io_uring_alloc_task_context()
78 ret = PTR_ERR(tctx->io_wq); in io_uring_alloc_task_context()
81 return ret; in io_uring_alloc_task_context()
98 int ret; in __io_uring_add_tctx_node() local
101 ret = io_uring_alloc_task_context(current, ctx); in __io_uring_add_tctx_node()
102 if (unlikely(ret)) in __io_uring_add_tctx_node()
103 return ret; in __io_uring_add_tctx_node()
[all …]
Drw.c80 int ret; in io_prep_rw() local
100 ret = ioprio_check_cap(ioprio); in io_prep_rw()
101 if (ret) in io_prep_rw()
102 return ret; in io_prep_rw()
117 ret = io_iov_buffer_select_prep(req); in io_prep_rw()
118 if (ret) in io_prep_rw()
119 return ret; in io_prep_rw()
132 static inline void io_rw_done(struct kiocb *kiocb, ssize_t ret) in io_rw_done() argument
134 switch (ret) { in io_rw_done()
146 ret = -EINTR; in io_rw_done()
[all …]
Dopenclose.c38 int ret; in __io_openat_prep() local
53 ret = PTR_ERR(open->filename); in __io_openat_prep()
55 return ret; in __io_openat_prep()
82 int ret; in io_openat2_prep() local
89 ret = copy_struct_from_user(&open->how, sizeof(open->how), how, len); in io_openat2_prep()
90 if (ret) in io_openat2_prep()
91 return ret; in io_openat2_prep()
103 int ret; in io_openat2() local
105 ret = build_open_flags(&open->how, &op); in io_openat2()
106 if (ret) in io_openat2()
[all …]
Drsrc.c76 int ret; in io_account_mem() local
79 ret = __io_account_mem(ctx->user, nr_pages); in io_account_mem()
80 if (ret) in io_account_mem()
81 return ret; in io_account_mem()
307 int ret; in io_rsrc_ref_quiesce() local
315 ret = io_rsrc_node_switch_start(ctx); in io_rsrc_ref_quiesce()
316 if (ret) in io_rsrc_ref_quiesce()
325 ret = wait_for_completion_interruptible(&data->done); in io_rsrc_ref_quiesce()
326 if (!ret) { in io_rsrc_ref_quiesce()
344 ret = io_run_task_work_sig(ctx); in io_rsrc_ref_quiesce()
[all …]
Dmsg_ring.c64 int ret; in io_msg_ring_data() local
71 ret = -EOVERFLOW; in io_msg_ring_data()
76 ret = 0; in io_msg_ring_data()
80 ret = 0; in io_msg_ring_data()
83 return ret; in io_msg_ring_data()
111 int ret; in io_msg_install_complete() local
116 ret = __io_fixed_fd_install(target_ctx, src_file, msg->dst_fd); in io_msg_install_complete()
117 if (ret < 0) in io_msg_install_complete()
131 if (!io_post_aux_cqe(target_ctx, msg->user_data, ret, 0, true)) in io_msg_install_complete()
132 ret = -EOVERFLOW; in io_msg_install_complete()
[all …]
Dio_uring.c646 bool ret = true; in io_cqring_overflow_flush() local
652 ret = __io_cqring_overflow_flush(ctx, false); in io_cqring_overflow_flush()
657 return ret; in io_cqring_overflow_flush()
915 int ret, i; in __io_alloc_req_refill() local
928 ret = kmem_cache_alloc_bulk(req_cachep, gfp, ARRAY_SIZE(reqs), reqs); in __io_alloc_req_refill()
934 if (unlikely(ret <= 0)) { in __io_alloc_req_refill()
938 ret = 1; in __io_alloc_req_refill()
941 percpu_ref_get_many(&ctx->refs, ret); in __io_alloc_req_refill()
942 for (i = 0; i < ret; i++) { in __io_alloc_req_refill()
1186 int ret; in __io_run_local_work() local
[all …]
During_cmd.c46 void io_uring_cmd_done(struct io_uring_cmd *ioucmd, ssize_t ret, ssize_t res2, in io_uring_cmd_done() argument
51 if (ret < 0) in io_uring_cmd_done()
54 io_req_set_res(req, ret, 0); in io_uring_cmd_done()
111 int ret; in io_uring_cmd() local
116 ret = security_uring_cmd(ioucmd); in io_uring_cmd()
117 if (ret) in io_uring_cmd()
118 return ret; in io_uring_cmd()
135 ret = file->f_op->uring_cmd(ioucmd, issue_flags); in io_uring_cmd()
136 if (ret == -EAGAIN) { in io_uring_cmd()
145 if (ret != -EIOCBQUEUED) { in io_uring_cmd()
[all …]
Dsync.c41 int ret; in io_sync_file_range() local
47 ret = sync_file_range(req->file, sync->off, sync->len, sync->flags); in io_sync_file_range()
48 io_req_set_res(req, ret, 0); in io_sync_file_range()
72 int ret; in io_fsync() local
78 ret = vfs_fsync_range(req->file, sync->off, end > 0 ? end : LLONG_MAX, in io_fsync()
80 io_req_set_res(req, ret, 0); in io_fsync()
100 int ret; in io_fallocate() local
105 ret = vfs_fallocate(req->file, sync->mode, sync->off, sync->len); in io_fallocate()
106 if (ret >= 0) in io_fallocate()
108 io_req_set_res(req, ret, 0); in io_fallocate()
Dsplice.c53 long ret = 0; in io_tee() local
63 ret = -EBADF; in io_tee()
68 ret = do_tee(in, out, sp->len, flags); in io_tee()
73 if (ret != sp->len) in io_tee()
75 io_req_set_res(req, ret, 0); in io_tee()
95 long ret = 0; in io_splice() local
105 ret = -EBADF; in io_splice()
113 ret = do_splice(in, poff_in, out, poff_out, sp->len, flags); in io_splice()
118 if (ret != sp->len) in io_splice()
120 io_req_set_res(req, ret, 0); in io_splice()
Dsqpoll.c170 int ret = 0; in __io_sq_thread() local
193 ret = io_submit_sqes(ctx, to_submit); in __io_sq_thread()
202 return ret; in __io_sq_thread()
250 int ret = __io_sq_thread(ctx, cap_entries); in io_sq_thread() local
252 if (!sqt_spin && (ret > 0 || !wq_list_empty(&ctx->iopoll_list))) in io_sq_thread()
340 int ret; in io_sq_offload_create() local
361 ret = security_uring_sqpoll(); in io_sq_offload_create()
362 if (ret) in io_sq_offload_create()
363 return ret; in io_sq_offload_create()
367 ret = PTR_ERR(sqd); in io_sq_offload_create()
[all …]
Dpoll.c307 int ret = io_poll_issue(req, locked); in io_poll_check_events() local
308 if (ret == IOU_STOP_MULTISHOT) in io_poll_check_events()
310 if (ret < 0) in io_poll_check_events()
311 return ret; in io_poll_check_events()
329 int ret; in io_poll_task_func() local
331 ret = io_poll_check_events(req, locked); in io_poll_task_func()
332 if (ret == IOU_POLL_NO_ACTION) in io_poll_task_func()
335 if (ret == IOU_POLL_DONE) { in io_poll_task_func()
338 } else if (ret == IOU_POLL_REISSUE) { in io_poll_task_func()
343 } else if (ret != IOU_POLL_REMOVE_POLL_USE_RES) { in io_poll_task_func()
[all …]
Dfs.c83 int ret; in io_renameat() local
88 ret = do_renameat2(ren->old_dfd, ren->oldpath, ren->new_dfd, in io_renameat()
92 io_req_set_res(req, ret, 0); in io_renameat()
132 int ret; in io_unlinkat() local
138 ret = do_rmdir(un->dfd, un->filename); in io_unlinkat()
140 ret = do_unlinkat(un->dfd, un->filename); in io_unlinkat()
143 io_req_set_res(req, ret, 0); in io_unlinkat()
179 int ret; in io_mkdirat() local
184 ret = do_mkdirat(mkd->dfd, mkd->filename, mkd->mode); in io_mkdirat()
187 io_req_set_res(req, ret, 0); in io_mkdirat()
[all …]
Dadvise.c52 int ret; in io_madvise() local
57 ret = do_madvise(current->mm, ma->addr, ma->len, ma->advice); in io_madvise()
58 io_req_set_res(req, ret, 0); in io_madvise()
81 int ret; in io_fadvise() local
94 ret = vfs_fadvise(req->file, fa->offset, fa->len, fa->advice); in io_fadvise()
95 if (ret < 0) in io_fadvise()
97 io_req_set_res(req, ret, 0); in io_fadvise()
Dkbuf.c179 void __user *ret = NULL; in io_buffer_select() local
186 ret = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select()
188 ret = io_provided_buffer_select(req, len, bl); in io_buffer_select()
191 return ret; in io_buffer_select()
301 int ret = 0; in io_remove_buffers() local
305 ret = -ENOENT; in io_remove_buffers()
308 ret = -EINVAL; in io_remove_buffers()
311 ret = __io_remove_buffers(ctx, bl, p->nbufs); in io_remove_buffers()
313 if (ret < 0) in io_remove_buffers()
317 io_req_set_res(req, ret, 0); in io_remove_buffers()
[all …]
Depoll.c49 int ret; in io_epoll_ctl() local
52 ret = do_epoll_ctl(ie->epfd, ie->op, ie->fd, &ie->event, force_nonblock); in io_epoll_ctl()
53 if (force_nonblock && ret == -EAGAIN) in io_epoll_ctl()
56 if (ret < 0) in io_epoll_ctl()
58 io_req_set_res(req, ret, 0); in io_epoll_ctl()
Dnotif.h34 int ret; in io_notif_account_mem() local
37 ret = __io_account_mem(ctx->user, nr_pages); in io_notif_account_mem()
38 if (ret) in io_notif_account_mem()
39 return ret; in io_notif_account_mem()
Dstatx.c44 int ret = PTR_ERR(sx->filename); in io_statx_prep() local
47 return ret; in io_statx_prep()
57 int ret; in io_statx() local
62 ret = do_statx(sx->dfd, sx->filename, sx->flags, sx->mask, sx->buffer); in io_statx()
63 io_req_set_res(req, ret, 0); in io_statx()
Dio_uring.h60 void io_req_task_queue_fail(struct io_kiocb *req, int ret);
285 int ret = 0; in io_run_task_work_ctx() local
289 ret = io_run_local_work(ctx); in io_run_task_work_ctx()
297 if (ret >= 0) in io_run_task_work_ctx()
298 ret += ret2; in io_run_task_work_ctx()
300 return ret; in io_run_task_work_ctx()
306 int ret; in io_run_local_work_locked() local
312 ret = __io_run_local_work(ctx, &locked); in io_run_local_work_locked()
316 return ret; in io_run_local_work_locked()
Dslist.h39 struct io_wq_work_node *ret; in wq_list_merge() local
42 ret = list1->first; in wq_list_merge()
44 ret = list0->first; in wq_list_merge()
49 return ret; in wq_list_merge()
Dio-wq.c254 bool ret = false; in io_acct_run_queue() local
259 ret = true; in io_acct_run_queue()
262 return ret; in io_acct_run_queue()
461 bool ret = false; in io_wait_on_hash() local
469 ret = true; in io_wait_on_hash()
473 return ret; in io_wait_on_hash()
638 long ret; in io_wqe_worker() local
657 ret = schedule_timeout(WORKER_IDLE_TIMEOUT); in io_wqe_worker()
665 last_timeout = !ret; in io_wqe_worker()
853 bool ret = false; in io_wq_for_each_worker() local
[all …]
Dtimeout.c275 int ret = -ENOENT; in io_req_task_link_timeout() local
284 ret = io_try_cancel(req->task->io_uring, &cd, issue_flags); in io_req_task_link_timeout()
286 io_req_set_res(req, ret ?: -ETIME, 0); in io_req_task_link_timeout()
435 int ret; in io_timeout_remove() local
441 ret = io_timeout_cancel(ctx, &cd); in io_timeout_remove()
448 ret = io_linked_timeout_update(ctx, tr->addr, &tr->ts, mode); in io_timeout_remove()
450 ret = io_timeout_update(ctx, tr->addr, &tr->ts, mode); in io_timeout_remove()
454 if (ret < 0) in io_timeout_remove()
456 io_req_set_res(req, ret, 0); in io_timeout_remove()

12