/io_uring/ |
D | xattr.c | 36 static void io_xattr_finish(struct io_kiocb *req, int ret) in io_xattr_finish() argument 41 io_req_set_res(req, ret, 0); in io_xattr_finish() 49 int ret; in __io_getxattr_prep() local 68 ret = strncpy_from_user(ix->ctx.kname->name, name, in __io_getxattr_prep() 70 if (!ret || ret == sizeof(ix->ctx.kname->name)) in __io_getxattr_prep() 71 ret = -ERANGE; in __io_getxattr_prep() 72 if (ret < 0) { in __io_getxattr_prep() 74 return ret; in __io_getxattr_prep() 90 int ret; in io_getxattr_prep() local 92 ret = __io_getxattr_prep(req, sqe); in io_getxattr_prep() [all …]
|
D | filetable.c | 20 int ret; in io_file_bitmap_get() local 26 ret = find_next_zero_bit(table->bitmap, nr, table->alloc_hint); in io_file_bitmap_get() 27 if (ret != nr) in io_file_bitmap_get() 28 return ret; in io_file_bitmap_get() 69 int ret; in io_install_fixed_file() local 84 ret = io_rsrc_node_switch_start(ctx); in io_install_fixed_file() 85 if (ret) in io_install_fixed_file() 89 ret = io_queue_rsrc_removal(ctx->file_data, slot_index, in io_install_fixed_file() 91 if (ret) in io_install_fixed_file() 98 ret = io_scm_file_account(ctx, file); in io_install_fixed_file() [all …]
|
D | net.c | 89 int ret; in io_shutdown() local 98 ret = __sys_shutdown_sock(sock, shutdown->how); in io_shutdown() 99 io_req_set_res(req, ret, 0); in io_shutdown() 190 int ret; in io_sendmsg_copy_hdr() local 194 ret = sendmsg_copy_msghdr(&iomsg->msg, sr->umsg, sr->msg_flags, in io_sendmsg_copy_hdr() 198 return ret; in io_sendmsg_copy_hdr() 205 int ret; in io_send_prep_async() local 212 ret = move_addr_to_kernel(zc->addr, zc->addr_len, &io->addr); in io_send_prep_async() 213 return ret; in io_send_prep_async() 234 int ret; in io_sendmsg_prep_async() local [all …]
|
D | cancel.c | 58 int ret = 0; in io_async_cancel_one() local 68 ret = 0; in io_async_cancel_one() 71 ret = -EALREADY; in io_async_cancel_one() 74 ret = -ENOENT; in io_async_cancel_one() 78 return ret; in io_async_cancel_one() 85 int ret; in io_try_cancel() local 89 ret = io_async_cancel_one(tctx, cd); in io_try_cancel() 94 if (!ret) in io_try_cancel() 97 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel() 98 if (ret != -ENOENT) in io_try_cancel() [all …]
|
D | tctx.c | 64 int ret; in io_uring_alloc_task_context() local 70 ret = percpu_counter_init(&tctx->inflight, 0, GFP_KERNEL); in io_uring_alloc_task_context() 71 if (unlikely(ret)) { in io_uring_alloc_task_context() 73 return ret; in io_uring_alloc_task_context() 78 ret = PTR_ERR(tctx->io_wq); in io_uring_alloc_task_context() 81 return ret; in io_uring_alloc_task_context() 98 int ret; in __io_uring_add_tctx_node() local 101 ret = io_uring_alloc_task_context(current, ctx); in __io_uring_add_tctx_node() 102 if (unlikely(ret)) in __io_uring_add_tctx_node() 103 return ret; in __io_uring_add_tctx_node() [all …]
|
D | rw.c | 80 int ret; in io_prep_rw() local 100 ret = ioprio_check_cap(ioprio); in io_prep_rw() 101 if (ret) in io_prep_rw() 102 return ret; in io_prep_rw() 117 ret = io_iov_buffer_select_prep(req); in io_prep_rw() 118 if (ret) in io_prep_rw() 119 return ret; in io_prep_rw() 132 static inline void io_rw_done(struct kiocb *kiocb, ssize_t ret) in io_rw_done() argument 134 switch (ret) { in io_rw_done() 146 ret = -EINTR; in io_rw_done() [all …]
|
D | openclose.c | 38 int ret; in __io_openat_prep() local 53 ret = PTR_ERR(open->filename); in __io_openat_prep() 55 return ret; in __io_openat_prep() 82 int ret; in io_openat2_prep() local 89 ret = copy_struct_from_user(&open->how, sizeof(open->how), how, len); in io_openat2_prep() 90 if (ret) in io_openat2_prep() 91 return ret; in io_openat2_prep() 103 int ret; in io_openat2() local 105 ret = build_open_flags(&open->how, &op); in io_openat2() 106 if (ret) in io_openat2() [all …]
|
D | rsrc.c | 76 int ret; in io_account_mem() local 79 ret = __io_account_mem(ctx->user, nr_pages); in io_account_mem() 80 if (ret) in io_account_mem() 81 return ret; in io_account_mem() 307 int ret; in io_rsrc_ref_quiesce() local 315 ret = io_rsrc_node_switch_start(ctx); in io_rsrc_ref_quiesce() 316 if (ret) in io_rsrc_ref_quiesce() 325 ret = wait_for_completion_interruptible(&data->done); in io_rsrc_ref_quiesce() 326 if (!ret) { in io_rsrc_ref_quiesce() 344 ret = io_run_task_work_sig(ctx); in io_rsrc_ref_quiesce() [all …]
|
D | msg_ring.c | 64 int ret; in io_msg_ring_data() local 71 ret = -EOVERFLOW; in io_msg_ring_data() 76 ret = 0; in io_msg_ring_data() 80 ret = 0; in io_msg_ring_data() 83 return ret; in io_msg_ring_data() 111 int ret; in io_msg_install_complete() local 116 ret = __io_fixed_fd_install(target_ctx, src_file, msg->dst_fd); in io_msg_install_complete() 117 if (ret < 0) in io_msg_install_complete() 131 if (!io_post_aux_cqe(target_ctx, msg->user_data, ret, 0, true)) in io_msg_install_complete() 132 ret = -EOVERFLOW; in io_msg_install_complete() [all …]
|
D | io_uring.c | 646 bool ret = true; in io_cqring_overflow_flush() local 652 ret = __io_cqring_overflow_flush(ctx, false); in io_cqring_overflow_flush() 657 return ret; in io_cqring_overflow_flush() 915 int ret, i; in __io_alloc_req_refill() local 928 ret = kmem_cache_alloc_bulk(req_cachep, gfp, ARRAY_SIZE(reqs), reqs); in __io_alloc_req_refill() 934 if (unlikely(ret <= 0)) { in __io_alloc_req_refill() 938 ret = 1; in __io_alloc_req_refill() 941 percpu_ref_get_many(&ctx->refs, ret); in __io_alloc_req_refill() 942 for (i = 0; i < ret; i++) { in __io_alloc_req_refill() 1186 int ret; in __io_run_local_work() local [all …]
|
D | uring_cmd.c | 46 void io_uring_cmd_done(struct io_uring_cmd *ioucmd, ssize_t ret, ssize_t res2, in io_uring_cmd_done() argument 51 if (ret < 0) in io_uring_cmd_done() 54 io_req_set_res(req, ret, 0); in io_uring_cmd_done() 111 int ret; in io_uring_cmd() local 116 ret = security_uring_cmd(ioucmd); in io_uring_cmd() 117 if (ret) in io_uring_cmd() 118 return ret; in io_uring_cmd() 135 ret = file->f_op->uring_cmd(ioucmd, issue_flags); in io_uring_cmd() 136 if (ret == -EAGAIN) { in io_uring_cmd() 145 if (ret != -EIOCBQUEUED) { in io_uring_cmd() [all …]
|
D | sync.c | 41 int ret; in io_sync_file_range() local 47 ret = sync_file_range(req->file, sync->off, sync->len, sync->flags); in io_sync_file_range() 48 io_req_set_res(req, ret, 0); in io_sync_file_range() 72 int ret; in io_fsync() local 78 ret = vfs_fsync_range(req->file, sync->off, end > 0 ? end : LLONG_MAX, in io_fsync() 80 io_req_set_res(req, ret, 0); in io_fsync() 100 int ret; in io_fallocate() local 105 ret = vfs_fallocate(req->file, sync->mode, sync->off, sync->len); in io_fallocate() 106 if (ret >= 0) in io_fallocate() 108 io_req_set_res(req, ret, 0); in io_fallocate()
|
D | splice.c | 53 long ret = 0; in io_tee() local 63 ret = -EBADF; in io_tee() 68 ret = do_tee(in, out, sp->len, flags); in io_tee() 73 if (ret != sp->len) in io_tee() 75 io_req_set_res(req, ret, 0); in io_tee() 95 long ret = 0; in io_splice() local 105 ret = -EBADF; in io_splice() 113 ret = do_splice(in, poff_in, out, poff_out, sp->len, flags); in io_splice() 118 if (ret != sp->len) in io_splice() 120 io_req_set_res(req, ret, 0); in io_splice()
|
D | sqpoll.c | 170 int ret = 0; in __io_sq_thread() local 193 ret = io_submit_sqes(ctx, to_submit); in __io_sq_thread() 202 return ret; in __io_sq_thread() 250 int ret = __io_sq_thread(ctx, cap_entries); in io_sq_thread() local 252 if (!sqt_spin && (ret > 0 || !wq_list_empty(&ctx->iopoll_list))) in io_sq_thread() 340 int ret; in io_sq_offload_create() local 361 ret = security_uring_sqpoll(); in io_sq_offload_create() 362 if (ret) in io_sq_offload_create() 363 return ret; in io_sq_offload_create() 367 ret = PTR_ERR(sqd); in io_sq_offload_create() [all …]
|
D | poll.c | 307 int ret = io_poll_issue(req, locked); in io_poll_check_events() local 308 if (ret == IOU_STOP_MULTISHOT) in io_poll_check_events() 310 if (ret < 0) in io_poll_check_events() 311 return ret; in io_poll_check_events() 329 int ret; in io_poll_task_func() local 331 ret = io_poll_check_events(req, locked); in io_poll_task_func() 332 if (ret == IOU_POLL_NO_ACTION) in io_poll_task_func() 335 if (ret == IOU_POLL_DONE) { in io_poll_task_func() 338 } else if (ret == IOU_POLL_REISSUE) { in io_poll_task_func() 343 } else if (ret != IOU_POLL_REMOVE_POLL_USE_RES) { in io_poll_task_func() [all …]
|
D | fs.c | 83 int ret; in io_renameat() local 88 ret = do_renameat2(ren->old_dfd, ren->oldpath, ren->new_dfd, in io_renameat() 92 io_req_set_res(req, ret, 0); in io_renameat() 132 int ret; in io_unlinkat() local 138 ret = do_rmdir(un->dfd, un->filename); in io_unlinkat() 140 ret = do_unlinkat(un->dfd, un->filename); in io_unlinkat() 143 io_req_set_res(req, ret, 0); in io_unlinkat() 179 int ret; in io_mkdirat() local 184 ret = do_mkdirat(mkd->dfd, mkd->filename, mkd->mode); in io_mkdirat() 187 io_req_set_res(req, ret, 0); in io_mkdirat() [all …]
|
D | advise.c | 52 int ret; in io_madvise() local 57 ret = do_madvise(current->mm, ma->addr, ma->len, ma->advice); in io_madvise() 58 io_req_set_res(req, ret, 0); in io_madvise() 81 int ret; in io_fadvise() local 94 ret = vfs_fadvise(req->file, fa->offset, fa->len, fa->advice); in io_fadvise() 95 if (ret < 0) in io_fadvise() 97 io_req_set_res(req, ret, 0); in io_fadvise()
|
D | kbuf.c | 179 void __user *ret = NULL; in io_buffer_select() local 186 ret = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select() 188 ret = io_provided_buffer_select(req, len, bl); in io_buffer_select() 191 return ret; in io_buffer_select() 301 int ret = 0; in io_remove_buffers() local 305 ret = -ENOENT; in io_remove_buffers() 308 ret = -EINVAL; in io_remove_buffers() 311 ret = __io_remove_buffers(ctx, bl, p->nbufs); in io_remove_buffers() 313 if (ret < 0) in io_remove_buffers() 317 io_req_set_res(req, ret, 0); in io_remove_buffers() [all …]
|
D | epoll.c | 49 int ret; in io_epoll_ctl() local 52 ret = do_epoll_ctl(ie->epfd, ie->op, ie->fd, &ie->event, force_nonblock); in io_epoll_ctl() 53 if (force_nonblock && ret == -EAGAIN) in io_epoll_ctl() 56 if (ret < 0) in io_epoll_ctl() 58 io_req_set_res(req, ret, 0); in io_epoll_ctl()
|
D | notif.h | 34 int ret; in io_notif_account_mem() local 37 ret = __io_account_mem(ctx->user, nr_pages); in io_notif_account_mem() 38 if (ret) in io_notif_account_mem() 39 return ret; in io_notif_account_mem()
|
D | statx.c | 44 int ret = PTR_ERR(sx->filename); in io_statx_prep() local 47 return ret; in io_statx_prep() 57 int ret; in io_statx() local 62 ret = do_statx(sx->dfd, sx->filename, sx->flags, sx->mask, sx->buffer); in io_statx() 63 io_req_set_res(req, ret, 0); in io_statx()
|
D | io_uring.h | 60 void io_req_task_queue_fail(struct io_kiocb *req, int ret); 285 int ret = 0; in io_run_task_work_ctx() local 289 ret = io_run_local_work(ctx); in io_run_task_work_ctx() 297 if (ret >= 0) in io_run_task_work_ctx() 298 ret += ret2; in io_run_task_work_ctx() 300 return ret; in io_run_task_work_ctx() 306 int ret; in io_run_local_work_locked() local 312 ret = __io_run_local_work(ctx, &locked); in io_run_local_work_locked() 316 return ret; in io_run_local_work_locked()
|
D | slist.h | 39 struct io_wq_work_node *ret; in wq_list_merge() local 42 ret = list1->first; in wq_list_merge() 44 ret = list0->first; in wq_list_merge() 49 return ret; in wq_list_merge()
|
D | io-wq.c | 254 bool ret = false; in io_acct_run_queue() local 259 ret = true; in io_acct_run_queue() 262 return ret; in io_acct_run_queue() 461 bool ret = false; in io_wait_on_hash() local 469 ret = true; in io_wait_on_hash() 473 return ret; in io_wait_on_hash() 638 long ret; in io_wqe_worker() local 657 ret = schedule_timeout(WORKER_IDLE_TIMEOUT); in io_wqe_worker() 665 last_timeout = !ret; in io_wqe_worker() 853 bool ret = false; in io_wq_for_each_worker() local [all …]
|
D | timeout.c | 275 int ret = -ENOENT; in io_req_task_link_timeout() local 284 ret = io_try_cancel(req->task->io_uring, &cd, issue_flags); in io_req_task_link_timeout() 286 io_req_set_res(req, ret ?: -ETIME, 0); in io_req_task_link_timeout() 435 int ret; in io_timeout_remove() local 441 ret = io_timeout_cancel(ctx, &cd); in io_timeout_remove() 448 ret = io_linked_timeout_update(ctx, tr->addr, &tr->ts, mode); in io_timeout_remove() 450 ret = io_timeout_update(ctx, tr->addr, &tr->ts, mode); in io_timeout_remove() 454 if (ret < 0) in io_timeout_remove() 456 io_req_set_res(req, ret, 0); in io_timeout_remove()
|