Home
last modified time | relevance | path

Searched refs:issue_flags (Results 1 – 25 of 41) sorted by relevance

12

/io_uring/
Dnet.h32 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags);
37 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags);
39 int io_send(struct io_kiocb *req, unsigned int issue_flags);
44 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags);
45 int io_recv(struct io_kiocb *req, unsigned int issue_flags);
50 int io_accept(struct io_kiocb *req, unsigned int issue_flags);
53 int io_socket(struct io_kiocb *req, unsigned int issue_flags);
57 int io_connect(struct io_kiocb *req, unsigned int issue_flags);
59 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags);
60 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags);
Dmsg_ring.c33 unsigned int issue_flags) in io_double_lock_ctx() argument
40 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_double_lock_ctx()
60 static int io_msg_ring_data(struct io_kiocb *req, unsigned int issue_flags) in io_msg_ring_data() argument
73 if (unlikely(io_double_lock_ctx(target_ctx, issue_flags))) in io_msg_ring_data()
86 static struct file *io_msg_grab_file(struct io_kiocb *req, unsigned int issue_flags) in io_msg_grab_file() argument
94 io_ring_submit_lock(ctx, issue_flags); in io_msg_grab_file()
102 io_ring_submit_unlock(ctx, issue_flags); in io_msg_grab_file()
106 static int io_msg_install_complete(struct io_kiocb *req, unsigned int issue_flags) in io_msg_install_complete() argument
113 if (unlikely(io_double_lock_ctx(target_ctx, issue_flags))) in io_msg_install_complete()
138 static int io_msg_send_fd(struct io_kiocb *req, unsigned int issue_flags) in io_msg_send_fd() argument
[all …]
Dkbuf.h37 unsigned int issue_flags);
41 int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags);
44 int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags);
49 unsigned int __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags);
51 void io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags);
87 static inline void io_kbuf_recycle(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle() argument
90 io_kbuf_recycle_legacy(req, issue_flags); in io_kbuf_recycle()
125 unsigned issue_flags) in io_put_kbuf() argument
130 return __io_put_kbuf(req, issue_flags); in io_put_kbuf()
Dopenclose.c96 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument
110 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2()
142 (!resolve_nonblock && (issue_flags & IO_URING_F_NONBLOCK))) in io_openat2()
147 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2()
154 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2()
165 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument
167 return io_openat2(req, issue_flags); in io_openat()
178 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, in __io_close_fixed() argument
183 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed()
185 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed()
[all …]
Dnet.c85 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument
91 if (issue_flags & IO_URING_F_NONBLOCK) in io_shutdown()
110 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
114 if (!req_has_async_data(req) || issue_flags & IO_URING_F_UNLOCKED) in io_netmsg_recycle()
125 unsigned int issue_flags) in io_msg_alloc_async() argument
131 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_msg_alloc_async()
158 unsigned int issue_flags) in io_setup_async_msg() argument
164 async_msg = io_msg_alloc_async(req, issue_flags); in io_setup_async_msg()
218 unsigned int issue_flags) in io_setup_async_addr() argument
225 io = io_msg_alloc_async(req, issue_flags); in io_setup_async_addr()
[all …]
Dfs.h4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags);
8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags);
12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags);
16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags);
19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
During_cmd.c18 unsigned issue_flags = *locked ? 0 : IO_URING_F_UNLOCKED; in io_uring_cmd_work() local
20 ioucmd->task_work_cb(ioucmd, issue_flags); in io_uring_cmd_work()
47 unsigned issue_flags) in io_uring_cmd_done() argument
61 __io_req_complete(req, issue_flags); in io_uring_cmd_done()
106 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags) in io_uring_cmd() argument
121 issue_flags |= IO_URING_F_SQE128; in io_uring_cmd()
123 issue_flags |= IO_URING_F_CQE32; in io_uring_cmd()
127 issue_flags |= IO_URING_F_IOPOLL; in io_uring_cmd()
135 ret = file->f_op->uring_cmd(ioucmd, issue_flags); in io_uring_cmd()
Dopenclose.h3 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags,
7 int io_openat(struct io_kiocb *req, unsigned int issue_flags);
11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags);
14 int io_close(struct io_kiocb *req, unsigned int issue_flags);
Dxattr.h6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags);
9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags);
12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags);
15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
Dpoll.h28 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags);
31 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags);
35 unsigned issue_flags);
36 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
Dkbuf.c53 void io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy() argument
68 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy()
76 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy()
80 unsigned int __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags) in __io_put_kbuf() argument
99 } else if (issue_flags & IO_URING_F_UNLOCKED) { in __io_put_kbuf()
133 unsigned int issue_flags) in io_ring_buffer_select() argument
157 if (issue_flags & IO_URING_F_UNLOCKED || !file_can_poll(req->file)) { in io_ring_buffer_select()
175 unsigned int issue_flags) in io_buffer_select() argument
181 io_ring_submit_lock(req->ctx, issue_flags); in io_buffer_select()
186 ret = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select()
[all …]
Dxattr.c107 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr() argument
112 if (issue_flags & IO_URING_F_NONBLOCK) in io_fgetxattr()
123 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr() argument
130 if (issue_flags & IO_URING_F_NONBLOCK) in io_getxattr()
208 static int __io_setxattr(struct io_kiocb *req, unsigned int issue_flags, in __io_setxattr() argument
223 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fsetxattr() argument
227 if (issue_flags & IO_URING_F_NONBLOCK) in io_fsetxattr()
230 ret = __io_setxattr(req, issue_flags, &req->file->f_path); in io_fsetxattr()
235 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags) in io_setxattr() argument
242 if (issue_flags & IO_URING_F_NONBLOCK) in io_setxattr()
[all …]
Dsync.c38 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument
44 if (issue_flags & IO_URING_F_NONBLOCK) in io_sync_file_range()
68 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync() argument
75 if (issue_flags & IO_URING_F_NONBLOCK) in io_fsync()
97 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate() argument
103 if (issue_flags & IO_URING_F_NONBLOCK) in io_fallocate()
Dsplice.c47 int io_tee(struct io_kiocb *req, unsigned int issue_flags) in io_tee() argument
55 if (issue_flags & IO_URING_F_NONBLOCK) in io_tee()
59 in = io_file_get_fixed(req, sp->splice_fd_in, issue_flags); in io_tee()
88 int io_splice(struct io_kiocb *req, unsigned int issue_flags) in io_splice() argument
97 if (issue_flags & IO_URING_F_NONBLOCK) in io_splice()
101 in = io_file_get_fixed(req, sp->splice_fd_in, issue_flags); in io_splice()
Dsync.h4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags);
7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags);
9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
Dfs.c80 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat() argument
85 if (issue_flags & IO_URING_F_NONBLOCK) in io_renameat()
129 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat() argument
134 if (issue_flags & IO_URING_F_NONBLOCK) in io_unlinkat()
176 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat() argument
181 if (issue_flags & IO_URING_F_NONBLOCK) in io_mkdirat()
226 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat() argument
231 if (issue_flags & IO_URING_F_NONBLOCK) in io_symlinkat()
271 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat() argument
276 if (issue_flags & IO_URING_F_NONBLOCK) in io_linkat()
Dcancel.c82 unsigned issue_flags) in io_try_cancel() argument
97 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel()
132 unsigned int issue_flags) in __io_async_cancel() argument
140 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel()
149 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel()
161 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel()
165 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument
182 issue_flags); in io_async_cancel()
193 ret = __io_async_cancel(&cd, tctx, issue_flags); in io_async_cancel()
Dadvise.c48 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument
54 if (issue_flags & IO_URING_F_NONBLOCK) in io_madvise()
78 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument
83 if (issue_flags & IO_URING_F_NONBLOCK) { in io_fadvise()
Dadvise.h4 int io_madvise(struct io_kiocb *req, unsigned int issue_flags);
7 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags);
Dsplice.h4 int io_tee(struct io_kiocb *req, unsigned int issue_flags);
7 int io_splice(struct io_kiocb *req, unsigned int issue_flags);
Dpoll.c585 unsigned issue_flags) in __io_arm_poll_handler() argument
610 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler()
614 if (issue_flags & IO_URING_F_UNLOCKED) in __io_arm_poll_handler()
681 unsigned issue_flags) in io_req_alloc_apoll() argument
690 } else if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_req_alloc_apoll()
710 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags) in io_arm_poll_handler() argument
743 apoll = io_req_alloc_apoll(req, issue_flags); in io_arm_poll_handler()
750 io_kbuf_recycle(req, issue_flags); in io_arm_poll_handler()
752 ret = __io_arm_poll_handler(req, &apoll->poll, &ipt, mask, issue_flags); in io_arm_poll_handler()
888 unsigned issue_flags) in io_poll_cancel() argument
[all …]
Dcancel.h17 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags);
20 unsigned int issue_flags);
Drw.h19 int io_read(struct io_kiocb *req, unsigned int issue_flags);
21 int io_write(struct io_kiocb *req, unsigned int issue_flags);
Drsrc.h136 unsigned int issue_flags) in io_req_set_rsrc_node() argument
139 io_ring_submit_lock(ctx, issue_flags); in io_req_set_rsrc_node()
145 io_ring_submit_unlock(ctx, issue_flags); in io_req_set_rsrc_node()
157 int io_files_update(struct io_kiocb *req, unsigned int issue_flags);
Dtimeout.h34 int io_timeout(struct io_kiocb *req, unsigned int issue_flags);
36 int io_timeout_remove(struct io_kiocb *req, unsigned int issue_flags);

12