Lines Matching refs:issue_flags
85 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument
91 if (issue_flags & IO_URING_F_NONBLOCK) in io_shutdown()
110 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
114 if (!req_has_async_data(req) || issue_flags & IO_URING_F_UNLOCKED) in io_netmsg_recycle()
125 unsigned int issue_flags) in io_msg_alloc_async() argument
131 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_msg_alloc_async()
158 unsigned int issue_flags) in io_setup_async_msg() argument
164 async_msg = io_msg_alloc_async(req, issue_flags); in io_setup_async_msg()
218 unsigned int issue_flags) in io_setup_async_addr() argument
225 io = io_msg_alloc_async(req, issue_flags); in io_setup_async_addr()
281 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg() argument
306 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
309 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg()
317 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg()
318 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
324 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
334 io_netmsg_recycle(req, issue_flags); in io_sendmsg()
343 int io_send(struct io_kiocb *req, unsigned int issue_flags) in io_send() argument
376 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
387 if (issue_flags & IO_URING_F_NONBLOCK) in io_send()
396 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_send()
397 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
404 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
622 unsigned issue_flags) in io_recv_finish() argument
645 if (issue_flags & IO_URING_F_MULTISHOT) in io_recv_finish()
736 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags) in io_recvmsg() argument
744 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_recvmsg()
762 return io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
769 buf = io_buffer_select(req, &len, issue_flags); in io_recvmsg()
776 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
806 ret = io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
807 if (ret == -EAGAIN && (issue_flags & IO_URING_F_MULTISHOT)) { in io_recvmsg()
808 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
816 return io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
830 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
832 cflags = io_put_kbuf(req, issue_flags); in io_recvmsg()
836 if (!io_recv_finish(req, &ret, cflags, mshot_finished, issue_flags)) in io_recvmsg()
843 io_netmsg_recycle(req, issue_flags); in io_recvmsg()
850 int io_recv(struct io_kiocb *req, unsigned int issue_flags) in io_recv() argument
859 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_recv()
874 buf = io_buffer_select(req, &len, issue_flags); in io_recv()
903 if (issue_flags & IO_URING_F_MULTISHOT) { in io_recv()
904 io_kbuf_recycle(req, issue_flags); in io_recv()
930 io_kbuf_recycle(req, issue_flags); in io_recv()
932 cflags = io_put_kbuf(req, issue_flags); in io_recv()
936 if (!io_recv_finish(req, &ret, cflags, ret <= 0, issue_flags)) in io_recv()
1081 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags) in io_send_zc() argument
1118 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1138 if (issue_flags & IO_URING_F_NONBLOCK) in io_send_zc()
1149 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_send_zc()
1150 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1157 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1173 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_send_zc()
1181 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg_zc() argument
1206 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1209 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg_zc()
1219 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg_zc()
1220 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1225 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1237 io_netmsg_recycle(req, issue_flags); in io_sendmsg_zc()
1247 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_sendmsg_zc()
1300 int io_accept(struct io_kiocb *req, unsigned int issue_flags) in io_accept() argument
1304 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_accept()
1328 if (issue_flags & IO_URING_F_MULTISHOT) in io_accept()
1339 ret = io_fixed_fd_install(req, issue_flags, file, in io_accept()
1378 int io_socket(struct io_kiocb *req, unsigned int issue_flags) in io_socket() argument
1395 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_socket()
1404 ret = io_fixed_fd_install(req, issue_flags, file, in io_socket()
1432 int io_connect(struct io_kiocb *req, unsigned int issue_flags) in io_connect() argument
1438 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_connect()