Lines Matching refs:issue_flags
82 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument
88 if (issue_flags & IO_URING_F_NONBLOCK) in io_shutdown()
107 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
111 if (!req_has_async_data(req) || issue_flags & IO_URING_F_UNLOCKED) in io_netmsg_recycle()
122 unsigned int issue_flags) in io_msg_alloc_async() argument
128 if (!(issue_flags & IO_URING_F_UNLOCKED) && in io_msg_alloc_async()
153 unsigned int issue_flags) in io_setup_async_msg() argument
159 async_msg = io_msg_alloc_async(req, issue_flags); in io_setup_async_msg()
205 unsigned int issue_flags) in io_setup_async_addr() argument
212 io = io_msg_alloc_async(req, issue_flags); in io_setup_async_addr()
268 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg() argument
292 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
295 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg()
303 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg()
304 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
308 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
318 io_netmsg_recycle(req, issue_flags); in io_sendmsg()
327 int io_send(struct io_kiocb *req, unsigned int issue_flags) in io_send() argument
360 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
371 if (issue_flags & IO_URING_F_NONBLOCK) in io_send()
379 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_send()
380 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
387 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
593 unsigned issue_flags) in io_recv_finish() argument
616 if (issue_flags & IO_URING_F_MULTISHOT) in io_recv_finish()
707 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags) in io_recvmsg() argument
715 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_recvmsg()
733 return io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
740 buf = io_buffer_select(req, &len, issue_flags); in io_recvmsg()
747 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
774 ret = io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
775 if (ret == -EAGAIN && (issue_flags & IO_URING_F_MULTISHOT)) { in io_recvmsg()
776 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
784 return io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
798 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
800 cflags = io_put_kbuf(req, issue_flags); in io_recvmsg()
804 if (!io_recv_finish(req, &ret, cflags, mshot_finished, issue_flags)) in io_recvmsg()
808 io_netmsg_recycle(req, issue_flags); in io_recvmsg()
818 int io_recv(struct io_kiocb *req, unsigned int issue_flags) in io_recv() argument
827 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_recv()
842 buf = io_buffer_select(req, &len, issue_flags); in io_recv()
870 if (issue_flags & IO_URING_F_MULTISHOT) { in io_recv()
871 io_kbuf_recycle(req, issue_flags); in io_recv()
897 io_kbuf_recycle(req, issue_flags); in io_recv()
899 cflags = io_put_kbuf(req, issue_flags); in io_recv()
903 if (!io_recv_finish(req, &ret, cflags, ret <= 0, issue_flags)) in io_recv()
1044 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags) in io_send_zc() argument
1081 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1101 if (issue_flags & IO_URING_F_NONBLOCK) in io_send_zc()
1111 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_send_zc()
1112 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1119 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1135 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_send_zc()
1143 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg_zc() argument
1168 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1171 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg_zc()
1181 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg_zc()
1182 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1187 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1199 io_netmsg_recycle(req, issue_flags); in io_sendmsg_zc()
1209 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_sendmsg_zc()
1262 int io_accept(struct io_kiocb *req, unsigned int issue_flags) in io_accept() argument
1266 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_accept()
1290 if (issue_flags & IO_URING_F_MULTISHOT) in io_accept()
1301 ret = io_fixed_fd_install(req, issue_flags, file, in io_accept()
1315 return (issue_flags & IO_URING_F_MULTISHOT) ? IOU_STOP_MULTISHOT : IOU_OK; in io_accept()
1339 int io_socket(struct io_kiocb *req, unsigned int issue_flags) in io_socket() argument
1356 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_socket()
1365 ret = io_fixed_fd_install(req, issue_flags, file, in io_socket()
1393 int io_connect(struct io_kiocb *req, unsigned int issue_flags) in io_connect() argument
1399 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_connect()