Commit bfe76559 authored by Pavel Begunkov's avatar Pavel Begunkov Committed by Jens Axboe

io_uring: decouple issuing and req preparation

io_issue_sqe() does two things at once, trying to prepare request and
issuing them. Split it in two and deduplicate with io_defer_prep().
Signed-off-by: default avatarPavel Begunkov <asml.silence@gmail.com>
Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent 73debe68
...@@ -5590,120 +5590,95 @@ static int io_files_update(struct io_kiocb *req, bool force_nonblock, ...@@ -5590,120 +5590,95 @@ static int io_files_update(struct io_kiocb *req, bool force_nonblock,
return 0; return 0;
} }
static int io_req_defer_prep(struct io_kiocb *req, static int io_req_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
const struct io_uring_sqe *sqe)
{ {
ssize_t ret = 0;
if (!sqe)
return 0;
if (io_alloc_async_data(req))
return -EAGAIN;
ret = io_prep_work_files(req);
if (unlikely(ret))
return ret;
io_prep_async_work(req);
switch (req->opcode) { switch (req->opcode) {
case IORING_OP_NOP: case IORING_OP_NOP:
break; return 0;
case IORING_OP_READV: case IORING_OP_READV:
case IORING_OP_READ_FIXED: case IORING_OP_READ_FIXED:
case IORING_OP_READ: case IORING_OP_READ:
ret = io_read_prep(req, sqe); return io_read_prep(req, sqe);
break;
case IORING_OP_WRITEV: case IORING_OP_WRITEV:
case IORING_OP_WRITE_FIXED: case IORING_OP_WRITE_FIXED:
case IORING_OP_WRITE: case IORING_OP_WRITE:
ret = io_write_prep(req, sqe); return io_write_prep(req, sqe);
break;
case IORING_OP_POLL_ADD: case IORING_OP_POLL_ADD:
ret = io_poll_add_prep(req, sqe); return io_poll_add_prep(req, sqe);
break;
case IORING_OP_POLL_REMOVE: case IORING_OP_POLL_REMOVE:
ret = io_poll_remove_prep(req, sqe); return io_poll_remove_prep(req, sqe);
break;
case IORING_OP_FSYNC: case IORING_OP_FSYNC:
ret = io_prep_fsync(req, sqe); return io_prep_fsync(req, sqe);
break;
case IORING_OP_SYNC_FILE_RANGE: case IORING_OP_SYNC_FILE_RANGE:
ret = io_prep_sfr(req, sqe); return io_prep_sfr(req, sqe);
break;
case IORING_OP_SENDMSG: case IORING_OP_SENDMSG:
case IORING_OP_SEND: case IORING_OP_SEND:
ret = io_sendmsg_prep(req, sqe); return io_sendmsg_prep(req, sqe);
break;
case IORING_OP_RECVMSG: case IORING_OP_RECVMSG:
case IORING_OP_RECV: case IORING_OP_RECV:
ret = io_recvmsg_prep(req, sqe); return io_recvmsg_prep(req, sqe);
break;
case IORING_OP_CONNECT: case IORING_OP_CONNECT:
ret = io_connect_prep(req, sqe); return io_connect_prep(req, sqe);
break;
case IORING_OP_TIMEOUT: case IORING_OP_TIMEOUT:
ret = io_timeout_prep(req, sqe, false); return io_timeout_prep(req, sqe, false);
break;
case IORING_OP_TIMEOUT_REMOVE: case IORING_OP_TIMEOUT_REMOVE:
ret = io_timeout_remove_prep(req, sqe); return io_timeout_remove_prep(req, sqe);
break;
case IORING_OP_ASYNC_CANCEL: case IORING_OP_ASYNC_CANCEL:
ret = io_async_cancel_prep(req, sqe); return io_async_cancel_prep(req, sqe);
break;
case IORING_OP_LINK_TIMEOUT: case IORING_OP_LINK_TIMEOUT:
ret = io_timeout_prep(req, sqe, true); return io_timeout_prep(req, sqe, true);
break;
case IORING_OP_ACCEPT: case IORING_OP_ACCEPT:
ret = io_accept_prep(req, sqe); return io_accept_prep(req, sqe);
break;
case IORING_OP_FALLOCATE: case IORING_OP_FALLOCATE:
ret = io_fallocate_prep(req, sqe); return io_fallocate_prep(req, sqe);
break;
case IORING_OP_OPENAT: case IORING_OP_OPENAT:
ret = io_openat_prep(req, sqe); return io_openat_prep(req, sqe);
break;
case IORING_OP_CLOSE: case IORING_OP_CLOSE:
ret = io_close_prep(req, sqe); return io_close_prep(req, sqe);
break;
case IORING_OP_FILES_UPDATE: case IORING_OP_FILES_UPDATE:
ret = io_files_update_prep(req, sqe); return io_files_update_prep(req, sqe);
break;
case IORING_OP_STATX: case IORING_OP_STATX:
ret = io_statx_prep(req, sqe); return io_statx_prep(req, sqe);
break;
case IORING_OP_FADVISE: case IORING_OP_FADVISE:
ret = io_fadvise_prep(req, sqe); return io_fadvise_prep(req, sqe);
break;
case IORING_OP_MADVISE: case IORING_OP_MADVISE:
ret = io_madvise_prep(req, sqe); return io_madvise_prep(req, sqe);
break;
case IORING_OP_OPENAT2: case IORING_OP_OPENAT2:
ret = io_openat2_prep(req, sqe); return io_openat2_prep(req, sqe);
break;
case IORING_OP_EPOLL_CTL: case IORING_OP_EPOLL_CTL:
ret = io_epoll_ctl_prep(req, sqe); return io_epoll_ctl_prep(req, sqe);
break;
case IORING_OP_SPLICE: case IORING_OP_SPLICE:
ret = io_splice_prep(req, sqe); return io_splice_prep(req, sqe);
break;
case IORING_OP_PROVIDE_BUFFERS: case IORING_OP_PROVIDE_BUFFERS:
ret = io_provide_buffers_prep(req, sqe); return io_provide_buffers_prep(req, sqe);
break;
case IORING_OP_REMOVE_BUFFERS: case IORING_OP_REMOVE_BUFFERS:
ret = io_remove_buffers_prep(req, sqe); return io_remove_buffers_prep(req, sqe);
break;
case IORING_OP_TEE: case IORING_OP_TEE:
ret = io_tee_prep(req, sqe); return io_tee_prep(req, sqe);
break;
default:
printk_once(KERN_WARNING "io_uring: unhandled opcode %d\n",
req->opcode);
ret = -EINVAL;
break;
} }
return ret; printk_once(KERN_WARNING "io_uring: unhandled opcode %d\n",
req->opcode);
return-EINVAL;
}
static int io_req_defer_prep(struct io_kiocb *req,
const struct io_uring_sqe *sqe)
{
int ret;
if (!sqe)
return 0;
if (io_alloc_async_data(req))
return -EAGAIN;
ret = io_prep_work_files(req);
if (unlikely(ret))
return ret;
io_prep_async_work(req);
return io_req_prep(req, sqe);
} }
static u32 io_get_sequence(struct io_kiocb *req) static u32 io_get_sequence(struct io_kiocb *req)
...@@ -5840,6 +5815,12 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe, ...@@ -5840,6 +5815,12 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe,
struct io_ring_ctx *ctx = req->ctx; struct io_ring_ctx *ctx = req->ctx;
int ret; int ret;
if (sqe) {
ret = io_req_prep(req, sqe);
if (unlikely(ret < 0))
return ret;
}
switch (req->opcode) { switch (req->opcode) {
case IORING_OP_NOP: case IORING_OP_NOP:
ret = io_nop(req, cs); ret = io_nop(req, cs);
...@@ -5847,62 +5828,27 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe, ...@@ -5847,62 +5828,27 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe,
case IORING_OP_READV: case IORING_OP_READV:
case IORING_OP_READ_FIXED: case IORING_OP_READ_FIXED:
case IORING_OP_READ: case IORING_OP_READ:
if (sqe) {
ret = io_read_prep(req, sqe);
if (ret < 0)
break;
}
ret = io_read(req, force_nonblock, cs); ret = io_read(req, force_nonblock, cs);
break; break;
case IORING_OP_WRITEV: case IORING_OP_WRITEV:
case IORING_OP_WRITE_FIXED: case IORING_OP_WRITE_FIXED:
case IORING_OP_WRITE: case IORING_OP_WRITE:
if (sqe) {
ret = io_write_prep(req, sqe);
if (ret < 0)
break;
}
ret = io_write(req, force_nonblock, cs); ret = io_write(req, force_nonblock, cs);
break; break;
case IORING_OP_FSYNC: case IORING_OP_FSYNC:
if (sqe) {
ret = io_prep_fsync(req, sqe);
if (ret < 0)
break;
}
ret = io_fsync(req, force_nonblock); ret = io_fsync(req, force_nonblock);
break; break;
case IORING_OP_POLL_ADD: case IORING_OP_POLL_ADD:
if (sqe) {
ret = io_poll_add_prep(req, sqe);
if (ret)
break;
}
ret = io_poll_add(req); ret = io_poll_add(req);
break; break;
case IORING_OP_POLL_REMOVE: case IORING_OP_POLL_REMOVE:
if (sqe) {
ret = io_poll_remove_prep(req, sqe);
if (ret < 0)
break;
}
ret = io_poll_remove(req); ret = io_poll_remove(req);
break; break;
case IORING_OP_SYNC_FILE_RANGE: case IORING_OP_SYNC_FILE_RANGE:
if (sqe) {
ret = io_prep_sfr(req, sqe);
if (ret < 0)
break;
}
ret = io_sync_file_range(req, force_nonblock); ret = io_sync_file_range(req, force_nonblock);
break; break;
case IORING_OP_SENDMSG: case IORING_OP_SENDMSG:
case IORING_OP_SEND: case IORING_OP_SEND:
if (sqe) {
ret = io_sendmsg_prep(req, sqe);
if (ret < 0)
break;
}
if (req->opcode == IORING_OP_SENDMSG) if (req->opcode == IORING_OP_SENDMSG)
ret = io_sendmsg(req, force_nonblock, cs); ret = io_sendmsg(req, force_nonblock, cs);
else else
...@@ -5910,158 +5856,63 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe, ...@@ -5910,158 +5856,63 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe,
break; break;
case IORING_OP_RECVMSG: case IORING_OP_RECVMSG:
case IORING_OP_RECV: case IORING_OP_RECV:
if (sqe) {
ret = io_recvmsg_prep(req, sqe);
if (ret)
break;
}
if (req->opcode == IORING_OP_RECVMSG) if (req->opcode == IORING_OP_RECVMSG)
ret = io_recvmsg(req, force_nonblock, cs); ret = io_recvmsg(req, force_nonblock, cs);
else else
ret = io_recv(req, force_nonblock, cs); ret = io_recv(req, force_nonblock, cs);
break; break;
case IORING_OP_TIMEOUT: case IORING_OP_TIMEOUT:
if (sqe) {
ret = io_timeout_prep(req, sqe, false);
if (ret)
break;
}
ret = io_timeout(req); ret = io_timeout(req);
break; break;
case IORING_OP_TIMEOUT_REMOVE: case IORING_OP_TIMEOUT_REMOVE:
if (sqe) {
ret = io_timeout_remove_prep(req, sqe);
if (ret)
break;
}
ret = io_timeout_remove(req); ret = io_timeout_remove(req);
break; break;
case IORING_OP_ACCEPT: case IORING_OP_ACCEPT:
if (sqe) {
ret = io_accept_prep(req, sqe);
if (ret)
break;
}
ret = io_accept(req, force_nonblock, cs); ret = io_accept(req, force_nonblock, cs);
break; break;
case IORING_OP_CONNECT: case IORING_OP_CONNECT:
if (sqe) {
ret = io_connect_prep(req, sqe);
if (ret)
break;
}
ret = io_connect(req, force_nonblock, cs); ret = io_connect(req, force_nonblock, cs);
break; break;
case IORING_OP_ASYNC_CANCEL: case IORING_OP_ASYNC_CANCEL:
if (sqe) {
ret = io_async_cancel_prep(req, sqe);
if (ret)
break;
}
ret = io_async_cancel(req); ret = io_async_cancel(req);
break; break;
case IORING_OP_FALLOCATE: case IORING_OP_FALLOCATE:
if (sqe) {
ret = io_fallocate_prep(req, sqe);
if (ret)
break;
}
ret = io_fallocate(req, force_nonblock); ret = io_fallocate(req, force_nonblock);
break; break;
case IORING_OP_OPENAT: case IORING_OP_OPENAT:
if (sqe) {
ret = io_openat_prep(req, sqe);
if (ret)
break;
}
ret = io_openat(req, force_nonblock); ret = io_openat(req, force_nonblock);
break; break;
case IORING_OP_CLOSE: case IORING_OP_CLOSE:
if (sqe) {
ret = io_close_prep(req, sqe);
if (ret)
break;
}
ret = io_close(req, force_nonblock, cs); ret = io_close(req, force_nonblock, cs);
break; break;
case IORING_OP_FILES_UPDATE: case IORING_OP_FILES_UPDATE:
if (sqe) {
ret = io_files_update_prep(req, sqe);
if (ret)
break;
}
ret = io_files_update(req, force_nonblock, cs); ret = io_files_update(req, force_nonblock, cs);
break; break;
case IORING_OP_STATX: case IORING_OP_STATX:
if (sqe) {
ret = io_statx_prep(req, sqe);
if (ret)
break;
}
ret = io_statx(req, force_nonblock); ret = io_statx(req, force_nonblock);
break; break;
case IORING_OP_FADVISE: case IORING_OP_FADVISE:
if (sqe) {
ret = io_fadvise_prep(req, sqe);
if (ret)
break;
}
ret = io_fadvise(req, force_nonblock); ret = io_fadvise(req, force_nonblock);
break; break;
case IORING_OP_MADVISE: case IORING_OP_MADVISE:
if (sqe) {
ret = io_madvise_prep(req, sqe);
if (ret)
break;
}
ret = io_madvise(req, force_nonblock); ret = io_madvise(req, force_nonblock);
break; break;
case IORING_OP_OPENAT2: case IORING_OP_OPENAT2:
if (sqe) {
ret = io_openat2_prep(req, sqe);
if (ret)
break;
}
ret = io_openat2(req, force_nonblock); ret = io_openat2(req, force_nonblock);
break; break;
case IORING_OP_EPOLL_CTL: case IORING_OP_EPOLL_CTL:
if (sqe) {
ret = io_epoll_ctl_prep(req, sqe);
if (ret)
break;
}
ret = io_epoll_ctl(req, force_nonblock, cs); ret = io_epoll_ctl(req, force_nonblock, cs);
break; break;
case IORING_OP_SPLICE: case IORING_OP_SPLICE:
if (sqe) {
ret = io_splice_prep(req, sqe);
if (ret < 0)
break;
}
ret = io_splice(req, force_nonblock); ret = io_splice(req, force_nonblock);
break; break;
case IORING_OP_PROVIDE_BUFFERS: case IORING_OP_PROVIDE_BUFFERS:
if (sqe) {
ret = io_provide_buffers_prep(req, sqe);
if (ret)
break;
}
ret = io_provide_buffers(req, force_nonblock, cs); ret = io_provide_buffers(req, force_nonblock, cs);
break; break;
case IORING_OP_REMOVE_BUFFERS: case IORING_OP_REMOVE_BUFFERS:
if (sqe) {
ret = io_remove_buffers_prep(req, sqe);
if (ret)
break;
}
ret = io_remove_buffers(req, force_nonblock, cs); ret = io_remove_buffers(req, force_nonblock, cs);
break; break;
case IORING_OP_TEE: case IORING_OP_TEE:
if (sqe) {
ret = io_tee_prep(req, sqe);
if (ret < 0)
break;
}
ret = io_tee(req, force_nonblock); ret = io_tee(req, force_nonblock);
break; break;
default: default:
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment