io_uring: implement multishot mode for accept
Refactor io_accept() to support multishot mode. theoretical analysis: 1) when connections come in fast - singleshot: add accept sqe(userspace) --> accept inline ^ | |-----------------| - multishot: add accept sqe(userspace) --> accept inline ^ | |--*--| we do accept repeatedly in * place until get EAGAIN 2) when connections come in at a low pressure similar thing like 1), we reduce a lot of userspace-kernel context switch and useless vfs_poll() tests: Did some tests, which goes in this way: server client(multiple) accept connect read write write read close close Basically, raise up a number of clients(on same machine with server) to connect to the server, and then write some data to it, the server will write those data back to the client after it receives them, and then close the connection after write return. Then the client will read the data and then close the connection. Here I test 10000 clients connect one server, data size 128 bytes. And each client has a go routine for it, so they come to the server in short time. test 20 times before/after this patchset, time spent:(unit cycle, which is the return value of clock()) before: 1930136+1940725+1907981+1947601+1923812+1928226+1911087+1905897+1941075 +1934374+1906614+1912504+1949110+1908790+1909951+1941672+1969525+1934984 +1934226+1914385)/20.0 = 1927633.75 after: 1858905+1917104+1895455+1963963+1892706+1889208+1874175+1904753+1874112 +1874985+1882706+1884642+1864694+1906508+1916150+1924250+1869060+1889506 +1871324+1940803)/20.0 = 1894750.45 (1927633.75 - 1894750.45) / 1927633.75 = 1.65% Signed-off-by: Hao Xu <howeyxu@tencent.com> Link: https://lore.kernel.org/r/20220514142046.58072-5-haoxu.linux@gmail.com Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
dbc2564cfe
commit
4e86a2c980
@ -1148,6 +1148,7 @@ static const struct io_op_def io_op_defs[] = {
|
||||
.unbound_nonreg_file = 1,
|
||||
.pollin = 1,
|
||||
.poll_exclusive = 1,
|
||||
.ioprio = 1, /* used for flags */
|
||||
},
|
||||
[IORING_OP_ASYNC_CANCEL] = {
|
||||
.audit_skip = 1,
|
||||
@ -5762,6 +5763,7 @@ out_free:
|
||||
static int io_accept_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
||||
{
|
||||
struct io_accept *accept = &req->accept;
|
||||
unsigned flags;
|
||||
|
||||
if (sqe->len || sqe->buf_index)
|
||||
return -EINVAL;
|
||||
@ -5770,19 +5772,30 @@ static int io_accept_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
||||
accept->addr_len = u64_to_user_ptr(READ_ONCE(sqe->addr2));
|
||||
accept->flags = READ_ONCE(sqe->accept_flags);
|
||||
accept->nofile = rlimit(RLIMIT_NOFILE);
|
||||
flags = READ_ONCE(sqe->ioprio);
|
||||
if (flags & ~IORING_ACCEPT_MULTISHOT)
|
||||
return -EINVAL;
|
||||
|
||||
accept->file_slot = READ_ONCE(sqe->file_index);
|
||||
if (accept->file_slot && (accept->flags & SOCK_CLOEXEC))
|
||||
return -EINVAL;
|
||||
if (accept->file_slot) {
|
||||
if (accept->flags & SOCK_CLOEXEC)
|
||||
return -EINVAL;
|
||||
if (flags & IORING_ACCEPT_MULTISHOT &&
|
||||
accept->file_slot != IORING_FILE_INDEX_ALLOC)
|
||||
return -EINVAL;
|
||||
}
|
||||
if (accept->flags & ~(SOCK_CLOEXEC | SOCK_NONBLOCK))
|
||||
return -EINVAL;
|
||||
if (SOCK_NONBLOCK != O_NONBLOCK && (accept->flags & SOCK_NONBLOCK))
|
||||
accept->flags = (accept->flags & ~SOCK_NONBLOCK) | O_NONBLOCK;
|
||||
if (flags & IORING_ACCEPT_MULTISHOT)
|
||||
req->flags |= REQ_F_APOLL_MULTISHOT;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int io_accept(struct io_kiocb *req, unsigned int issue_flags)
|
||||
{
|
||||
struct io_ring_ctx *ctx = req->ctx;
|
||||
struct io_accept *accept = &req->accept;
|
||||
bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK;
|
||||
unsigned int file_flags = force_nonblock ? O_NONBLOCK : 0;
|
||||
@ -5790,6 +5803,7 @@ static int io_accept(struct io_kiocb *req, unsigned int issue_flags)
|
||||
struct file *file;
|
||||
int ret, fd;
|
||||
|
||||
retry:
|
||||
if (!fixed) {
|
||||
fd = __get_unused_fd_flags(accept->flags, accept->nofile);
|
||||
if (unlikely(fd < 0))
|
||||
@ -5801,8 +5815,17 @@ static int io_accept(struct io_kiocb *req, unsigned int issue_flags)
|
||||
if (!fixed)
|
||||
put_unused_fd(fd);
|
||||
ret = PTR_ERR(file);
|
||||
if (ret == -EAGAIN && force_nonblock)
|
||||
return -EAGAIN;
|
||||
if (ret == -EAGAIN && force_nonblock) {
|
||||
/*
|
||||
* if it's multishot and polled, we don't need to
|
||||
* return EAGAIN to arm the poll infra since it
|
||||
* has already been done
|
||||
*/
|
||||
if ((req->flags & IO_APOLL_MULTI_POLLED) ==
|
||||
IO_APOLL_MULTI_POLLED)
|
||||
ret = 0;
|
||||
return ret;
|
||||
}
|
||||
if (ret == -ERESTARTSYS)
|
||||
ret = -EINTR;
|
||||
req_set_fail(req);
|
||||
@ -5813,8 +5836,27 @@ static int io_accept(struct io_kiocb *req, unsigned int issue_flags)
|
||||
ret = io_fixed_fd_install(req, issue_flags, file,
|
||||
accept->file_slot);
|
||||
}
|
||||
__io_req_complete(req, issue_flags, ret, 0);
|
||||
return 0;
|
||||
|
||||
if (!(req->flags & REQ_F_APOLL_MULTISHOT)) {
|
||||
__io_req_complete(req, issue_flags, ret, 0);
|
||||
return 0;
|
||||
}
|
||||
if (ret >= 0) {
|
||||
bool filled;
|
||||
|
||||
spin_lock(&ctx->completion_lock);
|
||||
filled = io_fill_cqe_aux(ctx, req->cqe.user_data, ret,
|
||||
IORING_CQE_F_MORE);
|
||||
io_commit_cqring(ctx);
|
||||
spin_unlock(&ctx->completion_lock);
|
||||
if (filled) {
|
||||
io_cqring_ev_posted(ctx);
|
||||
goto retry;
|
||||
}
|
||||
ret = -ECANCELED;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int io_connect_prep_async(struct io_kiocb *req)
|
||||
|
Loading…
Reference in New Issue
Block a user