io_uring: run linked timeouts from task_work
This is in preparation to making the completion lock work outside of hard/soft IRQ context. Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
89850fce16
commit
89b263f6d5
@ -538,6 +538,8 @@ struct io_timeout {
|
|||||||
struct list_head list;
|
struct list_head list;
|
||||||
/* head of the link, used by linked timeouts only */
|
/* head of the link, used by linked timeouts only */
|
||||||
struct io_kiocb *head;
|
struct io_kiocb *head;
|
||||||
|
/* for linked completions */
|
||||||
|
struct io_kiocb *prev;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct io_timeout_rem {
|
struct io_timeout_rem {
|
||||||
@ -1848,6 +1850,7 @@ static inline void io_remove_next_linked(struct io_kiocb *req)
|
|||||||
|
|
||||||
static bool io_kill_linked_timeout(struct io_kiocb *req)
|
static bool io_kill_linked_timeout(struct io_kiocb *req)
|
||||||
__must_hold(&req->ctx->completion_lock)
|
__must_hold(&req->ctx->completion_lock)
|
||||||
|
__must_hold(&req->ctx->timeout_lock)
|
||||||
{
|
{
|
||||||
struct io_kiocb *link = req->link;
|
struct io_kiocb *link = req->link;
|
||||||
|
|
||||||
@ -1892,8 +1895,13 @@ static bool io_disarm_next(struct io_kiocb *req)
|
|||||||
{
|
{
|
||||||
bool posted = false;
|
bool posted = false;
|
||||||
|
|
||||||
if (likely(req->flags & REQ_F_LINK_TIMEOUT))
|
if (likely(req->flags & REQ_F_LINK_TIMEOUT)) {
|
||||||
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
|
|
||||||
|
spin_lock_irq(&ctx->timeout_lock);
|
||||||
posted = io_kill_linked_timeout(req);
|
posted = io_kill_linked_timeout(req);
|
||||||
|
spin_unlock_irq(&ctx->timeout_lock);
|
||||||
|
}
|
||||||
if (unlikely((req->flags & REQ_F_FAIL) &&
|
if (unlikely((req->flags & REQ_F_FAIL) &&
|
||||||
!(req->flags & REQ_F_HARDLINK))) {
|
!(req->flags & REQ_F_HARDLINK))) {
|
||||||
posted |= (req->link != NULL);
|
posted |= (req->link != NULL);
|
||||||
@ -6359,6 +6367,20 @@ static inline struct file *io_file_get(struct io_ring_ctx *ctx,
|
|||||||
return io_file_get_normal(ctx, req, fd);
|
return io_file_get_normal(ctx, req, fd);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void io_req_task_link_timeout(struct io_kiocb *req)
|
||||||
|
{
|
||||||
|
struct io_kiocb *prev = req->timeout.prev;
|
||||||
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
|
|
||||||
|
if (prev) {
|
||||||
|
io_async_find_and_cancel(ctx, req, prev->user_data, -ETIME);
|
||||||
|
io_put_req(prev);
|
||||||
|
io_put_req(req);
|
||||||
|
} else {
|
||||||
|
io_req_complete_post(req, -ETIME, 0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static enum hrtimer_restart io_link_timeout_fn(struct hrtimer *timer)
|
static enum hrtimer_restart io_link_timeout_fn(struct hrtimer *timer)
|
||||||
{
|
{
|
||||||
struct io_timeout_data *data = container_of(timer,
|
struct io_timeout_data *data = container_of(timer,
|
||||||
@ -6367,7 +6389,7 @@ static enum hrtimer_restart io_link_timeout_fn(struct hrtimer *timer)
|
|||||||
struct io_ring_ctx *ctx = req->ctx;
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
spin_lock_irqsave(&ctx->completion_lock, flags);
|
spin_lock_irqsave(&ctx->timeout_lock, flags);
|
||||||
prev = req->timeout.head;
|
prev = req->timeout.head;
|
||||||
req->timeout.head = NULL;
|
req->timeout.head = NULL;
|
||||||
|
|
||||||
@ -6380,15 +6402,11 @@ static enum hrtimer_restart io_link_timeout_fn(struct hrtimer *timer)
|
|||||||
if (!req_ref_inc_not_zero(prev))
|
if (!req_ref_inc_not_zero(prev))
|
||||||
prev = NULL;
|
prev = NULL;
|
||||||
}
|
}
|
||||||
spin_unlock_irqrestore(&ctx->completion_lock, flags);
|
req->timeout.prev = prev;
|
||||||
|
spin_unlock_irqrestore(&ctx->timeout_lock, flags);
|
||||||
|
|
||||||
if (prev) {
|
req->io_task_work.func = io_req_task_link_timeout;
|
||||||
io_async_find_and_cancel(ctx, req, prev->user_data, -ETIME);
|
io_req_task_work_add(req);
|
||||||
io_put_req_deferred(prev, 1);
|
|
||||||
io_put_req_deferred(req, 1);
|
|
||||||
} else {
|
|
||||||
io_req_complete_post(req, -ETIME, 0);
|
|
||||||
}
|
|
||||||
return HRTIMER_NORESTART;
|
return HRTIMER_NORESTART;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -6396,7 +6414,7 @@ static void io_queue_linked_timeout(struct io_kiocb *req)
|
|||||||
{
|
{
|
||||||
struct io_ring_ctx *ctx = req->ctx;
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
|
|
||||||
spin_lock_irq(&ctx->completion_lock);
|
spin_lock_irq(&ctx->timeout_lock);
|
||||||
/*
|
/*
|
||||||
* If the back reference is NULL, then our linked request finished
|
* If the back reference is NULL, then our linked request finished
|
||||||
* before we got a chance to setup the timer
|
* before we got a chance to setup the timer
|
||||||
@ -6408,7 +6426,7 @@ static void io_queue_linked_timeout(struct io_kiocb *req)
|
|||||||
hrtimer_start(&data->timer, timespec64_to_ktime(data->ts),
|
hrtimer_start(&data->timer, timespec64_to_ktime(data->ts),
|
||||||
data->mode);
|
data->mode);
|
||||||
}
|
}
|
||||||
spin_unlock_irq(&ctx->completion_lock);
|
spin_unlock_irq(&ctx->timeout_lock);
|
||||||
/* drop submission reference */
|
/* drop submission reference */
|
||||||
io_put_req(req);
|
io_put_req(req);
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user