@@ -20,6 +20,7 @@
#include "io-wq.h"
#include "slist.h"
#include "io_uring.h"
+#include "tctx.h"
#define WORKER_IDLE_TIMEOUT (5 * HZ)
@@ -617,19 +618,12 @@ static void io_worker_handle_work(struct io_worker *worker)
} while (1);
}
-static int io_wqe_worker(void *data)
+static void io_wqe_worker_normal(struct io_worker *worker)
{
- struct io_worker *worker = data;
struct io_wqe_acct *acct = io_wqe_get_acct(worker);
struct io_wqe *wqe = worker->wqe;
struct io_wq *wq = wqe->wq;
bool last_timeout = false;
- char buf[TASK_COMM_LEN];
-
- worker->flags |= (IO_WORKER_F_UP | IO_WORKER_F_RUNNING);
-
- snprintf(buf, sizeof(buf), "iou-wrk-%d", wq->task->pid);
- set_task_comm(current, buf);
while (!test_bit(IO_WQ_BIT_EXIT, &wq->state)) {
long ret;
@@ -664,6 +658,78 @@ static int io_wqe_worker(void *data)
if (test_bit(IO_WQ_BIT_EXIT, &wq->state))
io_worker_handle_work(worker);
+}
+
+#define IO_URINGLET_EMPTY_LIMIT 100000
+#define URINGLET_WORKER_IDLE_TIMEOUT 1
+
+static void io_wqe_worker_let(struct io_worker *worker)
+{
+ struct io_wqe *wqe = worker->wqe;
+ struct io_wq *wq = wqe->wq;
+
+ /* TODO this one breaks encapsulation */
+ if (unlikely(io_uring_add_tctx_node(wq->private)))
+ goto out;
+
+ while (!test_bit(IO_WQ_BIT_EXIT, &wq->state)) {
+ unsigned int empty_count = 0;
+
+ __io_worker_busy(wqe, worker);
+ set_current_state(TASK_INTERRUPTIBLE);
+
+ do {
+ enum io_uringlet_state submit_state;
+
+ submit_state = wq->do_work(wq->private);
+ if (submit_state == IO_URINGLET_SCHEDULED) {
+ empty_count = 0;
+ break;
+ } else if (submit_state == IO_URINGLET_EMPTY) {
+ if (++empty_count > IO_URINGLET_EMPTY_LIMIT)
+ break;
+ } else {
+ empty_count = 0;
+ }
+ cond_resched();
+ } while (1);
+
+ raw_spin_lock(&wqe->lock);
+ __io_worker_idle(wqe, worker);
+ raw_spin_unlock(&wqe->lock);
+ schedule_timeout(URINGLET_WORKER_IDLE_TIMEOUT);
+ if (signal_pending(current)) {
+ struct ksignal ksig;
+
+ if (!get_signal(&ksig))
+ continue;
+ break;
+ }
+ }
+
+ __set_current_state(TASK_RUNNING);
+out:
+ wq->free_work(NULL);
+}
+
+static int io_wqe_worker(void *data)
+{
+ struct io_worker *worker = data;
+ struct io_wqe *wqe = worker->wqe;
+ struct io_wq *wq = wqe->wq;
+ bool uringlet = io_wq_is_uringlet(wq);
+ char buf[TASK_COMM_LEN];
+
+ worker->flags |= (IO_WORKER_F_UP | IO_WORKER_F_RUNNING);
+
+ snprintf(buf, sizeof(buf), uringlet ? "iou-let-%d" : "iou-wrk-%d",
+ wq->task->pid);
+ set_task_comm(current, buf);
+
+ if (uringlet)
+ io_wqe_worker_let(worker);
+ else
+ io_wqe_worker_normal(worker);
io_worker_exit(worker);
return 0;
@@ -21,8 +21,14 @@ enum io_wq_cancel {
IO_WQ_CANCEL_NOTFOUND, /* work not found */
};
+enum io_uringlet_state {
+ IO_URINGLET_INLINE,
+ IO_URINGLET_EMPTY,
+ IO_URINGLET_SCHEDULED,
+};
+
typedef struct io_wq_work *(free_work_fn)(struct io_wq_work *);
-typedef void (io_wq_work_fn)(struct io_wq_work *);
+typedef int (io_wq_work_fn)(struct io_wq_work *);
struct io_wq_hash {
refcount_t refs;
@@ -1612,7 +1612,7 @@ struct io_wq_work *io_wq_free_work(struct io_wq_work *work)
return req ? &req->work : NULL;
}
-void io_wq_submit_work(struct io_wq_work *work)
+int io_wq_submit_work(struct io_wq_work *work)
{
struct io_kiocb *req = container_of(work, struct io_kiocb, work);
const struct io_op_def *def = &io_op_defs[req->opcode];
@@ -1632,7 +1632,7 @@ void io_wq_submit_work(struct io_wq_work *work)
if (work->flags & IO_WQ_WORK_CANCEL) {
fail:
io_req_task_queue_fail(req, err);
- return;
+ return 0;
}
if (!io_assign_file(req, issue_flags)) {
err = -EBADF;
@@ -1666,7 +1666,7 @@ void io_wq_submit_work(struct io_wq_work *work)
}
if (io_arm_poll_handler(req, issue_flags) == IO_APOLL_OK)
- return;
+ return 0;
/* aborted or ready, in either case retry blocking */
needs_poll = false;
issue_flags &= ~IO_URING_F_NONBLOCK;
@@ -1675,6 +1675,8 @@ void io_wq_submit_work(struct io_wq_work *work)
/* avoid locking problems by failing it from a clean context */
if (ret < 0)
io_req_task_queue_fail(req, ret);
+
+ return 0;
}
inline struct file *io_file_get_fixed(struct io_kiocb *req, int fd,
@@ -69,7 +69,7 @@ void io_free_batch_list(struct io_ring_ctx *ctx, struct io_wq_work_node *node);
int io_req_prep_async(struct io_kiocb *req);
struct io_wq_work *io_wq_free_work(struct io_wq_work *work);
-void io_wq_submit_work(struct io_wq_work *work);
+int io_wq_submit_work(struct io_wq_work *work);
void io_free_req(struct io_kiocb *req);
void io_queue_next(struct io_kiocb *req);