@@ -516,6 +516,10 @@ enum {
/* register a range of fixed file slots for automatic slot allocation */
IORING_REGISTER_FILE_ALLOC_RANGE = 25,
+ /* set/clear busy poll settings */
+ IORING_REGISTER_NAPI = 26,
+ IORING_UNREGISTER_NAPI = 27,
+
/* this goes last */
IORING_REGISTER_LAST
};
@@ -638,6 +642,13 @@ struct io_uring_buf_reg {
__u64 resv[3];
};
+/* argument for IORING_(UN)REGISTER_NAPI */
+struct io_uring_napi {
+ __u32 busy_poll_to;
+ __u32 pad;
+ __u64 resv;
+};
+
/*
* io_uring_restriction->opcode values
*/
@@ -4246,6 +4246,49 @@ static __cold int io_register_iowq_max_workers(struct io_ring_ctx *ctx,
return ret;
}
+static int io_register_napi(struct io_ring_ctx *ctx, void __user *arg)
+{
+#ifdef CONFIG_NET_RX_BUSY_POLL
+ const struct io_uring_napi curr = {
+ .busy_poll_to = ctx->napi_busy_poll_to,
+ };
+ struct io_uring_napi napi;
+
+ if (copy_from_user(&napi, arg, sizeof(napi)))
+ return -EFAULT;
+ if (napi.pad || napi.resv)
+ return -EINVAL;
+
+ WRITE_ONCE(ctx->napi_busy_poll_to, napi.busy_poll_to);
+
+ if (copy_to_user(arg, &curr, sizeof(curr)))
+ return -EFAULT;
+
+ return 0;
+#else
+ return -EOPNOTSUPP;
+#endif
+}
+
+static int io_unregister_napi(struct io_ring_ctx *ctx, void __user *arg)
+{
+#ifdef CONFIG_NET_RX_BUSY_POLL
+ const struct io_uring_napi curr = {
+ .busy_poll_to = ctx->napi_busy_poll_to,
+ };
+
+ if (arg) {
+ if (copy_to_user(arg, &curr, sizeof(curr)))
+ return -EFAULT;
+ }
+
+ WRITE_ONCE(ctx->napi_busy_poll_to, 0);
+ return 0;
+#else
+ return -EOPNOTSUPP;
+#endif
+}
+
static int __io_uring_register(struct io_ring_ctx *ctx, unsigned opcode,
void __user *arg, unsigned nr_args)
__releases(ctx->uring_lock)
@@ -4404,6 +4447,15 @@ static int __io_uring_register(struct io_ring_ctx *ctx, unsigned opcode,
break;
ret = io_register_file_alloc_range(ctx, arg);
break;
+ case IORING_REGISTER_NAPI:
+ ret = -EINVAL;
+ if (!arg)
+ break;
+ ret = io_register_napi(ctx, arg);
+ break;
+ case IORING_UNREGISTER_NAPI:
+ ret = io_unregister_napi(ctx, arg);
+ break;
default:
ret = -EINVAL;
break;