mbox series

[v2,00/15] caching and SQ/CQ optimisations

Message ID cover.1692916914.git.asml.silence@gmail.com (mailing list archive)
Headers show
Series caching and SQ/CQ optimisations | expand

Message

Pavel Begunkov Aug. 24, 2023, 10:53 p.m. UTC
Patch 1-5 optimise io_fill_cqe_req

Patch 6-7 combine iopoll and normal completion paths

Patch 8 inlines io_fill_cqe_req.

Patch 9 should improve CPU caching of SQ/CQ pointers

Patch 10 removes conditionally SQ indirection (->sq_array). Assuming we'll
make it a default in liburing, Patch 10 optimises it with static_key.

Patch 11-15 shuffle io_ring_ctx fields.

Testing with t/io_uring nops only for now

                QD2     QD4     QD8     QD16    QD32
baseline:       17.3    26.6    36.4    43.7    49.4
Patches 1-15:   17.8    27.4    37.9    45.8    51.2
Patches 1-16:   17.9    28.2    39.3    47.8    54

v2:
  removed static_key, it'll be submitted later after it rolls out well
  minor description changes

Pavel Begunkov (15):
  io_uring: improve cqe !tracing hot path
  io_uring: cqe init hardening
  io_uring: simplify big_cqe handling
  io_uring: refactor __io_get_cqe()
  io_uring: optimise extra io_get_cqe null check
  io_uring: reorder cqring_flush and wakeups
  io_uring: merge iopoll and normal completion paths
  io_uring: force inline io_fill_cqe_req
  io_uring: compact SQ/CQ heads/tails
  io_uring: add option to remove SQ indirection
  io_uring: move non aligned field to the end
  io_uring: banish non-hot data to end of io_ring_ctx
  io_uring: separate task_work/waiting cache line
  io_uring: move multishot cqe cache in ctx
  io_uring: move iopoll ctx fields around

 include/linux/io_uring_types.h | 129 +++++++++++++++++----------------
 include/uapi/linux/io_uring.h  |   5 ++
 io_uring/io_uring.c            | 120 +++++++++++++++---------------
 io_uring/io_uring.h            |  58 +++++++--------
 io_uring/rw.c                  |  24 ++----
 io_uring/uring_cmd.c           |   5 +-
 6 files changed, 163 insertions(+), 178 deletions(-)

Comments

Jens Axboe Aug. 24, 2023, 11:16 p.m. UTC | #1
On Thu, 24 Aug 2023 23:53:22 +0100, Pavel Begunkov wrote:
> Patch 1-5 optimise io_fill_cqe_req
> 
> Patch 6-7 combine iopoll and normal completion paths
> 
> Patch 8 inlines io_fill_cqe_req.
> 
> Patch 9 should improve CPU caching of SQ/CQ pointers
> 
> [...]

Applied, thanks!

[01/15] io_uring: improve cqe !tracing hot path
        commit: a0727c738309a06ef5579c1742f8f0def63aa883
[02/15] io_uring: cqe init hardening
        commit: 31d3ba924fd86add6d14f9085fdd2f4ec0879631
[03/15] io_uring: simplify big_cqe handling
        commit: b24c5d752962fa0970cd7e3d74b1cd0e843358de
[04/15] io_uring: refactor __io_get_cqe()
        commit: 20d6b633870495fda1d92d283ebf890d80f68ecd
[05/15] io_uring: optimise extra io_get_cqe null check
        commit: 59fbc409e71649f558fb4578cdbfac67acb824dc
[06/15] io_uring: reorder cqring_flush and wakeups
        commit: 54927baf6c195fb512ac38b26a041ca44edb2e29
[07/15] io_uring: merge iopoll and normal completion paths
        commit: ec26c225f06f5993f8891fa6c79fab3c92981181
[08/15] io_uring: force inline io_fill_cqe_req
        commit: 093a650b757210bc856ca7f5349fb5a4bb9d4bd6
[09/15] io_uring: compact SQ/CQ heads/tails
        commit: e5598d6ae62626d261b046a2f19347c38681ff51
[10/15] io_uring: add option to remove SQ indirection
        commit: 2af89abda7d9c2aeb573677e2c498ddb09f8058a
[11/15] io_uring: move non aligned field to the end
        commit: d7f06fea5d6be78403d42c9637f67bc883870094
[12/15] io_uring: banish non-hot data to end of io_ring_ctx
        commit: 18df385f42f0b3310ed2e4a3e39264bf5e784692
[13/15] io_uring: separate task_work/waiting cache line
        commit: c9def23dde5238184777340ad811e4903f216a2d
[14/15] io_uring: move multishot cqe cache in ctx
        commit: 0aa7aa5f766933d4f91b22d9658cd688e1f15dab
[15/15] io_uring: move iopoll ctx fields around
        commit: 644c4a7a721fb90356cdd42219c9928a3c386230

Best regards,