diff mbox series

[bpf-next,v5] libbpf: Expose API to consume one ring at a time

Message ID 20230728093346.673994-1-adam@wowsignal.io (mailing list archive)
State New, archived
Delegated to: BPF
Headers show
Series [bpf-next,v5] libbpf: Expose API to consume one ring at a time | expand

Checks

Context Check Description
netdev/series_format success Single patches do not need cover letters
netdev/tree_selection success Clearly marked for bpf-next
netdev/fixes_present success Fixes tag not required for -next series
netdev/header_inline success No static functions without inline keyword in header files
netdev/build_32bit success Errors and warnings before: 9 this patch: 9
netdev/cc_maintainers warning 13 maintainers not CCed: daniel@iogearbox.net kpsingh@kernel.org martin.lau@linux.dev john.fastabend@gmail.com song@kernel.org sdf@google.com andrii@kernel.org yonghong.song@linux.dev shuah@kernel.org mykolal@fb.com linux-kselftest@vger.kernel.org jolsa@kernel.org haoluo@google.com
netdev/build_clang success Errors and warnings before: 9 this patch: 9
netdev/verify_signedoff success Signed-off-by tag matches author and committer
netdev/deprecated_api success None detected
netdev/check_selftest success No net selftest shell script
netdev/verify_fixes success No Fixes tag
netdev/build_allmodconfig_warn success Errors and warnings before: 9 this patch: 9
netdev/checkpatch warning CHECK: Alignment should match open parenthesis WARNING: line length of 81 exceeds 80 columns WARNING: line length of 83 exceeds 80 columns WARNING: line length of 86 exceeds 80 columns WARNING: line length of 92 exceeds 80 columns
netdev/kdoc success Errors and warnings before: 116 this patch: 116
netdev/source_inline success Was 0 now: 0
bpf/vmtest-bpf-next-PR success PR summary
bpf/vmtest-bpf-next-VM_Test-1 success Logs for ShellCheck
bpf/vmtest-bpf-next-VM_Test-6 success Logs for set-matrix
bpf/vmtest-bpf-next-VM_Test-2 success Logs for build for aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-4 success Logs for build for x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-5 success Logs for build for x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-3 success Logs for build for s390x with gcc
bpf/vmtest-bpf-next-VM_Test-25 success Logs for test_verifier on aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-7 success Logs for test_maps on aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-9 success Logs for test_maps on x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-10 success Logs for test_maps on x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-11 success Logs for test_progs on aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-13 success Logs for test_progs on x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-14 success Logs for test_progs on x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-15 success Logs for test_progs_no_alu32 on aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-17 success Logs for test_progs_no_alu32 on x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-18 success Logs for test_progs_no_alu32 on x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-19 success Logs for test_progs_no_alu32_parallel on aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-20 success Logs for test_progs_no_alu32_parallel on x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-21 success Logs for test_progs_no_alu32_parallel on x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-22 success Logs for test_progs_parallel on aarch64 with gcc
bpf/vmtest-bpf-next-VM_Test-23 success Logs for test_progs_parallel on x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-24 success Logs for test_progs_parallel on x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-27 success Logs for test_verifier on x86_64 with gcc
bpf/vmtest-bpf-next-VM_Test-28 success Logs for test_verifier on x86_64 with llvm-16
bpf/vmtest-bpf-next-VM_Test-29 success Logs for veristat
bpf/vmtest-bpf-next-VM_Test-12 success Logs for test_progs on s390x with gcc
bpf/vmtest-bpf-next-VM_Test-16 fail Logs for test_progs_no_alu32 on s390x with gcc
bpf/vmtest-bpf-next-VM_Test-26 success Logs for test_verifier on s390x with gcc
bpf/vmtest-bpf-next-VM_Test-8 success Logs for test_maps on s390x with gcc

Commit Message

Adam Sindelar July 28, 2023, 9:33 a.m. UTC
We already provide ring_buffer__epoll_fd to enable use of external
polling systems. However, the only API available to consume the ring
buffer is ring_buffer__consume, which always checks all rings. When
polling for many events, this can be wasteful.

Signed-off-by: Adam Sindelar <adam@wowsignal.io>
---
v1->v2: Added entry to libbpf.map
v2->v3: Correctly set errno and handle overflow
v3->v4: Fixed an embarrasing typo from zealous autocomplete
v4->v5: Added a selftest to show usage

 tools/lib/bpf/libbpf.h                        |  1 +
 tools/lib/bpf/libbpf.map                      |  1 +
 tools/lib/bpf/ringbuf.c                       | 22 ++++++++++++++++
 .../selftests/bpf/prog_tests/ringbuf_multi.c  | 26 +++++++++++++++++++
 4 files changed, 50 insertions(+)

Comments

Hou Tao July 28, 2023, 10:51 a.m. UTC | #1
On 7/28/2023 5:33 PM, Adam Sindelar wrote:
> We already provide ring_buffer__epoll_fd to enable use of external
> polling systems. However, the only API available to consume the ring
> buffer is ring_buffer__consume, which always checks all rings. When
> polling for many events, this can be wasteful.
>
> Signed-off-by: Adam Sindelar <adam@wowsignal.io>

Acked-by: Hou Tao <houtao1@huawei.com>
Adam Sindelar Aug. 16, 2023, 7:22 a.m. UTC | #2
On Fri, Jul 28, 2023 at 06:51:25PM +0800, Hou Tao wrote:
> 
Hi, sorry for potentially dumb question, but should I do anything else
after someone acks it? This is a minor patch for a userland component,
but it's really helpful IMO - is anything preventing this getting merged
at this point?

Thanks,
Adam

> On 7/28/2023 5:33 PM, Adam Sindelar wrote:
> > We already provide ring_buffer__epoll_fd to enable use of external
> > polling systems. However, the only API available to consume the ring
> > buffer is ring_buffer__consume, which always checks all rings. When
> > polling for many events, this can be wasteful.
> >
> > Signed-off-by: Adam Sindelar <adam@wowsignal.io>
> 
> Acked-by: Hou Tao <houtao1@huawei.com>
>
Hou Tao Aug. 16, 2023, 7:57 a.m. UTC | #3
Hi,

On 8/16/2023 3:22 PM, Adam Sindelar wrote:
> On Fri, Jul 28, 2023 at 06:51:25PM +0800, Hou Tao wrote:
> Hi, sorry for potentially dumb question, but should I do anything else
> after someone acks it? This is a minor patch for a userland component,
> but it's really helpful IMO - is anything preventing this getting merged
> at this point?

According to BPF CI [0], the main reason that it was closed was due to
the failure of test_progs_no_alu32 on s390x, but it seems the failure
was due to BPF CI itself instead of the modification in the patchset, so
I think resend v5 to BPF mail list will be enough. After the pass of BPF
CI, I think the maintainer will merge it if there is no disagreement.

[0]:
https://github.com/kernel-patches/bpf/actions/runs/5829217685/job/15808898814
> Thanks,
> Adam
>
>> On 7/28/2023 5:33 PM, Adam Sindelar wrote:
>>> We already provide ring_buffer__epoll_fd to enable use of external
>>> polling systems. However, the only API available to consume the ring
>>> buffer is ring_buffer__consume, which always checks all rings. When
>>> polling for many events, this can be wasteful.
>>>
>>> Signed-off-by: Adam Sindelar <adam@wowsignal.io>
>> Acked-by: Hou Tao <houtao1@huawei.com>
>>
diff mbox series

Patch

diff --git a/tools/lib/bpf/libbpf.h b/tools/lib/bpf/libbpf.h
index 55b97b2087540..20ccc65eb3f9d 100644
--- a/tools/lib/bpf/libbpf.h
+++ b/tools/lib/bpf/libbpf.h
@@ -1195,6 +1195,7 @@  LIBBPF_API int ring_buffer__add(struct ring_buffer *rb, int map_fd,
 				ring_buffer_sample_fn sample_cb, void *ctx);
 LIBBPF_API int ring_buffer__poll(struct ring_buffer *rb, int timeout_ms);
 LIBBPF_API int ring_buffer__consume(struct ring_buffer *rb);
+LIBBPF_API int ring_buffer__consume_ring(struct ring_buffer *rb, uint32_t ring_id);
 LIBBPF_API int ring_buffer__epoll_fd(const struct ring_buffer *rb);
 
 struct user_ring_buffer_opts {
diff --git a/tools/lib/bpf/libbpf.map b/tools/lib/bpf/libbpf.map
index 9c7538dd5835e..42dc418b4672f 100644
--- a/tools/lib/bpf/libbpf.map
+++ b/tools/lib/bpf/libbpf.map
@@ -398,4 +398,5 @@  LIBBPF_1.3.0 {
 		bpf_prog_detach_opts;
 		bpf_program__attach_netfilter;
 		bpf_program__attach_tcx;
+		ring_buffer__consume_ring;
 } LIBBPF_1.2.0;
diff --git a/tools/lib/bpf/ringbuf.c b/tools/lib/bpf/ringbuf.c
index 02199364db136..457469fc7d71e 100644
--- a/tools/lib/bpf/ringbuf.c
+++ b/tools/lib/bpf/ringbuf.c
@@ -290,6 +290,28 @@  int ring_buffer__consume(struct ring_buffer *rb)
 	return res;
 }
 
+/* Consume available data from a single RINGBUF map identified by its ID.
+ * The ring ID is returned in epoll_data by epoll_wait when called with
+ * ring_buffer__epoll_fd.
+ */
+int ring_buffer__consume_ring(struct ring_buffer *rb, uint32_t ring_id)
+{
+	struct ring *ring;
+	int64_t res;
+
+	if (ring_id >= rb->ring_cnt)
+		return libbpf_err(-EINVAL);
+
+	ring = &rb->rings[ring_id];
+	res = ringbuf_process_ring(ring);
+	if (res < 0)
+		return libbpf_err(res);
+
+	if (res > INT_MAX)
+		return INT_MAX;
+	return res;
+}
+
 /* Poll for available data and consume records, if any are available.
  * Returns number of records consumed (or INT_MAX, whichever is less), or
  * negative number, if any of the registered callbacks returned error.
diff --git a/tools/testing/selftests/bpf/prog_tests/ringbuf_multi.c b/tools/testing/selftests/bpf/prog_tests/ringbuf_multi.c
index 1455911d9fcbe..8123efc94d1a8 100644
--- a/tools/testing/selftests/bpf/prog_tests/ringbuf_multi.c
+++ b/tools/testing/selftests/bpf/prog_tests/ringbuf_multi.c
@@ -29,6 +29,11 @@  static int process_sample(void *ctx, void *data, size_t len)
 		CHECK(s->value != 777, "sample2_value", "exp %ld, got %ld\n",
 		      777L, s->value);
 		break;
+	case 2:
+		CHECK(ring != 2, "sample3_ring", "exp %d, got %d\n", 2, ring);
+		CHECK(s->value != 1337, "sample3_value", "exp %ld, got %ld\n",
+		      1337L, s->value);
+		break;
 	default:
 		CHECK(true, "extra_sample", "unexpected sample seq %d, val %ld\n",
 		      s->seq, s->value);
@@ -45,6 +50,8 @@  void test_ringbuf_multi(void)
 	int err;
 	int page_size = getpagesize();
 	int proto_fd = -1;
+	int epoll_fd;
+	struct epoll_event events[2];
 
 	skel = test_ringbuf_multi__open();
 	if (CHECK(!skel, "skel_open", "skeleton open failed\n"))
@@ -124,6 +131,25 @@  void test_ringbuf_multi(void)
 	CHECK(skel->bss->total != 2, "err_total", "exp %ld, got %ld\n",
 	      2L, skel->bss->total);
 
+	/* validate APIs to support external polling */
+	epoll_fd = ring_buffer__epoll_fd(ringbuf);
+
+	/* expect events on either ring to trigger through the epoll_fd */
+	skel->bss->target_ring = 2;
+	skel->bss->value = 1337;
+	syscall(__NR_getpgid);
+
+	err = epoll_wait(epoll_fd, events, sizeof(events) / sizeof(struct epoll_event), -1);
+	if (CHECK(err != 1, "epoll_wait", "epoll_wait exp %d, got %d\n", 1, err))
+		goto cleanup;
+	if (CHECK(!(events[0].events & EPOLLIN), "epoll_event", "expected EPOLLIN\n"))
+		goto cleanup;
+
+	/* epoll data can be used to consume only the affected ring */
+	err = ring_buffer__consume_ring(ringbuf, events[0].data.u32);
+	CHECK(err != 1, "consume_ring", "consume_ring %u exp %d, got %d\n",
+		  events[0].data.u32, 1, err);
+
 cleanup:
 	if (proto_fd >= 0)
 		close(proto_fd);