1 // SPDX-License-Identifier: GPL-2.0
3 #include <linux/compiler.h>
4 #include <asm/barrier.h>
5 #include <test_progs.h>
12 #include <sys/sysinfo.h>
13 #include <linux/perf_event.h>
14 #include <linux/ring_buffer.h>
15 #include "test_ringbuf.lskel.h"
16 #include "test_ringbuf_n.lskel.h"
17 #include "test_ringbuf_map_key.lskel.h"
21 static int duration = 0;
30 static int sample_cnt;
32 static void atomic_inc(int *cnt)
34 __atomic_add_fetch(cnt, 1, __ATOMIC_SEQ_CST);
37 static int atomic_xchg(int *cnt, int val)
39 return __atomic_exchange_n(cnt, val, __ATOMIC_SEQ_CST);
42 static int process_sample(void *ctx, void *data, size_t len)
44 struct sample *s = data;
46 atomic_inc(&sample_cnt);
50 CHECK(s->value != 333, "sample1_value", "exp %ld, got %ld\n",
54 CHECK(s->value != 777, "sample2_value", "exp %ld, got %ld\n",
58 /* we don't care about the rest */
63 static struct test_ringbuf_map_key_lskel *skel_map_key;
64 static struct test_ringbuf_lskel *skel;
65 static struct ring_buffer *ringbuf;
67 static void trigger_samples()
69 skel->bss->dropped = 0;
71 skel->bss->discarded = 0;
73 /* trigger exactly two samples */
74 skel->bss->value = 333;
75 syscall(__NR_getpgid);
76 skel->bss->value = 777;
77 syscall(__NR_getpgid);
80 static void *poll_thread(void *input)
82 long timeout = (long)input;
84 return (void *)(long)ring_buffer__poll(ringbuf, timeout);
87 static void ringbuf_subtest(void)
89 const size_t rec_sz = BPF_RINGBUF_HDR_SZ + sizeof(struct sample);
93 int page_size = getpagesize();
94 void *mmap_ptr, *tmp_ptr;
97 unsigned long avail_data, ring_size, cons_pos, prod_pos;
99 skel = test_ringbuf_lskel__open();
100 if (CHECK(!skel, "skel_open", "skeleton open failed\n"))
103 skel->maps.ringbuf.max_entries = page_size;
105 err = test_ringbuf_lskel__load(skel);
106 if (CHECK(err != 0, "skel_load", "skeleton load failed\n"))
109 rb_fd = skel->maps.ringbuf.map_fd;
110 /* good read/write cons_pos */
111 mmap_ptr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, rb_fd, 0);
112 ASSERT_OK_PTR(mmap_ptr, "rw_cons_pos");
113 tmp_ptr = mremap(mmap_ptr, page_size, 2 * page_size, MREMAP_MAYMOVE);
114 if (!ASSERT_ERR_PTR(tmp_ptr, "rw_extend"))
116 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_cons_pos_protect");
117 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_rw");
119 /* bad writeable prod_pos */
120 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, page_size);
122 ASSERT_ERR_PTR(mmap_ptr, "wr_prod_pos");
123 ASSERT_EQ(err, -EPERM, "wr_prod_pos_err");
125 /* bad writeable data pages */
126 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size);
128 ASSERT_ERR_PTR(mmap_ptr, "wr_data_page_one");
129 ASSERT_EQ(err, -EPERM, "wr_data_page_one_err");
130 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 3 * page_size);
131 ASSERT_ERR_PTR(mmap_ptr, "wr_data_page_two");
132 mmap_ptr = mmap(NULL, 2 * page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size);
133 ASSERT_ERR_PTR(mmap_ptr, "wr_data_page_all");
135 /* good read-only pages */
136 mmap_ptr = mmap(NULL, 4 * page_size, PROT_READ, MAP_SHARED, rb_fd, 0);
137 if (!ASSERT_OK_PTR(mmap_ptr, "ro_prod_pos"))
140 ASSERT_ERR(mprotect(mmap_ptr, 4 * page_size, PROT_WRITE), "write_protect");
141 ASSERT_ERR(mprotect(mmap_ptr, 4 * page_size, PROT_EXEC), "exec_protect");
142 ASSERT_ERR_PTR(mremap(mmap_ptr, 0, 4 * page_size, MREMAP_MAYMOVE), "ro_remap");
143 ASSERT_OK(munmap(mmap_ptr, 4 * page_size), "unmap_ro");
145 /* good read-only pages with initial offset */
146 mmap_ptr = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rb_fd, page_size);
147 if (!ASSERT_OK_PTR(mmap_ptr, "ro_prod_pos"))
150 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_WRITE), "write_protect");
151 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_protect");
152 ASSERT_ERR_PTR(mremap(mmap_ptr, 0, 3 * page_size, MREMAP_MAYMOVE), "ro_remap");
153 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_ro");
155 /* only trigger BPF program for current process */
156 skel->bss->pid = getpid();
158 ringbuf = ring_buffer__new(skel->maps.ringbuf.map_fd,
159 process_sample, NULL, NULL);
160 if (CHECK(!ringbuf, "ringbuf_create", "failed to create ringbuf\n"))
163 err = test_ringbuf_lskel__attach(skel);
164 if (CHECK(err, "skel_attach", "skeleton attachment failed: %d\n", err))
169 ring = ring_buffer__ring(ringbuf, 0);
170 if (!ASSERT_OK_PTR(ring, "ring_buffer__ring_idx_0"))
173 map_fd = ring__map_fd(ring);
174 ASSERT_EQ(map_fd, skel->maps.ringbuf.map_fd, "ring_map_fd");
176 /* 2 submitted + 1 discarded records */
177 CHECK(skel->bss->avail_data != 3 * rec_sz,
178 "err_avail_size", "exp %ld, got %ld\n",
179 3L * rec_sz, skel->bss->avail_data);
180 CHECK(skel->bss->ring_size != page_size,
181 "err_ring_size", "exp %ld, got %ld\n",
182 (long)page_size, skel->bss->ring_size);
183 CHECK(skel->bss->cons_pos != 0,
184 "err_cons_pos", "exp %ld, got %ld\n",
185 0L, skel->bss->cons_pos);
186 CHECK(skel->bss->prod_pos != 3 * rec_sz,
187 "err_prod_pos", "exp %ld, got %ld\n",
188 3L * rec_sz, skel->bss->prod_pos);
190 /* verify getting this data directly via the ring object yields the same
193 avail_data = ring__avail_data_size(ring);
194 ASSERT_EQ(avail_data, 3 * rec_sz, "ring_avail_size");
195 ring_size = ring__size(ring);
196 ASSERT_EQ(ring_size, page_size, "ring_ring_size");
197 cons_pos = ring__consumer_pos(ring);
198 ASSERT_EQ(cons_pos, 0, "ring_cons_pos");
199 prod_pos = ring__producer_pos(ring);
200 ASSERT_EQ(prod_pos, 3 * rec_sz, "ring_prod_pos");
202 /* poll for samples */
203 err = ring_buffer__poll(ringbuf, -1);
205 /* -EDONE is used as an indicator that we are done */
206 if (CHECK(err != -EDONE, "err_done", "done err: %d\n", err))
208 cnt = atomic_xchg(&sample_cnt, 0);
209 CHECK(cnt != 2, "cnt", "exp %d samples, got %d\n", 2, cnt);
211 /* we expect extra polling to return nothing */
212 err = ring_buffer__poll(ringbuf, 0);
213 if (CHECK(err != 0, "extra_samples", "poll result: %d\n", err))
215 cnt = atomic_xchg(&sample_cnt, 0);
216 CHECK(cnt != 0, "cnt", "exp %d samples, got %d\n", 0, cnt);
218 CHECK(skel->bss->dropped != 0, "err_dropped", "exp %ld, got %ld\n",
219 0L, skel->bss->dropped);
220 CHECK(skel->bss->total != 2, "err_total", "exp %ld, got %ld\n",
221 2L, skel->bss->total);
222 CHECK(skel->bss->discarded != 1, "err_discarded", "exp %ld, got %ld\n",
223 1L, skel->bss->discarded);
225 /* now validate consumer position is updated and returned */
227 CHECK(skel->bss->cons_pos != 3 * rec_sz,
228 "err_cons_pos", "exp %ld, got %ld\n",
229 3L * rec_sz, skel->bss->cons_pos);
230 err = ring_buffer__poll(ringbuf, -1);
231 CHECK(err <= 0, "poll_err", "err %d\n", err);
232 cnt = atomic_xchg(&sample_cnt, 0);
233 CHECK(cnt != 2, "cnt", "exp %d samples, got %d\n", 2, cnt);
235 /* start poll in background w/ long timeout */
236 err = pthread_create(&thread, NULL, poll_thread, (void *)(long)10000);
237 if (CHECK(err, "bg_poll", "pthread_create failed: %d\n", err))
240 /* turn off notifications now */
241 skel->bss->flags = BPF_RB_NO_WAKEUP;
243 /* give background thread a bit of a time */
246 /* sleeping arbitrarily is bad, but no better way to know that
247 * epoll_wait() **DID NOT** unblock in background thread
250 /* background poll should still be blocked */
251 err = pthread_tryjoin_np(thread, (void **)&bg_ret);
252 if (CHECK(err != EBUSY, "try_join", "err %d\n", err))
255 /* BPF side did everything right */
256 CHECK(skel->bss->dropped != 0, "err_dropped", "exp %ld, got %ld\n",
257 0L, skel->bss->dropped);
258 CHECK(skel->bss->total != 2, "err_total", "exp %ld, got %ld\n",
259 2L, skel->bss->total);
260 CHECK(skel->bss->discarded != 1, "err_discarded", "exp %ld, got %ld\n",
261 1L, skel->bss->discarded);
262 cnt = atomic_xchg(&sample_cnt, 0);
263 CHECK(cnt != 0, "cnt", "exp %d samples, got %d\n", 0, cnt);
265 /* clear flags to return to "adaptive" notification mode */
266 skel->bss->flags = 0;
268 /* produce new samples, no notification should be triggered, because
269 * consumer is now behind
273 /* background poll should still be blocked */
274 err = pthread_tryjoin_np(thread, (void **)&bg_ret);
275 if (CHECK(err != EBUSY, "try_join", "err %d\n", err))
278 /* still no samples, because consumer is behind */
279 cnt = atomic_xchg(&sample_cnt, 0);
280 CHECK(cnt != 0, "cnt", "exp %d samples, got %d\n", 0, cnt);
282 skel->bss->dropped = 0;
283 skel->bss->total = 0;
284 skel->bss->discarded = 0;
286 skel->bss->value = 333;
287 syscall(__NR_getpgid);
288 /* now force notifications */
289 skel->bss->flags = BPF_RB_FORCE_WAKEUP;
290 skel->bss->value = 777;
291 syscall(__NR_getpgid);
293 /* now we should get a pending notification */
295 err = pthread_tryjoin_np(thread, (void **)&bg_ret);
296 if (CHECK(err, "join_bg", "err %d\n", err))
299 if (CHECK(bg_ret <= 0, "bg_ret", "epoll_wait result: %ld", bg_ret))
302 /* due to timing variations, there could still be non-notified
303 * samples, so consume them here to collect all the samples
305 err = ring_buffer__consume(ringbuf);
306 CHECK(err < 0, "rb_consume", "failed: %d\b", err);
308 /* also consume using ring__consume to make sure it works the same */
309 err = ring__consume(ring);
310 ASSERT_GE(err, 0, "ring_consume");
312 /* 3 rounds, 2 samples each */
313 cnt = atomic_xchg(&sample_cnt, 0);
314 CHECK(cnt != 6, "cnt", "exp %d samples, got %d\n", 6, cnt);
316 /* BPF side did everything right */
317 CHECK(skel->bss->dropped != 0, "err_dropped", "exp %ld, got %ld\n",
318 0L, skel->bss->dropped);
319 CHECK(skel->bss->total != 2, "err_total", "exp %ld, got %ld\n",
320 2L, skel->bss->total);
321 CHECK(skel->bss->discarded != 1, "err_discarded", "exp %ld, got %ld\n",
322 1L, skel->bss->discarded);
324 test_ringbuf_lskel__detach(skel);
326 ring_buffer__free(ringbuf);
327 test_ringbuf_lskel__destroy(skel);
331 * Test ring_buffer__consume_n() by producing N_TOT_SAMPLES samples in the ring
332 * buffer, via getpid(), and consuming them in chunks of N_SAMPLES.
334 #define N_TOT_SAMPLES 32
337 /* Sample value to verify the callback validity */
338 #define SAMPLE_VALUE 42L
340 static int process_n_sample(void *ctx, void *data, size_t len)
342 struct sample *s = data;
344 ASSERT_EQ(s->value, SAMPLE_VALUE, "sample_value");
349 static void ringbuf_n_subtest(void)
351 struct test_ringbuf_n_lskel *skel_n;
354 skel_n = test_ringbuf_n_lskel__open();
355 if (!ASSERT_OK_PTR(skel_n, "test_ringbuf_n_lskel__open"))
358 skel_n->maps.ringbuf.max_entries = getpagesize();
359 skel_n->bss->pid = getpid();
361 err = test_ringbuf_n_lskel__load(skel_n);
362 if (!ASSERT_OK(err, "test_ringbuf_n_lskel__load"))
365 ringbuf = ring_buffer__new(skel_n->maps.ringbuf.map_fd,
366 process_n_sample, NULL, NULL);
367 if (!ASSERT_OK_PTR(ringbuf, "ring_buffer__new"))
370 err = test_ringbuf_n_lskel__attach(skel_n);
371 if (!ASSERT_OK(err, "test_ringbuf_n_lskel__attach"))
372 goto cleanup_ringbuf;
374 /* Produce N_TOT_SAMPLES samples in the ring buffer by calling getpid() */
375 skel_n->bss->value = SAMPLE_VALUE;
376 for (i = 0; i < N_TOT_SAMPLES; i++)
377 syscall(__NR_getpgid);
379 /* Consume all samples from the ring buffer in batches of N_SAMPLES */
380 for (i = 0; i < N_TOT_SAMPLES; i += err) {
381 err = ring_buffer__consume_n(ringbuf, N_SAMPLES);
382 if (!ASSERT_EQ(err, N_SAMPLES, "rb_consume"))
383 goto cleanup_ringbuf;
387 ring_buffer__free(ringbuf);
389 test_ringbuf_n_lskel__destroy(skel_n);
392 static int process_map_key_sample(void *ctx, void *data, size_t len)
400 ASSERT_EQ(s->value, 42, "sample_value");
401 err = bpf_map_lookup_elem(skel_map_key->maps.hash_map.map_fd,
403 ASSERT_OK(err, "hash_map bpf_map_lookup_elem");
404 ASSERT_EQ(val, 1, "hash_map val");
411 static void ringbuf_map_key_subtest(void)
415 skel_map_key = test_ringbuf_map_key_lskel__open();
416 if (!ASSERT_OK_PTR(skel_map_key, "test_ringbuf_map_key_lskel__open"))
419 skel_map_key->maps.ringbuf.max_entries = getpagesize();
420 skel_map_key->bss->pid = getpid();
422 err = test_ringbuf_map_key_lskel__load(skel_map_key);
423 if (!ASSERT_OK(err, "test_ringbuf_map_key_lskel__load"))
426 ringbuf = ring_buffer__new(skel_map_key->maps.ringbuf.map_fd,
427 process_map_key_sample, NULL, NULL);
428 if (!ASSERT_OK_PTR(ringbuf, "ring_buffer__new"))
431 err = test_ringbuf_map_key_lskel__attach(skel_map_key);
432 if (!ASSERT_OK(err, "test_ringbuf_map_key_lskel__attach"))
433 goto cleanup_ringbuf;
435 syscall(__NR_getpgid);
436 ASSERT_EQ(skel_map_key->bss->seq, 1, "skel_map_key->bss->seq");
437 err = ring_buffer__poll(ringbuf, -1);
438 ASSERT_EQ(err, -EDONE, "ring_buffer__poll");
441 ring_buffer__free(ringbuf);
443 test_ringbuf_map_key_lskel__destroy(skel_map_key);
446 void test_ringbuf(void)
448 if (test__start_subtest("ringbuf"))
450 if (test__start_subtest("ringbuf_n"))
452 if (test__start_subtest("ringbuf_map_key"))
453 ringbuf_map_key_subtest();