]> Git Repo - linux.git/blob - tools/testing/selftests/bpf/prog_tests/bpf_cookie.c
Linux 6.14-rc3
[linux.git] / tools / testing / selftests / bpf / prog_tests / bpf_cookie.c
1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (c) 2021 Facebook */
3 #define _GNU_SOURCE
4 #include <pthread.h>
5 #include <sched.h>
6 #include <sys/syscall.h>
7 #include <sys/mman.h>
8 #include <unistd.h>
9 #include <test_progs.h>
10 #include <network_helpers.h>
11 #include <bpf/btf.h>
12 #include "test_bpf_cookie.skel.h"
13 #include "kprobe_multi.skel.h"
14 #include "uprobe_multi.skel.h"
15
16 /* uprobe attach point */
17 static noinline void trigger_func(void)
18 {
19         asm volatile ("");
20 }
21
22 static void kprobe_subtest(struct test_bpf_cookie *skel)
23 {
24         DECLARE_LIBBPF_OPTS(bpf_kprobe_opts, opts);
25         struct bpf_link *link1 = NULL, *link2 = NULL;
26         struct bpf_link *retlink1 = NULL, *retlink2 = NULL;
27
28         /* attach two kprobes */
29         opts.bpf_cookie = 0x1;
30         opts.retprobe = false;
31         link1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe,
32                                                  SYS_NANOSLEEP_KPROBE_NAME, &opts);
33         if (!ASSERT_OK_PTR(link1, "link1"))
34                 goto cleanup;
35
36         opts.bpf_cookie = 0x2;
37         opts.retprobe = false;
38         link2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe,
39                                                  SYS_NANOSLEEP_KPROBE_NAME, &opts);
40         if (!ASSERT_OK_PTR(link2, "link2"))
41                 goto cleanup;
42
43         /* attach two kretprobes */
44         opts.bpf_cookie = 0x10;
45         opts.retprobe = true;
46         retlink1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe,
47                                                     SYS_NANOSLEEP_KPROBE_NAME, &opts);
48         if (!ASSERT_OK_PTR(retlink1, "retlink1"))
49                 goto cleanup;
50
51         opts.bpf_cookie = 0x20;
52         opts.retprobe = true;
53         retlink2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe,
54                                                     SYS_NANOSLEEP_KPROBE_NAME, &opts);
55         if (!ASSERT_OK_PTR(retlink2, "retlink2"))
56                 goto cleanup;
57
58         /* trigger kprobe && kretprobe */
59         usleep(1);
60
61         ASSERT_EQ(skel->bss->kprobe_res, 0x1 | 0x2, "kprobe_res");
62         ASSERT_EQ(skel->bss->kretprobe_res, 0x10 | 0x20, "kretprobe_res");
63
64 cleanup:
65         bpf_link__destroy(link1);
66         bpf_link__destroy(link2);
67         bpf_link__destroy(retlink1);
68         bpf_link__destroy(retlink2);
69 }
70
71 static void kprobe_multi_test_run(struct kprobe_multi *skel)
72 {
73         LIBBPF_OPTS(bpf_test_run_opts, topts);
74         int err, prog_fd;
75
76         prog_fd = bpf_program__fd(skel->progs.trigger);
77         err = bpf_prog_test_run_opts(prog_fd, &topts);
78         ASSERT_OK(err, "test_run");
79         ASSERT_EQ(topts.retval, 0, "test_run");
80
81         ASSERT_EQ(skel->bss->kprobe_test1_result, 1, "kprobe_test1_result");
82         ASSERT_EQ(skel->bss->kprobe_test2_result, 1, "kprobe_test2_result");
83         ASSERT_EQ(skel->bss->kprobe_test3_result, 1, "kprobe_test3_result");
84         ASSERT_EQ(skel->bss->kprobe_test4_result, 1, "kprobe_test4_result");
85         ASSERT_EQ(skel->bss->kprobe_test5_result, 1, "kprobe_test5_result");
86         ASSERT_EQ(skel->bss->kprobe_test6_result, 1, "kprobe_test6_result");
87         ASSERT_EQ(skel->bss->kprobe_test7_result, 1, "kprobe_test7_result");
88         ASSERT_EQ(skel->bss->kprobe_test8_result, 1, "kprobe_test8_result");
89
90         ASSERT_EQ(skel->bss->kretprobe_test1_result, 1, "kretprobe_test1_result");
91         ASSERT_EQ(skel->bss->kretprobe_test2_result, 1, "kretprobe_test2_result");
92         ASSERT_EQ(skel->bss->kretprobe_test3_result, 1, "kretprobe_test3_result");
93         ASSERT_EQ(skel->bss->kretprobe_test4_result, 1, "kretprobe_test4_result");
94         ASSERT_EQ(skel->bss->kretprobe_test5_result, 1, "kretprobe_test5_result");
95         ASSERT_EQ(skel->bss->kretprobe_test6_result, 1, "kretprobe_test6_result");
96         ASSERT_EQ(skel->bss->kretprobe_test7_result, 1, "kretprobe_test7_result");
97         ASSERT_EQ(skel->bss->kretprobe_test8_result, 1, "kretprobe_test8_result");
98 }
99
100 static void kprobe_multi_link_api_subtest(void)
101 {
102         int prog_fd, link1_fd = -1, link2_fd = -1;
103         struct kprobe_multi *skel = NULL;
104         LIBBPF_OPTS(bpf_link_create_opts, opts);
105         unsigned long long addrs[8];
106         __u64 cookies[8];
107
108         if (!ASSERT_OK(load_kallsyms(), "load_kallsyms"))
109                 goto cleanup;
110
111         skel = kprobe_multi__open_and_load();
112         if (!ASSERT_OK_PTR(skel, "fentry_raw_skel_load"))
113                 goto cleanup;
114
115         skel->bss->pid = getpid();
116         skel->bss->test_cookie = true;
117
118 #define GET_ADDR(__sym, __addr) ({                              \
119         __addr = ksym_get_addr(__sym);                          \
120         if (!ASSERT_NEQ(__addr, 0, "ksym_get_addr " #__sym))    \
121                 goto cleanup;                                   \
122 })
123
124         GET_ADDR("bpf_fentry_test1", addrs[0]);
125         GET_ADDR("bpf_fentry_test3", addrs[1]);
126         GET_ADDR("bpf_fentry_test4", addrs[2]);
127         GET_ADDR("bpf_fentry_test5", addrs[3]);
128         GET_ADDR("bpf_fentry_test6", addrs[4]);
129         GET_ADDR("bpf_fentry_test7", addrs[5]);
130         GET_ADDR("bpf_fentry_test2", addrs[6]);
131         GET_ADDR("bpf_fentry_test8", addrs[7]);
132
133 #undef GET_ADDR
134
135         cookies[0] = 1; /* bpf_fentry_test1 */
136         cookies[1] = 2; /* bpf_fentry_test3 */
137         cookies[2] = 3; /* bpf_fentry_test4 */
138         cookies[3] = 4; /* bpf_fentry_test5 */
139         cookies[4] = 5; /* bpf_fentry_test6 */
140         cookies[5] = 6; /* bpf_fentry_test7 */
141         cookies[6] = 7; /* bpf_fentry_test2 */
142         cookies[7] = 8; /* bpf_fentry_test8 */
143
144         opts.kprobe_multi.addrs = (const unsigned long *) &addrs;
145         opts.kprobe_multi.cnt = ARRAY_SIZE(addrs);
146         opts.kprobe_multi.cookies = (const __u64 *) &cookies;
147         prog_fd = bpf_program__fd(skel->progs.test_kprobe);
148
149         link1_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_KPROBE_MULTI, &opts);
150         if (!ASSERT_GE(link1_fd, 0, "link1_fd"))
151                 goto cleanup;
152
153         cookies[0] = 8; /* bpf_fentry_test1 */
154         cookies[1] = 7; /* bpf_fentry_test3 */
155         cookies[2] = 6; /* bpf_fentry_test4 */
156         cookies[3] = 5; /* bpf_fentry_test5 */
157         cookies[4] = 4; /* bpf_fentry_test6 */
158         cookies[5] = 3; /* bpf_fentry_test7 */
159         cookies[6] = 2; /* bpf_fentry_test2 */
160         cookies[7] = 1; /* bpf_fentry_test8 */
161
162         opts.kprobe_multi.flags = BPF_F_KPROBE_MULTI_RETURN;
163         prog_fd = bpf_program__fd(skel->progs.test_kretprobe);
164
165         link2_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_KPROBE_MULTI, &opts);
166         if (!ASSERT_GE(link2_fd, 0, "link2_fd"))
167                 goto cleanup;
168
169         kprobe_multi_test_run(skel);
170
171 cleanup:
172         close(link1_fd);
173         close(link2_fd);
174         kprobe_multi__destroy(skel);
175 }
176
177 static void kprobe_multi_attach_api_subtest(void)
178 {
179         struct bpf_link *link1 = NULL, *link2 = NULL;
180         LIBBPF_OPTS(bpf_kprobe_multi_opts, opts);
181         LIBBPF_OPTS(bpf_test_run_opts, topts);
182         struct kprobe_multi *skel = NULL;
183         const char *syms[8] = {
184                 "bpf_fentry_test1",
185                 "bpf_fentry_test3",
186                 "bpf_fentry_test4",
187                 "bpf_fentry_test5",
188                 "bpf_fentry_test6",
189                 "bpf_fentry_test7",
190                 "bpf_fentry_test2",
191                 "bpf_fentry_test8",
192         };
193         __u64 cookies[8];
194
195         skel = kprobe_multi__open_and_load();
196         if (!ASSERT_OK_PTR(skel, "fentry_raw_skel_load"))
197                 goto cleanup;
198
199         skel->bss->pid = getpid();
200         skel->bss->test_cookie = true;
201
202         cookies[0] = 1; /* bpf_fentry_test1 */
203         cookies[1] = 2; /* bpf_fentry_test3 */
204         cookies[2] = 3; /* bpf_fentry_test4 */
205         cookies[3] = 4; /* bpf_fentry_test5 */
206         cookies[4] = 5; /* bpf_fentry_test6 */
207         cookies[5] = 6; /* bpf_fentry_test7 */
208         cookies[6] = 7; /* bpf_fentry_test2 */
209         cookies[7] = 8; /* bpf_fentry_test8 */
210
211         opts.syms = syms;
212         opts.cnt = ARRAY_SIZE(syms);
213         opts.cookies = cookies;
214
215         link1 = bpf_program__attach_kprobe_multi_opts(skel->progs.test_kprobe,
216                                                       NULL, &opts);
217         if (!ASSERT_OK_PTR(link1, "bpf_program__attach_kprobe_multi_opts"))
218                 goto cleanup;
219
220         cookies[0] = 8; /* bpf_fentry_test1 */
221         cookies[1] = 7; /* bpf_fentry_test3 */
222         cookies[2] = 6; /* bpf_fentry_test4 */
223         cookies[3] = 5; /* bpf_fentry_test5 */
224         cookies[4] = 4; /* bpf_fentry_test6 */
225         cookies[5] = 3; /* bpf_fentry_test7 */
226         cookies[6] = 2; /* bpf_fentry_test2 */
227         cookies[7] = 1; /* bpf_fentry_test8 */
228
229         opts.retprobe = true;
230
231         link2 = bpf_program__attach_kprobe_multi_opts(skel->progs.test_kretprobe,
232                                                       NULL, &opts);
233         if (!ASSERT_OK_PTR(link2, "bpf_program__attach_kprobe_multi_opts"))
234                 goto cleanup;
235
236         kprobe_multi_test_run(skel);
237
238 cleanup:
239         bpf_link__destroy(link2);
240         bpf_link__destroy(link1);
241         kprobe_multi__destroy(skel);
242 }
243
244 /* defined in prog_tests/uprobe_multi_test.c */
245 void uprobe_multi_func_1(void);
246 void uprobe_multi_func_2(void);
247 void uprobe_multi_func_3(void);
248
249 static void uprobe_multi_test_run(struct uprobe_multi *skel)
250 {
251         skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1;
252         skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2;
253         skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3;
254
255         skel->bss->pid = getpid();
256         skel->bss->test_cookie = true;
257
258         uprobe_multi_func_1();
259         uprobe_multi_func_2();
260         uprobe_multi_func_3();
261
262         ASSERT_EQ(skel->bss->uprobe_multi_func_1_result, 1, "uprobe_multi_func_1_result");
263         ASSERT_EQ(skel->bss->uprobe_multi_func_2_result, 1, "uprobe_multi_func_2_result");
264         ASSERT_EQ(skel->bss->uprobe_multi_func_3_result, 1, "uprobe_multi_func_3_result");
265
266         ASSERT_EQ(skel->bss->uretprobe_multi_func_1_result, 1, "uretprobe_multi_func_1_result");
267         ASSERT_EQ(skel->bss->uretprobe_multi_func_2_result, 1, "uretprobe_multi_func_2_result");
268         ASSERT_EQ(skel->bss->uretprobe_multi_func_3_result, 1, "uretprobe_multi_func_3_result");
269 }
270
271 static void uprobe_multi_attach_api_subtest(void)
272 {
273         struct bpf_link *link1 = NULL, *link2 = NULL;
274         struct uprobe_multi *skel = NULL;
275         LIBBPF_OPTS(bpf_uprobe_multi_opts, opts);
276         const char *syms[3] = {
277                 "uprobe_multi_func_1",
278                 "uprobe_multi_func_2",
279                 "uprobe_multi_func_3",
280         };
281         __u64 cookies[3];
282
283         cookies[0] = 3; /* uprobe_multi_func_1 */
284         cookies[1] = 1; /* uprobe_multi_func_2 */
285         cookies[2] = 2; /* uprobe_multi_func_3 */
286
287         opts.syms = syms;
288         opts.cnt = ARRAY_SIZE(syms);
289         opts.cookies = &cookies[0];
290
291         skel = uprobe_multi__open_and_load();
292         if (!ASSERT_OK_PTR(skel, "uprobe_multi"))
293                 goto cleanup;
294
295         link1 = bpf_program__attach_uprobe_multi(skel->progs.uprobe, -1,
296                                                  "/proc/self/exe", NULL, &opts);
297         if (!ASSERT_OK_PTR(link1, "bpf_program__attach_uprobe_multi"))
298                 goto cleanup;
299
300         cookies[0] = 2; /* uprobe_multi_func_1 */
301         cookies[1] = 3; /* uprobe_multi_func_2 */
302         cookies[2] = 1; /* uprobe_multi_func_3 */
303
304         opts.retprobe = true;
305         link2 = bpf_program__attach_uprobe_multi(skel->progs.uretprobe, -1,
306                                                       "/proc/self/exe", NULL, &opts);
307         if (!ASSERT_OK_PTR(link2, "bpf_program__attach_uprobe_multi_retprobe"))
308                 goto cleanup;
309
310         uprobe_multi_test_run(skel);
311
312 cleanup:
313         bpf_link__destroy(link2);
314         bpf_link__destroy(link1);
315         uprobe_multi__destroy(skel);
316 }
317
318 static void uprobe_subtest(struct test_bpf_cookie *skel)
319 {
320         DECLARE_LIBBPF_OPTS(bpf_uprobe_opts, opts);
321         struct bpf_link *link1 = NULL, *link2 = NULL;
322         struct bpf_link *retlink1 = NULL, *retlink2 = NULL;
323         ssize_t uprobe_offset;
324
325         uprobe_offset = get_uprobe_offset(&trigger_func);
326         if (!ASSERT_GE(uprobe_offset, 0, "uprobe_offset"))
327                 goto cleanup;
328
329         /* attach two uprobes */
330         opts.bpf_cookie = 0x100;
331         opts.retprobe = false;
332         link1 = bpf_program__attach_uprobe_opts(skel->progs.handle_uprobe, 0 /* self pid */,
333                                                 "/proc/self/exe", uprobe_offset, &opts);
334         if (!ASSERT_OK_PTR(link1, "link1"))
335                 goto cleanup;
336
337         opts.bpf_cookie = 0x200;
338         opts.retprobe = false;
339         link2 = bpf_program__attach_uprobe_opts(skel->progs.handle_uprobe, -1 /* any pid */,
340                                                 "/proc/self/exe", uprobe_offset, &opts);
341         if (!ASSERT_OK_PTR(link2, "link2"))
342                 goto cleanup;
343
344         /* attach two uretprobes */
345         opts.bpf_cookie = 0x1000;
346         opts.retprobe = true;
347         retlink1 = bpf_program__attach_uprobe_opts(skel->progs.handle_uretprobe, -1 /* any pid */,
348                                                    "/proc/self/exe", uprobe_offset, &opts);
349         if (!ASSERT_OK_PTR(retlink1, "retlink1"))
350                 goto cleanup;
351
352         opts.bpf_cookie = 0x2000;
353         opts.retprobe = true;
354         retlink2 = bpf_program__attach_uprobe_opts(skel->progs.handle_uretprobe, 0 /* self pid */,
355                                                    "/proc/self/exe", uprobe_offset, &opts);
356         if (!ASSERT_OK_PTR(retlink2, "retlink2"))
357                 goto cleanup;
358
359         /* trigger uprobe && uretprobe */
360         trigger_func();
361
362         ASSERT_EQ(skel->bss->uprobe_res, 0x100 | 0x200, "uprobe_res");
363         ASSERT_EQ(skel->bss->uretprobe_res, 0x1000 | 0x2000, "uretprobe_res");
364
365 cleanup:
366         bpf_link__destroy(link1);
367         bpf_link__destroy(link2);
368         bpf_link__destroy(retlink1);
369         bpf_link__destroy(retlink2);
370 }
371
372 static void tp_subtest(struct test_bpf_cookie *skel)
373 {
374         DECLARE_LIBBPF_OPTS(bpf_tracepoint_opts, opts);
375         struct bpf_link *link1 = NULL, *link2 = NULL, *link3 = NULL;
376
377         /* attach first tp prog */
378         opts.bpf_cookie = 0x10000;
379         link1 = bpf_program__attach_tracepoint_opts(skel->progs.handle_tp1,
380                                                     "syscalls", "sys_enter_nanosleep", &opts);
381         if (!ASSERT_OK_PTR(link1, "link1"))
382                 goto cleanup;
383
384         /* attach second tp prog */
385         opts.bpf_cookie = 0x20000;
386         link2 = bpf_program__attach_tracepoint_opts(skel->progs.handle_tp2,
387                                                     "syscalls", "sys_enter_nanosleep", &opts);
388         if (!ASSERT_OK_PTR(link2, "link2"))
389                 goto cleanup;
390
391         /* trigger tracepoints */
392         usleep(1);
393
394         ASSERT_EQ(skel->bss->tp_res, 0x10000 | 0x20000, "tp_res1");
395
396         /* now we detach first prog and will attach third one, which causes
397          * two internal calls to bpf_prog_array_copy(), shuffling
398          * bpf_prog_array_items around. We test here that we don't lose track
399          * of associated bpf_cookies.
400          */
401         bpf_link__destroy(link1);
402         link1 = NULL;
403         kern_sync_rcu();
404         skel->bss->tp_res = 0;
405
406         /* attach third tp prog */
407         opts.bpf_cookie = 0x40000;
408         link3 = bpf_program__attach_tracepoint_opts(skel->progs.handle_tp3,
409                                                     "syscalls", "sys_enter_nanosleep", &opts);
410         if (!ASSERT_OK_PTR(link3, "link3"))
411                 goto cleanup;
412
413         /* trigger tracepoints */
414         usleep(1);
415
416         ASSERT_EQ(skel->bss->tp_res, 0x20000 | 0x40000, "tp_res2");
417
418 cleanup:
419         bpf_link__destroy(link1);
420         bpf_link__destroy(link2);
421         bpf_link__destroy(link3);
422 }
423
424 static void burn_cpu(void)
425 {
426         volatile int j = 0;
427         cpu_set_t cpu_set;
428         int i, err;
429
430         /* generate some branches on cpu 0 */
431         CPU_ZERO(&cpu_set);
432         CPU_SET(0, &cpu_set);
433         err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
434         ASSERT_OK(err, "set_thread_affinity");
435
436         /* spin the loop for a while (random high number) */
437         for (i = 0; i < 1000000; ++i)
438                 ++j;
439 }
440
441 static void pe_subtest(struct test_bpf_cookie *skel)
442 {
443         DECLARE_LIBBPF_OPTS(bpf_perf_event_opts, opts);
444         struct bpf_link *link = NULL;
445         struct perf_event_attr attr;
446         int pfd = -1;
447
448         /* create perf event */
449         memset(&attr, 0, sizeof(attr));
450         attr.size = sizeof(attr);
451         attr.type = PERF_TYPE_SOFTWARE;
452         attr.config = PERF_COUNT_SW_CPU_CLOCK;
453         attr.freq = 1;
454         attr.sample_freq = 10000;
455         pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
456         if (!ASSERT_GE(pfd, 0, "perf_fd"))
457                 goto cleanup;
458
459         opts.bpf_cookie = 0x100000;
460         link = bpf_program__attach_perf_event_opts(skel->progs.handle_pe, pfd, &opts);
461         if (!ASSERT_OK_PTR(link, "link1"))
462                 goto cleanup;
463
464         burn_cpu(); /* trigger BPF prog */
465
466         ASSERT_EQ(skel->bss->pe_res, 0x100000, "pe_res1");
467
468         /* prevent bpf_link__destroy() closing pfd itself */
469         bpf_link__disconnect(link);
470         /* close BPF link's FD explicitly */
471         close(bpf_link__fd(link));
472         /* free up memory used by struct bpf_link */
473         bpf_link__destroy(link);
474         link = NULL;
475         kern_sync_rcu();
476         skel->bss->pe_res = 0;
477
478         opts.bpf_cookie = 0x200000;
479         link = bpf_program__attach_perf_event_opts(skel->progs.handle_pe, pfd, &opts);
480         if (!ASSERT_OK_PTR(link, "link2"))
481                 goto cleanup;
482
483         burn_cpu(); /* trigger BPF prog */
484
485         ASSERT_EQ(skel->bss->pe_res, 0x200000, "pe_res2");
486
487 cleanup:
488         close(pfd);
489         bpf_link__destroy(link);
490 }
491
492 static void tracing_subtest(struct test_bpf_cookie *skel)
493 {
494         __u64 cookie;
495         int prog_fd;
496         int fentry_fd = -1, fexit_fd = -1, fmod_ret_fd = -1;
497         LIBBPF_OPTS(bpf_test_run_opts, opts);
498         LIBBPF_OPTS(bpf_link_create_opts, link_opts);
499
500         skel->bss->fentry_res = 0;
501         skel->bss->fexit_res = 0;
502
503         cookie = 0x10000000000000L;
504         prog_fd = bpf_program__fd(skel->progs.fentry_test1);
505         link_opts.tracing.cookie = cookie;
506         fentry_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_FENTRY, &link_opts);
507         if (!ASSERT_GE(fentry_fd, 0, "fentry.link_create"))
508                 goto cleanup;
509
510         cookie = 0x20000000000000L;
511         prog_fd = bpf_program__fd(skel->progs.fexit_test1);
512         link_opts.tracing.cookie = cookie;
513         fexit_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_FEXIT, &link_opts);
514         if (!ASSERT_GE(fexit_fd, 0, "fexit.link_create"))
515                 goto cleanup;
516
517         cookie = 0x30000000000000L;
518         prog_fd = bpf_program__fd(skel->progs.fmod_ret_test);
519         link_opts.tracing.cookie = cookie;
520         fmod_ret_fd = bpf_link_create(prog_fd, 0, BPF_MODIFY_RETURN, &link_opts);
521         if (!ASSERT_GE(fmod_ret_fd, 0, "fmod_ret.link_create"))
522                 goto cleanup;
523
524         prog_fd = bpf_program__fd(skel->progs.fentry_test1);
525         bpf_prog_test_run_opts(prog_fd, &opts);
526
527         prog_fd = bpf_program__fd(skel->progs.fmod_ret_test);
528         bpf_prog_test_run_opts(prog_fd, &opts);
529
530         ASSERT_EQ(skel->bss->fentry_res, 0x10000000000000L, "fentry_res");
531         ASSERT_EQ(skel->bss->fexit_res, 0x20000000000000L, "fexit_res");
532         ASSERT_EQ(skel->bss->fmod_ret_res, 0x30000000000000L, "fmod_ret_res");
533
534 cleanup:
535         if (fentry_fd >= 0)
536                 close(fentry_fd);
537         if (fexit_fd >= 0)
538                 close(fexit_fd);
539         if (fmod_ret_fd >= 0)
540                 close(fmod_ret_fd);
541 }
542
543 int stack_mprotect(void);
544
545 static void lsm_subtest(struct test_bpf_cookie *skel)
546 {
547         __u64 cookie;
548         int prog_fd;
549         int lsm_fd = -1;
550         LIBBPF_OPTS(bpf_link_create_opts, link_opts);
551         int err;
552
553         skel->bss->lsm_res = 0;
554
555         cookie = 0x90000000000090L;
556         prog_fd = bpf_program__fd(skel->progs.test_int_hook);
557         link_opts.tracing.cookie = cookie;
558         lsm_fd = bpf_link_create(prog_fd, 0, BPF_LSM_MAC, &link_opts);
559         if (!ASSERT_GE(lsm_fd, 0, "lsm.link_create"))
560                 goto cleanup;
561
562         err = stack_mprotect();
563         if (!ASSERT_EQ(err, -1, "stack_mprotect") ||
564             !ASSERT_EQ(errno, EPERM, "stack_mprotect"))
565                 goto cleanup;
566
567         usleep(1);
568
569         ASSERT_EQ(skel->bss->lsm_res, 0x90000000000090L, "fentry_res");
570
571 cleanup:
572         if (lsm_fd >= 0)
573                 close(lsm_fd);
574 }
575
576 static void tp_btf_subtest(struct test_bpf_cookie *skel)
577 {
578         __u64 cookie;
579         int prog_fd, link_fd = -1;
580         struct bpf_link *link = NULL;
581         LIBBPF_OPTS(bpf_link_create_opts, link_opts);
582         LIBBPF_OPTS(bpf_raw_tp_opts, raw_tp_opts);
583         LIBBPF_OPTS(bpf_trace_opts, trace_opts);
584
585         /* There are three different ways to attach tp_btf (BTF-aware raw
586          * tracepoint) programs. Let's test all of them.
587          */
588         prog_fd = bpf_program__fd(skel->progs.handle_tp_btf);
589
590         /* low-level BPF_RAW_TRACEPOINT_OPEN command wrapper */
591         skel->bss->tp_btf_res = 0;
592
593         raw_tp_opts.cookie = cookie = 0x11000000000000L;
594         link_fd = bpf_raw_tracepoint_open_opts(prog_fd, &raw_tp_opts);
595         if (!ASSERT_GE(link_fd, 0, "bpf_raw_tracepoint_open_opts"))
596                 goto cleanup;
597
598         usleep(1); /* trigger */
599         close(link_fd); /* detach */
600         link_fd = -1;
601
602         ASSERT_EQ(skel->bss->tp_btf_res, cookie, "raw_tp_open_res");
603
604         /* low-level generic bpf_link_create() API */
605         skel->bss->tp_btf_res = 0;
606
607         link_opts.tracing.cookie = cookie = 0x22000000000000L;
608         link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_RAW_TP, &link_opts);
609         if (!ASSERT_GE(link_fd, 0, "bpf_link_create"))
610                 goto cleanup;
611
612         usleep(1); /* trigger */
613         close(link_fd); /* detach */
614         link_fd = -1;
615
616         ASSERT_EQ(skel->bss->tp_btf_res, cookie, "link_create_res");
617
618         /* high-level bpf_link-based bpf_program__attach_trace_opts() API */
619         skel->bss->tp_btf_res = 0;
620
621         trace_opts.cookie = cookie = 0x33000000000000L;
622         link = bpf_program__attach_trace_opts(skel->progs.handle_tp_btf, &trace_opts);
623         if (!ASSERT_OK_PTR(link, "attach_trace_opts"))
624                 goto cleanup;
625
626         usleep(1); /* trigger */
627         bpf_link__destroy(link); /* detach */
628         link = NULL;
629
630         ASSERT_EQ(skel->bss->tp_btf_res, cookie, "attach_trace_opts_res");
631
632 cleanup:
633         if (link_fd >= 0)
634                 close(link_fd);
635         bpf_link__destroy(link);
636 }
637
638 static void raw_tp_subtest(struct test_bpf_cookie *skel)
639 {
640         __u64 cookie;
641         int prog_fd, link_fd = -1;
642         struct bpf_link *link = NULL;
643         LIBBPF_OPTS(bpf_raw_tp_opts, raw_tp_opts);
644         LIBBPF_OPTS(bpf_raw_tracepoint_opts, opts);
645
646         /* There are two different ways to attach raw_tp programs */
647         prog_fd = bpf_program__fd(skel->progs.handle_raw_tp);
648
649         /* low-level BPF_RAW_TRACEPOINT_OPEN command wrapper */
650         skel->bss->raw_tp_res = 0;
651
652         raw_tp_opts.tp_name = "sys_enter";
653         raw_tp_opts.cookie = cookie = 0x55000000000000L;
654         link_fd = bpf_raw_tracepoint_open_opts(prog_fd, &raw_tp_opts);
655         if (!ASSERT_GE(link_fd, 0, "bpf_raw_tracepoint_open_opts"))
656                 goto cleanup;
657
658         usleep(1); /* trigger */
659         close(link_fd); /* detach */
660         link_fd = -1;
661
662         ASSERT_EQ(skel->bss->raw_tp_res, cookie, "raw_tp_open_res");
663
664         /* high-level bpf_link-based bpf_program__attach_raw_tracepoint_opts() API */
665         skel->bss->raw_tp_res = 0;
666
667         opts.cookie = cookie = 0x66000000000000L;
668         link = bpf_program__attach_raw_tracepoint_opts(skel->progs.handle_raw_tp,
669                                                        "sys_enter", &opts);
670         if (!ASSERT_OK_PTR(link, "attach_raw_tp_opts"))
671                 goto cleanup;
672
673         usleep(1); /* trigger */
674         bpf_link__destroy(link); /* detach */
675         link = NULL;
676
677         ASSERT_EQ(skel->bss->raw_tp_res, cookie, "attach_raw_tp_opts_res");
678
679 cleanup:
680         if (link_fd >= 0)
681                 close(link_fd);
682         bpf_link__destroy(link);
683 }
684
685 void test_bpf_cookie(void)
686 {
687         struct test_bpf_cookie *skel;
688
689         skel = test_bpf_cookie__open_and_load();
690         if (!ASSERT_OK_PTR(skel, "skel_open"))
691                 return;
692
693         skel->bss->my_tid = sys_gettid();
694
695         if (test__start_subtest("kprobe"))
696                 kprobe_subtest(skel);
697         if (test__start_subtest("multi_kprobe_link_api"))
698                 kprobe_multi_link_api_subtest();
699         if (test__start_subtest("multi_kprobe_attach_api"))
700                 kprobe_multi_attach_api_subtest();
701         if (test__start_subtest("uprobe"))
702                 uprobe_subtest(skel);
703         if (test__start_subtest("multi_uprobe_attach_api"))
704                 uprobe_multi_attach_api_subtest();
705         if (test__start_subtest("tracepoint"))
706                 tp_subtest(skel);
707         if (test__start_subtest("perf_event"))
708                 pe_subtest(skel);
709         if (test__start_subtest("trampoline"))
710                 tracing_subtest(skel);
711         if (test__start_subtest("lsm"))
712                 lsm_subtest(skel);
713         if (test__start_subtest("tp_btf"))
714                 tp_btf_subtest(skel);
715         if (test__start_subtest("raw_tp"))
716                 raw_tp_subtest(skel);
717         test_bpf_cookie__destroy(skel);
718 }
This page took 0.073709 seconds and 4 git commands to generate.