Add fmod_ret BPF program to existing test_overhead selftest. Also re-implement user-space benchmarking part into benchmark runner to compare results. Results with ./bench are consistently somewhat lower than test_overhead's, but relative performance of various types of BPF programs stay consisten (e.g., kretprobe is noticeably slower). This slowdown seems to be coming from the fact that test_overhead is single-threaded, while benchmark always spins off at least one thread for producer. This has been confirmed by hacking multi-threaded test_overhead variant and also single-threaded bench variant. Resutls are below. run_bench_rename.sh script from benchs/ subdirectory was used to produce results for ./bench. Single-threaded implementations =============================== /* bench: single-threaded, atomics */ base : 4.622 ± 0.049M/s kprobe : 3.673 ± 0.052M/s kretprobe : 2.625 ± 0.052M/s rawtp : 4.369 ± 0.089M/s fentry : 4.201 ± 0.558M/s fexit : 4.309 ± 0.148M/s fmodret : 4.314 ± 0.203M/s /* selftest: single-threaded, no atomics */ task_rename base 4555K events per sec task_rename kprobe 3643K events per sec task_rename kretprobe 2506K events per sec task_rename raw_tp 4303K events per sec task_rename fentry 4307K events per sec task_rename fexit 4010K events per sec task_rename fmod_ret 3984K events per sec Multi-threaded implementations ============================== /* bench: multi-threaded w/ atomics */ base : 3.910 ± 0.023M/s kprobe : 3.048 ± 0.037M/s kretprobe : 2.300 ± 0.015M/s rawtp : 3.687 ± 0.034M/s fentry : 3.740 ± 0.087M/s fexit : 3.510 ± 0.009M/s fmodret : 3.485 ± 0.050M/s /* selftest: multi-threaded w/ atomics */ task_rename base 3872K events per sec task_rename kprobe 3068K events per sec task_rename kretprobe 2350K events per sec task_rename raw_tp 3731K events per sec task_rename fentry 3639K events per sec task_rename fexit 3558K events per sec task_rename fmod_ret 3511K events per sec /* selftest: multi-threaded, no atomics */ task_rename base 3945K events per sec task_rename kprobe 3298K events per sec task_rename kretprobe 2451K events per sec task_rename raw_tp 3718K events per sec task_rename fentry 3782K events per sec task_rename fexit 3543K events per sec task_rename fmod_ret 3526K events per sec Note that the fact that ./bench benchmark always uses atomic increments for counting, while test_overhead doesn't, doesn't influence test results all that much. Signed-off-by: Andrii Nakryiko <andriin@fb.com> Signed-off-by: Alexei Starovoitov <ast@kernel.org> Acked-by: John Fastabend <john.fastabend@gmail.com> Acked-by: Yonghong Song <yhs@fb.com> Link: https://lore.kernel.org/bpf/20200512192445.2351848-4-andriin@fb.com
161 lines
4.4 KiB
C
161 lines
4.4 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (c) 2019 Facebook */
|
|
#define _GNU_SOURCE
|
|
#include <sched.h>
|
|
#include <sys/prctl.h>
|
|
#include <test_progs.h>
|
|
|
|
#define MAX_CNT 100000
|
|
|
|
static __u64 time_get_ns(void)
|
|
{
|
|
struct timespec ts;
|
|
|
|
clock_gettime(CLOCK_MONOTONIC, &ts);
|
|
return ts.tv_sec * 1000000000ull + ts.tv_nsec;
|
|
}
|
|
|
|
static int test_task_rename(const char *prog)
|
|
{
|
|
int i, fd, duration = 0, err;
|
|
char buf[] = "test_overhead";
|
|
__u64 start_time;
|
|
|
|
fd = open("/proc/self/comm", O_WRONLY|O_TRUNC);
|
|
if (CHECK(fd < 0, "open /proc", "err %d", errno))
|
|
return -1;
|
|
start_time = time_get_ns();
|
|
for (i = 0; i < MAX_CNT; i++) {
|
|
err = write(fd, buf, sizeof(buf));
|
|
if (err < 0) {
|
|
CHECK(err < 0, "task rename", "err %d", errno);
|
|
close(fd);
|
|
return -1;
|
|
}
|
|
}
|
|
printf("task_rename %s\t%lluK events per sec\n", prog,
|
|
MAX_CNT * 1000000ll / (time_get_ns() - start_time));
|
|
close(fd);
|
|
return 0;
|
|
}
|
|
|
|
static void test_run(const char *prog)
|
|
{
|
|
test_task_rename(prog);
|
|
}
|
|
|
|
static void setaffinity(void)
|
|
{
|
|
cpu_set_t cpuset;
|
|
int cpu = 0;
|
|
|
|
CPU_ZERO(&cpuset);
|
|
CPU_SET(cpu, &cpuset);
|
|
sched_setaffinity(0, sizeof(cpuset), &cpuset);
|
|
}
|
|
|
|
void test_test_overhead(void)
|
|
{
|
|
const char *kprobe_name = "kprobe/__set_task_comm";
|
|
const char *kretprobe_name = "kretprobe/__set_task_comm";
|
|
const char *raw_tp_name = "raw_tp/task_rename";
|
|
const char *fentry_name = "fentry/__set_task_comm";
|
|
const char *fexit_name = "fexit/__set_task_comm";
|
|
const char *fmodret_name = "fmod_ret/__set_task_comm";
|
|
const char *kprobe_func = "__set_task_comm";
|
|
struct bpf_program *kprobe_prog, *kretprobe_prog, *raw_tp_prog;
|
|
struct bpf_program *fentry_prog, *fexit_prog, *fmodret_prog;
|
|
struct bpf_object *obj;
|
|
struct bpf_link *link;
|
|
int err, duration = 0;
|
|
char comm[16] = {};
|
|
|
|
if (CHECK_FAIL(prctl(PR_GET_NAME, comm, 0L, 0L, 0L)))
|
|
return;
|
|
|
|
obj = bpf_object__open_file("./test_overhead.o", NULL);
|
|
if (CHECK(IS_ERR(obj), "obj_open_file", "err %ld\n", PTR_ERR(obj)))
|
|
return;
|
|
|
|
kprobe_prog = bpf_object__find_program_by_title(obj, kprobe_name);
|
|
if (CHECK(!kprobe_prog, "find_probe",
|
|
"prog '%s' not found\n", kprobe_name))
|
|
goto cleanup;
|
|
kretprobe_prog = bpf_object__find_program_by_title(obj, kretprobe_name);
|
|
if (CHECK(!kretprobe_prog, "find_probe",
|
|
"prog '%s' not found\n", kretprobe_name))
|
|
goto cleanup;
|
|
raw_tp_prog = bpf_object__find_program_by_title(obj, raw_tp_name);
|
|
if (CHECK(!raw_tp_prog, "find_probe",
|
|
"prog '%s' not found\n", raw_tp_name))
|
|
goto cleanup;
|
|
fentry_prog = bpf_object__find_program_by_title(obj, fentry_name);
|
|
if (CHECK(!fentry_prog, "find_probe",
|
|
"prog '%s' not found\n", fentry_name))
|
|
goto cleanup;
|
|
fexit_prog = bpf_object__find_program_by_title(obj, fexit_name);
|
|
if (CHECK(!fexit_prog, "find_probe",
|
|
"prog '%s' not found\n", fexit_name))
|
|
goto cleanup;
|
|
fmodret_prog = bpf_object__find_program_by_title(obj, fmodret_name);
|
|
if (CHECK(!fmodret_prog, "find_probe",
|
|
"prog '%s' not found\n", fmodret_name))
|
|
goto cleanup;
|
|
|
|
err = bpf_object__load(obj);
|
|
if (CHECK(err, "obj_load", "err %d\n", err))
|
|
goto cleanup;
|
|
|
|
setaffinity();
|
|
|
|
/* base line run */
|
|
test_run("base");
|
|
|
|
/* attach kprobe */
|
|
link = bpf_program__attach_kprobe(kprobe_prog, false /* retprobe */,
|
|
kprobe_func);
|
|
if (CHECK(IS_ERR(link), "attach_kprobe", "err %ld\n", PTR_ERR(link)))
|
|
goto cleanup;
|
|
test_run("kprobe");
|
|
bpf_link__destroy(link);
|
|
|
|
/* attach kretprobe */
|
|
link = bpf_program__attach_kprobe(kretprobe_prog, true /* retprobe */,
|
|
kprobe_func);
|
|
if (CHECK(IS_ERR(link), "attach kretprobe", "err %ld\n", PTR_ERR(link)))
|
|
goto cleanup;
|
|
test_run("kretprobe");
|
|
bpf_link__destroy(link);
|
|
|
|
/* attach raw_tp */
|
|
link = bpf_program__attach_raw_tracepoint(raw_tp_prog, "task_rename");
|
|
if (CHECK(IS_ERR(link), "attach fentry", "err %ld\n", PTR_ERR(link)))
|
|
goto cleanup;
|
|
test_run("raw_tp");
|
|
bpf_link__destroy(link);
|
|
|
|
/* attach fentry */
|
|
link = bpf_program__attach_trace(fentry_prog);
|
|
if (CHECK(IS_ERR(link), "attach fentry", "err %ld\n", PTR_ERR(link)))
|
|
goto cleanup;
|
|
test_run("fentry");
|
|
bpf_link__destroy(link);
|
|
|
|
/* attach fexit */
|
|
link = bpf_program__attach_trace(fexit_prog);
|
|
if (CHECK(IS_ERR(link), "attach fexit", "err %ld\n", PTR_ERR(link)))
|
|
goto cleanup;
|
|
test_run("fexit");
|
|
bpf_link__destroy(link);
|
|
|
|
/* attach fmod_ret */
|
|
link = bpf_program__attach_trace(fmodret_prog);
|
|
if (CHECK(IS_ERR(link), "attach fmod_ret", "err %ld\n", PTR_ERR(link)))
|
|
goto cleanup;
|
|
test_run("fmod_ret");
|
|
bpf_link__destroy(link);
|
|
cleanup:
|
|
prctl(PR_SET_NAME, comm, 0L, 0L, 0L);
|
|
bpf_object__close(obj);
|
|
}
|