97c11d6e31
Instead of adding the whole test to DENYLIST.s390x, which also has success test cases that should be run, just skip over failure test cases in case the JIT does not support kfuncs. Signed-off-by: Kumar Kartikeya Dwivedi <memxor@gmail.com> Link: https://lore.kernel.org/r/20221118185938.2139616-3-memxor@gmail.com Signed-off-by: Alexei Starovoitov <ast@kernel.org>
143 lines
4.8 KiB
C
143 lines
4.8 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <test_progs.h>
|
|
#include <network_helpers.h>
|
|
|
|
#include "test_spin_lock.skel.h"
|
|
#include "test_spin_lock_fail.skel.h"
|
|
|
|
static char log_buf[1024 * 1024];
|
|
|
|
static struct {
|
|
const char *prog_name;
|
|
const char *err_msg;
|
|
} spin_lock_fail_tests[] = {
|
|
{ "lock_id_kptr_preserve",
|
|
"5: (bf) r1 = r0 ; R0_w=ptr_foo(id=2,ref_obj_id=2,off=0,imm=0) "
|
|
"R1_w=ptr_foo(id=2,ref_obj_id=2,off=0,imm=0) refs=2\n6: (85) call bpf_this_cpu_ptr#154\n"
|
|
"R1 type=ptr_ expected=percpu_ptr_" },
|
|
{ "lock_id_global_zero",
|
|
"; R1_w=map_value(off=0,ks=4,vs=4,imm=0)\n2: (85) call bpf_this_cpu_ptr#154\n"
|
|
"R1 type=map_value expected=percpu_ptr_" },
|
|
{ "lock_id_mapval_preserve",
|
|
"8: (bf) r1 = r0 ; R0_w=map_value(id=1,off=0,ks=4,vs=8,imm=0) "
|
|
"R1_w=map_value(id=1,off=0,ks=4,vs=8,imm=0)\n9: (85) call bpf_this_cpu_ptr#154\n"
|
|
"R1 type=map_value expected=percpu_ptr_" },
|
|
{ "lock_id_innermapval_preserve",
|
|
"13: (bf) r1 = r0 ; R0=map_value(id=2,off=0,ks=4,vs=8,imm=0) "
|
|
"R1_w=map_value(id=2,off=0,ks=4,vs=8,imm=0)\n14: (85) call bpf_this_cpu_ptr#154\n"
|
|
"R1 type=map_value expected=percpu_ptr_" },
|
|
{ "lock_id_mismatch_kptr_kptr", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_kptr_global", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_kptr_mapval", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_kptr_innermapval", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_global_global", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_global_kptr", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_global_mapval", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_global_innermapval", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_mapval_mapval", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_mapval_kptr", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_mapval_global", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_mapval_innermapval", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_innermapval_innermapval1", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_innermapval_innermapval2", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_innermapval_kptr", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_innermapval_global", "bpf_spin_unlock of different lock" },
|
|
{ "lock_id_mismatch_innermapval_mapval", "bpf_spin_unlock of different lock" },
|
|
};
|
|
|
|
static void test_spin_lock_fail_prog(const char *prog_name, const char *err_msg)
|
|
{
|
|
LIBBPF_OPTS(bpf_object_open_opts, opts, .kernel_log_buf = log_buf,
|
|
.kernel_log_size = sizeof(log_buf),
|
|
.kernel_log_level = 1);
|
|
struct test_spin_lock_fail *skel;
|
|
struct bpf_program *prog;
|
|
int ret;
|
|
|
|
skel = test_spin_lock_fail__open_opts(&opts);
|
|
if (!ASSERT_OK_PTR(skel, "test_spin_lock_fail__open_opts"))
|
|
return;
|
|
|
|
prog = bpf_object__find_program_by_name(skel->obj, prog_name);
|
|
if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name"))
|
|
goto end;
|
|
|
|
bpf_program__set_autoload(prog, true);
|
|
|
|
ret = test_spin_lock_fail__load(skel);
|
|
if (!ASSERT_ERR(ret, "test_spin_lock_fail__load must fail"))
|
|
goto end;
|
|
|
|
/* Skip check if JIT does not support kfuncs */
|
|
if (strstr(log_buf, "JIT does not support calling kernel function")) {
|
|
test__skip();
|
|
goto end;
|
|
}
|
|
|
|
if (!ASSERT_OK_PTR(strstr(log_buf, err_msg), "expected error message")) {
|
|
fprintf(stderr, "Expected: %s\n", err_msg);
|
|
fprintf(stderr, "Verifier: %s\n", log_buf);
|
|
}
|
|
|
|
end:
|
|
test_spin_lock_fail__destroy(skel);
|
|
}
|
|
|
|
static void *spin_lock_thread(void *arg)
|
|
{
|
|
int err, prog_fd = *(u32 *) arg;
|
|
LIBBPF_OPTS(bpf_test_run_opts, topts,
|
|
.data_in = &pkt_v4,
|
|
.data_size_in = sizeof(pkt_v4),
|
|
.repeat = 10000,
|
|
);
|
|
|
|
err = bpf_prog_test_run_opts(prog_fd, &topts);
|
|
ASSERT_OK(err, "test_run");
|
|
ASSERT_OK(topts.retval, "test_run retval");
|
|
pthread_exit(arg);
|
|
}
|
|
|
|
void test_spin_lock_success(void)
|
|
{
|
|
struct test_spin_lock *skel;
|
|
pthread_t thread_id[4];
|
|
int prog_fd, i;
|
|
void *ret;
|
|
|
|
skel = test_spin_lock__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "test_spin_lock__open_and_load"))
|
|
return;
|
|
prog_fd = bpf_program__fd(skel->progs.bpf_spin_lock_test);
|
|
for (i = 0; i < 4; i++) {
|
|
int err;
|
|
|
|
err = pthread_create(&thread_id[i], NULL, &spin_lock_thread, &prog_fd);
|
|
if (!ASSERT_OK(err, "pthread_create"))
|
|
goto end;
|
|
}
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
if (!ASSERT_OK(pthread_join(thread_id[i], &ret), "pthread_join"))
|
|
goto end;
|
|
if (!ASSERT_EQ(ret, &prog_fd, "ret == prog_fd"))
|
|
goto end;
|
|
}
|
|
end:
|
|
test_spin_lock__destroy(skel);
|
|
}
|
|
|
|
void test_spin_lock(void)
|
|
{
|
|
int i;
|
|
|
|
test_spin_lock_success();
|
|
|
|
for (i = 0; i < ARRAY_SIZE(spin_lock_fail_tests); i++) {
|
|
if (!test__start_subtest(spin_lock_fail_tests[i].prog_name))
|
|
continue;
|
|
test_spin_lock_fail_prog(spin_lock_fail_tests[i].prog_name,
|
|
spin_lock_fail_tests[i].err_msg);
|
|
}
|
|
}
|