Despite multiple attempts to get the syscall number assignment right for the newly added uretprobe syscall, we ended up with a bit of a mess: - The number is defined as 467 based on the assumption that the xattrat family of syscalls would use 463 through 466, but those did not make it into 6.11. - The include/uapi/asm-generic/unistd.h file still lists the number 463, but the new scripts/syscall.tbl that was supposed to have the same data lists 467 instead as the number for arc, arm64, csky, hexagon, loongarch, nios2, openrisc and riscv. None of these architectures actually provide a uretprobe syscall. - All the other architectures (powerpc, arm, mips, ...) don't list this syscall at all. There are two ways to make it consistent again: either list it with the same syscall number on all architectures, or only list it on x86 but not in scripts/syscall.tbl and asm-generic/unistd.h. Based on the most recent discussion, it seems like we won't need it anywhere else, so just remove the inconsistent assignment and instead move the x86 number to the next available one in the architecture specific range, which is 335. Fixes: 5c28424e9a34 ("syscalls: Fix to add sys_uretprobe to syscall.tbl") Fixes: 190fec72df4a ("uprobe: Wire up uretprobe system call") Fixes: 63ded110979b ("uprobe: Change uretprobe syscall scope and number") Acked-by: Masami Hiramatsu (Google) <mhiramat@kernel.org> Reviewed-by: Jiri Olsa <jolsa@kernel.org> Signed-off-by: Arnd Bergmann <arnd@arndb.de>
386 lines
8.9 KiB
C
386 lines
8.9 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
#include <test_progs.h>
|
|
|
|
#ifdef __x86_64__
|
|
|
|
#include <unistd.h>
|
|
#include <asm/ptrace.h>
|
|
#include <linux/compiler.h>
|
|
#include <linux/stringify.h>
|
|
#include <sys/wait.h>
|
|
#include <sys/syscall.h>
|
|
#include <sys/prctl.h>
|
|
#include <asm/prctl.h>
|
|
#include "uprobe_syscall.skel.h"
|
|
#include "uprobe_syscall_executed.skel.h"
|
|
|
|
__naked unsigned long uretprobe_regs_trigger(void)
|
|
{
|
|
asm volatile (
|
|
"movq $0xdeadbeef, %rax\n"
|
|
"ret\n"
|
|
);
|
|
}
|
|
|
|
__naked void uretprobe_regs(struct pt_regs *before, struct pt_regs *after)
|
|
{
|
|
asm volatile (
|
|
"movq %r15, 0(%rdi)\n"
|
|
"movq %r14, 8(%rdi)\n"
|
|
"movq %r13, 16(%rdi)\n"
|
|
"movq %r12, 24(%rdi)\n"
|
|
"movq %rbp, 32(%rdi)\n"
|
|
"movq %rbx, 40(%rdi)\n"
|
|
"movq %r11, 48(%rdi)\n"
|
|
"movq %r10, 56(%rdi)\n"
|
|
"movq %r9, 64(%rdi)\n"
|
|
"movq %r8, 72(%rdi)\n"
|
|
"movq %rax, 80(%rdi)\n"
|
|
"movq %rcx, 88(%rdi)\n"
|
|
"movq %rdx, 96(%rdi)\n"
|
|
"movq %rsi, 104(%rdi)\n"
|
|
"movq %rdi, 112(%rdi)\n"
|
|
"movq $0, 120(%rdi)\n" /* orig_rax */
|
|
"movq $0, 128(%rdi)\n" /* rip */
|
|
"movq $0, 136(%rdi)\n" /* cs */
|
|
"pushf\n"
|
|
"pop %rax\n"
|
|
"movq %rax, 144(%rdi)\n" /* eflags */
|
|
"movq %rsp, 152(%rdi)\n" /* rsp */
|
|
"movq $0, 160(%rdi)\n" /* ss */
|
|
|
|
/* save 2nd argument */
|
|
"pushq %rsi\n"
|
|
"call uretprobe_regs_trigger\n"
|
|
|
|
/* save return value and load 2nd argument pointer to rax */
|
|
"pushq %rax\n"
|
|
"movq 8(%rsp), %rax\n"
|
|
|
|
"movq %r15, 0(%rax)\n"
|
|
"movq %r14, 8(%rax)\n"
|
|
"movq %r13, 16(%rax)\n"
|
|
"movq %r12, 24(%rax)\n"
|
|
"movq %rbp, 32(%rax)\n"
|
|
"movq %rbx, 40(%rax)\n"
|
|
"movq %r11, 48(%rax)\n"
|
|
"movq %r10, 56(%rax)\n"
|
|
"movq %r9, 64(%rax)\n"
|
|
"movq %r8, 72(%rax)\n"
|
|
"movq %rcx, 88(%rax)\n"
|
|
"movq %rdx, 96(%rax)\n"
|
|
"movq %rsi, 104(%rax)\n"
|
|
"movq %rdi, 112(%rax)\n"
|
|
"movq $0, 120(%rax)\n" /* orig_rax */
|
|
"movq $0, 128(%rax)\n" /* rip */
|
|
"movq $0, 136(%rax)\n" /* cs */
|
|
|
|
/* restore return value and 2nd argument */
|
|
"pop %rax\n"
|
|
"pop %rsi\n"
|
|
|
|
"movq %rax, 80(%rsi)\n"
|
|
|
|
"pushf\n"
|
|
"pop %rax\n"
|
|
|
|
"movq %rax, 144(%rsi)\n" /* eflags */
|
|
"movq %rsp, 152(%rsi)\n" /* rsp */
|
|
"movq $0, 160(%rsi)\n" /* ss */
|
|
"ret\n"
|
|
);
|
|
}
|
|
|
|
static void test_uretprobe_regs_equal(void)
|
|
{
|
|
struct uprobe_syscall *skel = NULL;
|
|
struct pt_regs before = {}, after = {};
|
|
unsigned long *pb = (unsigned long *) &before;
|
|
unsigned long *pa = (unsigned long *) &after;
|
|
unsigned long *pp;
|
|
unsigned int i, cnt;
|
|
int err;
|
|
|
|
skel = uprobe_syscall__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_syscall__open_and_load"))
|
|
goto cleanup;
|
|
|
|
err = uprobe_syscall__attach(skel);
|
|
if (!ASSERT_OK(err, "uprobe_syscall__attach"))
|
|
goto cleanup;
|
|
|
|
uretprobe_regs(&before, &after);
|
|
|
|
pp = (unsigned long *) &skel->bss->regs;
|
|
cnt = sizeof(before)/sizeof(*pb);
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
unsigned int offset = i * sizeof(unsigned long);
|
|
|
|
/*
|
|
* Check register before and after uretprobe_regs_trigger call
|
|
* that triggers the uretprobe.
|
|
*/
|
|
switch (offset) {
|
|
case offsetof(struct pt_regs, rax):
|
|
ASSERT_EQ(pa[i], 0xdeadbeef, "return value");
|
|
break;
|
|
default:
|
|
if (!ASSERT_EQ(pb[i], pa[i], "register before-after value check"))
|
|
fprintf(stdout, "failed register offset %u\n", offset);
|
|
}
|
|
|
|
/*
|
|
* Check register seen from bpf program and register after
|
|
* uretprobe_regs_trigger call
|
|
*/
|
|
switch (offset) {
|
|
/*
|
|
* These values will be different (not set in uretprobe_regs),
|
|
* we don't care.
|
|
*/
|
|
case offsetof(struct pt_regs, orig_rax):
|
|
case offsetof(struct pt_regs, rip):
|
|
case offsetof(struct pt_regs, cs):
|
|
case offsetof(struct pt_regs, rsp):
|
|
case offsetof(struct pt_regs, ss):
|
|
break;
|
|
default:
|
|
if (!ASSERT_EQ(pp[i], pa[i], "register prog-after value check"))
|
|
fprintf(stdout, "failed register offset %u\n", offset);
|
|
}
|
|
}
|
|
|
|
cleanup:
|
|
uprobe_syscall__destroy(skel);
|
|
}
|
|
|
|
#define BPF_TESTMOD_UPROBE_TEST_FILE "/sys/kernel/bpf_testmod_uprobe"
|
|
|
|
static int write_bpf_testmod_uprobe(unsigned long offset)
|
|
{
|
|
size_t n, ret;
|
|
char buf[30];
|
|
int fd;
|
|
|
|
n = sprintf(buf, "%lu", offset);
|
|
|
|
fd = open(BPF_TESTMOD_UPROBE_TEST_FILE, O_WRONLY);
|
|
if (fd < 0)
|
|
return -errno;
|
|
|
|
ret = write(fd, buf, n);
|
|
close(fd);
|
|
return ret != n ? (int) ret : 0;
|
|
}
|
|
|
|
static void test_uretprobe_regs_change(void)
|
|
{
|
|
struct pt_regs before = {}, after = {};
|
|
unsigned long *pb = (unsigned long *) &before;
|
|
unsigned long *pa = (unsigned long *) &after;
|
|
unsigned long cnt = sizeof(before)/sizeof(*pb);
|
|
unsigned int i, err, offset;
|
|
|
|
offset = get_uprobe_offset(uretprobe_regs_trigger);
|
|
|
|
err = write_bpf_testmod_uprobe(offset);
|
|
if (!ASSERT_OK(err, "register_uprobe"))
|
|
return;
|
|
|
|
uretprobe_regs(&before, &after);
|
|
|
|
err = write_bpf_testmod_uprobe(0);
|
|
if (!ASSERT_OK(err, "unregister_uprobe"))
|
|
return;
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
unsigned int offset = i * sizeof(unsigned long);
|
|
|
|
switch (offset) {
|
|
case offsetof(struct pt_regs, rax):
|
|
ASSERT_EQ(pa[i], 0x12345678deadbeef, "rax");
|
|
break;
|
|
case offsetof(struct pt_regs, rcx):
|
|
ASSERT_EQ(pa[i], 0x87654321feebdaed, "rcx");
|
|
break;
|
|
case offsetof(struct pt_regs, r11):
|
|
ASSERT_EQ(pa[i], (__u64) -1, "r11");
|
|
break;
|
|
default:
|
|
if (!ASSERT_EQ(pa[i], pb[i], "register before-after value check"))
|
|
fprintf(stdout, "failed register offset %u\n", offset);
|
|
}
|
|
}
|
|
}
|
|
|
|
#ifndef __NR_uretprobe
|
|
#define __NR_uretprobe 335
|
|
#endif
|
|
|
|
__naked unsigned long uretprobe_syscall_call_1(void)
|
|
{
|
|
/*
|
|
* Pretend we are uretprobe trampoline to trigger the return
|
|
* probe invocation in order to verify we get SIGILL.
|
|
*/
|
|
asm volatile (
|
|
"pushq %rax\n"
|
|
"pushq %rcx\n"
|
|
"pushq %r11\n"
|
|
"movq $" __stringify(__NR_uretprobe) ", %rax\n"
|
|
"syscall\n"
|
|
"popq %r11\n"
|
|
"popq %rcx\n"
|
|
"retq\n"
|
|
);
|
|
}
|
|
|
|
__naked unsigned long uretprobe_syscall_call(void)
|
|
{
|
|
asm volatile (
|
|
"call uretprobe_syscall_call_1\n"
|
|
"retq\n"
|
|
);
|
|
}
|
|
|
|
static void test_uretprobe_syscall_call(void)
|
|
{
|
|
LIBBPF_OPTS(bpf_uprobe_multi_opts, opts,
|
|
.retprobe = true,
|
|
);
|
|
struct uprobe_syscall_executed *skel;
|
|
int pid, status, err, go[2], c;
|
|
|
|
if (ASSERT_OK(pipe(go), "pipe"))
|
|
return;
|
|
|
|
skel = uprobe_syscall_executed__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_syscall_executed__open_and_load"))
|
|
goto cleanup;
|
|
|
|
pid = fork();
|
|
if (!ASSERT_GE(pid, 0, "fork"))
|
|
goto cleanup;
|
|
|
|
/* child */
|
|
if (pid == 0) {
|
|
close(go[1]);
|
|
|
|
/* wait for parent's kick */
|
|
err = read(go[0], &c, 1);
|
|
if (err != 1)
|
|
exit(-1);
|
|
|
|
uretprobe_syscall_call();
|
|
_exit(0);
|
|
}
|
|
|
|
skel->links.test = bpf_program__attach_uprobe_multi(skel->progs.test, pid,
|
|
"/proc/self/exe",
|
|
"uretprobe_syscall_call", &opts);
|
|
if (!ASSERT_OK_PTR(skel->links.test, "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
|
|
/* kick the child */
|
|
write(go[1], &c, 1);
|
|
err = waitpid(pid, &status, 0);
|
|
ASSERT_EQ(err, pid, "waitpid");
|
|
|
|
/* verify the child got killed with SIGILL */
|
|
ASSERT_EQ(WIFSIGNALED(status), 1, "WIFSIGNALED");
|
|
ASSERT_EQ(WTERMSIG(status), SIGILL, "WTERMSIG");
|
|
|
|
/* verify the uretprobe program wasn't called */
|
|
ASSERT_EQ(skel->bss->executed, 0, "executed");
|
|
|
|
cleanup:
|
|
uprobe_syscall_executed__destroy(skel);
|
|
close(go[1]);
|
|
close(go[0]);
|
|
}
|
|
|
|
/*
|
|
* Borrowed from tools/testing/selftests/x86/test_shadow_stack.c.
|
|
*
|
|
* For use in inline enablement of shadow stack.
|
|
*
|
|
* The program can't return from the point where shadow stack gets enabled
|
|
* because there will be no address on the shadow stack. So it can't use
|
|
* syscall() for enablement, since it is a function.
|
|
*
|
|
* Based on code from nolibc.h. Keep a copy here because this can't pull
|
|
* in all of nolibc.h.
|
|
*/
|
|
#define ARCH_PRCTL(arg1, arg2) \
|
|
({ \
|
|
long _ret; \
|
|
register long _num asm("eax") = __NR_arch_prctl; \
|
|
register long _arg1 asm("rdi") = (long)(arg1); \
|
|
register long _arg2 asm("rsi") = (long)(arg2); \
|
|
\
|
|
asm volatile ( \
|
|
"syscall\n" \
|
|
: "=a"(_ret) \
|
|
: "r"(_arg1), "r"(_arg2), \
|
|
"0"(_num) \
|
|
: "rcx", "r11", "memory", "cc" \
|
|
); \
|
|
_ret; \
|
|
})
|
|
|
|
#ifndef ARCH_SHSTK_ENABLE
|
|
#define ARCH_SHSTK_ENABLE 0x5001
|
|
#define ARCH_SHSTK_DISABLE 0x5002
|
|
#define ARCH_SHSTK_SHSTK (1ULL << 0)
|
|
#endif
|
|
|
|
static void test_uretprobe_shadow_stack(void)
|
|
{
|
|
if (ARCH_PRCTL(ARCH_SHSTK_ENABLE, ARCH_SHSTK_SHSTK)) {
|
|
test__skip();
|
|
return;
|
|
}
|
|
|
|
/* Run all of the uretprobe tests. */
|
|
test_uretprobe_regs_equal();
|
|
test_uretprobe_regs_change();
|
|
test_uretprobe_syscall_call();
|
|
|
|
ARCH_PRCTL(ARCH_SHSTK_DISABLE, ARCH_SHSTK_SHSTK);
|
|
}
|
|
#else
|
|
static void test_uretprobe_regs_equal(void)
|
|
{
|
|
test__skip();
|
|
}
|
|
|
|
static void test_uretprobe_regs_change(void)
|
|
{
|
|
test__skip();
|
|
}
|
|
|
|
static void test_uretprobe_syscall_call(void)
|
|
{
|
|
test__skip();
|
|
}
|
|
|
|
static void test_uretprobe_shadow_stack(void)
|
|
{
|
|
test__skip();
|
|
}
|
|
#endif
|
|
|
|
void test_uprobe_syscall(void)
|
|
{
|
|
if (test__start_subtest("uretprobe_regs_equal"))
|
|
test_uretprobe_regs_equal();
|
|
if (test__start_subtest("uretprobe_regs_change"))
|
|
test_uretprobe_regs_change();
|
|
if (test__start_subtest("uretprobe_syscall_call"))
|
|
test_uretprobe_syscall_call();
|
|
if (test__start_subtest("uretprobe_shadow_stack"))
|
|
test_uretprobe_shadow_stack();
|
|
}
|