Static ksyms often have problems because the number of symbols exceeds the MAX_SYMS limit. Like changing the MAX_SYMS from 300000 to 400000 in commit e76a014334a6("selftests/bpf: Bump and validate MAX_SYMS") solves the problem somewhat, but it's not the perfect way. This commit uses dynamic memory allocation, which completely solves the problem caused by the limitation of the number of kallsyms. At the same time, add APIs: load_kallsyms_local() ksym_search_local() ksym_get_addr_local() free_kallsyms_local() There are used to solve the problem of selftests/bpf updating kallsyms after attach new symbols during testmod testing. Signed-off-by: Rong Tao <rongtao@cestc.cn> Signed-off-by: Andrii Nakryiko <andrii@kernel.org> Acked-by: Jiri Olsa <jolsa@kernel.org> Acked-by: Stanislav Fomichev <sdf@google.com> Link: https://lore.kernel.org/bpf/tencent_C9BDA68F9221F21BE4081566A55D66A9700A@qq.com
36 lines
928 B
C
36 lines
928 B
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __TRACE_HELPER_H
|
|
#define __TRACE_HELPER_H
|
|
|
|
#include <bpf/libbpf.h>
|
|
|
|
#define __ALIGN_MASK(x, mask) (((x)+(mask))&~(mask))
|
|
#define ALIGN(x, a) __ALIGN_MASK(x, (typeof(x))(a)-1)
|
|
|
|
struct ksym {
|
|
long addr;
|
|
char *name;
|
|
};
|
|
struct ksyms;
|
|
|
|
int load_kallsyms(void);
|
|
struct ksym *ksym_search(long key);
|
|
long ksym_get_addr(const char *name);
|
|
|
|
struct ksyms *load_kallsyms_local(void);
|
|
struct ksym *ksym_search_local(struct ksyms *ksyms, long key);
|
|
long ksym_get_addr_local(struct ksyms *ksyms, const char *name);
|
|
void free_kallsyms_local(struct ksyms *ksyms);
|
|
|
|
/* open kallsyms and find addresses on the fly, faster than load + search. */
|
|
int kallsyms_find(const char *sym, unsigned long long *addr);
|
|
|
|
void read_trace_pipe(void);
|
|
|
|
ssize_t get_uprobe_offset(const void *addr);
|
|
ssize_t get_rel_offset(uintptr_t addr);
|
|
|
|
int read_build_id(const char *path, char *build_id, size_t size);
|
|
|
|
#endif
|