3de7ae0b2a
Threads synthesized from /proc have comms with a start time of zero, and not marked as "exec". Currently, there can be 2 such comms. The first is created by processing a synthesized fork event and is set to the parent's comm string, and the second by processing a synthesized comm event set to the thread's current comm string. In the absence of an "exec" comm, thread__exec_comm() picks the last (oldest) comm, which, in the case above, is the parent's comm string. For a main thread, that is very probably wrong. Use the second-to-last in that case. This affects only db-export because it is the only user of thread__exec_comm(). Example: $ sudo perf record -a -o pt-a-sleep-1 -e intel_pt//u -- sleep 1 $ sudo chown ahunter pt-a-sleep-1 Before: $ perf script -i pt-a-sleep-1 --itrace=bep -s tools/perf/scripts/python/export-to-sqlite.py pt-a-sleep-1.db branches calls $ sqlite3 -header -column pt-a-sleep-1.db 'select * from comm_threads_view' comm_id command thread_id pid tid ---------- ---------- ---------- ---------- ---------- 1 swapper 1 0 0 2 rcu_sched 2 10 10 3 kthreadd 3 78 78 5 sudo 4 15180 15180 5 sudo 5 15180 15182 7 kworker/4: 6 10335 10335 8 kthreadd 7 55 55 10 systemd 8 865 865 10 systemd 9 865 875 13 perf 10 15181 15181 15 sleep 10 15181 15181 16 kworker/3: 11 14179 14179 17 kthreadd 12 29376 29376 19 systemd 13 746 746 21 systemd 14 401 401 23 systemd 15 879 879 23 systemd 16 879 945 25 kthreadd 17 556 556 27 kworker/u1 18 14136 14136 28 kworker/u1 19 15021 15021 29 kthreadd 20 509 509 31 systemd 21 836 836 31 systemd 22 836 967 33 systemd 23 1148 1148 33 systemd 24 1148 1163 35 kworker/2: 25 17988 17988 36 kworker/0: 26 13478 13478 After: $ perf script -i pt-a-sleep-1 --itrace=bep -s tools/perf/scripts/python/export-to-sqlite.py pt-a-sleep-1b.db branches calls $ sqlite3 -header -column pt-a-sleep-1b.db 'select * from comm_threads_view' comm_id command thread_id pid tid ---------- ---------- ---------- ---------- ---------- 1 swapper 1 0 0 2 rcu_sched 2 10 10 3 kswapd0 3 78 78 4 perf 4 15180 15180 4 perf 5 15180 15182 6 kworker/4: 6 10335 10335 7 kcompactd0 7 55 55 8 accounts-d 8 865 865 8 accounts-d 9 865 875 10 perf 10 15181 15181 12 sleep 10 15181 15181 13 kworker/3: 11 14179 14179 14 kworker/1: 12 29376 29376 15 haveged 13 746 746 16 systemd-jo 14 401 401 17 NetworkMan 15 879 879 17 NetworkMan 16 879 945 19 irq/131-iw 17 556 556 20 kworker/u1 18 14136 14136 21 kworker/u1 19 15021 15021 22 kworker/u1 20 509 509 23 thermald 21 836 836 23 thermald 22 836 967 25 unity-sett 23 1148 1148 25 unity-sett 24 1148 1163 27 kworker/2: 25 17988 17988 28 kworker/0: 26 13478 13478 Signed-off-by: Adrian Hunter <adrian.hunter@intel.com> Cc: Jiri Olsa <jolsa@redhat.com> Cc: stable@vger.kernel.org Fixes: 65de51f93ebf ("perf tools: Identify which comms are from exec") Link: http://lkml.kernel.org/r/20190808064823.14846-1-adrian.hunter@intel.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
458 lines
11 KiB
C
458 lines
11 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include "../perf.h"
|
|
#include <errno.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/zalloc.h>
|
|
#include "session.h"
|
|
#include "thread.h"
|
|
#include "thread-stack.h"
|
|
#include "debug.h"
|
|
#include "namespaces.h"
|
|
#include "comm.h"
|
|
#include "map.h"
|
|
#include "symbol.h"
|
|
#include "unwind.h"
|
|
#include "callchain.h"
|
|
|
|
#include <api/fs/fs.h>
|
|
|
|
int thread__init_map_groups(struct thread *thread, struct machine *machine)
|
|
{
|
|
pid_t pid = thread->pid_;
|
|
|
|
if (pid == thread->tid || pid == -1) {
|
|
thread->mg = map_groups__new(machine);
|
|
} else {
|
|
struct thread *leader = __machine__findnew_thread(machine, pid, pid);
|
|
if (leader) {
|
|
thread->mg = map_groups__get(leader->mg);
|
|
thread__put(leader);
|
|
}
|
|
}
|
|
|
|
return thread->mg ? 0 : -1;
|
|
}
|
|
|
|
struct thread *thread__new(pid_t pid, pid_t tid)
|
|
{
|
|
char *comm_str;
|
|
struct comm *comm;
|
|
struct thread *thread = zalloc(sizeof(*thread));
|
|
|
|
if (thread != NULL) {
|
|
thread->pid_ = pid;
|
|
thread->tid = tid;
|
|
thread->ppid = -1;
|
|
thread->cpu = -1;
|
|
INIT_LIST_HEAD(&thread->namespaces_list);
|
|
INIT_LIST_HEAD(&thread->comm_list);
|
|
init_rwsem(&thread->namespaces_lock);
|
|
init_rwsem(&thread->comm_lock);
|
|
|
|
comm_str = malloc(32);
|
|
if (!comm_str)
|
|
goto err_thread;
|
|
|
|
snprintf(comm_str, 32, ":%d", tid);
|
|
comm = comm__new(comm_str, 0, false);
|
|
free(comm_str);
|
|
if (!comm)
|
|
goto err_thread;
|
|
|
|
list_add(&comm->list, &thread->comm_list);
|
|
refcount_set(&thread->refcnt, 1);
|
|
RB_CLEAR_NODE(&thread->rb_node);
|
|
/* Thread holds first ref to nsdata. */
|
|
thread->nsinfo = nsinfo__new(pid);
|
|
srccode_state_init(&thread->srccode_state);
|
|
}
|
|
|
|
return thread;
|
|
|
|
err_thread:
|
|
free(thread);
|
|
return NULL;
|
|
}
|
|
|
|
void thread__delete(struct thread *thread)
|
|
{
|
|
struct namespaces *namespaces, *tmp_namespaces;
|
|
struct comm *comm, *tmp_comm;
|
|
|
|
BUG_ON(!RB_EMPTY_NODE(&thread->rb_node));
|
|
|
|
thread_stack__free(thread);
|
|
|
|
if (thread->mg) {
|
|
map_groups__put(thread->mg);
|
|
thread->mg = NULL;
|
|
}
|
|
down_write(&thread->namespaces_lock);
|
|
list_for_each_entry_safe(namespaces, tmp_namespaces,
|
|
&thread->namespaces_list, list) {
|
|
list_del_init(&namespaces->list);
|
|
namespaces__free(namespaces);
|
|
}
|
|
up_write(&thread->namespaces_lock);
|
|
|
|
down_write(&thread->comm_lock);
|
|
list_for_each_entry_safe(comm, tmp_comm, &thread->comm_list, list) {
|
|
list_del_init(&comm->list);
|
|
comm__free(comm);
|
|
}
|
|
up_write(&thread->comm_lock);
|
|
|
|
unwind__finish_access(thread);
|
|
nsinfo__zput(thread->nsinfo);
|
|
srccode_state_free(&thread->srccode_state);
|
|
|
|
exit_rwsem(&thread->namespaces_lock);
|
|
exit_rwsem(&thread->comm_lock);
|
|
free(thread);
|
|
}
|
|
|
|
struct thread *thread__get(struct thread *thread)
|
|
{
|
|
if (thread)
|
|
refcount_inc(&thread->refcnt);
|
|
return thread;
|
|
}
|
|
|
|
void thread__put(struct thread *thread)
|
|
{
|
|
if (thread && refcount_dec_and_test(&thread->refcnt)) {
|
|
/*
|
|
* Remove it from the dead threads list, as last reference is
|
|
* gone, if it is in a dead threads list.
|
|
*
|
|
* We may not be there anymore if say, the machine where it was
|
|
* stored was already deleted, so we already removed it from
|
|
* the dead threads and some other piece of code still keeps a
|
|
* reference.
|
|
*
|
|
* This is what 'perf sched' does and finally drops it in
|
|
* perf_sched__lat(), where it calls perf_sched__read_events(),
|
|
* that processes the events by creating a session and deleting
|
|
* it, which ends up destroying the list heads for the dead
|
|
* threads, but before it does that it removes all threads from
|
|
* it using list_del_init().
|
|
*
|
|
* So we need to check here if it is in a dead threads list and
|
|
* if so, remove it before finally deleting the thread, to avoid
|
|
* an use after free situation.
|
|
*/
|
|
if (!list_empty(&thread->node))
|
|
list_del_init(&thread->node);
|
|
thread__delete(thread);
|
|
}
|
|
}
|
|
|
|
static struct namespaces *__thread__namespaces(const struct thread *thread)
|
|
{
|
|
if (list_empty(&thread->namespaces_list))
|
|
return NULL;
|
|
|
|
return list_first_entry(&thread->namespaces_list, struct namespaces, list);
|
|
}
|
|
|
|
struct namespaces *thread__namespaces(struct thread *thread)
|
|
{
|
|
struct namespaces *ns;
|
|
|
|
down_read(&thread->namespaces_lock);
|
|
ns = __thread__namespaces(thread);
|
|
up_read(&thread->namespaces_lock);
|
|
|
|
return ns;
|
|
}
|
|
|
|
static int __thread__set_namespaces(struct thread *thread, u64 timestamp,
|
|
struct namespaces_event *event)
|
|
{
|
|
struct namespaces *new, *curr = __thread__namespaces(thread);
|
|
|
|
new = namespaces__new(event);
|
|
if (!new)
|
|
return -ENOMEM;
|
|
|
|
list_add(&new->list, &thread->namespaces_list);
|
|
|
|
if (timestamp && curr) {
|
|
/*
|
|
* setns syscall must have changed few or all the namespaces
|
|
* of this thread. Update end time for the namespaces
|
|
* previously used.
|
|
*/
|
|
curr = list_next_entry(new, list);
|
|
curr->end_time = timestamp;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int thread__set_namespaces(struct thread *thread, u64 timestamp,
|
|
struct namespaces_event *event)
|
|
{
|
|
int ret;
|
|
|
|
down_write(&thread->namespaces_lock);
|
|
ret = __thread__set_namespaces(thread, timestamp, event);
|
|
up_write(&thread->namespaces_lock);
|
|
return ret;
|
|
}
|
|
|
|
struct comm *thread__comm(const struct thread *thread)
|
|
{
|
|
if (list_empty(&thread->comm_list))
|
|
return NULL;
|
|
|
|
return list_first_entry(&thread->comm_list, struct comm, list);
|
|
}
|
|
|
|
struct comm *thread__exec_comm(const struct thread *thread)
|
|
{
|
|
struct comm *comm, *last = NULL, *second_last = NULL;
|
|
|
|
list_for_each_entry(comm, &thread->comm_list, list) {
|
|
if (comm->exec)
|
|
return comm;
|
|
second_last = last;
|
|
last = comm;
|
|
}
|
|
|
|
/*
|
|
* 'last' with no start time might be the parent's comm of a synthesized
|
|
* thread (created by processing a synthesized fork event). For a main
|
|
* thread, that is very probably wrong. Prefer a later comm to avoid
|
|
* that case.
|
|
*/
|
|
if (second_last && !last->start && thread->pid_ == thread->tid)
|
|
return second_last;
|
|
|
|
return last;
|
|
}
|
|
|
|
static int ____thread__set_comm(struct thread *thread, const char *str,
|
|
u64 timestamp, bool exec)
|
|
{
|
|
struct comm *new, *curr = thread__comm(thread);
|
|
|
|
/* Override the default :tid entry */
|
|
if (!thread->comm_set) {
|
|
int err = comm__override(curr, str, timestamp, exec);
|
|
if (err)
|
|
return err;
|
|
} else {
|
|
new = comm__new(str, timestamp, exec);
|
|
if (!new)
|
|
return -ENOMEM;
|
|
list_add(&new->list, &thread->comm_list);
|
|
|
|
if (exec)
|
|
unwind__flush_access(thread);
|
|
}
|
|
|
|
thread->comm_set = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int __thread__set_comm(struct thread *thread, const char *str, u64 timestamp,
|
|
bool exec)
|
|
{
|
|
int ret;
|
|
|
|
down_write(&thread->comm_lock);
|
|
ret = ____thread__set_comm(thread, str, timestamp, exec);
|
|
up_write(&thread->comm_lock);
|
|
return ret;
|
|
}
|
|
|
|
int thread__set_comm_from_proc(struct thread *thread)
|
|
{
|
|
char path[64];
|
|
char *comm = NULL;
|
|
size_t sz;
|
|
int err = -1;
|
|
|
|
if (!(snprintf(path, sizeof(path), "%d/task/%d/comm",
|
|
thread->pid_, thread->tid) >= (int)sizeof(path)) &&
|
|
procfs__read_str(path, &comm, &sz) == 0) {
|
|
comm[sz - 1] = '\0';
|
|
err = thread__set_comm(thread, comm, 0);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static const char *__thread__comm_str(const struct thread *thread)
|
|
{
|
|
const struct comm *comm = thread__comm(thread);
|
|
|
|
if (!comm)
|
|
return NULL;
|
|
|
|
return comm__str(comm);
|
|
}
|
|
|
|
const char *thread__comm_str(struct thread *thread)
|
|
{
|
|
const char *str;
|
|
|
|
down_read(&thread->comm_lock);
|
|
str = __thread__comm_str(thread);
|
|
up_read(&thread->comm_lock);
|
|
|
|
return str;
|
|
}
|
|
|
|
/* CHECKME: it should probably better return the max comm len from its comm list */
|
|
int thread__comm_len(struct thread *thread)
|
|
{
|
|
if (!thread->comm_len) {
|
|
const char *comm = thread__comm_str(thread);
|
|
if (!comm)
|
|
return 0;
|
|
thread->comm_len = strlen(comm);
|
|
}
|
|
|
|
return thread->comm_len;
|
|
}
|
|
|
|
size_t thread__fprintf(struct thread *thread, FILE *fp)
|
|
{
|
|
return fprintf(fp, "Thread %d %s\n", thread->tid, thread__comm_str(thread)) +
|
|
map_groups__fprintf(thread->mg, fp);
|
|
}
|
|
|
|
int thread__insert_map(struct thread *thread, struct map *map)
|
|
{
|
|
int ret;
|
|
|
|
ret = unwind__prepare_access(thread, map, NULL);
|
|
if (ret)
|
|
return ret;
|
|
|
|
map_groups__fixup_overlappings(thread->mg, map, stderr);
|
|
map_groups__insert(thread->mg, map);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __thread__prepare_access(struct thread *thread)
|
|
{
|
|
bool initialized = false;
|
|
int err = 0;
|
|
struct maps *maps = &thread->mg->maps;
|
|
struct map *map;
|
|
|
|
down_read(&maps->lock);
|
|
|
|
for (map = maps__first(maps); map; map = map__next(map)) {
|
|
err = unwind__prepare_access(thread, map, &initialized);
|
|
if (err || initialized)
|
|
break;
|
|
}
|
|
|
|
up_read(&maps->lock);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int thread__prepare_access(struct thread *thread)
|
|
{
|
|
int err = 0;
|
|
|
|
if (dwarf_callchain_users)
|
|
err = __thread__prepare_access(thread);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int thread__clone_map_groups(struct thread *thread,
|
|
struct thread *parent,
|
|
bool do_maps_clone)
|
|
{
|
|
/* This is new thread, we share map groups for process. */
|
|
if (thread->pid_ == parent->pid_)
|
|
return thread__prepare_access(thread);
|
|
|
|
if (thread->mg == parent->mg) {
|
|
pr_debug("broken map groups on thread %d/%d parent %d/%d\n",
|
|
thread->pid_, thread->tid, parent->pid_, parent->tid);
|
|
return 0;
|
|
}
|
|
/* But this one is new process, copy maps. */
|
|
return do_maps_clone ? map_groups__clone(thread, parent->mg) : 0;
|
|
}
|
|
|
|
int thread__fork(struct thread *thread, struct thread *parent, u64 timestamp, bool do_maps_clone)
|
|
{
|
|
if (parent->comm_set) {
|
|
const char *comm = thread__comm_str(parent);
|
|
int err;
|
|
if (!comm)
|
|
return -ENOMEM;
|
|
err = thread__set_comm(thread, comm, timestamp);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
thread->ppid = parent->tid;
|
|
return thread__clone_map_groups(thread, parent, do_maps_clone);
|
|
}
|
|
|
|
void thread__find_cpumode_addr_location(struct thread *thread, u64 addr,
|
|
struct addr_location *al)
|
|
{
|
|
size_t i;
|
|
const u8 cpumodes[] = {
|
|
PERF_RECORD_MISC_USER,
|
|
PERF_RECORD_MISC_KERNEL,
|
|
PERF_RECORD_MISC_GUEST_USER,
|
|
PERF_RECORD_MISC_GUEST_KERNEL
|
|
};
|
|
|
|
for (i = 0; i < ARRAY_SIZE(cpumodes); i++) {
|
|
thread__find_symbol(thread, cpumodes[i], addr, al);
|
|
if (al->map)
|
|
break;
|
|
}
|
|
}
|
|
|
|
struct thread *thread__main_thread(struct machine *machine, struct thread *thread)
|
|
{
|
|
if (thread->pid_ == thread->tid)
|
|
return thread__get(thread);
|
|
|
|
if (thread->pid_ == -1)
|
|
return NULL;
|
|
|
|
return machine__find_thread(machine, thread->pid_, thread->pid_);
|
|
}
|
|
|
|
int thread__memcpy(struct thread *thread, struct machine *machine,
|
|
void *buf, u64 ip, int len, bool *is64bit)
|
|
{
|
|
u8 cpumode = PERF_RECORD_MISC_USER;
|
|
struct addr_location al;
|
|
long offset;
|
|
|
|
if (machine__kernel_ip(machine, ip))
|
|
cpumode = PERF_RECORD_MISC_KERNEL;
|
|
|
|
if (!thread__find_map(thread, cpumode, ip, &al) || !al.map->dso ||
|
|
al.map->dso->data.status == DSO_DATA_STATUS_ERROR ||
|
|
map__load(al.map) < 0)
|
|
return -1;
|
|
|
|
offset = al.map->map_ip(al.map, ip);
|
|
if (is64bit)
|
|
*is64bit = al.map->dso->is_64_bit;
|
|
|
|
return dso__data_read_offset(al.map->dso, machine, offset, buf, len);
|
|
}
|