2008-05-12 23:20:42 +04:00
/*
* ring buffer based function tracer
*
* Copyright ( C ) 2007 - 2008 Steven Rostedt < srostedt @ redhat . com >
* Copyright ( C ) 2008 Ingo Molnar < mingo @ redhat . com >
*
* Originally taken from the RT patch by :
* Arnaldo Carvalho de Melo < acme @ redhat . com >
*
* Based on code from the latency_tracer , that is :
* Copyright ( C ) 2004 - 2006 Ingo Molnar
* Copyright ( C ) 2004 William Lee Irwin III
*/
# include <linux/utsrelease.h>
# include <linux/kallsyms.h>
# include <linux/seq_file.h>
2008-07-31 06:36:46 +04:00
# include <linux/notifier.h>
2008-05-12 23:20:42 +04:00
# include <linux/debugfs.h>
2008-05-12 23:20:43 +04:00
# include <linux/pagemap.h>
2008-05-12 23:20:42 +04:00
# include <linux/hardirq.h>
# include <linux/linkage.h>
# include <linux/uaccess.h>
# include <linux/ftrace.h>
# include <linux/module.h>
# include <linux/percpu.h>
2008-07-31 06:36:46 +04:00
# include <linux/kdebug.h>
2008-05-12 23:20:42 +04:00
# include <linux/ctype.h>
# include <linux/init.h>
2008-05-12 23:20:49 +04:00
# include <linux/poll.h>
2008-05-12 23:20:42 +04:00
# include <linux/gfp.h>
# include <linux/fs.h>
2008-05-27 22:33:18 +04:00
# include <linux/kprobes.h>
2008-05-12 23:21:04 +04:00
# include <linux/writeback.h>
2009-02-09 09:15:56 +03:00
# include <linux/splice.h>
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:51 +04:00
# include <linux/stacktrace.h>
2008-09-30 07:02:41 +04:00
# include <linux/ring_buffer.h>
2008-10-28 11:43:26 +03:00
# include <linux/irqflags.h>
2008-05-12 23:20:51 +04:00
2008-05-12 23:20:42 +04:00
# include "trace.h"
2008-12-24 07:24:12 +03:00
# include "trace_output.h"
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
# define TRACE_BUFFER_FLAGS (RB_FL_OVERWRITE)
2009-01-16 07:40:11 +03:00
unsigned long __read_mostly tracing_max_latency ;
2008-05-12 23:20:42 +04:00
unsigned long __read_mostly tracing_thresh ;
2008-12-06 05:41:33 +03:00
/*
* We need to change this state when a selftest is running .
2008-12-05 01:47:35 +03:00
* A selftest will lurk into the ring - buffer to count the
* entries inserted during the selftest although some concurrent
* insertions into the ring - buffer such as ftrace_printk could occurred
* at the same time , giving false positive or negative results .
*/
2008-12-06 05:41:33 +03:00
static bool __read_mostly tracing_selftest_running ;
2008-12-05 01:47:35 +03:00
2009-02-03 05:38:32 +03:00
/*
* If a tracer is running , we do not want to run SELFTEST .
*/
static bool __read_mostly tracing_selftest_disabled ;
2008-11-17 21:23:42 +03:00
/* For tracers that don't implement custom flags */
static struct tracer_opt dummy_tracer_opt [ ] = {
{ }
} ;
static struct tracer_flags dummy_tracer_flags = {
. val = 0 ,
. opts = dummy_tracer_opt
} ;
static int dummy_set_flag ( u32 old_flags , u32 bit , int set )
{
return 0 ;
}
2008-11-06 00:05:44 +03:00
/*
* Kill all tracing for good ( never come back ) .
* It is initialized to 1 but will turn to zero if the initialization
* of the tracer is successful . But that is the only place that sets
* this back to zero .
*/
int tracing_disabled = 1 ;
2008-10-01 08:29:53 +04:00
static DEFINE_PER_CPU ( local_t , ftrace_cpu_disabled ) ;
static inline void ftrace_disable_cpu ( void )
{
preempt_disable ( ) ;
local_inc ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ;
}
static inline void ftrace_enable_cpu ( void )
{
local_dec ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ;
preempt_enable ( ) ;
}
2009-01-01 02:42:22 +03:00
static cpumask_var_t __read_mostly tracing_buffer_mask ;
2008-05-12 23:21:00 +04:00
# define for_each_tracing_cpu(cpu) \
2009-01-01 02:42:22 +03:00
for_each_cpu ( cpu , tracing_buffer_mask )
2008-05-12 23:21:00 +04:00
2008-10-24 03:26:08 +04:00
/*
* ftrace_dump_on_oops - variable to dump ftrace buffer on oops
*
* If there is an oops ( or kernel panic ) and the ftrace_dump_on_oops
* is set , then ftrace_dump is called . This will output the contents
* of the ftrace buffers to the console . This is very useful for
* capturing traces that lead to crashes and outputing it to a
* serial console .
*
* It is default off , but you can enable it with either specifying
* " ftrace_dump_on_oops " in the kernel command line , or setting
* / proc / sys / kernel / ftrace_dump_on_oops to true .
*/
int ftrace_dump_on_oops ;
2009-02-03 05:38:32 +03:00
static int tracing_set_tracer ( const char * buf ) ;
# define BOOTUP_TRACER_SIZE 100
static char bootup_tracer_buf [ BOOTUP_TRACER_SIZE ] __initdata ;
static char * default_bootup_tracer ;
2008-11-01 21:57:37 +03:00
static int __init set_ftrace ( char * str )
{
2009-02-03 05:38:32 +03:00
strncpy ( bootup_tracer_buf , str , BOOTUP_TRACER_SIZE ) ;
default_bootup_tracer = bootup_tracer_buf ;
2008-11-01 21:57:37 +03:00
return 1 ;
}
2009-02-03 05:38:32 +03:00
__setup ( " ftrace= " , set_ftrace ) ;
2008-11-01 21:57:37 +03:00
2008-10-24 03:26:08 +04:00
static int __init set_ftrace_dump_on_oops ( char * str )
{
ftrace_dump_on_oops = 1 ;
return 1 ;
}
__setup ( " ftrace_dump_on_oops " , set_ftrace_dump_on_oops ) ;
2008-05-12 23:20:44 +04:00
2008-05-23 23:37:28 +04:00
long
2008-05-12 23:20:42 +04:00
ns2usecs ( cycle_t nsec )
{
nsec + = 500 ;
do_div ( nsec , 1000 ) ;
return nsec ;
}
2008-05-12 23:20:51 +04:00
cycle_t ftrace_now ( int cpu )
2008-05-12 23:20:46 +04:00
{
2008-09-30 07:02:41 +04:00
u64 ts = ring_buffer_time_stamp ( cpu ) ;
ring_buffer_normalize_time_stamp ( cpu , & ts ) ;
return ts ;
2008-05-12 23:20:46 +04:00
}
2008-05-12 23:21:00 +04:00
/*
* The global_trace is the descriptor that holds the tracing
* buffers for the live tracing . For each CPU , it contains
* a link list of pages that will store trace entries . The
* page descriptor of the pages in the memory is used to hold
* the link list by linking the lru item in the page descriptor
* to each of the pages in the buffer per CPU .
*
* For each active CPU there is a data field that holds the
* pages for the buffer for that CPU . Each CPU has the same number
* of pages allocated for its buffer .
*/
2008-05-12 23:20:42 +04:00
static struct trace_array global_trace ;
static DEFINE_PER_CPU ( struct trace_array_cpu , global_trace_cpu ) ;
2008-05-12 23:21:00 +04:00
/*
* The max_tr is used to snapshot the global_trace when a maximum
* latency is reached . Some tracers will use this to store a maximum
* trace while it continues examining live traces .
*
* The buffers for the max_tr are set up the same as the global_trace .
* When a snapshot is taken , the link list of the max_tr is swapped
* with the link list of the global_trace and the buffers are reset for
* the global_trace so the tracing can continue .
*/
2008-05-12 23:20:42 +04:00
static struct trace_array max_tr ;
static DEFINE_PER_CPU ( struct trace_array_cpu , max_data ) ;
2008-05-12 23:21:00 +04:00
/* tracer_enabled is used to toggle activation of a tracer */
2008-05-12 23:20:48 +04:00
static int tracer_enabled = 1 ;
2008-05-12 23:21:00 +04:00
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
/**
* tracing_is_enabled - return tracer_enabled status
*
* This function is used by other tracers to know the status
* of the tracer_enabled flag . Tracers may use this function
* to know if it should enable their features when starting
* up . See irqsoff tracer for an example ( start_irqsoff_tracer ) .
*/
int tracing_is_enabled ( void )
{
return tracer_enabled ;
}
2008-05-12 23:21:00 +04:00
/*
2008-09-30 07:02:41 +04:00
* trace_buf_size is the size in bytes that is allocated
* for a buffer . Note , the number of bytes is always rounded
* to page size .
2008-07-31 06:36:46 +04:00
*
* This number is purposely set to a low number of 16384.
* If the dump on oops happens , it will be much appreciated
* to not have to wait for all that output . Anyway this can be
* boot time and run time configurable .
2008-05-12 23:21:00 +04:00
*/
2008-09-30 07:02:41 +04:00
# define TRACE_BUF_SIZE_DEFAULT 1441792UL /* 16384 * 88 (sizeof(entry)) */
2008-07-31 06:36:46 +04:00
2008-09-30 07:02:41 +04:00
static unsigned long trace_buf_size = TRACE_BUF_SIZE_DEFAULT ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:21:00 +04:00
/* trace_types holds a link list of available tracers. */
2008-05-12 23:20:42 +04:00
static struct tracer * trace_types __read_mostly ;
2008-05-12 23:21:00 +04:00
/* current_trace points to the tracer that is currently active */
2008-05-12 23:20:42 +04:00
static struct tracer * current_trace __read_mostly ;
2008-05-12 23:21:00 +04:00
/*
* max_tracer_type_len is used to simplify the allocating of
* buffers to read userspace tracer names . We keep track of
* the longest tracer name registered .
*/
2008-05-12 23:20:42 +04:00
static int max_tracer_type_len ;
2008-05-12 23:21:00 +04:00
/*
* trace_types_lock is used to protect the trace_types list .
* This lock is also used to keep user access serialized .
* Accesses from userspace will grab this lock while userspace
* activities happen inside the kernel .
*/
2008-05-12 23:20:42 +04:00
static DEFINE_MUTEX ( trace_types_lock ) ;
2008-05-12 23:21:00 +04:00
/* trace_wait is a waitqueue for tasks blocked on trace_poll */
2008-05-12 23:20:52 +04:00
static DECLARE_WAIT_QUEUE_HEAD ( trace_wait ) ;
2008-11-13 01:52:37 +03:00
/* trace_flags holds trace_options default values */
2008-11-13 01:52:38 +03:00
unsigned long trace_flags = TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK |
2009-02-03 01:29:21 +03:00
TRACE_ITER_ANNOTATE | TRACE_ITER_CONTEXT_INFO ;
2008-05-12 23:20:52 +04:00
2008-05-12 23:21:00 +04:00
/**
* trace_wake_up - wake up tasks waiting for trace input
*
* Simply wakes up any task that is blocked on the trace_wait
* queue . These is used with trace_poll for tasks polling the trace .
*/
2008-05-12 23:20:52 +04:00
void trace_wake_up ( void )
{
2008-05-12 23:20:52 +04:00
/*
* The runqueue_is_locked ( ) can fail , but this is the best we
* have for now :
*/
if ( ! ( trace_flags & TRACE_ITER_BLOCK ) & & ! runqueue_is_locked ( ) )
2008-05-12 23:20:52 +04:00
wake_up ( & trace_wait ) ;
}
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
static int __init set_buf_size ( char * str )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
unsigned long buf_size ;
2008-05-12 23:21:00 +04:00
int ret ;
2008-05-12 23:20:42 +04:00
if ( ! str )
return 0 ;
2008-09-30 07:02:41 +04:00
ret = strict_strtoul ( str , 0 , & buf_size ) ;
2008-05-12 23:21:00 +04:00
/* nr_entries can not be zero */
2008-09-30 07:02:41 +04:00
if ( ret < 0 | | buf_size = = 0 )
2008-05-12 23:21:00 +04:00
return 0 ;
2008-09-30 07:02:41 +04:00
trace_buf_size = buf_size ;
2008-05-12 23:20:42 +04:00
return 1 ;
}
2008-09-30 07:02:41 +04:00
__setup ( " trace_buf_size= " , set_buf_size ) ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:44 +04:00
unsigned long nsecs_to_usecs ( unsigned long nsecs )
{
return nsecs / 1000 ;
}
2008-05-12 23:21:00 +04:00
/* These must match the bit postions in trace_iterator_flags */
2008-05-12 23:20:42 +04:00
static const char * trace_options [ ] = {
" print-parent " ,
" sym-offset " ,
" sym-addr " ,
" verbose " ,
2008-05-12 23:20:47 +04:00
" raw " ,
2008-05-12 23:20:49 +04:00
" hex " ,
2008-05-12 23:20:47 +04:00
" bin " ,
2008-05-12 23:20:49 +04:00
" block " ,
2008-05-12 23:20:51 +04:00
" stacktrace " ,
2008-05-12 23:20:52 +04:00
" sched-tree " ,
2008-09-04 12:24:14 +04:00
" ftrace_printk " ,
ftrace: function tracer with irqs disabled
Impact: disable interrupts during trace entry creation (as opposed to preempt)
To help with performance, I set the ftracer to not disable interrupts,
and only to disable preemption. If an interrupt occurred, it would not
be traced, because the function tracer protects itself from recursion.
This may be faster, but the trace output might miss some traces.
This patch makes the fuction trace disable interrupts, but it also
adds a runtime feature to disable preemption instead. It does this by
having two different tracer functions. When the function tracer is
enabled, it will check to see which version is requested (irqs disabled
or preemption disabled). Then it will use the corresponding function
as the tracer.
Irq disabling is the default behaviour, but if the user wants better
performance, with the chance of missing traces, then they can choose
the preempt disabled version.
Running hackbench 3 times with the irqs disabled and 3 times with
the preempt disabled function tracer yielded:
tracing type times entries recorded
------------ -------- ----------------
irq disabled 43.393 166433066
43.282 166172618
43.298 166256704
preempt disabled 38.969 159871710
38.943 159972935
39.325 161056510
Average:
irqs disabled: 43.324 166287462
preempt disabled: 39.079 160300385
preempt is 10.8 percent faster than irqs disabled.
I wrote a patch to count function trace recursion and reran hackbench.
With irq disabled: 1,150 times the function tracer did not trace due to
recursion.
with preempt disabled: 5,117,718 times.
The thousand times with irq disabled could be due to NMIs, or simply a case
where it called a function that was not protected by notrace.
But we also see that a large amount of the trace is lost with the
preempt version.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 07:15:57 +03:00
" ftrace_preempt " ,
2008-11-12 23:24:24 +03:00
" branch " ,
2008-11-13 01:52:38 +03:00
" annotate " ,
2008-11-22 14:28:47 +03:00
" userstacktrace " ,
2008-11-22 14:28:48 +03:00
" sym-userobj " ,
2008-12-13 22:18:13 +03:00
" printk-msg-only " ,
2009-02-03 01:29:21 +03:00
" context-info " ,
2008-05-12 23:20:42 +04:00
NULL
} ;
2008-05-12 23:21:00 +04:00
/*
* ftrace_max_lock is used to protect the swapping of buffers
* when taking a max snapshot . The buffers themselves are
* protected by per_cpu spinlocks . But the action of the swap
* needs its own lock .
*
* This is defined as a raw_spinlock_t in order to help
* with performance when lockdep debugging is enabled .
*/
2008-05-12 23:20:55 +04:00
static raw_spinlock_t ftrace_max_lock =
( raw_spinlock_t ) __RAW_SPIN_LOCK_UNLOCKED ;
2008-05-12 23:20:42 +04:00
/*
* Copy the new maximum trace into the separate maximum - trace
* structure . ( this way the maximum trace is permanently saved ,
* for later retrieval via / debugfs / tracing / latency_trace )
*/
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:42 +04:00
__update_max_tr ( struct trace_array * tr , struct task_struct * tsk , int cpu )
{
struct trace_array_cpu * data = tr - > data [ cpu ] ;
max_tr . cpu = cpu ;
max_tr . time_start = data - > preempt_timestamp ;
data = max_tr . data [ cpu ] ;
data - > saved_latency = tracing_max_latency ;
memcpy ( data - > comm , tsk - > comm , TASK_COMM_LEN ) ;
data - > pid = tsk - > pid ;
2008-11-14 02:39:16 +03:00
data - > uid = task_uid ( tsk ) ;
2008-05-12 23:20:42 +04:00
data - > nice = tsk - > static_prio - 20 - MAX_RT_PRIO ;
data - > policy = tsk - > policy ;
data - > rt_priority = tsk - > rt_priority ;
/* record this tasks comm */
tracing_record_cmdline ( current ) ;
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:46 +04:00
trace_seq_reset ( struct trace_seq * s )
{
s - > len = 0 ;
2008-05-12 23:21:02 +04:00
s - > readpos = 0 ;
}
ssize_t trace_seq_to_user ( struct trace_seq * s , char __user * ubuf , size_t cnt )
{
int len ;
int ret ;
if ( s - > len < = s - > readpos )
return - EBUSY ;
len = s - > len - s - > readpos ;
if ( cnt > len )
cnt = len ;
ret = copy_to_user ( ubuf , s - > buffer + s - > readpos , cnt ) ;
if ( ret )
return - EFAULT ;
s - > readpos + = len ;
return cnt ;
2008-05-12 23:20:46 +04:00
}
2009-02-09 09:15:56 +03:00
ssize_t trace_seq_to_buffer ( struct trace_seq * s , void * buf , size_t cnt )
{
int len ;
void * ret ;
if ( s - > len < = s - > readpos )
return - EBUSY ;
len = s - > len - s - > readpos ;
if ( cnt > len )
cnt = len ;
ret = memcpy ( buf , s - > buffer + s - > readpos , cnt ) ;
if ( ! ret )
return - EFAULT ;
s - > readpos + = len ;
return cnt ;
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:46 +04:00
trace_print_seq ( struct seq_file * m , struct trace_seq * s )
{
int len = s - > len > = PAGE_SIZE ? PAGE_SIZE - 1 : s - > len ;
s - > buffer [ len ] = 0 ;
seq_puts ( m , s - > buffer ) ;
trace_seq_reset ( s ) ;
}
2008-05-12 23:21:00 +04:00
/**
* update_max_tr - snapshot all trace buffers from global_trace to max_tr
* @ tr : tracer
* @ tsk : the task with the latency
* @ cpu : The cpu that initiated the trace .
*
* Flip the buffers between the @ tr and the max_tr and record information
* about which task was the cause of this latency .
*/
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:42 +04:00
update_max_tr ( struct trace_array * tr , struct task_struct * tsk , int cpu )
{
2008-09-30 07:02:41 +04:00
struct ring_buffer * buf = tr - > buffer ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:43 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_lock ( & ftrace_max_lock ) ;
2008-09-30 07:02:41 +04:00
tr - > buffer = max_tr . buffer ;
max_tr . buffer = buf ;
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_reset ( tr - > buffer ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
__update_max_tr ( tr , tsk , cpu ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_unlock ( & ftrace_max_lock ) ;
2008-05-12 23:20:42 +04:00
}
/**
* update_max_tr_single - only copy one trace over , and reset the rest
* @ tr - tracer
* @ tsk - task with the latency
* @ cpu - the cpu of the buffer to copy .
2008-05-12 23:21:00 +04:00
*
* Flip the trace of a single CPU buffer between the @ tr and the max_tr .
2008-05-12 23:20:42 +04:00
*/
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:42 +04:00
update_max_tr_single ( struct trace_array * tr , struct task_struct * tsk , int cpu )
{
2008-09-30 07:02:41 +04:00
int ret ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:43 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_lock ( & ftrace_max_lock ) ;
2008-05-12 23:20:42 +04:00
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_reset ( max_tr . buffer ) ;
ret = ring_buffer_swap_cpu ( max_tr . buffer , tr - > buffer , cpu ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2009-01-21 23:24:56 +03:00
WARN_ON_ONCE ( ret & & ret ! = - EAGAIN ) ;
2008-05-12 23:20:42 +04:00
__update_max_tr ( tr , tsk , cpu ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_unlock ( & ftrace_max_lock ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:21:00 +04:00
/**
* register_tracer - register a tracer with the ftrace system .
* @ type - the plugin for the tracer
*
* Register a new plugin tracer .
*/
2008-05-12 23:20:42 +04:00
int register_tracer ( struct tracer * type )
{
struct tracer * t ;
int len ;
int ret = 0 ;
if ( ! type - > name ) {
pr_info ( " Tracer must have a name \n " ) ;
return - 1 ;
}
2008-11-19 12:00:15 +03:00
/*
* When this gets called we hold the BKL which means that
* preemption is disabled . Various trace selftests however
* need to disable and enable preemption for successful tests .
* So we drop the BKL here and grab it after the tests again .
*/
unlock_kernel ( ) ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-11-19 12:00:15 +03:00
2008-12-06 05:41:33 +03:00
tracing_selftest_running = true ;
2008-05-12 23:20:42 +04:00
for ( t = trace_types ; t ; t = t - > next ) {
if ( strcmp ( type - > name , t - > name ) = = 0 ) {
/* already found */
pr_info ( " Trace %s already registered \n " ,
type - > name ) ;
ret = - 1 ;
goto out ;
}
}
2008-11-17 21:23:42 +03:00
if ( ! type - > set_flag )
type - > set_flag = & dummy_set_flag ;
if ( ! type - > flags )
type - > flags = & dummy_tracer_flags ;
else
if ( ! type - > flags - > opts )
type - > flags - > opts = dummy_tracer_opt ;
2008-05-12 23:20:44 +04:00
# ifdef CONFIG_FTRACE_STARTUP_TEST
2009-02-03 05:38:32 +03:00
if ( type - > selftest & & ! tracing_selftest_disabled ) {
2008-05-12 23:20:44 +04:00
struct tracer * saved_tracer = current_trace ;
struct trace_array * tr = & global_trace ;
int i ;
2008-12-05 01:47:35 +03:00
2008-05-12 23:20:44 +04:00
/*
* Run a selftest on this tracer .
* Here we reset the trace buffer , and set the current
* tracer to be this tracer . The tracer can then run some
* internal tracing to verify that everything is in order .
* If we fail , we do not register this tracer .
*/
2008-11-19 12:00:15 +03:00
for_each_tracing_cpu ( i )
2008-09-30 07:02:41 +04:00
tracing_reset ( tr , i ) ;
2008-11-19 12:00:15 +03:00
2008-05-12 23:20:44 +04:00
current_trace = type ;
/* the test is responsible for initializing and enabling */
pr_info ( " Testing tracer %s: " , type - > name ) ;
ret = type - > selftest ( type , tr ) ;
/* the test is responsible for resetting too */
current_trace = saved_tracer ;
if ( ret ) {
printk ( KERN_CONT " FAILED! \n " ) ;
goto out ;
}
2008-05-12 23:20:45 +04:00
/* Only reset on passing, to avoid touching corrupted buffers */
2008-11-19 12:00:15 +03:00
for_each_tracing_cpu ( i )
2008-09-30 07:02:41 +04:00
tracing_reset ( tr , i ) ;
2008-11-19 12:00:15 +03:00
2008-05-12 23:20:44 +04:00
printk ( KERN_CONT " PASSED \n " ) ;
}
# endif
2008-05-12 23:20:42 +04:00
type - > next = trace_types ;
trace_types = type ;
len = strlen ( type - > name ) ;
if ( len > max_tracer_type_len )
max_tracer_type_len = len ;
2008-05-12 23:20:44 +04:00
2008-05-12 23:20:42 +04:00
out :
2008-12-06 05:41:33 +03:00
tracing_selftest_running = false ;
2008-05-12 23:20:42 +04:00
mutex_unlock ( & trace_types_lock ) ;
2009-02-05 09:13:38 +03:00
if ( ret | | ! default_bootup_tracer )
goto out_unlock ;
if ( strncmp ( default_bootup_tracer , type - > name , BOOTUP_TRACER_SIZE ) )
goto out_unlock ;
printk ( KERN_INFO " Starting tracer '%s' \n " , type - > name ) ;
/* Do we want this tracer to start on bootup? */
tracing_set_tracer ( type - > name ) ;
default_bootup_tracer = NULL ;
/* disable other selftests, since this will break it. */
tracing_selftest_disabled = 1 ;
2009-02-03 05:38:32 +03:00
# ifdef CONFIG_FTRACE_STARTUP_TEST
2009-02-05 09:13:38 +03:00
printk ( KERN_INFO " Disabling FTRACE selftests due to running tracer '%s' \n " ,
type - > name ) ;
2009-02-03 05:38:32 +03:00
# endif
2009-02-05 09:13:38 +03:00
out_unlock :
2009-02-03 05:38:32 +03:00
lock_kernel ( ) ;
2008-05-12 23:20:42 +04:00
return ret ;
}
void unregister_tracer ( struct tracer * type )
{
struct tracer * * t ;
int len ;
mutex_lock ( & trace_types_lock ) ;
for ( t = & trace_types ; * t ; t = & ( * t ) - > next ) {
if ( * t = = type )
goto found ;
}
pr_info ( " Trace %s not registered \n " , type - > name ) ;
goto out ;
found :
* t = ( * t ) - > next ;
2009-02-07 23:52:59 +03:00
if ( type = = current_trace & & tracer_enabled ) {
tracer_enabled = 0 ;
tracing_stop ( ) ;
if ( current_trace - > stop )
current_trace - > stop ( & global_trace ) ;
current_trace = & nop_trace ;
}
2008-05-12 23:20:42 +04:00
if ( strlen ( type - > name ) ! = max_tracer_type_len )
goto out ;
max_tracer_type_len = 0 ;
for ( t = & trace_types ; * t ; t = & ( * t ) - > next ) {
len = strlen ( ( * t ) - > name ) ;
if ( len > max_tracer_type_len )
max_tracer_type_len = len ;
}
out :
mutex_unlock ( & trace_types_lock ) ;
}
2008-09-30 07:02:41 +04:00
void tracing_reset ( struct trace_array * tr , int cpu )
2008-05-12 23:20:42 +04:00
{
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_reset_cpu ( tr - > buffer , cpu ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
}
2008-12-19 13:08:39 +03:00
void tracing_reset_online_cpus ( struct trace_array * tr )
{
int cpu ;
tr - > time_start = ftrace_now ( tr - > cpu ) ;
for_each_online_cpu ( cpu )
tracing_reset ( tr , cpu ) ;
}
2008-05-12 23:20:42 +04:00
# define SAVED_CMDLINES 128
static unsigned map_pid_to_cmdline [ PID_MAX_DEFAULT + 1 ] ;
static unsigned map_cmdline_to_pid [ SAVED_CMDLINES ] ;
static char saved_cmdlines [ SAVED_CMDLINES ] [ TASK_COMM_LEN ] ;
static int cmdline_idx ;
static DEFINE_SPINLOCK ( trace_cmdline_lock ) ;
2008-05-12 23:21:00 +04:00
/* temporary disable recording */
atomic_t trace_record_cmdline_disabled __read_mostly ;
2008-05-12 23:20:42 +04:00
static void trace_init_cmdlines ( void )
{
memset ( & map_pid_to_cmdline , - 1 , sizeof ( map_pid_to_cmdline ) ) ;
memset ( & map_cmdline_to_pid , - 1 , sizeof ( map_cmdline_to_pid ) ) ;
cmdline_idx = 0 ;
}
2008-11-06 00:05:44 +03:00
static int trace_stop_count ;
static DEFINE_SPINLOCK ( tracing_start_lock ) ;
2008-11-21 20:59:38 +03:00
/**
* ftrace_off_permanent - disable all ftrace code permanently
*
* This should only be called when a serious anomally has
* been detected . This will turn off the function tracing ,
* ring buffers , and other tracing utilites . It takes no
* locks and can be called from any context .
*/
void ftrace_off_permanent ( void )
{
tracing_disabled = 1 ;
ftrace_stop ( ) ;
tracing_off_permanent ( ) ;
}
2008-11-06 00:05:44 +03:00
/**
* tracing_start - quick start of the tracer
*
* If tracing is enabled but was stopped by tracing_stop ,
* this will start the tracer back up .
*/
void tracing_start ( void )
{
struct ring_buffer * buffer ;
unsigned long flags ;
if ( tracing_disabled )
return ;
spin_lock_irqsave ( & tracing_start_lock , flags ) ;
2009-01-22 22:26:15 +03:00
if ( - - trace_stop_count ) {
if ( trace_stop_count < 0 ) {
/* Someone screwed up their debugging */
WARN_ON_ONCE ( 1 ) ;
trace_stop_count = 0 ;
}
2008-11-06 00:05:44 +03:00
goto out ;
}
buffer = global_trace . buffer ;
if ( buffer )
ring_buffer_record_enable ( buffer ) ;
buffer = max_tr . buffer ;
if ( buffer )
ring_buffer_record_enable ( buffer ) ;
ftrace_start ( ) ;
out :
spin_unlock_irqrestore ( & tracing_start_lock , flags ) ;
}
/**
* tracing_stop - quick stop of the tracer
*
* Light weight way to stop tracing . Use in conjunction with
* tracing_start .
*/
void tracing_stop ( void )
{
struct ring_buffer * buffer ;
unsigned long flags ;
ftrace_stop ( ) ;
spin_lock_irqsave ( & tracing_start_lock , flags ) ;
if ( trace_stop_count + + )
goto out ;
buffer = global_trace . buffer ;
if ( buffer )
ring_buffer_record_disable ( buffer ) ;
buffer = max_tr . buffer ;
if ( buffer )
ring_buffer_record_disable ( buffer ) ;
out :
spin_unlock_irqrestore ( & tracing_start_lock , flags ) ;
}
2008-05-12 23:20:51 +04:00
void trace_stop_cmdline_recording ( void ) ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:51 +04:00
static void trace_save_cmdline ( struct task_struct * tsk )
2008-05-12 23:20:42 +04:00
{
unsigned map ;
unsigned idx ;
if ( ! tsk - > pid | | unlikely ( tsk - > pid > PID_MAX_DEFAULT ) )
return ;
/*
* It ' s not the end of the world if we don ' t get
* the lock , but we also don ' t want to spin
* nor do we want to disable interrupts ,
* so if we miss here , then better luck next time .
*/
if ( ! spin_trylock ( & trace_cmdline_lock ) )
return ;
idx = map_pid_to_cmdline [ tsk - > pid ] ;
if ( idx > = SAVED_CMDLINES ) {
idx = ( cmdline_idx + 1 ) % SAVED_CMDLINES ;
map = map_cmdline_to_pid [ idx ] ;
if ( map < = PID_MAX_DEFAULT )
map_pid_to_cmdline [ map ] = ( unsigned ) - 1 ;
map_pid_to_cmdline [ tsk - > pid ] = idx ;
cmdline_idx = idx ;
}
memcpy ( & saved_cmdlines [ idx ] , tsk - > comm , TASK_COMM_LEN ) ;
spin_unlock ( & trace_cmdline_lock ) ;
}
2008-11-26 08:16:26 +03:00
char * trace_find_cmdline ( int pid )
2008-05-12 23:20:42 +04:00
{
char * cmdline = " <...> " ;
unsigned map ;
if ( ! pid )
return " <idle> " ;
if ( pid > PID_MAX_DEFAULT )
goto out ;
map = map_pid_to_cmdline [ pid ] ;
if ( map > = SAVED_CMDLINES )
goto out ;
cmdline = saved_cmdlines [ map ] ;
out :
return cmdline ;
}
2008-05-12 23:20:51 +04:00
void tracing_record_cmdline ( struct task_struct * tsk )
2008-05-12 23:20:42 +04:00
{
if ( atomic_read ( & trace_record_cmdline_disabled ) )
return ;
trace_save_cmdline ( tsk ) ;
}
2008-09-16 22:56:41 +04:00
void
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( struct trace_entry * entry , unsigned long flags ,
int pc )
2008-05-12 23:20:42 +04:00
{
struct task_struct * tsk = current ;
2008-09-30 07:02:42 +04:00
entry - > preempt_count = pc & 0xff ;
entry - > pid = ( tsk ) ? tsk - > pid : 0 ;
2008-11-22 14:28:48 +03:00
entry - > tgid = ( tsk ) ? tsk - > tgid : 0 ;
2008-09-30 07:02:42 +04:00
entry - > flags =
2008-10-24 17:42:59 +04:00
# ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT
2008-08-01 20:26:40 +04:00
( irqs_disabled_flags ( flags ) ? TRACE_FLAG_IRQS_OFF : 0 ) |
2008-10-24 17:42:59 +04:00
# else
TRACE_FLAG_IRQS_NOSUPPORT |
# endif
2008-05-12 23:20:42 +04:00
( ( pc & HARDIRQ_MASK ) ? TRACE_FLAG_HARDIRQ : 0 ) |
( ( pc & SOFTIRQ_MASK ) ? TRACE_FLAG_SOFTIRQ : 0 ) |
( need_resched ( ) ? TRACE_FLAG_NEED_RESCHED : 0 ) ;
}
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
struct ring_buffer_event * trace_buffer_lock_reserve ( struct trace_array * tr ,
unsigned char type ,
unsigned long len ,
unsigned long flags , int pc )
{
struct ring_buffer_event * event ;
event = ring_buffer_lock_reserve ( tr - > buffer , len ) ;
if ( event ! = NULL ) {
struct trace_entry * ent = ring_buffer_event_data ( event ) ;
tracing_generic_entry_update ( ent , flags , pc ) ;
ent - > type = type ;
}
return event ;
}
static void ftrace_trace_stack ( struct trace_array * tr ,
unsigned long flags , int skip , int pc ) ;
static void ftrace_trace_userstack ( struct trace_array * tr ,
unsigned long flags , int pc ) ;
void trace_buffer_unlock_commit ( struct trace_array * tr ,
struct ring_buffer_event * event ,
unsigned long flags , int pc )
{
ring_buffer_unlock_commit ( tr - > buffer , event ) ;
ftrace_trace_stack ( tr , flags , 6 , pc ) ;
ftrace_trace_userstack ( tr , flags , pc ) ;
trace_wake_up ( ) ;
}
2008-05-12 23:20:51 +04:00
void
2009-02-05 09:13:37 +03:00
trace_function ( struct trace_array * tr ,
2008-10-01 21:14:09 +04:00
unsigned long ip , unsigned long parent_ip , unsigned long flags ,
int pc )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct ftrace_entry * entry ;
2008-05-12 23:20:42 +04:00
2008-10-01 08:29:53 +04:00
/* If we are reading the ring buffer, don't trace */
if ( unlikely ( local_read ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ) )
return ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_FN , sizeof ( * entry ) ,
flags , pc ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-09-30 07:02:42 +04:00
entry - > ip = ip ;
entry - > parent_ip = parent_ip ;
2009-02-05 21:12:56 +03:00
ring_buffer_unlock_commit ( tr - > buffer , event ) ;
2008-05-12 23:20:42 +04:00
}
2008-11-25 23:07:04 +03:00
# ifdef CONFIG_FUNCTION_GRAPH_TRACER
2008-11-26 02:57:25 +03:00
static void __trace_graph_entry ( struct trace_array * tr ,
struct ftrace_graph_ent * trace ,
unsigned long flags ,
int pc )
{
struct ring_buffer_event * event ;
struct ftrace_graph_ent_entry * entry ;
if ( unlikely ( local_read ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ) )
return ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( & global_trace , TRACE_GRAPH_ENT ,
sizeof ( * entry ) , flags , pc ) ;
2008-11-26 02:57:25 +03:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
entry - > graph_ent = * trace ;
2009-02-05 21:12:56 +03:00
ring_buffer_unlock_commit ( global_trace . buffer , event ) ;
2008-11-26 02:57:25 +03:00
}
static void __trace_graph_return ( struct trace_array * tr ,
2008-11-25 23:07:04 +03:00
struct ftrace_graph_ret * trace ,
2008-11-11 09:14:25 +03:00
unsigned long flags ,
int pc )
{
struct ring_buffer_event * event ;
2008-11-26 02:57:25 +03:00
struct ftrace_graph_ret_entry * entry ;
2008-11-11 09:14:25 +03:00
if ( unlikely ( local_read ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ) )
return ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( & global_trace , TRACE_GRAPH_RET ,
sizeof ( * entry ) , flags , pc ) ;
2008-11-11 09:14:25 +03:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-11-26 02:57:25 +03:00
entry - > ret = * trace ;
2009-02-05 21:12:56 +03:00
ring_buffer_unlock_commit ( global_trace . buffer , event ) ;
2008-11-11 09:14:25 +03:00
}
# endif
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:49 +04:00
ftrace ( struct trace_array * tr , struct trace_array_cpu * data ,
2008-10-01 21:14:09 +04:00
unsigned long ip , unsigned long parent_ip , unsigned long flags ,
int pc )
2008-05-12 23:20:49 +04:00
{
if ( likely ( ! atomic_read ( & data - > disabled ) ) )
2009-02-05 09:13:37 +03:00
trace_function ( tr , ip , parent_ip , flags , pc ) ;
2008-05-12 23:20:49 +04:00
}
2009-01-16 03:12:40 +03:00
static void __ftrace_trace_stack ( struct trace_array * tr ,
unsigned long flags ,
int skip , int pc )
2008-05-12 23:20:51 +04:00
{
2008-10-31 22:50:41 +03:00
# ifdef CONFIG_STACKTRACE
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct stack_entry * entry ;
2008-05-12 23:20:51 +04:00
struct stack_trace trace ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_STACK ,
sizeof ( * entry ) , flags , pc ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-09-30 07:02:42 +04:00
memset ( & entry - > caller , 0 , sizeof ( entry - > caller ) ) ;
2008-05-12 23:20:51 +04:00
trace . nr_entries = 0 ;
trace . max_entries = FTRACE_STACK_ENTRIES ;
trace . skip = skip ;
2008-09-30 07:02:42 +04:00
trace . entries = entry - > caller ;
2008-05-12 23:20:51 +04:00
save_stack_trace ( & trace ) ;
2009-02-05 21:12:56 +03:00
ring_buffer_unlock_commit ( tr - > buffer , event ) ;
2008-10-31 22:50:41 +03:00
# endif
2008-05-12 23:20:47 +04:00
}
2009-01-16 03:12:40 +03:00
static void ftrace_trace_stack ( struct trace_array * tr ,
unsigned long flags ,
int skip , int pc )
{
if ( ! ( trace_flags & TRACE_ITER_STACKTRACE ) )
return ;
2009-02-05 09:13:37 +03:00
__ftrace_trace_stack ( tr , flags , skip , pc ) ;
2009-01-16 03:12:40 +03:00
}
2008-10-01 21:14:09 +04:00
void __trace_stack ( struct trace_array * tr ,
unsigned long flags ,
2009-01-16 03:12:40 +03:00
int skip , int pc )
2008-10-01 21:14:09 +04:00
{
2009-02-05 09:13:37 +03:00
__ftrace_trace_stack ( tr , flags , skip , pc ) ;
2008-10-01 21:14:09 +04:00
}
2008-11-22 14:28:47 +03:00
static void ftrace_trace_userstack ( struct trace_array * tr ,
2009-02-05 09:13:37 +03:00
unsigned long flags , int pc )
2008-11-22 14:28:47 +03:00
{
2008-11-28 12:17:56 +03:00
# ifdef CONFIG_STACKTRACE
2008-11-23 13:39:06 +03:00
struct ring_buffer_event * event ;
2008-11-22 14:28:47 +03:00
struct userstack_entry * entry ;
struct stack_trace trace ;
if ( ! ( trace_flags & TRACE_ITER_USERSTACKTRACE ) )
return ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_USER_STACK ,
sizeof ( * entry ) , flags , pc ) ;
2008-11-22 14:28:47 +03:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
memset ( & entry - > caller , 0 , sizeof ( entry - > caller ) ) ;
trace . nr_entries = 0 ;
trace . max_entries = FTRACE_STACK_ENTRIES ;
trace . skip = 0 ;
trace . entries = entry - > caller ;
save_stack_trace_user ( & trace ) ;
2009-02-05 21:12:56 +03:00
ring_buffer_unlock_commit ( tr - > buffer , event ) ;
2008-11-28 12:17:56 +03:00
# endif
2008-11-22 14:28:47 +03:00
}
2009-02-05 09:13:37 +03:00
void __trace_userstack ( struct trace_array * tr , unsigned long flags )
2008-11-22 14:28:47 +03:00
{
2009-02-05 09:13:37 +03:00
ftrace_trace_userstack ( tr , flags , preempt_count ( ) ) ;
2008-11-22 14:28:47 +03:00
}
2008-10-01 21:14:09 +04:00
static void
2009-02-05 09:13:37 +03:00
ftrace_trace_special ( void * __tr ,
2008-10-01 21:14:09 +04:00
unsigned long arg1 , unsigned long arg2 , unsigned long arg3 ,
int pc )
2008-05-12 23:21:02 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-05-12 23:21:02 +04:00
struct trace_array * tr = __tr ;
2008-09-30 07:02:42 +04:00
struct special_entry * entry ;
2008-05-12 23:21:02 +04:00
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_SPECIAL ,
sizeof ( * entry ) , 0 , pc ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-09-30 07:02:42 +04:00
entry - > arg1 = arg1 ;
entry - > arg2 = arg2 ;
entry - > arg3 = arg3 ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
trace_buffer_unlock_commit ( tr , event , 0 , pc ) ;
2008-05-12 23:21:02 +04:00
}
2008-10-01 21:14:09 +04:00
void
__trace_special ( void * __tr , void * __data ,
unsigned long arg1 , unsigned long arg2 , unsigned long arg3 )
{
2009-02-05 09:13:37 +03:00
ftrace_trace_special ( __tr , arg1 , arg2 , arg3 , preempt_count ( ) ) ;
2008-10-01 21:14:09 +04:00
}
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:42 +04:00
tracing_sched_switch_trace ( struct trace_array * tr ,
2008-05-12 23:20:51 +04:00
struct task_struct * prev ,
struct task_struct * next ,
2008-10-01 21:14:09 +04:00
unsigned long flags , int pc )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct ctx_switch_entry * entry ;
2008-05-12 23:20:42 +04:00
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_CTX ,
sizeof ( * entry ) , flags , pc ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-09-30 07:02:42 +04:00
entry - > prev_pid = prev - > pid ;
entry - > prev_prio = prev - > prio ;
entry - > prev_state = prev - > state ;
entry - > next_pid = next - > pid ;
entry - > next_prio = next - > prio ;
entry - > next_state = next - > state ;
entry - > next_cpu = task_cpu ( next ) ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
trace_buffer_unlock_commit ( tr , event , flags , pc ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
void
tracing_sched_wakeup_trace ( struct trace_array * tr ,
2008-05-12 23:20:51 +04:00
struct task_struct * wakee ,
struct task_struct * curr ,
2008-10-01 21:14:09 +04:00
unsigned long flags , int pc )
2008-05-12 23:20:51 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct ctx_switch_entry * entry ;
2008-05-12 23:20:51 +04:00
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_WAKE ,
sizeof ( * entry ) , flags , pc ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-09-30 07:02:42 +04:00
entry - > prev_pid = curr - > pid ;
entry - > prev_prio = curr - > prio ;
entry - > prev_state = curr - > state ;
entry - > next_pid = wakee - > pid ;
entry - > next_prio = wakee - > prio ;
entry - > next_state = wakee - > state ;
entry - > next_cpu = task_cpu ( wakee ) ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
trace_buffer_unlock_commit ( tr , event , flags , pc ) ;
2008-05-12 23:20:51 +04:00
}
2008-05-22 08:22:18 +04:00
void
ftrace_special ( unsigned long arg1 , unsigned long arg2 , unsigned long arg3 )
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
2008-11-11 07:07:30 +03:00
unsigned long flags ;
2008-05-22 08:22:18 +04:00
int cpu ;
2008-10-01 21:14:09 +04:00
int pc ;
2008-05-22 08:22:18 +04:00
2008-11-08 06:36:02 +03:00
if ( tracing_disabled )
2008-05-22 08:22:18 +04:00
return ;
2008-10-01 21:14:09 +04:00
pc = preempt_count ( ) ;
2008-11-11 07:07:30 +03:00
local_irq_save ( flags ) ;
2008-05-22 08:22:18 +04:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
2008-11-11 07:07:30 +03:00
if ( likely ( atomic_inc_return ( & data - > disabled ) = = 1 ) )
2009-02-05 09:13:37 +03:00
ftrace_trace_special ( tr , arg1 , arg2 , arg3 , pc ) ;
2008-05-22 08:22:18 +04:00
2008-11-11 07:07:30 +03:00
atomic_dec ( & data - > disabled ) ;
local_irq_restore ( flags ) ;
2008-05-22 08:22:18 +04:00
}
2008-11-25 23:07:04 +03:00
# ifdef CONFIG_FUNCTION_GRAPH_TRACER
2008-12-03 07:50:05 +03:00
int trace_graph_entry ( struct ftrace_graph_ent * trace )
2008-11-11 09:14:25 +03:00
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
unsigned long flags ;
long disabled ;
int cpu ;
int pc ;
2008-12-03 23:36:59 +03:00
if ( ! ftrace_trace_task ( current ) )
return 0 ;
2008-12-03 23:36:57 +03:00
if ( ! ftrace_graph_addr ( trace - > func ) )
return 0 ;
2008-12-02 23:34:05 +03:00
local_irq_save ( flags ) ;
2008-11-11 09:14:25 +03:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
disabled = atomic_inc_return ( & data - > disabled ) ;
if ( likely ( disabled = = 1 ) ) {
pc = preempt_count ( ) ;
2009-02-05 09:13:37 +03:00
__trace_graph_entry ( tr , trace , flags , pc ) ;
2008-11-26 02:57:25 +03:00
}
2008-12-03 23:36:57 +03:00
/* Only do the atomic if it is not already set */
if ( ! test_tsk_trace_graph ( current ) )
set_tsk_trace_graph ( current ) ;
2008-11-26 02:57:25 +03:00
atomic_dec ( & data - > disabled ) ;
2008-12-02 23:34:05 +03:00
local_irq_restore ( flags ) ;
2008-12-03 07:50:05 +03:00
return 1 ;
2008-11-26 02:57:25 +03:00
}
void trace_graph_return ( struct ftrace_graph_ret * trace )
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
unsigned long flags ;
long disabled ;
int cpu ;
int pc ;
2008-12-02 23:34:05 +03:00
local_irq_save ( flags ) ;
2008-11-26 02:57:25 +03:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
disabled = atomic_inc_return ( & data - > disabled ) ;
if ( likely ( disabled = = 1 ) ) {
pc = preempt_count ( ) ;
2009-02-05 09:13:37 +03:00
__trace_graph_return ( tr , trace , flags , pc ) ;
2008-11-11 09:14:25 +03:00
}
2008-12-03 23:36:57 +03:00
if ( ! trace - > depth )
clear_tsk_trace_graph ( current ) ;
2008-11-11 09:14:25 +03:00
atomic_dec ( & data - > disabled ) ;
2008-12-02 23:34:05 +03:00
local_irq_restore ( flags ) ;
2008-11-11 09:14:25 +03:00
}
2008-11-25 23:07:04 +03:00
# endif /* CONFIG_FUNCTION_GRAPH_TRACER */
2008-11-11 09:14:25 +03:00
2008-05-12 23:20:42 +04:00
enum trace_file_type {
TRACE_FILE_LAT_FMT = 1 ,
2008-11-13 01:52:38 +03:00
TRACE_FILE_ANNOTATE = 2 ,
2008-05-12 23:20:42 +04:00
} ;
2008-11-12 14:59:32 +03:00
static void trace_iterator_increment ( struct trace_iterator * iter )
2008-09-04 01:42:51 +04:00
{
2008-10-01 08:29:53 +04:00
/* Don't allow ftrace to trace into the ring buffers */
ftrace_disable_cpu ( ) ;
2008-09-04 01:42:51 +04:00
iter - > idx + + ;
2008-10-01 08:29:53 +04:00
if ( iter - > buffer_iter [ iter - > cpu ] )
ring_buffer_read ( iter - > buffer_iter [ iter - > cpu ] , NULL ) ;
ftrace_enable_cpu ( ) ;
2008-09-04 01:42:51 +04:00
}
2008-05-12 23:20:51 +04:00
static struct trace_entry *
2008-09-30 07:02:41 +04:00
peek_next_entry ( struct trace_iterator * iter , int cpu , u64 * ts )
2008-08-01 20:26:41 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
struct ring_buffer_iter * buf_iter = iter - > buffer_iter [ cpu ] ;
2008-08-01 20:26:41 +04:00
2008-10-01 08:29:53 +04:00
/* Don't allow ftrace to trace into the ring buffers */
ftrace_disable_cpu ( ) ;
if ( buf_iter )
event = ring_buffer_iter_peek ( buf_iter , ts ) ;
else
event = ring_buffer_peek ( iter - > tr - > buffer , cpu , ts ) ;
ftrace_enable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
return event ? ring_buffer_event_data ( event ) : NULL ;
2008-08-01 20:26:41 +04:00
}
2008-10-01 08:29:53 +04:00
2008-08-01 20:26:41 +04:00
static struct trace_entry *
2008-09-30 07:02:41 +04:00
__find_next_entry ( struct trace_iterator * iter , int * ent_cpu , u64 * ent_ts )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer * buffer = iter - > tr - > buffer ;
2008-05-12 23:20:42 +04:00
struct trace_entry * ent , * next = NULL ;
2008-09-30 07:02:41 +04:00
u64 next_ts = 0 , ts ;
2008-05-12 23:20:42 +04:00
int next_cpu = - 1 ;
int cpu ;
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( cpu ) {
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
if ( ring_buffer_empty_cpu ( buffer , cpu ) )
continue ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
ent = peek_next_entry ( iter , cpu , & ts ) ;
2008-08-01 20:26:41 +04:00
2008-05-12 23:20:46 +04:00
/*
* Pick the entry with the smallest timestamp :
*/
2008-09-30 07:02:41 +04:00
if ( ent & & ( ! next | | ts < next_ts ) ) {
2008-05-12 23:20:42 +04:00
next = ent ;
next_cpu = cpu ;
2008-09-30 07:02:41 +04:00
next_ts = ts ;
2008-05-12 23:20:42 +04:00
}
}
if ( ent_cpu )
* ent_cpu = next_cpu ;
2008-09-30 07:02:41 +04:00
if ( ent_ts )
* ent_ts = next_ts ;
2008-05-12 23:20:42 +04:00
return next ;
}
2008-08-01 20:26:41 +04:00
/* Find the next real entry, without updating the iterator itself */
2009-02-03 01:29:21 +03:00
struct trace_entry * trace_find_next_entry ( struct trace_iterator * iter ,
int * ent_cpu , u64 * ent_ts )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
return __find_next_entry ( iter , ent_cpu , ent_ts ) ;
2008-08-01 20:26:41 +04:00
}
/* Find the next real entry, and increment the iterator to the next entry */
static void * find_next_entry_inc ( struct trace_iterator * iter )
{
2008-09-30 07:02:41 +04:00
iter - > ent = __find_next_entry ( iter , & iter - > cpu , & iter - > ts ) ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
if ( iter - > ent )
2008-11-12 14:59:32 +03:00
trace_iterator_increment ( iter ) ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
return iter - > ent ? iter : NULL ;
2008-05-12 23:20:46 +04:00
}
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:51 +04:00
static void trace_consume ( struct trace_iterator * iter )
2008-05-12 23:20:46 +04:00
{
2008-10-01 08:29:53 +04:00
/* Don't allow ftrace to trace into the ring buffers */
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_consume ( iter - > tr - > buffer , iter - > cpu , & iter - > ts ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void * s_next ( struct seq_file * m , void * v , loff_t * pos )
2008-05-12 23:20:42 +04:00
{
struct trace_iterator * iter = m - > private ;
int i = ( int ) * pos ;
2008-05-12 23:20:45 +04:00
void * ent ;
2008-05-12 23:20:42 +04:00
( * pos ) + + ;
/* can't go backwards */
if ( iter - > idx > i )
return NULL ;
if ( iter - > idx < 0 )
ent = find_next_entry_inc ( iter ) ;
else
ent = iter ;
while ( ent & & iter - > idx < i )
ent = find_next_entry_inc ( iter ) ;
iter - > pos = * pos ;
return ent ;
}
static void * s_start ( struct seq_file * m , loff_t * pos )
{
struct trace_iterator * iter = m - > private ;
void * p = NULL ;
loff_t l = 0 ;
2008-09-30 07:02:41 +04:00
int cpu ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-05-12 23:20:56 +04:00
if ( ! current_trace | | current_trace ! = iter - > trace ) {
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:20:42 +04:00
return NULL ;
2008-05-12 23:20:56 +04:00
}
2008-05-12 23:20:42 +04:00
atomic_inc ( & trace_record_cmdline_disabled ) ;
if ( * pos ! = iter - > pos ) {
iter - > ent = NULL ;
iter - > cpu = 0 ;
iter - > idx = - 1 ;
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
for_each_tracing_cpu ( cpu ) {
ring_buffer_iter_reset ( iter - > buffer_iter [ cpu ] ) ;
2008-05-12 23:20:43 +04:00
}
2008-05-12 23:20:42 +04:00
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
for ( p = iter ; p & & l < * pos ; p = s_next ( m , p , & l ) )
;
} else {
2008-05-12 23:20:43 +04:00
l = * pos - 1 ;
2008-05-12 23:20:42 +04:00
p = s_next ( m , p , & l ) ;
}
return p ;
}
static void s_stop ( struct seq_file * m , void * p )
{
atomic_dec ( & trace_record_cmdline_disabled ) ;
mutex_unlock ( & trace_types_lock ) ;
}
2008-05-12 23:20:51 +04:00
static void print_lat_help_header ( struct seq_file * m )
2008-05-12 23:20:42 +04:00
{
2008-08-21 03:36:11 +04:00
seq_puts ( m , " # _------=> CPU# \n " ) ;
seq_puts ( m , " # / _-----=> irqs-off \n " ) ;
seq_puts ( m , " # | / _----=> need-resched \n " ) ;
seq_puts ( m , " # || / _---=> hardirq/softirq \n " ) ;
seq_puts ( m , " # ||| / _--=> preempt-depth \n " ) ;
seq_puts ( m , " # |||| / \n " ) ;
seq_puts ( m , " # ||||| delay \n " ) ;
seq_puts ( m , " # cmd pid ||||| time | caller \n " ) ;
seq_puts ( m , " # \\ / ||||| \\ | / \n " ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void print_func_help_header ( struct seq_file * m )
2008-05-12 23:20:42 +04:00
{
2008-08-21 03:36:11 +04:00
seq_puts ( m , " # TASK-PID CPU# TIMESTAMP FUNCTION \n " ) ;
seq_puts ( m , " # | | | | | \n " ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:42 +04:00
print_trace_header ( struct seq_file * m , struct trace_iterator * iter )
{
unsigned long sym_flags = ( trace_flags & TRACE_ITER_SYM_MASK ) ;
struct trace_array * tr = iter - > tr ;
struct trace_array_cpu * data = tr - > data [ tr - > cpu ] ;
struct tracer * type = current_trace ;
2008-09-30 07:02:41 +04:00
unsigned long total ;
unsigned long entries ;
2008-05-12 23:20:42 +04:00
const char * name = " preemption " ;
if ( type )
name = type - > name ;
2008-09-30 07:02:41 +04:00
entries = ring_buffer_entries ( iter - > tr - > buffer ) ;
total = entries +
ring_buffer_overruns ( iter - > tr - > buffer ) ;
2008-05-12 23:20:42 +04:00
seq_printf ( m , " %s latency trace v1.1.5 on %s \n " ,
name , UTS_RELEASE ) ;
seq_puts ( m , " ----------------------------------- "
" --------------------------------- \n " ) ;
seq_printf ( m , " latency: %lu us, #%lu/%lu, CPU#%d | "
" (M:%s VP:%d, KP:%d, SP:%d HP:%d " ,
2008-05-12 23:20:44 +04:00
nsecs_to_usecs ( data - > saved_latency ) ,
2008-05-12 23:20:42 +04:00
entries ,
2008-05-12 23:20:43 +04:00
total ,
2008-05-12 23:20:42 +04:00
tr - > cpu ,
# if defined(CONFIG_PREEMPT_NONE)
" server " ,
# elif defined(CONFIG_PREEMPT_VOLUNTARY)
" desktop " ,
2008-07-11 04:58:12 +04:00
# elif defined(CONFIG_PREEMPT)
2008-05-12 23:20:42 +04:00
" preempt " ,
# else
" unknown " ,
# endif
/* These are reserved for later use */
0 , 0 , 0 , 0 ) ;
# ifdef CONFIG_SMP
seq_printf ( m , " #P:%d) \n " , num_online_cpus ( ) ) ;
# else
seq_puts ( m , " ) \n " ) ;
# endif
seq_puts ( m , " ----------------- \n " ) ;
seq_printf ( m , " | task: %.16s-%d "
" (uid:%d nice:%ld policy:%ld rt_prio:%ld) \n " ,
data - > comm , data - > pid , data - > uid , data - > nice ,
data - > policy , data - > rt_priority ) ;
seq_puts ( m , " ----------------- \n " ) ;
if ( data - > critical_start ) {
seq_puts ( m , " => started at: " ) ;
2008-05-12 23:20:46 +04:00
seq_print_ip_sym ( & iter - > seq , data - > critical_start , sym_flags ) ;
trace_print_seq ( m , & iter - > seq ) ;
2008-05-12 23:20:42 +04:00
seq_puts ( m , " \n => ended at: " ) ;
2008-05-12 23:20:46 +04:00
seq_print_ip_sym ( & iter - > seq , data - > critical_end , sym_flags ) ;
trace_print_seq ( m , & iter - > seq ) ;
2008-05-12 23:20:42 +04:00
seq_puts ( m , " \n " ) ;
}
seq_puts ( m , " \n " ) ;
}
2008-11-08 06:36:02 +03:00
static void test_cpu_buff_start ( struct trace_iterator * iter )
{
struct trace_seq * s = & iter - > seq ;
2008-11-13 01:52:38 +03:00
if ( ! ( trace_flags & TRACE_ITER_ANNOTATE ) )
return ;
if ( ! ( iter - > iter_flags & TRACE_FILE_ANNOTATE ) )
return ;
2009-01-01 02:42:23 +03:00
if ( cpumask_test_cpu ( iter - > cpu , iter - > started ) )
2008-11-08 06:36:02 +03:00
return ;
2009-01-01 02:42:23 +03:00
cpumask_set_cpu ( iter - > cpu , iter - > started ) ;
2008-11-08 06:36:02 +03:00
trace_seq_printf ( s , " ##### CPU %u buffer started #### \n " , iter - > cpu ) ;
}
2009-02-03 01:29:21 +03:00
static enum print_line_t print_lat_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:46 +04:00
struct trace_seq * s = & iter - > seq ;
2008-05-12 23:20:42 +04:00
unsigned long sym_flags = ( trace_flags & TRACE_ITER_SYM_MASK ) ;
2008-12-24 07:24:13 +03:00
struct trace_event * event ;
2008-05-12 23:20:42 +04:00
struct trace_entry * entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
2008-11-08 06:36:02 +03:00
test_cpu_buff_start ( iter ) ;
2009-02-03 01:29:21 +03:00
event = ftrace_find_event ( entry - > type ) ;
if ( trace_flags & TRACE_ITER_CONTEXT_INFO ) {
2009-02-04 01:20:41 +03:00
if ( ! trace_print_lat_context ( iter ) )
goto partial ;
2008-05-12 23:20:42 +04:00
}
2008-09-30 07:02:42 +04:00
2009-02-05 01:16:39 +03:00
if ( event )
2009-02-04 01:20:41 +03:00
return event - > latency_trace ( iter , sym_flags ) ;
if ( ! trace_seq_printf ( s , " Unknown type %d \n " , entry - > type ) )
goto partial ;
2008-11-22 14:28:47 +03:00
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2009-02-04 01:20:41 +03:00
partial :
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_trace_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:46 +04:00
struct trace_seq * s = & iter - > seq ;
2008-05-12 23:20:42 +04:00
unsigned long sym_flags = ( trace_flags & TRACE_ITER_SYM_MASK ) ;
2008-05-12 23:20:45 +04:00
struct trace_entry * entry ;
2008-12-24 07:24:13 +03:00
struct trace_event * event ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:45 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
2008-11-08 06:36:02 +03:00
test_cpu_buff_start ( iter ) ;
2009-02-03 01:29:21 +03:00
event = ftrace_find_event ( entry - > type ) ;
2008-05-12 23:20:42 +04:00
2009-02-03 01:29:21 +03:00
if ( trace_flags & TRACE_ITER_CONTEXT_INFO ) {
2009-02-04 01:20:41 +03:00
if ( ! trace_print_context ( iter ) )
goto partial ;
2009-02-03 01:29:21 +03:00
}
2008-05-12 23:20:42 +04:00
2009-02-05 01:16:39 +03:00
if ( event )
2009-02-04 01:20:41 +03:00
return event - > trace ( iter , sym_flags ) ;
if ( ! trace_seq_printf ( s , " Unknown type %d \n " , entry - > type ) )
goto partial ;
2008-11-22 14:28:47 +03:00
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2009-02-04 01:20:41 +03:00
partial :
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_raw_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:47 +04:00
{
struct trace_seq * s = & iter - > seq ;
struct trace_entry * entry ;
2008-12-24 07:24:13 +03:00
struct trace_event * event ;
2008-05-12 23:20:47 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
2009-02-03 01:29:21 +03:00
if ( trace_flags & TRACE_ITER_CONTEXT_INFO ) {
2009-02-04 01:20:41 +03:00
if ( ! trace_seq_printf ( s , " %d %d %llu " ,
entry - > pid , iter - > cpu , iter - > ts ) )
goto partial ;
2009-02-03 01:29:21 +03:00
}
2008-05-12 23:20:47 +04:00
2008-12-24 07:24:13 +03:00
event = ftrace_find_event ( entry - > type ) ;
2009-02-05 01:16:39 +03:00
if ( event )
2009-02-04 01:20:41 +03:00
return event - > raw ( iter , 0 ) ;
if ( ! trace_seq_printf ( s , " %d ? \n " , entry - > type ) )
goto partial ;
2008-09-30 07:02:42 +04:00
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2009-02-04 01:20:41 +03:00
partial :
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:47 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_hex_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:49 +04:00
{
struct trace_seq * s = & iter - > seq ;
unsigned char newline = ' \n ' ;
struct trace_entry * entry ;
2008-12-24 07:24:13 +03:00
struct trace_event * event ;
2008-05-12 23:20:49 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
2009-02-03 01:29:21 +03:00
if ( trace_flags & TRACE_ITER_CONTEXT_INFO ) {
SEQ_PUT_HEX_FIELD_RET ( s , entry - > pid ) ;
SEQ_PUT_HEX_FIELD_RET ( s , iter - > cpu ) ;
SEQ_PUT_HEX_FIELD_RET ( s , iter - > ts ) ;
}
2008-05-12 23:20:49 +04:00
2008-12-24 07:24:13 +03:00
event = ftrace_find_event ( entry - > type ) ;
2009-02-05 01:16:39 +03:00
if ( event ) {
2009-02-04 03:05:50 +03:00
enum print_line_t ret = event - > hex ( iter , 0 ) ;
2009-02-04 01:20:41 +03:00
if ( ret ! = TRACE_TYPE_HANDLED )
return ret ;
}
2008-10-01 18:52:51 +04:00
2008-05-12 23:20:49 +04:00
SEQ_PUT_FIELD_RET ( s , newline ) ;
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-05-12 23:20:49 +04:00
}
2008-12-13 22:18:13 +03:00
static enum print_line_t print_printk_msg_only ( struct trace_iterator * iter )
{
struct trace_seq * s = & iter - > seq ;
struct trace_entry * entry = iter - > ent ;
struct print_entry * field ;
int ret ;
trace_assign_type ( field , entry ) ;
2009-02-03 01:29:21 +03:00
ret = trace_seq_printf ( s , " %s " , field - > buf ) ;
2008-12-13 22:18:13 +03:00
if ( ! ret )
return TRACE_TYPE_PARTIAL_LINE ;
return TRACE_TYPE_HANDLED ;
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_bin_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:47 +04:00
{
struct trace_seq * s = & iter - > seq ;
struct trace_entry * entry ;
2008-12-24 07:24:13 +03:00
struct trace_event * event ;
2008-05-12 23:20:47 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
2009-02-03 01:29:21 +03:00
if ( trace_flags & TRACE_ITER_CONTEXT_INFO ) {
SEQ_PUT_FIELD_RET ( s , entry - > pid ) ;
2009-02-08 03:38:43 +03:00
SEQ_PUT_FIELD_RET ( s , iter - > cpu ) ;
2009-02-03 01:29:21 +03:00
SEQ_PUT_FIELD_RET ( s , iter - > ts ) ;
}
2008-05-12 23:20:47 +04:00
2008-12-24 07:24:13 +03:00
event = ftrace_find_event ( entry - > type ) ;
2009-02-05 01:16:39 +03:00
return event ? event - > binary ( iter , 0 ) : TRACE_TYPE_HANDLED ;
2008-05-12 23:20:47 +04:00
}
2008-05-12 23:20:42 +04:00
static int trace_empty ( struct trace_iterator * iter )
{
int cpu ;
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( cpu ) {
2008-10-01 08:29:53 +04:00
if ( iter - > buffer_iter [ cpu ] ) {
if ( ! ring_buffer_iter_empty ( iter - > buffer_iter [ cpu ] ) )
return 0 ;
} else {
if ( ! ring_buffer_empty_cpu ( iter - > tr - > buffer , cpu ) )
return 0 ;
}
2008-05-12 23:20:42 +04:00
}
2008-10-01 08:29:53 +04:00
2008-09-30 20:13:45 +04:00
return 1 ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_trace_line ( struct trace_iterator * iter )
2008-05-12 23:20:47 +04:00
{
2008-09-29 22:18:34 +04:00
enum print_line_t ret ;
if ( iter - > trace & & iter - > trace - > print_line ) {
ret = iter - > trace - > print_line ( iter ) ;
if ( ret ! = TRACE_TYPE_UNHANDLED )
return ret ;
}
2008-05-23 23:37:28 +04:00
2008-12-13 22:18:13 +03:00
if ( iter - > ent - > type = = TRACE_PRINT & &
trace_flags & TRACE_ITER_PRINTK & &
trace_flags & TRACE_ITER_PRINTK_MSGONLY )
return print_printk_msg_only ( iter ) ;
2008-05-12 23:20:47 +04:00
if ( trace_flags & TRACE_ITER_BIN )
return print_bin_fmt ( iter ) ;
2008-05-12 23:20:49 +04:00
if ( trace_flags & TRACE_ITER_HEX )
return print_hex_fmt ( iter ) ;
2008-05-12 23:20:47 +04:00
if ( trace_flags & TRACE_ITER_RAW )
return print_raw_fmt ( iter ) ;
if ( iter - > iter_flags & TRACE_FILE_LAT_FMT )
2009-02-03 01:29:21 +03:00
return print_lat_fmt ( iter ) ;
2008-05-12 23:20:47 +04:00
return print_trace_fmt ( iter ) ;
}
2008-05-12 23:20:42 +04:00
static int s_show ( struct seq_file * m , void * v )
{
struct trace_iterator * iter = v ;
if ( iter - > ent = = NULL ) {
if ( iter - > tr ) {
seq_printf ( m , " # tracer: %s \n " , iter - > trace - > name ) ;
seq_puts ( m , " # \n " ) ;
}
2008-11-25 11:12:31 +03:00
if ( iter - > trace & & iter - > trace - > print_header )
iter - > trace - > print_header ( m ) ;
else if ( iter - > iter_flags & TRACE_FILE_LAT_FMT ) {
2008-05-12 23:20:42 +04:00
/* print nothing if the buffers are empty */
if ( trace_empty ( iter ) )
return 0 ;
print_trace_header ( m , iter ) ;
if ( ! ( trace_flags & TRACE_ITER_VERBOSE ) )
print_lat_help_header ( m ) ;
} else {
if ( ! ( trace_flags & TRACE_ITER_VERBOSE ) )
print_func_help_header ( m ) ;
}
} else {
2008-05-12 23:20:47 +04:00
print_trace_line ( iter ) ;
2008-05-12 23:20:46 +04:00
trace_print_seq ( m , & iter - > seq ) ;
2008-05-12 23:20:42 +04:00
}
return 0 ;
}
static struct seq_operations tracer_seq_ops = {
2008-05-12 23:20:46 +04:00
. start = s_start ,
. next = s_next ,
. stop = s_stop ,
. show = s_show ,
2008-05-12 23:20:42 +04:00
} ;
2008-05-12 23:20:51 +04:00
static struct trace_iterator *
2008-05-12 23:20:42 +04:00
__tracing_open ( struct inode * inode , struct file * file , int * ret )
{
struct trace_iterator * iter ;
2008-09-30 07:02:41 +04:00
struct seq_file * m ;
int cpu ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:44 +04:00
if ( tracing_disabled ) {
* ret = - ENODEV ;
return NULL ;
}
2008-05-12 23:20:42 +04:00
iter = kzalloc ( sizeof ( * iter ) , GFP_KERNEL ) ;
if ( ! iter ) {
* ret = - ENOMEM ;
goto out ;
}
mutex_lock ( & trace_types_lock ) ;
if ( current_trace & & current_trace - > print_max )
iter - > tr = & max_tr ;
else
iter - > tr = inode - > i_private ;
iter - > trace = current_trace ;
iter - > pos = - 1 ;
2008-11-25 11:12:31 +03:00
/* Notify the tracer early; before we stop tracing. */
if ( iter - > trace & & iter - > trace - > open )
2008-12-11 15:53:26 +03:00
iter - > trace - > open ( iter ) ;
2008-11-25 11:12:31 +03:00
2008-11-13 01:52:38 +03:00
/* Annotate start of buffers if we had overruns */
if ( ring_buffer_overruns ( iter - > tr - > buffer ) )
iter - > iter_flags | = TRACE_FILE_ANNOTATE ;
2008-09-30 07:02:41 +04:00
for_each_tracing_cpu ( cpu ) {
2008-10-01 08:29:53 +04:00
2008-09-30 07:02:41 +04:00
iter - > buffer_iter [ cpu ] =
ring_buffer_read_start ( iter - > tr - > buffer , cpu ) ;
2008-10-01 08:29:53 +04:00
2008-09-30 07:02:41 +04:00
if ( ! iter - > buffer_iter [ cpu ] )
goto fail_buffer ;
}
2008-05-12 23:20:42 +04:00
/* TODO stop tracer */
* ret = seq_open ( file , & tracer_seq_ops ) ;
2008-09-30 07:02:41 +04:00
if ( * ret )
goto fail_buffer ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
m = file - > private_data ;
m - > private = iter ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
/* stop the trace while dumping */
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
tracing_stop ( ) ;
2008-09-30 07:02:41 +04:00
2008-05-12 23:20:42 +04:00
mutex_unlock ( & trace_types_lock ) ;
out :
return iter ;
2008-09-30 07:02:41 +04:00
fail_buffer :
for_each_tracing_cpu ( cpu ) {
if ( iter - > buffer_iter [ cpu ] )
ring_buffer_read_finish ( iter - > buffer_iter [ cpu ] ) ;
}
mutex_unlock ( & trace_types_lock ) ;
2008-11-14 21:05:31 +03:00
kfree ( iter ) ;
2008-09-30 07:02:41 +04:00
return ERR_PTR ( - ENOMEM ) ;
2008-05-12 23:20:42 +04:00
}
int tracing_open_generic ( struct inode * inode , struct file * filp )
{
2008-05-12 23:20:44 +04:00
if ( tracing_disabled )
return - ENODEV ;
2008-05-12 23:20:42 +04:00
filp - > private_data = inode - > i_private ;
return 0 ;
}
int tracing_release ( struct inode * inode , struct file * file )
{
struct seq_file * m = ( struct seq_file * ) file - > private_data ;
struct trace_iterator * iter = m - > private ;
2008-09-30 07:02:41 +04:00
int cpu ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-09-30 07:02:41 +04:00
for_each_tracing_cpu ( cpu ) {
if ( iter - > buffer_iter [ cpu ] )
ring_buffer_read_finish ( iter - > buffer_iter [ cpu ] ) ;
}
2008-05-12 23:20:42 +04:00
if ( iter - > trace & & iter - > trace - > close )
iter - > trace - > close ( iter ) ;
/* reenable tracing if it was previously enabled */
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
tracing_start ( ) ;
2008-05-12 23:20:42 +04:00
mutex_unlock ( & trace_types_lock ) ;
seq_release ( inode , file ) ;
kfree ( iter ) ;
return 0 ;
}
static int tracing_open ( struct inode * inode , struct file * file )
{
int ret ;
__tracing_open ( inode , file , & ret ) ;
return ret ;
}
static int tracing_lt_open ( struct inode * inode , struct file * file )
{
struct trace_iterator * iter ;
int ret ;
iter = __tracing_open ( inode , file , & ret ) ;
if ( ! ret )
iter - > iter_flags | = TRACE_FILE_LAT_FMT ;
return ret ;
}
2008-05-12 23:20:51 +04:00
static void *
2008-05-12 23:20:42 +04:00
t_next ( struct seq_file * m , void * v , loff_t * pos )
{
struct tracer * t = m - > private ;
( * pos ) + + ;
if ( t )
t = t - > next ;
m - > private = t ;
return t ;
}
static void * t_start ( struct seq_file * m , loff_t * pos )
{
struct tracer * t = m - > private ;
loff_t l = 0 ;
mutex_lock ( & trace_types_lock ) ;
for ( ; t & & l < * pos ; t = t_next ( m , t , & l ) )
;
return t ;
}
static void t_stop ( struct seq_file * m , void * p )
{
mutex_unlock ( & trace_types_lock ) ;
}
static int t_show ( struct seq_file * m , void * v )
{
struct tracer * t = v ;
if ( ! t )
return 0 ;
seq_printf ( m , " %s " , t - > name ) ;
if ( t - > next )
seq_putc ( m , ' ' ) ;
else
seq_putc ( m , ' \n ' ) ;
return 0 ;
}
static struct seq_operations show_traces_seq_ops = {
2008-05-12 23:20:46 +04:00
. start = t_start ,
. next = t_next ,
. stop = t_stop ,
. show = t_show ,
2008-05-12 23:20:42 +04:00
} ;
static int show_traces_open ( struct inode * inode , struct file * file )
{
int ret ;
2008-05-12 23:20:44 +04:00
if ( tracing_disabled )
return - ENODEV ;
2008-05-12 23:20:42 +04:00
ret = seq_open ( file , & show_traces_seq_ops ) ;
if ( ! ret ) {
struct seq_file * m = file - > private_data ;
m - > private = trace_types ;
}
return ret ;
}
static struct file_operations tracing_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open ,
. read = seq_read ,
. llseek = seq_lseek ,
. release = tracing_release ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations tracing_lt_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_lt_open ,
. read = seq_read ,
. llseek = seq_lseek ,
. release = tracing_release ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations show_traces_fops = {
2008-05-12 23:20:52 +04:00
. open = show_traces_open ,
. read = seq_read ,
. release = seq_release ,
} ;
2008-05-12 23:20:52 +04:00
/*
* Only trace on a CPU if the bitmask is set :
*/
2009-01-01 02:42:22 +03:00
static cpumask_var_t tracing_cpumask ;
2008-05-12 23:20:52 +04:00
/*
* The tracer itself will not take this lock , but still we want
* to provide a consistent cpumask to user - space :
*/
static DEFINE_MUTEX ( tracing_cpumask_update_lock ) ;
/*
* Temporary storage for the character representation of the
* CPU bitmask ( and one more byte for the newline ) :
*/
static char mask_str [ NR_CPUS + 1 ] ;
2008-05-12 23:20:52 +04:00
static ssize_t
tracing_cpumask_read ( struct file * filp , char __user * ubuf ,
size_t count , loff_t * ppos )
{
2008-05-12 23:20:52 +04:00
int len ;
2008-05-12 23:20:52 +04:00
mutex_lock ( & tracing_cpumask_update_lock ) ;
2008-05-12 23:20:52 +04:00
2009-01-01 02:42:22 +03:00
len = cpumask_scnprintf ( mask_str , count , tracing_cpumask ) ;
2008-05-12 23:20:52 +04:00
if ( count - len < 2 ) {
count = - EINVAL ;
goto out_err ;
}
len + = sprintf ( mask_str + len , " \n " ) ;
count = simple_read_from_buffer ( ubuf , count , ppos , mask_str , NR_CPUS + 1 ) ;
out_err :
2008-05-12 23:20:52 +04:00
mutex_unlock ( & tracing_cpumask_update_lock ) ;
return count ;
}
static ssize_t
tracing_cpumask_write ( struct file * filp , const char __user * ubuf ,
size_t count , loff_t * ppos )
{
2008-05-12 23:20:52 +04:00
int err , cpu ;
2009-01-01 02:42:22 +03:00
cpumask_var_t tracing_cpumask_new ;
if ( ! alloc_cpumask_var ( & tracing_cpumask_new , GFP_KERNEL ) )
return - ENOMEM ;
2008-05-12 23:20:52 +04:00
mutex_lock ( & tracing_cpumask_update_lock ) ;
2009-01-01 02:42:22 +03:00
err = cpumask_parse_user ( ubuf , count , tracing_cpumask_new ) ;
2008-05-12 23:20:52 +04:00
if ( err )
2008-05-12 23:20:52 +04:00
goto err_unlock ;
2008-12-02 23:34:05 +03:00
local_irq_disable ( ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_lock ( & ftrace_max_lock ) ;
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( cpu ) {
2008-05-12 23:20:52 +04:00
/*
* Increase / decrease the disabled counter if we are
* about to flip a bit in the cpumask :
*/
2009-01-01 02:42:22 +03:00
if ( cpumask_test_cpu ( cpu , tracing_cpumask ) & &
! cpumask_test_cpu ( cpu , tracing_cpumask_new ) ) {
2008-05-12 23:20:52 +04:00
atomic_inc ( & global_trace . data [ cpu ] - > disabled ) ;
}
2009-01-01 02:42:22 +03:00
if ( ! cpumask_test_cpu ( cpu , tracing_cpumask ) & &
cpumask_test_cpu ( cpu , tracing_cpumask_new ) ) {
2008-05-12 23:20:52 +04:00
atomic_dec ( & global_trace . data [ cpu ] - > disabled ) ;
}
}
2008-05-12 23:20:55 +04:00
__raw_spin_unlock ( & ftrace_max_lock ) ;
2008-12-02 23:34:05 +03:00
local_irq_enable ( ) ;
2008-05-12 23:20:52 +04:00
2009-01-01 02:42:22 +03:00
cpumask_copy ( tracing_cpumask , tracing_cpumask_new ) ;
2008-05-12 23:20:52 +04:00
mutex_unlock ( & tracing_cpumask_update_lock ) ;
2009-01-01 02:42:22 +03:00
free_cpumask_var ( tracing_cpumask_new ) ;
2008-05-12 23:20:52 +04:00
return count ;
2008-05-12 23:20:52 +04:00
err_unlock :
mutex_unlock ( & tracing_cpumask_update_lock ) ;
2009-01-01 02:42:22 +03:00
free_cpumask_var ( tracing_cpumask ) ;
2008-05-12 23:20:52 +04:00
return err ;
2008-05-12 23:20:52 +04:00
}
static struct file_operations tracing_cpumask_fops = {
. open = tracing_open_generic ,
. read = tracing_cpumask_read ,
. write = tracing_cpumask_write ,
2008-05-12 23:20:42 +04:00
} ;
static ssize_t
2008-11-13 01:52:37 +03:00
tracing_trace_options_read ( struct file * filp , char __user * ubuf ,
2008-05-12 23:20:42 +04:00
size_t cnt , loff_t * ppos )
{
2008-11-17 21:23:42 +03:00
int i ;
2008-05-12 23:20:42 +04:00
char * buf ;
int r = 0 ;
int len = 0 ;
2008-11-17 21:23:42 +03:00
u32 tracer_flags = current_trace - > flags - > val ;
struct tracer_opt * trace_opts = current_trace - > flags - > opts ;
2008-05-12 23:20:42 +04:00
/* calulate max size */
for ( i = 0 ; trace_options [ i ] ; i + + ) {
len + = strlen ( trace_options [ i ] ) ;
len + = 3 ; /* "no" and space */
}
2008-11-17 21:23:42 +03:00
/*
* Increase the size with names of options specific
* of the current tracer .
*/
for ( i = 0 ; trace_opts [ i ] . name ; i + + ) {
len + = strlen ( trace_opts [ i ] . name ) ;
len + = 3 ; /* "no" and space */
}
2008-05-12 23:20:42 +04:00
/* +2 for \n and \0 */
buf = kmalloc ( len + 2 , GFP_KERNEL ) ;
if ( ! buf )
return - ENOMEM ;
for ( i = 0 ; trace_options [ i ] ; i + + ) {
if ( trace_flags & ( 1 < < i ) )
r + = sprintf ( buf + r , " %s " , trace_options [ i ] ) ;
else
r + = sprintf ( buf + r , " no%s " , trace_options [ i ] ) ;
}
2008-11-17 21:23:42 +03:00
for ( i = 0 ; trace_opts [ i ] . name ; i + + ) {
if ( tracer_flags & trace_opts [ i ] . bit )
r + = sprintf ( buf + r , " %s " ,
trace_opts [ i ] . name ) ;
else
r + = sprintf ( buf + r , " no%s " ,
trace_opts [ i ] . name ) ;
}
2008-05-12 23:20:42 +04:00
r + = sprintf ( buf + r , " \n " ) ;
WARN_ON ( r > = len + 2 ) ;
2008-05-12 23:20:52 +04:00
r = simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
kfree ( buf ) ;
return r ;
}
2008-11-17 21:23:42 +03:00
/* Try to assign a tracer specific option */
static int set_tracer_option ( struct tracer * trace , char * cmp , int neg )
{
struct tracer_flags * trace_flags = trace - > flags ;
struct tracer_opt * opts = NULL ;
int ret = 0 , i = 0 ;
int len ;
for ( i = 0 ; trace_flags - > opts [ i ] . name ; i + + ) {
opts = & trace_flags - > opts [ i ] ;
len = strlen ( opts - > name ) ;
if ( strncmp ( cmp , opts - > name , len ) = = 0 ) {
ret = trace - > set_flag ( trace_flags - > val ,
opts - > bit , ! neg ) ;
break ;
}
}
/* Not found */
if ( ! trace_flags - > opts [ i ] . name )
return - EINVAL ;
/* Refused to handle */
if ( ret )
return ret ;
if ( neg )
trace_flags - > val & = ~ opts - > bit ;
else
trace_flags - > val | = opts - > bit ;
return 0 ;
}
2008-05-12 23:20:42 +04:00
static ssize_t
2008-11-13 01:52:37 +03:00
tracing_trace_options_write ( struct file * filp , const char __user * ubuf ,
2008-05-12 23:20:42 +04:00
size_t cnt , loff_t * ppos )
{
char buf [ 64 ] ;
char * cmp = buf ;
int neg = 0 ;
2008-11-17 21:23:42 +03:00
int ret ;
2008-05-12 23:20:42 +04:00
int i ;
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
if ( strncmp ( buf , " no " , 2 ) = = 0 ) {
neg = 1 ;
cmp + = 2 ;
}
for ( i = 0 ; trace_options [ i ] ; i + + ) {
int len = strlen ( trace_options [ i ] ) ;
if ( strncmp ( cmp , trace_options [ i ] , len ) = = 0 ) {
if ( neg )
trace_flags & = ~ ( 1 < < i ) ;
else
trace_flags | = ( 1 < < i ) ;
break ;
}
}
2008-11-17 21:23:42 +03:00
/* If no option could be set, test the specific tracer options */
if ( ! trace_options [ i ] ) {
ret = set_tracer_option ( current_trace , cmp , neg ) ;
if ( ret )
return ret ;
}
2008-05-12 23:20:42 +04:00
filp - > f_pos + = cnt ;
return cnt ;
}
static struct file_operations tracing_iter_fops = {
2008-05-12 23:20:52 +04:00
. open = tracing_open_generic ,
2008-11-13 01:52:37 +03:00
. read = tracing_trace_options_read ,
. write = tracing_trace_options_write ,
2008-05-12 23:20:42 +04:00
} ;
2008-05-12 23:20:45 +04:00
static const char readme_msg [ ] =
" tracing mini-HOWTO: \n \n "
" # mkdir /debug \n "
" # mount -t debugfs nodev /debug \n \n "
" # cat /debug/tracing/available_tracers \n "
" wakeup preemptirqsoff preemptoff irqsoff ftrace sched_switch none \n \n "
" # cat /debug/tracing/current_tracer \n "
" none \n "
" # echo sched_switch > /debug/tracing/current_tracer \n "
" # cat /debug/tracing/current_tracer \n "
" sched_switch \n "
2008-11-13 01:52:37 +03:00
" # cat /debug/tracing/trace_options \n "
2008-05-12 23:20:45 +04:00
" noprint-parent nosym-offset nosym-addr noverbose \n "
2008-11-13 01:52:37 +03:00
" # echo print-parent > /debug/tracing/trace_options \n "
2008-05-12 23:20:45 +04:00
" # echo 1 > /debug/tracing/tracing_enabled \n "
" # cat /debug/tracing/trace > /tmp/trace.txt \n "
" echo 0 > /debug/tracing/tracing_enabled \n "
;
static ssize_t
tracing_readme_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
return simple_read_from_buffer ( ubuf , cnt , ppos ,
readme_msg , strlen ( readme_msg ) ) ;
}
static struct file_operations tracing_readme_fops = {
2008-05-12 23:20:52 +04:00
. open = tracing_open_generic ,
. read = tracing_readme_read ,
2008-05-12 23:20:45 +04:00
} ;
2008-05-12 23:20:42 +04:00
static ssize_t
tracing_ctrl_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ 64 ] ;
int r ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
r = sprintf ( buf , " %u \n " , tracer_enabled ) ;
2008-05-12 23:20:45 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
}
static ssize_t
tracing_ctrl_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
struct trace_array * tr = filp - > private_data ;
char buf [ 64 ] ;
2008-05-12 23:21:00 +04:00
long val ;
int ret ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
2008-05-12 23:21:00 +04:00
ret = strict_strtoul ( buf , 10 , & val ) ;
if ( ret < 0 )
return ret ;
2008-05-12 23:20:42 +04:00
val = ! ! val ;
mutex_lock ( & trace_types_lock ) ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
if ( tracer_enabled ^ val ) {
if ( val ) {
2008-05-12 23:20:42 +04:00
tracer_enabled = 1 ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
if ( current_trace - > start )
current_trace - > start ( tr ) ;
tracing_start ( ) ;
} else {
2008-05-12 23:20:42 +04:00
tracer_enabled = 0 ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
tracing_stop ( ) ;
if ( current_trace - > stop )
current_trace - > stop ( tr ) ;
}
2008-05-12 23:20:42 +04:00
}
mutex_unlock ( & trace_types_lock ) ;
filp - > f_pos + = cnt ;
return cnt ;
}
static ssize_t
tracing_set_trace_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ max_tracer_type_len + 2 ] ;
int r ;
mutex_lock ( & trace_types_lock ) ;
if ( current_trace )
r = sprintf ( buf , " %s \n " , current_trace - > name ) ;
else
r = sprintf ( buf , " \n " ) ;
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:20:46 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
}
2009-02-05 23:02:00 +03:00
int tracer_init ( struct tracer * t , struct trace_array * tr )
{
tracing_reset_online_cpus ( tr ) ;
return t - > init ( tr ) ;
}
2009-02-03 05:38:32 +03:00
static int tracing_set_tracer ( const char * buf )
2008-05-12 23:20:42 +04:00
{
struct trace_array * tr = & global_trace ;
struct tracer * t ;
2008-11-01 21:57:37 +03:00
int ret = 0 ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
for ( t = trace_types ; t ; t = t - > next ) {
if ( strcmp ( t - > name , buf ) = = 0 )
break ;
}
2008-10-05 00:04:44 +04:00
if ( ! t ) {
ret = - EINVAL ;
goto out ;
}
if ( t = = current_trace )
2008-05-12 23:20:42 +04:00
goto out ;
2008-11-12 23:24:24 +03:00
trace_branch_disable ( ) ;
2008-05-12 23:20:42 +04:00
if ( current_trace & & current_trace - > reset )
current_trace - > reset ( tr ) ;
current_trace = t ;
2008-11-16 07:57:26 +03:00
if ( t - > init ) {
2009-02-05 23:02:00 +03:00
ret = tracer_init ( t , tr ) ;
2008-11-16 07:57:26 +03:00
if ( ret )
goto out ;
}
2008-05-12 23:20:42 +04:00
2008-11-12 23:24:24 +03:00
trace_branch_enable ( tr ) ;
2008-05-12 23:20:42 +04:00
out :
mutex_unlock ( & trace_types_lock ) ;
2008-11-01 21:57:37 +03:00
return ret ;
}
static ssize_t
tracing_set_trace_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ max_tracer_type_len + 1 ] ;
int i ;
size_t ret ;
2008-11-16 07:53:19 +03:00
int err ;
ret = cnt ;
2008-11-01 21:57:37 +03:00
if ( cnt > max_tracer_type_len )
cnt = max_tracer_type_len ;
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
/* strip ending whitespace. */
for ( i = cnt - 1 ; i > 0 & & isspace ( buf [ i ] ) ; i - - )
buf [ i ] = 0 ;
2008-11-16 07:53:19 +03:00
err = tracing_set_tracer ( buf ) ;
if ( err )
return err ;
2008-11-01 21:57:37 +03:00
2008-11-16 07:53:19 +03:00
filp - > f_pos + = ret ;
2008-05-12 23:20:42 +04:00
2008-10-05 00:04:44 +04:00
return ret ;
2008-05-12 23:20:42 +04:00
}
static ssize_t
tracing_max_lat_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
unsigned long * ptr = filp - > private_data ;
char buf [ 64 ] ;
int r ;
2008-05-12 23:21:00 +04:00
r = snprintf ( buf , sizeof ( buf ) , " %ld \n " ,
2008-05-12 23:20:42 +04:00
* ptr = = ( unsigned long ) - 1 ? - 1 : nsecs_to_usecs ( * ptr ) ) ;
2008-05-12 23:21:00 +04:00
if ( r > sizeof ( buf ) )
r = sizeof ( buf ) ;
2008-05-12 23:20:46 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
}
static ssize_t
tracing_max_lat_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
long * ptr = filp - > private_data ;
char buf [ 64 ] ;
2008-05-12 23:21:00 +04:00
long val ;
int ret ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
2008-05-12 23:21:00 +04:00
ret = strict_strtoul ( buf , 10 , & val ) ;
if ( ret < 0 )
return ret ;
2008-05-12 23:20:42 +04:00
* ptr = val * 1000 ;
return cnt ;
}
2008-05-12 23:20:46 +04:00
static atomic_t tracing_reader ;
static int tracing_open_pipe ( struct inode * inode , struct file * filp )
{
struct trace_iterator * iter ;
if ( tracing_disabled )
return - ENODEV ;
/* We only allow for reader of the pipe */
if ( atomic_inc_return ( & tracing_reader ) ! = 1 ) {
atomic_dec ( & tracing_reader ) ;
return - EBUSY ;
}
/* create a buffer to store the information to pass to userspace */
iter = kzalloc ( sizeof ( * iter ) , GFP_KERNEL ) ;
if ( ! iter )
return - ENOMEM ;
2009-01-01 02:42:23 +03:00
if ( ! alloc_cpumask_var ( & iter - > started , GFP_KERNEL ) ) {
kfree ( iter ) ;
return - ENOMEM ;
}
2008-05-12 23:21:01 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-11-08 06:36:02 +03:00
/* trace pipe does not show start of buffer */
2009-01-01 02:42:23 +03:00
cpumask_setall ( iter - > started ) ;
2008-11-08 06:36:02 +03:00
2008-05-12 23:20:46 +04:00
iter - > tr = & global_trace ;
2008-05-23 23:37:28 +04:00
iter - > trace = current_trace ;
2008-05-12 23:20:46 +04:00
filp - > private_data = iter ;
2008-05-12 23:21:01 +04:00
if ( iter - > trace - > pipe_open )
iter - > trace - > pipe_open ( iter ) ;
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:20:46 +04:00
return 0 ;
}
static int tracing_release_pipe ( struct inode * inode , struct file * file )
{
struct trace_iterator * iter = file - > private_data ;
2009-01-01 02:42:23 +03:00
free_cpumask_var ( iter - > started ) ;
2008-05-12 23:20:46 +04:00
kfree ( iter ) ;
atomic_dec ( & tracing_reader ) ;
return 0 ;
}
2008-05-12 23:20:49 +04:00
static unsigned int
tracing_poll_pipe ( struct file * filp , poll_table * poll_table )
{
struct trace_iterator * iter = filp - > private_data ;
if ( trace_flags & TRACE_ITER_BLOCK ) {
/*
* Always select as readable when in blocking mode
*/
return POLLIN | POLLRDNORM ;
2008-05-12 23:21:00 +04:00
} else {
2008-05-12 23:20:49 +04:00
if ( ! trace_empty ( iter ) )
return POLLIN | POLLRDNORM ;
poll_wait ( filp , & trace_wait , poll_table ) ;
if ( ! trace_empty ( iter ) )
return POLLIN | POLLRDNORM ;
return 0 ;
}
}
2008-05-12 23:20:46 +04:00
/*
* Consumer reader .
*/
static ssize_t
tracing_read_pipe ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
struct trace_iterator * iter = filp - > private_data ;
2008-05-12 23:21:02 +04:00
ssize_t sret ;
2008-05-12 23:20:46 +04:00
/* return any leftover data */
2008-05-12 23:21:02 +04:00
sret = trace_seq_to_user ( & iter - > seq , ubuf , cnt ) ;
if ( sret ! = - EBUSY )
return sret ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:02 +04:00
trace_seq_reset ( & iter - > seq ) ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:01 +04:00
mutex_lock ( & trace_types_lock ) ;
if ( iter - > trace - > read ) {
2008-05-12 23:21:02 +04:00
sret = iter - > trace - > read ( iter , filp , ubuf , cnt , ppos ) ;
if ( sret )
2008-05-12 23:21:01 +04:00
goto out ;
}
2008-09-29 22:23:48 +04:00
waitagain :
sret = 0 ;
2008-05-12 23:20:46 +04:00
while ( trace_empty ( iter ) ) {
2008-05-12 23:20:58 +04:00
2008-05-12 23:21:01 +04:00
if ( ( filp - > f_flags & O_NONBLOCK ) ) {
2008-05-12 23:21:02 +04:00
sret = - EAGAIN ;
2008-05-12 23:21:01 +04:00
goto out ;
}
2008-05-12 23:20:58 +04:00
2008-05-12 23:20:46 +04:00
/*
* This is a make - shift waitqueue . The reason we don ' t use
* an actual wait queue is because :
* 1 ) we only ever have one waiter
* 2 ) the tracing , traces all functions , we don ' t want
* the overhead of calling wake_up and friends
* ( and tracing them too )
* Anyway , this is really very primitive wakeup .
*/
set_current_state ( TASK_INTERRUPTIBLE ) ;
iter - > tr - > waiter = current ;
2008-05-12 23:21:01 +04:00
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:21:02 +04:00
/* sleep for 100 msecs, and try again. */
schedule_timeout ( HZ / 10 ) ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:01 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-05-12 23:20:46 +04:00
iter - > tr - > waiter = NULL ;
2008-05-12 23:21:01 +04:00
if ( signal_pending ( current ) ) {
2008-05-12 23:21:02 +04:00
sret = - EINTR ;
2008-05-12 23:21:01 +04:00
goto out ;
}
2008-05-12 23:20:46 +04:00
2008-05-12 23:20:58 +04:00
if ( iter - > trace ! = current_trace )
2008-05-12 23:21:01 +04:00
goto out ;
2008-05-12 23:20:58 +04:00
2008-05-12 23:20:46 +04:00
/*
* We block until we read something and tracing is disabled .
* We still block if tracing is disabled , but we have never
* read anything . This allows a user to cat this file , and
* then enable tracing . But after we have read something ,
* we give an EOF when tracing is again disabled .
*
* iter - > pos will be 0 if we haven ' t read anything .
*/
if ( ! tracer_enabled & & iter - > pos )
break ;
continue ;
}
/* stop when tracing is finished */
if ( trace_empty ( iter ) )
2008-05-12 23:21:01 +04:00
goto out ;
2008-05-12 23:20:46 +04:00
if ( cnt > = PAGE_SIZE )
cnt = PAGE_SIZE - 1 ;
2008-05-12 23:21:01 +04:00
/* reset all but tr, trace, and overruns */
memset ( & iter - > seq , 0 ,
sizeof ( struct trace_iterator ) -
offsetof ( struct trace_iterator , seq ) ) ;
2008-05-12 23:21:01 +04:00
iter - > pos = - 1 ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:20:48 +04:00
while ( find_next_entry_inc ( iter ) ! = NULL ) {
2008-09-29 22:18:34 +04:00
enum print_line_t ret ;
2008-05-12 23:20:48 +04:00
int len = iter - > seq . len ;
2008-05-12 23:20:47 +04:00
ret = print_trace_line ( iter ) ;
2008-09-29 22:18:34 +04:00
if ( ret = = TRACE_TYPE_PARTIAL_LINE ) {
2008-05-12 23:20:48 +04:00
/* don't print partial lines */
iter - > seq . len = len ;
2008-05-12 23:20:46 +04:00
break ;
2008-05-12 23:20:48 +04:00
}
2009-02-06 20:30:44 +03:00
if ( ret ! = TRACE_TYPE_NO_CONSUME )
trace_consume ( iter ) ;
2008-05-12 23:20:46 +04:00
if ( iter - > seq . len > = cnt )
break ;
}
/* Now copy what we have to the user */
2008-05-12 23:21:02 +04:00
sret = trace_seq_to_user ( & iter - > seq , ubuf , cnt ) ;
if ( iter - > seq . readpos > = iter - > seq . len )
2008-05-12 23:20:46 +04:00
trace_seq_reset ( & iter - > seq ) ;
2008-09-29 22:23:48 +04:00
/*
* If there was nothing to send to user , inspite of consuming trace
* entries , go back to wait for more entries .
*/
2008-05-12 23:21:02 +04:00
if ( sret = = - EBUSY )
2008-09-29 22:23:48 +04:00
goto waitagain ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:01 +04:00
out :
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:21:02 +04:00
return sret ;
2008-05-12 23:20:46 +04:00
}
2009-02-09 09:15:56 +03:00
static void tracing_pipe_buf_release ( struct pipe_inode_info * pipe ,
struct pipe_buffer * buf )
{
__free_page ( buf - > page ) ;
}
static void tracing_spd_release_pipe ( struct splice_pipe_desc * spd ,
unsigned int idx )
{
__free_page ( spd - > pages [ idx ] ) ;
}
static struct pipe_buf_operations tracing_pipe_buf_ops = {
. can_merge = 0 ,
. map = generic_pipe_buf_map ,
. unmap = generic_pipe_buf_unmap ,
. confirm = generic_pipe_buf_confirm ,
. release = tracing_pipe_buf_release ,
. steal = generic_pipe_buf_steal ,
. get = generic_pipe_buf_get ,
} ;
static ssize_t tracing_splice_read_pipe ( struct file * filp ,
loff_t * ppos ,
struct pipe_inode_info * pipe ,
size_t len ,
unsigned int flags )
{
struct page * pages [ PIPE_BUFFERS ] ;
struct partial_page partial [ PIPE_BUFFERS ] ;
struct trace_iterator * iter = filp - > private_data ;
struct splice_pipe_desc spd = {
. pages = pages ,
. partial = partial ,
. nr_pages = 0 , /* This gets updated below. */
. flags = flags ,
. ops = & tracing_pipe_buf_ops ,
. spd_release = tracing_spd_release_pipe ,
} ;
ssize_t ret ;
size_t count , rem ;
unsigned int i ;
mutex_lock ( & trace_types_lock ) ;
if ( iter - > trace - > splice_read ) {
ret = iter - > trace - > splice_read ( iter , filp ,
ppos , pipe , len , flags ) ;
if ( ret )
goto out ;
}
ret = tracing_wait_pipe ( filp ) ;
if ( ret < = 0 )
goto out ;
if ( ! iter - > ent & & ! find_next_entry_inc ( iter ) ) {
ret = - EFAULT ;
goto out ;
}
/* Fill as many pages as possible. */
for ( i = 0 , rem = len ; i < PIPE_BUFFERS & & rem ; i + + ) {
pages [ i ] = alloc_page ( GFP_KERNEL ) ;
/* Seq buffer is page-sized, exactly what we need. */
for ( ; ; ) {
count = iter - > seq . len ;
ret = print_trace_line ( iter ) ;
count = iter - > seq . len - count ;
if ( rem < count ) {
rem = 0 ;
iter - > seq . len - = count ;
break ;
}
if ( ret = = TRACE_TYPE_PARTIAL_LINE ) {
iter - > seq . len - = count ;
break ;
}
trace_consume ( iter ) ;
rem - = count ;
if ( ! find_next_entry_inc ( iter ) ) {
rem = 0 ;
iter - > ent = NULL ;
break ;
}
}
/* Copy the data into the page, so we can start over. */
ret = trace_seq_to_buffer ( & iter - > seq ,
page_address ( pages [ i ] ) ,
iter - > seq . len ) ;
if ( ret < 0 ) {
__free_page ( pages [ i ] ) ;
break ;
}
partial [ i ] . offset = 0 ;
partial [ i ] . len = iter - > seq . len ;
trace_seq_reset ( & iter - > seq ) ;
}
mutex_unlock ( & trace_types_lock ) ;
spd . nr_pages = i ;
return splice_to_pipe ( pipe , & spd ) ;
out :
mutex_unlock ( & trace_types_lock ) ;
return ret ;
}
2008-05-12 23:20:59 +04:00
static ssize_t
tracing_entries_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
struct trace_array * tr = filp - > private_data ;
char buf [ 64 ] ;
int r ;
2008-11-13 08:09:35 +03:00
r = sprintf ( buf , " %lu \n " , tr - > entries > > 10 ) ;
2008-05-12 23:20:59 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
}
static ssize_t
tracing_entries_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
unsigned long val ;
char buf [ 64 ] ;
2008-11-11 05:46:00 +03:00
int ret , cpu ;
2008-05-12 23:20:59 +04:00
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:59 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
2008-05-12 23:21:00 +04:00
ret = strict_strtoul ( buf , 10 , & val ) ;
if ( ret < 0 )
return ret ;
2008-05-12 23:20:59 +04:00
/* must have at least 1 entry */
if ( ! val )
return - EINVAL ;
mutex_lock ( & trace_types_lock ) ;
2008-11-08 06:36:02 +03:00
tracing_stop ( ) ;
2008-05-12 23:20:59 +04:00
2008-11-11 05:46:00 +03:00
/* disable all cpu buffers */
for_each_tracing_cpu ( cpu ) {
if ( global_trace . data [ cpu ] )
atomic_inc ( & global_trace . data [ cpu ] - > disabled ) ;
if ( max_tr . data [ cpu ] )
atomic_inc ( & max_tr . data [ cpu ] - > disabled ) ;
}
2008-11-13 08:09:35 +03:00
/* value is in KB */
val < < = 10 ;
2008-09-30 07:02:41 +04:00
if ( val ! = global_trace . entries ) {
ret = ring_buffer_resize ( global_trace . buffer , val ) ;
if ( ret < 0 ) {
cnt = ret ;
2008-05-12 23:21:04 +04:00
goto out ;
}
2008-09-30 07:02:41 +04:00
ret = ring_buffer_resize ( max_tr . buffer , val ) ;
if ( ret < 0 ) {
int r ;
cnt = ret ;
r = ring_buffer_resize ( global_trace . buffer ,
global_trace . entries ) ;
if ( r < 0 ) {
/* AARGH! We are left with different
* size max buffer ! ! ! ! */
WARN_ON ( 1 ) ;
tracing_disabled = 1 ;
2008-05-12 23:20:59 +04:00
}
2008-09-30 07:02:41 +04:00
goto out ;
2008-05-12 23:20:59 +04:00
}
2008-05-12 23:21:04 +04:00
2008-09-30 07:02:41 +04:00
global_trace . entries = val ;
2008-05-12 23:20:59 +04:00
}
filp - > f_pos + = cnt ;
2008-05-22 08:22:16 +04:00
/* If check pages failed, return ENOMEM */
if ( tracing_disabled )
cnt = - ENOMEM ;
2008-05-12 23:20:59 +04:00
out :
2008-11-11 05:46:00 +03:00
for_each_tracing_cpu ( cpu ) {
if ( global_trace . data [ cpu ] )
atomic_dec ( & global_trace . data [ cpu ] - > disabled ) ;
if ( max_tr . data [ cpu ] )
atomic_dec ( & max_tr . data [ cpu ] - > disabled ) ;
}
2008-11-08 06:36:02 +03:00
tracing_start ( ) ;
2008-05-12 23:20:59 +04:00
max_tr . entries = global_trace . entries ;
mutex_unlock ( & trace_types_lock ) ;
return cnt ;
}
2008-09-16 23:06:42 +04:00
static int mark_printk ( const char * fmt , . . . )
{
int ret ;
va_list args ;
va_start ( args , fmt ) ;
2008-12-04 01:45:11 +03:00
ret = trace_vprintk ( 0 , - 1 , fmt , args ) ;
2008-09-16 23:06:42 +04:00
va_end ( args ) ;
return ret ;
}
static ssize_t
tracing_mark_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * fpos )
{
char * buf ;
char * end ;
2008-11-08 06:36:02 +03:00
if ( tracing_disabled )
2008-09-16 23:06:42 +04:00
return - EINVAL ;
if ( cnt > TRACE_BUF_SIZE )
cnt = TRACE_BUF_SIZE ;
buf = kmalloc ( cnt + 1 , GFP_KERNEL ) ;
if ( buf = = NULL )
return - ENOMEM ;
if ( copy_from_user ( buf , ubuf , cnt ) ) {
kfree ( buf ) ;
return - EFAULT ;
}
/* Cut from the first nil or newline. */
buf [ cnt ] = ' \0 ' ;
end = strchr ( buf , ' \n ' ) ;
if ( end )
* end = ' \0 ' ;
cnt = mark_printk ( " %s \n " , buf ) ;
kfree ( buf ) ;
* fpos + = cnt ;
return cnt ;
}
2008-05-12 23:20:42 +04:00
static struct file_operations tracing_max_lat_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
. read = tracing_max_lat_read ,
. write = tracing_max_lat_write ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations tracing_ctrl_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
. read = tracing_ctrl_read ,
. write = tracing_ctrl_write ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations set_tracer_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
. read = tracing_set_trace_read ,
. write = tracing_set_trace_write ,
2008-05-12 23:20:42 +04:00
} ;
2008-05-12 23:20:46 +04:00
static struct file_operations tracing_pipe_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_pipe ,
2008-05-12 23:20:49 +04:00
. poll = tracing_poll_pipe ,
2008-05-12 23:20:46 +04:00
. read = tracing_read_pipe ,
2009-02-09 09:15:56 +03:00
. splice_read = tracing_splice_read_pipe ,
2008-05-12 23:20:46 +04:00
. release = tracing_release_pipe ,
2008-05-12 23:20:46 +04:00
} ;
2008-05-12 23:20:59 +04:00
static struct file_operations tracing_entries_fops = {
. open = tracing_open_generic ,
. read = tracing_entries_read ,
. write = tracing_entries_write ,
} ;
2008-09-16 23:06:42 +04:00
static struct file_operations tracing_mark_fops = {
2008-09-21 22:16:30 +04:00
. open = tracing_open_generic ,
2008-09-16 23:06:42 +04:00
. write = tracing_mark_write ,
} ;
2008-05-12 23:20:42 +04:00
# ifdef CONFIG_DYNAMIC_FTRACE
2008-10-30 23:08:33 +03:00
int __weak ftrace_arch_read_dyn_info ( char * buf , int size )
{
return 0 ;
}
2008-05-12 23:20:42 +04:00
static ssize_t
2008-10-30 23:08:33 +03:00
tracing_read_dyn_info ( struct file * filp , char __user * ubuf ,
2008-05-12 23:20:42 +04:00
size_t cnt , loff_t * ppos )
{
2008-10-31 07:03:22 +03:00
static char ftrace_dyn_info_buffer [ 1024 ] ;
static DEFINE_MUTEX ( dyn_info_mutex ) ;
2008-05-12 23:20:42 +04:00
unsigned long * p = filp - > private_data ;
2008-10-30 23:08:33 +03:00
char * buf = ftrace_dyn_info_buffer ;
2008-10-31 07:03:22 +03:00
int size = ARRAY_SIZE ( ftrace_dyn_info_buffer ) ;
2008-05-12 23:20:42 +04:00
int r ;
2008-10-30 23:08:33 +03:00
mutex_lock ( & dyn_info_mutex ) ;
r = sprintf ( buf , " %ld " , * p ) ;
2008-05-12 23:20:46 +04:00
2008-10-31 07:03:22 +03:00
r + = ftrace_arch_read_dyn_info ( buf + r , ( size - 1 ) - r ) ;
2008-10-30 23:08:33 +03:00
buf [ r + + ] = ' \n ' ;
r = simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
mutex_unlock ( & dyn_info_mutex ) ;
return r ;
2008-05-12 23:20:42 +04:00
}
2008-10-30 23:08:33 +03:00
static struct file_operations tracing_dyn_info_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
2008-10-30 23:08:33 +03:00
. read = tracing_read_dyn_info ,
2008-05-12 23:20:42 +04:00
} ;
# endif
static struct dentry * d_tracer ;
struct dentry * tracing_init_dentry ( void )
{
static int once ;
if ( d_tracer )
return d_tracer ;
d_tracer = debugfs_create_dir ( " tracing " , NULL ) ;
if ( ! d_tracer & & ! once ) {
once = 1 ;
pr_warning ( " Could not create debugfs directory 'tracing' \n " ) ;
return NULL ;
}
return d_tracer ;
}
2008-05-12 23:20:44 +04:00
# ifdef CONFIG_FTRACE_SELFTEST
/* Let selftest have access to static functions in this file */
# include "trace_selftest.c"
# endif
2008-09-23 14:34:32 +04:00
static __init int tracer_init_debugfs ( void )
2008-05-12 23:20:42 +04:00
{
struct dentry * d_tracer ;
struct dentry * entry ;
d_tracer = tracing_init_dentry ( ) ;
entry = debugfs_create_file ( " tracing_enabled " , 0644 , d_tracer ,
& global_trace , & tracing_ctrl_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'tracing_enabled' entry \n " ) ;
2008-11-13 01:52:37 +03:00
entry = debugfs_create_file ( " trace_options " , 0644 , d_tracer ,
2008-05-12 23:20:42 +04:00
NULL , & tracing_iter_fops ) ;
if ( ! entry )
2008-11-13 01:52:37 +03:00
pr_warning ( " Could not create debugfs 'trace_options' entry \n " ) ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:52 +04:00
entry = debugfs_create_file ( " tracing_cpumask " , 0644 , d_tracer ,
NULL , & tracing_cpumask_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'tracing_cpumask' entry \n " ) ;
2008-05-12 23:20:42 +04:00
entry = debugfs_create_file ( " latency_trace " , 0444 , d_tracer ,
& global_trace , & tracing_lt_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'latency_trace' entry \n " ) ;
entry = debugfs_create_file ( " trace " , 0444 , d_tracer ,
& global_trace , & tracing_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'trace' entry \n " ) ;
entry = debugfs_create_file ( " available_tracers " , 0444 , d_tracer ,
& global_trace , & show_traces_fops ) ;
if ( ! entry )
2008-08-15 23:08:22 +04:00
pr_warning ( " Could not create debugfs 'available_tracers' entry \n " ) ;
2008-05-12 23:20:42 +04:00
entry = debugfs_create_file ( " current_tracer " , 0444 , d_tracer ,
& global_trace , & set_tracer_fops ) ;
if ( ! entry )
2008-08-15 23:08:22 +04:00
pr_warning ( " Could not create debugfs 'current_tracer' entry \n " ) ;
2008-05-12 23:20:42 +04:00
entry = debugfs_create_file ( " tracing_max_latency " , 0644 , d_tracer ,
& tracing_max_latency ,
& tracing_max_lat_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
" 'tracing_max_latency' entry \n " ) ;
entry = debugfs_create_file ( " tracing_thresh " , 0644 , d_tracer ,
& tracing_thresh , & tracing_max_lat_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
2008-08-15 23:08:22 +04:00
" 'tracing_thresh' entry \n " ) ;
2008-05-12 23:20:45 +04:00
entry = debugfs_create_file ( " README " , 0644 , d_tracer ,
NULL , & tracing_readme_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'README' entry \n " ) ;
2008-05-12 23:20:46 +04:00
entry = debugfs_create_file ( " trace_pipe " , 0644 , d_tracer ,
NULL , & tracing_pipe_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
2008-08-15 23:08:22 +04:00
" 'trace_pipe' entry \n " ) ;
2008-05-12 23:20:42 +04:00
2008-11-13 01:52:36 +03:00
entry = debugfs_create_file ( " buffer_size_kb " , 0644 , d_tracer ,
2008-05-12 23:20:59 +04:00
& global_trace , & tracing_entries_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
2008-11-13 01:52:36 +03:00
" 'buffer_size_kb' entry \n " ) ;
2008-05-12 23:20:59 +04:00
2008-09-16 23:06:42 +04:00
entry = debugfs_create_file ( " trace_marker " , 0220 , d_tracer ,
NULL , & tracing_mark_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
" 'trace_marker' entry \n " ) ;
2008-05-12 23:20:42 +04:00
# ifdef CONFIG_DYNAMIC_FTRACE
entry = debugfs_create_file ( " dyn_ftrace_total_info " , 0444 , d_tracer ,
& ftrace_update_tot_cnt ,
2008-10-30 23:08:33 +03:00
& tracing_dyn_info_fops ) ;
2008-05-12 23:20:42 +04:00
if ( ! entry )
pr_warning ( " Could not create debugfs "
" 'dyn_ftrace_total_info' entry \n " ) ;
# endif
2008-05-12 23:20:49 +04:00
# ifdef CONFIG_SYSPROF_TRACER
init_tracer_sysprof_debugfs ( d_tracer ) ;
# endif
2008-09-23 14:34:32 +04:00
return 0 ;
2008-05-12 23:20:42 +04:00
}
2008-12-04 01:45:11 +03:00
int trace_vprintk ( unsigned long ip , int depth , const char * fmt , va_list args )
2008-08-01 20:26:41 +04:00
{
static DEFINE_SPINLOCK ( trace_buf_lock ) ;
static char trace_buf [ TRACE_BUF_SIZE ] ;
2008-09-04 12:24:14 +04:00
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-04 12:24:14 +04:00
struct trace_array * tr = & global_trace ;
2008-08-01 20:26:41 +04:00
struct trace_array_cpu * data ;
2008-10-01 21:14:09 +04:00
int cpu , len = 0 , size , pc ;
2008-12-08 18:55:53 +03:00
struct print_entry * entry ;
unsigned long irq_flags ;
2008-08-01 20:26:41 +04:00
2008-12-06 05:41:33 +03:00
if ( tracing_disabled | | tracing_selftest_running )
2008-08-01 20:26:41 +04:00
return 0 ;
2008-10-01 21:14:09 +04:00
pc = preempt_count ( ) ;
preempt_disable_notrace ( ) ;
2008-08-01 20:26:41 +04:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
2008-10-04 10:01:00 +04:00
if ( unlikely ( atomic_read ( & data - > disabled ) ) )
2008-08-01 20:26:41 +04:00
goto out ;
2008-12-06 05:43:41 +03:00
pause_graph_tracing ( ) ;
spin_lock_irqsave ( & trace_buf_lock , irq_flags ) ;
2008-09-16 22:58:24 +04:00
len = vsnprintf ( trace_buf , TRACE_BUF_SIZE , fmt , args ) ;
2008-08-01 20:26:41 +04:00
len = min ( len , TRACE_BUF_SIZE - 1 ) ;
trace_buf [ len ] = 0 ;
2008-09-30 07:02:42 +04:00
size = sizeof ( * entry ) + len + 1 ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
event = trace_buffer_lock_reserve ( tr , TRACE_PRINT , size , irq_flags , pc ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
goto out_unlock ;
2008-09-30 07:02:42 +04:00
entry = ring_buffer_event_data ( event ) ;
entry - > ip = ip ;
2008-12-04 01:45:11 +03:00
entry - > depth = depth ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:42 +04:00
memcpy ( & entry - > buf , trace_buf , len ) ;
entry - > buf [ len ] = 0 ;
2009-02-05 21:12:56 +03:00
ring_buffer_unlock_commit ( tr - > buffer , event ) ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
out_unlock :
2008-12-06 05:43:41 +03:00
spin_unlock_irqrestore ( & trace_buf_lock , irq_flags ) ;
unpause_graph_tracing ( ) ;
2008-08-01 20:26:41 +04:00
out :
2008-10-01 21:14:09 +04:00
preempt_enable_notrace ( ) ;
2008-08-01 20:26:41 +04:00
return len ;
}
2008-09-16 22:58:24 +04:00
EXPORT_SYMBOL_GPL ( trace_vprintk ) ;
int __ftrace_printk ( unsigned long ip , const char * fmt , . . . )
{
int ret ;
va_list ap ;
if ( ! ( trace_flags & TRACE_ITER_PRINTK ) )
return 0 ;
va_start ( ap , fmt ) ;
2008-12-05 01:51:23 +03:00
ret = trace_vprintk ( ip , task_curr_ret_stack ( current ) , fmt , ap ) ;
2008-09-16 22:58:24 +04:00
va_end ( ap ) ;
return ret ;
}
2008-08-01 20:26:41 +04:00
EXPORT_SYMBOL_GPL ( __ftrace_printk ) ;
2009-01-23 17:06:23 +03:00
int __ftrace_vprintk ( unsigned long ip , const char * fmt , va_list ap )
{
if ( ! ( trace_flags & TRACE_ITER_PRINTK ) )
return 0 ;
return trace_vprintk ( ip , task_curr_ret_stack ( current ) , fmt , ap ) ;
}
EXPORT_SYMBOL_GPL ( __ftrace_vprintk ) ;
2008-07-31 06:36:46 +04:00
static int trace_panic_handler ( struct notifier_block * this ,
unsigned long event , void * unused )
{
2008-10-24 03:26:08 +04:00
if ( ftrace_dump_on_oops )
ftrace_dump ( ) ;
2008-07-31 06:36:46 +04:00
return NOTIFY_OK ;
}
static struct notifier_block trace_panic_notifier = {
. notifier_call = trace_panic_handler ,
. next = NULL ,
. priority = 150 /* priority: INT_MAX >= x >= 0 */
} ;
static int trace_die_handler ( struct notifier_block * self ,
unsigned long val ,
void * data )
{
switch ( val ) {
case DIE_OOPS :
2008-10-24 03:26:08 +04:00
if ( ftrace_dump_on_oops )
ftrace_dump ( ) ;
2008-07-31 06:36:46 +04:00
break ;
default :
break ;
}
return NOTIFY_OK ;
}
static struct notifier_block trace_die_notifier = {
. notifier_call = trace_die_handler ,
. priority = 200
} ;
/*
* printk is set to max of 1024 , we really don ' t need it that big .
* Nothing should be printing 1000 characters anyway .
*/
# define TRACE_MAX_PRINT 1000
/*
* Define here KERN_TRACE so that we have one place to modify
* it if we decide to change what log level the ftrace dump
* should be at .
*/
2009-01-14 20:24:42 +03:00
# define KERN_TRACE KERN_EMERG
2008-07-31 06:36:46 +04:00
static void
trace_printk_seq ( struct trace_seq * s )
{
/* Probably should print a warning here. */
if ( s - > len > = 1000 )
s - > len = 1000 ;
/* should be zero ended, but we are paranoid. */
s - > buffer [ s - > len ] = 0 ;
printk ( KERN_TRACE " %s " , s - > buffer ) ;
trace_seq_reset ( s ) ;
}
void ftrace_dump ( void )
{
static DEFINE_SPINLOCK ( ftrace_dump_lock ) ;
/* use static because iter can be a bit big for the stack */
static struct trace_iterator iter ;
static int dump_ran ;
2008-10-01 08:29:53 +04:00
unsigned long flags ;
int cnt = 0 , cpu ;
2008-07-31 06:36:46 +04:00
/* only one dump */
spin_lock_irqsave ( & ftrace_dump_lock , flags ) ;
if ( dump_ran )
goto out ;
dump_ran = 1 ;
/* No turning back! */
2009-01-14 22:50:19 +03:00
tracing_off ( ) ;
2008-10-23 17:33:02 +04:00
ftrace_kill ( ) ;
2008-07-31 06:36:46 +04:00
2008-10-01 08:29:53 +04:00
for_each_tracing_cpu ( cpu ) {
atomic_inc ( & global_trace . data [ cpu ] - > disabled ) ;
}
2008-11-22 14:28:48 +03:00
/* don't look at user memory in panic mode */
trace_flags & = ~ TRACE_ITER_SYM_USEROBJ ;
2008-07-31 06:36:46 +04:00
printk ( KERN_TRACE " Dumping ftrace buffer: \n " ) ;
iter . tr = & global_trace ;
iter . trace = current_trace ;
/*
* We need to stop all tracing on all CPUS to read the
* the next buffer . This is a bit expensive , but is
* not done often . We fill all what we can read ,
* and then release the locks again .
*/
while ( ! trace_empty ( & iter ) ) {
if ( ! cnt )
printk ( KERN_TRACE " --------------------------------- \n " ) ;
cnt + + ;
/* reset all but tr, trace, and overruns */
memset ( & iter . seq , 0 ,
sizeof ( struct trace_iterator ) -
offsetof ( struct trace_iterator , seq ) ) ;
iter . iter_flags | = TRACE_FILE_LAT_FMT ;
iter . pos = - 1 ;
if ( find_next_entry_inc ( & iter ) ! = NULL ) {
print_trace_line ( & iter ) ;
trace_consume ( & iter ) ;
}
trace_printk_seq ( & iter . seq ) ;
}
if ( ! cnt )
printk ( KERN_TRACE " (ftrace buffer empty) \n " ) ;
else
printk ( KERN_TRACE " --------------------------------- \n " ) ;
out :
spin_unlock_irqrestore ( & ftrace_dump_lock , flags ) ;
}
2008-09-30 07:02:41 +04:00
__init static int tracer_alloc_buffers ( void )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:43 +04:00
struct trace_array_cpu * data ;
int i ;
2009-01-01 02:42:22 +03:00
int ret = - ENOMEM ;
2008-05-12 23:20:43 +04:00
2009-01-01 02:42:22 +03:00
if ( ! alloc_cpumask_var ( & tracing_buffer_mask , GFP_KERNEL ) )
goto out ;
if ( ! alloc_cpumask_var ( & tracing_cpumask , GFP_KERNEL ) )
goto out_free_buffer_mask ;
2008-05-12 23:20:43 +04:00
2009-01-01 02:42:22 +03:00
cpumask_copy ( tracing_buffer_mask , cpu_possible_mask ) ;
cpumask_copy ( tracing_cpumask , cpu_all_mask ) ;
/* TODO: make the number of buffers hot pluggable with CPUS */
2008-09-30 07:02:41 +04:00
global_trace . buffer = ring_buffer_alloc ( trace_buf_size ,
TRACE_BUFFER_FLAGS ) ;
if ( ! global_trace . buffer ) {
printk ( KERN_ERR " tracer: failed to allocate ring buffer! \n " ) ;
WARN_ON ( 1 ) ;
2009-01-01 02:42:22 +03:00
goto out_free_cpumask ;
2008-05-12 23:20:43 +04:00
}
2008-09-30 07:02:41 +04:00
global_trace . entries = ring_buffer_size ( global_trace . buffer ) ;
2008-05-12 23:20:43 +04:00
2009-01-01 02:42:22 +03:00
2008-05-12 23:20:43 +04:00
# ifdef CONFIG_TRACER_MAX_TRACE
2008-09-30 07:02:41 +04:00
max_tr . buffer = ring_buffer_alloc ( trace_buf_size ,
TRACE_BUFFER_FLAGS ) ;
if ( ! max_tr . buffer ) {
printk ( KERN_ERR " tracer: failed to allocate max ring buffer! \n " ) ;
WARN_ON ( 1 ) ;
ring_buffer_free ( global_trace . buffer ) ;
2009-01-01 02:42:22 +03:00
goto out_free_cpumask ;
2008-05-12 23:20:43 +04:00
}
2008-09-30 07:02:41 +04:00
max_tr . entries = ring_buffer_size ( max_tr . buffer ) ;
WARN_ON ( max_tr . entries ! = global_trace . entries ) ;
2008-05-12 23:20:59 +04:00
# endif
2008-05-12 23:21:00 +04:00
2008-05-12 23:20:43 +04:00
/* Allocate the first page for all buffers */
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( i ) {
2008-05-12 23:20:43 +04:00
data = global_trace . data [ i ] = & per_cpu ( global_trace_cpu , i ) ;
2008-05-12 23:20:42 +04:00
max_tr . data [ i ] = & per_cpu ( max_data , i ) ;
2008-05-12 23:20:43 +04:00
}
2008-05-12 23:20:42 +04:00
trace_init_cmdlines ( ) ;
2008-09-21 22:16:30 +04:00
register_tracer ( & nop_trace ) ;
2009-02-03 05:38:33 +03:00
current_trace = & nop_trace ;
2008-09-23 14:34:32 +04:00
# ifdef CONFIG_BOOT_TRACER
register_tracer ( & boot_tracer ) ;
# endif
2008-05-12 23:20:44 +04:00
/* All seems OK, enable tracing */
tracing_disabled = 0 ;
2008-09-30 07:02:41 +04:00
2008-07-31 06:36:46 +04:00
atomic_notifier_chain_register ( & panic_notifier_list ,
& trace_panic_notifier ) ;
register_die_notifier ( & trace_die_notifier ) ;
2009-01-01 02:42:22 +03:00
ret = 0 ;
2008-07-31 06:36:46 +04:00
2009-01-01 02:42:22 +03:00
out_free_cpumask :
free_cpumask_var ( tracing_cpumask ) ;
out_free_buffer_mask :
free_cpumask_var ( tracing_buffer_mask ) ;
out :
return ret ;
2008-05-12 23:20:42 +04:00
}
2009-02-03 05:38:32 +03:00
__init static int clear_boot_tracer ( void )
{
/*
* The default tracer at boot buffer is an init section .
* This function is called in lateinit . If we did not
* find the boot tracer , then clear it out , to prevent
* later registration from accessing the buffer that is
* about to be freed .
*/
if ( ! default_bootup_tracer )
return 0 ;
printk ( KERN_INFO " ftrace bootup tracer '%s' not registered. \n " ,
default_bootup_tracer ) ;
default_bootup_tracer = NULL ;
return 0 ;
}
2008-09-23 14:34:32 +04:00
early_initcall ( tracer_alloc_buffers ) ;
fs_initcall ( tracer_init_debugfs ) ;
2009-02-03 05:38:32 +03:00
late_initcall ( clear_boot_tracer ) ;