2008-05-12 23:20:42 +04:00
/*
* ring buffer based function tracer
*
* Copyright ( C ) 2007 - 2008 Steven Rostedt < srostedt @ redhat . com >
* Copyright ( C ) 2008 Ingo Molnar < mingo @ redhat . com >
*
* Originally taken from the RT patch by :
* Arnaldo Carvalho de Melo < acme @ redhat . com >
*
* Based on code from the latency_tracer , that is :
* Copyright ( C ) 2004 - 2006 Ingo Molnar
* Copyright ( C ) 2004 William Lee Irwin III
*/
# include <linux/utsrelease.h>
# include <linux/kallsyms.h>
# include <linux/seq_file.h>
2008-07-31 06:36:46 +04:00
# include <linux/notifier.h>
2008-05-12 23:20:42 +04:00
# include <linux/debugfs.h>
2008-05-12 23:20:43 +04:00
# include <linux/pagemap.h>
2008-05-12 23:20:42 +04:00
# include <linux/hardirq.h>
# include <linux/linkage.h>
# include <linux/uaccess.h>
# include <linux/ftrace.h>
# include <linux/module.h>
# include <linux/percpu.h>
2008-07-31 06:36:46 +04:00
# include <linux/kdebug.h>
2008-05-12 23:20:42 +04:00
# include <linux/ctype.h>
# include <linux/init.h>
2008-05-12 23:20:49 +04:00
# include <linux/poll.h>
2008-05-12 23:20:42 +04:00
# include <linux/gfp.h>
# include <linux/fs.h>
2008-05-27 22:33:18 +04:00
# include <linux/kprobes.h>
2008-05-12 23:21:04 +04:00
# include <linux/writeback.h>
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:51 +04:00
# include <linux/stacktrace.h>
2008-09-30 07:02:41 +04:00
# include <linux/ring_buffer.h>
2008-10-28 11:43:26 +03:00
# include <linux/irqflags.h>
2008-05-12 23:20:51 +04:00
2008-05-12 23:20:42 +04:00
# include "trace.h"
2008-09-30 07:02:41 +04:00
# define TRACE_BUFFER_FLAGS (RB_FL_OVERWRITE)
2008-05-12 23:20:42 +04:00
unsigned long __read_mostly tracing_max_latency = ( cycle_t ) ULONG_MAX ;
unsigned long __read_mostly tracing_thresh ;
2008-11-06 00:05:44 +03:00
/*
* Kill all tracing for good ( never come back ) .
* It is initialized to 1 but will turn to zero if the initialization
* of the tracer is successful . But that is the only place that sets
* this back to zero .
*/
int tracing_disabled = 1 ;
2008-10-01 08:29:53 +04:00
static DEFINE_PER_CPU ( local_t , ftrace_cpu_disabled ) ;
static inline void ftrace_disable_cpu ( void )
{
preempt_disable ( ) ;
local_inc ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ;
}
static inline void ftrace_enable_cpu ( void )
{
local_dec ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ;
preempt_enable ( ) ;
}
2008-05-12 23:21:00 +04:00
static cpumask_t __read_mostly tracing_buffer_mask ;
# define for_each_tracing_cpu(cpu) \
for_each_cpu_mask ( cpu , tracing_buffer_mask )
2008-10-24 03:26:08 +04:00
/*
* ftrace_dump_on_oops - variable to dump ftrace buffer on oops
*
* If there is an oops ( or kernel panic ) and the ftrace_dump_on_oops
* is set , then ftrace_dump is called . This will output the contents
* of the ftrace buffers to the console . This is very useful for
* capturing traces that lead to crashes and outputing it to a
* serial console .
*
* It is default off , but you can enable it with either specifying
* " ftrace_dump_on_oops " in the kernel command line , or setting
* / proc / sys / kernel / ftrace_dump_on_oops to true .
*/
int ftrace_dump_on_oops ;
2008-11-01 21:57:37 +03:00
static int tracing_set_tracer ( char * buf ) ;
static int __init set_ftrace ( char * str )
{
tracing_set_tracer ( str ) ;
return 1 ;
}
__setup ( " ftrace " , set_ftrace ) ;
2008-10-24 03:26:08 +04:00
static int __init set_ftrace_dump_on_oops ( char * str )
{
ftrace_dump_on_oops = 1 ;
return 1 ;
}
__setup ( " ftrace_dump_on_oops " , set_ftrace_dump_on_oops ) ;
2008-05-23 23:37:28 +04:00
long
2008-05-12 23:20:42 +04:00
ns2usecs ( cycle_t nsec )
{
nsec + = 500 ;
do_div ( nsec , 1000 ) ;
return nsec ;
}
2008-05-12 23:20:51 +04:00
cycle_t ftrace_now ( int cpu )
2008-05-12 23:20:46 +04:00
{
2008-09-30 07:02:41 +04:00
u64 ts = ring_buffer_time_stamp ( cpu ) ;
ring_buffer_normalize_time_stamp ( cpu , & ts ) ;
return ts ;
2008-05-12 23:20:46 +04:00
}
2008-05-12 23:21:00 +04:00
/*
* The global_trace is the descriptor that holds the tracing
* buffers for the live tracing . For each CPU , it contains
* a link list of pages that will store trace entries . The
* page descriptor of the pages in the memory is used to hold
* the link list by linking the lru item in the page descriptor
* to each of the pages in the buffer per CPU .
*
* For each active CPU there is a data field that holds the
* pages for the buffer for that CPU . Each CPU has the same number
* of pages allocated for its buffer .
*/
2008-05-12 23:20:42 +04:00
static struct trace_array global_trace ;
static DEFINE_PER_CPU ( struct trace_array_cpu , global_trace_cpu ) ;
2008-05-12 23:21:00 +04:00
/*
* The max_tr is used to snapshot the global_trace when a maximum
* latency is reached . Some tracers will use this to store a maximum
* trace while it continues examining live traces .
*
* The buffers for the max_tr are set up the same as the global_trace .
* When a snapshot is taken , the link list of the max_tr is swapped
* with the link list of the global_trace and the buffers are reset for
* the global_trace so the tracing can continue .
*/
2008-05-12 23:20:42 +04:00
static struct trace_array max_tr ;
static DEFINE_PER_CPU ( struct trace_array_cpu , max_data ) ;
2008-05-12 23:21:00 +04:00
/* tracer_enabled is used to toggle activation of a tracer */
2008-05-12 23:20:48 +04:00
static int tracer_enabled = 1 ;
2008-05-12 23:21:00 +04:00
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
/**
* tracing_is_enabled - return tracer_enabled status
*
* This function is used by other tracers to know the status
* of the tracer_enabled flag . Tracers may use this function
* to know if it should enable their features when starting
* up . See irqsoff tracer for an example ( start_irqsoff_tracer ) .
*/
int tracing_is_enabled ( void )
{
return tracer_enabled ;
}
2008-07-11 04:58:16 +04:00
/* function tracing enabled */
int ftrace_function_enabled ;
2008-05-12 23:21:00 +04:00
/*
2008-09-30 07:02:41 +04:00
* trace_buf_size is the size in bytes that is allocated
* for a buffer . Note , the number of bytes is always rounded
* to page size .
2008-07-31 06:36:46 +04:00
*
* This number is purposely set to a low number of 16384.
* If the dump on oops happens , it will be much appreciated
* to not have to wait for all that output . Anyway this can be
* boot time and run time configurable .
2008-05-12 23:21:00 +04:00
*/
2008-09-30 07:02:41 +04:00
# define TRACE_BUF_SIZE_DEFAULT 1441792UL /* 16384 * 88 (sizeof(entry)) */
2008-07-31 06:36:46 +04:00
2008-09-30 07:02:41 +04:00
static unsigned long trace_buf_size = TRACE_BUF_SIZE_DEFAULT ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:21:00 +04:00
/* trace_types holds a link list of available tracers. */
2008-05-12 23:20:42 +04:00
static struct tracer * trace_types __read_mostly ;
2008-05-12 23:21:00 +04:00
/* current_trace points to the tracer that is currently active */
2008-05-12 23:20:42 +04:00
static struct tracer * current_trace __read_mostly ;
2008-05-12 23:21:00 +04:00
/*
* max_tracer_type_len is used to simplify the allocating of
* buffers to read userspace tracer names . We keep track of
* the longest tracer name registered .
*/
2008-05-12 23:20:42 +04:00
static int max_tracer_type_len ;
2008-05-12 23:21:00 +04:00
/*
* trace_types_lock is used to protect the trace_types list .
* This lock is also used to keep user access serialized .
* Accesses from userspace will grab this lock while userspace
* activities happen inside the kernel .
*/
2008-05-12 23:20:42 +04:00
static DEFINE_MUTEX ( trace_types_lock ) ;
2008-05-12 23:21:00 +04:00
/* trace_wait is a waitqueue for tasks blocked on trace_poll */
2008-05-12 23:20:52 +04:00
static DECLARE_WAIT_QUEUE_HEAD ( trace_wait ) ;
2008-05-12 23:21:00 +04:00
/* trace_flags holds iter_ctrl options */
2008-11-08 06:36:02 +03:00
unsigned long trace_flags = TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK ;
2008-05-12 23:20:52 +04:00
2008-05-12 23:21:00 +04:00
/**
* trace_wake_up - wake up tasks waiting for trace input
*
* Simply wakes up any task that is blocked on the trace_wait
* queue . These is used with trace_poll for tasks polling the trace .
*/
2008-05-12 23:20:52 +04:00
void trace_wake_up ( void )
{
2008-05-12 23:20:52 +04:00
/*
* The runqueue_is_locked ( ) can fail , but this is the best we
* have for now :
*/
if ( ! ( trace_flags & TRACE_ITER_BLOCK ) & & ! runqueue_is_locked ( ) )
2008-05-12 23:20:52 +04:00
wake_up ( & trace_wait ) ;
}
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
static int __init set_buf_size ( char * str )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
unsigned long buf_size ;
2008-05-12 23:21:00 +04:00
int ret ;
2008-05-12 23:20:42 +04:00
if ( ! str )
return 0 ;
2008-09-30 07:02:41 +04:00
ret = strict_strtoul ( str , 0 , & buf_size ) ;
2008-05-12 23:21:00 +04:00
/* nr_entries can not be zero */
2008-09-30 07:02:41 +04:00
if ( ret < 0 | | buf_size = = 0 )
2008-05-12 23:21:00 +04:00
return 0 ;
2008-09-30 07:02:41 +04:00
trace_buf_size = buf_size ;
2008-05-12 23:20:42 +04:00
return 1 ;
}
2008-09-30 07:02:41 +04:00
__setup ( " trace_buf_size= " , set_buf_size ) ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:44 +04:00
unsigned long nsecs_to_usecs ( unsigned long nsecs )
{
return nsecs / 1000 ;
}
2008-05-12 23:21:00 +04:00
/* These must match the bit postions in trace_iterator_flags */
2008-05-12 23:20:42 +04:00
static const char * trace_options [ ] = {
" print-parent " ,
" sym-offset " ,
" sym-addr " ,
" verbose " ,
2008-05-12 23:20:47 +04:00
" raw " ,
2008-05-12 23:20:49 +04:00
" hex " ,
2008-05-12 23:20:47 +04:00
" bin " ,
2008-05-12 23:20:49 +04:00
" block " ,
2008-05-12 23:20:51 +04:00
" stacktrace " ,
2008-05-12 23:20:52 +04:00
" sched-tree " ,
2008-09-04 12:24:14 +04:00
" ftrace_printk " ,
ftrace: function tracer with irqs disabled
Impact: disable interrupts during trace entry creation (as opposed to preempt)
To help with performance, I set the ftracer to not disable interrupts,
and only to disable preemption. If an interrupt occurred, it would not
be traced, because the function tracer protects itself from recursion.
This may be faster, but the trace output might miss some traces.
This patch makes the fuction trace disable interrupts, but it also
adds a runtime feature to disable preemption instead. It does this by
having two different tracer functions. When the function tracer is
enabled, it will check to see which version is requested (irqs disabled
or preemption disabled). Then it will use the corresponding function
as the tracer.
Irq disabling is the default behaviour, but if the user wants better
performance, with the chance of missing traces, then they can choose
the preempt disabled version.
Running hackbench 3 times with the irqs disabled and 3 times with
the preempt disabled function tracer yielded:
tracing type times entries recorded
------------ -------- ----------------
irq disabled 43.393 166433066
43.282 166172618
43.298 166256704
preempt disabled 38.969 159871710
38.943 159972935
39.325 161056510
Average:
irqs disabled: 43.324 166287462
preempt disabled: 39.079 160300385
preempt is 10.8 percent faster than irqs disabled.
I wrote a patch to count function trace recursion and reran hackbench.
With irq disabled: 1,150 times the function tracer did not trace due to
recursion.
with preempt disabled: 5,117,718 times.
The thousand times with irq disabled could be due to NMIs, or simply a case
where it called a function that was not protected by notrace.
But we also see that a large amount of the trace is lost with the
preempt version.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 07:15:57 +03:00
" ftrace_preempt " ,
2008-11-12 23:24:24 +03:00
# ifdef CONFIG_BRANCH_TRACER
2008-11-12 23:24:24 +03:00
" branch " ,
2008-11-12 08:14:40 +03:00
# endif
2008-05-12 23:20:42 +04:00
NULL
} ;
2008-05-12 23:21:00 +04:00
/*
* ftrace_max_lock is used to protect the swapping of buffers
* when taking a max snapshot . The buffers themselves are
* protected by per_cpu spinlocks . But the action of the swap
* needs its own lock .
*
* This is defined as a raw_spinlock_t in order to help
* with performance when lockdep debugging is enabled .
*/
2008-05-12 23:20:55 +04:00
static raw_spinlock_t ftrace_max_lock =
( raw_spinlock_t ) __RAW_SPIN_LOCK_UNLOCKED ;
2008-05-12 23:20:42 +04:00
/*
* Copy the new maximum trace into the separate maximum - trace
* structure . ( this way the maximum trace is permanently saved ,
* for later retrieval via / debugfs / tracing / latency_trace )
*/
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:42 +04:00
__update_max_tr ( struct trace_array * tr , struct task_struct * tsk , int cpu )
{
struct trace_array_cpu * data = tr - > data [ cpu ] ;
max_tr . cpu = cpu ;
max_tr . time_start = data - > preempt_timestamp ;
data = max_tr . data [ cpu ] ;
data - > saved_latency = tracing_max_latency ;
memcpy ( data - > comm , tsk - > comm , TASK_COMM_LEN ) ;
data - > pid = tsk - > pid ;
data - > uid = tsk - > uid ;
data - > nice = tsk - > static_prio - 20 - MAX_RT_PRIO ;
data - > policy = tsk - > policy ;
data - > rt_priority = tsk - > rt_priority ;
/* record this tasks comm */
tracing_record_cmdline ( current ) ;
}
2008-05-12 23:21:00 +04:00
/**
* trace_seq_printf - sequence printing of trace information
* @ s : trace sequence descriptor
* @ fmt : printf format string
*
* The tracer may use either sequence operations or its own
* copy to user routines . To simplify formating of a trace
* trace_seq_printf is used to store strings into a special
* buffer ( @ s ) . Then the output may be either used by
* the sequencer or pulled into another buffer .
*/
2008-05-23 23:37:28 +04:00
int
2008-05-12 23:20:46 +04:00
trace_seq_printf ( struct trace_seq * s , const char * fmt , . . . )
{
int len = ( PAGE_SIZE - 1 ) - s - > len ;
va_list ap ;
2008-05-12 23:20:46 +04:00
int ret ;
2008-05-12 23:20:46 +04:00
if ( ! len )
return 0 ;
va_start ( ap , fmt ) ;
2008-05-12 23:20:46 +04:00
ret = vsnprintf ( s - > buffer + s - > len , len , fmt , ap ) ;
2008-05-12 23:20:46 +04:00
va_end ( ap ) ;
2008-05-12 23:20:46 +04:00
/* If we can't write it all, don't bother writing anything */
2008-05-23 23:37:28 +04:00
if ( ret > = len )
2008-05-12 23:20:46 +04:00
return 0 ;
s - > len + = ret ;
2008-05-12 23:20:46 +04:00
return len ;
}
2008-05-12 23:21:00 +04:00
/**
* trace_seq_puts - trace sequence printing of simple string
* @ s : trace sequence descriptor
* @ str : simple string to record
*
* The tracer may use either the sequence operations or its own
* copy to user routines . This function records a simple string
* into a special buffer ( @ s ) for later retrieval by a sequencer
* or other mechanism .
*/
2008-05-12 23:20:51 +04:00
static int
2008-05-12 23:20:46 +04:00
trace_seq_puts ( struct trace_seq * s , const char * str )
{
int len = strlen ( str ) ;
if ( len > ( ( PAGE_SIZE - 1 ) - s - > len ) )
2008-05-12 23:20:46 +04:00
return 0 ;
2008-05-12 23:20:46 +04:00
memcpy ( s - > buffer + s - > len , str , len ) ;
s - > len + = len ;
return len ;
}
2008-05-12 23:20:51 +04:00
static int
2008-05-12 23:20:46 +04:00
trace_seq_putc ( struct trace_seq * s , unsigned char c )
{
if ( s - > len > = ( PAGE_SIZE - 1 ) )
return 0 ;
s - > buffer [ s - > len + + ] = c ;
return 1 ;
}
2008-05-12 23:20:51 +04:00
static int
2008-05-12 23:20:47 +04:00
trace_seq_putmem ( struct trace_seq * s , void * mem , size_t len )
{
if ( len > ( ( PAGE_SIZE - 1 ) - s - > len ) )
return 0 ;
memcpy ( s - > buffer + s - > len , mem , len ) ;
s - > len + = len ;
return len ;
}
2008-10-09 04:44:55 +04:00
# define MAX_MEMHEX_BYTES 8
# define HEX_CHARS (MAX_MEMHEX_BYTES*2 + 1)
2008-05-12 23:20:49 +04:00
2008-05-12 23:20:51 +04:00
static int
2008-05-12 23:20:49 +04:00
trace_seq_putmem_hex ( struct trace_seq * s , void * mem , size_t len )
{
unsigned char hex [ HEX_CHARS ] ;
2008-05-12 23:21:00 +04:00
unsigned char * data = mem ;
2008-05-12 23:20:49 +04:00
int i , j ;
# ifdef __BIG_ENDIAN
for ( i = 0 , j = 0 ; i < len ; i + + ) {
# else
for ( i = len - 1 , j = 0 ; i > = 0 ; i - - ) {
# endif
2008-10-09 03:51:49 +04:00
hex [ j + + ] = hex_asc_hi ( data [ i ] ) ;
hex [ j + + ] = hex_asc_lo ( data [ i ] ) ;
2008-05-12 23:20:49 +04:00
}
2008-05-12 23:21:00 +04:00
hex [ j + + ] = ' ' ;
2008-05-12 23:20:49 +04:00
return trace_seq_putmem ( s , hex , j ) ;
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:46 +04:00
trace_seq_reset ( struct trace_seq * s )
{
s - > len = 0 ;
2008-05-12 23:21:02 +04:00
s - > readpos = 0 ;
}
ssize_t trace_seq_to_user ( struct trace_seq * s , char __user * ubuf , size_t cnt )
{
int len ;
int ret ;
if ( s - > len < = s - > readpos )
return - EBUSY ;
len = s - > len - s - > readpos ;
if ( cnt > len )
cnt = len ;
ret = copy_to_user ( ubuf , s - > buffer + s - > readpos , cnt ) ;
if ( ret )
return - EFAULT ;
s - > readpos + = len ;
return cnt ;
2008-05-12 23:20:46 +04:00
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:46 +04:00
trace_print_seq ( struct seq_file * m , struct trace_seq * s )
{
int len = s - > len > = PAGE_SIZE ? PAGE_SIZE - 1 : s - > len ;
s - > buffer [ len ] = 0 ;
seq_puts ( m , s - > buffer ) ;
trace_seq_reset ( s ) ;
}
2008-05-12 23:21:00 +04:00
/**
* update_max_tr - snapshot all trace buffers from global_trace to max_tr
* @ tr : tracer
* @ tsk : the task with the latency
* @ cpu : The cpu that initiated the trace .
*
* Flip the buffers between the @ tr and the max_tr and record information
* about which task was the cause of this latency .
*/
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:42 +04:00
update_max_tr ( struct trace_array * tr , struct task_struct * tsk , int cpu )
{
2008-09-30 07:02:41 +04:00
struct ring_buffer * buf = tr - > buffer ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:43 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_lock ( & ftrace_max_lock ) ;
2008-09-30 07:02:41 +04:00
tr - > buffer = max_tr . buffer ;
max_tr . buffer = buf ;
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_reset ( tr - > buffer ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
__update_max_tr ( tr , tsk , cpu ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_unlock ( & ftrace_max_lock ) ;
2008-05-12 23:20:42 +04:00
}
/**
* update_max_tr_single - only copy one trace over , and reset the rest
* @ tr - tracer
* @ tsk - task with the latency
* @ cpu - the cpu of the buffer to copy .
2008-05-12 23:21:00 +04:00
*
* Flip the trace of a single CPU buffer between the @ tr and the max_tr .
2008-05-12 23:20:42 +04:00
*/
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:42 +04:00
update_max_tr_single ( struct trace_array * tr , struct task_struct * tsk , int cpu )
{
2008-09-30 07:02:41 +04:00
int ret ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:43 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_lock ( & ftrace_max_lock ) ;
2008-05-12 23:20:42 +04:00
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_reset ( max_tr . buffer ) ;
ret = ring_buffer_swap_cpu ( max_tr . buffer , tr - > buffer , cpu ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
WARN_ON_ONCE ( ret ) ;
2008-05-12 23:20:42 +04:00
__update_max_tr ( tr , tsk , cpu ) ;
2008-05-12 23:20:55 +04:00
__raw_spin_unlock ( & ftrace_max_lock ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:21:00 +04:00
/**
* register_tracer - register a tracer with the ftrace system .
* @ type - the plugin for the tracer
*
* Register a new plugin tracer .
*/
2008-05-12 23:20:42 +04:00
int register_tracer ( struct tracer * type )
{
struct tracer * t ;
int len ;
int ret = 0 ;
if ( ! type - > name ) {
pr_info ( " Tracer must have a name \n " ) ;
return - 1 ;
}
mutex_lock ( & trace_types_lock ) ;
for ( t = trace_types ; t ; t = t - > next ) {
if ( strcmp ( type - > name , t - > name ) = = 0 ) {
/* already found */
pr_info ( " Trace %s already registered \n " ,
type - > name ) ;
ret = - 1 ;
goto out ;
}
}
2008-05-12 23:20:44 +04:00
# ifdef CONFIG_FTRACE_STARTUP_TEST
if ( type - > selftest ) {
struct tracer * saved_tracer = current_trace ;
struct trace_array * tr = & global_trace ;
int i ;
/*
* Run a selftest on this tracer .
* Here we reset the trace buffer , and set the current
* tracer to be this tracer . The tracer can then run some
* internal tracing to verify that everything is in order .
* If we fail , we do not register this tracer .
*/
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( i ) {
2008-09-30 07:02:41 +04:00
tracing_reset ( tr , i ) ;
2008-05-12 23:20:44 +04:00
}
current_trace = type ;
/* the test is responsible for initializing and enabling */
pr_info ( " Testing tracer %s: " , type - > name ) ;
ret = type - > selftest ( type , tr ) ;
/* the test is responsible for resetting too */
current_trace = saved_tracer ;
if ( ret ) {
printk ( KERN_CONT " FAILED! \n " ) ;
goto out ;
}
2008-05-12 23:20:45 +04:00
/* Only reset on passing, to avoid touching corrupted buffers */
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( i ) {
2008-09-30 07:02:41 +04:00
tracing_reset ( tr , i ) ;
2008-05-12 23:20:45 +04:00
}
2008-05-12 23:20:44 +04:00
printk ( KERN_CONT " PASSED \n " ) ;
}
# endif
2008-05-12 23:20:42 +04:00
type - > next = trace_types ;
trace_types = type ;
len = strlen ( type - > name ) ;
if ( len > max_tracer_type_len )
max_tracer_type_len = len ;
2008-05-12 23:20:44 +04:00
2008-05-12 23:20:42 +04:00
out :
mutex_unlock ( & trace_types_lock ) ;
return ret ;
}
void unregister_tracer ( struct tracer * type )
{
struct tracer * * t ;
int len ;
mutex_lock ( & trace_types_lock ) ;
for ( t = & trace_types ; * t ; t = & ( * t ) - > next ) {
if ( * t = = type )
goto found ;
}
pr_info ( " Trace %s not registered \n " , type - > name ) ;
goto out ;
found :
* t = ( * t ) - > next ;
if ( strlen ( type - > name ) ! = max_tracer_type_len )
goto out ;
max_tracer_type_len = 0 ;
for ( t = & trace_types ; * t ; t = & ( * t ) - > next ) {
len = strlen ( ( * t ) - > name ) ;
if ( len > max_tracer_type_len )
max_tracer_type_len = len ;
}
out :
mutex_unlock ( & trace_types_lock ) ;
}
2008-09-30 07:02:41 +04:00
void tracing_reset ( struct trace_array * tr , int cpu )
2008-05-12 23:20:42 +04:00
{
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_reset_cpu ( tr - > buffer , cpu ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
}
# define SAVED_CMDLINES 128
static unsigned map_pid_to_cmdline [ PID_MAX_DEFAULT + 1 ] ;
static unsigned map_cmdline_to_pid [ SAVED_CMDLINES ] ;
static char saved_cmdlines [ SAVED_CMDLINES ] [ TASK_COMM_LEN ] ;
static int cmdline_idx ;
static DEFINE_SPINLOCK ( trace_cmdline_lock ) ;
2008-05-12 23:21:00 +04:00
/* temporary disable recording */
atomic_t trace_record_cmdline_disabled __read_mostly ;
2008-05-12 23:20:42 +04:00
static void trace_init_cmdlines ( void )
{
memset ( & map_pid_to_cmdline , - 1 , sizeof ( map_pid_to_cmdline ) ) ;
memset ( & map_cmdline_to_pid , - 1 , sizeof ( map_cmdline_to_pid ) ) ;
cmdline_idx = 0 ;
}
2008-11-06 00:05:44 +03:00
static int trace_stop_count ;
static DEFINE_SPINLOCK ( tracing_start_lock ) ;
/**
* tracing_start - quick start of the tracer
*
* If tracing is enabled but was stopped by tracing_stop ,
* this will start the tracer back up .
*/
void tracing_start ( void )
{
struct ring_buffer * buffer ;
unsigned long flags ;
if ( tracing_disabled )
return ;
spin_lock_irqsave ( & tracing_start_lock , flags ) ;
if ( - - trace_stop_count )
goto out ;
if ( trace_stop_count < 0 ) {
/* Someone screwed up their debugging */
WARN_ON_ONCE ( 1 ) ;
trace_stop_count = 0 ;
goto out ;
}
buffer = global_trace . buffer ;
if ( buffer )
ring_buffer_record_enable ( buffer ) ;
buffer = max_tr . buffer ;
if ( buffer )
ring_buffer_record_enable ( buffer ) ;
ftrace_start ( ) ;
out :
spin_unlock_irqrestore ( & tracing_start_lock , flags ) ;
}
/**
* tracing_stop - quick stop of the tracer
*
* Light weight way to stop tracing . Use in conjunction with
* tracing_start .
*/
void tracing_stop ( void )
{
struct ring_buffer * buffer ;
unsigned long flags ;
ftrace_stop ( ) ;
spin_lock_irqsave ( & tracing_start_lock , flags ) ;
if ( trace_stop_count + + )
goto out ;
buffer = global_trace . buffer ;
if ( buffer )
ring_buffer_record_disable ( buffer ) ;
buffer = max_tr . buffer ;
if ( buffer )
ring_buffer_record_disable ( buffer ) ;
out :
spin_unlock_irqrestore ( & tracing_start_lock , flags ) ;
}
2008-05-12 23:20:51 +04:00
void trace_stop_cmdline_recording ( void ) ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:51 +04:00
static void trace_save_cmdline ( struct task_struct * tsk )
2008-05-12 23:20:42 +04:00
{
unsigned map ;
unsigned idx ;
if ( ! tsk - > pid | | unlikely ( tsk - > pid > PID_MAX_DEFAULT ) )
return ;
/*
* It ' s not the end of the world if we don ' t get
* the lock , but we also don ' t want to spin
* nor do we want to disable interrupts ,
* so if we miss here , then better luck next time .
*/
if ( ! spin_trylock ( & trace_cmdline_lock ) )
return ;
idx = map_pid_to_cmdline [ tsk - > pid ] ;
if ( idx > = SAVED_CMDLINES ) {
idx = ( cmdline_idx + 1 ) % SAVED_CMDLINES ;
map = map_cmdline_to_pid [ idx ] ;
if ( map < = PID_MAX_DEFAULT )
map_pid_to_cmdline [ map ] = ( unsigned ) - 1 ;
map_pid_to_cmdline [ tsk - > pid ] = idx ;
cmdline_idx = idx ;
}
memcpy ( & saved_cmdlines [ idx ] , tsk - > comm , TASK_COMM_LEN ) ;
spin_unlock ( & trace_cmdline_lock ) ;
}
2008-05-12 23:20:51 +04:00
static char * trace_find_cmdline ( int pid )
2008-05-12 23:20:42 +04:00
{
char * cmdline = " <...> " ;
unsigned map ;
if ( ! pid )
return " <idle> " ;
if ( pid > PID_MAX_DEFAULT )
goto out ;
map = map_pid_to_cmdline [ pid ] ;
if ( map > = SAVED_CMDLINES )
goto out ;
cmdline = saved_cmdlines [ map ] ;
out :
return cmdline ;
}
2008-05-12 23:20:51 +04:00
void tracing_record_cmdline ( struct task_struct * tsk )
2008-05-12 23:20:42 +04:00
{
if ( atomic_read ( & trace_record_cmdline_disabled ) )
return ;
trace_save_cmdline ( tsk ) ;
}
2008-09-16 22:56:41 +04:00
void
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( struct trace_entry * entry , unsigned long flags ,
int pc )
2008-05-12 23:20:42 +04:00
{
struct task_struct * tsk = current ;
2008-09-30 07:02:42 +04:00
entry - > preempt_count = pc & 0xff ;
entry - > pid = ( tsk ) ? tsk - > pid : 0 ;
entry - > flags =
2008-10-24 17:42:59 +04:00
# ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT
2008-08-01 20:26:40 +04:00
( irqs_disabled_flags ( flags ) ? TRACE_FLAG_IRQS_OFF : 0 ) |
2008-10-24 17:42:59 +04:00
# else
TRACE_FLAG_IRQS_NOSUPPORT |
# endif
2008-05-12 23:20:42 +04:00
( ( pc & HARDIRQ_MASK ) ? TRACE_FLAG_HARDIRQ : 0 ) |
( ( pc & SOFTIRQ_MASK ) ? TRACE_FLAG_SOFTIRQ : 0 ) |
( need_resched ( ) ? TRACE_FLAG_NEED_RESCHED : 0 ) ;
}
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:49 +04:00
trace_function ( struct trace_array * tr , struct trace_array_cpu * data ,
2008-10-01 21:14:09 +04:00
unsigned long ip , unsigned long parent_ip , unsigned long flags ,
int pc )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct ftrace_entry * entry ;
2008-05-12 23:20:48 +04:00
unsigned long irq_flags ;
2008-05-12 23:20:42 +04:00
2008-10-01 08:29:53 +04:00
/* If we are reading the ring buffer, don't trace */
if ( unlikely ( local_read ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ) )
return ;
2008-09-30 07:02:41 +04:00
event = ring_buffer_lock_reserve ( tr - > buffer , sizeof ( * entry ) ,
& irq_flags ) ;
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( & entry - > ent , flags , pc ) ;
2008-09-30 07:02:42 +04:00
entry - > ent . type = TRACE_FN ;
entry - > ip = ip ;
entry - > parent_ip = parent_ip ;
2008-09-30 07:02:41 +04:00
ring_buffer_unlock_commit ( tr - > buffer , event , irq_flags ) ;
2008-05-12 23:20:42 +04:00
}
2008-11-11 09:14:25 +03:00
# ifdef CONFIG_FUNCTION_RET_TRACER
static void __trace_function_return ( struct trace_array * tr ,
struct trace_array_cpu * data ,
struct ftrace_retfunc * trace ,
unsigned long flags ,
int pc )
{
struct ring_buffer_event * event ;
struct ftrace_ret_entry * entry ;
unsigned long irq_flags ;
if ( unlikely ( local_read ( & __get_cpu_var ( ftrace_cpu_disabled ) ) ) )
return ;
event = ring_buffer_lock_reserve ( global_trace . buffer , sizeof ( * entry ) ,
& irq_flags ) ;
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
tracing_generic_entry_update ( & entry - > ent , flags , pc ) ;
entry - > ent . type = TRACE_FN_RET ;
entry - > ip = trace - > func ;
entry - > parent_ip = trace - > ret ;
entry - > rettime = trace - > rettime ;
entry - > calltime = trace - > calltime ;
ring_buffer_unlock_commit ( global_trace . buffer , event , irq_flags ) ;
}
# endif
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:49 +04:00
ftrace ( struct trace_array * tr , struct trace_array_cpu * data ,
2008-10-01 21:14:09 +04:00
unsigned long ip , unsigned long parent_ip , unsigned long flags ,
int pc )
2008-05-12 23:20:49 +04:00
{
if ( likely ( ! atomic_read ( & data - > disabled ) ) )
2008-10-01 21:14:09 +04:00
trace_function ( tr , data , ip , parent_ip , flags , pc ) ;
2008-05-12 23:20:49 +04:00
}
2008-10-01 21:14:09 +04:00
static void ftrace_trace_stack ( struct trace_array * tr ,
struct trace_array_cpu * data ,
unsigned long flags ,
int skip , int pc )
2008-05-12 23:20:51 +04:00
{
2008-10-31 22:50:41 +03:00
# ifdef CONFIG_STACKTRACE
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct stack_entry * entry ;
2008-05-12 23:20:51 +04:00
struct stack_trace trace ;
2008-09-30 07:02:41 +04:00
unsigned long irq_flags ;
2008-05-12 23:20:51 +04:00
if ( ! ( trace_flags & TRACE_ITER_STACKTRACE ) )
return ;
2008-09-30 07:02:41 +04:00
event = ring_buffer_lock_reserve ( tr - > buffer , sizeof ( * entry ) ,
& irq_flags ) ;
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( & entry - > ent , flags , pc ) ;
2008-09-30 07:02:42 +04:00
entry - > ent . type = TRACE_STACK ;
2008-05-12 23:20:51 +04:00
2008-09-30 07:02:42 +04:00
memset ( & entry - > caller , 0 , sizeof ( entry - > caller ) ) ;
2008-05-12 23:20:51 +04:00
trace . nr_entries = 0 ;
trace . max_entries = FTRACE_STACK_ENTRIES ;
trace . skip = skip ;
2008-09-30 07:02:42 +04:00
trace . entries = entry - > caller ;
2008-05-12 23:20:51 +04:00
save_stack_trace ( & trace ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_unlock_commit ( tr - > buffer , event , irq_flags ) ;
2008-10-31 22:50:41 +03:00
# endif
2008-05-12 23:20:47 +04:00
}
2008-10-01 21:14:09 +04:00
void __trace_stack ( struct trace_array * tr ,
struct trace_array_cpu * data ,
unsigned long flags ,
int skip )
{
ftrace_trace_stack ( tr , data , flags , skip , preempt_count ( ) ) ;
}
static void
ftrace_trace_special ( void * __tr , void * __data ,
unsigned long arg1 , unsigned long arg2 , unsigned long arg3 ,
int pc )
2008-05-12 23:21:02 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-05-12 23:21:02 +04:00
struct trace_array_cpu * data = __data ;
struct trace_array * tr = __tr ;
2008-09-30 07:02:42 +04:00
struct special_entry * entry ;
2008-05-12 23:21:02 +04:00
unsigned long irq_flags ;
2008-09-30 07:02:41 +04:00
event = ring_buffer_lock_reserve ( tr - > buffer , sizeof ( * entry ) ,
& irq_flags ) ;
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( & entry - > ent , 0 , pc ) ;
2008-09-30 07:02:42 +04:00
entry - > ent . type = TRACE_SPECIAL ;
entry - > arg1 = arg1 ;
entry - > arg2 = arg2 ;
entry - > arg3 = arg3 ;
2008-09-30 07:02:41 +04:00
ring_buffer_unlock_commit ( tr - > buffer , event , irq_flags ) ;
2008-10-01 21:14:09 +04:00
ftrace_trace_stack ( tr , data , irq_flags , 4 , pc ) ;
2008-05-12 23:21:02 +04:00
trace_wake_up ( ) ;
}
2008-10-01 21:14:09 +04:00
void
__trace_special ( void * __tr , void * __data ,
unsigned long arg1 , unsigned long arg2 , unsigned long arg3 )
{
ftrace_trace_special ( __tr , __data , arg1 , arg2 , arg3 , preempt_count ( ) ) ;
}
2008-05-12 23:20:51 +04:00
void
2008-05-12 23:20:42 +04:00
tracing_sched_switch_trace ( struct trace_array * tr ,
struct trace_array_cpu * data ,
2008-05-12 23:20:51 +04:00
struct task_struct * prev ,
struct task_struct * next ,
2008-10-01 21:14:09 +04:00
unsigned long flags , int pc )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct ctx_switch_entry * entry ;
2008-05-12 23:20:48 +04:00
unsigned long irq_flags ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
event = ring_buffer_lock_reserve ( tr - > buffer , sizeof ( * entry ) ,
& irq_flags ) ;
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( & entry - > ent , flags , pc ) ;
2008-09-30 07:02:42 +04:00
entry - > ent . type = TRACE_CTX ;
entry - > prev_pid = prev - > pid ;
entry - > prev_prio = prev - > prio ;
entry - > prev_state = prev - > state ;
entry - > next_pid = next - > pid ;
entry - > next_prio = next - > prio ;
entry - > next_state = next - > state ;
entry - > next_cpu = task_cpu ( next ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_unlock_commit ( tr - > buffer , event , irq_flags ) ;
2008-10-01 21:14:09 +04:00
ftrace_trace_stack ( tr , data , flags , 5 , pc ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
void
tracing_sched_wakeup_trace ( struct trace_array * tr ,
struct trace_array_cpu * data ,
2008-05-12 23:20:51 +04:00
struct task_struct * wakee ,
struct task_struct * curr ,
2008-10-01 21:14:09 +04:00
unsigned long flags , int pc )
2008-05-12 23:20:51 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-30 07:02:42 +04:00
struct ctx_switch_entry * entry ;
2008-05-12 23:20:51 +04:00
unsigned long irq_flags ;
2008-09-30 07:02:41 +04:00
event = ring_buffer_lock_reserve ( tr - > buffer , sizeof ( * entry ) ,
& irq_flags ) ;
if ( ! event )
return ;
entry = ring_buffer_event_data ( event ) ;
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( & entry - > ent , flags , pc ) ;
2008-09-30 07:02:42 +04:00
entry - > ent . type = TRACE_WAKE ;
entry - > prev_pid = curr - > pid ;
entry - > prev_prio = curr - > prio ;
entry - > prev_state = curr - > state ;
entry - > next_pid = wakee - > pid ;
entry - > next_prio = wakee - > prio ;
entry - > next_state = wakee - > state ;
entry - > next_cpu = task_cpu ( wakee ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_unlock_commit ( tr - > buffer , event , irq_flags ) ;
2008-10-01 21:14:09 +04:00
ftrace_trace_stack ( tr , data , flags , 6 , pc ) ;
2008-05-12 23:20:52 +04:00
trace_wake_up ( ) ;
2008-05-12 23:20:51 +04:00
}
2008-05-22 08:22:18 +04:00
void
ftrace_special ( unsigned long arg1 , unsigned long arg2 , unsigned long arg3 )
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
2008-11-11 07:07:30 +03:00
unsigned long flags ;
2008-05-22 08:22:18 +04:00
int cpu ;
2008-10-01 21:14:09 +04:00
int pc ;
2008-05-22 08:22:18 +04:00
2008-11-08 06:36:02 +03:00
if ( tracing_disabled )
2008-05-22 08:22:18 +04:00
return ;
2008-10-01 21:14:09 +04:00
pc = preempt_count ( ) ;
2008-11-11 07:07:30 +03:00
local_irq_save ( flags ) ;
2008-05-22 08:22:18 +04:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
2008-11-11 07:07:30 +03:00
if ( likely ( atomic_inc_return ( & data - > disabled ) = = 1 ) )
2008-10-01 21:14:09 +04:00
ftrace_trace_special ( tr , data , arg1 , arg2 , arg3 , pc ) ;
2008-05-22 08:22:18 +04:00
2008-11-11 07:07:30 +03:00
atomic_dec ( & data - > disabled ) ;
local_irq_restore ( flags ) ;
2008-05-22 08:22:18 +04:00
}
2008-10-07 03:06:12 +04:00
# ifdef CONFIG_FUNCTION_TRACER
2008-05-12 23:20:51 +04:00
static void
ftrace: function tracer with irqs disabled
Impact: disable interrupts during trace entry creation (as opposed to preempt)
To help with performance, I set the ftracer to not disable interrupts,
and only to disable preemption. If an interrupt occurred, it would not
be traced, because the function tracer protects itself from recursion.
This may be faster, but the trace output might miss some traces.
This patch makes the fuction trace disable interrupts, but it also
adds a runtime feature to disable preemption instead. It does this by
having two different tracer functions. When the function tracer is
enabled, it will check to see which version is requested (irqs disabled
or preemption disabled). Then it will use the corresponding function
as the tracer.
Irq disabling is the default behaviour, but if the user wants better
performance, with the chance of missing traces, then they can choose
the preempt disabled version.
Running hackbench 3 times with the irqs disabled and 3 times with
the preempt disabled function tracer yielded:
tracing type times entries recorded
------------ -------- ----------------
irq disabled 43.393 166433066
43.282 166172618
43.298 166256704
preempt disabled 38.969 159871710
38.943 159972935
39.325 161056510
Average:
irqs disabled: 43.324 166287462
preempt disabled: 39.079 160300385
preempt is 10.8 percent faster than irqs disabled.
I wrote a patch to count function trace recursion and reran hackbench.
With irq disabled: 1,150 times the function tracer did not trace due to
recursion.
with preempt disabled: 5,117,718 times.
The thousand times with irq disabled could be due to NMIs, or simply a case
where it called a function that was not protected by notrace.
But we also see that a large amount of the trace is lost with the
preempt version.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 07:15:57 +03:00
function_trace_call_preempt_only ( unsigned long ip , unsigned long parent_ip )
2008-05-12 23:20:49 +04:00
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
unsigned long flags ;
long disabled ;
2008-10-01 21:14:09 +04:00
int cpu , resched ;
int pc ;
2008-05-12 23:20:49 +04:00
2008-07-11 04:58:16 +04:00
if ( unlikely ( ! ftrace_function_enabled ) )
2008-05-12 23:20:49 +04:00
return ;
2008-10-01 21:14:09 +04:00
pc = preempt_count ( ) ;
2008-11-04 07:15:56 +03:00
resched = ftrace_preempt_disable ( ) ;
2008-10-01 21:14:09 +04:00
local_save_flags ( flags ) ;
2008-05-12 23:20:49 +04:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
disabled = atomic_inc_return ( & data - > disabled ) ;
if ( likely ( disabled = = 1 ) )
2008-10-01 21:14:09 +04:00
trace_function ( tr , data , ip , parent_ip , flags , pc ) ;
2008-05-12 23:20:49 +04:00
atomic_dec ( & data - > disabled ) ;
2008-11-04 07:15:56 +03:00
ftrace_preempt_enable ( resched ) ;
2008-05-12 23:20:49 +04:00
}
ftrace: function tracer with irqs disabled
Impact: disable interrupts during trace entry creation (as opposed to preempt)
To help with performance, I set the ftracer to not disable interrupts,
and only to disable preemption. If an interrupt occurred, it would not
be traced, because the function tracer protects itself from recursion.
This may be faster, but the trace output might miss some traces.
This patch makes the fuction trace disable interrupts, but it also
adds a runtime feature to disable preemption instead. It does this by
having two different tracer functions. When the function tracer is
enabled, it will check to see which version is requested (irqs disabled
or preemption disabled). Then it will use the corresponding function
as the tracer.
Irq disabling is the default behaviour, but if the user wants better
performance, with the chance of missing traces, then they can choose
the preempt disabled version.
Running hackbench 3 times with the irqs disabled and 3 times with
the preempt disabled function tracer yielded:
tracing type times entries recorded
------------ -------- ----------------
irq disabled 43.393 166433066
43.282 166172618
43.298 166256704
preempt disabled 38.969 159871710
38.943 159972935
39.325 161056510
Average:
irqs disabled: 43.324 166287462
preempt disabled: 39.079 160300385
preempt is 10.8 percent faster than irqs disabled.
I wrote a patch to count function trace recursion and reran hackbench.
With irq disabled: 1,150 times the function tracer did not trace due to
recursion.
with preempt disabled: 5,117,718 times.
The thousand times with irq disabled could be due to NMIs, or simply a case
where it called a function that was not protected by notrace.
But we also see that a large amount of the trace is lost with the
preempt version.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 07:15:57 +03:00
static void
function_trace_call ( unsigned long ip , unsigned long parent_ip )
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
unsigned long flags ;
long disabled ;
int cpu ;
int pc ;
if ( unlikely ( ! ftrace_function_enabled ) )
return ;
/*
* Need to use raw , since this must be called before the
* recursive protection is performed .
*/
raw_local_irq_save ( flags ) ;
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
disabled = atomic_inc_return ( & data - > disabled ) ;
if ( likely ( disabled = = 1 ) ) {
pc = preempt_count ( ) ;
trace_function ( tr , data , ip , parent_ip , flags , pc ) ;
}
atomic_dec ( & data - > disabled ) ;
raw_local_irq_restore ( flags ) ;
}
2008-11-11 09:14:25 +03:00
# ifdef CONFIG_FUNCTION_RET_TRACER
void trace_function_return ( struct ftrace_retfunc * trace )
{
struct trace_array * tr = & global_trace ;
struct trace_array_cpu * data ;
unsigned long flags ;
long disabled ;
int cpu ;
int pc ;
raw_local_irq_save ( flags ) ;
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
disabled = atomic_inc_return ( & data - > disabled ) ;
if ( likely ( disabled = = 1 ) ) {
pc = preempt_count ( ) ;
__trace_function_return ( tr , data , trace , flags , pc ) ;
}
atomic_dec ( & data - > disabled ) ;
raw_local_irq_restore ( flags ) ;
}
# endif /* CONFIG_FUNCTION_RET_TRACER */
2008-05-12 23:20:49 +04:00
static struct ftrace_ops trace_ops __read_mostly =
{
. func = function_trace_call ,
} ;
2008-05-12 23:20:51 +04:00
void tracing_start_function_trace ( void )
2008-05-12 23:20:49 +04:00
{
2008-07-11 04:58:16 +04:00
ftrace_function_enabled = 0 ;
ftrace: function tracer with irqs disabled
Impact: disable interrupts during trace entry creation (as opposed to preempt)
To help with performance, I set the ftracer to not disable interrupts,
and only to disable preemption. If an interrupt occurred, it would not
be traced, because the function tracer protects itself from recursion.
This may be faster, but the trace output might miss some traces.
This patch makes the fuction trace disable interrupts, but it also
adds a runtime feature to disable preemption instead. It does this by
having two different tracer functions. When the function tracer is
enabled, it will check to see which version is requested (irqs disabled
or preemption disabled). Then it will use the corresponding function
as the tracer.
Irq disabling is the default behaviour, but if the user wants better
performance, with the chance of missing traces, then they can choose
the preempt disabled version.
Running hackbench 3 times with the irqs disabled and 3 times with
the preempt disabled function tracer yielded:
tracing type times entries recorded
------------ -------- ----------------
irq disabled 43.393 166433066
43.282 166172618
43.298 166256704
preempt disabled 38.969 159871710
38.943 159972935
39.325 161056510
Average:
irqs disabled: 43.324 166287462
preempt disabled: 39.079 160300385
preempt is 10.8 percent faster than irqs disabled.
I wrote a patch to count function trace recursion and reran hackbench.
With irq disabled: 1,150 times the function tracer did not trace due to
recursion.
with preempt disabled: 5,117,718 times.
The thousand times with irq disabled could be due to NMIs, or simply a case
where it called a function that was not protected by notrace.
But we also see that a large amount of the trace is lost with the
preempt version.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 07:15:57 +03:00
if ( trace_flags & TRACE_ITER_PREEMPTONLY )
trace_ops . func = function_trace_call_preempt_only ;
else
trace_ops . func = function_trace_call ;
2008-05-12 23:20:49 +04:00
register_ftrace_function ( & trace_ops ) ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
ftrace_function_enabled = 1 ;
2008-05-12 23:20:49 +04:00
}
2008-05-12 23:20:51 +04:00
void tracing_stop_function_trace ( void )
2008-05-12 23:20:49 +04:00
{
2008-07-11 04:58:16 +04:00
ftrace_function_enabled = 0 ;
2008-05-12 23:20:49 +04:00
unregister_ftrace_function ( & trace_ops ) ;
}
# endif
2008-05-12 23:20:42 +04:00
enum trace_file_type {
TRACE_FILE_LAT_FMT = 1 ,
} ;
2008-09-04 01:42:51 +04:00
static void trace_iterator_increment ( struct trace_iterator * iter , int cpu )
{
2008-10-01 08:29:53 +04:00
/* Don't allow ftrace to trace into the ring buffers */
ftrace_disable_cpu ( ) ;
2008-09-04 01:42:51 +04:00
iter - > idx + + ;
2008-10-01 08:29:53 +04:00
if ( iter - > buffer_iter [ iter - > cpu ] )
ring_buffer_read ( iter - > buffer_iter [ iter - > cpu ] , NULL ) ;
ftrace_enable_cpu ( ) ;
2008-09-04 01:42:51 +04:00
}
2008-05-12 23:20:51 +04:00
static struct trace_entry *
2008-09-30 07:02:41 +04:00
peek_next_entry ( struct trace_iterator * iter , int cpu , u64 * ts )
2008-08-01 20:26:41 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
struct ring_buffer_iter * buf_iter = iter - > buffer_iter [ cpu ] ;
2008-08-01 20:26:41 +04:00
2008-10-01 08:29:53 +04:00
/* Don't allow ftrace to trace into the ring buffers */
ftrace_disable_cpu ( ) ;
if ( buf_iter )
event = ring_buffer_iter_peek ( buf_iter , ts ) ;
else
event = ring_buffer_peek ( iter - > tr - > buffer , cpu , ts ) ;
ftrace_enable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
return event ? ring_buffer_event_data ( event ) : NULL ;
2008-08-01 20:26:41 +04:00
}
2008-10-01 08:29:53 +04:00
2008-08-01 20:26:41 +04:00
static struct trace_entry *
2008-09-30 07:02:41 +04:00
__find_next_entry ( struct trace_iterator * iter , int * ent_cpu , u64 * ent_ts )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
struct ring_buffer * buffer = iter - > tr - > buffer ;
2008-05-12 23:20:42 +04:00
struct trace_entry * ent , * next = NULL ;
2008-09-30 07:02:41 +04:00
u64 next_ts = 0 , ts ;
2008-05-12 23:20:42 +04:00
int next_cpu = - 1 ;
int cpu ;
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( cpu ) {
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
if ( ring_buffer_empty_cpu ( buffer , cpu ) )
continue ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
ent = peek_next_entry ( iter , cpu , & ts ) ;
2008-08-01 20:26:41 +04:00
2008-05-12 23:20:46 +04:00
/*
* Pick the entry with the smallest timestamp :
*/
2008-09-30 07:02:41 +04:00
if ( ent & & ( ! next | | ts < next_ts ) ) {
2008-05-12 23:20:42 +04:00
next = ent ;
next_cpu = cpu ;
2008-09-30 07:02:41 +04:00
next_ts = ts ;
2008-05-12 23:20:42 +04:00
}
}
if ( ent_cpu )
* ent_cpu = next_cpu ;
2008-09-30 07:02:41 +04:00
if ( ent_ts )
* ent_ts = next_ts ;
2008-05-12 23:20:42 +04:00
return next ;
}
2008-08-01 20:26:41 +04:00
/* Find the next real entry, without updating the iterator itself */
static struct trace_entry *
2008-09-30 07:02:41 +04:00
find_next_entry ( struct trace_iterator * iter , int * ent_cpu , u64 * ent_ts )
2008-05-12 23:20:42 +04:00
{
2008-09-30 07:02:41 +04:00
return __find_next_entry ( iter , ent_cpu , ent_ts ) ;
2008-08-01 20:26:41 +04:00
}
/* Find the next real entry, and increment the iterator to the next entry */
static void * find_next_entry_inc ( struct trace_iterator * iter )
{
2008-09-30 07:02:41 +04:00
iter - > ent = __find_next_entry ( iter , & iter - > cpu , & iter - > ts ) ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
if ( iter - > ent )
2008-08-01 20:26:41 +04:00
trace_iterator_increment ( iter , iter - > cpu ) ;
2008-09-30 07:02:41 +04:00
return iter - > ent ? iter : NULL ;
2008-05-12 23:20:46 +04:00
}
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:51 +04:00
static void trace_consume ( struct trace_iterator * iter )
2008-05-12 23:20:46 +04:00
{
2008-10-01 08:29:53 +04:00
/* Don't allow ftrace to trace into the ring buffers */
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ring_buffer_consume ( iter - > tr - > buffer , iter - > cpu , & iter - > ts ) ;
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void * s_next ( struct seq_file * m , void * v , loff_t * pos )
2008-05-12 23:20:42 +04:00
{
struct trace_iterator * iter = m - > private ;
int i = ( int ) * pos ;
2008-05-12 23:20:45 +04:00
void * ent ;
2008-05-12 23:20:42 +04:00
( * pos ) + + ;
/* can't go backwards */
if ( iter - > idx > i )
return NULL ;
if ( iter - > idx < 0 )
ent = find_next_entry_inc ( iter ) ;
else
ent = iter ;
while ( ent & & iter - > idx < i )
ent = find_next_entry_inc ( iter ) ;
iter - > pos = * pos ;
return ent ;
}
static void * s_start ( struct seq_file * m , loff_t * pos )
{
struct trace_iterator * iter = m - > private ;
void * p = NULL ;
loff_t l = 0 ;
2008-09-30 07:02:41 +04:00
int cpu ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-05-12 23:20:56 +04:00
if ( ! current_trace | | current_trace ! = iter - > trace ) {
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:20:42 +04:00
return NULL ;
2008-05-12 23:20:56 +04:00
}
2008-05-12 23:20:42 +04:00
atomic_inc ( & trace_record_cmdline_disabled ) ;
if ( * pos ! = iter - > pos ) {
iter - > ent = NULL ;
iter - > cpu = 0 ;
iter - > idx = - 1 ;
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
for_each_tracing_cpu ( cpu ) {
ring_buffer_iter_reset ( iter - > buffer_iter [ cpu ] ) ;
2008-05-12 23:20:43 +04:00
}
2008-05-12 23:20:42 +04:00
2008-10-01 08:29:53 +04:00
ftrace_enable_cpu ( ) ;
2008-05-12 23:20:42 +04:00
for ( p = iter ; p & & l < * pos ; p = s_next ( m , p , & l ) )
;
} else {
2008-05-12 23:20:43 +04:00
l = * pos - 1 ;
2008-05-12 23:20:42 +04:00
p = s_next ( m , p , & l ) ;
}
return p ;
}
static void s_stop ( struct seq_file * m , void * p )
{
atomic_dec ( & trace_record_cmdline_disabled ) ;
mutex_unlock ( & trace_types_lock ) ;
}
2008-05-27 22:33:18 +04:00
# ifdef CONFIG_KRETPROBES
2008-10-31 22:44:07 +03:00
static inline const char * kretprobed ( const char * name )
2008-05-27 22:33:18 +04:00
{
2008-10-31 22:44:07 +03:00
static const char tramp_name [ ] = " kretprobe_trampoline " ;
int size = sizeof ( tramp_name ) ;
if ( strncmp ( tramp_name , name , size ) = = 0 )
return " [unknown/kretprobe'd] " ;
return name ;
2008-05-27 22:33:18 +04:00
}
# else
2008-10-31 22:44:07 +03:00
static inline const char * kretprobed ( const char * name )
2008-05-27 22:33:18 +04:00
{
2008-10-31 22:44:07 +03:00
return name ;
2008-05-27 22:33:18 +04:00
}
# endif /* CONFIG_KRETPROBES */
2008-05-12 23:20:46 +04:00
static int
2008-05-12 23:20:46 +04:00
seq_print_sym_short ( struct trace_seq * s , const char * fmt , unsigned long address )
2008-05-12 23:20:42 +04:00
{
# ifdef CONFIG_KALLSYMS
char str [ KSYM_SYMBOL_LEN ] ;
2008-10-31 22:44:07 +03:00
const char * name ;
2008-05-12 23:20:42 +04:00
kallsyms_lookup ( address , NULL , NULL , NULL , str ) ;
2008-10-31 22:44:07 +03:00
name = kretprobed ( str ) ;
return trace_seq_printf ( s , fmt , name ) ;
2008-05-12 23:20:42 +04:00
# endif
2008-05-12 23:20:46 +04:00
return 1 ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:46 +04:00
static int
2008-05-12 23:20:46 +04:00
seq_print_sym_offset ( struct trace_seq * s , const char * fmt ,
unsigned long address )
2008-05-12 23:20:42 +04:00
{
# ifdef CONFIG_KALLSYMS
char str [ KSYM_SYMBOL_LEN ] ;
2008-10-31 22:44:07 +03:00
const char * name ;
2008-05-12 23:20:42 +04:00
sprint_symbol ( str , address ) ;
2008-10-31 22:44:07 +03:00
name = kretprobed ( str ) ;
return trace_seq_printf ( s , fmt , name ) ;
2008-05-12 23:20:42 +04:00
# endif
2008-05-12 23:20:46 +04:00
return 1 ;
2008-05-12 23:20:42 +04:00
}
# ifndef CONFIG_64BIT
# define IP_FMT "%08lx"
# else
# define IP_FMT "%016lx"
# endif
2008-11-11 09:14:25 +03:00
int
2008-05-12 23:20:46 +04:00
seq_print_ip_sym ( struct trace_seq * s , unsigned long ip , unsigned long sym_flags )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:46 +04:00
int ret ;
if ( ! ip )
return trace_seq_printf ( s , " 0 " ) ;
2008-05-12 23:20:42 +04:00
if ( sym_flags & TRACE_ITER_SYM_OFFSET )
2008-05-12 23:20:46 +04:00
ret = seq_print_sym_offset ( s , " %s " , ip ) ;
2008-05-12 23:20:42 +04:00
else
2008-05-12 23:20:46 +04:00
ret = seq_print_sym_short ( s , " %s " , ip ) ;
if ( ! ret )
return 0 ;
2008-05-12 23:20:42 +04:00
if ( sym_flags & TRACE_ITER_SYM_ADDR )
2008-05-12 23:20:46 +04:00
ret = trace_seq_printf ( s , " < " IP_FMT " > " , ip ) ;
return ret ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void print_lat_help_header ( struct seq_file * m )
2008-05-12 23:20:42 +04:00
{
2008-08-21 03:36:11 +04:00
seq_puts ( m , " # _------=> CPU# \n " ) ;
seq_puts ( m , " # / _-----=> irqs-off \n " ) ;
seq_puts ( m , " # | / _----=> need-resched \n " ) ;
seq_puts ( m , " # || / _---=> hardirq/softirq \n " ) ;
seq_puts ( m , " # ||| / _--=> preempt-depth \n " ) ;
seq_puts ( m , " # |||| / \n " ) ;
seq_puts ( m , " # ||||| delay \n " ) ;
seq_puts ( m , " # cmd pid ||||| time | caller \n " ) ;
seq_puts ( m , " # \\ / ||||| \\ | / \n " ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void print_func_help_header ( struct seq_file * m )
2008-05-12 23:20:42 +04:00
{
2008-08-21 03:36:11 +04:00
seq_puts ( m , " # TASK-PID CPU# TIMESTAMP FUNCTION \n " ) ;
seq_puts ( m , " # | | | | | \n " ) ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:42 +04:00
print_trace_header ( struct seq_file * m , struct trace_iterator * iter )
{
unsigned long sym_flags = ( trace_flags & TRACE_ITER_SYM_MASK ) ;
struct trace_array * tr = iter - > tr ;
struct trace_array_cpu * data = tr - > data [ tr - > cpu ] ;
struct tracer * type = current_trace ;
2008-09-30 07:02:41 +04:00
unsigned long total ;
unsigned long entries ;
2008-05-12 23:20:42 +04:00
const char * name = " preemption " ;
if ( type )
name = type - > name ;
2008-09-30 07:02:41 +04:00
entries = ring_buffer_entries ( iter - > tr - > buffer ) ;
total = entries +
ring_buffer_overruns ( iter - > tr - > buffer ) ;
2008-05-12 23:20:42 +04:00
seq_printf ( m , " %s latency trace v1.1.5 on %s \n " ,
name , UTS_RELEASE ) ;
seq_puts ( m , " ----------------------------------- "
" --------------------------------- \n " ) ;
seq_printf ( m , " latency: %lu us, #%lu/%lu, CPU#%d | "
" (M:%s VP:%d, KP:%d, SP:%d HP:%d " ,
2008-05-12 23:20:44 +04:00
nsecs_to_usecs ( data - > saved_latency ) ,
2008-05-12 23:20:42 +04:00
entries ,
2008-05-12 23:20:43 +04:00
total ,
2008-05-12 23:20:42 +04:00
tr - > cpu ,
# if defined(CONFIG_PREEMPT_NONE)
" server " ,
# elif defined(CONFIG_PREEMPT_VOLUNTARY)
" desktop " ,
2008-07-11 04:58:12 +04:00
# elif defined(CONFIG_PREEMPT)
2008-05-12 23:20:42 +04:00
" preempt " ,
# else
" unknown " ,
# endif
/* These are reserved for later use */
0 , 0 , 0 , 0 ) ;
# ifdef CONFIG_SMP
seq_printf ( m , " #P:%d) \n " , num_online_cpus ( ) ) ;
# else
seq_puts ( m , " ) \n " ) ;
# endif
seq_puts ( m , " ----------------- \n " ) ;
seq_printf ( m , " | task: %.16s-%d "
" (uid:%d nice:%ld policy:%ld rt_prio:%ld) \n " ,
data - > comm , data - > pid , data - > uid , data - > nice ,
data - > policy , data - > rt_priority ) ;
seq_puts ( m , " ----------------- \n " ) ;
if ( data - > critical_start ) {
seq_puts ( m , " => started at: " ) ;
2008-05-12 23:20:46 +04:00
seq_print_ip_sym ( & iter - > seq , data - > critical_start , sym_flags ) ;
trace_print_seq ( m , & iter - > seq ) ;
2008-05-12 23:20:42 +04:00
seq_puts ( m , " \n => ended at: " ) ;
2008-05-12 23:20:46 +04:00
seq_print_ip_sym ( & iter - > seq , data - > critical_end , sym_flags ) ;
trace_print_seq ( m , & iter - > seq ) ;
2008-05-12 23:20:42 +04:00
seq_puts ( m , " \n " ) ;
}
seq_puts ( m , " \n " ) ;
}
2008-05-12 23:20:51 +04:00
static void
2008-05-12 23:20:46 +04:00
lat_print_generic ( struct trace_seq * s , struct trace_entry * entry , int cpu )
2008-05-12 23:20:42 +04:00
{
int hardirq , softirq ;
char * comm ;
2008-09-30 07:02:42 +04:00
comm = trace_find_cmdline ( entry - > pid ) ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:42 +04:00
trace_seq_printf ( s , " %8.8s-%-5d " , comm , entry - > pid ) ;
2008-08-21 03:36:11 +04:00
trace_seq_printf ( s , " %3d " , cpu ) ;
2008-05-12 23:20:46 +04:00
trace_seq_printf ( s , " %c%c " ,
2008-10-24 17:42:59 +04:00
( entry - > flags & TRACE_FLAG_IRQS_OFF ) ? ' d ' :
( entry - > flags & TRACE_FLAG_IRQS_NOSUPPORT ) ? ' X ' : ' . ' ,
2008-09-30 07:02:42 +04:00
( ( entry - > flags & TRACE_FLAG_NEED_RESCHED ) ? ' N ' : ' . ' ) ) ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:42 +04:00
hardirq = entry - > flags & TRACE_FLAG_HARDIRQ ;
softirq = entry - > flags & TRACE_FLAG_SOFTIRQ ;
2008-05-12 23:21:00 +04:00
if ( hardirq & & softirq ) {
2008-05-12 23:20:46 +04:00
trace_seq_putc ( s , ' H ' ) ;
2008-05-12 23:21:00 +04:00
} else {
if ( hardirq ) {
2008-05-12 23:20:46 +04:00
trace_seq_putc ( s , ' h ' ) ;
2008-05-12 23:21:00 +04:00
} else {
2008-05-12 23:20:42 +04:00
if ( softirq )
2008-05-12 23:20:46 +04:00
trace_seq_putc ( s , ' s ' ) ;
2008-05-12 23:20:42 +04:00
else
2008-05-12 23:20:46 +04:00
trace_seq_putc ( s , ' . ' ) ;
2008-05-12 23:20:42 +04:00
}
}
2008-09-30 07:02:42 +04:00
if ( entry - > preempt_count )
trace_seq_printf ( s , " %x " , entry - > preempt_count ) ;
2008-05-12 23:20:42 +04:00
else
2008-05-12 23:20:46 +04:00
trace_seq_puts ( s , " . " ) ;
2008-05-12 23:20:42 +04:00
}
unsigned long preempt_mark_thresh = 100 ;
2008-05-12 23:20:51 +04:00
static void
2008-09-30 07:02:41 +04:00
lat_print_timestamp ( struct trace_seq * s , u64 abs_usecs ,
2008-05-12 23:20:42 +04:00
unsigned long rel_usecs )
{
2008-05-12 23:20:46 +04:00
trace_seq_printf ( s , " %4lldus " , abs_usecs ) ;
2008-05-12 23:20:42 +04:00
if ( rel_usecs > preempt_mark_thresh )
2008-05-12 23:20:46 +04:00
trace_seq_puts ( s , " !: " ) ;
2008-05-12 23:20:42 +04:00
else if ( rel_usecs > 1 )
2008-05-12 23:20:46 +04:00
trace_seq_puts ( s , " +: " ) ;
2008-05-12 23:20:42 +04:00
else
2008-05-12 23:20:46 +04:00
trace_seq_puts ( s , " : " ) ;
2008-05-12 23:20:42 +04:00
}
static const char state_to_char [ ] = TASK_STATE_TO_CHAR_STR ;
2008-09-16 23:02:27 +04:00
/*
* The message is supposed to contain an ending newline .
* If the printing stops prematurely , try to add a newline of our own .
*/
void trace_seq_print_cont ( struct trace_seq * s , struct trace_iterator * iter )
2008-08-01 20:26:41 +04:00
{
struct trace_entry * ent ;
2008-09-30 07:02:42 +04:00
struct trace_field_cont * cont ;
2008-09-16 23:02:27 +04:00
bool ok = true ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
ent = peek_next_entry ( iter , iter - > cpu , NULL ) ;
2008-08-01 20:26:41 +04:00
if ( ! ent | | ent - > type ! = TRACE_CONT ) {
trace_seq_putc ( s , ' \n ' ) ;
return ;
}
do {
2008-09-30 07:02:42 +04:00
cont = ( struct trace_field_cont * ) ent ;
2008-09-16 23:02:27 +04:00
if ( ok )
2008-09-30 07:02:42 +04:00
ok = ( trace_seq_printf ( s , " %s " , cont - > buf ) > 0 ) ;
2008-10-01 08:29:53 +04:00
ftrace_disable_cpu ( ) ;
if ( iter - > buffer_iter [ iter - > cpu ] )
ring_buffer_read ( iter - > buffer_iter [ iter - > cpu ] , NULL ) ;
else
ring_buffer_consume ( iter - > tr - > buffer , iter - > cpu , NULL ) ;
ftrace_enable_cpu ( ) ;
2008-09-30 07:02:41 +04:00
ent = peek_next_entry ( iter , iter - > cpu , NULL ) ;
2008-08-01 20:26:41 +04:00
} while ( ent & & ent - > type = = TRACE_CONT ) ;
2008-09-16 23:02:27 +04:00
if ( ! ok )
trace_seq_putc ( s , ' \n ' ) ;
2008-08-01 20:26:41 +04:00
}
2008-11-08 06:36:02 +03:00
static void test_cpu_buff_start ( struct trace_iterator * iter )
{
struct trace_seq * s = & iter - > seq ;
if ( cpu_isset ( iter - > cpu , iter - > started ) )
return ;
cpu_set ( iter - > cpu , iter - > started ) ;
trace_seq_printf ( s , " ##### CPU %u buffer started #### \n " , iter - > cpu ) ;
}
2008-09-29 22:18:34 +04:00
static enum print_line_t
2008-05-12 23:20:46 +04:00
print_lat_fmt ( struct trace_iterator * iter , unsigned int trace_idx , int cpu )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:46 +04:00
struct trace_seq * s = & iter - > seq ;
2008-05-12 23:20:42 +04:00
unsigned long sym_flags = ( trace_flags & TRACE_ITER_SYM_MASK ) ;
2008-09-30 07:02:41 +04:00
struct trace_entry * next_entry ;
2008-05-12 23:20:42 +04:00
unsigned long verbose = ( trace_flags & TRACE_ITER_VERBOSE ) ;
struct trace_entry * entry = iter - > ent ;
unsigned long abs_usecs ;
unsigned long rel_usecs ;
2008-09-30 07:02:41 +04:00
u64 next_ts ;
2008-05-12 23:20:42 +04:00
char * comm ;
2008-05-12 23:20:53 +04:00
int S , T ;
2008-05-12 23:20:51 +04:00
int i ;
2008-05-12 23:20:58 +04:00
unsigned state ;
2008-05-12 23:20:42 +04:00
2008-08-01 20:26:41 +04:00
if ( entry - > type = = TRACE_CONT )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-08-01 20:26:41 +04:00
2008-11-08 06:36:02 +03:00
test_cpu_buff_start ( iter ) ;
2008-09-30 07:02:41 +04:00
next_entry = find_next_entry ( iter , NULL , & next_ts ) ;
if ( ! next_entry )
next_ts = iter - > ts ;
rel_usecs = ns2usecs ( next_ts - iter - > ts ) ;
abs_usecs = ns2usecs ( iter - > ts - iter - > tr - > time_start ) ;
2008-05-12 23:20:42 +04:00
if ( verbose ) {
2008-09-30 07:02:42 +04:00
comm = trace_find_cmdline ( entry - > pid ) ;
2008-08-21 03:36:11 +04:00
trace_seq_printf ( s , " %16s %5d %3d %d %08x %08x [%08lx] "
2008-05-12 23:20:46 +04:00
" %ld.%03ldms (+%ld.%03ldms): " ,
comm ,
2008-09-30 07:02:42 +04:00
entry - > pid , cpu , entry - > flags ,
entry - > preempt_count , trace_idx ,
2008-09-30 07:02:41 +04:00
ns2usecs ( iter - > ts ) ,
2008-05-12 23:20:46 +04:00
abs_usecs / 1000 ,
abs_usecs % 1000 , rel_usecs / 1000 ,
rel_usecs % 1000 ) ;
2008-05-12 23:20:42 +04:00
} else {
2008-05-12 23:20:53 +04:00
lat_print_generic ( s , entry , cpu ) ;
lat_print_timestamp ( s , abs_usecs , rel_usecs ) ;
2008-05-12 23:20:42 +04:00
}
switch ( entry - > type ) {
2008-09-30 07:02:42 +04:00
case TRACE_FN : {
2008-10-01 18:52:51 +04:00
struct ftrace_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
seq_print_ip_sym ( s , field - > ip , sym_flags ) ;
2008-05-12 23:20:46 +04:00
trace_seq_puts ( s , " ( " ) ;
2008-10-31 22:44:07 +03:00
seq_print_ip_sym ( s , field - > parent_ip , sym_flags ) ;
2008-05-12 23:20:46 +04:00
trace_seq_puts ( s , " ) \n " ) ;
2008-05-12 23:20:42 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:42 +04:00
case TRACE_CTX :
2008-09-30 07:02:42 +04:00
case TRACE_WAKE : {
2008-10-01 18:52:51 +04:00
struct ctx_switch_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
T = field - > next_state < sizeof ( state_to_char ) ?
state_to_char [ field - > next_state ] : ' X ' ;
2008-05-12 23:20:53 +04:00
2008-09-30 07:02:42 +04:00
state = field - > prev_state ?
__ffs ( field - > prev_state ) + 1 : 0 ;
2008-05-12 23:20:58 +04:00
S = state < sizeof ( state_to_char ) - 1 ? state_to_char [ state ] : ' X ' ;
2008-09-30 07:02:42 +04:00
comm = trace_find_cmdline ( field - > next_pid ) ;
2008-09-04 12:24:16 +04:00
trace_seq_printf ( s , " %5d:%3d:%c %s [%03d] %5d:%3d:%c %s \n " ,
2008-09-30 07:02:42 +04:00
field - > prev_pid ,
field - > prev_prio ,
2008-05-12 23:20:51 +04:00
S , entry - > type = = TRACE_CTX ? " ==> " : " + " ,
2008-09-30 07:02:42 +04:00
field - > next_cpu ,
field - > next_pid ,
field - > next_prio ,
2008-05-12 23:20:53 +04:00
T , comm ) ;
2008-05-12 23:20:42 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_SPECIAL : {
2008-10-01 18:52:51 +04:00
struct special_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
2008-05-12 23:20:53 +04:00
trace_seq_printf ( s , " # %ld %ld %ld \n " ,
2008-09-30 07:02:42 +04:00
field - > arg1 ,
field - > arg2 ,
field - > arg3 ) ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_STACK : {
2008-10-01 18:52:51 +04:00
struct stack_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
2008-05-12 23:20:51 +04:00
for ( i = 0 ; i < FTRACE_STACK_ENTRIES ; i + + ) {
if ( i )
trace_seq_puts ( s , " <= " ) ;
2008-09-30 07:02:42 +04:00
seq_print_ip_sym ( s , field - > caller [ i ] , sym_flags ) ;
2008-05-12 23:20:51 +04:00
}
trace_seq_puts ( s , " \n " ) ;
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_PRINT : {
2008-10-01 18:52:51 +04:00
struct print_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
seq_print_ip_sym ( s , field - > ip , sym_flags ) ;
trace_seq_printf ( s , " : %s " , field - > buf ) ;
if ( entry - > flags & TRACE_FLAG_CONT )
2008-08-01 20:26:41 +04:00
trace_seq_print_cont ( s , iter ) ;
break ;
2008-09-30 07:02:42 +04:00
}
2008-11-12 23:24:24 +03:00
case TRACE_BRANCH : {
struct trace_branch * field ;
2008-11-12 08:14:40 +03:00
trace_assign_type ( field , entry ) ;
trace_seq_printf ( s , " [%s] %s:%s:%d \n " ,
2008-11-12 16:09:30 +03:00
field - > correct ? " ok " : " MISS " ,
2008-11-12 08:14:40 +03:00
field - > func ,
field - > file ,
field - > line ) ;
break ;
}
2008-05-12 23:20:44 +04:00
default :
2008-05-12 23:20:46 +04:00
trace_seq_printf ( s , " Unknown type %d \n " , entry - > type ) ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_trace_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:46 +04:00
struct trace_seq * s = & iter - > seq ;
2008-05-12 23:20:42 +04:00
unsigned long sym_flags = ( trace_flags & TRACE_ITER_SYM_MASK ) ;
2008-05-12 23:20:45 +04:00
struct trace_entry * entry ;
2008-05-12 23:20:42 +04:00
unsigned long usec_rem ;
unsigned long long t ;
unsigned long secs ;
char * comm ;
2008-05-12 23:20:46 +04:00
int ret ;
2008-05-12 23:20:53 +04:00
int S , T ;
2008-05-12 23:20:51 +04:00
int i ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:45 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
if ( entry - > type = = TRACE_CONT )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-08-01 20:26:41 +04:00
2008-11-08 06:36:02 +03:00
test_cpu_buff_start ( iter ) ;
2008-09-30 07:02:42 +04:00
comm = trace_find_cmdline ( iter - > ent - > pid ) ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
t = ns2usecs ( iter - > ts ) ;
2008-05-12 23:20:42 +04:00
usec_rem = do_div ( t , 1000000ULL ) ;
secs = ( unsigned long ) t ;
2008-09-30 07:02:42 +04:00
ret = trace_seq_printf ( s , " %16s-%-5d " , comm , entry - > pid ) ;
2008-05-12 23:20:53 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-08-21 03:36:11 +04:00
ret = trace_seq_printf ( s , " [%03d] " , iter - > cpu ) ;
2008-05-12 23:20:53 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:53 +04:00
ret = trace_seq_printf ( s , " %5lu.%06lu: " , secs , usec_rem ) ;
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
switch ( entry - > type ) {
2008-09-30 07:02:42 +04:00
case TRACE_FN : {
2008-10-01 18:52:51 +04:00
struct ftrace_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
ret = seq_print_ip_sym ( s , field - > ip , sym_flags ) ;
2008-05-12 23:20:46 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
if ( ( sym_flags & TRACE_ITER_PRINT_PARENT ) & &
2008-09-30 07:02:42 +04:00
field - > parent_ip ) {
2008-05-12 23:20:46 +04:00
ret = trace_seq_printf ( s , " <- " ) ;
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-10-31 22:44:07 +03:00
ret = seq_print_ip_sym ( s ,
field - > parent_ip ,
sym_flags ) ;
2008-05-12 23:20:46 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
}
2008-05-12 23:20:46 +04:00
ret = trace_seq_printf ( s , " \n " ) ;
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:42 +04:00
case TRACE_CTX :
2008-09-30 07:02:42 +04:00
case TRACE_WAKE : {
2008-10-01 18:52:51 +04:00
struct ctx_switch_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
S = field - > prev_state < sizeof ( state_to_char ) ?
state_to_char [ field - > prev_state ] : ' X ' ;
T = field - > next_state < sizeof ( state_to_char ) ?
state_to_char [ field - > next_state ] : ' X ' ;
2008-09-04 12:24:16 +04:00
ret = trace_seq_printf ( s , " %5d:%3d:%c %s [%03d] %5d:%3d:%c \n " ,
2008-09-30 07:02:42 +04:00
field - > prev_pid ,
field - > prev_prio ,
2008-05-12 23:20:46 +04:00
S ,
2008-05-12 23:20:51 +04:00
entry - > type = = TRACE_CTX ? " ==> " : " + " ,
2008-09-30 07:02:42 +04:00
field - > next_cpu ,
field - > next_pid ,
field - > next_prio ,
2008-05-12 23:20:53 +04:00
T ) ;
2008-05-12 23:20:46 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:42 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_SPECIAL : {
2008-10-01 18:52:51 +04:00
struct special_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
2008-05-12 23:20:53 +04:00
ret = trace_seq_printf ( s , " # %ld %ld %ld \n " ,
2008-09-30 07:02:42 +04:00
field - > arg1 ,
field - > arg2 ,
field - > arg3 ) ;
2008-05-12 23:20:47 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_STACK : {
2008-10-01 18:52:51 +04:00
struct stack_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
2008-05-12 23:20:51 +04:00
for ( i = 0 ; i < FTRACE_STACK_ENTRIES ; i + + ) {
if ( i ) {
ret = trace_seq_puts ( s , " <= " ) ;
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:51 +04:00
}
2008-09-30 07:02:42 +04:00
ret = seq_print_ip_sym ( s , field - > caller [ i ] ,
2008-05-12 23:20:51 +04:00
sym_flags ) ;
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:51 +04:00
}
ret = trace_seq_puts ( s , " \n " ) ;
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:51 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_PRINT : {
2008-10-01 18:52:51 +04:00
struct print_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
seq_print_ip_sym ( s , field - > ip , sym_flags ) ;
trace_seq_printf ( s , " : %s " , field - > buf ) ;
if ( entry - > flags & TRACE_FLAG_CONT )
2008-08-01 20:26:41 +04:00
trace_seq_print_cont ( s , iter ) ;
break ;
2008-05-12 23:20:42 +04:00
}
2008-11-11 09:14:25 +03:00
case TRACE_FN_RET : {
return print_return_function ( iter ) ;
break ;
}
2008-11-12 23:24:24 +03:00
case TRACE_BRANCH : {
struct trace_branch * field ;
2008-11-12 08:14:40 +03:00
trace_assign_type ( field , entry ) ;
trace_seq_printf ( s , " [%s] %s:%s:%d \n " ,
2008-11-12 16:09:30 +03:00
field - > correct ? " ok " : " MISS " ,
2008-11-12 08:14:40 +03:00
field - > func ,
field - > file ,
field - > line ) ;
break ;
}
2008-09-30 07:02:42 +04:00
}
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_raw_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:47 +04:00
{
struct trace_seq * s = & iter - > seq ;
struct trace_entry * entry ;
int ret ;
2008-05-12 23:20:53 +04:00
int S , T ;
2008-05-12 23:20:47 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
if ( entry - > type = = TRACE_CONT )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-08-01 20:26:41 +04:00
2008-05-12 23:20:47 +04:00
ret = trace_seq_printf ( s , " %d %d %llu " ,
2008-09-30 07:02:42 +04:00
entry - > pid , iter - > cpu , iter - > ts ) ;
2008-05-12 23:20:47 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:47 +04:00
switch ( entry - > type ) {
2008-09-30 07:02:42 +04:00
case TRACE_FN : {
2008-10-01 18:52:51 +04:00
struct ftrace_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
2008-05-12 23:20:47 +04:00
ret = trace_seq_printf ( s , " %x %x \n " ,
2008-09-30 07:02:42 +04:00
field - > ip ,
field - > parent_ip ) ;
2008-05-12 23:20:47 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:47 +04:00
case TRACE_CTX :
2008-09-30 07:02:42 +04:00
case TRACE_WAKE : {
2008-10-01 18:52:51 +04:00
struct ctx_switch_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
S = field - > prev_state < sizeof ( state_to_char ) ?
state_to_char [ field - > prev_state ] : ' X ' ;
T = field - > next_state < sizeof ( state_to_char ) ?
state_to_char [ field - > next_state ] : ' X ' ;
2008-05-12 23:20:51 +04:00
if ( entry - > type = = TRACE_WAKE )
S = ' + ' ;
2008-09-04 12:24:16 +04:00
ret = trace_seq_printf ( s , " %d %d %c %d %d %d %c \n " ,
2008-09-30 07:02:42 +04:00
field - > prev_pid ,
field - > prev_prio ,
2008-05-12 23:20:47 +04:00
S ,
2008-09-30 07:02:42 +04:00
field - > next_cpu ,
field - > next_pid ,
field - > next_prio ,
2008-05-12 23:20:53 +04:00
T ) ;
2008-05-12 23:20:47 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:47 +04:00
case TRACE_SPECIAL :
2008-09-30 07:02:42 +04:00
case TRACE_STACK : {
2008-10-01 18:52:51 +04:00
struct special_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
2008-05-12 23:20:53 +04:00
ret = trace_seq_printf ( s , " # %ld %ld %ld \n " ,
2008-09-30 07:02:42 +04:00
field - > arg1 ,
field - > arg2 ,
field - > arg3 ) ;
2008-05-12 23:20:47 +04:00
if ( ! ret )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_PARTIAL_LINE ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_PRINT : {
2008-10-01 18:52:51 +04:00
struct print_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
trace_seq_printf ( s , " # %lx %s " , field - > ip , field - > buf ) ;
if ( entry - > flags & TRACE_FLAG_CONT )
2008-08-01 20:26:41 +04:00
trace_seq_print_cont ( s , iter ) ;
break ;
2008-05-12 23:20:47 +04:00
}
2008-09-30 07:02:42 +04:00
}
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-05-12 23:20:47 +04:00
}
2008-05-12 23:20:47 +04:00
# define SEQ_PUT_FIELD_RET(s, x) \
do { \
if ( ! trace_seq_putmem ( s , & ( x ) , sizeof ( x ) ) ) \
return 0 ; \
} while ( 0 )
2008-05-12 23:20:49 +04:00
# define SEQ_PUT_HEX_FIELD_RET(s, x) \
do { \
2008-10-09 04:44:55 +04:00
BUILD_BUG_ON ( sizeof ( x ) > MAX_MEMHEX_BYTES ) ; \
2008-05-12 23:20:49 +04:00
if ( ! trace_seq_putmem_hex ( s , & ( x ) , sizeof ( x ) ) ) \
return 0 ; \
} while ( 0 )
2008-09-29 22:18:34 +04:00
static enum print_line_t print_hex_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:49 +04:00
{
struct trace_seq * s = & iter - > seq ;
unsigned char newline = ' \n ' ;
struct trace_entry * entry ;
2008-05-12 23:20:53 +04:00
int S , T ;
2008-05-12 23:20:49 +04:00
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
if ( entry - > type = = TRACE_CONT )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:42 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , entry - > pid ) ;
2008-05-12 23:20:49 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , iter - > cpu ) ;
2008-09-30 07:02:41 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , iter - > ts ) ;
2008-05-12 23:20:49 +04:00
switch ( entry - > type ) {
2008-09-30 07:02:42 +04:00
case TRACE_FN : {
2008-10-01 18:52:51 +04:00
struct ftrace_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , field - > ip ) ;
SEQ_PUT_HEX_FIELD_RET ( s , field - > parent_ip ) ;
2008-05-12 23:20:49 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:49 +04:00
case TRACE_CTX :
2008-09-30 07:02:42 +04:00
case TRACE_WAKE : {
2008-10-01 18:52:51 +04:00
struct ctx_switch_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
S = field - > prev_state < sizeof ( state_to_char ) ?
state_to_char [ field - > prev_state ] : ' X ' ;
T = field - > next_state < sizeof ( state_to_char ) ?
state_to_char [ field - > next_state ] : ' X ' ;
2008-05-12 23:20:51 +04:00
if ( entry - > type = = TRACE_WAKE )
S = ' + ' ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , field - > prev_pid ) ;
SEQ_PUT_HEX_FIELD_RET ( s , field - > prev_prio ) ;
2008-05-12 23:20:49 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , S ) ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , field - > next_cpu ) ;
SEQ_PUT_HEX_FIELD_RET ( s , field - > next_pid ) ;
SEQ_PUT_HEX_FIELD_RET ( s , field - > next_prio ) ;
2008-05-12 23:20:53 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , T ) ;
2008-05-12 23:20:49 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:49 +04:00
case TRACE_SPECIAL :
2008-09-30 07:02:42 +04:00
case TRACE_STACK : {
2008-10-01 18:52:51 +04:00
struct special_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_HEX_FIELD_RET ( s , field - > arg1 ) ;
SEQ_PUT_HEX_FIELD_RET ( s , field - > arg2 ) ;
SEQ_PUT_HEX_FIELD_RET ( s , field - > arg3 ) ;
2008-05-12 23:20:49 +04:00
break ;
}
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:49 +04:00
SEQ_PUT_FIELD_RET ( s , newline ) ;
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-05-12 23:20:49 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_bin_fmt ( struct trace_iterator * iter )
2008-05-12 23:20:47 +04:00
{
struct trace_seq * s = & iter - > seq ;
struct trace_entry * entry ;
entry = iter - > ent ;
2008-08-01 20:26:41 +04:00
if ( entry - > type = = TRACE_CONT )
2008-09-29 22:18:34 +04:00
return TRACE_TYPE_HANDLED ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:42 +04:00
SEQ_PUT_FIELD_RET ( s , entry - > pid ) ;
2008-10-27 01:26:57 +03:00
SEQ_PUT_FIELD_RET ( s , entry - > cpu ) ;
2008-09-30 07:02:41 +04:00
SEQ_PUT_FIELD_RET ( s , iter - > ts ) ;
2008-05-12 23:20:47 +04:00
switch ( entry - > type ) {
2008-09-30 07:02:42 +04:00
case TRACE_FN : {
2008-10-01 18:52:51 +04:00
struct ftrace_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_FIELD_RET ( s , field - > ip ) ;
SEQ_PUT_FIELD_RET ( s , field - > parent_ip ) ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
case TRACE_CTX : {
2008-10-01 18:52:51 +04:00
struct ctx_switch_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_FIELD_RET ( s , field - > prev_pid ) ;
SEQ_PUT_FIELD_RET ( s , field - > prev_prio ) ;
SEQ_PUT_FIELD_RET ( s , field - > prev_state ) ;
SEQ_PUT_FIELD_RET ( s , field - > next_pid ) ;
SEQ_PUT_FIELD_RET ( s , field - > next_prio ) ;
SEQ_PUT_FIELD_RET ( s , field - > next_state ) ;
2008-05-12 23:20:47 +04:00
break ;
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:47 +04:00
case TRACE_SPECIAL :
2008-09-30 07:02:42 +04:00
case TRACE_STACK : {
2008-10-01 18:52:51 +04:00
struct special_entry * field ;
trace_assign_type ( field , entry ) ;
2008-09-30 07:02:42 +04:00
SEQ_PUT_FIELD_RET ( s , field - > arg1 ) ;
SEQ_PUT_FIELD_RET ( s , field - > arg2 ) ;
SEQ_PUT_FIELD_RET ( s , field - > arg3 ) ;
2008-05-12 23:20:47 +04:00
break ;
2008-05-12 23:20:47 +04:00
}
2008-09-30 07:02:42 +04:00
}
2008-05-12 23:20:47 +04:00
return 1 ;
}
2008-05-12 23:20:42 +04:00
static int trace_empty ( struct trace_iterator * iter )
{
int cpu ;
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( cpu ) {
2008-10-01 08:29:53 +04:00
if ( iter - > buffer_iter [ cpu ] ) {
if ( ! ring_buffer_iter_empty ( iter - > buffer_iter [ cpu ] ) )
return 0 ;
} else {
if ( ! ring_buffer_empty_cpu ( iter - > tr - > buffer , cpu ) )
return 0 ;
}
2008-05-12 23:20:42 +04:00
}
2008-10-01 08:29:53 +04:00
2008-09-30 20:13:45 +04:00
return 1 ;
2008-05-12 23:20:42 +04:00
}
2008-09-29 22:18:34 +04:00
static enum print_line_t print_trace_line ( struct trace_iterator * iter )
2008-05-12 23:20:47 +04:00
{
2008-09-29 22:18:34 +04:00
enum print_line_t ret ;
if ( iter - > trace & & iter - > trace - > print_line ) {
ret = iter - > trace - > print_line ( iter ) ;
if ( ret ! = TRACE_TYPE_UNHANDLED )
return ret ;
}
2008-05-23 23:37:28 +04:00
2008-05-12 23:20:47 +04:00
if ( trace_flags & TRACE_ITER_BIN )
return print_bin_fmt ( iter ) ;
2008-05-12 23:20:49 +04:00
if ( trace_flags & TRACE_ITER_HEX )
return print_hex_fmt ( iter ) ;
2008-05-12 23:20:47 +04:00
if ( trace_flags & TRACE_ITER_RAW )
return print_raw_fmt ( iter ) ;
if ( iter - > iter_flags & TRACE_FILE_LAT_FMT )
return print_lat_fmt ( iter , iter - > idx , iter - > cpu ) ;
return print_trace_fmt ( iter ) ;
}
2008-05-12 23:20:42 +04:00
static int s_show ( struct seq_file * m , void * v )
{
struct trace_iterator * iter = v ;
if ( iter - > ent = = NULL ) {
if ( iter - > tr ) {
seq_printf ( m , " # tracer: %s \n " , iter - > trace - > name ) ;
seq_puts ( m , " # \n " ) ;
}
if ( iter - > iter_flags & TRACE_FILE_LAT_FMT ) {
/* print nothing if the buffers are empty */
if ( trace_empty ( iter ) )
return 0 ;
print_trace_header ( m , iter ) ;
if ( ! ( trace_flags & TRACE_ITER_VERBOSE ) )
print_lat_help_header ( m ) ;
} else {
if ( ! ( trace_flags & TRACE_ITER_VERBOSE ) )
print_func_help_header ( m ) ;
}
} else {
2008-05-12 23:20:47 +04:00
print_trace_line ( iter ) ;
2008-05-12 23:20:46 +04:00
trace_print_seq ( m , & iter - > seq ) ;
2008-05-12 23:20:42 +04:00
}
return 0 ;
}
static struct seq_operations tracer_seq_ops = {
2008-05-12 23:20:46 +04:00
. start = s_start ,
. next = s_next ,
. stop = s_stop ,
. show = s_show ,
2008-05-12 23:20:42 +04:00
} ;
2008-05-12 23:20:51 +04:00
static struct trace_iterator *
2008-05-12 23:20:42 +04:00
__tracing_open ( struct inode * inode , struct file * file , int * ret )
{
struct trace_iterator * iter ;
2008-09-30 07:02:41 +04:00
struct seq_file * m ;
int cpu ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:44 +04:00
if ( tracing_disabled ) {
* ret = - ENODEV ;
return NULL ;
}
2008-05-12 23:20:42 +04:00
iter = kzalloc ( sizeof ( * iter ) , GFP_KERNEL ) ;
if ( ! iter ) {
* ret = - ENOMEM ;
goto out ;
}
mutex_lock ( & trace_types_lock ) ;
if ( current_trace & & current_trace - > print_max )
iter - > tr = & max_tr ;
else
iter - > tr = inode - > i_private ;
iter - > trace = current_trace ;
iter - > pos = - 1 ;
2008-09-30 07:02:41 +04:00
for_each_tracing_cpu ( cpu ) {
2008-10-01 08:29:53 +04:00
2008-09-30 07:02:41 +04:00
iter - > buffer_iter [ cpu ] =
ring_buffer_read_start ( iter - > tr - > buffer , cpu ) ;
2008-10-01 08:29:53 +04:00
2008-09-30 07:02:41 +04:00
if ( ! iter - > buffer_iter [ cpu ] )
goto fail_buffer ;
}
2008-05-12 23:20:42 +04:00
/* TODO stop tracer */
* ret = seq_open ( file , & tracer_seq_ops ) ;
2008-09-30 07:02:41 +04:00
if ( * ret )
goto fail_buffer ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
m = file - > private_data ;
m - > private = iter ;
2008-05-12 23:20:42 +04:00
2008-09-30 07:02:41 +04:00
/* stop the trace while dumping */
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
tracing_stop ( ) ;
2008-09-30 07:02:41 +04:00
if ( iter - > trace & & iter - > trace - > open )
iter - > trace - > open ( iter ) ;
2008-05-12 23:20:42 +04:00
mutex_unlock ( & trace_types_lock ) ;
out :
return iter ;
2008-09-30 07:02:41 +04:00
fail_buffer :
for_each_tracing_cpu ( cpu ) {
if ( iter - > buffer_iter [ cpu ] )
ring_buffer_read_finish ( iter - > buffer_iter [ cpu ] ) ;
}
mutex_unlock ( & trace_types_lock ) ;
return ERR_PTR ( - ENOMEM ) ;
2008-05-12 23:20:42 +04:00
}
int tracing_open_generic ( struct inode * inode , struct file * filp )
{
2008-05-12 23:20:44 +04:00
if ( tracing_disabled )
return - ENODEV ;
2008-05-12 23:20:42 +04:00
filp - > private_data = inode - > i_private ;
return 0 ;
}
int tracing_release ( struct inode * inode , struct file * file )
{
struct seq_file * m = ( struct seq_file * ) file - > private_data ;
struct trace_iterator * iter = m - > private ;
2008-09-30 07:02:41 +04:00
int cpu ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-09-30 07:02:41 +04:00
for_each_tracing_cpu ( cpu ) {
if ( iter - > buffer_iter [ cpu ] )
ring_buffer_read_finish ( iter - > buffer_iter [ cpu ] ) ;
}
2008-05-12 23:20:42 +04:00
if ( iter - > trace & & iter - > trace - > close )
iter - > trace - > close ( iter ) ;
/* reenable tracing if it was previously enabled */
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
tracing_start ( ) ;
2008-05-12 23:20:42 +04:00
mutex_unlock ( & trace_types_lock ) ;
seq_release ( inode , file ) ;
kfree ( iter ) ;
return 0 ;
}
static int tracing_open ( struct inode * inode , struct file * file )
{
int ret ;
__tracing_open ( inode , file , & ret ) ;
return ret ;
}
static int tracing_lt_open ( struct inode * inode , struct file * file )
{
struct trace_iterator * iter ;
int ret ;
iter = __tracing_open ( inode , file , & ret ) ;
if ( ! ret )
iter - > iter_flags | = TRACE_FILE_LAT_FMT ;
return ret ;
}
2008-05-12 23:20:51 +04:00
static void *
2008-05-12 23:20:42 +04:00
t_next ( struct seq_file * m , void * v , loff_t * pos )
{
struct tracer * t = m - > private ;
( * pos ) + + ;
if ( t )
t = t - > next ;
m - > private = t ;
return t ;
}
static void * t_start ( struct seq_file * m , loff_t * pos )
{
struct tracer * t = m - > private ;
loff_t l = 0 ;
mutex_lock ( & trace_types_lock ) ;
for ( ; t & & l < * pos ; t = t_next ( m , t , & l ) )
;
return t ;
}
static void t_stop ( struct seq_file * m , void * p )
{
mutex_unlock ( & trace_types_lock ) ;
}
static int t_show ( struct seq_file * m , void * v )
{
struct tracer * t = v ;
if ( ! t )
return 0 ;
seq_printf ( m , " %s " , t - > name ) ;
if ( t - > next )
seq_putc ( m , ' ' ) ;
else
seq_putc ( m , ' \n ' ) ;
return 0 ;
}
static struct seq_operations show_traces_seq_ops = {
2008-05-12 23:20:46 +04:00
. start = t_start ,
. next = t_next ,
. stop = t_stop ,
. show = t_show ,
2008-05-12 23:20:42 +04:00
} ;
static int show_traces_open ( struct inode * inode , struct file * file )
{
int ret ;
2008-05-12 23:20:44 +04:00
if ( tracing_disabled )
return - ENODEV ;
2008-05-12 23:20:42 +04:00
ret = seq_open ( file , & show_traces_seq_ops ) ;
if ( ! ret ) {
struct seq_file * m = file - > private_data ;
m - > private = trace_types ;
}
return ret ;
}
static struct file_operations tracing_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open ,
. read = seq_read ,
. llseek = seq_lseek ,
. release = tracing_release ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations tracing_lt_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_lt_open ,
. read = seq_read ,
. llseek = seq_lseek ,
. release = tracing_release ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations show_traces_fops = {
2008-05-12 23:20:52 +04:00
. open = show_traces_open ,
. read = seq_read ,
. release = seq_release ,
} ;
2008-05-12 23:20:52 +04:00
/*
* Only trace on a CPU if the bitmask is set :
*/
static cpumask_t tracing_cpumask = CPU_MASK_ALL ;
/*
* When tracing / tracing_cpu_mask is modified then this holds
* the new bitmask we are about to install :
*/
static cpumask_t tracing_cpumask_new ;
/*
* The tracer itself will not take this lock , but still we want
* to provide a consistent cpumask to user - space :
*/
static DEFINE_MUTEX ( tracing_cpumask_update_lock ) ;
/*
* Temporary storage for the character representation of the
* CPU bitmask ( and one more byte for the newline ) :
*/
static char mask_str [ NR_CPUS + 1 ] ;
2008-05-12 23:20:52 +04:00
static ssize_t
tracing_cpumask_read ( struct file * filp , char __user * ubuf ,
size_t count , loff_t * ppos )
{
2008-05-12 23:20:52 +04:00
int len ;
2008-05-12 23:20:52 +04:00
mutex_lock ( & tracing_cpumask_update_lock ) ;
2008-05-12 23:20:52 +04:00
len = cpumask_scnprintf ( mask_str , count , tracing_cpumask ) ;
if ( count - len < 2 ) {
count = - EINVAL ;
goto out_err ;
}
len + = sprintf ( mask_str + len , " \n " ) ;
count = simple_read_from_buffer ( ubuf , count , ppos , mask_str , NR_CPUS + 1 ) ;
out_err :
2008-05-12 23:20:52 +04:00
mutex_unlock ( & tracing_cpumask_update_lock ) ;
return count ;
}
static ssize_t
tracing_cpumask_write ( struct file * filp , const char __user * ubuf ,
size_t count , loff_t * ppos )
{
2008-05-12 23:20:52 +04:00
int err , cpu ;
2008-05-12 23:20:52 +04:00
mutex_lock ( & tracing_cpumask_update_lock ) ;
2008-05-12 23:20:52 +04:00
err = cpumask_parse_user ( ubuf , count , tracing_cpumask_new ) ;
2008-05-12 23:20:52 +04:00
if ( err )
2008-05-12 23:20:52 +04:00
goto err_unlock ;
2008-05-12 23:20:55 +04:00
raw_local_irq_disable ( ) ;
__raw_spin_lock ( & ftrace_max_lock ) ;
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( cpu ) {
2008-05-12 23:20:52 +04:00
/*
* Increase / decrease the disabled counter if we are
* about to flip a bit in the cpumask :
*/
if ( cpu_isset ( cpu , tracing_cpumask ) & &
! cpu_isset ( cpu , tracing_cpumask_new ) ) {
atomic_inc ( & global_trace . data [ cpu ] - > disabled ) ;
}
if ( ! cpu_isset ( cpu , tracing_cpumask ) & &
cpu_isset ( cpu , tracing_cpumask_new ) ) {
atomic_dec ( & global_trace . data [ cpu ] - > disabled ) ;
}
}
2008-05-12 23:20:55 +04:00
__raw_spin_unlock ( & ftrace_max_lock ) ;
raw_local_irq_enable ( ) ;
2008-05-12 23:20:52 +04:00
tracing_cpumask = tracing_cpumask_new ;
mutex_unlock ( & tracing_cpumask_update_lock ) ;
2008-05-12 23:20:52 +04:00
return count ;
2008-05-12 23:20:52 +04:00
err_unlock :
mutex_unlock ( & tracing_cpumask_update_lock ) ;
return err ;
2008-05-12 23:20:52 +04:00
}
static struct file_operations tracing_cpumask_fops = {
. open = tracing_open_generic ,
. read = tracing_cpumask_read ,
. write = tracing_cpumask_write ,
2008-05-12 23:20:42 +04:00
} ;
static ssize_t
tracing_iter_ctrl_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char * buf ;
int r = 0 ;
int len = 0 ;
int i ;
/* calulate max size */
for ( i = 0 ; trace_options [ i ] ; i + + ) {
len + = strlen ( trace_options [ i ] ) ;
len + = 3 ; /* "no" and space */
}
/* +2 for \n and \0 */
buf = kmalloc ( len + 2 , GFP_KERNEL ) ;
if ( ! buf )
return - ENOMEM ;
for ( i = 0 ; trace_options [ i ] ; i + + ) {
if ( trace_flags & ( 1 < < i ) )
r + = sprintf ( buf + r , " %s " , trace_options [ i ] ) ;
else
r + = sprintf ( buf + r , " no%s " , trace_options [ i ] ) ;
}
r + = sprintf ( buf + r , " \n " ) ;
WARN_ON ( r > = len + 2 ) ;
2008-05-12 23:20:52 +04:00
r = simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
kfree ( buf ) ;
return r ;
}
static ssize_t
tracing_iter_ctrl_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ 64 ] ;
char * cmp = buf ;
int neg = 0 ;
int i ;
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
if ( strncmp ( buf , " no " , 2 ) = = 0 ) {
neg = 1 ;
cmp + = 2 ;
}
for ( i = 0 ; trace_options [ i ] ; i + + ) {
int len = strlen ( trace_options [ i ] ) ;
if ( strncmp ( cmp , trace_options [ i ] , len ) = = 0 ) {
if ( neg )
trace_flags & = ~ ( 1 < < i ) ;
else
trace_flags | = ( 1 < < i ) ;
break ;
}
}
2008-05-12 23:20:53 +04:00
/*
* If no option could be set , return an error :
*/
if ( ! trace_options [ i ] )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
filp - > f_pos + = cnt ;
return cnt ;
}
static struct file_operations tracing_iter_fops = {
2008-05-12 23:20:52 +04:00
. open = tracing_open_generic ,
. read = tracing_iter_ctrl_read ,
. write = tracing_iter_ctrl_write ,
2008-05-12 23:20:42 +04:00
} ;
2008-05-12 23:20:45 +04:00
static const char readme_msg [ ] =
" tracing mini-HOWTO: \n \n "
" # mkdir /debug \n "
" # mount -t debugfs nodev /debug \n \n "
" # cat /debug/tracing/available_tracers \n "
" wakeup preemptirqsoff preemptoff irqsoff ftrace sched_switch none \n \n "
" # cat /debug/tracing/current_tracer \n "
" none \n "
" # echo sched_switch > /debug/tracing/current_tracer \n "
" # cat /debug/tracing/current_tracer \n "
" sched_switch \n "
" # cat /debug/tracing/iter_ctrl \n "
" noprint-parent nosym-offset nosym-addr noverbose \n "
" # echo print-parent > /debug/tracing/iter_ctrl \n "
" # echo 1 > /debug/tracing/tracing_enabled \n "
" # cat /debug/tracing/trace > /tmp/trace.txt \n "
" echo 0 > /debug/tracing/tracing_enabled \n "
;
static ssize_t
tracing_readme_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
return simple_read_from_buffer ( ubuf , cnt , ppos ,
readme_msg , strlen ( readme_msg ) ) ;
}
static struct file_operations tracing_readme_fops = {
2008-05-12 23:20:52 +04:00
. open = tracing_open_generic ,
. read = tracing_readme_read ,
2008-05-12 23:20:45 +04:00
} ;
2008-05-12 23:20:42 +04:00
static ssize_t
tracing_ctrl_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ 64 ] ;
int r ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
r = sprintf ( buf , " %u \n " , tracer_enabled ) ;
2008-05-12 23:20:45 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
}
static ssize_t
tracing_ctrl_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
struct trace_array * tr = filp - > private_data ;
char buf [ 64 ] ;
2008-05-12 23:21:00 +04:00
long val ;
int ret ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
2008-05-12 23:21:00 +04:00
ret = strict_strtoul ( buf , 10 , & val ) ;
if ( ret < 0 )
return ret ;
2008-05-12 23:20:42 +04:00
val = ! ! val ;
mutex_lock ( & trace_types_lock ) ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
if ( tracer_enabled ^ val ) {
if ( val ) {
2008-05-12 23:20:42 +04:00
tracer_enabled = 1 ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
if ( current_trace - > start )
current_trace - > start ( tr ) ;
tracing_start ( ) ;
} else {
2008-05-12 23:20:42 +04:00
tracer_enabled = 0 ;
ftrace: restructure tracing start/stop infrastructure
Impact: change where tracing is started up and stopped
Currently, when a new tracer is selected via echo'ing a tracer name into
the current_tracer file, the startup is only done if tracing_enabled is
set to one. If tracing_enabled is changed to zero (by echo'ing 0 into
the tracing_enabled file) a full shutdown is performed.
The full startup and shutdown of a tracer can be expensive and the
user can lose out traces when echo'ing in 0 to the tracing_enabled file,
because the process takes too long. There can also be places that
the user would like to start and stop the tracer several times and
doing the full startup and shutdown of a tracer might be too expensive.
This patch performs the full startup and shutdown when a tracer is
selected. It also adds a way to do a quick start or stop of a tracer.
The quick version is just a flag that prevents the tracing from
taking place, but the overhead of the code is still there.
For example, the startup of a tracer may enable tracepoints, or enable
the function tracer. The stop and start will just set a flag to
have the tracer ignore the calls when the tracepoint or function trace
is called. The overhead of the tracer may still be present when
the tracer is stopped, but no tracing will occur. Setting the tracer
to the 'nop' tracer (or any other tracer) will perform the shutdown
of the tracer which will disable the tracepoint or disable the
function tracer.
The tracing_enabled file will simply start or stop tracing.
This change is all internal. The end result for the user should be the same
as before. If tracing_enabled is not set, no trace will happen.
If tracing_enabled is set, then the trace will happen. The tracing_enabled
variable is static between tracers. Enabling tracing_enabled and
going to another tracer will keep tracing_enabled enabled. Same
is true with disabling tracing_enabled.
This patch will now provide a fast start/stop method to the users
for enabling or disabling tracing.
Note: There were two methods to the struct tracer that were never
used: The methods start and stop. These were to be used as a hook
to the reading of the trace output, but ended up not being
necessary. These two methods are now used to enable the start
and stop of each tracer, in case the tracer needs to do more than
just not write into the buffer. For example, the irqsoff tracer
must stop recording max latencies when tracing is stopped.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-06 00:05:44 +03:00
tracing_stop ( ) ;
if ( current_trace - > stop )
current_trace - > stop ( tr ) ;
}
2008-05-12 23:20:42 +04:00
}
mutex_unlock ( & trace_types_lock ) ;
filp - > f_pos + = cnt ;
return cnt ;
}
static ssize_t
tracing_set_trace_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ max_tracer_type_len + 2 ] ;
int r ;
mutex_lock ( & trace_types_lock ) ;
if ( current_trace )
r = sprintf ( buf , " %s \n " , current_trace - > name ) ;
else
r = sprintf ( buf , " \n " ) ;
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:20:46 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
}
2008-11-01 21:57:37 +03:00
static int tracing_set_tracer ( char * buf )
2008-05-12 23:20:42 +04:00
{
struct trace_array * tr = & global_trace ;
struct tracer * t ;
2008-11-01 21:57:37 +03:00
int ret = 0 ;
2008-05-12 23:20:42 +04:00
mutex_lock ( & trace_types_lock ) ;
for ( t = trace_types ; t ; t = t - > next ) {
if ( strcmp ( t - > name , buf ) = = 0 )
break ;
}
2008-10-05 00:04:44 +04:00
if ( ! t ) {
ret = - EINVAL ;
goto out ;
}
if ( t = = current_trace )
2008-05-12 23:20:42 +04:00
goto out ;
2008-11-12 23:24:24 +03:00
trace_branch_disable ( ) ;
2008-05-12 23:20:42 +04:00
if ( current_trace & & current_trace - > reset )
current_trace - > reset ( tr ) ;
current_trace = t ;
if ( t - > init )
t - > init ( tr ) ;
2008-11-12 23:24:24 +03:00
trace_branch_enable ( tr ) ;
2008-05-12 23:20:42 +04:00
out :
mutex_unlock ( & trace_types_lock ) ;
2008-11-01 21:57:37 +03:00
return ret ;
}
static ssize_t
tracing_set_trace_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
char buf [ max_tracer_type_len + 1 ] ;
int i ;
size_t ret ;
if ( cnt > max_tracer_type_len )
cnt = max_tracer_type_len ;
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
/* strip ending whitespace. */
for ( i = cnt - 1 ; i > 0 & & isspace ( buf [ i ] ) ; i - - )
buf [ i ] = 0 ;
ret = tracing_set_tracer ( buf ) ;
if ( ! ret )
ret = cnt ;
2008-10-28 17:44:24 +03:00
if ( ret > 0 )
filp - > f_pos + = ret ;
2008-05-12 23:20:42 +04:00
2008-10-05 00:04:44 +04:00
return ret ;
2008-05-12 23:20:42 +04:00
}
static ssize_t
tracing_max_lat_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
unsigned long * ptr = filp - > private_data ;
char buf [ 64 ] ;
int r ;
2008-05-12 23:21:00 +04:00
r = snprintf ( buf , sizeof ( buf ) , " %ld \n " ,
2008-05-12 23:20:42 +04:00
* ptr = = ( unsigned long ) - 1 ? - 1 : nsecs_to_usecs ( * ptr ) ) ;
2008-05-12 23:21:00 +04:00
if ( r > sizeof ( buf ) )
r = sizeof ( buf ) ;
2008-05-12 23:20:46 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
2008-05-12 23:20:42 +04:00
}
static ssize_t
tracing_max_lat_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
long * ptr = filp - > private_data ;
char buf [ 64 ] ;
2008-05-12 23:21:00 +04:00
long val ;
int ret ;
2008-05-12 23:20:42 +04:00
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:42 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
2008-05-12 23:21:00 +04:00
ret = strict_strtoul ( buf , 10 , & val ) ;
if ( ret < 0 )
return ret ;
2008-05-12 23:20:42 +04:00
* ptr = val * 1000 ;
return cnt ;
}
2008-05-12 23:20:46 +04:00
static atomic_t tracing_reader ;
static int tracing_open_pipe ( struct inode * inode , struct file * filp )
{
struct trace_iterator * iter ;
if ( tracing_disabled )
return - ENODEV ;
/* We only allow for reader of the pipe */
if ( atomic_inc_return ( & tracing_reader ) ! = 1 ) {
atomic_dec ( & tracing_reader ) ;
return - EBUSY ;
}
/* create a buffer to store the information to pass to userspace */
iter = kzalloc ( sizeof ( * iter ) , GFP_KERNEL ) ;
if ( ! iter )
return - ENOMEM ;
2008-05-12 23:21:01 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-11-08 06:36:02 +03:00
/* trace pipe does not show start of buffer */
cpus_setall ( iter - > started ) ;
2008-05-12 23:20:46 +04:00
iter - > tr = & global_trace ;
2008-05-23 23:37:28 +04:00
iter - > trace = current_trace ;
2008-05-12 23:20:46 +04:00
filp - > private_data = iter ;
2008-05-12 23:21:01 +04:00
if ( iter - > trace - > pipe_open )
iter - > trace - > pipe_open ( iter ) ;
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:20:46 +04:00
return 0 ;
}
static int tracing_release_pipe ( struct inode * inode , struct file * file )
{
struct trace_iterator * iter = file - > private_data ;
kfree ( iter ) ;
atomic_dec ( & tracing_reader ) ;
return 0 ;
}
2008-05-12 23:20:49 +04:00
static unsigned int
tracing_poll_pipe ( struct file * filp , poll_table * poll_table )
{
struct trace_iterator * iter = filp - > private_data ;
if ( trace_flags & TRACE_ITER_BLOCK ) {
/*
* Always select as readable when in blocking mode
*/
return POLLIN | POLLRDNORM ;
2008-05-12 23:21:00 +04:00
} else {
2008-05-12 23:20:49 +04:00
if ( ! trace_empty ( iter ) )
return POLLIN | POLLRDNORM ;
poll_wait ( filp , & trace_wait , poll_table ) ;
if ( ! trace_empty ( iter ) )
return POLLIN | POLLRDNORM ;
return 0 ;
}
}
2008-05-12 23:20:46 +04:00
/*
* Consumer reader .
*/
static ssize_t
tracing_read_pipe ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
struct trace_iterator * iter = filp - > private_data ;
2008-05-12 23:21:02 +04:00
ssize_t sret ;
2008-05-12 23:20:46 +04:00
/* return any leftover data */
2008-05-12 23:21:02 +04:00
sret = trace_seq_to_user ( & iter - > seq , ubuf , cnt ) ;
if ( sret ! = - EBUSY )
return sret ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:02 +04:00
trace_seq_reset ( & iter - > seq ) ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:01 +04:00
mutex_lock ( & trace_types_lock ) ;
if ( iter - > trace - > read ) {
2008-05-12 23:21:02 +04:00
sret = iter - > trace - > read ( iter , filp , ubuf , cnt , ppos ) ;
if ( sret )
2008-05-12 23:21:01 +04:00
goto out ;
}
2008-09-29 22:23:48 +04:00
waitagain :
sret = 0 ;
2008-05-12 23:20:46 +04:00
while ( trace_empty ( iter ) ) {
2008-05-12 23:20:58 +04:00
2008-05-12 23:21:01 +04:00
if ( ( filp - > f_flags & O_NONBLOCK ) ) {
2008-05-12 23:21:02 +04:00
sret = - EAGAIN ;
2008-05-12 23:21:01 +04:00
goto out ;
}
2008-05-12 23:20:58 +04:00
2008-05-12 23:20:46 +04:00
/*
* This is a make - shift waitqueue . The reason we don ' t use
* an actual wait queue is because :
* 1 ) we only ever have one waiter
* 2 ) the tracing , traces all functions , we don ' t want
* the overhead of calling wake_up and friends
* ( and tracing them too )
* Anyway , this is really very primitive wakeup .
*/
set_current_state ( TASK_INTERRUPTIBLE ) ;
iter - > tr - > waiter = current ;
2008-05-12 23:21:01 +04:00
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:21:02 +04:00
/* sleep for 100 msecs, and try again. */
schedule_timeout ( HZ / 10 ) ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:01 +04:00
mutex_lock ( & trace_types_lock ) ;
2008-05-12 23:20:46 +04:00
iter - > tr - > waiter = NULL ;
2008-05-12 23:21:01 +04:00
if ( signal_pending ( current ) ) {
2008-05-12 23:21:02 +04:00
sret = - EINTR ;
2008-05-12 23:21:01 +04:00
goto out ;
}
2008-05-12 23:20:46 +04:00
2008-05-12 23:20:58 +04:00
if ( iter - > trace ! = current_trace )
2008-05-12 23:21:01 +04:00
goto out ;
2008-05-12 23:20:58 +04:00
2008-05-12 23:20:46 +04:00
/*
* We block until we read something and tracing is disabled .
* We still block if tracing is disabled , but we have never
* read anything . This allows a user to cat this file , and
* then enable tracing . But after we have read something ,
* we give an EOF when tracing is again disabled .
*
* iter - > pos will be 0 if we haven ' t read anything .
*/
if ( ! tracer_enabled & & iter - > pos )
break ;
continue ;
}
/* stop when tracing is finished */
if ( trace_empty ( iter ) )
2008-05-12 23:21:01 +04:00
goto out ;
2008-05-12 23:20:46 +04:00
if ( cnt > = PAGE_SIZE )
cnt = PAGE_SIZE - 1 ;
2008-05-12 23:21:01 +04:00
/* reset all but tr, trace, and overruns */
memset ( & iter - > seq , 0 ,
sizeof ( struct trace_iterator ) -
offsetof ( struct trace_iterator , seq ) ) ;
2008-05-12 23:21:01 +04:00
iter - > pos = - 1 ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:20:48 +04:00
while ( find_next_entry_inc ( iter ) ! = NULL ) {
2008-09-29 22:18:34 +04:00
enum print_line_t ret ;
2008-05-12 23:20:48 +04:00
int len = iter - > seq . len ;
2008-05-12 23:20:47 +04:00
ret = print_trace_line ( iter ) ;
2008-09-29 22:18:34 +04:00
if ( ret = = TRACE_TYPE_PARTIAL_LINE ) {
2008-05-12 23:20:48 +04:00
/* don't print partial lines */
iter - > seq . len = len ;
2008-05-12 23:20:46 +04:00
break ;
2008-05-12 23:20:48 +04:00
}
2008-05-12 23:20:46 +04:00
trace_consume ( iter ) ;
if ( iter - > seq . len > = cnt )
break ;
}
/* Now copy what we have to the user */
2008-05-12 23:21:02 +04:00
sret = trace_seq_to_user ( & iter - > seq , ubuf , cnt ) ;
if ( iter - > seq . readpos > = iter - > seq . len )
2008-05-12 23:20:46 +04:00
trace_seq_reset ( & iter - > seq ) ;
2008-09-29 22:23:48 +04:00
/*
* If there was nothing to send to user , inspite of consuming trace
* entries , go back to wait for more entries .
*/
2008-05-12 23:21:02 +04:00
if ( sret = = - EBUSY )
2008-09-29 22:23:48 +04:00
goto waitagain ;
2008-05-12 23:20:46 +04:00
2008-05-12 23:21:01 +04:00
out :
mutex_unlock ( & trace_types_lock ) ;
2008-05-12 23:21:02 +04:00
return sret ;
2008-05-12 23:20:46 +04:00
}
2008-05-12 23:20:59 +04:00
static ssize_t
tracing_entries_read ( struct file * filp , char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
struct trace_array * tr = filp - > private_data ;
char buf [ 64 ] ;
int r ;
2008-11-13 08:09:35 +03:00
r = sprintf ( buf , " %lu \n " , tr - > entries > > 10 ) ;
2008-05-12 23:20:59 +04:00
return simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
}
static ssize_t
tracing_entries_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * ppos )
{
unsigned long val ;
char buf [ 64 ] ;
2008-11-11 05:46:00 +03:00
int ret , cpu ;
2008-05-12 23:20:59 +04:00
2008-05-12 23:21:00 +04:00
if ( cnt > = sizeof ( buf ) )
return - EINVAL ;
2008-05-12 23:20:59 +04:00
if ( copy_from_user ( & buf , ubuf , cnt ) )
return - EFAULT ;
buf [ cnt ] = 0 ;
2008-05-12 23:21:00 +04:00
ret = strict_strtoul ( buf , 10 , & val ) ;
if ( ret < 0 )
return ret ;
2008-05-12 23:20:59 +04:00
/* must have at least 1 entry */
if ( ! val )
return - EINVAL ;
mutex_lock ( & trace_types_lock ) ;
2008-11-08 06:36:02 +03:00
tracing_stop ( ) ;
2008-05-12 23:20:59 +04:00
2008-11-11 05:46:00 +03:00
/* disable all cpu buffers */
for_each_tracing_cpu ( cpu ) {
if ( global_trace . data [ cpu ] )
atomic_inc ( & global_trace . data [ cpu ] - > disabled ) ;
if ( max_tr . data [ cpu ] )
atomic_inc ( & max_tr . data [ cpu ] - > disabled ) ;
}
2008-11-13 08:09:35 +03:00
/* value is in KB */
val < < = 10 ;
2008-09-30 07:02:41 +04:00
if ( val ! = global_trace . entries ) {
ret = ring_buffer_resize ( global_trace . buffer , val ) ;
if ( ret < 0 ) {
cnt = ret ;
2008-05-12 23:21:04 +04:00
goto out ;
}
2008-09-30 07:02:41 +04:00
ret = ring_buffer_resize ( max_tr . buffer , val ) ;
if ( ret < 0 ) {
int r ;
cnt = ret ;
r = ring_buffer_resize ( global_trace . buffer ,
global_trace . entries ) ;
if ( r < 0 ) {
/* AARGH! We are left with different
* size max buffer ! ! ! ! */
WARN_ON ( 1 ) ;
tracing_disabled = 1 ;
2008-05-12 23:20:59 +04:00
}
2008-09-30 07:02:41 +04:00
goto out ;
2008-05-12 23:20:59 +04:00
}
2008-05-12 23:21:04 +04:00
2008-09-30 07:02:41 +04:00
global_trace . entries = val ;
2008-05-12 23:20:59 +04:00
}
filp - > f_pos + = cnt ;
2008-05-22 08:22:16 +04:00
/* If check pages failed, return ENOMEM */
if ( tracing_disabled )
cnt = - ENOMEM ;
2008-05-12 23:20:59 +04:00
out :
2008-11-11 05:46:00 +03:00
for_each_tracing_cpu ( cpu ) {
if ( global_trace . data [ cpu ] )
atomic_dec ( & global_trace . data [ cpu ] - > disabled ) ;
if ( max_tr . data [ cpu ] )
atomic_dec ( & max_tr . data [ cpu ] - > disabled ) ;
}
2008-11-08 06:36:02 +03:00
tracing_start ( ) ;
2008-05-12 23:20:59 +04:00
max_tr . entries = global_trace . entries ;
mutex_unlock ( & trace_types_lock ) ;
return cnt ;
}
2008-09-16 23:06:42 +04:00
static int mark_printk ( const char * fmt , . . . )
{
int ret ;
va_list args ;
va_start ( args , fmt ) ;
ret = trace_vprintk ( 0 , fmt , args ) ;
va_end ( args ) ;
return ret ;
}
static ssize_t
tracing_mark_write ( struct file * filp , const char __user * ubuf ,
size_t cnt , loff_t * fpos )
{
char * buf ;
char * end ;
2008-11-08 06:36:02 +03:00
if ( tracing_disabled )
2008-09-16 23:06:42 +04:00
return - EINVAL ;
if ( cnt > TRACE_BUF_SIZE )
cnt = TRACE_BUF_SIZE ;
buf = kmalloc ( cnt + 1 , GFP_KERNEL ) ;
if ( buf = = NULL )
return - ENOMEM ;
if ( copy_from_user ( buf , ubuf , cnt ) ) {
kfree ( buf ) ;
return - EFAULT ;
}
/* Cut from the first nil or newline. */
buf [ cnt ] = ' \0 ' ;
end = strchr ( buf , ' \n ' ) ;
if ( end )
* end = ' \0 ' ;
cnt = mark_printk ( " %s \n " , buf ) ;
kfree ( buf ) ;
* fpos + = cnt ;
return cnt ;
}
2008-05-12 23:20:42 +04:00
static struct file_operations tracing_max_lat_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
. read = tracing_max_lat_read ,
. write = tracing_max_lat_write ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations tracing_ctrl_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
. read = tracing_ctrl_read ,
. write = tracing_ctrl_write ,
2008-05-12 23:20:42 +04:00
} ;
static struct file_operations set_tracer_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
. read = tracing_set_trace_read ,
. write = tracing_set_trace_write ,
2008-05-12 23:20:42 +04:00
} ;
2008-05-12 23:20:46 +04:00
static struct file_operations tracing_pipe_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_pipe ,
2008-05-12 23:20:49 +04:00
. poll = tracing_poll_pipe ,
2008-05-12 23:20:46 +04:00
. read = tracing_read_pipe ,
. release = tracing_release_pipe ,
2008-05-12 23:20:46 +04:00
} ;
2008-05-12 23:20:59 +04:00
static struct file_operations tracing_entries_fops = {
. open = tracing_open_generic ,
. read = tracing_entries_read ,
. write = tracing_entries_write ,
} ;
2008-09-16 23:06:42 +04:00
static struct file_operations tracing_mark_fops = {
2008-09-21 22:16:30 +04:00
. open = tracing_open_generic ,
2008-09-16 23:06:42 +04:00
. write = tracing_mark_write ,
} ;
2008-05-12 23:20:42 +04:00
# ifdef CONFIG_DYNAMIC_FTRACE
2008-10-30 23:08:33 +03:00
int __weak ftrace_arch_read_dyn_info ( char * buf , int size )
{
return 0 ;
}
2008-05-12 23:20:42 +04:00
static ssize_t
2008-10-30 23:08:33 +03:00
tracing_read_dyn_info ( struct file * filp , char __user * ubuf ,
2008-05-12 23:20:42 +04:00
size_t cnt , loff_t * ppos )
{
2008-10-31 07:03:22 +03:00
static char ftrace_dyn_info_buffer [ 1024 ] ;
static DEFINE_MUTEX ( dyn_info_mutex ) ;
2008-05-12 23:20:42 +04:00
unsigned long * p = filp - > private_data ;
2008-10-30 23:08:33 +03:00
char * buf = ftrace_dyn_info_buffer ;
2008-10-31 07:03:22 +03:00
int size = ARRAY_SIZE ( ftrace_dyn_info_buffer ) ;
2008-05-12 23:20:42 +04:00
int r ;
2008-10-30 23:08:33 +03:00
mutex_lock ( & dyn_info_mutex ) ;
r = sprintf ( buf , " %ld " , * p ) ;
2008-05-12 23:20:46 +04:00
2008-10-31 07:03:22 +03:00
r + = ftrace_arch_read_dyn_info ( buf + r , ( size - 1 ) - r ) ;
2008-10-30 23:08:33 +03:00
buf [ r + + ] = ' \n ' ;
r = simple_read_from_buffer ( ubuf , cnt , ppos , buf , r ) ;
mutex_unlock ( & dyn_info_mutex ) ;
return r ;
2008-05-12 23:20:42 +04:00
}
2008-10-30 23:08:33 +03:00
static struct file_operations tracing_dyn_info_fops = {
2008-05-12 23:20:46 +04:00
. open = tracing_open_generic ,
2008-10-30 23:08:33 +03:00
. read = tracing_read_dyn_info ,
2008-05-12 23:20:42 +04:00
} ;
# endif
static struct dentry * d_tracer ;
struct dentry * tracing_init_dentry ( void )
{
static int once ;
if ( d_tracer )
return d_tracer ;
d_tracer = debugfs_create_dir ( " tracing " , NULL ) ;
if ( ! d_tracer & & ! once ) {
once = 1 ;
pr_warning ( " Could not create debugfs directory 'tracing' \n " ) ;
return NULL ;
}
return d_tracer ;
}
2008-05-12 23:20:44 +04:00
# ifdef CONFIG_FTRACE_SELFTEST
/* Let selftest have access to static functions in this file */
# include "trace_selftest.c"
# endif
2008-09-23 14:34:32 +04:00
static __init int tracer_init_debugfs ( void )
2008-05-12 23:20:42 +04:00
{
struct dentry * d_tracer ;
struct dentry * entry ;
d_tracer = tracing_init_dentry ( ) ;
entry = debugfs_create_file ( " tracing_enabled " , 0644 , d_tracer ,
& global_trace , & tracing_ctrl_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'tracing_enabled' entry \n " ) ;
entry = debugfs_create_file ( " iter_ctrl " , 0644 , d_tracer ,
NULL , & tracing_iter_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'iter_ctrl' entry \n " ) ;
2008-05-12 23:20:52 +04:00
entry = debugfs_create_file ( " tracing_cpumask " , 0644 , d_tracer ,
NULL , & tracing_cpumask_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'tracing_cpumask' entry \n " ) ;
2008-05-12 23:20:42 +04:00
entry = debugfs_create_file ( " latency_trace " , 0444 , d_tracer ,
& global_trace , & tracing_lt_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'latency_trace' entry \n " ) ;
entry = debugfs_create_file ( " trace " , 0444 , d_tracer ,
& global_trace , & tracing_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'trace' entry \n " ) ;
entry = debugfs_create_file ( " available_tracers " , 0444 , d_tracer ,
& global_trace , & show_traces_fops ) ;
if ( ! entry )
2008-08-15 23:08:22 +04:00
pr_warning ( " Could not create debugfs 'available_tracers' entry \n " ) ;
2008-05-12 23:20:42 +04:00
entry = debugfs_create_file ( " current_tracer " , 0444 , d_tracer ,
& global_trace , & set_tracer_fops ) ;
if ( ! entry )
2008-08-15 23:08:22 +04:00
pr_warning ( " Could not create debugfs 'current_tracer' entry \n " ) ;
2008-05-12 23:20:42 +04:00
entry = debugfs_create_file ( " tracing_max_latency " , 0644 , d_tracer ,
& tracing_max_latency ,
& tracing_max_lat_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
" 'tracing_max_latency' entry \n " ) ;
entry = debugfs_create_file ( " tracing_thresh " , 0644 , d_tracer ,
& tracing_thresh , & tracing_max_lat_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
2008-08-15 23:08:22 +04:00
" 'tracing_thresh' entry \n " ) ;
2008-05-12 23:20:45 +04:00
entry = debugfs_create_file ( " README " , 0644 , d_tracer ,
NULL , & tracing_readme_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs 'README' entry \n " ) ;
2008-05-12 23:20:46 +04:00
entry = debugfs_create_file ( " trace_pipe " , 0644 , d_tracer ,
NULL , & tracing_pipe_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
2008-08-15 23:08:22 +04:00
" 'trace_pipe' entry \n " ) ;
2008-05-12 23:20:42 +04:00
2008-11-13 01:52:36 +03:00
entry = debugfs_create_file ( " buffer_size_kb " , 0644 , d_tracer ,
2008-05-12 23:20:59 +04:00
& global_trace , & tracing_entries_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
2008-11-13 01:52:36 +03:00
" 'buffer_size_kb' entry \n " ) ;
2008-05-12 23:20:59 +04:00
2008-09-16 23:06:42 +04:00
entry = debugfs_create_file ( " trace_marker " , 0220 , d_tracer ,
NULL , & tracing_mark_fops ) ;
if ( ! entry )
pr_warning ( " Could not create debugfs "
" 'trace_marker' entry \n " ) ;
2008-05-12 23:20:42 +04:00
# ifdef CONFIG_DYNAMIC_FTRACE
entry = debugfs_create_file ( " dyn_ftrace_total_info " , 0444 , d_tracer ,
& ftrace_update_tot_cnt ,
2008-10-30 23:08:33 +03:00
& tracing_dyn_info_fops ) ;
2008-05-12 23:20:42 +04:00
if ( ! entry )
pr_warning ( " Could not create debugfs "
" 'dyn_ftrace_total_info' entry \n " ) ;
# endif
2008-05-12 23:20:49 +04:00
# ifdef CONFIG_SYSPROF_TRACER
init_tracer_sysprof_debugfs ( d_tracer ) ;
# endif
2008-09-23 14:34:32 +04:00
return 0 ;
2008-05-12 23:20:42 +04:00
}
2008-09-16 22:58:24 +04:00
int trace_vprintk ( unsigned long ip , const char * fmt , va_list args )
2008-08-01 20:26:41 +04:00
{
static DEFINE_SPINLOCK ( trace_buf_lock ) ;
static char trace_buf [ TRACE_BUF_SIZE ] ;
2008-09-04 12:24:14 +04:00
2008-09-30 07:02:41 +04:00
struct ring_buffer_event * event ;
2008-09-04 12:24:14 +04:00
struct trace_array * tr = & global_trace ;
2008-08-01 20:26:41 +04:00
struct trace_array_cpu * data ;
2008-09-30 07:02:42 +04:00
struct print_entry * entry ;
2008-09-30 07:02:41 +04:00
unsigned long flags , irq_flags ;
2008-10-01 21:14:09 +04:00
int cpu , len = 0 , size , pc ;
2008-08-01 20:26:41 +04:00
2008-11-08 06:36:02 +03:00
if ( tracing_disabled )
2008-08-01 20:26:41 +04:00
return 0 ;
2008-10-01 21:14:09 +04:00
pc = preempt_count ( ) ;
preempt_disable_notrace ( ) ;
2008-08-01 20:26:41 +04:00
cpu = raw_smp_processor_id ( ) ;
data = tr - > data [ cpu ] ;
2008-10-04 10:01:00 +04:00
if ( unlikely ( atomic_read ( & data - > disabled ) ) )
2008-08-01 20:26:41 +04:00
goto out ;
2008-10-01 21:14:09 +04:00
spin_lock_irqsave ( & trace_buf_lock , flags ) ;
2008-09-16 22:58:24 +04:00
len = vsnprintf ( trace_buf , TRACE_BUF_SIZE , fmt , args ) ;
2008-08-01 20:26:41 +04:00
len = min ( len , TRACE_BUF_SIZE - 1 ) ;
trace_buf [ len ] = 0 ;
2008-09-30 07:02:42 +04:00
size = sizeof ( * entry ) + len + 1 ;
event = ring_buffer_lock_reserve ( tr - > buffer , size , & irq_flags ) ;
2008-09-30 07:02:41 +04:00
if ( ! event )
goto out_unlock ;
2008-09-30 07:02:42 +04:00
entry = ring_buffer_event_data ( event ) ;
2008-10-01 21:14:09 +04:00
tracing_generic_entry_update ( & entry - > ent , flags , pc ) ;
2008-09-30 07:02:42 +04:00
entry - > ent . type = TRACE_PRINT ;
entry - > ip = ip ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:42 +04:00
memcpy ( & entry - > buf , trace_buf , len ) ;
entry - > buf [ len ] = 0 ;
2008-09-30 07:02:41 +04:00
ring_buffer_unlock_commit ( tr - > buffer , event , irq_flags ) ;
2008-08-01 20:26:41 +04:00
2008-09-30 07:02:41 +04:00
out_unlock :
2008-10-01 21:14:09 +04:00
spin_unlock_irqrestore ( & trace_buf_lock , flags ) ;
2008-08-01 20:26:41 +04:00
out :
2008-10-01 21:14:09 +04:00
preempt_enable_notrace ( ) ;
2008-08-01 20:26:41 +04:00
return len ;
}
2008-09-16 22:58:24 +04:00
EXPORT_SYMBOL_GPL ( trace_vprintk ) ;
int __ftrace_printk ( unsigned long ip , const char * fmt , . . . )
{
int ret ;
va_list ap ;
if ( ! ( trace_flags & TRACE_ITER_PRINTK ) )
return 0 ;
va_start ( ap , fmt ) ;
ret = trace_vprintk ( ip , fmt , ap ) ;
va_end ( ap ) ;
return ret ;
}
2008-08-01 20:26:41 +04:00
EXPORT_SYMBOL_GPL ( __ftrace_printk ) ;
2008-07-31 06:36:46 +04:00
static int trace_panic_handler ( struct notifier_block * this ,
unsigned long event , void * unused )
{
2008-10-24 03:26:08 +04:00
if ( ftrace_dump_on_oops )
ftrace_dump ( ) ;
2008-07-31 06:36:46 +04:00
return NOTIFY_OK ;
}
static struct notifier_block trace_panic_notifier = {
. notifier_call = trace_panic_handler ,
. next = NULL ,
. priority = 150 /* priority: INT_MAX >= x >= 0 */
} ;
static int trace_die_handler ( struct notifier_block * self ,
unsigned long val ,
void * data )
{
switch ( val ) {
case DIE_OOPS :
2008-10-24 03:26:08 +04:00
if ( ftrace_dump_on_oops )
ftrace_dump ( ) ;
2008-07-31 06:36:46 +04:00
break ;
default :
break ;
}
return NOTIFY_OK ;
}
static struct notifier_block trace_die_notifier = {
. notifier_call = trace_die_handler ,
. priority = 200
} ;
/*
* printk is set to max of 1024 , we really don ' t need it that big .
* Nothing should be printing 1000 characters anyway .
*/
# define TRACE_MAX_PRINT 1000
/*
* Define here KERN_TRACE so that we have one place to modify
* it if we decide to change what log level the ftrace dump
* should be at .
*/
# define KERN_TRACE KERN_INFO
static void
trace_printk_seq ( struct trace_seq * s )
{
/* Probably should print a warning here. */
if ( s - > len > = 1000 )
s - > len = 1000 ;
/* should be zero ended, but we are paranoid. */
s - > buffer [ s - > len ] = 0 ;
printk ( KERN_TRACE " %s " , s - > buffer ) ;
trace_seq_reset ( s ) ;
}
void ftrace_dump ( void )
{
static DEFINE_SPINLOCK ( ftrace_dump_lock ) ;
/* use static because iter can be a bit big for the stack */
static struct trace_iterator iter ;
static cpumask_t mask ;
static int dump_ran ;
2008-10-01 08:29:53 +04:00
unsigned long flags ;
int cnt = 0 , cpu ;
2008-07-31 06:36:46 +04:00
/* only one dump */
spin_lock_irqsave ( & ftrace_dump_lock , flags ) ;
if ( dump_ran )
goto out ;
dump_ran = 1 ;
/* No turning back! */
2008-10-23 17:33:02 +04:00
ftrace_kill ( ) ;
2008-07-31 06:36:46 +04:00
2008-10-01 08:29:53 +04:00
for_each_tracing_cpu ( cpu ) {
atomic_inc ( & global_trace . data [ cpu ] - > disabled ) ;
}
2008-07-31 06:36:46 +04:00
printk ( KERN_TRACE " Dumping ftrace buffer: \n " ) ;
iter . tr = & global_trace ;
iter . trace = current_trace ;
/*
* We need to stop all tracing on all CPUS to read the
* the next buffer . This is a bit expensive , but is
* not done often . We fill all what we can read ,
* and then release the locks again .
*/
cpus_clear ( mask ) ;
while ( ! trace_empty ( & iter ) ) {
if ( ! cnt )
printk ( KERN_TRACE " --------------------------------- \n " ) ;
cnt + + ;
/* reset all but tr, trace, and overruns */
memset ( & iter . seq , 0 ,
sizeof ( struct trace_iterator ) -
offsetof ( struct trace_iterator , seq ) ) ;
iter . iter_flags | = TRACE_FILE_LAT_FMT ;
iter . pos = - 1 ;
if ( find_next_entry_inc ( & iter ) ! = NULL ) {
print_trace_line ( & iter ) ;
trace_consume ( & iter ) ;
}
trace_printk_seq ( & iter . seq ) ;
}
if ( ! cnt )
printk ( KERN_TRACE " (ftrace buffer empty) \n " ) ;
else
printk ( KERN_TRACE " --------------------------------- \n " ) ;
out :
spin_unlock_irqrestore ( & ftrace_dump_lock , flags ) ;
}
2008-09-30 07:02:41 +04:00
__init static int tracer_alloc_buffers ( void )
2008-05-12 23:20:42 +04:00
{
2008-05-12 23:20:43 +04:00
struct trace_array_cpu * data ;
int i ;
2008-09-30 07:02:41 +04:00
/* TODO: make the number of buffers hot pluggable with CPUS */
tracing_buffer_mask = cpu_possible_map ;
2008-05-12 23:20:43 +04:00
2008-09-30 07:02:41 +04:00
global_trace . buffer = ring_buffer_alloc ( trace_buf_size ,
TRACE_BUFFER_FLAGS ) ;
if ( ! global_trace . buffer ) {
printk ( KERN_ERR " tracer: failed to allocate ring buffer! \n " ) ;
WARN_ON ( 1 ) ;
return 0 ;
2008-05-12 23:20:43 +04:00
}
2008-09-30 07:02:41 +04:00
global_trace . entries = ring_buffer_size ( global_trace . buffer ) ;
2008-05-12 23:20:43 +04:00
# ifdef CONFIG_TRACER_MAX_TRACE
2008-09-30 07:02:41 +04:00
max_tr . buffer = ring_buffer_alloc ( trace_buf_size ,
TRACE_BUFFER_FLAGS ) ;
if ( ! max_tr . buffer ) {
printk ( KERN_ERR " tracer: failed to allocate max ring buffer! \n " ) ;
WARN_ON ( 1 ) ;
ring_buffer_free ( global_trace . buffer ) ;
return 0 ;
2008-05-12 23:20:43 +04:00
}
2008-09-30 07:02:41 +04:00
max_tr . entries = ring_buffer_size ( max_tr . buffer ) ;
WARN_ON ( max_tr . entries ! = global_trace . entries ) ;
2008-05-12 23:20:59 +04:00
# endif
2008-05-12 23:21:00 +04:00
2008-05-12 23:20:43 +04:00
/* Allocate the first page for all buffers */
2008-05-12 23:21:00 +04:00
for_each_tracing_cpu ( i ) {
2008-05-12 23:20:43 +04:00
data = global_trace . data [ i ] = & per_cpu ( global_trace_cpu , i ) ;
2008-05-12 23:20:42 +04:00
max_tr . data [ i ] = & per_cpu ( max_data , i ) ;
2008-05-12 23:20:43 +04:00
}
2008-05-12 23:20:42 +04:00
trace_init_cmdlines ( ) ;
2008-09-21 22:16:30 +04:00
register_tracer ( & nop_trace ) ;
2008-09-23 14:34:32 +04:00
# ifdef CONFIG_BOOT_TRACER
register_tracer ( & boot_tracer ) ;
current_trace = & boot_tracer ;
current_trace - > init ( & global_trace ) ;
# else
2008-09-21 22:16:30 +04:00
current_trace = & nop_trace ;
2008-09-23 14:34:32 +04:00
# endif
2008-05-12 23:20:42 +04:00
2008-05-12 23:20:44 +04:00
/* All seems OK, enable tracing */
tracing_disabled = 0 ;
2008-09-30 07:02:41 +04:00
2008-07-31 06:36:46 +04:00
atomic_notifier_chain_register ( & panic_notifier_list ,
& trace_panic_notifier ) ;
register_die_notifier ( & trace_die_notifier ) ;
2008-05-12 23:20:42 +04:00
return 0 ;
}
2008-09-23 14:34:32 +04:00
early_initcall ( tracer_alloc_buffers ) ;
fs_initcall ( tracer_init_debugfs ) ;