License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 17:07:57 +03:00
// SPDX-License-Identifier: GPL-2.0
2008-11-12 08:14:39 +03:00
/*
* unlikely profiler
*
* Copyright ( C ) 2008 Steven Rostedt < srostedt @ redhat . com >
*/
# include <linux/kallsyms.h>
# include <linux/seq_file.h>
# include <linux/spinlock.h>
2008-11-29 06:12:46 +03:00
# include <linux/irqflags.h>
2008-11-12 08:14:39 +03:00
# include <linux/uaccess.h>
# include <linux/module.h>
# include <linux/ftrace.h>
# include <linux/hash.h>
# include <linux/fs.h>
# include <asm/local.h>
2008-12-24 07:24:13 +03:00
2008-11-12 08:14:39 +03:00
# include "trace.h"
2009-01-10 22:34:13 +03:00
# include "trace_stat.h"
2008-12-24 07:24:13 +03:00
# include "trace_output.h"
2008-11-12 08:14:39 +03:00
2008-11-12 23:24:24 +03:00
# ifdef CONFIG_BRANCH_TRACER
2008-11-12 08:14:40 +03:00
2009-01-10 22:34:13 +03:00
static struct tracer branch_trace ;
2008-11-12 23:24:24 +03:00
static int branch_tracing_enabled __read_mostly ;
static DEFINE_MUTEX ( branch_tracing_mutex ) ;
2008-12-28 01:25:38 +03:00
2008-11-12 23:24:24 +03:00
static struct trace_array * branch_tracer ;
2008-11-12 08:14:40 +03:00
static void
2017-01-19 16:57:41 +03:00
probe_likely_condition ( struct ftrace_likely_data * f , int val , int expect )
2008-11-12 08:14:40 +03:00
{
2015-05-05 18:45:27 +03:00
struct trace_event_call * call = & event_branch ;
2008-11-12 23:24:24 +03:00
struct trace_array * tr = branch_tracer ;
2019-12-13 21:58:57 +03:00
struct trace_buffer * buffer ;
2012-08-07 00:24:11 +04:00
struct trace_array_cpu * data ;
2008-11-12 08:14:40 +03:00
struct ring_buffer_event * event ;
2008-11-12 23:24:24 +03:00
struct trace_branch * entry ;
2009-02-05 21:12:56 +03:00
unsigned long flags ;
2015-07-07 22:05:03 +03:00
int pc ;
2008-11-12 08:14:40 +03:00
const char * p ;
2015-07-07 22:05:03 +03:00
if ( current - > trace_recursion & TRACE_BRANCH_BIT )
return ;
2008-11-12 08:14:40 +03:00
/*
* I would love to save just the ftrace_likely_data pointer , but
* this code can also be used by modules . Ugly things can happen
* if the module is unloaded , and then we go and read the
* pointer . This is slower , but much safer .
*/
if ( unlikely ( ! tr ) )
return ;
2015-07-07 22:05:03 +03:00
raw_local_irq_save ( flags ) ;
current - > trace_recursion | = TRACE_BRANCH_BIT ;
2020-01-10 02:53:48 +03:00
data = this_cpu_ptr ( tr - > array_buffer . data ) ;
2015-07-07 22:05:03 +03:00
if ( atomic_read ( & data - > disabled ) )
2008-11-12 08:14:40 +03:00
goto out ;
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
pc = preempt_count ( ) ;
2020-01-10 02:53:48 +03:00
buffer = tr - > array_buffer . buffer ;
2009-10-08 05:53:41 +04:00
event = trace_buffer_lock_reserve ( buffer , TRACE_BRANCH ,
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-05 21:14:13 +03:00
sizeof ( * entry ) , flags , pc ) ;
2008-11-12 08:14:40 +03:00
if ( ! event )
goto out ;
entry = ring_buffer_event_data ( event ) ;
/* Strip off the path, only save the file */
2017-01-19 16:57:41 +03:00
p = f - > data . file + strlen ( f - > data . file ) ;
while ( p > = f - > data . file & & * p ! = ' / ' )
2008-11-12 08:14:40 +03:00
p - - ;
p + + ;
2017-01-19 16:57:41 +03:00
strncpy ( entry - > func , f - > data . func , TRACE_FUNC_SIZE ) ;
2008-11-12 08:14:40 +03:00
strncpy ( entry - > file , p , TRACE_FILE_SIZE ) ;
entry - > func [ TRACE_FUNC_SIZE ] = 0 ;
entry - > file [ TRACE_FILE_SIZE ] = 0 ;
2017-01-19 16:57:41 +03:00
entry - > constant = f - > constant ;
entry - > line = f - > data . line ;
2008-11-12 08:14:40 +03:00
entry - > correct = val = = expect ;
2013-10-24 17:34:17 +04:00
if ( ! call_filter_check_discard ( call , entry , buffer , event ) )
2016-11-24 04:28:38 +03:00
trace_buffer_unlock_commit_nostack ( buffer , event ) ;
2008-11-12 08:14:40 +03:00
out :
2015-07-07 22:05:03 +03:00
current - > trace_recursion & = ~ TRACE_BRANCH_BIT ;
raw_local_irq_restore ( flags ) ;
2008-11-12 08:14:40 +03:00
}
static inline
2017-01-19 16:57:41 +03:00
void trace_likely_condition ( struct ftrace_likely_data * f , int val , int expect )
2008-11-12 08:14:40 +03:00
{
2008-11-12 23:24:24 +03:00
if ( ! branch_tracing_enabled )
2008-11-12 08:14:40 +03:00
return ;
probe_likely_condition ( f , val , expect ) ;
}
2008-11-12 23:24:24 +03:00
int enable_branch_tracing ( struct trace_array * tr )
2008-11-12 08:14:40 +03:00
{
2008-11-12 23:24:24 +03:00
mutex_lock ( & branch_tracing_mutex ) ;
branch_tracer = tr ;
2008-11-12 08:14:40 +03:00
/*
* Must be seen before enabling . The reader is a condition
* where we do not need a matching rmb ( )
*/
smp_wmb ( ) ;
2008-11-12 23:24:24 +03:00
branch_tracing_enabled + + ;
mutex_unlock ( & branch_tracing_mutex ) ;
2008-11-12 08:14:40 +03:00
2009-02-10 09:02:46 +03:00
return 0 ;
2008-11-12 08:14:40 +03:00
}
2008-11-12 23:24:24 +03:00
void disable_branch_tracing ( void )
2008-11-12 08:14:40 +03:00
{
2008-11-12 23:24:24 +03:00
mutex_lock ( & branch_tracing_mutex ) ;
2008-11-12 08:14:40 +03:00
2008-11-12 23:24:24 +03:00
if ( ! branch_tracing_enabled )
2008-11-12 08:14:40 +03:00
goto out_unlock ;
2008-11-12 23:24:24 +03:00
branch_tracing_enabled - - ;
2008-11-12 08:14:40 +03:00
out_unlock :
2008-11-12 23:24:24 +03:00
mutex_unlock ( & branch_tracing_mutex ) ;
2008-11-12 08:14:40 +03:00
}
2008-11-12 23:24:24 +03:00
2008-11-16 07:57:26 +03:00
static int branch_trace_init ( struct trace_array * tr )
2008-11-12 23:24:24 +03:00
{
2015-10-16 16:04:49 +03:00
return enable_branch_tracing ( tr ) ;
2008-11-12 23:24:24 +03:00
}
static void branch_trace_reset ( struct trace_array * tr )
{
2015-10-16 16:04:49 +03:00
disable_branch_tracing ( ) ;
2008-11-12 23:24:24 +03:00
}
2009-02-04 03:05:50 +03:00
static enum print_line_t trace_branch_print ( struct trace_iterator * iter ,
2010-04-23 02:46:14 +04:00
int flags , struct trace_event * event )
2008-12-24 07:24:13 +03:00
{
struct trace_branch * field ;
2009-02-03 01:30:12 +03:00
trace_assign_type ( field , iter - > ent ) ;
2008-12-24 07:24:13 +03:00
2014-11-12 21:19:06 +03:00
trace_seq_printf ( & iter - > seq , " [%s] %s:%s:%d \n " ,
field - > correct ? " ok " : " MISS " ,
field - > func ,
field - > file ,
field - > line ) ;
return trace_handle_return ( & iter - > seq ) ;
2008-12-24 07:24:13 +03:00
}
2009-04-13 12:02:34 +04:00
static void branch_print_header ( struct seq_file * s )
{
seq_puts ( s , " # TASK-PID CPU# TIMESTAMP CORRECT "
2014-11-08 23:42:11 +03:00
" FUNC:FILE:LINE \n "
" # | | | | | "
" | \n " ) ;
2009-04-13 12:02:34 +04:00
}
2008-12-28 01:25:38 +03:00
2010-04-23 02:46:14 +04:00
static struct trace_event_functions trace_branch_funcs = {
. trace = trace_branch_print ,
} ;
2008-12-24 07:24:13 +03:00
static struct trace_event trace_branch_event = {
2009-03-10 21:10:56 +03:00
. type = TRACE_BRANCH ,
2010-04-23 02:46:14 +04:00
. funcs = & trace_branch_funcs ,
2008-12-24 07:24:13 +03:00
} ;
2009-01-10 22:34:13 +03:00
static struct tracer branch_trace __read_mostly =
{
. name = " branch " ,
. init = branch_trace_init ,
. reset = branch_trace_reset ,
# ifdef CONFIG_FTRACE_SELFTEST
. selftest = trace_selftest_startup_branch ,
# endif /* CONFIG_FTRACE_SELFTEST */
2009-04-13 12:02:34 +04:00
. print_header = branch_print_header ,
2009-01-10 22:34:13 +03:00
} ;
__init static int init_branch_tracer ( void )
{
int ret ;
2015-05-05 16:39:12 +03:00
ret = register_trace_event ( & trace_branch_event ) ;
2009-01-10 22:34:13 +03:00
if ( ! ret ) {
printk ( KERN_WARNING " Warning: could not register "
" branch events \n " ) ;
return 1 ;
}
return register_tracer ( & branch_trace ) ;
}
2012-10-05 20:13:07 +04:00
core_initcall ( init_branch_tracer ) ;
2009-01-10 22:34:13 +03:00
2008-11-12 08:14:40 +03:00
# else
static inline
2017-01-19 16:57:41 +03:00
void trace_likely_condition ( struct ftrace_likely_data * f , int val , int expect )
2008-11-12 08:14:40 +03:00
{
}
2008-11-12 23:24:24 +03:00
# endif /* CONFIG_BRANCH_TRACER */
2008-11-12 08:14:40 +03:00
2017-01-19 16:57:14 +03:00
void ftrace_likely_update ( struct ftrace_likely_data * f , int val ,
tracing: Process constants for (un)likely() profiler
When running the likely/unlikely profiler, one of the results did not look
accurate. It noted that the unlikely() in link_path_walk() was 100%
incorrect. When I added a trace_printk() to see what was happening there, it
became 80% correct! Looking deeper into what whas happening, I found that
gcc split that if statement into two paths. One where the if statement
became a constant, the other path a variable. The other path had the if
statement always hit (making the unlikely there, always false), but since
the #define unlikely() has:
#define unlikely() (__builtin_constant_p(x) ? !!(x) : __branch_check__(x, 0))
Where constants are ignored by the branch profiler, the "constant" path
made by the compiler was ignored, even though it was hit 80% of the time.
By just passing the constant value to the __branch_check__() function and
tracing it out of line (as always correct, as likely/unlikely isn't a factor
for constants), then we get back the accurate readings of branches that were
optimized by gcc causing part of the execution to become constant.
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
2017-01-17 20:29:35 +03:00
int expect , int is_constant )
2008-11-12 08:14:39 +03:00
{
2019-03-07 13:09:13 +03:00
unsigned long flags = user_access_save ( ) ;
tracing: Process constants for (un)likely() profiler
When running the likely/unlikely profiler, one of the results did not look
accurate. It noted that the unlikely() in link_path_walk() was 100%
incorrect. When I added a trace_printk() to see what was happening there, it
became 80% correct! Looking deeper into what whas happening, I found that
gcc split that if statement into two paths. One where the if statement
became a constant, the other path a variable. The other path had the if
statement always hit (making the unlikely there, always false), but since
the #define unlikely() has:
#define unlikely() (__builtin_constant_p(x) ? !!(x) : __branch_check__(x, 0))
Where constants are ignored by the branch profiler, the "constant" path
made by the compiler was ignored, even though it was hit 80% of the time.
By just passing the constant value to the __branch_check__() function and
tracing it out of line (as always correct, as likely/unlikely isn't a factor
for constants), then we get back the accurate readings of branches that were
optimized by gcc causing part of the execution to become constant.
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
2017-01-17 20:29:35 +03:00
/* A constant is always correct */
2017-01-19 16:57:14 +03:00
if ( is_constant ) {
f - > constant + + ;
tracing: Process constants for (un)likely() profiler
When running the likely/unlikely profiler, one of the results did not look
accurate. It noted that the unlikely() in link_path_walk() was 100%
incorrect. When I added a trace_printk() to see what was happening there, it
became 80% correct! Looking deeper into what whas happening, I found that
gcc split that if statement into two paths. One where the if statement
became a constant, the other path a variable. The other path had the if
statement always hit (making the unlikely there, always false), but since
the #define unlikely() has:
#define unlikely() (__builtin_constant_p(x) ? !!(x) : __branch_check__(x, 0))
Where constants are ignored by the branch profiler, the "constant" path
made by the compiler was ignored, even though it was hit 80% of the time.
By just passing the constant value to the __branch_check__() function and
tracing it out of line (as always correct, as likely/unlikely isn't a factor
for constants), then we get back the accurate readings of branches that were
optimized by gcc causing part of the execution to become constant.
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
2017-01-17 20:29:35 +03:00
val = expect ;
2017-01-19 16:57:14 +03:00
}
2008-11-12 08:14:40 +03:00
/*
* I would love to have a trace point here instead , but the
* trace point code is so inundated with unlikely and likely
* conditions that the recursive nightmare that exists is too
* much to try to get working . At least for now .
*/
2017-01-19 16:57:41 +03:00
trace_likely_condition ( f , val , expect ) ;
2008-11-12 08:14:40 +03:00
2008-11-12 08:14:39 +03:00
/* FIXME: Make this atomic! */
if ( val = = expect )
2017-01-19 16:57:14 +03:00
f - > data . correct + + ;
2008-11-12 08:14:39 +03:00
else
2017-01-19 16:57:14 +03:00
f - > data . incorrect + + ;
2019-03-07 13:09:13 +03:00
user_access_restore ( flags ) ;
2008-11-12 08:14:39 +03:00
}
EXPORT_SYMBOL ( ftrace_likely_update ) ;
2008-12-28 01:25:38 +03:00
extern unsigned long __start_annotated_branch_profile [ ] ;
extern unsigned long __stop_annotated_branch_profile [ ] ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
static int annotated_branch_stat_headers ( struct seq_file * m )
2008-11-12 08:14:39 +03:00
{
2014-11-08 23:42:11 +03:00
seq_puts ( m , " correct incorrect % "
" Function "
2014-11-08 23:42:10 +03:00
" File Line \n "
" ------- --------- - "
" -------- "
" ---- ---- \n " ) ;
2008-12-28 01:25:38 +03:00
return 0 ;
2008-11-12 08:14:39 +03:00
}
2019-10-07 16:56:56 +03:00
static inline long get_incorrect_percent ( const struct ftrace_branch_data * p )
2008-11-12 08:14:39 +03:00
{
2008-12-28 01:25:38 +03:00
long percent ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
if ( p - > correct ) {
percent = p - > incorrect * 100 ;
percent / = p - > correct + p - > incorrect ;
} else
percent = p - > incorrect ? 100 : - 1 ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
return percent ;
2008-11-12 08:14:39 +03:00
}
2017-01-19 16:57:14 +03:00
static const char * branch_stat_process_file ( struct ftrace_branch_data * p )
2008-11-12 08:14:39 +03:00
{
const char * f ;
/* Only print the file, not the path */
f = p - > file + strlen ( p - > file ) ;
while ( f > = p - > file & & * f ! = ' / ' )
f - - ;
2017-01-19 16:57:14 +03:00
return + + f ;
}
static void branch_stat_show ( struct seq_file * m ,
struct ftrace_branch_data * p , const char * f )
{
long percent ;
2008-11-12 08:14:39 +03:00
2008-11-21 09:30:54 +03:00
/*
* The miss is overlayed on correct , and hit on incorrect .
*/
2008-12-28 01:25:38 +03:00
percent = get_incorrect_percent ( p ) ;
2008-11-12 08:14:39 +03:00
2008-11-21 09:51:53 +03:00
if ( percent < 0 )
2014-11-08 23:42:10 +03:00
seq_puts ( m , " X " ) ;
2008-11-21 09:51:53 +03:00
else
seq_printf ( m , " %3ld " , percent ) ;
2017-01-19 16:57:14 +03:00
2008-11-12 08:14:39 +03:00
seq_printf ( m , " %-30.30s %-20.20s %d \n " , p - > func , f , p - > line ) ;
2017-01-19 16:57:14 +03:00
}
static int branch_stat_show_normal ( struct seq_file * m ,
struct ftrace_branch_data * p , const char * f )
{
seq_printf ( m , " %8lu %8lu " , p - > correct , p - > incorrect ) ;
branch_stat_show ( m , p , f ) ;
return 0 ;
}
static int annotate_branch_stat_show ( struct seq_file * m , void * v )
{
struct ftrace_likely_data * p = v ;
const char * f ;
int l ;
f = branch_stat_process_file ( & p - > data ) ;
if ( ! p - > constant )
return branch_stat_show_normal ( m , & p - > data , f ) ;
l = snprintf ( NULL , 0 , " /%lu " , p - > constant ) ;
l = l > 8 ? 0 : 8 - l ;
seq_printf ( m , " %8lu/%lu %*lu " ,
p - > data . correct , p - > constant , l , p - > data . incorrect ) ;
branch_stat_show ( m , & p - > data , f ) ;
2008-11-12 08:14:39 +03:00
return 0 ;
}
2009-03-24 20:38:36 +03:00
static void * annotated_branch_stat_start ( struct tracer_stat * trace )
2008-12-28 01:25:38 +03:00
{
return __start_annotated_branch_profile ;
}
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
static void *
annotated_branch_stat_next ( void * v , int idx )
2008-11-12 08:14:39 +03:00
{
2017-01-19 16:57:14 +03:00
struct ftrace_likely_data * p = v ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
+ + p ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
if ( ( void * ) p > = ( void * ) __stop_annotated_branch_profile )
return NULL ;
return p ;
2008-11-12 08:14:39 +03:00
}
2019-10-07 16:56:56 +03:00
static int annotated_branch_stat_cmp ( const void * p1 , const void * p2 )
2008-12-28 01:25:38 +03:00
{
2019-10-07 16:56:56 +03:00
const struct ftrace_branch_data * a = p1 ;
const struct ftrace_branch_data * b = p2 ;
2008-12-28 01:25:38 +03:00
long percent_a , percent_b ;
percent_a = get_incorrect_percent ( a ) ;
percent_b = get_incorrect_percent ( b ) ;
if ( percent_a < percent_b )
return - 1 ;
if ( percent_a > percent_b )
return 1 ;
2010-01-27 19:25:54 +03:00
if ( a - > incorrect < b - > incorrect )
return - 1 ;
if ( a - > incorrect > b - > incorrect )
return 1 ;
/*
* Since the above shows worse ( incorrect ) cases
* first , we continue that by showing best ( correct )
* cases last .
*/
if ( a - > correct > b - > correct )
return - 1 ;
if ( a - > correct < b - > correct )
return 1 ;
return 0 ;
2008-12-28 01:25:38 +03:00
}
2008-11-12 08:14:39 +03:00
2009-01-10 22:34:13 +03:00
static struct tracer_stat annotated_branch_stats = {
. name = " branch_annotated " ,
. stat_start = annotated_branch_stat_start ,
. stat_next = annotated_branch_stat_next ,
. stat_cmp = annotated_branch_stat_cmp ,
. stat_headers = annotated_branch_stat_headers ,
2017-01-19 16:57:14 +03:00
. stat_show = annotate_branch_stat_show
2009-01-10 22:34:13 +03:00
} ;
__init static int init_annotated_branch_stats ( void )
{
int ret ;
ret = register_stat_tracer ( & annotated_branch_stats ) ;
if ( ! ret ) {
printk ( KERN_WARNING " Warning: could not register "
" annotated branches stats \n " ) ;
return 1 ;
}
return 0 ;
}
fs_initcall ( init_annotated_branch_stats ) ;
2008-11-21 09:30:54 +03:00
# ifdef CONFIG_PROFILE_ALL_BRANCHES
2008-12-28 01:25:38 +03:00
extern unsigned long __start_branch_profile [ ] ;
extern unsigned long __stop_branch_profile [ ] ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
static int all_branch_stat_headers ( struct seq_file * m )
{
2014-11-08 23:42:11 +03:00
seq_puts ( m , " miss hit % "
" Function "
2014-11-08 23:42:10 +03:00
" File Line \n "
" ------- --------- - "
" -------- "
" ---- ---- \n " ) ;
2008-12-28 01:25:38 +03:00
return 0 ;
}
2008-11-12 08:14:39 +03:00
2009-03-24 20:38:36 +03:00
static void * all_branch_stat_start ( struct tracer_stat * trace )
2008-11-12 08:14:39 +03:00
{
2008-12-28 01:25:38 +03:00
return __start_branch_profile ;
}
static void *
all_branch_stat_next ( void * v , int idx )
{
struct ftrace_branch_data * p = v ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
+ + p ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
if ( ( void * ) p > = ( void * ) __stop_branch_profile )
return NULL ;
2008-11-12 08:14:39 +03:00
2008-12-28 01:25:38 +03:00
return p ;
}
2008-11-21 09:30:54 +03:00
2017-01-19 16:57:14 +03:00
static int all_branch_stat_show ( struct seq_file * m , void * v )
{
struct ftrace_branch_data * p = v ;
const char * f ;
f = branch_stat_process_file ( p ) ;
return branch_stat_show_normal ( m , p , f ) ;
}
2009-01-10 22:34:13 +03:00
static struct tracer_stat all_branch_stats = {
. name = " branch_all " ,
2009-01-08 21:03:56 +03:00
. stat_start = all_branch_stat_start ,
. stat_next = all_branch_stat_next ,
. stat_headers = all_branch_stat_headers ,
2017-01-19 16:57:14 +03:00
. stat_show = all_branch_stat_show
2009-01-08 21:03:56 +03:00
} ;
2008-12-28 01:25:38 +03:00
2009-01-10 22:34:13 +03:00
__init static int all_annotated_branch_stats ( void )
2008-12-28 01:25:38 +03:00
{
int ret ;
2009-01-10 22:34:13 +03:00
ret = register_stat_tracer ( & all_branch_stats ) ;
2008-12-28 01:25:38 +03:00
if ( ! ret ) {
2009-01-10 22:34:13 +03:00
printk ( KERN_WARNING " Warning: could not register "
" all branches stats \n " ) ;
2008-12-28 01:25:38 +03:00
return 1 ;
}
2009-01-10 22:34:13 +03:00
return 0 ;
2008-12-28 01:25:38 +03:00
}
2009-01-10 22:34:13 +03:00
fs_initcall ( all_annotated_branch_stats ) ;
# endif /* CONFIG_PROFILE_ALL_BRANCHES */