License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 17:07:57 +03:00
// SPDX-License-Identifier: GPL-2.0
2006-07-03 11:24:48 +04:00
/*
* lib / locking - selftest . c
*
* Testsuite for various locking APIs : spinlocks , rwlocks ,
* mutexes and rw - semaphores .
*
* It is checking both false positives and false negatives .
*
* Started by Ingo Molnar :
*
* Copyright ( C ) 2006 Red Hat , Inc . , Ingo Molnar < mingo @ redhat . com >
*/
# include <linux/rwsem.h>
# include <linux/mutex.h>
2013-07-05 11:29:32 +04:00
# include <linux/ww_mutex.h>
2006-07-03 11:24:48 +04:00
# include <linux/sched.h>
# include <linux/delay.h>
[PATCH] lockdep: core
Do 'make oldconfig' and accept all the defaults for new config options -
reboot into the kernel and if everything goes well it should boot up fine and
you should have /proc/lockdep and /proc/lockdep_stats files.
Typically if the lock validator finds some problem it will print out
voluminous debug output that begins with "BUG: ..." and which syslog output
can be used by kernel developers to figure out the precise locking scenario.
What does the lock validator do? It "observes" and maps all locking rules as
they occur dynamically (as triggered by the kernel's natural use of spinlocks,
rwlocks, mutexes and rwsems). Whenever the lock validator subsystem detects a
new locking scenario, it validates this new rule against the existing set of
rules. If this new rule is consistent with the existing set of rules then the
new rule is added transparently and the kernel continues as normal. If the
new rule could create a deadlock scenario then this condition is printed out.
When determining validity of locking, all possible "deadlock scenarios" are
considered: assuming arbitrary number of CPUs, arbitrary irq context and task
context constellations, running arbitrary combinations of all the existing
locking scenarios. In a typical system this means millions of separate
scenarios. This is why we call it a "locking correctness" validator - for all
rules that are observed the lock validator proves it with mathematical
certainty that a deadlock could not occur (assuming that the lock validator
implementation itself is correct and its internal data structures are not
corrupted by some other kernel subsystem). [see more details and conditionals
of this statement in include/linux/lockdep.h and
Documentation/lockdep-design.txt]
Furthermore, this "all possible scenarios" property of the validator also
enables the finding of complex, highly unlikely multi-CPU multi-context races
via single single-context rules, increasing the likelyhood of finding bugs
drastically. In practical terms: the lock validator already found a bug in
the upstream kernel that could only occur on systems with 3 or more CPUs, and
which needed 3 very unlikely code sequences to occur at once on the 3 CPUs.
That bug was found and reported on a single-CPU system (!). So in essence a
race will be found "piecemail-wise", triggering all the necessary components
for the race, without having to reproduce the race scenario itself! In its
short existence the lock validator found and reported many bugs before they
actually caused a real deadlock.
To further increase the efficiency of the validator, the mapping is not per
"lock instance", but per "lock-class". For example, all struct inode objects
in the kernel have inode->inotify_mutex. If there are 10,000 inodes cached,
then there are 10,000 lock objects. But ->inotify_mutex is a single "lock
type", and all locking activities that occur against ->inotify_mutex are
"unified" into this single lock-class. The advantage of the lock-class
approach is that all historical ->inotify_mutex uses are mapped into a single
(and as narrow as possible) set of locking rules - regardless of how many
different tasks or inode structures it took to build this set of rules. The
set of rules persist during the lifetime of the kernel.
To see the rough magnitude of checking that the lock validator does, here's a
portion of /proc/lockdep_stats, fresh after bootup:
lock-classes: 694 [max: 2048]
direct dependencies: 1598 [max: 8192]
indirect dependencies: 17896
all direct dependencies: 16206
dependency chains: 1910 [max: 8192]
in-hardirq chains: 17
in-softirq chains: 105
in-process chains: 1065
stack-trace entries: 38761 [max: 131072]
combined max dependencies: 2033928
hardirq-safe locks: 24
hardirq-unsafe locks: 176
softirq-safe locks: 53
softirq-unsafe locks: 137
irq-safe locks: 59
irq-unsafe locks: 176
The lock validator has observed 1598 actual single-thread locking patterns,
and has validated all possible 2033928 distinct locking scenarios.
More details about the design of the lock validator can be found in
Documentation/lockdep-design.txt, which can also found at:
http://redhat.com/~mingo/lockdep-patches/lockdep-design.txt
[bunk@stusta.de: cleanups]
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Arjan van de Ven <arjan@linux.intel.com>
Signed-off-by: Adrian Bunk <bunk@stusta.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-07-03 11:24:50 +04:00
# include <linux/lockdep.h>
2006-07-03 11:24:48 +04:00
# include <linux/spinlock.h>
# include <linux/kallsyms.h>
# include <linux/interrupt.h>
# include <linux/debug_locks.h>
# include <linux/irqflags.h>
2017-05-31 18:11:49 +03:00
# include <linux/rtmutex.h>
2006-07-03 11:24:48 +04:00
/*
* Change this to 1 if you want to see the failure printouts :
*/
static unsigned int debug_locks_verbose ;
2018-06-15 11:17:38 +03:00
static DEFINE_WD_CLASS ( ww_lockdep ) ;
2013-06-20 15:31:24 +04:00
2006-07-03 11:24:48 +04:00
static int __init setup_debug_locks_verbose ( char * str )
{
get_option ( & str , & debug_locks_verbose ) ;
return 1 ;
}
__setup ( " debug_locks_verbose= " , setup_debug_locks_verbose ) ;
# define FAILURE 0
# define SUCCESS 1
# define LOCKTYPE_SPIN 0x1
# define LOCKTYPE_RWLOCK 0x2
# define LOCKTYPE_MUTEX 0x4
# define LOCKTYPE_RWSEM 0x8
2013-06-20 15:31:24 +04:00
# define LOCKTYPE_WW 0x10
2017-05-31 18:11:49 +03:00
# define LOCKTYPE_RTMUTEX 0x20
2013-06-20 15:31:24 +04:00
static struct ww_acquire_ctx t , t2 ;
2013-06-20 15:31:42 +04:00
static struct ww_mutex o , o2 , o3 ;
2006-07-03 11:24:48 +04:00
/*
* Normal standalone locks , for the circular and irq - context
* dependency tests :
*/
2012-04-16 11:01:55 +04:00
static DEFINE_RAW_SPINLOCK ( lock_A ) ;
static DEFINE_RAW_SPINLOCK ( lock_B ) ;
static DEFINE_RAW_SPINLOCK ( lock_C ) ;
static DEFINE_RAW_SPINLOCK ( lock_D ) ;
2006-07-03 11:24:48 +04:00
static DEFINE_RWLOCK ( rwlock_A ) ;
static DEFINE_RWLOCK ( rwlock_B ) ;
static DEFINE_RWLOCK ( rwlock_C ) ;
static DEFINE_RWLOCK ( rwlock_D ) ;
static DEFINE_MUTEX ( mutex_A ) ;
static DEFINE_MUTEX ( mutex_B ) ;
static DEFINE_MUTEX ( mutex_C ) ;
static DEFINE_MUTEX ( mutex_D ) ;
static DECLARE_RWSEM ( rwsem_A ) ;
static DECLARE_RWSEM ( rwsem_B ) ;
static DECLARE_RWSEM ( rwsem_C ) ;
static DECLARE_RWSEM ( rwsem_D ) ;
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
static DEFINE_RT_MUTEX ( rtmutex_A ) ;
static DEFINE_RT_MUTEX ( rtmutex_B ) ;
static DEFINE_RT_MUTEX ( rtmutex_C ) ;
static DEFINE_RT_MUTEX ( rtmutex_D ) ;
# endif
2006-07-03 11:24:48 +04:00
/*
* Locks that we initialize dynamically as well so that
* e . g . X1 and X2 becomes two instances of the same class ,
* but X * and Y * are different classes . We do this so that
* we do not trigger a real lockup :
*/
2012-04-16 11:01:55 +04:00
static DEFINE_RAW_SPINLOCK ( lock_X1 ) ;
static DEFINE_RAW_SPINLOCK ( lock_X2 ) ;
static DEFINE_RAW_SPINLOCK ( lock_Y1 ) ;
static DEFINE_RAW_SPINLOCK ( lock_Y2 ) ;
static DEFINE_RAW_SPINLOCK ( lock_Z1 ) ;
static DEFINE_RAW_SPINLOCK ( lock_Z2 ) ;
2006-07-03 11:24:48 +04:00
static DEFINE_RWLOCK ( rwlock_X1 ) ;
static DEFINE_RWLOCK ( rwlock_X2 ) ;
static DEFINE_RWLOCK ( rwlock_Y1 ) ;
static DEFINE_RWLOCK ( rwlock_Y2 ) ;
static DEFINE_RWLOCK ( rwlock_Z1 ) ;
static DEFINE_RWLOCK ( rwlock_Z2 ) ;
static DEFINE_MUTEX ( mutex_X1 ) ;
static DEFINE_MUTEX ( mutex_X2 ) ;
static DEFINE_MUTEX ( mutex_Y1 ) ;
static DEFINE_MUTEX ( mutex_Y2 ) ;
static DEFINE_MUTEX ( mutex_Z1 ) ;
static DEFINE_MUTEX ( mutex_Z2 ) ;
static DECLARE_RWSEM ( rwsem_X1 ) ;
static DECLARE_RWSEM ( rwsem_X2 ) ;
static DECLARE_RWSEM ( rwsem_Y1 ) ;
static DECLARE_RWSEM ( rwsem_Y2 ) ;
static DECLARE_RWSEM ( rwsem_Z1 ) ;
static DECLARE_RWSEM ( rwsem_Z2 ) ;
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
static DEFINE_RT_MUTEX ( rtmutex_X1 ) ;
static DEFINE_RT_MUTEX ( rtmutex_X2 ) ;
static DEFINE_RT_MUTEX ( rtmutex_Y1 ) ;
static DEFINE_RT_MUTEX ( rtmutex_Y2 ) ;
static DEFINE_RT_MUTEX ( rtmutex_Z1 ) ;
static DEFINE_RT_MUTEX ( rtmutex_Z2 ) ;
# endif
2006-07-03 11:24:48 +04:00
/*
* non - inlined runtime initializers , to let separate locks share
* the same lock - class :
*/
# define INIT_CLASS_FUNC(class) \
static noinline void \
2012-04-16 11:01:55 +04:00
init_class_ # # class ( raw_spinlock_t * lock , rwlock_t * rwlock , \
struct mutex * mutex , struct rw_semaphore * rwsem ) \
2006-07-03 11:24:48 +04:00
{ \
2012-04-16 11:01:55 +04:00
raw_spin_lock_init ( lock ) ; \
2006-07-03 11:24:48 +04:00
rwlock_init ( rwlock ) ; \
mutex_init ( mutex ) ; \
init_rwsem ( rwsem ) ; \
}
INIT_CLASS_FUNC ( X )
INIT_CLASS_FUNC ( Y )
INIT_CLASS_FUNC ( Z )
static void init_shared_classes ( void )
{
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
static struct lock_class_key rt_X , rt_Y , rt_Z ;
__rt_mutex_init ( & rtmutex_X1 , __func__ , & rt_X ) ;
__rt_mutex_init ( & rtmutex_X2 , __func__ , & rt_X ) ;
__rt_mutex_init ( & rtmutex_Y1 , __func__ , & rt_Y ) ;
__rt_mutex_init ( & rtmutex_Y2 , __func__ , & rt_Y ) ;
__rt_mutex_init ( & rtmutex_Z1 , __func__ , & rt_Z ) ;
__rt_mutex_init ( & rtmutex_Z2 , __func__ , & rt_Z ) ;
# endif
2006-07-03 11:24:48 +04:00
init_class_X ( & lock_X1 , & rwlock_X1 , & mutex_X1 , & rwsem_X1 ) ;
init_class_X ( & lock_X2 , & rwlock_X2 , & mutex_X2 , & rwsem_X2 ) ;
init_class_Y ( & lock_Y1 , & rwlock_Y1 , & mutex_Y1 , & rwsem_Y1 ) ;
init_class_Y ( & lock_Y2 , & rwlock_Y2 , & mutex_Y2 , & rwsem_Y2 ) ;
init_class_Z ( & lock_Z1 , & rwlock_Z1 , & mutex_Z1 , & rwsem_Z1 ) ;
init_class_Z ( & lock_Z2 , & rwlock_Z2 , & mutex_Z2 , & rwsem_Z2 ) ;
}
/*
* For spinlocks and rwlocks we also do hardirq - safe / softirq - safe tests .
* The following functions use a lock from a simulated hardirq / softirq
* context , causing the locks to be marked as hardirq - safe / softirq - safe :
*/
# define HARDIRQ_DISABLE local_irq_disable
# define HARDIRQ_ENABLE local_irq_enable
# define HARDIRQ_ENTER() \
local_irq_disable ( ) ; \
2011-05-20 04:09:54 +04:00
__irq_enter ( ) ; \
2006-07-03 11:24:48 +04:00
WARN_ON ( ! in_irq ( ) ) ;
# define HARDIRQ_EXIT() \
__irq_exit ( ) ; \
local_irq_enable ( ) ;
# define SOFTIRQ_DISABLE local_bh_disable
# define SOFTIRQ_ENABLE local_bh_enable
# define SOFTIRQ_ENTER() \
local_bh_disable ( ) ; \
local_irq_disable ( ) ; \
2009-03-13 03:30:40 +03:00
lockdep_softirq_enter ( ) ; \
2006-07-03 11:24:48 +04:00
WARN_ON ( ! in_softirq ( ) ) ;
# define SOFTIRQ_EXIT() \
2009-03-13 03:30:40 +03:00
lockdep_softirq_exit ( ) ; \
2006-07-03 11:24:48 +04:00
local_irq_enable ( ) ; \
local_bh_enable ( ) ;
/*
* Shortcuts for lock / unlock API variants , to keep
* the testcases compact :
*/
2012-04-16 11:01:55 +04:00
# define L(x) raw_spin_lock(&lock_##x)
# define U(x) raw_spin_unlock(&lock_##x)
2006-07-03 11:24:48 +04:00
# define LU(x) L(x); U(x)
2012-04-16 11:01:55 +04:00
# define SI(x) raw_spin_lock_init(&lock_##x)
2006-07-03 11:24:48 +04:00
# define WL(x) write_lock(&rwlock_##x)
# define WU(x) write_unlock(&rwlock_##x)
# define WLU(x) WL(x); WU(x)
# define RL(x) read_lock(&rwlock_##x)
# define RU(x) read_unlock(&rwlock_##x)
# define RLU(x) RL(x); RU(x)
# define RWI(x) rwlock_init(&rwlock_##x)
# define ML(x) mutex_lock(&mutex_##x)
# define MU(x) mutex_unlock(&mutex_##x)
# define MI(x) mutex_init(&mutex_##x)
2017-05-31 18:11:49 +03:00
# define RTL(x) rt_mutex_lock(&rtmutex_##x)
# define RTU(x) rt_mutex_unlock(&rtmutex_##x)
# define RTI(x) rt_mutex_init(&rtmutex_##x)
2006-07-03 11:24:48 +04:00
# define WSL(x) down_write(&rwsem_##x)
# define WSU(x) up_write(&rwsem_##x)
# define RSL(x) down_read(&rwsem_##x)
# define RSU(x) up_read(&rwsem_##x)
# define RWSI(x) init_rwsem(&rwsem_##x)
2013-06-20 15:31:24 +04:00
# ifndef CONFIG_DEBUG_WW_MUTEX_SLOWPATH
# define WWAI(x) ww_acquire_init(x, &ww_lockdep)
# else
# define WWAI(x) do { ww_acquire_init(x, &ww_lockdep); (x)->deadlock_inject_countdown = ~0U; } while (0)
# endif
# define WWAD(x) ww_acquire_done(x)
# define WWAF(x) ww_acquire_fini(x)
# define WWL(x, c) ww_mutex_lock(x, c)
# define WWT(x) ww_mutex_trylock(x)
# define WWL1(x) ww_mutex_lock(x, NULL)
# define WWU(x) ww_mutex_unlock(x)
2006-07-03 11:24:48 +04:00
# define LOCK_UNLOCK_2(x,y) LOCK(x); LOCK(y); UNLOCK(y); UNLOCK(x)
/*
* Generate different permutations of the same testcase , using
* the same basic lock - dependency / state events :
*/
# define GENERATE_TESTCASE(name) \
\
static void name ( void ) { E ( ) ; }
# define GENERATE_PERMUTATIONS_2_EVENTS(name) \
\
static void name # # _12 ( void ) { E1 ( ) ; E2 ( ) ; } \
static void name # # _21 ( void ) { E2 ( ) ; E1 ( ) ; }
# define GENERATE_PERMUTATIONS_3_EVENTS(name) \
\
static void name # # _123 ( void ) { E1 ( ) ; E2 ( ) ; E3 ( ) ; } \
static void name # # _132 ( void ) { E1 ( ) ; E3 ( ) ; E2 ( ) ; } \
static void name # # _213 ( void ) { E2 ( ) ; E1 ( ) ; E3 ( ) ; } \
static void name # # _231 ( void ) { E2 ( ) ; E3 ( ) ; E1 ( ) ; } \
static void name # # _312 ( void ) { E3 ( ) ; E1 ( ) ; E2 ( ) ; } \
static void name # # _321 ( void ) { E3 ( ) ; E2 ( ) ; E1 ( ) ; }
/*
* AA deadlock :
*/
# define E() \
\
LOCK ( X1 ) ; \
LOCK ( X2 ) ; /* this one should fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( AA_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( AA_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( AA_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( AA_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( AA_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( AA_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( AA_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* Special - case for read - locking , they are
2006-07-03 11:24:51 +04:00
* allowed to recurse on the same lock class :
2006-07-03 11:24:48 +04:00
*/
static void rlock_AA1 ( void )
{
RL ( X1 ) ;
RL ( X1 ) ; // this one should NOT fail
}
static void rlock_AA1B ( void )
{
RL ( X1 ) ;
2006-07-03 11:24:51 +04:00
RL ( X2 ) ; // this one should NOT fail
2006-07-03 11:24:48 +04:00
}
static void rsem_AA1 ( void )
{
RSL ( X1 ) ;
RSL ( X1 ) ; // this one should fail
}
static void rsem_AA1B ( void )
{
RSL ( X1 ) ;
RSL ( X2 ) ; // this one should fail
}
/*
* The mixing of read and write locks is not allowed :
*/
static void rlock_AA2 ( void )
{
RL ( X1 ) ;
WL ( X2 ) ; // this one should fail
}
static void rsem_AA2 ( void )
{
RSL ( X1 ) ;
WSL ( X2 ) ; // this one should fail
}
static void rlock_AA3 ( void )
{
WL ( X1 ) ;
RL ( X2 ) ; // this one should fail
}
static void rsem_AA3 ( void )
{
WSL ( X1 ) ;
RSL ( X2 ) ; // this one should fail
}
2017-08-23 14:13:11 +03:00
/*
* read_lock ( A )
* spin_lock ( B )
* spin_lock ( B )
* write_lock ( A )
*/
static void rlock_ABBA1 ( void )
{
RL ( X1 ) ;
L ( Y1 ) ;
U ( Y1 ) ;
RU ( X1 ) ;
L ( Y1 ) ;
WL ( X1 ) ;
WU ( X1 ) ;
U ( Y1 ) ; // should fail
}
static void rwsem_ABBA1 ( void )
{
RSL ( X1 ) ;
ML ( Y1 ) ;
MU ( Y1 ) ;
RSU ( X1 ) ;
ML ( Y1 ) ;
WSL ( X1 ) ;
WSU ( X1 ) ;
MU ( Y1 ) ; // should fail
}
/*
* read_lock ( A )
* spin_lock ( B )
* spin_lock ( B )
* read_lock ( A )
*/
static void rlock_ABBA2 ( void )
{
RL ( X1 ) ;
L ( Y1 ) ;
U ( Y1 ) ;
RU ( X1 ) ;
L ( Y1 ) ;
RL ( X1 ) ;
RU ( X1 ) ;
U ( Y1 ) ; // should NOT fail
}
static void rwsem_ABBA2 ( void )
{
RSL ( X1 ) ;
ML ( Y1 ) ;
MU ( Y1 ) ;
RSU ( X1 ) ;
ML ( Y1 ) ;
RSL ( X1 ) ;
RSU ( X1 ) ;
MU ( Y1 ) ; // should fail
}
/*
* write_lock ( A )
* spin_lock ( B )
* spin_lock ( B )
* write_lock ( A )
*/
static void rlock_ABBA3 ( void )
{
WL ( X1 ) ;
L ( Y1 ) ;
U ( Y1 ) ;
WU ( X1 ) ;
L ( Y1 ) ;
WL ( X1 ) ;
WU ( X1 ) ;
U ( Y1 ) ; // should fail
}
static void rwsem_ABBA3 ( void )
{
WSL ( X1 ) ;
ML ( Y1 ) ;
MU ( Y1 ) ;
WSU ( X1 ) ;
ML ( Y1 ) ;
WSL ( X1 ) ;
WSU ( X1 ) ;
MU ( Y1 ) ; // should fail
}
2006-07-03 11:24:48 +04:00
/*
* ABBA deadlock :
*/
# define E() \
\
LOCK_UNLOCK_2 ( A , B ) ; \
LOCK_UNLOCK_2 ( B , A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( ABBA_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( ABBA_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( ABBA_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( ABBA_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( ABBA_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( ABBA_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( ABBA_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* AB BC CA deadlock :
*/
# define E() \
\
LOCK_UNLOCK_2 ( A , B ) ; \
LOCK_UNLOCK_2 ( B , C ) ; \
LOCK_UNLOCK_2 ( C , A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( ABBCCA_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( ABBCCA_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( ABBCCA_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( ABBCCA_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( ABBCCA_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( ABBCCA_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( ABBCCA_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* AB CA BC deadlock :
*/
# define E() \
\
LOCK_UNLOCK_2 ( A , B ) ; \
LOCK_UNLOCK_2 ( C , A ) ; \
LOCK_UNLOCK_2 ( B , C ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( ABCABC_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( ABCABC_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( ABCABC_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( ABCABC_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( ABCABC_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( ABCABC_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( ABCABC_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* AB BC CD DA deadlock :
*/
# define E() \
\
LOCK_UNLOCK_2 ( A , B ) ; \
LOCK_UNLOCK_2 ( B , C ) ; \
LOCK_UNLOCK_2 ( C , D ) ; \
LOCK_UNLOCK_2 ( D , A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( ABBCCDDA_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( ABBCCDDA_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( ABBCCDDA_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( ABBCCDDA_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( ABBCCDDA_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( ABBCCDDA_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( ABBCCDDA_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* AB CD BD DA deadlock :
*/
# define E() \
\
LOCK_UNLOCK_2 ( A , B ) ; \
LOCK_UNLOCK_2 ( C , D ) ; \
LOCK_UNLOCK_2 ( B , D ) ; \
LOCK_UNLOCK_2 ( D , A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( ABCDBDDA_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( ABCDBDDA_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( ABCDBDDA_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( ABCDBDDA_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( ABCDBDDA_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( ABCDBDDA_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( ABCDBDDA_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* AB CD BC DA deadlock :
*/
# define E() \
\
LOCK_UNLOCK_2 ( A , B ) ; \
LOCK_UNLOCK_2 ( C , D ) ; \
LOCK_UNLOCK_2 ( B , C ) ; \
LOCK_UNLOCK_2 ( D , A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( ABCDBCDA_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( ABCDBCDA_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( ABCDBCDA_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( ABCDBCDA_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( ABCDBCDA_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( ABCDBCDA_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( ABCDBCDA_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* Double unlock :
*/
# define E() \
\
LOCK ( A ) ; \
UNLOCK ( A ) ; \
UNLOCK ( A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( double_unlock_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( double_unlock_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( double_unlock_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( double_unlock_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( double_unlock_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( double_unlock_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( double_unlock_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* initializing a held lock :
*/
# define E() \
\
LOCK ( A ) ; \
INIT ( A ) ; /* fail */
/*
* 6 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_TESTCASE ( init_held_spin )
# include "locking-selftest-wlock.h"
GENERATE_TESTCASE ( init_held_wlock )
# include "locking-selftest-rlock.h"
GENERATE_TESTCASE ( init_held_rlock )
# include "locking-selftest-mutex.h"
GENERATE_TESTCASE ( init_held_mutex )
# include "locking-selftest-wsem.h"
GENERATE_TESTCASE ( init_held_wsem )
# include "locking-selftest-rsem.h"
GENERATE_TESTCASE ( init_held_rsem )
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# include "locking-selftest-rtmutex.h"
GENERATE_TESTCASE ( init_held_rtmutex ) ;
# endif
2006-07-03 11:24:48 +04:00
# undef E
/*
* locking an irq - safe lock with irqs enabled :
*/
# define E1() \
\
IRQ_ENTER ( ) ; \
LOCK ( A ) ; \
UNLOCK ( A ) ; \
IRQ_EXIT ( ) ;
# define E2() \
\
LOCK ( A ) ; \
UNLOCK ( A ) ;
/*
* Generate 24 testcases :
*/
# include "locking-selftest-spin-hardirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe1_hard_spin )
# include "locking-selftest-rlock-hardirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe1_hard_rlock )
# include "locking-selftest-wlock-hardirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe1_hard_wlock )
# include "locking-selftest-spin-softirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe1_soft_spin )
# include "locking-selftest-rlock-softirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe1_soft_rlock )
# include "locking-selftest-wlock-softirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe1_soft_wlock )
# undef E1
# undef E2
/*
* Enabling hardirqs with a softirq - safe lock held :
*/
# define E1() \
\
SOFTIRQ_ENTER ( ) ; \
LOCK ( A ) ; \
UNLOCK ( A ) ; \
SOFTIRQ_EXIT ( ) ;
# define E2() \
\
HARDIRQ_DISABLE ( ) ; \
LOCK ( A ) ; \
HARDIRQ_ENABLE ( ) ; \
UNLOCK ( A ) ;
/*
* Generate 12 testcases :
*/
# include "locking-selftest-spin.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2A_spin )
# include "locking-selftest-wlock.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2A_wlock )
# include "locking-selftest-rlock.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2A_rlock )
# undef E1
# undef E2
/*
* Enabling irqs with an irq - safe lock held :
*/
# define E1() \
\
IRQ_ENTER ( ) ; \
LOCK ( A ) ; \
UNLOCK ( A ) ; \
IRQ_EXIT ( ) ;
# define E2() \
\
IRQ_DISABLE ( ) ; \
LOCK ( A ) ; \
IRQ_ENABLE ( ) ; \
UNLOCK ( A ) ;
/*
* Generate 24 testcases :
*/
# include "locking-selftest-spin-hardirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2B_hard_spin )
# include "locking-selftest-rlock-hardirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2B_hard_rlock )
# include "locking-selftest-wlock-hardirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2B_hard_wlock )
# include "locking-selftest-spin-softirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2B_soft_spin )
# include "locking-selftest-rlock-softirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2B_soft_rlock )
# include "locking-selftest-wlock-softirq.h"
GENERATE_PERMUTATIONS_2_EVENTS ( irqsafe2B_soft_wlock )
# undef E1
# undef E2
/*
* Acquiring a irq - unsafe lock while holding an irq - safe - lock :
*/
# define E1() \
\
LOCK ( A ) ; \
LOCK ( B ) ; \
UNLOCK ( B ) ; \
UNLOCK ( A ) ; \
# define E2() \
\
LOCK ( B ) ; \
UNLOCK ( B ) ;
# define E3() \
\
IRQ_ENTER ( ) ; \
LOCK ( A ) ; \
UNLOCK ( A ) ; \
IRQ_EXIT ( ) ;
/*
* Generate 36 testcases :
*/
# include "locking-selftest-spin-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe3_hard_spin )
# include "locking-selftest-rlock-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe3_hard_rlock )
# include "locking-selftest-wlock-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe3_hard_wlock )
# include "locking-selftest-spin-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe3_soft_spin )
# include "locking-selftest-rlock-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe3_soft_rlock )
# include "locking-selftest-wlock-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe3_soft_wlock )
# undef E1
# undef E2
# undef E3
/*
* If a lock turns into softirq - safe , but earlier it took
* a softirq - unsafe lock :
*/
# define E1() \
IRQ_DISABLE ( ) ; \
LOCK ( A ) ; \
LOCK ( B ) ; \
UNLOCK ( B ) ; \
UNLOCK ( A ) ; \
IRQ_ENABLE ( ) ;
# define E2() \
LOCK ( B ) ; \
UNLOCK ( B ) ;
# define E3() \
IRQ_ENTER ( ) ; \
LOCK ( A ) ; \
UNLOCK ( A ) ; \
IRQ_EXIT ( ) ;
/*
* Generate 36 testcases :
*/
# include "locking-selftest-spin-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe4_hard_spin )
# include "locking-selftest-rlock-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe4_hard_rlock )
# include "locking-selftest-wlock-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe4_hard_wlock )
# include "locking-selftest-spin-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe4_soft_spin )
# include "locking-selftest-rlock-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe4_soft_rlock )
# include "locking-selftest-wlock-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irqsafe4_soft_wlock )
# undef E1
# undef E2
# undef E3
/*
* read - lock / write - lock irq inversion .
*
* Deadlock scenario :
*
* CPU # 1 is at # 1 , i . e . it has write - locked A , but has not
* taken B yet .
*
* CPU # 2 is at # 2 , i . e . it has locked B .
*
* Hardirq hits CPU # 2 at point # 2 and is trying to read - lock A .
*
* The deadlock occurs because CPU # 1 will spin on B , and CPU # 2
* will spin on A .
*/
# define E1() \
\
IRQ_DISABLE ( ) ; \
WL ( A ) ; \
LOCK ( B ) ; \
UNLOCK ( B ) ; \
WU ( A ) ; \
IRQ_ENABLE ( ) ;
# define E2() \
\
LOCK ( B ) ; \
UNLOCK ( B ) ;
# define E3() \
\
IRQ_ENTER ( ) ; \
RL ( A ) ; \
RU ( A ) ; \
IRQ_EXIT ( ) ;
/*
* Generate 36 testcases :
*/
# include "locking-selftest-spin-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_inversion_hard_spin )
# include "locking-selftest-rlock-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_inversion_hard_rlock )
# include "locking-selftest-wlock-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_inversion_hard_wlock )
# include "locking-selftest-spin-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_inversion_soft_spin )
# include "locking-selftest-rlock-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_inversion_soft_rlock )
# include "locking-selftest-wlock-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_inversion_soft_wlock )
# undef E1
# undef E2
# undef E3
/*
* read - lock / write - lock recursion that is actually safe .
*/
# define E1() \
\
IRQ_DISABLE ( ) ; \
WL ( A ) ; \
WU ( A ) ; \
IRQ_ENABLE ( ) ;
# define E2() \
\
RL ( A ) ; \
RU ( A ) ; \
# define E3() \
\
IRQ_ENTER ( ) ; \
RL ( A ) ; \
L ( B ) ; \
U ( B ) ; \
RU ( A ) ; \
IRQ_EXIT ( ) ;
/*
* Generate 12 testcases :
*/
# include "locking-selftest-hardirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_read_recursion_hard )
# include "locking-selftest-softirq.h"
GENERATE_PERMUTATIONS_3_EVENTS ( irq_read_recursion_soft )
# undef E1
# undef E2
# undef E3
/*
* read - lock / write - lock recursion that is unsafe .
*/
# define E1() \
\
IRQ_DISABLE ( ) ; \
L ( B ) ; \
WL ( A ) ; \
WU ( A ) ; \
U ( B ) ; \
IRQ_ENABLE ( ) ;
# define E2() \
\
RL ( A ) ; \
RU ( A ) ; \
# define E3() \
\
IRQ_ENTER ( ) ; \
L ( B ) ; \
U ( B ) ; \
IRQ_EXIT ( ) ;
/*
* Generate 12 testcases :
*/
# include "locking-selftest-hardirq.h"
// GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_hard)
# include "locking-selftest-softirq.h"
// GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_soft)
# ifdef CONFIG_DEBUG_LOCK_ALLOC
# define I_SPINLOCK(x) lockdep_reset_lock(&lock_##x.dep_map)
# define I_RWLOCK(x) lockdep_reset_lock(&rwlock_##x.dep_map)
# define I_MUTEX(x) lockdep_reset_lock(&mutex_##x.dep_map)
# define I_RWSEM(x) lockdep_reset_lock(&rwsem_##x.dep_map)
2013-06-20 15:31:24 +04:00
# define I_WW(x) lockdep_reset_lock(&x.dep_map)
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# define I_RTMUTEX(x) lockdep_reset_lock(&rtmutex_##x.dep_map)
# endif
2006-07-03 11:24:48 +04:00
# else
# define I_SPINLOCK(x)
# define I_RWLOCK(x)
# define I_MUTEX(x)
# define I_RWSEM(x)
2013-06-20 15:31:24 +04:00
# define I_WW(x)
2006-07-03 11:24:48 +04:00
# endif
2017-05-31 18:11:49 +03:00
# ifndef I_RTMUTEX
# define I_RTMUTEX(x)
# endif
# ifdef CONFIG_RT_MUTEXES
# define I2_RTMUTEX(x) rt_mutex_init(&rtmutex_##x)
# else
# define I2_RTMUTEX(x)
# endif
2006-07-03 11:24:48 +04:00
# define I1(x) \
do { \
I_SPINLOCK ( x ) ; \
I_RWLOCK ( x ) ; \
I_MUTEX ( x ) ; \
I_RWSEM ( x ) ; \
2017-05-31 18:11:49 +03:00
I_RTMUTEX ( x ) ; \
2006-07-03 11:24:48 +04:00
} while ( 0 )
# define I2(x) \
do { \
2012-04-16 11:01:55 +04:00
raw_spin_lock_init ( & lock_ # # x ) ; \
2006-07-03 11:24:48 +04:00
rwlock_init ( & rwlock_ # # x ) ; \
mutex_init ( & mutex_ # # x ) ; \
init_rwsem ( & rwsem_ # # x ) ; \
2017-05-31 18:11:49 +03:00
I2_RTMUTEX ( x ) ; \
2006-07-03 11:24:48 +04:00
} while ( 0 )
static void reset_locks ( void )
{
local_irq_disable ( ) ;
2013-06-20 15:31:24 +04:00
lockdep_free_key_range ( & ww_lockdep . acquire_key , 1 ) ;
lockdep_free_key_range ( & ww_lockdep . mutex_key , 1 ) ;
2006-07-03 11:24:48 +04:00
I1 ( A ) ; I1 ( B ) ; I1 ( C ) ; I1 ( D ) ;
I1 ( X1 ) ; I1 ( X2 ) ; I1 ( Y1 ) ; I1 ( Y2 ) ; I1 ( Z1 ) ; I1 ( Z2 ) ;
2013-06-20 15:31:42 +04:00
I_WW ( t ) ; I_WW ( t2 ) ; I_WW ( o . base ) ; I_WW ( o2 . base ) ; I_WW ( o3 . base ) ;
2006-07-03 11:24:48 +04:00
lockdep_reset ( ) ;
I2 ( A ) ; I2 ( B ) ; I2 ( C ) ; I2 ( D ) ;
init_shared_classes ( ) ;
2013-06-20 15:31:24 +04:00
2013-06-20 15:31:42 +04:00
ww_mutex_init ( & o , & ww_lockdep ) ; ww_mutex_init ( & o2 , & ww_lockdep ) ; ww_mutex_init ( & o3 , & ww_lockdep ) ;
2013-06-20 15:31:24 +04:00
memset ( & t , 0 , sizeof ( t ) ) ; memset ( & t2 , 0 , sizeof ( t2 ) ) ;
memset ( & ww_lockdep . acquire_key , 0 , sizeof ( ww_lockdep . acquire_key ) ) ;
memset ( & ww_lockdep . mutex_key , 0 , sizeof ( ww_lockdep . mutex_key ) ) ;
2006-07-03 11:24:48 +04:00
local_irq_enable ( ) ;
}
# undef I
static int testcase_total ;
static int testcase_successes ;
static int expected_testcase_failures ;
static int unexpected_testcase_failures ;
static void dotest ( void ( * testcase_fn ) ( void ) , int expected , int lockclass_mask )
{
unsigned long saved_preempt_count = preempt_count ( ) ;
WARN_ON ( irqs_disabled ( ) ) ;
testcase_fn ( ) ;
/*
* Filter out expected failures :
*/
# ifndef CONFIG_PROVE_LOCKING
2013-06-20 15:31:51 +04:00
if ( expected = = FAILURE & & debug_locks ) {
2013-06-20 15:31:24 +04:00
expected_testcase_failures + + ;
2016-11-25 01:45:28 +03:00
pr_cont ( " failed| " ) ;
2013-06-20 15:31:51 +04:00
}
else
# endif
if ( debug_locks ! = expected ) {
2013-06-20 15:31:24 +04:00
unexpected_testcase_failures + + ;
2016-11-25 01:45:28 +03:00
pr_cont ( " FAILED| " ) ;
2006-07-03 11:24:48 +04:00
} else {
testcase_successes + + ;
2016-11-25 01:45:28 +03:00
pr_cont ( " ok | " ) ;
2006-07-03 11:24:48 +04:00
}
testcase_total + + ;
if ( debug_locks_verbose )
2016-11-25 01:45:28 +03:00
pr_cont ( " lockclass mask: %x, debug_locks: %d, expected: %d \n " ,
2006-07-03 11:24:48 +04:00
lockclass_mask , debug_locks , expected ) ;
/*
* Some tests ( e . g . double - unlock ) might corrupt the preemption
* count , so restore it :
*/
2013-08-14 16:55:24 +04:00
preempt_count_set ( saved_preempt_count ) ;
2006-07-03 11:24:48 +04:00
# ifdef CONFIG_TRACE_IRQFLAGS
if ( softirq_count ( ) )
current - > softirqs_enabled = 0 ;
else
current - > softirqs_enabled = 1 ;
# endif
reset_locks ( ) ;
}
2017-05-31 18:11:49 +03:00
# ifdef CONFIG_RT_MUTEXES
# define dotest_rt(fn, e, m) dotest((fn), (e), (m))
# else
# define dotest_rt(fn, e, m)
# endif
2006-07-03 11:24:48 +04:00
static inline void print_testname ( const char * testname )
{
printk ( " %33s: " , testname ) ;
}
# define DO_TESTCASE_1(desc, name, nr) \
print_testname ( desc " / " # nr ) ; \
dotest ( name # # _ # # nr , SUCCESS , LOCKTYPE_RWLOCK ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
# define DO_TESTCASE_1B(desc, name, nr) \
print_testname ( desc " / " # nr ) ; \
dotest ( name # # _ # # nr , FAILURE , LOCKTYPE_RWLOCK ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
# define DO_TESTCASE_3(desc, name, nr) \
print_testname ( desc " / " # nr ) ; \
dotest ( name # # _spin_ # # nr , FAILURE , LOCKTYPE_SPIN ) ; \
dotest ( name # # _wlock_ # # nr , FAILURE , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _rlock_ # # nr , SUCCESS , LOCKTYPE_RWLOCK ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
# define DO_TESTCASE_3RW(desc, name, nr) \
print_testname ( desc " / " # nr ) ; \
dotest ( name # # _spin_ # # nr , FAILURE , LOCKTYPE_SPIN | LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _wlock_ # # nr , FAILURE , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _rlock_ # # nr , SUCCESS , LOCKTYPE_RWLOCK ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
# define DO_TESTCASE_6(desc, name) \
print_testname ( desc ) ; \
dotest ( name # # _spin , FAILURE , LOCKTYPE_SPIN ) ; \
dotest ( name # # _wlock , FAILURE , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _rlock , FAILURE , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _mutex , FAILURE , LOCKTYPE_MUTEX ) ; \
dotest ( name # # _wsem , FAILURE , LOCKTYPE_RWSEM ) ; \
dotest ( name # # _rsem , FAILURE , LOCKTYPE_RWSEM ) ; \
2017-05-31 18:11:49 +03:00
dotest_rt ( name # # _rtmutex , FAILURE , LOCKTYPE_RTMUTEX ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
# define DO_TESTCASE_6_SUCCESS(desc, name) \
print_testname ( desc ) ; \
dotest ( name # # _spin , SUCCESS , LOCKTYPE_SPIN ) ; \
dotest ( name # # _wlock , SUCCESS , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _rlock , SUCCESS , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _mutex , SUCCESS , LOCKTYPE_MUTEX ) ; \
dotest ( name # # _wsem , SUCCESS , LOCKTYPE_RWSEM ) ; \
dotest ( name # # _rsem , SUCCESS , LOCKTYPE_RWSEM ) ; \
2017-05-31 18:11:49 +03:00
dotest_rt ( name # # _rtmutex , SUCCESS , LOCKTYPE_RTMUTEX ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
/*
* ' read ' variant : rlocks must not trigger .
*/
# define DO_TESTCASE_6R(desc, name) \
print_testname ( desc ) ; \
dotest ( name # # _spin , FAILURE , LOCKTYPE_SPIN ) ; \
dotest ( name # # _wlock , FAILURE , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _rlock , SUCCESS , LOCKTYPE_RWLOCK ) ; \
dotest ( name # # _mutex , FAILURE , LOCKTYPE_MUTEX ) ; \
dotest ( name # # _wsem , FAILURE , LOCKTYPE_RWSEM ) ; \
dotest ( name # # _rsem , FAILURE , LOCKTYPE_RWSEM ) ; \
2017-05-31 18:11:49 +03:00
dotest_rt ( name # # _rtmutex , FAILURE , LOCKTYPE_RTMUTEX ) ; \
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
# define DO_TESTCASE_2I(desc, name, nr) \
DO_TESTCASE_1 ( " hard- " desc , name # # _hard , nr ) ; \
DO_TESTCASE_1 ( " soft- " desc , name # # _soft , nr ) ;
# define DO_TESTCASE_2IB(desc, name, nr) \
DO_TESTCASE_1B ( " hard- " desc , name # # _hard , nr ) ; \
DO_TESTCASE_1B ( " soft- " desc , name # # _soft , nr ) ;
# define DO_TESTCASE_6I(desc, name, nr) \
DO_TESTCASE_3 ( " hard- " desc , name # # _hard , nr ) ; \
DO_TESTCASE_3 ( " soft- " desc , name # # _soft , nr ) ;
# define DO_TESTCASE_6IRW(desc, name, nr) \
DO_TESTCASE_3RW ( " hard- " desc , name # # _hard , nr ) ; \
DO_TESTCASE_3RW ( " soft- " desc , name # # _soft , nr ) ;
# define DO_TESTCASE_2x3(desc, name) \
DO_TESTCASE_3 ( desc , name , 12 ) ; \
DO_TESTCASE_3 ( desc , name , 21 ) ;
# define DO_TESTCASE_2x6(desc, name) \
DO_TESTCASE_6I ( desc , name , 12 ) ; \
DO_TESTCASE_6I ( desc , name , 21 ) ;
# define DO_TESTCASE_6x2(desc, name) \
DO_TESTCASE_2I ( desc , name , 123 ) ; \
DO_TESTCASE_2I ( desc , name , 132 ) ; \
DO_TESTCASE_2I ( desc , name , 213 ) ; \
DO_TESTCASE_2I ( desc , name , 231 ) ; \
DO_TESTCASE_2I ( desc , name , 312 ) ; \
DO_TESTCASE_2I ( desc , name , 321 ) ;
# define DO_TESTCASE_6x2B(desc, name) \
DO_TESTCASE_2IB ( desc , name , 123 ) ; \
DO_TESTCASE_2IB ( desc , name , 132 ) ; \
DO_TESTCASE_2IB ( desc , name , 213 ) ; \
DO_TESTCASE_2IB ( desc , name , 231 ) ; \
DO_TESTCASE_2IB ( desc , name , 312 ) ; \
DO_TESTCASE_2IB ( desc , name , 321 ) ;
# define DO_TESTCASE_6x6(desc, name) \
DO_TESTCASE_6I ( desc , name , 123 ) ; \
DO_TESTCASE_6I ( desc , name , 132 ) ; \
DO_TESTCASE_6I ( desc , name , 213 ) ; \
DO_TESTCASE_6I ( desc , name , 231 ) ; \
DO_TESTCASE_6I ( desc , name , 312 ) ; \
DO_TESTCASE_6I ( desc , name , 321 ) ;
# define DO_TESTCASE_6x6RW(desc, name) \
DO_TESTCASE_6IRW ( desc , name , 123 ) ; \
DO_TESTCASE_6IRW ( desc , name , 132 ) ; \
DO_TESTCASE_6IRW ( desc , name , 213 ) ; \
DO_TESTCASE_6IRW ( desc , name , 231 ) ; \
DO_TESTCASE_6IRW ( desc , name , 312 ) ; \
DO_TESTCASE_6IRW ( desc , name , 321 ) ;
2013-06-20 15:31:24 +04:00
static void ww_test_fail_acquire ( void )
{
int ret ;
WWAI ( & t ) ;
t . stamp + + ;
ret = WWL ( & o , & t ) ;
if ( WARN_ON ( ! o . ctx ) | |
WARN_ON ( ret ) )
return ;
/* No lockdep test, pure API */
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ! = - EALREADY ) ;
ret = WWT ( & o ) ;
WARN_ON ( ret ) ;
t2 = t ;
t2 . stamp + + ;
ret = WWL ( & o , & t2 ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
WWU ( & o ) ;
if ( WWT ( & o ) )
WWU ( & o ) ;
# ifdef CONFIG_DEBUG_LOCK_ALLOC
else
DEBUG_LOCKS_WARN_ON ( 1 ) ;
# endif
}
2013-06-20 15:31:30 +04:00
static void ww_test_normal ( void )
{
int ret ;
WWAI ( & t ) ;
/*
* None of the ww_mutex codepaths should be taken in the ' normal '
* mutex calls . The easiest way to verify this is by using the
* normal mutex calls , and making sure o . ctx is unmodified .
*/
/* mutex_lock (and indirectly, mutex_lock_nested) */
o . ctx = ( void * ) ~ 0UL ;
mutex_lock ( & o . base ) ;
mutex_unlock ( & o . base ) ;
WARN_ON ( o . ctx ! = ( void * ) ~ 0UL ) ;
/* mutex_lock_interruptible (and *_nested) */
o . ctx = ( void * ) ~ 0UL ;
ret = mutex_lock_interruptible ( & o . base ) ;
if ( ! ret )
mutex_unlock ( & o . base ) ;
else
WARN_ON ( 1 ) ;
WARN_ON ( o . ctx ! = ( void * ) ~ 0UL ) ;
/* mutex_lock_killable (and *_nested) */
o . ctx = ( void * ) ~ 0UL ;
ret = mutex_lock_killable ( & o . base ) ;
if ( ! ret )
mutex_unlock ( & o . base ) ;
else
WARN_ON ( 1 ) ;
WARN_ON ( o . ctx ! = ( void * ) ~ 0UL ) ;
/* trylock, succeeding */
o . ctx = ( void * ) ~ 0UL ;
ret = mutex_trylock ( & o . base ) ;
WARN_ON ( ! ret ) ;
if ( ret )
mutex_unlock ( & o . base ) ;
else
WARN_ON ( 1 ) ;
WARN_ON ( o . ctx ! = ( void * ) ~ 0UL ) ;
/* trylock, failing */
o . ctx = ( void * ) ~ 0UL ;
mutex_lock ( & o . base ) ;
ret = mutex_trylock ( & o . base ) ;
WARN_ON ( ret ) ;
mutex_unlock ( & o . base ) ;
WARN_ON ( o . ctx ! = ( void * ) ~ 0UL ) ;
/* nest_lock */
o . ctx = ( void * ) ~ 0UL ;
mutex_lock_nest_lock ( & o . base , & t ) ;
mutex_unlock ( & o . base ) ;
WARN_ON ( o . ctx ! = ( void * ) ~ 0UL ) ;
}
2013-06-20 15:31:24 +04:00
static void ww_test_two_contexts ( void )
{
WWAI ( & t ) ;
WWAI ( & t2 ) ;
}
static void ww_test_diff_class ( void )
{
WWAI ( & t ) ;
# ifdef CONFIG_DEBUG_MUTEXES
t . ww_class = NULL ;
# endif
WWL ( & o , & t ) ;
}
static void ww_test_context_done_twice ( void )
{
WWAI ( & t ) ;
WWAD ( & t ) ;
WWAD ( & t ) ;
WWAF ( & t ) ;
}
static void ww_test_context_unlock_twice ( void )
{
WWAI ( & t ) ;
WWAD ( & t ) ;
WWAF ( & t ) ;
WWAF ( & t ) ;
}
static void ww_test_context_fini_early ( void )
{
WWAI ( & t ) ;
WWL ( & o , & t ) ;
WWAD ( & t ) ;
WWAF ( & t ) ;
}
static void ww_test_context_lock_after_done ( void )
{
WWAI ( & t ) ;
WWAD ( & t ) ;
WWL ( & o , & t ) ;
}
static void ww_test_object_unlock_twice ( void )
{
WWL1 ( & o ) ;
WWU ( & o ) ;
WWU ( & o ) ;
}
static void ww_test_object_lock_unbalanced ( void )
{
WWAI ( & t ) ;
WWL ( & o , & t ) ;
t . acquired = 0 ;
WWU ( & o ) ;
WWAF ( & t ) ;
}
static void ww_test_object_lock_stale_context ( void )
{
WWAI ( & t ) ;
o . ctx = & t2 ;
WWL ( & o , & t ) ;
}
2013-06-20 15:31:42 +04:00
static void ww_test_edeadlk_normal ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
o2 . ctx = & t2 ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
o2 . ctx = NULL ;
mutex_acquire ( & o2 . base . dep_map , 0 , 1 , _THIS_IP_ ) ;
mutex_unlock ( & o2 . base ) ;
WWU ( & o ) ;
WWL ( & o2 , & t ) ;
}
static void ww_test_edeadlk_normal_slow ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
o2 . ctx = NULL ;
mutex_acquire ( & o2 . base . dep_map , 0 , 1 , _THIS_IP_ ) ;
mutex_unlock ( & o2 . base ) ;
WWU ( & o ) ;
ww_mutex_lock_slow ( & o2 , & t ) ;
}
static void ww_test_edeadlk_no_unlock ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
o2 . ctx = & t2 ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
o2 . ctx = NULL ;
mutex_acquire ( & o2 . base . dep_map , 0 , 1 , _THIS_IP_ ) ;
mutex_unlock ( & o2 . base ) ;
WWL ( & o2 , & t ) ;
}
static void ww_test_edeadlk_no_unlock_slow ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
o2 . ctx = NULL ;
mutex_acquire ( & o2 . base . dep_map , 0 , 1 , _THIS_IP_ ) ;
mutex_unlock ( & o2 . base ) ;
ww_mutex_lock_slow ( & o2 , & t ) ;
}
static void ww_test_edeadlk_acquire_more ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
ret = WWL ( & o3 , & t ) ;
}
static void ww_test_edeadlk_acquire_more_slow ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
ww_mutex_lock_slow ( & o3 , & t ) ;
}
static void ww_test_edeadlk_acquire_more_edeadlk ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
mutex_lock ( & o3 . base ) ;
mutex_release ( & o3 . base . dep_map , 1 , _THIS_IP_ ) ;
o3 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
ret = WWL ( & o3 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
}
static void ww_test_edeadlk_acquire_more_edeadlk_slow ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
mutex_lock ( & o3 . base ) ;
mutex_release ( & o3 . base . dep_map , 1 , _THIS_IP_ ) ;
o3 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
ww_mutex_lock_slow ( & o3 , & t ) ;
}
static void ww_test_edeadlk_acquire_wrong ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
if ( ! ret )
WWU ( & o2 ) ;
WWU ( & o ) ;
ret = WWL ( & o3 , & t ) ;
}
static void ww_test_edeadlk_acquire_wrong_slow ( void )
{
int ret ;
mutex_lock ( & o2 . base ) ;
mutex_release ( & o2 . base . dep_map , 1 , _THIS_IP_ ) ;
o2 . ctx = & t2 ;
WWAI ( & t ) ;
t2 = t ;
t2 . stamp - - ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ! = - EDEADLK ) ;
if ( ! ret )
WWU ( & o2 ) ;
WWU ( & o ) ;
ww_mutex_lock_slow ( & o3 , & t ) ;
}
2013-06-20 15:31:24 +04:00
static void ww_test_spin_nest_unlocked ( void )
{
raw_spin_lock_nest_lock ( & lock_A , & o . base ) ;
U ( A ) ;
}
static void ww_test_unneeded_slow ( void )
{
WWAI ( & t ) ;
ww_mutex_lock_slow ( & o , & t ) ;
}
static void ww_test_context_block ( void )
{
int ret ;
WWAI ( & t ) ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
WWL1 ( & o2 ) ;
}
static void ww_test_context_try ( void )
{
int ret ;
WWAI ( & t ) ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWT ( & o2 ) ;
WARN_ON ( ! ret ) ;
WWU ( & o2 ) ;
WWU ( & o ) ;
}
static void ww_test_context_context ( void )
{
int ret ;
WWAI ( & t ) ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ) ;
WWU ( & o2 ) ;
WWU ( & o ) ;
}
static void ww_test_try_block ( void )
{
bool ret ;
ret = WWT ( & o ) ;
WARN_ON ( ! ret ) ;
WWL1 ( & o2 ) ;
WWU ( & o2 ) ;
WWU ( & o ) ;
}
static void ww_test_try_try ( void )
{
bool ret ;
ret = WWT ( & o ) ;
WARN_ON ( ! ret ) ;
ret = WWT ( & o2 ) ;
WARN_ON ( ! ret ) ;
WWU ( & o2 ) ;
WWU ( & o ) ;
}
static void ww_test_try_context ( void )
{
int ret ;
ret = WWT ( & o ) ;
WARN_ON ( ! ret ) ;
WWAI ( & t ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ) ;
}
static void ww_test_block_block ( void )
{
WWL1 ( & o ) ;
WWL1 ( & o2 ) ;
}
static void ww_test_block_try ( void )
{
bool ret ;
WWL1 ( & o ) ;
ret = WWT ( & o2 ) ;
WARN_ON ( ! ret ) ;
}
static void ww_test_block_context ( void )
{
int ret ;
WWL1 ( & o ) ;
WWAI ( & t ) ;
ret = WWL ( & o2 , & t ) ;
WARN_ON ( ret ) ;
}
static void ww_test_spin_block ( void )
{
L ( A ) ;
U ( A ) ;
WWL1 ( & o ) ;
L ( A ) ;
U ( A ) ;
WWU ( & o ) ;
L ( A ) ;
WWL1 ( & o ) ;
WWU ( & o ) ;
U ( A ) ;
}
static void ww_test_spin_try ( void )
{
bool ret ;
L ( A ) ;
U ( A ) ;
ret = WWT ( & o ) ;
WARN_ON ( ! ret ) ;
L ( A ) ;
U ( A ) ;
WWU ( & o ) ;
L ( A ) ;
ret = WWT ( & o ) ;
WARN_ON ( ! ret ) ;
WWU ( & o ) ;
U ( A ) ;
}
static void ww_test_spin_context ( void )
{
int ret ;
L ( A ) ;
U ( A ) ;
WWAI ( & t ) ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
L ( A ) ;
U ( A ) ;
WWU ( & o ) ;
L ( A ) ;
ret = WWL ( & o , & t ) ;
WARN_ON ( ret ) ;
WWU ( & o ) ;
U ( A ) ;
}
static void ww_tests ( void )
{
printk ( " -------------------------------------------------------------------------- \n " ) ;
printk ( " | Wound/wait tests | \n " ) ;
printk ( " --------------------- \n " ) ;
print_testname ( " ww api failures " ) ;
dotest ( ww_test_fail_acquire , SUCCESS , LOCKTYPE_WW ) ;
2013-06-20 15:31:30 +04:00
dotest ( ww_test_normal , SUCCESS , LOCKTYPE_WW ) ;
2013-06-20 15:31:24 +04:00
dotest ( ww_test_unneeded_slow , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
print_testname ( " ww contexts mixing " ) ;
dotest ( ww_test_two_contexts , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_diff_class , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
print_testname ( " finishing ww context " ) ;
dotest ( ww_test_context_done_twice , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_context_unlock_twice , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_context_fini_early , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_context_lock_after_done , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
print_testname ( " locking mismatches " ) ;
dotest ( ww_test_object_unlock_twice , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_object_lock_unbalanced , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_object_lock_stale_context , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
2013-06-20 15:31:42 +04:00
print_testname ( " EDEADLK handling " ) ;
dotest ( ww_test_edeadlk_normal , SUCCESS , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_normal_slow , SUCCESS , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_no_unlock , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_no_unlock_slow , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_acquire_more , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_acquire_more_slow , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_acquire_more_edeadlk , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_acquire_more_edeadlk_slow , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_acquire_wrong , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_edeadlk_acquire_wrong_slow , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:42 +04:00
2013-06-20 15:31:24 +04:00
print_testname ( " spinlock nest unlocked " ) ;
dotest ( ww_test_spin_nest_unlocked , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
printk ( " ----------------------------------------------------- \n " ) ;
printk ( " |block | try |context| \n " ) ;
printk ( " ----------------------------------------------------- \n " ) ;
print_testname ( " context " ) ;
dotest ( ww_test_context_block , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_context_try , SUCCESS , LOCKTYPE_WW ) ;
dotest ( ww_test_context_context , SUCCESS , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
print_testname ( " try " ) ;
dotest ( ww_test_try_block , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_try_try , SUCCESS , LOCKTYPE_WW ) ;
dotest ( ww_test_try_context , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
print_testname ( " block " ) ;
dotest ( ww_test_block_block , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_block_try , SUCCESS , LOCKTYPE_WW ) ;
dotest ( ww_test_block_context , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
print_testname ( " spinlock " ) ;
dotest ( ww_test_spin_block , FAILURE , LOCKTYPE_WW ) ;
dotest ( ww_test_spin_try , SUCCESS , LOCKTYPE_WW ) ;
dotest ( ww_test_spin_context , FAILURE , LOCKTYPE_WW ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2013-06-20 15:31:24 +04:00
}
2006-07-03 11:24:48 +04:00
void locking_selftest ( void )
{
/*
* Got a locking failure before the selftest ran ?
*/
if ( ! debug_locks ) {
printk ( " ---------------------------------- \n " ) ;
printk ( " | Locking API testsuite disabled | \n " ) ;
printk ( " ---------------------------------- \n " ) ;
return ;
}
/*
* Run the testsuite :
*/
printk ( " ------------------------ \n " ) ;
printk ( " | Locking API testsuite: \n " ) ;
printk ( " ---------------------------------------------------------------------------- \n " ) ;
printk ( " | spin |wlock |rlock |mutex | wsem | rsem | \n " ) ;
printk ( " -------------------------------------------------------------------------- \n " ) ;
init_shared_classes ( ) ;
debug_locks_silent = ! debug_locks_verbose ;
2019-02-15 02:00:44 +03:00
lockdep_set_selftest_task ( current ) ;
2006-07-03 11:24:48 +04:00
2006-07-03 11:24:51 +04:00
DO_TESTCASE_6R ( " A-A deadlock " , AA ) ;
2006-07-03 11:24:48 +04:00
DO_TESTCASE_6R ( " A-B-B-A deadlock " , ABBA ) ;
DO_TESTCASE_6R ( " A-B-B-C-C-A deadlock " , ABBCCA ) ;
DO_TESTCASE_6R ( " A-B-C-A-B-C deadlock " , ABCABC ) ;
DO_TESTCASE_6R ( " A-B-B-C-C-D-D-A deadlock " , ABBCCDDA ) ;
DO_TESTCASE_6R ( " A-B-C-D-B-D-D-A deadlock " , ABCDBDDA ) ;
DO_TESTCASE_6R ( " A-B-C-D-B-C-D-A deadlock " , ABCDBCDA ) ;
DO_TESTCASE_6 ( " double unlock " , double_unlock ) ;
DO_TESTCASE_6 ( " initialize held " , init_held ) ;
printk ( " -------------------------------------------------------------------------- \n " ) ;
print_testname ( " recursive read-lock " ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rlock_AA1 , SUCCESS , LOCKTYPE_RWLOCK ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rsem_AA1 , FAILURE , LOCKTYPE_RWSEM ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
print_testname ( " recursive read-lock #2 " ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:51 +04:00
dotest ( rlock_AA1B , SUCCESS , LOCKTYPE_RWLOCK ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rsem_AA1B , FAILURE , LOCKTYPE_RWSEM ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
print_testname ( " mixed read-write-lock " ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rlock_AA2 , FAILURE , LOCKTYPE_RWLOCK ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rsem_AA2 , FAILURE , LOCKTYPE_RWSEM ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
print_testname ( " mixed write-read-lock " ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rlock_AA3 , FAILURE , LOCKTYPE_RWLOCK ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " | " ) ;
2006-07-03 11:24:48 +04:00
dotest ( rsem_AA3 , FAILURE , LOCKTYPE_RWSEM ) ;
2016-11-25 01:45:28 +03:00
pr_cont ( " \n " ) ;
2006-07-03 11:24:48 +04:00
2017-08-23 14:13:11 +03:00
print_testname ( " mixed read-lock/lock-write ABBA " ) ;
pr_cont ( " | " ) ;
dotest ( rlock_ABBA1 , FAILURE , LOCKTYPE_RWLOCK ) ;
2017-10-04 11:50:25 +03:00
# ifdef CONFIG_PROVE_LOCKING
2017-08-28 15:42:45 +03:00
/*
* Lockdep does indeed fail here , but there ' s nothing we can do about
* that now . Don ' t kill lockdep for it .
*/
unexpected_testcase_failures - - ;
2017-10-04 11:50:25 +03:00
# endif
2017-08-28 15:42:45 +03:00
2017-08-23 14:13:11 +03:00
pr_cont ( " | " ) ;
dotest ( rwsem_ABBA1 , FAILURE , LOCKTYPE_RWSEM ) ;
print_testname ( " mixed read-lock/lock-read ABBA " ) ;
pr_cont ( " | " ) ;
dotest ( rlock_ABBA2 , SUCCESS , LOCKTYPE_RWLOCK ) ;
pr_cont ( " | " ) ;
dotest ( rwsem_ABBA2 , FAILURE , LOCKTYPE_RWSEM ) ;
print_testname ( " mixed write-lock/lock-write ABBA " ) ;
pr_cont ( " | " ) ;
dotest ( rlock_ABBA3 , FAILURE , LOCKTYPE_RWLOCK ) ;
pr_cont ( " | " ) ;
dotest ( rwsem_ABBA3 , FAILURE , LOCKTYPE_RWSEM ) ;
2006-07-03 11:24:48 +04:00
printk ( " -------------------------------------------------------------------------- \n " ) ;
/*
* irq - context testcases :
*/
DO_TESTCASE_2x6 ( " irqs-on + irq-safe-A " , irqsafe1 ) ;
DO_TESTCASE_2x3 ( " sirq-safe-A => hirqs-on " , irqsafe2A ) ;
DO_TESTCASE_2x6 ( " safe-A + irqs-on " , irqsafe2B ) ;
DO_TESTCASE_6x6 ( " safe-A + unsafe-B #1 " , irqsafe3 ) ;
DO_TESTCASE_6x6 ( " safe-A + unsafe-B #2 " , irqsafe4 ) ;
DO_TESTCASE_6x6RW ( " irq lock-inversion " , irq_inversion ) ;
DO_TESTCASE_6x2 ( " irq read-recursion " , irq_read_recursion ) ;
// DO_TESTCASE_6x2B("irq read-recursion #2", irq_read_recursion2);
2013-06-20 15:31:24 +04:00
ww_tests ( ) ;
2006-07-03 11:24:48 +04:00
if ( unexpected_testcase_failures ) {
printk ( " ----------------------------------------------------------------- \n " ) ;
debug_locks = 0 ;
printk ( " BUG: %3d unexpected failures (out of %3d) - debugging disabled! | \n " ,
unexpected_testcase_failures , testcase_total ) ;
printk ( " ----------------------------------------------------------------- \n " ) ;
} else if ( expected_testcase_failures & & testcase_successes ) {
printk ( " -------------------------------------------------------- \n " ) ;
printk ( " %3d out of %3d testcases failed, as expected. | \n " ,
expected_testcase_failures , testcase_total ) ;
printk ( " ---------------------------------------------------- \n " ) ;
debug_locks = 1 ;
} else if ( expected_testcase_failures & & ! testcase_successes ) {
printk ( " -------------------------------------------------------- \n " ) ;
printk ( " All %3d testcases failed, as expected. | \n " ,
expected_testcase_failures ) ;
printk ( " ---------------------------------------- \n " ) ;
debug_locks = 1 ;
} else {
printk ( " ------------------------------------------------------- \n " ) ;
printk ( " Good, all %3d testcases passed! | \n " ,
testcase_successes ) ;
printk ( " --------------------------------- \n " ) ;
debug_locks = 1 ;
}
2019-02-15 02:00:44 +03:00
lockdep_set_selftest_task ( NULL ) ;
2006-07-03 11:24:48 +04:00
debug_locks_silent = 0 ;
}