License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 14:07:57 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* lib/locking-selftest.c
|
|
|
|
*
|
|
|
|
* Testsuite for various locking APIs: spinlocks, rwlocks,
|
|
|
|
* mutexes and rw-semaphores.
|
|
|
|
*
|
|
|
|
* It is checking both false positives and false negatives.
|
|
|
|
*
|
|
|
|
* Started by Ingo Molnar:
|
|
|
|
*
|
|
|
|
* Copyright (C) 2006 Red Hat, Inc., Ingo Molnar <mingo@redhat.com>
|
|
|
|
*/
|
|
|
|
#include <linux/rwsem.h>
|
|
|
|
#include <linux/mutex.h>
|
2013-07-05 07:29:32 +00:00
|
|
|
#include <linux/ww_mutex.h>
|
2006-07-03 07:24:48 +00:00
|
|
|
#include <linux/sched.h>
|
2020-12-15 03:08:38 +00:00
|
|
|
#include <linux/sched/mm.h>
|
2006-07-03 07:24:48 +00:00
|
|
|
#include <linux/delay.h>
|
[PATCH] lockdep: core
Do 'make oldconfig' and accept all the defaults for new config options -
reboot into the kernel and if everything goes well it should boot up fine and
you should have /proc/lockdep and /proc/lockdep_stats files.
Typically if the lock validator finds some problem it will print out
voluminous debug output that begins with "BUG: ..." and which syslog output
can be used by kernel developers to figure out the precise locking scenario.
What does the lock validator do? It "observes" and maps all locking rules as
they occur dynamically (as triggered by the kernel's natural use of spinlocks,
rwlocks, mutexes and rwsems). Whenever the lock validator subsystem detects a
new locking scenario, it validates this new rule against the existing set of
rules. If this new rule is consistent with the existing set of rules then the
new rule is added transparently and the kernel continues as normal. If the
new rule could create a deadlock scenario then this condition is printed out.
When determining validity of locking, all possible "deadlock scenarios" are
considered: assuming arbitrary number of CPUs, arbitrary irq context and task
context constellations, running arbitrary combinations of all the existing
locking scenarios. In a typical system this means millions of separate
scenarios. This is why we call it a "locking correctness" validator - for all
rules that are observed the lock validator proves it with mathematical
certainty that a deadlock could not occur (assuming that the lock validator
implementation itself is correct and its internal data structures are not
corrupted by some other kernel subsystem). [see more details and conditionals
of this statement in include/linux/lockdep.h and
Documentation/lockdep-design.txt]
Furthermore, this "all possible scenarios" property of the validator also
enables the finding of complex, highly unlikely multi-CPU multi-context races
via single single-context rules, increasing the likelyhood of finding bugs
drastically. In practical terms: the lock validator already found a bug in
the upstream kernel that could only occur on systems with 3 or more CPUs, and
which needed 3 very unlikely code sequences to occur at once on the 3 CPUs.
That bug was found and reported on a single-CPU system (!). So in essence a
race will be found "piecemail-wise", triggering all the necessary components
for the race, without having to reproduce the race scenario itself! In its
short existence the lock validator found and reported many bugs before they
actually caused a real deadlock.
To further increase the efficiency of the validator, the mapping is not per
"lock instance", but per "lock-class". For example, all struct inode objects
in the kernel have inode->inotify_mutex. If there are 10,000 inodes cached,
then there are 10,000 lock objects. But ->inotify_mutex is a single "lock
type", and all locking activities that occur against ->inotify_mutex are
"unified" into this single lock-class. The advantage of the lock-class
approach is that all historical ->inotify_mutex uses are mapped into a single
(and as narrow as possible) set of locking rules - regardless of how many
different tasks or inode structures it took to build this set of rules. The
set of rules persist during the lifetime of the kernel.
To see the rough magnitude of checking that the lock validator does, here's a
portion of /proc/lockdep_stats, fresh after bootup:
lock-classes: 694 [max: 2048]
direct dependencies: 1598 [max: 8192]
indirect dependencies: 17896
all direct dependencies: 16206
dependency chains: 1910 [max: 8192]
in-hardirq chains: 17
in-softirq chains: 105
in-process chains: 1065
stack-trace entries: 38761 [max: 131072]
combined max dependencies: 2033928
hardirq-safe locks: 24
hardirq-unsafe locks: 176
softirq-safe locks: 53
softirq-unsafe locks: 137
irq-safe locks: 59
irq-unsafe locks: 176
The lock validator has observed 1598 actual single-thread locking patterns,
and has validated all possible 2033928 distinct locking scenarios.
More details about the design of the lock validator can be found in
Documentation/lockdep-design.txt, which can also found at:
http://redhat.com/~mingo/lockdep-patches/lockdep-design.txt
[bunk@stusta.de: cleanups]
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Arjan van de Ven <arjan@linux.intel.com>
Signed-off-by: Adrian Bunk <bunk@stusta.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-07-03 07:24:50 +00:00
|
|
|
#include <linux/lockdep.h>
|
2006-07-03 07:24:48 +00:00
|
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <linux/kallsyms.h>
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/debug_locks.h>
|
|
|
|
#include <linux/irqflags.h>
|
2017-05-31 15:11:49 +00:00
|
|
|
#include <linux/rtmutex.h>
|
2020-12-09 15:06:06 +00:00
|
|
|
#include <linux/local_lock.h>
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifdef CONFIG_PREEMPT_RT
|
|
|
|
# define NON_RT(...)
|
|
|
|
#else
|
|
|
|
# define NON_RT(...) __VA_ARGS__
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* Change this to 1 if you want to see the failure printouts:
|
|
|
|
*/
|
|
|
|
static unsigned int debug_locks_verbose;
|
2020-08-07 07:42:20 +00:00
|
|
|
unsigned int force_read_lock_recursive;
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2018-06-15 08:17:38 +00:00
|
|
|
static DEFINE_WD_CLASS(ww_lockdep);
|
2013-06-20 11:31:24 +00:00
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
static int __init setup_debug_locks_verbose(char *str)
|
|
|
|
{
|
|
|
|
get_option(&str, &debug_locks_verbose);
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
__setup("debug_locks_verbose=", setup_debug_locks_verbose);
|
|
|
|
|
|
|
|
#define FAILURE 0
|
|
|
|
#define SUCCESS 1
|
|
|
|
|
|
|
|
#define LOCKTYPE_SPIN 0x1
|
|
|
|
#define LOCKTYPE_RWLOCK 0x2
|
|
|
|
#define LOCKTYPE_MUTEX 0x4
|
|
|
|
#define LOCKTYPE_RWSEM 0x8
|
2013-06-20 11:31:24 +00:00
|
|
|
#define LOCKTYPE_WW 0x10
|
2017-05-31 15:11:49 +00:00
|
|
|
#define LOCKTYPE_RTMUTEX 0x20
|
2020-12-09 15:06:06 +00:00
|
|
|
#define LOCKTYPE_LL 0x40
|
2021-06-18 17:01:10 +00:00
|
|
|
#define LOCKTYPE_SPECIAL 0x80
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
static struct ww_acquire_ctx t, t2;
|
2013-06-20 11:31:42 +00:00
|
|
|
static struct ww_mutex o, o2, o3;
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Normal standalone locks, for the circular and irq-context
|
|
|
|
* dependency tests:
|
|
|
|
*/
|
2020-10-30 11:37:43 +00:00
|
|
|
static DEFINE_SPINLOCK(lock_A);
|
|
|
|
static DEFINE_SPINLOCK(lock_B);
|
|
|
|
static DEFINE_SPINLOCK(lock_C);
|
|
|
|
static DEFINE_SPINLOCK(lock_D);
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2020-12-08 10:31:12 +00:00
|
|
|
static DEFINE_RAW_SPINLOCK(raw_lock_A);
|
|
|
|
static DEFINE_RAW_SPINLOCK(raw_lock_B);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
static DEFINE_RWLOCK(rwlock_A);
|
|
|
|
static DEFINE_RWLOCK(rwlock_B);
|
|
|
|
static DEFINE_RWLOCK(rwlock_C);
|
|
|
|
static DEFINE_RWLOCK(rwlock_D);
|
|
|
|
|
|
|
|
static DEFINE_MUTEX(mutex_A);
|
|
|
|
static DEFINE_MUTEX(mutex_B);
|
|
|
|
static DEFINE_MUTEX(mutex_C);
|
|
|
|
static DEFINE_MUTEX(mutex_D);
|
|
|
|
|
|
|
|
static DECLARE_RWSEM(rwsem_A);
|
|
|
|
static DECLARE_RWSEM(rwsem_B);
|
|
|
|
static DECLARE_RWSEM(rwsem_C);
|
|
|
|
static DECLARE_RWSEM(rwsem_D);
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_A);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_B);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_C);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_D);
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* Locks that we initialize dynamically as well so that
|
|
|
|
* e.g. X1 and X2 becomes two instances of the same class,
|
|
|
|
* but X* and Y* are different classes. We do this so that
|
|
|
|
* we do not trigger a real lockup:
|
|
|
|
*/
|
2020-10-30 11:37:43 +00:00
|
|
|
static DEFINE_SPINLOCK(lock_X1);
|
|
|
|
static DEFINE_SPINLOCK(lock_X2);
|
|
|
|
static DEFINE_SPINLOCK(lock_Y1);
|
|
|
|
static DEFINE_SPINLOCK(lock_Y2);
|
|
|
|
static DEFINE_SPINLOCK(lock_Z1);
|
|
|
|
static DEFINE_SPINLOCK(lock_Z2);
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
static DEFINE_RWLOCK(rwlock_X1);
|
|
|
|
static DEFINE_RWLOCK(rwlock_X2);
|
|
|
|
static DEFINE_RWLOCK(rwlock_Y1);
|
|
|
|
static DEFINE_RWLOCK(rwlock_Y2);
|
|
|
|
static DEFINE_RWLOCK(rwlock_Z1);
|
|
|
|
static DEFINE_RWLOCK(rwlock_Z2);
|
|
|
|
|
|
|
|
static DEFINE_MUTEX(mutex_X1);
|
|
|
|
static DEFINE_MUTEX(mutex_X2);
|
|
|
|
static DEFINE_MUTEX(mutex_Y1);
|
|
|
|
static DEFINE_MUTEX(mutex_Y2);
|
|
|
|
static DEFINE_MUTEX(mutex_Z1);
|
|
|
|
static DEFINE_MUTEX(mutex_Z2);
|
|
|
|
|
|
|
|
static DECLARE_RWSEM(rwsem_X1);
|
|
|
|
static DECLARE_RWSEM(rwsem_X2);
|
|
|
|
static DECLARE_RWSEM(rwsem_Y1);
|
|
|
|
static DECLARE_RWSEM(rwsem_Y2);
|
|
|
|
static DECLARE_RWSEM(rwsem_Z1);
|
|
|
|
static DECLARE_RWSEM(rwsem_Z2);
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_X1);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_X2);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_Y1);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_Y2);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_Z1);
|
|
|
|
static DEFINE_RT_MUTEX(rtmutex_Z2);
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2021-11-29 17:46:49 +00:00
|
|
|
static DEFINE_PER_CPU(local_lock_t, local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* non-inlined runtime initializers, to let separate locks share
|
|
|
|
* the same lock-class:
|
|
|
|
*/
|
|
|
|
#define INIT_CLASS_FUNC(class) \
|
|
|
|
static noinline void \
|
2020-10-30 11:37:43 +00:00
|
|
|
init_class_##class(spinlock_t *lock, rwlock_t *rwlock, \
|
2012-04-16 07:01:55 +00:00
|
|
|
struct mutex *mutex, struct rw_semaphore *rwsem)\
|
2006-07-03 07:24:48 +00:00
|
|
|
{ \
|
2020-10-30 11:37:43 +00:00
|
|
|
spin_lock_init(lock); \
|
2006-07-03 07:24:48 +00:00
|
|
|
rwlock_init(rwlock); \
|
|
|
|
mutex_init(mutex); \
|
|
|
|
init_rwsem(rwsem); \
|
|
|
|
}
|
|
|
|
|
|
|
|
INIT_CLASS_FUNC(X)
|
|
|
|
INIT_CLASS_FUNC(Y)
|
|
|
|
INIT_CLASS_FUNC(Z)
|
|
|
|
|
|
|
|
static void init_shared_classes(void)
|
|
|
|
{
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
static struct lock_class_key rt_X, rt_Y, rt_Z;
|
|
|
|
|
|
|
|
__rt_mutex_init(&rtmutex_X1, __func__, &rt_X);
|
|
|
|
__rt_mutex_init(&rtmutex_X2, __func__, &rt_X);
|
|
|
|
__rt_mutex_init(&rtmutex_Y1, __func__, &rt_Y);
|
|
|
|
__rt_mutex_init(&rtmutex_Y2, __func__, &rt_Y);
|
|
|
|
__rt_mutex_init(&rtmutex_Z1, __func__, &rt_Z);
|
|
|
|
__rt_mutex_init(&rtmutex_Z2, __func__, &rt_Z);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
init_class_X(&lock_X1, &rwlock_X1, &mutex_X1, &rwsem_X1);
|
|
|
|
init_class_X(&lock_X2, &rwlock_X2, &mutex_X2, &rwsem_X2);
|
|
|
|
|
|
|
|
init_class_Y(&lock_Y1, &rwlock_Y1, &mutex_Y1, &rwsem_Y1);
|
|
|
|
init_class_Y(&lock_Y2, &rwlock_Y2, &mutex_Y2, &rwsem_Y2);
|
|
|
|
|
|
|
|
init_class_Z(&lock_Z1, &rwlock_Z1, &mutex_Z1, &rwsem_Z1);
|
|
|
|
init_class_Z(&lock_Z2, &rwlock_Z2, &mutex_Z2, &rwsem_Z2);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For spinlocks and rwlocks we also do hardirq-safe / softirq-safe tests.
|
|
|
|
* The following functions use a lock from a simulated hardirq/softirq
|
|
|
|
* context, causing the locks to be marked as hardirq-safe/softirq-safe:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define HARDIRQ_DISABLE local_irq_disable
|
|
|
|
#define HARDIRQ_ENABLE local_irq_enable
|
|
|
|
|
|
|
|
#define HARDIRQ_ENTER() \
|
|
|
|
local_irq_disable(); \
|
2011-05-20 00:09:54 +00:00
|
|
|
__irq_enter(); \
|
2021-06-17 18:57:19 +00:00
|
|
|
lockdep_hardirq_threaded(); \
|
2006-07-03 07:24:48 +00:00
|
|
|
WARN_ON(!in_irq());
|
|
|
|
|
|
|
|
#define HARDIRQ_EXIT() \
|
|
|
|
__irq_exit(); \
|
|
|
|
local_irq_enable();
|
|
|
|
|
|
|
|
#define SOFTIRQ_DISABLE local_bh_disable
|
|
|
|
#define SOFTIRQ_ENABLE local_bh_enable
|
|
|
|
|
|
|
|
#define SOFTIRQ_ENTER() \
|
|
|
|
local_bh_disable(); \
|
|
|
|
local_irq_disable(); \
|
2009-03-13 00:30:40 +00:00
|
|
|
lockdep_softirq_enter(); \
|
2006-07-03 07:24:48 +00:00
|
|
|
WARN_ON(!in_softirq());
|
|
|
|
|
|
|
|
#define SOFTIRQ_EXIT() \
|
2009-03-13 00:30:40 +00:00
|
|
|
lockdep_softirq_exit(); \
|
2006-07-03 07:24:48 +00:00
|
|
|
local_irq_enable(); \
|
|
|
|
local_bh_enable();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Shortcuts for lock/unlock API variants, to keep
|
|
|
|
* the testcases compact:
|
|
|
|
*/
|
2020-10-30 11:37:43 +00:00
|
|
|
#define L(x) spin_lock(&lock_##x)
|
|
|
|
#define U(x) spin_unlock(&lock_##x)
|
2006-07-03 07:24:48 +00:00
|
|
|
#define LU(x) L(x); U(x)
|
2020-10-30 11:37:43 +00:00
|
|
|
#define SI(x) spin_lock_init(&lock_##x)
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define WL(x) write_lock(&rwlock_##x)
|
|
|
|
#define WU(x) write_unlock(&rwlock_##x)
|
|
|
|
#define WLU(x) WL(x); WU(x)
|
|
|
|
|
|
|
|
#define RL(x) read_lock(&rwlock_##x)
|
|
|
|
#define RU(x) read_unlock(&rwlock_##x)
|
|
|
|
#define RLU(x) RL(x); RU(x)
|
|
|
|
#define RWI(x) rwlock_init(&rwlock_##x)
|
|
|
|
|
|
|
|
#define ML(x) mutex_lock(&mutex_##x)
|
|
|
|
#define MU(x) mutex_unlock(&mutex_##x)
|
|
|
|
#define MI(x) mutex_init(&mutex_##x)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#define RTL(x) rt_mutex_lock(&rtmutex_##x)
|
|
|
|
#define RTU(x) rt_mutex_unlock(&rtmutex_##x)
|
|
|
|
#define RTI(x) rt_mutex_init(&rtmutex_##x)
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#define WSL(x) down_write(&rwsem_##x)
|
|
|
|
#define WSU(x) up_write(&rwsem_##x)
|
|
|
|
|
|
|
|
#define RSL(x) down_read(&rwsem_##x)
|
|
|
|
#define RSU(x) up_read(&rwsem_##x)
|
|
|
|
#define RWSI(x) init_rwsem(&rwsem_##x)
|
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
#ifndef CONFIG_DEBUG_WW_MUTEX_SLOWPATH
|
|
|
|
#define WWAI(x) ww_acquire_init(x, &ww_lockdep)
|
|
|
|
#else
|
|
|
|
#define WWAI(x) do { ww_acquire_init(x, &ww_lockdep); (x)->deadlock_inject_countdown = ~0U; } while (0)
|
|
|
|
#endif
|
|
|
|
#define WWAD(x) ww_acquire_done(x)
|
|
|
|
#define WWAF(x) ww_acquire_fini(x)
|
|
|
|
|
|
|
|
#define WWL(x, c) ww_mutex_lock(x, c)
|
kernel/locking: Add context to ww_mutex_trylock()
i915 will soon gain an eviction path that trylock a whole lot of locks
for eviction, getting dmesg failures like below:
BUG: MAX_LOCK_DEPTH too low!
turning off the locking correctness validator.
depth: 48 max: 48!
48 locks held by i915_selftest/5776:
#0: ffff888101a79240 (&dev->mutex){....}-{3:3}, at: __driver_attach+0x88/0x160
#1: ffffc900009778c0 (reservation_ww_class_acquire){+.+.}-{0:0}, at: i915_vma_pin.constprop.63+0x39/0x1b0 [i915]
#2: ffff88800cf74de8 (reservation_ww_class_mutex){+.+.}-{3:3}, at: i915_vma_pin.constprop.63+0x5f/0x1b0 [i915]
#3: ffff88810c7f9e38 (&vm->mutex/1){+.+.}-{3:3}, at: i915_vma_pin_ww+0x1c4/0x9d0 [i915]
#4: ffff88810bad5768 (reservation_ww_class_mutex){+.+.}-{3:3}, at: i915_gem_evict_something+0x110/0x860 [i915]
#5: ffff88810bad60e8 (reservation_ww_class_mutex){+.+.}-{3:3}, at: i915_gem_evict_something+0x110/0x860 [i915]
...
#46: ffff88811964d768 (reservation_ww_class_mutex){+.+.}-{3:3}, at: i915_gem_evict_something+0x110/0x860 [i915]
#47: ffff88811964e0e8 (reservation_ww_class_mutex){+.+.}-{3:3}, at: i915_gem_evict_something+0x110/0x860 [i915]
INFO: lockdep is turned off.
Fixing eviction to nest into ww_class_acquire is a high priority, but
it requires a rework of the entire driver, which can only be done one
step at a time.
As an intermediate solution, add an acquire context to
ww_mutex_trylock, which allows us to do proper nesting annotations on
the trylocks, making the above lockdep splat disappear.
This is also useful in regulator_lock_nested, which may avoid dropping
regulator_nesting_mutex in the uncontended path, so use it there.
TTM may be another user for this, where we could lock a buffer in a
fastpath with list locks held, without dropping all locks we hold.
[peterz: rework actual ww_mutex_trylock() implementations]
Signed-off-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Link: https://lkml.kernel.org/r/YUBGPdDDjKlxAuXJ@hirez.programming.kicks-ass.net
2021-09-09 09:32:18 +00:00
|
|
|
#define WWT(x) ww_mutex_trylock(x, NULL)
|
2013-06-20 11:31:24 +00:00
|
|
|
#define WWL1(x) ww_mutex_lock(x, NULL)
|
|
|
|
#define WWU(x) ww_mutex_unlock(x)
|
|
|
|
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#define LOCK_UNLOCK_2(x,y) LOCK(x); LOCK(y); UNLOCK(y); UNLOCK(x)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate different permutations of the same testcase, using
|
|
|
|
* the same basic lock-dependency/state events:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define GENERATE_TESTCASE(name) \
|
|
|
|
\
|
|
|
|
static void name(void) { E(); }
|
|
|
|
|
|
|
|
#define GENERATE_PERMUTATIONS_2_EVENTS(name) \
|
|
|
|
\
|
|
|
|
static void name##_12(void) { E1(); E2(); } \
|
|
|
|
static void name##_21(void) { E2(); E1(); }
|
|
|
|
|
|
|
|
#define GENERATE_PERMUTATIONS_3_EVENTS(name) \
|
|
|
|
\
|
|
|
|
static void name##_123(void) { E1(); E2(); E3(); } \
|
|
|
|
static void name##_132(void) { E1(); E3(); E2(); } \
|
|
|
|
static void name##_213(void) { E2(); E1(); E3(); } \
|
|
|
|
static void name##_231(void) { E2(); E3(); E1(); } \
|
|
|
|
static void name##_312(void) { E3(); E1(); E2(); } \
|
|
|
|
static void name##_321(void) { E3(); E2(); E1(); }
|
|
|
|
|
|
|
|
/*
|
|
|
|
* AA deadlock:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK(X1); \
|
|
|
|
LOCK(X2); /* this one should fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(AA_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(AA_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(AA_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(AA_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(AA_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(AA_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(AA_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Special-case for read-locking, they are
|
2006-07-03 07:24:51 +00:00
|
|
|
* allowed to recurse on the same lock class:
|
2006-07-03 07:24:48 +00:00
|
|
|
*/
|
|
|
|
static void rlock_AA1(void)
|
|
|
|
{
|
|
|
|
RL(X1);
|
|
|
|
RL(X1); // this one should NOT fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rlock_AA1B(void)
|
|
|
|
{
|
|
|
|
RL(X1);
|
2006-07-03 07:24:51 +00:00
|
|
|
RL(X2); // this one should NOT fail
|
2006-07-03 07:24:48 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void rsem_AA1(void)
|
|
|
|
{
|
|
|
|
RSL(X1);
|
|
|
|
RSL(X1); // this one should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rsem_AA1B(void)
|
|
|
|
{
|
|
|
|
RSL(X1);
|
|
|
|
RSL(X2); // this one should fail
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* The mixing of read and write locks is not allowed:
|
|
|
|
*/
|
|
|
|
static void rlock_AA2(void)
|
|
|
|
{
|
|
|
|
RL(X1);
|
|
|
|
WL(X2); // this one should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rsem_AA2(void)
|
|
|
|
{
|
|
|
|
RSL(X1);
|
|
|
|
WSL(X2); // this one should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rlock_AA3(void)
|
|
|
|
{
|
|
|
|
WL(X1);
|
|
|
|
RL(X2); // this one should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rsem_AA3(void)
|
|
|
|
{
|
|
|
|
WSL(X1);
|
|
|
|
RSL(X2); // this one should fail
|
|
|
|
}
|
|
|
|
|
2017-08-23 11:13:11 +00:00
|
|
|
/*
|
|
|
|
* read_lock(A)
|
|
|
|
* spin_lock(B)
|
|
|
|
* spin_lock(B)
|
|
|
|
* write_lock(A)
|
|
|
|
*/
|
|
|
|
static void rlock_ABBA1(void)
|
|
|
|
{
|
|
|
|
RL(X1);
|
|
|
|
L(Y1);
|
|
|
|
U(Y1);
|
|
|
|
RU(X1);
|
|
|
|
|
|
|
|
L(Y1);
|
|
|
|
WL(X1);
|
|
|
|
WU(X1);
|
|
|
|
U(Y1); // should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rwsem_ABBA1(void)
|
|
|
|
{
|
|
|
|
RSL(X1);
|
|
|
|
ML(Y1);
|
|
|
|
MU(Y1);
|
|
|
|
RSU(X1);
|
|
|
|
|
|
|
|
ML(Y1);
|
|
|
|
WSL(X1);
|
|
|
|
WSU(X1);
|
|
|
|
MU(Y1); // should fail
|
|
|
|
}
|
|
|
|
|
2020-08-07 07:42:32 +00:00
|
|
|
/*
|
|
|
|
* read_lock(A)
|
|
|
|
* spin_lock(B)
|
|
|
|
* spin_lock(B)
|
|
|
|
* write_lock(A)
|
|
|
|
*
|
|
|
|
* This test case is aimed at poking whether the chain cache prevents us from
|
|
|
|
* detecting a read-lock/lock-write deadlock: if the chain cache doesn't differ
|
|
|
|
* read/write locks, the following case may happen
|
|
|
|
*
|
|
|
|
* { read_lock(A)->lock(B) dependency exists }
|
|
|
|
*
|
|
|
|
* P0:
|
|
|
|
* lock(B);
|
|
|
|
* read_lock(A);
|
|
|
|
*
|
|
|
|
* { Not a deadlock, B -> A is added in the chain cache }
|
|
|
|
*
|
|
|
|
* P1:
|
|
|
|
* lock(B);
|
|
|
|
* write_lock(A);
|
|
|
|
*
|
|
|
|
* { B->A found in chain cache, not reported as a deadlock }
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
static void rlock_chaincache_ABBA1(void)
|
|
|
|
{
|
|
|
|
RL(X1);
|
|
|
|
L(Y1);
|
|
|
|
U(Y1);
|
|
|
|
RU(X1);
|
|
|
|
|
|
|
|
L(Y1);
|
|
|
|
RL(X1);
|
|
|
|
RU(X1);
|
|
|
|
U(Y1);
|
|
|
|
|
|
|
|
L(Y1);
|
|
|
|
WL(X1);
|
|
|
|
WU(X1);
|
|
|
|
U(Y1); // should fail
|
|
|
|
}
|
|
|
|
|
2017-08-23 11:13:11 +00:00
|
|
|
/*
|
|
|
|
* read_lock(A)
|
|
|
|
* spin_lock(B)
|
|
|
|
* spin_lock(B)
|
|
|
|
* read_lock(A)
|
|
|
|
*/
|
|
|
|
static void rlock_ABBA2(void)
|
|
|
|
{
|
|
|
|
RL(X1);
|
|
|
|
L(Y1);
|
|
|
|
U(Y1);
|
|
|
|
RU(X1);
|
|
|
|
|
|
|
|
L(Y1);
|
|
|
|
RL(X1);
|
|
|
|
RU(X1);
|
|
|
|
U(Y1); // should NOT fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rwsem_ABBA2(void)
|
|
|
|
{
|
|
|
|
RSL(X1);
|
|
|
|
ML(Y1);
|
|
|
|
MU(Y1);
|
|
|
|
RSU(X1);
|
|
|
|
|
|
|
|
ML(Y1);
|
|
|
|
RSL(X1);
|
|
|
|
RSU(X1);
|
|
|
|
MU(Y1); // should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* write_lock(A)
|
|
|
|
* spin_lock(B)
|
|
|
|
* spin_lock(B)
|
|
|
|
* write_lock(A)
|
|
|
|
*/
|
|
|
|
static void rlock_ABBA3(void)
|
|
|
|
{
|
|
|
|
WL(X1);
|
|
|
|
L(Y1);
|
|
|
|
U(Y1);
|
|
|
|
WU(X1);
|
|
|
|
|
|
|
|
L(Y1);
|
|
|
|
WL(X1);
|
|
|
|
WU(X1);
|
|
|
|
U(Y1); // should fail
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rwsem_ABBA3(void)
|
|
|
|
{
|
|
|
|
WSL(X1);
|
|
|
|
ML(Y1);
|
|
|
|
MU(Y1);
|
|
|
|
WSU(X1);
|
|
|
|
|
|
|
|
ML(Y1);
|
|
|
|
WSL(X1);
|
|
|
|
WSU(X1);
|
|
|
|
MU(Y1); // should fail
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* ABBA deadlock:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK_UNLOCK_2(A, B); \
|
|
|
|
LOCK_UNLOCK_2(B, A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(ABBA_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* AB BC CA deadlock:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK_UNLOCK_2(A, B); \
|
|
|
|
LOCK_UNLOCK_2(B, C); \
|
|
|
|
LOCK_UNLOCK_2(C, A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCA_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* AB CA BC deadlock:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK_UNLOCK_2(A, B); \
|
|
|
|
LOCK_UNLOCK_2(C, A); \
|
|
|
|
LOCK_UNLOCK_2(B, C); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(ABCABC_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* AB BC CD DA deadlock:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK_UNLOCK_2(A, B); \
|
|
|
|
LOCK_UNLOCK_2(B, C); \
|
|
|
|
LOCK_UNLOCK_2(C, D); \
|
|
|
|
LOCK_UNLOCK_2(D, A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(ABBCCDDA_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* AB CD BD DA deadlock:
|
|
|
|
*/
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK_UNLOCK_2(A, B); \
|
|
|
|
LOCK_UNLOCK_2(C, D); \
|
|
|
|
LOCK_UNLOCK_2(B, D); \
|
|
|
|
LOCK_UNLOCK_2(D, A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBDDA_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* AB CD BC DA deadlock:
|
|
|
|
*/
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK_UNLOCK_2(A, B); \
|
|
|
|
LOCK_UNLOCK_2(C, D); \
|
|
|
|
LOCK_UNLOCK_2(B, C); \
|
|
|
|
LOCK_UNLOCK_2(D, A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(ABCDBCDA_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
2021-11-29 17:46:50 +00:00
|
|
|
#ifdef CONFIG_PREEMPT_RT
|
|
|
|
# define RT_PREPARE_DBL_UNLOCK() { migrate_disable(); rcu_read_lock(); }
|
|
|
|
#else
|
|
|
|
# define RT_PREPARE_DBL_UNLOCK()
|
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* Double unlock:
|
|
|
|
*/
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK(A); \
|
2021-11-29 17:46:50 +00:00
|
|
|
RT_PREPARE_DBL_UNLOCK(); \
|
2006-07-03 07:24:48 +00:00
|
|
|
UNLOCK(A); \
|
|
|
|
UNLOCK(A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(double_unlock_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* initializing a held lock:
|
|
|
|
*/
|
|
|
|
#define E() \
|
|
|
|
\
|
|
|
|
LOCK(A); \
|
|
|
|
INIT(A); /* fail */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 6 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_TESTCASE(init_held_spin)
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_TESTCASE(init_held_wlock)
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_TESTCASE(init_held_rlock)
|
|
|
|
#include "locking-selftest-mutex.h"
|
|
|
|
GENERATE_TESTCASE(init_held_mutex)
|
|
|
|
#include "locking-selftest-wsem.h"
|
|
|
|
GENERATE_TESTCASE(init_held_wsem)
|
|
|
|
#include "locking-selftest-rsem.h"
|
|
|
|
GENERATE_TESTCASE(init_held_rsem)
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#include "locking-selftest-rtmutex.h"
|
|
|
|
GENERATE_TESTCASE(init_held_rtmutex);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#undef E
|
|
|
|
|
|
|
|
/*
|
|
|
|
* locking an irq-safe lock with irqs enabled:
|
|
|
|
*/
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 24 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_hard_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_hard_wlock)
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-spin-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_soft_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* Enabling hardirqs with a softirq-safe lock held:
|
|
|
|
*/
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
SOFTIRQ_ENTER(); \
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
SOFTIRQ_EXIT();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
HARDIRQ_DISABLE(); \
|
|
|
|
LOCK(A); \
|
|
|
|
HARDIRQ_ENABLE(); \
|
|
|
|
UNLOCK(A);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 12 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2A_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2A_wlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2A_rlock)
|
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* Enabling irqs with an irq-safe lock held:
|
|
|
|
*/
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
IRQ_DISABLE(); \
|
|
|
|
LOCK(A); \
|
|
|
|
IRQ_ENABLE(); \
|
|
|
|
UNLOCK(A);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 24 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_hard_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_hard_wlock)
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-spin-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_soft_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Acquiring a irq-unsafe lock while holding an irq-safe-lock:
|
|
|
|
*/
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
LOCK(A); \
|
|
|
|
LOCK(B); \
|
|
|
|
UNLOCK(B); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
LOCK(B); \
|
|
|
|
UNLOCK(B);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 36 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_hard_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_hard_wlock)
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-spin-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_soft_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If a lock turns into softirq-safe, but earlier it took
|
|
|
|
* a softirq-unsafe lock:
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
IRQ_DISABLE(); \
|
|
|
|
LOCK(A); \
|
|
|
|
LOCK(B); \
|
|
|
|
UNLOCK(B); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
IRQ_ENABLE();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
LOCK(B); \
|
|
|
|
UNLOCK(B);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 36 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_hard_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_hard_wlock)
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-spin-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_soft_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
|
|
|
/*
|
|
|
|
* read-lock / write-lock irq inversion.
|
|
|
|
*
|
|
|
|
* Deadlock scenario:
|
|
|
|
*
|
|
|
|
* CPU#1 is at #1, i.e. it has write-locked A, but has not
|
|
|
|
* taken B yet.
|
|
|
|
*
|
|
|
|
* CPU#2 is at #2, i.e. it has locked B.
|
|
|
|
*
|
|
|
|
* Hardirq hits CPU#2 at point #2 and is trying to read-lock A.
|
|
|
|
*
|
|
|
|
* The deadlock occurs because CPU#1 will spin on B, and CPU#2
|
|
|
|
* will spin on A.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
IRQ_DISABLE(); \
|
|
|
|
WL(A); \
|
|
|
|
LOCK(B); \
|
|
|
|
UNLOCK(B); \
|
|
|
|
WU(A); \
|
|
|
|
IRQ_ENABLE();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
LOCK(B); \
|
|
|
|
UNLOCK(B);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
RL(A); \
|
|
|
|
RU(A); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 36 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-spin-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_hard_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-hardirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_hard_wlock)
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-spin-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_soft_spin)
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock-softirq.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
2020-08-07 07:42:35 +00:00
|
|
|
/*
|
|
|
|
* write-read / write-read / write-read deadlock even if read is recursive
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
WL(X1); \
|
|
|
|
RL(Y1); \
|
|
|
|
RU(Y1); \
|
|
|
|
WU(X1);
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
WL(Y1); \
|
|
|
|
RL(Z1); \
|
|
|
|
RU(Z1); \
|
|
|
|
WU(Y1);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
WL(Z1); \
|
|
|
|
RL(X1); \
|
|
|
|
RU(X1); \
|
|
|
|
WU(Z1);
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(W1R2_W2R3_W3R1)
|
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
|
|
|
/*
|
|
|
|
* write-write / read-read / write-read deadlock even if read is recursive
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
WL(X1); \
|
|
|
|
WL(Y1); \
|
|
|
|
WU(Y1); \
|
|
|
|
WU(X1);
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
RL(Y1); \
|
|
|
|
RL(Z1); \
|
|
|
|
RU(Z1); \
|
|
|
|
RU(Y1);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
WL(Z1); \
|
|
|
|
RL(X1); \
|
|
|
|
RU(X1); \
|
|
|
|
WU(Z1);
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(W1W2_R2R3_W3R1)
|
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
|
|
|
/*
|
|
|
|
* write-write / read-read / read-write is not deadlock when read is recursive
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
WL(X1); \
|
|
|
|
WL(Y1); \
|
|
|
|
WU(Y1); \
|
|
|
|
WU(X1);
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
RL(Y1); \
|
|
|
|
RL(Z1); \
|
|
|
|
RU(Z1); \
|
|
|
|
RU(Y1);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
RL(Z1); \
|
|
|
|
WL(X1); \
|
|
|
|
WU(X1); \
|
|
|
|
RU(Z1);
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(W1R2_R2R3_W3W1)
|
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
|
|
|
/*
|
|
|
|
* write-read / read-read / write-write is not deadlock when read is recursive
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
WL(X1); \
|
|
|
|
RL(Y1); \
|
|
|
|
RU(Y1); \
|
|
|
|
WU(X1);
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
RL(Y1); \
|
|
|
|
RL(Z1); \
|
|
|
|
RU(Z1); \
|
|
|
|
RU(Y1);
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
WL(Z1); \
|
|
|
|
WL(X1); \
|
|
|
|
WU(X1); \
|
|
|
|
WU(Z1);
|
|
|
|
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(W1W2_R2R3_R3W1)
|
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* read-lock / write-lock recursion that is actually safe.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
IRQ_DISABLE(); \
|
|
|
|
WL(A); \
|
|
|
|
WU(A); \
|
|
|
|
IRQ_ENABLE();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
RL(A); \
|
|
|
|
RU(A); \
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
2020-08-07 07:42:34 +00:00
|
|
|
LOCK(A); \
|
2006-07-03 07:24:48 +00:00
|
|
|
L(B); \
|
|
|
|
U(B); \
|
2020-08-07 07:42:34 +00:00
|
|
|
UNLOCK(A); \
|
2006-07-03 07:24:48 +00:00
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
/*
|
2020-08-07 07:42:34 +00:00
|
|
|
* Generate 24 testcases:
|
2006-07-03 07:24:48 +00:00
|
|
|
*/
|
|
|
|
#include "locking-selftest-hardirq.h"
|
2020-08-07 07:42:34 +00:00
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_hard_wlock)
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-softirq.h"
|
2020-08-07 07:42:34 +00:00
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
|
|
|
|
/*
|
|
|
|
* read-lock / write-lock recursion that is unsafe.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
IRQ_DISABLE(); \
|
|
|
|
L(B); \
|
2020-08-07 07:42:34 +00:00
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
2006-07-03 07:24:48 +00:00
|
|
|
U(B); \
|
|
|
|
IRQ_ENABLE();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
RL(A); \
|
|
|
|
RU(A); \
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
L(B); \
|
|
|
|
U(B); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
/*
|
2020-08-07 07:42:34 +00:00
|
|
|
* Generate 24 testcases:
|
2006-07-03 07:24:48 +00:00
|
|
|
*/
|
|
|
|
#include "locking-selftest-hardirq.h"
|
2020-08-07 07:42:34 +00:00
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_hard_wlock)
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2006-07-03 07:24:48 +00:00
|
|
|
#include "locking-selftest-softirq.h"
|
2020-08-07 07:42:34 +00:00
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2020-08-07 07:42:38 +00:00
|
|
|
#undef E1
|
|
|
|
#undef E2
|
|
|
|
#undef E3
|
|
|
|
/*
|
|
|
|
* read-lock / write-lock recursion that is unsafe.
|
|
|
|
*
|
|
|
|
* A is a ENABLED_*_READ lock
|
|
|
|
* B is a USED_IN_*_READ lock
|
|
|
|
*
|
|
|
|
* read_lock(A);
|
|
|
|
* write_lock(B);
|
|
|
|
* <interrupt>
|
|
|
|
* read_lock(B);
|
|
|
|
* write_lock(A); // if this one is read_lock(), no deadlock
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define E1() \
|
|
|
|
\
|
|
|
|
IRQ_DISABLE(); \
|
|
|
|
WL(B); \
|
|
|
|
LOCK(A); \
|
|
|
|
UNLOCK(A); \
|
|
|
|
WU(B); \
|
|
|
|
IRQ_ENABLE();
|
|
|
|
|
|
|
|
#define E2() \
|
|
|
|
\
|
|
|
|
RL(A); \
|
|
|
|
RU(A); \
|
|
|
|
|
|
|
|
#define E3() \
|
|
|
|
\
|
|
|
|
IRQ_ENTER(); \
|
|
|
|
RL(B); \
|
|
|
|
RU(B); \
|
|
|
|
IRQ_EXIT();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Generate 24 testcases:
|
|
|
|
*/
|
|
|
|
#include "locking-selftest-hardirq.h"
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_hard_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_hard_wlock)
|
|
|
|
|
2021-11-29 17:46:51 +00:00
|
|
|
#ifndef CONFIG_PREEMPT_RT
|
2020-08-07 07:42:38 +00:00
|
|
|
#include "locking-selftest-softirq.h"
|
|
|
|
#include "locking-selftest-rlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_soft_rlock)
|
|
|
|
|
|
|
|
#include "locking-selftest-wlock.h"
|
|
|
|
GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_soft_wlock)
|
2021-11-29 17:46:51 +00:00
|
|
|
#endif
|
2020-08-07 07:42:38 +00:00
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
|
|
|
# define I_SPINLOCK(x) lockdep_reset_lock(&lock_##x.dep_map)
|
2020-12-08 10:31:12 +00:00
|
|
|
# define I_RAW_SPINLOCK(x) lockdep_reset_lock(&raw_lock_##x.dep_map)
|
2006-07-03 07:24:48 +00:00
|
|
|
# define I_RWLOCK(x) lockdep_reset_lock(&rwlock_##x.dep_map)
|
|
|
|
# define I_MUTEX(x) lockdep_reset_lock(&mutex_##x.dep_map)
|
|
|
|
# define I_RWSEM(x) lockdep_reset_lock(&rwsem_##x.dep_map)
|
2013-06-20 11:31:24 +00:00
|
|
|
# define I_WW(x) lockdep_reset_lock(&x.dep_map)
|
2021-11-29 17:46:49 +00:00
|
|
|
# define I_LOCAL_LOCK(x) lockdep_reset_lock(this_cpu_ptr(&local_##x.dep_map))
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
# define I_RTMUTEX(x) lockdep_reset_lock(&rtmutex_##x.dep_map)
|
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
#else
|
|
|
|
# define I_SPINLOCK(x)
|
2020-12-08 10:31:12 +00:00
|
|
|
# define I_RAW_SPINLOCK(x)
|
2006-07-03 07:24:48 +00:00
|
|
|
# define I_RWLOCK(x)
|
|
|
|
# define I_MUTEX(x)
|
|
|
|
# define I_RWSEM(x)
|
2013-06-20 11:31:24 +00:00
|
|
|
# define I_WW(x)
|
2020-12-09 15:06:06 +00:00
|
|
|
# define I_LOCAL_LOCK(x)
|
2006-07-03 07:24:48 +00:00
|
|
|
#endif
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifndef I_RTMUTEX
|
|
|
|
# define I_RTMUTEX(x)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#define I2_RTMUTEX(x) rt_mutex_init(&rtmutex_##x)
|
|
|
|
#else
|
|
|
|
#define I2_RTMUTEX(x)
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#define I1(x) \
|
|
|
|
do { \
|
|
|
|
I_SPINLOCK(x); \
|
|
|
|
I_RWLOCK(x); \
|
|
|
|
I_MUTEX(x); \
|
|
|
|
I_RWSEM(x); \
|
2017-05-31 15:11:49 +00:00
|
|
|
I_RTMUTEX(x); \
|
2006-07-03 07:24:48 +00:00
|
|
|
} while (0)
|
|
|
|
|
|
|
|
#define I2(x) \
|
|
|
|
do { \
|
2020-10-30 11:37:43 +00:00
|
|
|
spin_lock_init(&lock_##x); \
|
2006-07-03 07:24:48 +00:00
|
|
|
rwlock_init(&rwlock_##x); \
|
|
|
|
mutex_init(&mutex_##x); \
|
|
|
|
init_rwsem(&rwsem_##x); \
|
2017-05-31 15:11:49 +00:00
|
|
|
I2_RTMUTEX(x); \
|
2006-07-03 07:24:48 +00:00
|
|
|
} while (0)
|
|
|
|
|
|
|
|
static void reset_locks(void)
|
|
|
|
{
|
|
|
|
local_irq_disable();
|
2013-06-20 11:31:24 +00:00
|
|
|
lockdep_free_key_range(&ww_lockdep.acquire_key, 1);
|
|
|
|
lockdep_free_key_range(&ww_lockdep.mutex_key, 1);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
I1(A); I1(B); I1(C); I1(D);
|
|
|
|
I1(X1); I1(X2); I1(Y1); I1(Y2); I1(Z1); I1(Z2);
|
2013-06-20 11:31:42 +00:00
|
|
|
I_WW(t); I_WW(t2); I_WW(o.base); I_WW(o2.base); I_WW(o3.base);
|
2020-12-08 10:31:12 +00:00
|
|
|
I_RAW_SPINLOCK(A); I_RAW_SPINLOCK(B);
|
2020-12-09 15:06:06 +00:00
|
|
|
I_LOCAL_LOCK(A);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
lockdep_reset();
|
2020-12-09 15:06:06 +00:00
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
I2(A); I2(B); I2(C); I2(D);
|
|
|
|
init_shared_classes();
|
2020-12-08 10:31:12 +00:00
|
|
|
raw_spin_lock_init(&raw_lock_A);
|
|
|
|
raw_spin_lock_init(&raw_lock_B);
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock_init(this_cpu_ptr(&local_A));
|
2013-06-20 11:31:24 +00:00
|
|
|
|
2013-06-20 11:31:42 +00:00
|
|
|
ww_mutex_init(&o, &ww_lockdep); ww_mutex_init(&o2, &ww_lockdep); ww_mutex_init(&o3, &ww_lockdep);
|
2013-06-20 11:31:24 +00:00
|
|
|
memset(&t, 0, sizeof(t)); memset(&t2, 0, sizeof(t2));
|
|
|
|
memset(&ww_lockdep.acquire_key, 0, sizeof(ww_lockdep.acquire_key));
|
|
|
|
memset(&ww_lockdep.mutex_key, 0, sizeof(ww_lockdep.mutex_key));
|
2006-07-03 07:24:48 +00:00
|
|
|
local_irq_enable();
|
|
|
|
}
|
|
|
|
|
|
|
|
#undef I
|
|
|
|
|
|
|
|
static int testcase_total;
|
|
|
|
static int testcase_successes;
|
|
|
|
static int expected_testcase_failures;
|
|
|
|
static int unexpected_testcase_failures;
|
|
|
|
|
|
|
|
static void dotest(void (*testcase_fn)(void), int expected, int lockclass_mask)
|
|
|
|
{
|
2021-11-29 17:46:50 +00:00
|
|
|
int saved_preempt_count = preempt_count();
|
|
|
|
#ifdef CONFIG_PREEMPT_RT
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
int saved_mgd_count = current->migration_disabled;
|
|
|
|
#endif
|
|
|
|
int saved_rcu_count = current->rcu_read_lock_nesting;
|
|
|
|
#endif
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
WARN_ON(irqs_disabled());
|
|
|
|
|
2020-12-09 15:42:57 +00:00
|
|
|
debug_locks_silent = !(debug_locks_verbose & lockclass_mask);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
testcase_fn();
|
|
|
|
/*
|
|
|
|
* Filter out expected failures:
|
|
|
|
*/
|
|
|
|
#ifndef CONFIG_PROVE_LOCKING
|
2013-06-20 11:31:51 +00:00
|
|
|
if (expected == FAILURE && debug_locks) {
|
2013-06-20 11:31:24 +00:00
|
|
|
expected_testcase_failures++;
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("failed|");
|
2013-06-20 11:31:51 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
#endif
|
|
|
|
if (debug_locks != expected) {
|
2013-06-20 11:31:24 +00:00
|
|
|
unexpected_testcase_failures++;
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("FAILED|");
|
2006-07-03 07:24:48 +00:00
|
|
|
} else {
|
|
|
|
testcase_successes++;
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" ok |");
|
2006-07-03 07:24:48 +00:00
|
|
|
}
|
|
|
|
testcase_total++;
|
|
|
|
|
2020-12-09 15:42:57 +00:00
|
|
|
if (debug_locks_verbose & lockclass_mask)
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" lockclass mask: %x, debug_locks: %d, expected: %d\n",
|
2006-07-03 07:24:48 +00:00
|
|
|
lockclass_mask, debug_locks, expected);
|
|
|
|
/*
|
|
|
|
* Some tests (e.g. double-unlock) might corrupt the preemption
|
|
|
|
* count, so restore it:
|
|
|
|
*/
|
2013-08-14 12:55:24 +00:00
|
|
|
preempt_count_set(saved_preempt_count);
|
2021-11-29 17:46:50 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_PREEMPT_RT
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
while (current->migration_disabled > saved_mgd_count)
|
|
|
|
migrate_enable();
|
|
|
|
#endif
|
|
|
|
|
|
|
|
while (current->rcu_read_lock_nesting > saved_rcu_count)
|
|
|
|
rcu_read_unlock();
|
|
|
|
WARN_ON_ONCE(current->rcu_read_lock_nesting < saved_rcu_count);
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#ifdef CONFIG_TRACE_IRQFLAGS
|
|
|
|
if (softirq_count())
|
|
|
|
current->softirqs_enabled = 0;
|
|
|
|
else
|
|
|
|
current->softirqs_enabled = 1;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
reset_locks();
|
|
|
|
}
|
|
|
|
|
2017-05-31 15:11:49 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
#define dotest_rt(fn, e, m) dotest((fn), (e), (m))
|
|
|
|
#else
|
|
|
|
#define dotest_rt(fn, e, m)
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
static inline void print_testname(const char *testname)
|
|
|
|
{
|
|
|
|
printk("%33s:", testname);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define DO_TESTCASE_1(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_1B(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2020-08-07 07:42:35 +00:00
|
|
|
#define DO_TESTCASE_1RR(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
pr_cont(" |"); \
|
|
|
|
dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
#define DO_TESTCASE_1RRB(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
pr_cont(" |"); \
|
|
|
|
dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#define DO_TESTCASE_3(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
dotest(name##_spin_##nr, FAILURE, LOCKTYPE_SPIN); \
|
|
|
|
dotest(name##_wlock_##nr, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_rlock_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_3RW(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
dotest(name##_spin_##nr, FAILURE, LOCKTYPE_SPIN|LOCKTYPE_RWLOCK);\
|
|
|
|
dotest(name##_wlock_##nr, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_rlock_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2020-08-07 07:42:34 +00:00
|
|
|
#define DO_TESTCASE_2RW(desc, name, nr) \
|
|
|
|
print_testname(desc"/"#nr); \
|
|
|
|
pr_cont(" |"); \
|
|
|
|
dotest(name##_wlock_##nr, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_rlock_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
#define DO_TESTCASE_2x2RW(desc, name, nr) \
|
|
|
|
DO_TESTCASE_2RW("hard-"desc, name##_hard, nr) \
|
2021-11-29 17:46:51 +00:00
|
|
|
NON_RT(DO_TESTCASE_2RW("soft-"desc, name##_soft, nr)) \
|
2020-08-07 07:42:34 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_6x2x2RW(desc, name) \
|
|
|
|
DO_TESTCASE_2x2RW(desc, name, 123); \
|
|
|
|
DO_TESTCASE_2x2RW(desc, name, 132); \
|
|
|
|
DO_TESTCASE_2x2RW(desc, name, 213); \
|
|
|
|
DO_TESTCASE_2x2RW(desc, name, 231); \
|
|
|
|
DO_TESTCASE_2x2RW(desc, name, 312); \
|
|
|
|
DO_TESTCASE_2x2RW(desc, name, 321);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#define DO_TESTCASE_6(desc, name) \
|
|
|
|
print_testname(desc); \
|
|
|
|
dotest(name##_spin, FAILURE, LOCKTYPE_SPIN); \
|
|
|
|
dotest(name##_wlock, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_rlock, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_mutex, FAILURE, LOCKTYPE_MUTEX); \
|
|
|
|
dotest(name##_wsem, FAILURE, LOCKTYPE_RWSEM); \
|
|
|
|
dotest(name##_rsem, FAILURE, LOCKTYPE_RWSEM); \
|
2017-05-31 15:11:49 +00:00
|
|
|
dotest_rt(name##_rtmutex, FAILURE, LOCKTYPE_RTMUTEX); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_6_SUCCESS(desc, name) \
|
|
|
|
print_testname(desc); \
|
|
|
|
dotest(name##_spin, SUCCESS, LOCKTYPE_SPIN); \
|
|
|
|
dotest(name##_wlock, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_rlock, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_mutex, SUCCESS, LOCKTYPE_MUTEX); \
|
|
|
|
dotest(name##_wsem, SUCCESS, LOCKTYPE_RWSEM); \
|
|
|
|
dotest(name##_rsem, SUCCESS, LOCKTYPE_RWSEM); \
|
2017-05-31 15:11:49 +00:00
|
|
|
dotest_rt(name##_rtmutex, SUCCESS, LOCKTYPE_RTMUTEX); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* 'read' variant: rlocks must not trigger.
|
|
|
|
*/
|
|
|
|
#define DO_TESTCASE_6R(desc, name) \
|
|
|
|
print_testname(desc); \
|
|
|
|
dotest(name##_spin, FAILURE, LOCKTYPE_SPIN); \
|
|
|
|
dotest(name##_wlock, FAILURE, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_rlock, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(name##_mutex, FAILURE, LOCKTYPE_MUTEX); \
|
|
|
|
dotest(name##_wsem, FAILURE, LOCKTYPE_RWSEM); \
|
|
|
|
dotest(name##_rsem, FAILURE, LOCKTYPE_RWSEM); \
|
2017-05-31 15:11:49 +00:00
|
|
|
dotest_rt(name##_rtmutex, FAILURE, LOCKTYPE_RTMUTEX); \
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_2I(desc, name, nr) \
|
|
|
|
DO_TESTCASE_1("hard-"desc, name##_hard, nr); \
|
2021-11-29 17:46:51 +00:00
|
|
|
NON_RT(DO_TESTCASE_1("soft-"desc, name##_soft, nr));
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_2IB(desc, name, nr) \
|
|
|
|
DO_TESTCASE_1B("hard-"desc, name##_hard, nr); \
|
2021-11-29 17:46:51 +00:00
|
|
|
NON_RT(DO_TESTCASE_1B("soft-"desc, name##_soft, nr));
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_6I(desc, name, nr) \
|
|
|
|
DO_TESTCASE_3("hard-"desc, name##_hard, nr); \
|
2021-11-29 17:46:51 +00:00
|
|
|
NON_RT(DO_TESTCASE_3("soft-"desc, name##_soft, nr));
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_6IRW(desc, name, nr) \
|
|
|
|
DO_TESTCASE_3RW("hard-"desc, name##_hard, nr); \
|
2021-11-29 17:46:51 +00:00
|
|
|
NON_RT(DO_TESTCASE_3RW("soft-"desc, name##_soft, nr));
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
#define DO_TESTCASE_2x3(desc, name) \
|
|
|
|
DO_TESTCASE_3(desc, name, 12); \
|
|
|
|
DO_TESTCASE_3(desc, name, 21);
|
|
|
|
|
|
|
|
#define DO_TESTCASE_2x6(desc, name) \
|
|
|
|
DO_TESTCASE_6I(desc, name, 12); \
|
|
|
|
DO_TESTCASE_6I(desc, name, 21);
|
|
|
|
|
|
|
|
#define DO_TESTCASE_6x2(desc, name) \
|
|
|
|
DO_TESTCASE_2I(desc, name, 123); \
|
|
|
|
DO_TESTCASE_2I(desc, name, 132); \
|
|
|
|
DO_TESTCASE_2I(desc, name, 213); \
|
|
|
|
DO_TESTCASE_2I(desc, name, 231); \
|
|
|
|
DO_TESTCASE_2I(desc, name, 312); \
|
|
|
|
DO_TESTCASE_2I(desc, name, 321);
|
|
|
|
|
|
|
|
#define DO_TESTCASE_6x2B(desc, name) \
|
|
|
|
DO_TESTCASE_2IB(desc, name, 123); \
|
|
|
|
DO_TESTCASE_2IB(desc, name, 132); \
|
|
|
|
DO_TESTCASE_2IB(desc, name, 213); \
|
|
|
|
DO_TESTCASE_2IB(desc, name, 231); \
|
|
|
|
DO_TESTCASE_2IB(desc, name, 312); \
|
|
|
|
DO_TESTCASE_2IB(desc, name, 321);
|
|
|
|
|
2020-08-07 07:42:35 +00:00
|
|
|
#define DO_TESTCASE_6x1RR(desc, name) \
|
|
|
|
DO_TESTCASE_1RR(desc, name, 123); \
|
|
|
|
DO_TESTCASE_1RR(desc, name, 132); \
|
|
|
|
DO_TESTCASE_1RR(desc, name, 213); \
|
|
|
|
DO_TESTCASE_1RR(desc, name, 231); \
|
|
|
|
DO_TESTCASE_1RR(desc, name, 312); \
|
|
|
|
DO_TESTCASE_1RR(desc, name, 321);
|
|
|
|
|
|
|
|
#define DO_TESTCASE_6x1RRB(desc, name) \
|
|
|
|
DO_TESTCASE_1RRB(desc, name, 123); \
|
|
|
|
DO_TESTCASE_1RRB(desc, name, 132); \
|
|
|
|
DO_TESTCASE_1RRB(desc, name, 213); \
|
|
|
|
DO_TESTCASE_1RRB(desc, name, 231); \
|
|
|
|
DO_TESTCASE_1RRB(desc, name, 312); \
|
|
|
|
DO_TESTCASE_1RRB(desc, name, 321);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
#define DO_TESTCASE_6x6(desc, name) \
|
|
|
|
DO_TESTCASE_6I(desc, name, 123); \
|
|
|
|
DO_TESTCASE_6I(desc, name, 132); \
|
|
|
|
DO_TESTCASE_6I(desc, name, 213); \
|
|
|
|
DO_TESTCASE_6I(desc, name, 231); \
|
|
|
|
DO_TESTCASE_6I(desc, name, 312); \
|
|
|
|
DO_TESTCASE_6I(desc, name, 321);
|
|
|
|
|
|
|
|
#define DO_TESTCASE_6x6RW(desc, name) \
|
|
|
|
DO_TESTCASE_6IRW(desc, name, 123); \
|
|
|
|
DO_TESTCASE_6IRW(desc, name, 132); \
|
|
|
|
DO_TESTCASE_6IRW(desc, name, 213); \
|
|
|
|
DO_TESTCASE_6IRW(desc, name, 231); \
|
|
|
|
DO_TESTCASE_6IRW(desc, name, 312); \
|
|
|
|
DO_TESTCASE_6IRW(desc, name, 321);
|
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
static void ww_test_fail_acquire(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t.stamp++;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
|
|
|
|
if (WARN_ON(!o.ctx) ||
|
|
|
|
WARN_ON(ret))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* No lockdep test, pure API */
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret != -EALREADY);
|
|
|
|
|
|
|
|
ret = WWT(&o);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp++;
|
|
|
|
ret = WWL(&o, &t2);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
if (WWT(&o))
|
|
|
|
WWU(&o);
|
|
|
|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
|
|
|
else
|
|
|
|
DEBUG_LOCKS_WARN_ON(1);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
#ifdef CONFIG_PREEMPT_RT
|
|
|
|
#define ww_mutex_base_lock(b) rt_mutex_lock(b)
|
|
|
|
#define ww_mutex_base_trylock(b) rt_mutex_trylock(b)
|
|
|
|
#define ww_mutex_base_lock_nest_lock(b, b2) rt_mutex_lock_nest_lock(b, b2)
|
|
|
|
#define ww_mutex_base_lock_interruptible(b) rt_mutex_lock_interruptible(b)
|
|
|
|
#define ww_mutex_base_lock_killable(b) rt_mutex_lock_killable(b)
|
|
|
|
#define ww_mutex_base_unlock(b) rt_mutex_unlock(b)
|
|
|
|
#else
|
|
|
|
#define ww_mutex_base_lock(b) mutex_lock(b)
|
|
|
|
#define ww_mutex_base_trylock(b) mutex_trylock(b)
|
|
|
|
#define ww_mutex_base_lock_nest_lock(b, b2) mutex_lock_nest_lock(b, b2)
|
|
|
|
#define ww_mutex_base_lock_interruptible(b) mutex_lock_interruptible(b)
|
|
|
|
#define ww_mutex_base_lock_killable(b) mutex_lock_killable(b)
|
|
|
|
#define ww_mutex_base_unlock(b) mutex_unlock(b)
|
|
|
|
#endif
|
|
|
|
|
2013-06-20 11:31:30 +00:00
|
|
|
static void ww_test_normal(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* None of the ww_mutex codepaths should be taken in the 'normal'
|
|
|
|
* mutex calls. The easiest way to verify this is by using the
|
|
|
|
* normal mutex calls, and making sure o.ctx is unmodified.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* mutex_lock (and indirectly, mutex_lock_nested) */
|
|
|
|
o.ctx = (void *)~0UL;
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o.base);
|
|
|
|
ww_mutex_base_unlock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
WARN_ON(o.ctx != (void *)~0UL);
|
|
|
|
|
|
|
|
/* mutex_lock_interruptible (and *_nested) */
|
|
|
|
o.ctx = (void *)~0UL;
|
2021-11-29 17:46:52 +00:00
|
|
|
ret = ww_mutex_base_lock_interruptible(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
if (!ret)
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
else
|
|
|
|
WARN_ON(1);
|
|
|
|
WARN_ON(o.ctx != (void *)~0UL);
|
|
|
|
|
|
|
|
/* mutex_lock_killable (and *_nested) */
|
|
|
|
o.ctx = (void *)~0UL;
|
2021-11-29 17:46:52 +00:00
|
|
|
ret = ww_mutex_base_lock_killable(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
if (!ret)
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
else
|
|
|
|
WARN_ON(1);
|
|
|
|
WARN_ON(o.ctx != (void *)~0UL);
|
|
|
|
|
|
|
|
/* trylock, succeeding */
|
|
|
|
o.ctx = (void *)~0UL;
|
2021-11-29 17:46:52 +00:00
|
|
|
ret = ww_mutex_base_trylock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
WARN_ON(!ret);
|
|
|
|
if (ret)
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
else
|
|
|
|
WARN_ON(1);
|
|
|
|
WARN_ON(o.ctx != (void *)~0UL);
|
|
|
|
|
|
|
|
/* trylock, failing */
|
|
|
|
o.ctx = (void *)~0UL;
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o.base);
|
|
|
|
ret = ww_mutex_base_trylock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
WARN_ON(ret);
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
WARN_ON(o.ctx != (void *)~0UL);
|
|
|
|
|
|
|
|
/* nest_lock */
|
|
|
|
o.ctx = (void *)~0UL;
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock_nest_lock(&o.base, &t);
|
|
|
|
ww_mutex_base_unlock(&o.base);
|
2013-06-20 11:31:30 +00:00
|
|
|
WARN_ON(o.ctx != (void *)~0UL);
|
|
|
|
}
|
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
static void ww_test_two_contexts(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
WWAI(&t2);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_diff_class(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
2021-11-29 17:46:52 +00:00
|
|
|
#ifdef DEBUG_WW_MUTEXES
|
2013-06-20 11:31:24 +00:00
|
|
|
t.ww_class = NULL;
|
|
|
|
#endif
|
|
|
|
WWL(&o, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_done_twice(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
WWAD(&t);
|
|
|
|
WWAD(&t);
|
|
|
|
WWAF(&t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_unlock_twice(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
WWAD(&t);
|
|
|
|
WWAF(&t);
|
|
|
|
WWAF(&t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_fini_early(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
WWL(&o, &t);
|
|
|
|
WWAD(&t);
|
|
|
|
WWAF(&t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_lock_after_done(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
WWAD(&t);
|
|
|
|
WWL(&o, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_object_unlock_twice(void)
|
|
|
|
{
|
|
|
|
WWL1(&o);
|
|
|
|
WWU(&o);
|
|
|
|
WWU(&o);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_object_lock_unbalanced(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
WWL(&o, &t);
|
|
|
|
t.acquired = 0;
|
|
|
|
WWU(&o);
|
|
|
|
WWAF(&t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_object_lock_stale_context(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
o.ctx = &t2;
|
|
|
|
WWL(&o, &t);
|
|
|
|
}
|
|
|
|
|
2013-06-20 11:31:42 +00:00
|
|
|
static void ww_test_edeadlk_normal(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
o2.ctx = NULL;
|
|
|
|
mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o2.base);
|
2013-06-20 11:31:42 +00:00
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
WWL(&o2, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_normal_slow(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
o2.ctx = NULL;
|
|
|
|
mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o2.base);
|
2013-06-20 11:31:42 +00:00
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
ww_mutex_lock_slow(&o2, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_no_unlock(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
o2.ctx = NULL;
|
|
|
|
mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o2.base);
|
2013-06-20 11:31:42 +00:00
|
|
|
|
|
|
|
WWL(&o2, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_no_unlock_slow(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
o2.ctx = NULL;
|
|
|
|
mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_unlock(&o2.base);
|
2013-06-20 11:31:42 +00:00
|
|
|
|
|
|
|
ww_mutex_lock_slow(&o2, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_acquire_more(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
ret = WWL(&o3, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_acquire_more_slow(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
ww_mutex_lock_slow(&o3, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_acquire_more_edeadlk(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o3.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o3.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o3.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
ret = WWL(&o3, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_acquire_more_edeadlk_slow(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o3.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o3.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o3.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
|
|
|
|
ww_mutex_lock_slow(&o3, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_acquire_wrong(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
if (!ret)
|
|
|
|
WWU(&o2);
|
|
|
|
|
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
ret = WWL(&o3, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_edeadlk_acquire_wrong_slow(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-11-29 17:46:52 +00:00
|
|
|
ww_mutex_base_lock(&o2.base);
|
2019-09-19 16:09:40 +00:00
|
|
|
mutex_release(&o2.base.dep_map, _THIS_IP_);
|
2013-06-20 11:31:42 +00:00
|
|
|
o2.ctx = &t2;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
t2 = t;
|
|
|
|
t2.stamp--;
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret != -EDEADLK);
|
|
|
|
if (!ret)
|
|
|
|
WWU(&o2);
|
|
|
|
|
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
ww_mutex_lock_slow(&o3, &t);
|
|
|
|
}
|
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
static void ww_test_spin_nest_unlocked(void)
|
|
|
|
{
|
2020-10-30 11:37:43 +00:00
|
|
|
spin_lock_nest_lock(&lock_A, &o.base);
|
2013-06-20 11:31:24 +00:00
|
|
|
U(A);
|
|
|
|
}
|
|
|
|
|
2020-11-02 05:37:42 +00:00
|
|
|
/* This is not a deadlock, because we have X1 to serialize Y1 and Y2 */
|
|
|
|
static void ww_test_spin_nest_lock(void)
|
|
|
|
{
|
|
|
|
spin_lock(&lock_X1);
|
|
|
|
spin_lock_nest_lock(&lock_Y1, &lock_X1);
|
|
|
|
spin_lock(&lock_A);
|
|
|
|
spin_lock_nest_lock(&lock_Y2, &lock_X1);
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
spin_unlock(&lock_Y2);
|
|
|
|
spin_unlock(&lock_Y1);
|
|
|
|
spin_unlock(&lock_X1);
|
|
|
|
}
|
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
static void ww_test_unneeded_slow(void)
|
|
|
|
{
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ww_mutex_lock_slow(&o, &t);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_block(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
WWL1(&o2);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_try(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWT(&o2);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
WWU(&o2);
|
|
|
|
WWU(&o);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_context_context(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
|
|
|
|
WWU(&o2);
|
|
|
|
WWU(&o);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_try_block(void)
|
|
|
|
{
|
|
|
|
bool ret;
|
|
|
|
|
|
|
|
ret = WWT(&o);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
|
|
|
|
WWL1(&o2);
|
|
|
|
WWU(&o2);
|
|
|
|
WWU(&o);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_try_try(void)
|
|
|
|
{
|
|
|
|
bool ret;
|
|
|
|
|
|
|
|
ret = WWT(&o);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
ret = WWT(&o2);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
WWU(&o2);
|
|
|
|
WWU(&o);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_try_context(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = WWT(&o);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_block_block(void)
|
|
|
|
{
|
|
|
|
WWL1(&o);
|
|
|
|
WWL1(&o2);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_block_try(void)
|
|
|
|
{
|
|
|
|
bool ret;
|
|
|
|
|
|
|
|
WWL1(&o);
|
|
|
|
ret = WWT(&o2);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_block_context(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WWL1(&o);
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ret = WWL(&o2, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_spin_block(void)
|
|
|
|
{
|
|
|
|
L(A);
|
|
|
|
U(A);
|
|
|
|
|
|
|
|
WWL1(&o);
|
|
|
|
L(A);
|
|
|
|
U(A);
|
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
L(A);
|
|
|
|
WWL1(&o);
|
|
|
|
WWU(&o);
|
|
|
|
U(A);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_spin_try(void)
|
|
|
|
{
|
|
|
|
bool ret;
|
|
|
|
|
|
|
|
L(A);
|
|
|
|
U(A);
|
|
|
|
|
|
|
|
ret = WWT(&o);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
L(A);
|
|
|
|
U(A);
|
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
L(A);
|
|
|
|
ret = WWT(&o);
|
|
|
|
WARN_ON(!ret);
|
|
|
|
WWU(&o);
|
|
|
|
U(A);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_test_spin_context(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
L(A);
|
|
|
|
U(A);
|
|
|
|
|
|
|
|
WWAI(&t);
|
|
|
|
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
L(A);
|
|
|
|
U(A);
|
|
|
|
WWU(&o);
|
|
|
|
|
|
|
|
L(A);
|
|
|
|
ret = WWL(&o, &t);
|
|
|
|
WARN_ON(ret);
|
|
|
|
WWU(&o);
|
|
|
|
U(A);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ww_tests(void)
|
|
|
|
{
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
printk(" | Wound/wait tests |\n");
|
|
|
|
printk(" ---------------------\n");
|
|
|
|
|
|
|
|
print_testname("ww api failures");
|
|
|
|
dotest(ww_test_fail_acquire, SUCCESS, LOCKTYPE_WW);
|
2013-06-20 11:31:30 +00:00
|
|
|
dotest(ww_test_normal, SUCCESS, LOCKTYPE_WW);
|
2013-06-20 11:31:24 +00:00
|
|
|
dotest(ww_test_unneeded_slow, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
print_testname("ww contexts mixing");
|
|
|
|
dotest(ww_test_two_contexts, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_diff_class, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
print_testname("finishing ww context");
|
|
|
|
dotest(ww_test_context_done_twice, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_context_unlock_twice, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_context_fini_early, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_context_lock_after_done, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
print_testname("locking mismatches");
|
|
|
|
dotest(ww_test_object_unlock_twice, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_object_lock_unbalanced, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_object_lock_stale_context, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
2013-06-20 11:31:42 +00:00
|
|
|
print_testname("EDEADLK handling");
|
|
|
|
dotest(ww_test_edeadlk_normal, SUCCESS, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_normal_slow, SUCCESS, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_no_unlock, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_no_unlock_slow, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_acquire_more, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_acquire_more_slow, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_acquire_more_edeadlk, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_acquire_more_edeadlk_slow, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_acquire_wrong, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_edeadlk_acquire_wrong_slow, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:42 +00:00
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
print_testname("spinlock nest unlocked");
|
|
|
|
dotest(ww_test_spin_nest_unlocked, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
2020-11-02 05:37:42 +00:00
|
|
|
print_testname("spinlock nest test");
|
|
|
|
dotest(ww_test_spin_nest_lock, SUCCESS, LOCKTYPE_WW);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
printk(" -----------------------------------------------------\n");
|
|
|
|
printk(" |block | try |context|\n");
|
|
|
|
printk(" -----------------------------------------------------\n");
|
|
|
|
|
|
|
|
print_testname("context");
|
|
|
|
dotest(ww_test_context_block, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_context_try, SUCCESS, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_context_context, SUCCESS, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
print_testname("try");
|
|
|
|
dotest(ww_test_try_block, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_try_try, SUCCESS, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_try_context, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
print_testname("block");
|
|
|
|
dotest(ww_test_block_block, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_block_try, SUCCESS, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_block_context, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
|
|
|
|
print_testname("spinlock");
|
|
|
|
dotest(ww_test_spin_block, FAILURE, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_spin_try, SUCCESS, LOCKTYPE_WW);
|
|
|
|
dotest(ww_test_spin_context, FAILURE, LOCKTYPE_WW);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2013-06-20 11:31:24 +00:00
|
|
|
}
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2020-08-07 07:42:37 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* <in hardirq handler>
|
|
|
|
* read_lock(&A);
|
|
|
|
* <hardirq disable>
|
|
|
|
* spin_lock(&B);
|
|
|
|
* spin_lock(&B);
|
|
|
|
* read_lock(&A);
|
|
|
|
*
|
|
|
|
* is a deadlock.
|
|
|
|
*/
|
|
|
|
static void queued_read_lock_hardirq_RE_Er(void)
|
|
|
|
{
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
read_lock(&rwlock_A);
|
|
|
|
LOCK(B);
|
|
|
|
UNLOCK(B);
|
|
|
|
read_unlock(&rwlock_A);
|
|
|
|
HARDIRQ_EXIT();
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
LOCK(B);
|
|
|
|
read_lock(&rwlock_A);
|
|
|
|
read_unlock(&rwlock_A);
|
|
|
|
UNLOCK(B);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* <in hardirq handler>
|
|
|
|
* spin_lock(&B);
|
|
|
|
* <hardirq disable>
|
|
|
|
* read_lock(&A);
|
|
|
|
* read_lock(&A);
|
|
|
|
* spin_lock(&B);
|
|
|
|
*
|
|
|
|
* is not a deadlock.
|
|
|
|
*/
|
|
|
|
static void queued_read_lock_hardirq_ER_rE(void)
|
|
|
|
{
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
LOCK(B);
|
|
|
|
read_lock(&rwlock_A);
|
|
|
|
read_unlock(&rwlock_A);
|
|
|
|
UNLOCK(B);
|
|
|
|
HARDIRQ_EXIT();
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
read_lock(&rwlock_A);
|
|
|
|
LOCK(B);
|
|
|
|
UNLOCK(B);
|
|
|
|
read_unlock(&rwlock_A);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* <hardirq disable>
|
|
|
|
* spin_lock(&B);
|
|
|
|
* read_lock(&A);
|
|
|
|
* <in hardirq handler>
|
|
|
|
* spin_lock(&B);
|
|
|
|
* read_lock(&A);
|
|
|
|
*
|
|
|
|
* is a deadlock. Because the two read_lock()s are both non-recursive readers.
|
|
|
|
*/
|
|
|
|
static void queued_read_lock_hardirq_inversion(void)
|
|
|
|
{
|
|
|
|
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
LOCK(B);
|
|
|
|
UNLOCK(B);
|
|
|
|
HARDIRQ_EXIT();
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
LOCK(B);
|
|
|
|
read_lock(&rwlock_A);
|
|
|
|
read_unlock(&rwlock_A);
|
|
|
|
UNLOCK(B);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
|
|
|
|
read_lock(&rwlock_A);
|
|
|
|
read_unlock(&rwlock_A);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void queued_read_lock_tests(void)
|
|
|
|
{
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
printk(" | queued read lock tests |\n");
|
|
|
|
printk(" ---------------------------\n");
|
|
|
|
print_testname("hardirq read-lock/lock-read");
|
|
|
|
dotest(queued_read_lock_hardirq_RE_Er, FAILURE, LOCKTYPE_RWLOCK);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("hardirq lock-read/read-lock");
|
|
|
|
dotest(queued_read_lock_hardirq_ER_rE, SUCCESS, LOCKTYPE_RWLOCK);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("hardirq inversion");
|
|
|
|
dotest(queued_read_lock_hardirq_inversion, FAILURE, LOCKTYPE_RWLOCK);
|
|
|
|
pr_cont("\n");
|
|
|
|
}
|
|
|
|
|
2020-12-15 03:08:38 +00:00
|
|
|
static void fs_reclaim_correct_nesting(void)
|
|
|
|
{
|
|
|
|
fs_reclaim_acquire(GFP_KERNEL);
|
|
|
|
might_alloc(GFP_NOFS);
|
|
|
|
fs_reclaim_release(GFP_KERNEL);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void fs_reclaim_wrong_nesting(void)
|
|
|
|
{
|
|
|
|
fs_reclaim_acquire(GFP_KERNEL);
|
|
|
|
might_alloc(GFP_KERNEL);
|
|
|
|
fs_reclaim_release(GFP_KERNEL);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void fs_reclaim_protected_nesting(void)
|
|
|
|
{
|
|
|
|
unsigned int flags;
|
|
|
|
|
|
|
|
fs_reclaim_acquire(GFP_KERNEL);
|
|
|
|
flags = memalloc_nofs_save();
|
|
|
|
might_alloc(GFP_KERNEL);
|
|
|
|
memalloc_nofs_restore(flags);
|
|
|
|
fs_reclaim_release(GFP_KERNEL);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void fs_reclaim_tests(void)
|
|
|
|
{
|
|
|
|
printk(" --------------------\n");
|
|
|
|
printk(" | fs_reclaim tests |\n");
|
|
|
|
printk(" --------------------\n");
|
|
|
|
|
|
|
|
print_testname("correct nesting");
|
|
|
|
dotest(fs_reclaim_correct_nesting, SUCCESS, 0);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("wrong nesting");
|
|
|
|
dotest(fs_reclaim_wrong_nesting, FAILURE, 0);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("protected nesting");
|
|
|
|
dotest(fs_reclaim_protected_nesting, SUCCESS, 0);
|
|
|
|
pr_cont("\n");
|
|
|
|
}
|
|
|
|
|
2020-12-08 10:31:12 +00:00
|
|
|
#define __guard(cleanup) __maybe_unused __attribute__((__cleanup__(cleanup)))
|
|
|
|
|
|
|
|
static void hardirq_exit(int *_)
|
|
|
|
{
|
|
|
|
HARDIRQ_EXIT();
|
|
|
|
}
|
|
|
|
|
|
|
|
#define HARDIRQ_CONTEXT(name, ...) \
|
|
|
|
int hardirq_guard_##name __guard(hardirq_exit); \
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
|
|
|
|
#define NOTTHREADED_HARDIRQ_CONTEXT(name, ...) \
|
|
|
|
int notthreaded_hardirq_guard_##name __guard(hardirq_exit); \
|
|
|
|
local_irq_disable(); \
|
|
|
|
__irq_enter(); \
|
|
|
|
WARN_ON(!in_irq());
|
|
|
|
|
|
|
|
static void softirq_exit(int *_)
|
|
|
|
{
|
|
|
|
SOFTIRQ_EXIT();
|
|
|
|
}
|
|
|
|
|
|
|
|
#define SOFTIRQ_CONTEXT(name, ...) \
|
|
|
|
int softirq_guard_##name __guard(softirq_exit); \
|
|
|
|
SOFTIRQ_ENTER();
|
|
|
|
|
|
|
|
static void rcu_exit(int *_)
|
|
|
|
{
|
|
|
|
rcu_read_unlock();
|
|
|
|
}
|
|
|
|
|
|
|
|
#define RCU_CONTEXT(name, ...) \
|
|
|
|
int rcu_guard_##name __guard(rcu_exit); \
|
|
|
|
rcu_read_lock();
|
|
|
|
|
|
|
|
static void rcu_bh_exit(int *_)
|
|
|
|
{
|
|
|
|
rcu_read_unlock_bh();
|
|
|
|
}
|
|
|
|
|
|
|
|
#define RCU_BH_CONTEXT(name, ...) \
|
|
|
|
int rcu_bh_guard_##name __guard(rcu_bh_exit); \
|
|
|
|
rcu_read_lock_bh();
|
|
|
|
|
|
|
|
static void rcu_sched_exit(int *_)
|
|
|
|
{
|
|
|
|
rcu_read_unlock_sched();
|
|
|
|
}
|
|
|
|
|
|
|
|
#define RCU_SCHED_CONTEXT(name, ...) \
|
|
|
|
int rcu_sched_guard_##name __guard(rcu_sched_exit); \
|
|
|
|
rcu_read_lock_sched();
|
|
|
|
|
|
|
|
static void raw_spinlock_exit(raw_spinlock_t **lock)
|
|
|
|
{
|
|
|
|
raw_spin_unlock(*lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define RAW_SPINLOCK_CONTEXT(name, lock) \
|
|
|
|
raw_spinlock_t *raw_spinlock_guard_##name __guard(raw_spinlock_exit) = &(lock); \
|
|
|
|
raw_spin_lock(&(lock));
|
|
|
|
|
|
|
|
static void spinlock_exit(spinlock_t **lock)
|
|
|
|
{
|
|
|
|
spin_unlock(*lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define SPINLOCK_CONTEXT(name, lock) \
|
|
|
|
spinlock_t *spinlock_guard_##name __guard(spinlock_exit) = &(lock); \
|
|
|
|
spin_lock(&(lock));
|
|
|
|
|
|
|
|
static void mutex_exit(struct mutex **lock)
|
|
|
|
{
|
|
|
|
mutex_unlock(*lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define MUTEX_CONTEXT(name, lock) \
|
|
|
|
struct mutex *mutex_guard_##name __guard(mutex_exit) = &(lock); \
|
|
|
|
mutex_lock(&(lock));
|
|
|
|
|
|
|
|
#define GENERATE_2_CONTEXT_TESTCASE(outer, outer_lock, inner, inner_lock) \
|
|
|
|
\
|
|
|
|
static void __maybe_unused inner##_in_##outer(void) \
|
|
|
|
{ \
|
|
|
|
outer##_CONTEXT(_, outer_lock); \
|
|
|
|
{ \
|
|
|
|
inner##_CONTEXT(_, inner_lock); \
|
|
|
|
} \
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* wait contexts (considering PREEMPT_RT)
|
|
|
|
*
|
|
|
|
* o: inner is allowed in outer
|
|
|
|
* x: inner is disallowed in outer
|
|
|
|
*
|
|
|
|
* \ inner | RCU | RAW_SPIN | SPIN | MUTEX
|
|
|
|
* outer \ | | | |
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* HARDIRQ | o | o | o | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* NOTTHREADED_IRQ| o | o | x | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* SOFTIRQ | o | o | o | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* RCU | o | o | o | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* RCU_BH | o | o | o | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* RCU_SCHED | o | o | x | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* RAW_SPIN | o | o | x | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* SPIN | o | o | o | x
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
* MUTEX | o | o | o | o
|
|
|
|
* ---------------+-------+----------+------+-------
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(HARDIRQ, , inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(NOTTHREADED_HARDIRQ, , inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(SOFTIRQ, , inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(RCU, , inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(RCU_BH, , inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(RCU_SCHED, , inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(RAW_SPINLOCK, raw_lock_A, inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(SPINLOCK, lock_A, inner, inner_lock) \
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE(MUTEX, mutex_A, inner, inner_lock)
|
|
|
|
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(RCU, )
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(RAW_SPINLOCK, raw_lock_B)
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(SPINLOCK, lock_B)
|
|
|
|
GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(MUTEX, mutex_B)
|
|
|
|
|
|
|
|
/* the outer context allows all kinds of preemption */
|
|
|
|
#define DO_CONTEXT_TESTCASE_OUTER_PREEMPTIBLE(outer) \
|
|
|
|
dotest(RCU_in_##outer, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(RAW_SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN); \
|
|
|
|
dotest(SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN); \
|
|
|
|
dotest(MUTEX_in_##outer, SUCCESS, LOCKTYPE_MUTEX); \
|
|
|
|
|
|
|
|
/*
|
|
|
|
* the outer context only allows the preemption introduced by spinlock_t (which
|
|
|
|
* is a sleepable lock for PREEMPT_RT)
|
|
|
|
*/
|
|
|
|
#define DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(outer) \
|
|
|
|
dotest(RCU_in_##outer, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(RAW_SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN); \
|
|
|
|
dotest(SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN); \
|
|
|
|
dotest(MUTEX_in_##outer, FAILURE, LOCKTYPE_MUTEX); \
|
|
|
|
|
|
|
|
/* the outer doesn't allows any kind of preemption */
|
|
|
|
#define DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(outer) \
|
|
|
|
dotest(RCU_in_##outer, SUCCESS, LOCKTYPE_RWLOCK); \
|
|
|
|
dotest(RAW_SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN); \
|
|
|
|
dotest(SPINLOCK_in_##outer, FAILURE, LOCKTYPE_SPIN); \
|
|
|
|
dotest(MUTEX_in_##outer, FAILURE, LOCKTYPE_MUTEX); \
|
|
|
|
|
|
|
|
static void wait_context_tests(void)
|
|
|
|
{
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
printk(" | wait context tests |\n");
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
printk(" | rcu | raw | spin |mutex |\n");
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
print_testname("in hardirq context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(HARDIRQ);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in hardirq context (not threaded)");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(NOTTHREADED_HARDIRQ);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in softirq context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(SOFTIRQ);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in RCU context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(RCU);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in RCU-bh context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(RCU_BH);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in RCU-sched context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(RCU_SCHED);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in RAW_SPINLOCK context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(RAW_SPINLOCK);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in SPINLOCK context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(SPINLOCK);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("in MUTEX context");
|
|
|
|
DO_CONTEXT_TESTCASE_OUTER_PREEMPTIBLE(MUTEX);
|
|
|
|
pr_cont("\n");
|
|
|
|
}
|
|
|
|
|
2020-12-09 15:06:06 +00:00
|
|
|
static void local_lock_2(void)
|
|
|
|
{
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock(&local_A); /* IRQ-ON */
|
|
|
|
local_unlock(&local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
spin_lock(&lock_A); /* IN-IRQ */
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_EXIT()
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_A);
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock(&local_A); /* IN-IRQ <-> IRQ-ON cycle, false */
|
|
|
|
local_unlock(&local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
}
|
|
|
|
|
|
|
|
static void local_lock_3A(void)
|
|
|
|
{
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock(&local_A); /* IRQ-ON */
|
2020-12-09 15:06:06 +00:00
|
|
|
spin_lock(&lock_B); /* IRQ-ON */
|
|
|
|
spin_unlock(&lock_B);
|
2021-11-29 17:46:49 +00:00
|
|
|
local_unlock(&local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
spin_lock(&lock_A); /* IN-IRQ */
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_EXIT()
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_A);
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock(&local_A); /* IN-IRQ <-> IRQ-ON cycle only if we count local_lock(), false */
|
|
|
|
local_unlock(&local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
}
|
|
|
|
|
|
|
|
static void local_lock_3B(void)
|
|
|
|
{
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock(&local_A); /* IRQ-ON */
|
2020-12-09 15:06:06 +00:00
|
|
|
spin_lock(&lock_B); /* IRQ-ON */
|
|
|
|
spin_unlock(&lock_B);
|
2021-11-29 17:46:49 +00:00
|
|
|
local_unlock(&local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
spin_lock(&lock_A); /* IN-IRQ */
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_EXIT()
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_A);
|
2021-11-29 17:46:49 +00:00
|
|
|
local_lock(&local_A); /* IN-IRQ <-> IRQ-ON cycle only if we count local_lock(), false */
|
|
|
|
local_unlock(&local_A);
|
2020-12-09 15:06:06 +00:00
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_A);
|
|
|
|
spin_lock(&lock_B); /* IN-IRQ <-> IRQ-ON cycle, true */
|
|
|
|
spin_unlock(&lock_B);
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
static void local_lock_tests(void)
|
|
|
|
{
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
printk(" | local_lock tests |\n");
|
|
|
|
printk(" ---------------------\n");
|
|
|
|
|
|
|
|
print_testname("local_lock inversion 2");
|
|
|
|
dotest(local_lock_2, SUCCESS, LOCKTYPE_LL);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("local_lock inversion 3A");
|
|
|
|
dotest(local_lock_3A, SUCCESS, LOCKTYPE_LL);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
|
|
|
print_testname("local_lock inversion 3B");
|
|
|
|
dotest(local_lock_3B, FAILURE, LOCKTYPE_LL);
|
|
|
|
pr_cont("\n");
|
|
|
|
}
|
|
|
|
|
2021-06-18 17:01:10 +00:00
|
|
|
static void hardirq_deadlock_softirq_not_deadlock(void)
|
|
|
|
{
|
|
|
|
/* mutex_A is hardirq-unsafe and softirq-unsafe */
|
|
|
|
/* mutex_A -> lock_C */
|
|
|
|
mutex_lock(&mutex_A);
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_C);
|
|
|
|
spin_unlock(&lock_C);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
mutex_unlock(&mutex_A);
|
|
|
|
|
|
|
|
/* lock_A is hardirq-safe */
|
|
|
|
HARDIRQ_ENTER();
|
|
|
|
spin_lock(&lock_A);
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_EXIT();
|
|
|
|
|
|
|
|
/* lock_A -> lock_B */
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_A);
|
|
|
|
spin_lock(&lock_B);
|
|
|
|
spin_unlock(&lock_B);
|
|
|
|
spin_unlock(&lock_A);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
|
|
|
|
/* lock_B -> lock_C */
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_B);
|
|
|
|
spin_lock(&lock_C);
|
|
|
|
spin_unlock(&lock_C);
|
|
|
|
spin_unlock(&lock_B);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
|
|
|
|
/* lock_D is softirq-safe */
|
|
|
|
SOFTIRQ_ENTER();
|
|
|
|
spin_lock(&lock_D);
|
|
|
|
spin_unlock(&lock_D);
|
|
|
|
SOFTIRQ_EXIT();
|
|
|
|
|
|
|
|
/* And lock_D is hardirq-unsafe */
|
|
|
|
SOFTIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_D);
|
|
|
|
spin_unlock(&lock_D);
|
|
|
|
SOFTIRQ_ENABLE();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* mutex_A -> lock_C -> lock_D is softirq-unsafe -> softirq-safe, not
|
|
|
|
* deadlock.
|
|
|
|
*
|
|
|
|
* lock_A -> lock_B -> lock_C -> lock_D is hardirq-safe ->
|
|
|
|
* hardirq-unsafe, deadlock.
|
|
|
|
*/
|
|
|
|
HARDIRQ_DISABLE();
|
|
|
|
spin_lock(&lock_C);
|
|
|
|
spin_lock(&lock_D);
|
|
|
|
spin_unlock(&lock_D);
|
|
|
|
spin_unlock(&lock_C);
|
|
|
|
HARDIRQ_ENABLE();
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
void locking_selftest(void)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Got a locking failure before the selftest ran?
|
|
|
|
*/
|
|
|
|
if (!debug_locks) {
|
|
|
|
printk("----------------------------------\n");
|
|
|
|
printk("| Locking API testsuite disabled |\n");
|
|
|
|
printk("----------------------------------\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2020-08-07 07:42:20 +00:00
|
|
|
/*
|
|
|
|
* treats read_lock() as recursive read locks for testing purpose
|
|
|
|
*/
|
|
|
|
force_read_lock_recursive = 1;
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
/*
|
|
|
|
* Run the testsuite:
|
|
|
|
*/
|
|
|
|
printk("------------------------\n");
|
|
|
|
printk("| Locking API testsuite:\n");
|
|
|
|
printk("----------------------------------------------------------------------------\n");
|
2021-11-29 17:46:49 +00:00
|
|
|
printk(" | spin |wlock |rlock |mutex | wsem | rsem |rtmutex\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
|
|
|
|
init_shared_classes();
|
2019-02-14 23:00:44 +00:00
|
|
|
lockdep_set_selftest_task(current);
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2006-07-03 07:24:51 +00:00
|
|
|
DO_TESTCASE_6R("A-A deadlock", AA);
|
2006-07-03 07:24:48 +00:00
|
|
|
DO_TESTCASE_6R("A-B-B-A deadlock", ABBA);
|
|
|
|
DO_TESTCASE_6R("A-B-B-C-C-A deadlock", ABBCCA);
|
|
|
|
DO_TESTCASE_6R("A-B-C-A-B-C deadlock", ABCABC);
|
|
|
|
DO_TESTCASE_6R("A-B-B-C-C-D-D-A deadlock", ABBCCDDA);
|
|
|
|
DO_TESTCASE_6R("A-B-C-D-B-D-D-A deadlock", ABCDBDDA);
|
|
|
|
DO_TESTCASE_6R("A-B-C-D-B-C-D-A deadlock", ABCDBCDA);
|
|
|
|
DO_TESTCASE_6("double unlock", double_unlock);
|
|
|
|
DO_TESTCASE_6("initialize held", init_held);
|
|
|
|
|
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
print_testname("recursive read-lock");
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rlock_AA1, SUCCESS, LOCKTYPE_RWLOCK);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rsem_AA1, FAILURE, LOCKTYPE_RWSEM);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
print_testname("recursive read-lock #2");
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:51 +00:00
|
|
|
dotest(rlock_AA1B, SUCCESS, LOCKTYPE_RWLOCK);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rsem_AA1B, FAILURE, LOCKTYPE_RWSEM);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
print_testname("mixed read-write-lock");
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rlock_AA2, FAILURE, LOCKTYPE_RWLOCK);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rsem_AA2, FAILURE, LOCKTYPE_RWSEM);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
|
|
|
print_testname("mixed write-read-lock");
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rlock_AA3, FAILURE, LOCKTYPE_RWLOCK);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont(" |");
|
2006-07-03 07:24:48 +00:00
|
|
|
dotest(rsem_AA3, FAILURE, LOCKTYPE_RWSEM);
|
2016-11-24 22:45:28 +00:00
|
|
|
pr_cont("\n");
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2017-08-23 11:13:11 +00:00
|
|
|
print_testname("mixed read-lock/lock-write ABBA");
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rlock_ABBA1, FAILURE, LOCKTYPE_RWLOCK);
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rwsem_ABBA1, FAILURE, LOCKTYPE_RWSEM);
|
|
|
|
|
|
|
|
print_testname("mixed read-lock/lock-read ABBA");
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rlock_ABBA2, SUCCESS, LOCKTYPE_RWLOCK);
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rwsem_ABBA2, FAILURE, LOCKTYPE_RWSEM);
|
|
|
|
|
|
|
|
print_testname("mixed write-lock/lock-write ABBA");
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rlock_ABBA3, FAILURE, LOCKTYPE_RWLOCK);
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rwsem_ABBA3, FAILURE, LOCKTYPE_RWSEM);
|
|
|
|
|
2020-08-07 07:42:32 +00:00
|
|
|
print_testname("chain cached mixed R-L/L-W ABBA");
|
|
|
|
pr_cont(" |");
|
|
|
|
dotest(rlock_chaincache_ABBA1, FAILURE, LOCKTYPE_RWLOCK);
|
|
|
|
|
2020-08-07 07:42:35 +00:00
|
|
|
DO_TESTCASE_6x1RRB("rlock W1R2/W2R3/W3R1", W1R2_W2R3_W3R1);
|
|
|
|
DO_TESTCASE_6x1RRB("rlock W1W2/R2R3/W3R1", W1W2_R2R3_W3R1);
|
|
|
|
DO_TESTCASE_6x1RR("rlock W1W2/R2R3/R3W1", W1W2_R2R3_R3W1);
|
|
|
|
DO_TESTCASE_6x1RR("rlock W1R2/R2R3/W3W1", W1R2_R2R3_W3W1);
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
printk(" --------------------------------------------------------------------------\n");
|
|
|
|
/*
|
|
|
|
* irq-context testcases:
|
|
|
|
*/
|
|
|
|
DO_TESTCASE_2x6("irqs-on + irq-safe-A", irqsafe1);
|
2021-11-29 17:46:51 +00:00
|
|
|
NON_RT(DO_TESTCASE_2x3("sirq-safe-A => hirqs-on", irqsafe2A));
|
2006-07-03 07:24:48 +00:00
|
|
|
DO_TESTCASE_2x6("safe-A + irqs-on", irqsafe2B);
|
|
|
|
DO_TESTCASE_6x6("safe-A + unsafe-B #1", irqsafe3);
|
|
|
|
DO_TESTCASE_6x6("safe-A + unsafe-B #2", irqsafe4);
|
|
|
|
DO_TESTCASE_6x6RW("irq lock-inversion", irq_inversion);
|
|
|
|
|
2020-08-07 07:42:34 +00:00
|
|
|
DO_TESTCASE_6x2x2RW("irq read-recursion", irq_read_recursion);
|
|
|
|
DO_TESTCASE_6x2x2RW("irq read-recursion #2", irq_read_recursion2);
|
2020-08-07 07:42:38 +00:00
|
|
|
DO_TESTCASE_6x2x2RW("irq read-recursion #3", irq_read_recursion3);
|
2006-07-03 07:24:48 +00:00
|
|
|
|
2013-06-20 11:31:24 +00:00
|
|
|
ww_tests();
|
|
|
|
|
2020-08-07 07:42:20 +00:00
|
|
|
force_read_lock_recursive = 0;
|
|
|
|
/*
|
|
|
|
* queued_read_lock() specific test cases can be put here
|
|
|
|
*/
|
2020-08-07 07:42:37 +00:00
|
|
|
if (IS_ENABLED(CONFIG_QUEUED_RWLOCKS))
|
|
|
|
queued_read_lock_tests();
|
2020-08-07 07:42:20 +00:00
|
|
|
|
2020-12-15 03:08:38 +00:00
|
|
|
fs_reclaim_tests();
|
|
|
|
|
2020-12-08 10:31:12 +00:00
|
|
|
/* Wait context test cases that are specific for RAW_LOCK_NESTING */
|
|
|
|
if (IS_ENABLED(CONFIG_PROVE_RAW_LOCK_NESTING))
|
|
|
|
wait_context_tests();
|
|
|
|
|
2020-12-09 15:06:06 +00:00
|
|
|
local_lock_tests();
|
|
|
|
|
2021-06-18 17:01:10 +00:00
|
|
|
print_testname("hardirq_unsafe_softirq_safe");
|
|
|
|
dotest(hardirq_deadlock_softirq_not_deadlock, FAILURE, LOCKTYPE_SPECIAL);
|
|
|
|
pr_cont("\n");
|
|
|
|
|
2006-07-03 07:24:48 +00:00
|
|
|
if (unexpected_testcase_failures) {
|
|
|
|
printk("-----------------------------------------------------------------\n");
|
|
|
|
debug_locks = 0;
|
|
|
|
printk("BUG: %3d unexpected failures (out of %3d) - debugging disabled! |\n",
|
|
|
|
unexpected_testcase_failures, testcase_total);
|
|
|
|
printk("-----------------------------------------------------------------\n");
|
|
|
|
} else if (expected_testcase_failures && testcase_successes) {
|
|
|
|
printk("--------------------------------------------------------\n");
|
|
|
|
printk("%3d out of %3d testcases failed, as expected. |\n",
|
|
|
|
expected_testcase_failures, testcase_total);
|
|
|
|
printk("----------------------------------------------------\n");
|
|
|
|
debug_locks = 1;
|
|
|
|
} else if (expected_testcase_failures && !testcase_successes) {
|
|
|
|
printk("--------------------------------------------------------\n");
|
|
|
|
printk("All %3d testcases failed, as expected. |\n",
|
|
|
|
expected_testcase_failures);
|
|
|
|
printk("----------------------------------------\n");
|
|
|
|
debug_locks = 1;
|
|
|
|
} else {
|
|
|
|
printk("-------------------------------------------------------\n");
|
|
|
|
printk("Good, all %3d testcases passed! |\n",
|
|
|
|
testcase_successes);
|
|
|
|
printk("---------------------------------\n");
|
|
|
|
debug_locks = 1;
|
|
|
|
}
|
2019-02-14 23:00:44 +00:00
|
|
|
lockdep_set_selftest_task(NULL);
|
2006-07-03 07:24:48 +00:00
|
|
|
debug_locks_silent = 0;
|
|
|
|
}
|