mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 08:38:45 +02:00 
			
		
		
		
	 9a75bd0c52
			
		
	
	
		9a75bd0c52
		
	
	
	
	
		
			
			The ww-mutex selftest operates directly on ww_mutex::base and assumes its type is struct mutex. This isn't true on PREEMPT_RT which turns the mutex into a rtmutex. Add a ww_mutex_base_ abstraction which maps to the relevant mutex_ or rt_mutex_ function. Change the CONFIG_DEBUG_MUTEXES ifdef to DEBUG_WW_MUTEXES. The latter is true for the MUTEX and RTMUTEX implementation of WW-MUTEX. The assignment is required in order to pass the tests. Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: https://lore.kernel.org/r/20211129174654.668506-10-bigeasy@linutronix.de
		
			
				
	
	
		
			3015 lines
		
	
	
	
		
			66 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			3015 lines
		
	
	
	
		
			66 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * lib/locking-selftest.c
 | |
|  *
 | |
|  * Testsuite for various locking APIs: spinlocks, rwlocks,
 | |
|  * mutexes and rw-semaphores.
 | |
|  *
 | |
|  * It is checking both false positives and false negatives.
 | |
|  *
 | |
|  * Started by Ingo Molnar:
 | |
|  *
 | |
|  *  Copyright (C) 2006 Red Hat, Inc., Ingo Molnar <mingo@redhat.com>
 | |
|  */
 | |
| #include <linux/rwsem.h>
 | |
| #include <linux/mutex.h>
 | |
| #include <linux/ww_mutex.h>
 | |
| #include <linux/sched.h>
 | |
| #include <linux/sched/mm.h>
 | |
| #include <linux/delay.h>
 | |
| #include <linux/lockdep.h>
 | |
| #include <linux/spinlock.h>
 | |
| #include <linux/kallsyms.h>
 | |
| #include <linux/interrupt.h>
 | |
| #include <linux/debug_locks.h>
 | |
| #include <linux/irqflags.h>
 | |
| #include <linux/rtmutex.h>
 | |
| #include <linux/local_lock.h>
 | |
| 
 | |
| #ifdef CONFIG_PREEMPT_RT
 | |
| # define NON_RT(...)
 | |
| #else
 | |
| # define NON_RT(...)	__VA_ARGS__
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * Change this to 1 if you want to see the failure printouts:
 | |
|  */
 | |
| static unsigned int debug_locks_verbose;
 | |
| unsigned int force_read_lock_recursive;
 | |
| 
 | |
| static DEFINE_WD_CLASS(ww_lockdep);
 | |
| 
 | |
| static int __init setup_debug_locks_verbose(char *str)
 | |
| {
 | |
| 	get_option(&str, &debug_locks_verbose);
 | |
| 
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| __setup("debug_locks_verbose=", setup_debug_locks_verbose);
 | |
| 
 | |
| #define FAILURE		0
 | |
| #define SUCCESS		1
 | |
| 
 | |
| #define LOCKTYPE_SPIN	0x1
 | |
| #define LOCKTYPE_RWLOCK	0x2
 | |
| #define LOCKTYPE_MUTEX	0x4
 | |
| #define LOCKTYPE_RWSEM	0x8
 | |
| #define LOCKTYPE_WW	0x10
 | |
| #define LOCKTYPE_RTMUTEX 0x20
 | |
| #define LOCKTYPE_LL	0x40
 | |
| #define LOCKTYPE_SPECIAL 0x80
 | |
| 
 | |
| static struct ww_acquire_ctx t, t2;
 | |
| static struct ww_mutex o, o2, o3;
 | |
| 
 | |
| /*
 | |
|  * Normal standalone locks, for the circular and irq-context
 | |
|  * dependency tests:
 | |
|  */
 | |
| static DEFINE_SPINLOCK(lock_A);
 | |
| static DEFINE_SPINLOCK(lock_B);
 | |
| static DEFINE_SPINLOCK(lock_C);
 | |
| static DEFINE_SPINLOCK(lock_D);
 | |
| 
 | |
| static DEFINE_RAW_SPINLOCK(raw_lock_A);
 | |
| static DEFINE_RAW_SPINLOCK(raw_lock_B);
 | |
| 
 | |
| static DEFINE_RWLOCK(rwlock_A);
 | |
| static DEFINE_RWLOCK(rwlock_B);
 | |
| static DEFINE_RWLOCK(rwlock_C);
 | |
| static DEFINE_RWLOCK(rwlock_D);
 | |
| 
 | |
| static DEFINE_MUTEX(mutex_A);
 | |
| static DEFINE_MUTEX(mutex_B);
 | |
| static DEFINE_MUTEX(mutex_C);
 | |
| static DEFINE_MUTEX(mutex_D);
 | |
| 
 | |
| static DECLARE_RWSEM(rwsem_A);
 | |
| static DECLARE_RWSEM(rwsem_B);
 | |
| static DECLARE_RWSEM(rwsem_C);
 | |
| static DECLARE_RWSEM(rwsem_D);
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| 
 | |
| static DEFINE_RT_MUTEX(rtmutex_A);
 | |
| static DEFINE_RT_MUTEX(rtmutex_B);
 | |
| static DEFINE_RT_MUTEX(rtmutex_C);
 | |
| static DEFINE_RT_MUTEX(rtmutex_D);
 | |
| 
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * Locks that we initialize dynamically as well so that
 | |
|  * e.g. X1 and X2 becomes two instances of the same class,
 | |
|  * but X* and Y* are different classes. We do this so that
 | |
|  * we do not trigger a real lockup:
 | |
|  */
 | |
| static DEFINE_SPINLOCK(lock_X1);
 | |
| static DEFINE_SPINLOCK(lock_X2);
 | |
| static DEFINE_SPINLOCK(lock_Y1);
 | |
| static DEFINE_SPINLOCK(lock_Y2);
 | |
| static DEFINE_SPINLOCK(lock_Z1);
 | |
| static DEFINE_SPINLOCK(lock_Z2);
 | |
| 
 | |
| static DEFINE_RWLOCK(rwlock_X1);
 | |
| static DEFINE_RWLOCK(rwlock_X2);
 | |
| static DEFINE_RWLOCK(rwlock_Y1);
 | |
| static DEFINE_RWLOCK(rwlock_Y2);
 | |
| static DEFINE_RWLOCK(rwlock_Z1);
 | |
| static DEFINE_RWLOCK(rwlock_Z2);
 | |
| 
 | |
| static DEFINE_MUTEX(mutex_X1);
 | |
| static DEFINE_MUTEX(mutex_X2);
 | |
| static DEFINE_MUTEX(mutex_Y1);
 | |
| static DEFINE_MUTEX(mutex_Y2);
 | |
| static DEFINE_MUTEX(mutex_Z1);
 | |
| static DEFINE_MUTEX(mutex_Z2);
 | |
| 
 | |
| static DECLARE_RWSEM(rwsem_X1);
 | |
| static DECLARE_RWSEM(rwsem_X2);
 | |
| static DECLARE_RWSEM(rwsem_Y1);
 | |
| static DECLARE_RWSEM(rwsem_Y2);
 | |
| static DECLARE_RWSEM(rwsem_Z1);
 | |
| static DECLARE_RWSEM(rwsem_Z2);
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| 
 | |
| static DEFINE_RT_MUTEX(rtmutex_X1);
 | |
| static DEFINE_RT_MUTEX(rtmutex_X2);
 | |
| static DEFINE_RT_MUTEX(rtmutex_Y1);
 | |
| static DEFINE_RT_MUTEX(rtmutex_Y2);
 | |
| static DEFINE_RT_MUTEX(rtmutex_Z1);
 | |
| static DEFINE_RT_MUTEX(rtmutex_Z2);
 | |
| 
 | |
| #endif
 | |
| 
 | |
| static DEFINE_PER_CPU(local_lock_t, local_A);
 | |
| 
 | |
| /*
 | |
|  * non-inlined runtime initializers, to let separate locks share
 | |
|  * the same lock-class:
 | |
|  */
 | |
| #define INIT_CLASS_FUNC(class) 				\
 | |
| static noinline void					\
 | |
| init_class_##class(spinlock_t *lock, rwlock_t *rwlock, \
 | |
| 	struct mutex *mutex, struct rw_semaphore *rwsem)\
 | |
| {							\
 | |
| 	spin_lock_init(lock);			\
 | |
| 	rwlock_init(rwlock);				\
 | |
| 	mutex_init(mutex);				\
 | |
| 	init_rwsem(rwsem);				\
 | |
| }
 | |
| 
 | |
| INIT_CLASS_FUNC(X)
 | |
| INIT_CLASS_FUNC(Y)
 | |
| INIT_CLASS_FUNC(Z)
 | |
| 
 | |
| static void init_shared_classes(void)
 | |
| {
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| 	static struct lock_class_key rt_X, rt_Y, rt_Z;
 | |
| 
 | |
| 	__rt_mutex_init(&rtmutex_X1, __func__, &rt_X);
 | |
| 	__rt_mutex_init(&rtmutex_X2, __func__, &rt_X);
 | |
| 	__rt_mutex_init(&rtmutex_Y1, __func__, &rt_Y);
 | |
| 	__rt_mutex_init(&rtmutex_Y2, __func__, &rt_Y);
 | |
| 	__rt_mutex_init(&rtmutex_Z1, __func__, &rt_Z);
 | |
| 	__rt_mutex_init(&rtmutex_Z2, __func__, &rt_Z);
 | |
| #endif
 | |
| 
 | |
| 	init_class_X(&lock_X1, &rwlock_X1, &mutex_X1, &rwsem_X1);
 | |
| 	init_class_X(&lock_X2, &rwlock_X2, &mutex_X2, &rwsem_X2);
 | |
| 
 | |
| 	init_class_Y(&lock_Y1, &rwlock_Y1, &mutex_Y1, &rwsem_Y1);
 | |
| 	init_class_Y(&lock_Y2, &rwlock_Y2, &mutex_Y2, &rwsem_Y2);
 | |
| 
 | |
| 	init_class_Z(&lock_Z1, &rwlock_Z1, &mutex_Z1, &rwsem_Z1);
 | |
| 	init_class_Z(&lock_Z2, &rwlock_Z2, &mutex_Z2, &rwsem_Z2);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * For spinlocks and rwlocks we also do hardirq-safe / softirq-safe tests.
 | |
|  * The following functions use a lock from a simulated hardirq/softirq
 | |
|  * context, causing the locks to be marked as hardirq-safe/softirq-safe:
 | |
|  */
 | |
| 
 | |
| #define HARDIRQ_DISABLE		local_irq_disable
 | |
| #define HARDIRQ_ENABLE		local_irq_enable
 | |
| 
 | |
| #define HARDIRQ_ENTER()				\
 | |
| 	local_irq_disable();			\
 | |
| 	__irq_enter();				\
 | |
| 	lockdep_hardirq_threaded();		\
 | |
| 	WARN_ON(!in_irq());
 | |
| 
 | |
| #define HARDIRQ_EXIT()				\
 | |
| 	__irq_exit();				\
 | |
| 	local_irq_enable();
 | |
| 
 | |
| #define SOFTIRQ_DISABLE		local_bh_disable
 | |
| #define SOFTIRQ_ENABLE		local_bh_enable
 | |
| 
 | |
| #define SOFTIRQ_ENTER()				\
 | |
| 		local_bh_disable();		\
 | |
| 		local_irq_disable();		\
 | |
| 		lockdep_softirq_enter();	\
 | |
| 		WARN_ON(!in_softirq());
 | |
| 
 | |
| #define SOFTIRQ_EXIT()				\
 | |
| 		lockdep_softirq_exit();		\
 | |
| 		local_irq_enable();		\
 | |
| 		local_bh_enable();
 | |
| 
 | |
| /*
 | |
|  * Shortcuts for lock/unlock API variants, to keep
 | |
|  * the testcases compact:
 | |
|  */
 | |
| #define L(x)			spin_lock(&lock_##x)
 | |
| #define U(x)			spin_unlock(&lock_##x)
 | |
| #define LU(x)			L(x); U(x)
 | |
| #define SI(x)			spin_lock_init(&lock_##x)
 | |
| 
 | |
| #define WL(x)			write_lock(&rwlock_##x)
 | |
| #define WU(x)			write_unlock(&rwlock_##x)
 | |
| #define WLU(x)			WL(x); WU(x)
 | |
| 
 | |
| #define RL(x)			read_lock(&rwlock_##x)
 | |
| #define RU(x)			read_unlock(&rwlock_##x)
 | |
| #define RLU(x)			RL(x); RU(x)
 | |
| #define RWI(x)			rwlock_init(&rwlock_##x)
 | |
| 
 | |
| #define ML(x)			mutex_lock(&mutex_##x)
 | |
| #define MU(x)			mutex_unlock(&mutex_##x)
 | |
| #define MI(x)			mutex_init(&mutex_##x)
 | |
| 
 | |
| #define RTL(x)			rt_mutex_lock(&rtmutex_##x)
 | |
| #define RTU(x)			rt_mutex_unlock(&rtmutex_##x)
 | |
| #define RTI(x)			rt_mutex_init(&rtmutex_##x)
 | |
| 
 | |
| #define WSL(x)			down_write(&rwsem_##x)
 | |
| #define WSU(x)			up_write(&rwsem_##x)
 | |
| 
 | |
| #define RSL(x)			down_read(&rwsem_##x)
 | |
| #define RSU(x)			up_read(&rwsem_##x)
 | |
| #define RWSI(x)			init_rwsem(&rwsem_##x)
 | |
| 
 | |
| #ifndef CONFIG_DEBUG_WW_MUTEX_SLOWPATH
 | |
| #define WWAI(x)			ww_acquire_init(x, &ww_lockdep)
 | |
| #else
 | |
| #define WWAI(x)			do { ww_acquire_init(x, &ww_lockdep); (x)->deadlock_inject_countdown = ~0U; } while (0)
 | |
| #endif
 | |
| #define WWAD(x)			ww_acquire_done(x)
 | |
| #define WWAF(x)			ww_acquire_fini(x)
 | |
| 
 | |
| #define WWL(x, c)		ww_mutex_lock(x, c)
 | |
| #define WWT(x)			ww_mutex_trylock(x, NULL)
 | |
| #define WWL1(x)			ww_mutex_lock(x, NULL)
 | |
| #define WWU(x)			ww_mutex_unlock(x)
 | |
| 
 | |
| 
 | |
| #define LOCK_UNLOCK_2(x,y)	LOCK(x); LOCK(y); UNLOCK(y); UNLOCK(x)
 | |
| 
 | |
| /*
 | |
|  * Generate different permutations of the same testcase, using
 | |
|  * the same basic lock-dependency/state events:
 | |
|  */
 | |
| 
 | |
| #define GENERATE_TESTCASE(name)			\
 | |
| 						\
 | |
| static void name(void) { E(); }
 | |
| 
 | |
| #define GENERATE_PERMUTATIONS_2_EVENTS(name)	\
 | |
| 						\
 | |
| static void name##_12(void) { E1(); E2(); }	\
 | |
| static void name##_21(void) { E2(); E1(); }
 | |
| 
 | |
| #define GENERATE_PERMUTATIONS_3_EVENTS(name)		\
 | |
| 							\
 | |
| static void name##_123(void) { E1(); E2(); E3(); }	\
 | |
| static void name##_132(void) { E1(); E3(); E2(); }	\
 | |
| static void name##_213(void) { E2(); E1(); E3(); }	\
 | |
| static void name##_231(void) { E2(); E3(); E1(); }	\
 | |
| static void name##_312(void) { E3(); E1(); E2(); }	\
 | |
| static void name##_321(void) { E3(); E2(); E1(); }
 | |
| 
 | |
| /*
 | |
|  * AA deadlock:
 | |
|  */
 | |
| 
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK(X1);				\
 | |
| 	LOCK(X2); /* this one should fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(AA_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(AA_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(AA_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(AA_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(AA_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(AA_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(AA_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * Special-case for read-locking, they are
 | |
|  * allowed to recurse on the same lock class:
 | |
|  */
 | |
| static void rlock_AA1(void)
 | |
| {
 | |
| 	RL(X1);
 | |
| 	RL(X1); // this one should NOT fail
 | |
| }
 | |
| 
 | |
| static void rlock_AA1B(void)
 | |
| {
 | |
| 	RL(X1);
 | |
| 	RL(X2); // this one should NOT fail
 | |
| }
 | |
| 
 | |
| static void rsem_AA1(void)
 | |
| {
 | |
| 	RSL(X1);
 | |
| 	RSL(X1); // this one should fail
 | |
| }
 | |
| 
 | |
| static void rsem_AA1B(void)
 | |
| {
 | |
| 	RSL(X1);
 | |
| 	RSL(X2); // this one should fail
 | |
| }
 | |
| /*
 | |
|  * The mixing of read and write locks is not allowed:
 | |
|  */
 | |
| static void rlock_AA2(void)
 | |
| {
 | |
| 	RL(X1);
 | |
| 	WL(X2); // this one should fail
 | |
| }
 | |
| 
 | |
| static void rsem_AA2(void)
 | |
| {
 | |
| 	RSL(X1);
 | |
| 	WSL(X2); // this one should fail
 | |
| }
 | |
| 
 | |
| static void rlock_AA3(void)
 | |
| {
 | |
| 	WL(X1);
 | |
| 	RL(X2); // this one should fail
 | |
| }
 | |
| 
 | |
| static void rsem_AA3(void)
 | |
| {
 | |
| 	WSL(X1);
 | |
| 	RSL(X2); // this one should fail
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * read_lock(A)
 | |
|  * spin_lock(B)
 | |
|  *		spin_lock(B)
 | |
|  *		write_lock(A)
 | |
|  */
 | |
| static void rlock_ABBA1(void)
 | |
| {
 | |
| 	RL(X1);
 | |
| 	L(Y1);
 | |
| 	U(Y1);
 | |
| 	RU(X1);
 | |
| 
 | |
| 	L(Y1);
 | |
| 	WL(X1);
 | |
| 	WU(X1);
 | |
| 	U(Y1); // should fail
 | |
| }
 | |
| 
 | |
| static void rwsem_ABBA1(void)
 | |
| {
 | |
| 	RSL(X1);
 | |
| 	ML(Y1);
 | |
| 	MU(Y1);
 | |
| 	RSU(X1);
 | |
| 
 | |
| 	ML(Y1);
 | |
| 	WSL(X1);
 | |
| 	WSU(X1);
 | |
| 	MU(Y1); // should fail
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * read_lock(A)
 | |
|  * spin_lock(B)
 | |
|  *		spin_lock(B)
 | |
|  *		write_lock(A)
 | |
|  *
 | |
|  * This test case is aimed at poking whether the chain cache prevents us from
 | |
|  * detecting a read-lock/lock-write deadlock: if the chain cache doesn't differ
 | |
|  * read/write locks, the following case may happen
 | |
|  *
 | |
|  * 	{ read_lock(A)->lock(B) dependency exists }
 | |
|  *
 | |
|  * 	P0:
 | |
|  * 	lock(B);
 | |
|  * 	read_lock(A);
 | |
|  *
 | |
|  *	{ Not a deadlock, B -> A is added in the chain cache }
 | |
|  *
 | |
|  *	P1:
 | |
|  *	lock(B);
 | |
|  *	write_lock(A);
 | |
|  *
 | |
|  *	{ B->A found in chain cache, not reported as a deadlock }
 | |
|  *
 | |
|  */
 | |
| static void rlock_chaincache_ABBA1(void)
 | |
| {
 | |
| 	RL(X1);
 | |
| 	L(Y1);
 | |
| 	U(Y1);
 | |
| 	RU(X1);
 | |
| 
 | |
| 	L(Y1);
 | |
| 	RL(X1);
 | |
| 	RU(X1);
 | |
| 	U(Y1);
 | |
| 
 | |
| 	L(Y1);
 | |
| 	WL(X1);
 | |
| 	WU(X1);
 | |
| 	U(Y1); // should fail
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * read_lock(A)
 | |
|  * spin_lock(B)
 | |
|  *		spin_lock(B)
 | |
|  *		read_lock(A)
 | |
|  */
 | |
| static void rlock_ABBA2(void)
 | |
| {
 | |
| 	RL(X1);
 | |
| 	L(Y1);
 | |
| 	U(Y1);
 | |
| 	RU(X1);
 | |
| 
 | |
| 	L(Y1);
 | |
| 	RL(X1);
 | |
| 	RU(X1);
 | |
| 	U(Y1); // should NOT fail
 | |
| }
 | |
| 
 | |
| static void rwsem_ABBA2(void)
 | |
| {
 | |
| 	RSL(X1);
 | |
| 	ML(Y1);
 | |
| 	MU(Y1);
 | |
| 	RSU(X1);
 | |
| 
 | |
| 	ML(Y1);
 | |
| 	RSL(X1);
 | |
| 	RSU(X1);
 | |
| 	MU(Y1); // should fail
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|  * write_lock(A)
 | |
|  * spin_lock(B)
 | |
|  *		spin_lock(B)
 | |
|  *		write_lock(A)
 | |
|  */
 | |
| static void rlock_ABBA3(void)
 | |
| {
 | |
| 	WL(X1);
 | |
| 	L(Y1);
 | |
| 	U(Y1);
 | |
| 	WU(X1);
 | |
| 
 | |
| 	L(Y1);
 | |
| 	WL(X1);
 | |
| 	WU(X1);
 | |
| 	U(Y1); // should fail
 | |
| }
 | |
| 
 | |
| static void rwsem_ABBA3(void)
 | |
| {
 | |
| 	WSL(X1);
 | |
| 	ML(Y1);
 | |
| 	MU(Y1);
 | |
| 	WSU(X1);
 | |
| 
 | |
| 	ML(Y1);
 | |
| 	WSL(X1);
 | |
| 	WSU(X1);
 | |
| 	MU(Y1); // should fail
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * ABBA deadlock:
 | |
|  */
 | |
| 
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK_UNLOCK_2(A, B);			\
 | |
| 	LOCK_UNLOCK_2(B, A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(ABBA_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(ABBA_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(ABBA_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(ABBA_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(ABBA_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(ABBA_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(ABBA_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * AB BC CA deadlock:
 | |
|  */
 | |
| 
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK_UNLOCK_2(A, B);			\
 | |
| 	LOCK_UNLOCK_2(B, C);			\
 | |
| 	LOCK_UNLOCK_2(C, A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(ABBCCA_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(ABBCCA_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(ABBCCA_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(ABBCCA_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(ABBCCA_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(ABBCCA_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(ABBCCA_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * AB CA BC deadlock:
 | |
|  */
 | |
| 
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK_UNLOCK_2(A, B);			\
 | |
| 	LOCK_UNLOCK_2(C, A);			\
 | |
| 	LOCK_UNLOCK_2(B, C); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(ABCABC_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(ABCABC_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(ABCABC_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(ABCABC_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(ABCABC_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(ABCABC_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(ABCABC_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * AB BC CD DA deadlock:
 | |
|  */
 | |
| 
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK_UNLOCK_2(A, B);			\
 | |
| 	LOCK_UNLOCK_2(B, C);			\
 | |
| 	LOCK_UNLOCK_2(C, D);			\
 | |
| 	LOCK_UNLOCK_2(D, A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(ABBCCDDA_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * AB CD BD DA deadlock:
 | |
|  */
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK_UNLOCK_2(A, B);			\
 | |
| 	LOCK_UNLOCK_2(C, D);			\
 | |
| 	LOCK_UNLOCK_2(B, D);			\
 | |
| 	LOCK_UNLOCK_2(D, A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(ABCDBDDA_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * AB CD BC DA deadlock:
 | |
|  */
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK_UNLOCK_2(A, B);			\
 | |
| 	LOCK_UNLOCK_2(C, D);			\
 | |
| 	LOCK_UNLOCK_2(B, C);			\
 | |
| 	LOCK_UNLOCK_2(D, A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(ABCDBCDA_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| #ifdef CONFIG_PREEMPT_RT
 | |
| # define RT_PREPARE_DBL_UNLOCK()	{ migrate_disable(); rcu_read_lock(); }
 | |
| #else
 | |
| # define RT_PREPARE_DBL_UNLOCK()
 | |
| #endif
 | |
| /*
 | |
|  * Double unlock:
 | |
|  */
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK(A);				\
 | |
| 	RT_PREPARE_DBL_UNLOCK();		\
 | |
| 	UNLOCK(A);				\
 | |
| 	UNLOCK(A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(double_unlock_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(double_unlock_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(double_unlock_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(double_unlock_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(double_unlock_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(double_unlock_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(double_unlock_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * initializing a held lock:
 | |
|  */
 | |
| #define E()					\
 | |
| 						\
 | |
| 	LOCK(A);				\
 | |
| 	INIT(A); /* fail */
 | |
| 
 | |
| /*
 | |
|  * 6 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_TESTCASE(init_held_spin)
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_TESTCASE(init_held_wlock)
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_TESTCASE(init_held_rlock)
 | |
| #include "locking-selftest-mutex.h"
 | |
| GENERATE_TESTCASE(init_held_mutex)
 | |
| #include "locking-selftest-wsem.h"
 | |
| GENERATE_TESTCASE(init_held_wsem)
 | |
| #include "locking-selftest-rsem.h"
 | |
| GENERATE_TESTCASE(init_held_rsem)
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #include "locking-selftest-rtmutex.h"
 | |
| GENERATE_TESTCASE(init_held_rtmutex);
 | |
| #endif
 | |
| 
 | |
| #undef E
 | |
| 
 | |
| /*
 | |
|  * locking an irq-safe lock with irqs enabled:
 | |
|  */
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);
 | |
| 
 | |
| /*
 | |
|  * Generate 24 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_hard_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-spin-softirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_soft_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe1_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| /*
 | |
|  * Enabling hardirqs with a softirq-safe lock held:
 | |
|  */
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	SOFTIRQ_ENTER();		\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	SOFTIRQ_EXIT();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	HARDIRQ_DISABLE();		\
 | |
| 	LOCK(A);			\
 | |
| 	HARDIRQ_ENABLE();		\
 | |
| 	UNLOCK(A);
 | |
| 
 | |
| /*
 | |
|  * Generate 12 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2A_spin)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2A_wlock)
 | |
| 
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2A_rlock)
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| 
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * Enabling irqs with an irq-safe lock held:
 | |
|  */
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	IRQ_DISABLE();			\
 | |
| 	LOCK(A);			\
 | |
| 	IRQ_ENABLE();			\
 | |
| 	UNLOCK(A);
 | |
| 
 | |
| /*
 | |
|  * Generate 24 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_hard_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-spin-softirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_soft_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_2_EVENTS(irqsafe2B_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| 
 | |
| /*
 | |
|  * Acquiring a irq-unsafe lock while holding an irq-safe-lock:
 | |
|  */
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	LOCK(A);			\
 | |
| 	LOCK(B);			\
 | |
| 	UNLOCK(B);			\
 | |
| 	UNLOCK(A);			\
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	LOCK(B);			\
 | |
| 	UNLOCK(B);
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| /*
 | |
|  * Generate 36 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_hard_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-spin-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_soft_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe3_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * If a lock turns into softirq-safe, but earlier it took
 | |
|  * a softirq-unsafe lock:
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 	IRQ_DISABLE();			\
 | |
| 	LOCK(A);			\
 | |
| 	LOCK(B);			\
 | |
| 	UNLOCK(B);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	IRQ_ENABLE();
 | |
| 
 | |
| #define E2()				\
 | |
| 	LOCK(B);			\
 | |
| 	UNLOCK(B);
 | |
| 
 | |
| #define E3()				\
 | |
| 	IRQ_ENTER();			\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| /*
 | |
|  * Generate 36 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_hard_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-spin-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_soft_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irqsafe4_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * read-lock / write-lock irq inversion.
 | |
|  *
 | |
|  * Deadlock scenario:
 | |
|  *
 | |
|  * CPU#1 is at #1, i.e. it has write-locked A, but has not
 | |
|  * taken B yet.
 | |
|  *
 | |
|  * CPU#2 is at #2, i.e. it has locked B.
 | |
|  *
 | |
|  * Hardirq hits CPU#2 at point #2 and is trying to read-lock A.
 | |
|  *
 | |
|  * The deadlock occurs because CPU#1 will spin on B, and CPU#2
 | |
|  * will spin on A.
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	IRQ_DISABLE();			\
 | |
| 	WL(A);				\
 | |
| 	LOCK(B);			\
 | |
| 	UNLOCK(B);			\
 | |
| 	WU(A);				\
 | |
| 	IRQ_ENABLE();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	LOCK(B);			\
 | |
| 	UNLOCK(B);
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	RL(A);				\
 | |
| 	RU(A);				\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| /*
 | |
|  * Generate 36 testcases:
 | |
|  */
 | |
| #include "locking-selftest-spin-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_hard_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-hardirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-spin-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_soft_spin)
 | |
| 
 | |
| #include "locking-selftest-rlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock-softirq.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_inversion_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * write-read / write-read / write-read deadlock even if read is recursive
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	WL(X1);				\
 | |
| 	RL(Y1);				\
 | |
| 	RU(Y1);				\
 | |
| 	WU(X1);
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	WL(Y1);				\
 | |
| 	RL(Z1);				\
 | |
| 	RU(Z1);				\
 | |
| 	WU(Y1);
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	WL(Z1);				\
 | |
| 	RL(X1);				\
 | |
| 	RU(X1);				\
 | |
| 	WU(Z1);
 | |
| 
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(W1R2_W2R3_W3R1)
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * write-write / read-read / write-read deadlock even if read is recursive
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	WL(X1);				\
 | |
| 	WL(Y1);				\
 | |
| 	WU(Y1);				\
 | |
| 	WU(X1);
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	RL(Y1);				\
 | |
| 	RL(Z1);				\
 | |
| 	RU(Z1);				\
 | |
| 	RU(Y1);
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	WL(Z1);				\
 | |
| 	RL(X1);				\
 | |
| 	RU(X1);				\
 | |
| 	WU(Z1);
 | |
| 
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(W1W2_R2R3_W3R1)
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * write-write / read-read / read-write is not deadlock when read is recursive
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	WL(X1);				\
 | |
| 	WL(Y1);				\
 | |
| 	WU(Y1);				\
 | |
| 	WU(X1);
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	RL(Y1);				\
 | |
| 	RL(Z1);				\
 | |
| 	RU(Z1);				\
 | |
| 	RU(Y1);
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	RL(Z1);				\
 | |
| 	WL(X1);				\
 | |
| 	WU(X1);				\
 | |
| 	RU(Z1);
 | |
| 
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(W1R2_R2R3_W3W1)
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * write-read / read-read / write-write is not deadlock when read is recursive
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	WL(X1);				\
 | |
| 	RL(Y1);				\
 | |
| 	RU(Y1);				\
 | |
| 	WU(X1);
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	RL(Y1);				\
 | |
| 	RL(Z1);				\
 | |
| 	RU(Z1);				\
 | |
| 	RU(Y1);
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	WL(Z1);				\
 | |
| 	WL(X1);				\
 | |
| 	WU(X1);				\
 | |
| 	WU(Z1);
 | |
| 
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(W1W2_R2R3_R3W1)
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| /*
 | |
|  * read-lock / write-lock recursion that is actually safe.
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	IRQ_DISABLE();			\
 | |
| 	WL(A);				\
 | |
| 	WU(A);				\
 | |
| 	IRQ_ENABLE();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	RL(A);				\
 | |
| 	RU(A);				\
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	LOCK(A);			\
 | |
| 	L(B);				\
 | |
| 	U(B);				\
 | |
| 	UNLOCK(A);			\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| /*
 | |
|  * Generate 24 testcases:
 | |
|  */
 | |
| #include "locking-selftest-hardirq.h"
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-softirq.h"
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| 
 | |
| /*
 | |
|  * read-lock / write-lock recursion that is unsafe.
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	IRQ_DISABLE();			\
 | |
| 	L(B);				\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	U(B);				\
 | |
| 	IRQ_ENABLE();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	RL(A);				\
 | |
| 	RU(A);				\
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	L(B);				\
 | |
| 	U(B);				\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| /*
 | |
|  * Generate 24 testcases:
 | |
|  */
 | |
| #include "locking-selftest-hardirq.h"
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-softirq.h"
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion2_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #undef E1
 | |
| #undef E2
 | |
| #undef E3
 | |
| /*
 | |
|  * read-lock / write-lock recursion that is unsafe.
 | |
|  *
 | |
|  * A is a ENABLED_*_READ lock
 | |
|  * B is a USED_IN_*_READ lock
 | |
|  *
 | |
|  * read_lock(A);
 | |
|  *			write_lock(B);
 | |
|  * <interrupt>
 | |
|  * read_lock(B);
 | |
|  * 			write_lock(A); // if this one is read_lock(), no deadlock
 | |
|  */
 | |
| 
 | |
| #define E1()				\
 | |
| 					\
 | |
| 	IRQ_DISABLE();			\
 | |
| 	WL(B);				\
 | |
| 	LOCK(A);			\
 | |
| 	UNLOCK(A);			\
 | |
| 	WU(B);				\
 | |
| 	IRQ_ENABLE();
 | |
| 
 | |
| #define E2()				\
 | |
| 					\
 | |
| 	RL(A);				\
 | |
| 	RU(A);				\
 | |
| 
 | |
| #define E3()				\
 | |
| 					\
 | |
| 	IRQ_ENTER();			\
 | |
| 	RL(B);				\
 | |
| 	RU(B);				\
 | |
| 	IRQ_EXIT();
 | |
| 
 | |
| /*
 | |
|  * Generate 24 testcases:
 | |
|  */
 | |
| #include "locking-selftest-hardirq.h"
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_hard_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_hard_wlock)
 | |
| 
 | |
| #ifndef CONFIG_PREEMPT_RT
 | |
| #include "locking-selftest-softirq.h"
 | |
| #include "locking-selftest-rlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_soft_rlock)
 | |
| 
 | |
| #include "locking-selftest-wlock.h"
 | |
| GENERATE_PERMUTATIONS_3_EVENTS(irq_read_recursion3_soft_wlock)
 | |
| #endif
 | |
| 
 | |
| #ifdef CONFIG_DEBUG_LOCK_ALLOC
 | |
| # define I_SPINLOCK(x)	lockdep_reset_lock(&lock_##x.dep_map)
 | |
| # define I_RAW_SPINLOCK(x)	lockdep_reset_lock(&raw_lock_##x.dep_map)
 | |
| # define I_RWLOCK(x)	lockdep_reset_lock(&rwlock_##x.dep_map)
 | |
| # define I_MUTEX(x)	lockdep_reset_lock(&mutex_##x.dep_map)
 | |
| # define I_RWSEM(x)	lockdep_reset_lock(&rwsem_##x.dep_map)
 | |
| # define I_WW(x)	lockdep_reset_lock(&x.dep_map)
 | |
| # define I_LOCAL_LOCK(x) lockdep_reset_lock(this_cpu_ptr(&local_##x.dep_map))
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| # define I_RTMUTEX(x)	lockdep_reset_lock(&rtmutex_##x.dep_map)
 | |
| #endif
 | |
| #else
 | |
| # define I_SPINLOCK(x)
 | |
| # define I_RAW_SPINLOCK(x)
 | |
| # define I_RWLOCK(x)
 | |
| # define I_MUTEX(x)
 | |
| # define I_RWSEM(x)
 | |
| # define I_WW(x)
 | |
| # define I_LOCAL_LOCK(x)
 | |
| #endif
 | |
| 
 | |
| #ifndef I_RTMUTEX
 | |
| # define I_RTMUTEX(x)
 | |
| #endif
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #define I2_RTMUTEX(x)	rt_mutex_init(&rtmutex_##x)
 | |
| #else
 | |
| #define I2_RTMUTEX(x)
 | |
| #endif
 | |
| 
 | |
| #define I1(x)					\
 | |
| 	do {					\
 | |
| 		I_SPINLOCK(x);			\
 | |
| 		I_RWLOCK(x);			\
 | |
| 		I_MUTEX(x);			\
 | |
| 		I_RWSEM(x);			\
 | |
| 		I_RTMUTEX(x);			\
 | |
| 	} while (0)
 | |
| 
 | |
| #define I2(x)					\
 | |
| 	do {					\
 | |
| 		spin_lock_init(&lock_##x);	\
 | |
| 		rwlock_init(&rwlock_##x);	\
 | |
| 		mutex_init(&mutex_##x);		\
 | |
| 		init_rwsem(&rwsem_##x);		\
 | |
| 		I2_RTMUTEX(x);			\
 | |
| 	} while (0)
 | |
| 
 | |
| static void reset_locks(void)
 | |
| {
 | |
| 	local_irq_disable();
 | |
| 	lockdep_free_key_range(&ww_lockdep.acquire_key, 1);
 | |
| 	lockdep_free_key_range(&ww_lockdep.mutex_key, 1);
 | |
| 
 | |
| 	I1(A); I1(B); I1(C); I1(D);
 | |
| 	I1(X1); I1(X2); I1(Y1); I1(Y2); I1(Z1); I1(Z2);
 | |
| 	I_WW(t); I_WW(t2); I_WW(o.base); I_WW(o2.base); I_WW(o3.base);
 | |
| 	I_RAW_SPINLOCK(A); I_RAW_SPINLOCK(B);
 | |
| 	I_LOCAL_LOCK(A);
 | |
| 
 | |
| 	lockdep_reset();
 | |
| 
 | |
| 	I2(A); I2(B); I2(C); I2(D);
 | |
| 	init_shared_classes();
 | |
| 	raw_spin_lock_init(&raw_lock_A);
 | |
| 	raw_spin_lock_init(&raw_lock_B);
 | |
| 	local_lock_init(this_cpu_ptr(&local_A));
 | |
| 
 | |
| 	ww_mutex_init(&o, &ww_lockdep); ww_mutex_init(&o2, &ww_lockdep); ww_mutex_init(&o3, &ww_lockdep);
 | |
| 	memset(&t, 0, sizeof(t)); memset(&t2, 0, sizeof(t2));
 | |
| 	memset(&ww_lockdep.acquire_key, 0, sizeof(ww_lockdep.acquire_key));
 | |
| 	memset(&ww_lockdep.mutex_key, 0, sizeof(ww_lockdep.mutex_key));
 | |
| 	local_irq_enable();
 | |
| }
 | |
| 
 | |
| #undef I
 | |
| 
 | |
| static int testcase_total;
 | |
| static int testcase_successes;
 | |
| static int expected_testcase_failures;
 | |
| static int unexpected_testcase_failures;
 | |
| 
 | |
| static void dotest(void (*testcase_fn)(void), int expected, int lockclass_mask)
 | |
| {
 | |
| 	int saved_preempt_count = preempt_count();
 | |
| #ifdef CONFIG_PREEMPT_RT
 | |
| #ifdef CONFIG_SMP
 | |
| 	int saved_mgd_count = current->migration_disabled;
 | |
| #endif
 | |
| 	int saved_rcu_count = current->rcu_read_lock_nesting;
 | |
| #endif
 | |
| 
 | |
| 	WARN_ON(irqs_disabled());
 | |
| 
 | |
| 	debug_locks_silent = !(debug_locks_verbose & lockclass_mask);
 | |
| 
 | |
| 	testcase_fn();
 | |
| 	/*
 | |
| 	 * Filter out expected failures:
 | |
| 	 */
 | |
| #ifndef CONFIG_PROVE_LOCKING
 | |
| 	if (expected == FAILURE && debug_locks) {
 | |
| 		expected_testcase_failures++;
 | |
| 		pr_cont("failed|");
 | |
| 	}
 | |
| 	else
 | |
| #endif
 | |
| 	if (debug_locks != expected) {
 | |
| 		unexpected_testcase_failures++;
 | |
| 		pr_cont("FAILED|");
 | |
| 	} else {
 | |
| 		testcase_successes++;
 | |
| 		pr_cont("  ok  |");
 | |
| 	}
 | |
| 	testcase_total++;
 | |
| 
 | |
| 	if (debug_locks_verbose & lockclass_mask)
 | |
| 		pr_cont(" lockclass mask: %x, debug_locks: %d, expected: %d\n",
 | |
| 			lockclass_mask, debug_locks, expected);
 | |
| 	/*
 | |
| 	 * Some tests (e.g. double-unlock) might corrupt the preemption
 | |
| 	 * count, so restore it:
 | |
| 	 */
 | |
| 	preempt_count_set(saved_preempt_count);
 | |
| 
 | |
| #ifdef CONFIG_PREEMPT_RT
 | |
| #ifdef CONFIG_SMP
 | |
| 	while (current->migration_disabled > saved_mgd_count)
 | |
| 		migrate_enable();
 | |
| #endif
 | |
| 
 | |
| 	while (current->rcu_read_lock_nesting > saved_rcu_count)
 | |
| 		rcu_read_unlock();
 | |
| 	WARN_ON_ONCE(current->rcu_read_lock_nesting < saved_rcu_count);
 | |
| #endif
 | |
| 
 | |
| #ifdef CONFIG_TRACE_IRQFLAGS
 | |
| 	if (softirq_count())
 | |
| 		current->softirqs_enabled = 0;
 | |
| 	else
 | |
| 		current->softirqs_enabled = 1;
 | |
| #endif
 | |
| 
 | |
| 	reset_locks();
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_RT_MUTEXES
 | |
| #define dotest_rt(fn, e, m)	dotest((fn), (e), (m))
 | |
| #else
 | |
| #define dotest_rt(fn, e, m)
 | |
| #endif
 | |
| 
 | |
| static inline void print_testname(const char *testname)
 | |
| {
 | |
| 	printk("%33s:", testname);
 | |
| }
 | |
| 
 | |
| #define DO_TESTCASE_1(desc, name, nr)				\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_1B(desc, name, nr)				\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK);		\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_1RR(desc, name, nr)				\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	pr_cont("             |");				\
 | |
| 	dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_1RRB(desc, name, nr)			\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	pr_cont("             |");				\
 | |
| 	dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK);		\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 
 | |
| #define DO_TESTCASE_3(desc, name, nr)				\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	dotest(name##_spin_##nr, FAILURE, LOCKTYPE_SPIN);	\
 | |
| 	dotest(name##_wlock_##nr, FAILURE, LOCKTYPE_RWLOCK);	\
 | |
| 	dotest(name##_rlock_##nr, SUCCESS, LOCKTYPE_RWLOCK);	\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_3RW(desc, name, nr)				\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	dotest(name##_spin_##nr, FAILURE, LOCKTYPE_SPIN|LOCKTYPE_RWLOCK);\
 | |
| 	dotest(name##_wlock_##nr, FAILURE, LOCKTYPE_RWLOCK);	\
 | |
| 	dotest(name##_rlock_##nr, SUCCESS, LOCKTYPE_RWLOCK);	\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_2RW(desc, name, nr)				\
 | |
| 	print_testname(desc"/"#nr);				\
 | |
| 	pr_cont("      |");					\
 | |
| 	dotest(name##_wlock_##nr, FAILURE, LOCKTYPE_RWLOCK);	\
 | |
| 	dotest(name##_rlock_##nr, SUCCESS, LOCKTYPE_RWLOCK);	\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_2x2RW(desc, name, nr)			\
 | |
| 	DO_TESTCASE_2RW("hard-"desc, name##_hard, nr)		\
 | |
| 	NON_RT(DO_TESTCASE_2RW("soft-"desc, name##_soft, nr))	\
 | |
| 
 | |
| #define DO_TESTCASE_6x2x2RW(desc, name)				\
 | |
| 	DO_TESTCASE_2x2RW(desc, name, 123);			\
 | |
| 	DO_TESTCASE_2x2RW(desc, name, 132);			\
 | |
| 	DO_TESTCASE_2x2RW(desc, name, 213);			\
 | |
| 	DO_TESTCASE_2x2RW(desc, name, 231);			\
 | |
| 	DO_TESTCASE_2x2RW(desc, name, 312);			\
 | |
| 	DO_TESTCASE_2x2RW(desc, name, 321);
 | |
| 
 | |
| #define DO_TESTCASE_6(desc, name)				\
 | |
| 	print_testname(desc);					\
 | |
| 	dotest(name##_spin, FAILURE, LOCKTYPE_SPIN);		\
 | |
| 	dotest(name##_wlock, FAILURE, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(name##_rlock, FAILURE, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(name##_mutex, FAILURE, LOCKTYPE_MUTEX);		\
 | |
| 	dotest(name##_wsem, FAILURE, LOCKTYPE_RWSEM);		\
 | |
| 	dotest(name##_rsem, FAILURE, LOCKTYPE_RWSEM);		\
 | |
| 	dotest_rt(name##_rtmutex, FAILURE, LOCKTYPE_RTMUTEX);	\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_6_SUCCESS(desc, name)			\
 | |
| 	print_testname(desc);					\
 | |
| 	dotest(name##_spin, SUCCESS, LOCKTYPE_SPIN);		\
 | |
| 	dotest(name##_wlock, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(name##_rlock, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(name##_mutex, SUCCESS, LOCKTYPE_MUTEX);		\
 | |
| 	dotest(name##_wsem, SUCCESS, LOCKTYPE_RWSEM);		\
 | |
| 	dotest(name##_rsem, SUCCESS, LOCKTYPE_RWSEM);		\
 | |
| 	dotest_rt(name##_rtmutex, SUCCESS, LOCKTYPE_RTMUTEX);	\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| /*
 | |
|  * 'read' variant: rlocks must not trigger.
 | |
|  */
 | |
| #define DO_TESTCASE_6R(desc, name)				\
 | |
| 	print_testname(desc);					\
 | |
| 	dotest(name##_spin, FAILURE, LOCKTYPE_SPIN);		\
 | |
| 	dotest(name##_wlock, FAILURE, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(name##_rlock, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(name##_mutex, FAILURE, LOCKTYPE_MUTEX);		\
 | |
| 	dotest(name##_wsem, FAILURE, LOCKTYPE_RWSEM);		\
 | |
| 	dotest(name##_rsem, FAILURE, LOCKTYPE_RWSEM);		\
 | |
| 	dotest_rt(name##_rtmutex, FAILURE, LOCKTYPE_RTMUTEX);	\
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| #define DO_TESTCASE_2I(desc, name, nr)				\
 | |
| 	DO_TESTCASE_1("hard-"desc, name##_hard, nr);		\
 | |
| 	NON_RT(DO_TESTCASE_1("soft-"desc, name##_soft, nr));
 | |
| 
 | |
| #define DO_TESTCASE_2IB(desc, name, nr)				\
 | |
| 	DO_TESTCASE_1B("hard-"desc, name##_hard, nr);		\
 | |
| 	NON_RT(DO_TESTCASE_1B("soft-"desc, name##_soft, nr));
 | |
| 
 | |
| #define DO_TESTCASE_6I(desc, name, nr)				\
 | |
| 	DO_TESTCASE_3("hard-"desc, name##_hard, nr);		\
 | |
| 	NON_RT(DO_TESTCASE_3("soft-"desc, name##_soft, nr));
 | |
| 
 | |
| #define DO_TESTCASE_6IRW(desc, name, nr)			\
 | |
| 	DO_TESTCASE_3RW("hard-"desc, name##_hard, nr);		\
 | |
| 	NON_RT(DO_TESTCASE_3RW("soft-"desc, name##_soft, nr));
 | |
| 
 | |
| #define DO_TESTCASE_2x3(desc, name)				\
 | |
| 	DO_TESTCASE_3(desc, name, 12);				\
 | |
| 	DO_TESTCASE_3(desc, name, 21);
 | |
| 
 | |
| #define DO_TESTCASE_2x6(desc, name)				\
 | |
| 	DO_TESTCASE_6I(desc, name, 12);				\
 | |
| 	DO_TESTCASE_6I(desc, name, 21);
 | |
| 
 | |
| #define DO_TESTCASE_6x2(desc, name)				\
 | |
| 	DO_TESTCASE_2I(desc, name, 123);			\
 | |
| 	DO_TESTCASE_2I(desc, name, 132);			\
 | |
| 	DO_TESTCASE_2I(desc, name, 213);			\
 | |
| 	DO_TESTCASE_2I(desc, name, 231);			\
 | |
| 	DO_TESTCASE_2I(desc, name, 312);			\
 | |
| 	DO_TESTCASE_2I(desc, name, 321);
 | |
| 
 | |
| #define DO_TESTCASE_6x2B(desc, name)				\
 | |
| 	DO_TESTCASE_2IB(desc, name, 123);			\
 | |
| 	DO_TESTCASE_2IB(desc, name, 132);			\
 | |
| 	DO_TESTCASE_2IB(desc, name, 213);			\
 | |
| 	DO_TESTCASE_2IB(desc, name, 231);			\
 | |
| 	DO_TESTCASE_2IB(desc, name, 312);			\
 | |
| 	DO_TESTCASE_2IB(desc, name, 321);
 | |
| 
 | |
| #define DO_TESTCASE_6x1RR(desc, name)				\
 | |
| 	DO_TESTCASE_1RR(desc, name, 123);			\
 | |
| 	DO_TESTCASE_1RR(desc, name, 132);			\
 | |
| 	DO_TESTCASE_1RR(desc, name, 213);			\
 | |
| 	DO_TESTCASE_1RR(desc, name, 231);			\
 | |
| 	DO_TESTCASE_1RR(desc, name, 312);			\
 | |
| 	DO_TESTCASE_1RR(desc, name, 321);
 | |
| 
 | |
| #define DO_TESTCASE_6x1RRB(desc, name)				\
 | |
| 	DO_TESTCASE_1RRB(desc, name, 123);			\
 | |
| 	DO_TESTCASE_1RRB(desc, name, 132);			\
 | |
| 	DO_TESTCASE_1RRB(desc, name, 213);			\
 | |
| 	DO_TESTCASE_1RRB(desc, name, 231);			\
 | |
| 	DO_TESTCASE_1RRB(desc, name, 312);			\
 | |
| 	DO_TESTCASE_1RRB(desc, name, 321);
 | |
| 
 | |
| #define DO_TESTCASE_6x6(desc, name)				\
 | |
| 	DO_TESTCASE_6I(desc, name, 123);			\
 | |
| 	DO_TESTCASE_6I(desc, name, 132);			\
 | |
| 	DO_TESTCASE_6I(desc, name, 213);			\
 | |
| 	DO_TESTCASE_6I(desc, name, 231);			\
 | |
| 	DO_TESTCASE_6I(desc, name, 312);			\
 | |
| 	DO_TESTCASE_6I(desc, name, 321);
 | |
| 
 | |
| #define DO_TESTCASE_6x6RW(desc, name)				\
 | |
| 	DO_TESTCASE_6IRW(desc, name, 123);			\
 | |
| 	DO_TESTCASE_6IRW(desc, name, 132);			\
 | |
| 	DO_TESTCASE_6IRW(desc, name, 213);			\
 | |
| 	DO_TESTCASE_6IRW(desc, name, 231);			\
 | |
| 	DO_TESTCASE_6IRW(desc, name, 312);			\
 | |
| 	DO_TESTCASE_6IRW(desc, name, 321);
 | |
| 
 | |
| static void ww_test_fail_acquire(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t.stamp++;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 
 | |
| 	if (WARN_ON(!o.ctx) ||
 | |
| 	    WARN_ON(ret))
 | |
| 		return;
 | |
| 
 | |
| 	/* No lockdep test, pure API */
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret != -EALREADY);
 | |
| 
 | |
| 	ret = WWT(&o);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	t2 = t;
 | |
| 	t2.stamp++;
 | |
| 	ret = WWL(&o, &t2);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	if (WWT(&o))
 | |
| 		WWU(&o);
 | |
| #ifdef CONFIG_DEBUG_LOCK_ALLOC
 | |
| 	else
 | |
| 		DEBUG_LOCKS_WARN_ON(1);
 | |
| #endif
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_PREEMPT_RT
 | |
| #define ww_mutex_base_lock(b)			rt_mutex_lock(b)
 | |
| #define ww_mutex_base_trylock(b)		rt_mutex_trylock(b)
 | |
| #define ww_mutex_base_lock_nest_lock(b, b2)	rt_mutex_lock_nest_lock(b, b2)
 | |
| #define ww_mutex_base_lock_interruptible(b)	rt_mutex_lock_interruptible(b)
 | |
| #define ww_mutex_base_lock_killable(b)		rt_mutex_lock_killable(b)
 | |
| #define ww_mutex_base_unlock(b)			rt_mutex_unlock(b)
 | |
| #else
 | |
| #define ww_mutex_base_lock(b)			mutex_lock(b)
 | |
| #define ww_mutex_base_trylock(b)		mutex_trylock(b)
 | |
| #define ww_mutex_base_lock_nest_lock(b, b2)	mutex_lock_nest_lock(b, b2)
 | |
| #define ww_mutex_base_lock_interruptible(b)	mutex_lock_interruptible(b)
 | |
| #define ww_mutex_base_lock_killable(b)		mutex_lock_killable(b)
 | |
| #define ww_mutex_base_unlock(b)			mutex_unlock(b)
 | |
| #endif
 | |
| 
 | |
| static void ww_test_normal(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	/*
 | |
| 	 * None of the ww_mutex codepaths should be taken in the 'normal'
 | |
| 	 * mutex calls. The easiest way to verify this is by using the
 | |
| 	 * normal mutex calls, and making sure o.ctx is unmodified.
 | |
| 	 */
 | |
| 
 | |
| 	/* mutex_lock (and indirectly, mutex_lock_nested) */
 | |
| 	o.ctx = (void *)~0UL;
 | |
| 	ww_mutex_base_lock(&o.base);
 | |
| 	ww_mutex_base_unlock(&o.base);
 | |
| 	WARN_ON(o.ctx != (void *)~0UL);
 | |
| 
 | |
| 	/* mutex_lock_interruptible (and *_nested) */
 | |
| 	o.ctx = (void *)~0UL;
 | |
| 	ret = ww_mutex_base_lock_interruptible(&o.base);
 | |
| 	if (!ret)
 | |
| 		ww_mutex_base_unlock(&o.base);
 | |
| 	else
 | |
| 		WARN_ON(1);
 | |
| 	WARN_ON(o.ctx != (void *)~0UL);
 | |
| 
 | |
| 	/* mutex_lock_killable (and *_nested) */
 | |
| 	o.ctx = (void *)~0UL;
 | |
| 	ret = ww_mutex_base_lock_killable(&o.base);
 | |
| 	if (!ret)
 | |
| 		ww_mutex_base_unlock(&o.base);
 | |
| 	else
 | |
| 		WARN_ON(1);
 | |
| 	WARN_ON(o.ctx != (void *)~0UL);
 | |
| 
 | |
| 	/* trylock, succeeding */
 | |
| 	o.ctx = (void *)~0UL;
 | |
| 	ret = ww_mutex_base_trylock(&o.base);
 | |
| 	WARN_ON(!ret);
 | |
| 	if (ret)
 | |
| 		ww_mutex_base_unlock(&o.base);
 | |
| 	else
 | |
| 		WARN_ON(1);
 | |
| 	WARN_ON(o.ctx != (void *)~0UL);
 | |
| 
 | |
| 	/* trylock, failing */
 | |
| 	o.ctx = (void *)~0UL;
 | |
| 	ww_mutex_base_lock(&o.base);
 | |
| 	ret = ww_mutex_base_trylock(&o.base);
 | |
| 	WARN_ON(ret);
 | |
| 	ww_mutex_base_unlock(&o.base);
 | |
| 	WARN_ON(o.ctx != (void *)~0UL);
 | |
| 
 | |
| 	/* nest_lock */
 | |
| 	o.ctx = (void *)~0UL;
 | |
| 	ww_mutex_base_lock_nest_lock(&o.base, &t);
 | |
| 	ww_mutex_base_unlock(&o.base);
 | |
| 	WARN_ON(o.ctx != (void *)~0UL);
 | |
| }
 | |
| 
 | |
| static void ww_test_two_contexts(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	WWAI(&t2);
 | |
| }
 | |
| 
 | |
| static void ww_test_diff_class(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| #ifdef DEBUG_WW_MUTEXES
 | |
| 	t.ww_class = NULL;
 | |
| #endif
 | |
| 	WWL(&o, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_done_twice(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	WWAD(&t);
 | |
| 	WWAD(&t);
 | |
| 	WWAF(&t);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_unlock_twice(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	WWAD(&t);
 | |
| 	WWAF(&t);
 | |
| 	WWAF(&t);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_fini_early(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	WWL(&o, &t);
 | |
| 	WWAD(&t);
 | |
| 	WWAF(&t);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_lock_after_done(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	WWAD(&t);
 | |
| 	WWL(&o, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_object_unlock_twice(void)
 | |
| {
 | |
| 	WWL1(&o);
 | |
| 	WWU(&o);
 | |
| 	WWU(&o);
 | |
| }
 | |
| 
 | |
| static void ww_test_object_lock_unbalanced(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	WWL(&o, &t);
 | |
| 	t.acquired = 0;
 | |
| 	WWU(&o);
 | |
| 	WWAF(&t);
 | |
| }
 | |
| 
 | |
| static void ww_test_object_lock_stale_context(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 	o.ctx = &t2;
 | |
| 	WWL(&o, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_normal(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	o2.ctx = &t2;
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	o2.ctx = NULL;
 | |
| 	mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
 | |
| 	ww_mutex_base_unlock(&o2.base);
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	WWL(&o2, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_normal_slow(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	o2.ctx = NULL;
 | |
| 	mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
 | |
| 	ww_mutex_base_unlock(&o2.base);
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	ww_mutex_lock_slow(&o2, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_no_unlock(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	o2.ctx = &t2;
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	o2.ctx = NULL;
 | |
| 	mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
 | |
| 	ww_mutex_base_unlock(&o2.base);
 | |
| 
 | |
| 	WWL(&o2, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_no_unlock_slow(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	o2.ctx = NULL;
 | |
| 	mutex_acquire(&o2.base.dep_map, 0, 1, _THIS_IP_);
 | |
| 	ww_mutex_base_unlock(&o2.base);
 | |
| 
 | |
| 	ww_mutex_lock_slow(&o2, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_acquire_more(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	ret = WWL(&o3, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_acquire_more_slow(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	ww_mutex_lock_slow(&o3, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_acquire_more_edeadlk(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o3.base);
 | |
| 	mutex_release(&o3.base.dep_map, _THIS_IP_);
 | |
| 	o3.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	ret = WWL(&o3, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_acquire_more_edeadlk_slow(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o3.base);
 | |
| 	mutex_release(&o3.base.dep_map, _THIS_IP_);
 | |
| 	o3.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 
 | |
| 	ww_mutex_lock_slow(&o3, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_acquire_wrong(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 	if (!ret)
 | |
| 		WWU(&o2);
 | |
| 
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	ret = WWL(&o3, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_edeadlk_acquire_wrong_slow(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ww_mutex_base_lock(&o2.base);
 | |
| 	mutex_release(&o2.base.dep_map, _THIS_IP_);
 | |
| 	o2.ctx = &t2;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 	t2 = t;
 | |
| 	t2.stamp--;
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret != -EDEADLK);
 | |
| 	if (!ret)
 | |
| 		WWU(&o2);
 | |
| 
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	ww_mutex_lock_slow(&o3, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_spin_nest_unlocked(void)
 | |
| {
 | |
| 	spin_lock_nest_lock(&lock_A, &o.base);
 | |
| 	U(A);
 | |
| }
 | |
| 
 | |
| /* This is not a deadlock, because we have X1 to serialize Y1 and Y2 */
 | |
| static void ww_test_spin_nest_lock(void)
 | |
| {
 | |
| 	spin_lock(&lock_X1);
 | |
| 	spin_lock_nest_lock(&lock_Y1, &lock_X1);
 | |
| 	spin_lock(&lock_A);
 | |
| 	spin_lock_nest_lock(&lock_Y2, &lock_X1);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	spin_unlock(&lock_Y2);
 | |
| 	spin_unlock(&lock_Y1);
 | |
| 	spin_unlock(&lock_X1);
 | |
| }
 | |
| 
 | |
| static void ww_test_unneeded_slow(void)
 | |
| {
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ww_mutex_lock_slow(&o, &t);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_block(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 	WWL1(&o2);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_try(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWT(&o2);
 | |
| 	WARN_ON(!ret);
 | |
| 	WWU(&o2);
 | |
| 	WWU(&o);
 | |
| }
 | |
| 
 | |
| static void ww_test_context_context(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret);
 | |
| 
 | |
| 	WWU(&o2);
 | |
| 	WWU(&o);
 | |
| }
 | |
| 
 | |
| static void ww_test_try_block(void)
 | |
| {
 | |
| 	bool ret;
 | |
| 
 | |
| 	ret = WWT(&o);
 | |
| 	WARN_ON(!ret);
 | |
| 
 | |
| 	WWL1(&o2);
 | |
| 	WWU(&o2);
 | |
| 	WWU(&o);
 | |
| }
 | |
| 
 | |
| static void ww_test_try_try(void)
 | |
| {
 | |
| 	bool ret;
 | |
| 
 | |
| 	ret = WWT(&o);
 | |
| 	WARN_ON(!ret);
 | |
| 	ret = WWT(&o2);
 | |
| 	WARN_ON(!ret);
 | |
| 	WWU(&o2);
 | |
| 	WWU(&o);
 | |
| }
 | |
| 
 | |
| static void ww_test_try_context(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = WWT(&o);
 | |
| 	WARN_ON(!ret);
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret);
 | |
| }
 | |
| 
 | |
| static void ww_test_block_block(void)
 | |
| {
 | |
| 	WWL1(&o);
 | |
| 	WWL1(&o2);
 | |
| }
 | |
| 
 | |
| static void ww_test_block_try(void)
 | |
| {
 | |
| 	bool ret;
 | |
| 
 | |
| 	WWL1(&o);
 | |
| 	ret = WWT(&o2);
 | |
| 	WARN_ON(!ret);
 | |
| }
 | |
| 
 | |
| static void ww_test_block_context(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	WWL1(&o);
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ret = WWL(&o2, &t);
 | |
| 	WARN_ON(ret);
 | |
| }
 | |
| 
 | |
| static void ww_test_spin_block(void)
 | |
| {
 | |
| 	L(A);
 | |
| 	U(A);
 | |
| 
 | |
| 	WWL1(&o);
 | |
| 	L(A);
 | |
| 	U(A);
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	L(A);
 | |
| 	WWL1(&o);
 | |
| 	WWU(&o);
 | |
| 	U(A);
 | |
| }
 | |
| 
 | |
| static void ww_test_spin_try(void)
 | |
| {
 | |
| 	bool ret;
 | |
| 
 | |
| 	L(A);
 | |
| 	U(A);
 | |
| 
 | |
| 	ret = WWT(&o);
 | |
| 	WARN_ON(!ret);
 | |
| 	L(A);
 | |
| 	U(A);
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	L(A);
 | |
| 	ret = WWT(&o);
 | |
| 	WARN_ON(!ret);
 | |
| 	WWU(&o);
 | |
| 	U(A);
 | |
| }
 | |
| 
 | |
| static void ww_test_spin_context(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	L(A);
 | |
| 	U(A);
 | |
| 
 | |
| 	WWAI(&t);
 | |
| 
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 	L(A);
 | |
| 	U(A);
 | |
| 	WWU(&o);
 | |
| 
 | |
| 	L(A);
 | |
| 	ret = WWL(&o, &t);
 | |
| 	WARN_ON(ret);
 | |
| 	WWU(&o);
 | |
| 	U(A);
 | |
| }
 | |
| 
 | |
| static void ww_tests(void)
 | |
| {
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	printk("  | Wound/wait tests |\n");
 | |
| 	printk("  ---------------------\n");
 | |
| 
 | |
| 	print_testname("ww api failures");
 | |
| 	dotest(ww_test_fail_acquire, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_normal, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_unneeded_slow, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("ww contexts mixing");
 | |
| 	dotest(ww_test_two_contexts, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_diff_class, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("finishing ww context");
 | |
| 	dotest(ww_test_context_done_twice, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_context_unlock_twice, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_context_fini_early, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_context_lock_after_done, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("locking mismatches");
 | |
| 	dotest(ww_test_object_unlock_twice, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_object_lock_unbalanced, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_object_lock_stale_context, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("EDEADLK handling");
 | |
| 	dotest(ww_test_edeadlk_normal, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_normal_slow, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_no_unlock, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_no_unlock_slow, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_acquire_more, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_acquire_more_slow, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_acquire_more_edeadlk, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_acquire_more_edeadlk_slow, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_acquire_wrong, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_edeadlk_acquire_wrong_slow, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("spinlock nest unlocked");
 | |
| 	dotest(ww_test_spin_nest_unlocked, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("spinlock nest test");
 | |
| 	dotest(ww_test_spin_nest_lock, SUCCESS, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	printk("  -----------------------------------------------------\n");
 | |
| 	printk("                                 |block | try  |context|\n");
 | |
| 	printk("  -----------------------------------------------------\n");
 | |
| 
 | |
| 	print_testname("context");
 | |
| 	dotest(ww_test_context_block, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_context_try, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_context_context, SUCCESS, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("try");
 | |
| 	dotest(ww_test_try_block, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_try_try, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_try_context, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("block");
 | |
| 	dotest(ww_test_block_block, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_block_try, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_block_context, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("spinlock");
 | |
| 	dotest(ww_test_spin_block, FAILURE, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_spin_try, SUCCESS, LOCKTYPE_WW);
 | |
| 	dotest(ww_test_spin_context, FAILURE, LOCKTYPE_WW);
 | |
| 	pr_cont("\n");
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|  * <in hardirq handler>
 | |
|  * read_lock(&A);
 | |
|  *			<hardirq disable>
 | |
|  *			spin_lock(&B);
 | |
|  * spin_lock(&B);
 | |
|  *			read_lock(&A);
 | |
|  *
 | |
|  * is a deadlock.
 | |
|  */
 | |
| static void queued_read_lock_hardirq_RE_Er(void)
 | |
| {
 | |
| 	HARDIRQ_ENTER();
 | |
| 	read_lock(&rwlock_A);
 | |
| 	LOCK(B);
 | |
| 	UNLOCK(B);
 | |
| 	read_unlock(&rwlock_A);
 | |
| 	HARDIRQ_EXIT();
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	LOCK(B);
 | |
| 	read_lock(&rwlock_A);
 | |
| 	read_unlock(&rwlock_A);
 | |
| 	UNLOCK(B);
 | |
| 	HARDIRQ_ENABLE();
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * <in hardirq handler>
 | |
|  * spin_lock(&B);
 | |
|  *			<hardirq disable>
 | |
|  *			read_lock(&A);
 | |
|  * read_lock(&A);
 | |
|  *			spin_lock(&B);
 | |
|  *
 | |
|  * is not a deadlock.
 | |
|  */
 | |
| static void queued_read_lock_hardirq_ER_rE(void)
 | |
| {
 | |
| 	HARDIRQ_ENTER();
 | |
| 	LOCK(B);
 | |
| 	read_lock(&rwlock_A);
 | |
| 	read_unlock(&rwlock_A);
 | |
| 	UNLOCK(B);
 | |
| 	HARDIRQ_EXIT();
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	read_lock(&rwlock_A);
 | |
| 	LOCK(B);
 | |
| 	UNLOCK(B);
 | |
| 	read_unlock(&rwlock_A);
 | |
| 	HARDIRQ_ENABLE();
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * <hardirq disable>
 | |
|  * spin_lock(&B);
 | |
|  *			read_lock(&A);
 | |
|  *			<in hardirq handler>
 | |
|  *			spin_lock(&B);
 | |
|  * read_lock(&A);
 | |
|  *
 | |
|  * is a deadlock. Because the two read_lock()s are both non-recursive readers.
 | |
|  */
 | |
| static void queued_read_lock_hardirq_inversion(void)
 | |
| {
 | |
| 
 | |
| 	HARDIRQ_ENTER();
 | |
| 	LOCK(B);
 | |
| 	UNLOCK(B);
 | |
| 	HARDIRQ_EXIT();
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	LOCK(B);
 | |
| 	read_lock(&rwlock_A);
 | |
| 	read_unlock(&rwlock_A);
 | |
| 	UNLOCK(B);
 | |
| 	HARDIRQ_ENABLE();
 | |
| 
 | |
| 	read_lock(&rwlock_A);
 | |
| 	read_unlock(&rwlock_A);
 | |
| }
 | |
| 
 | |
| static void queued_read_lock_tests(void)
 | |
| {
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	printk("  | queued read lock tests |\n");
 | |
| 	printk("  ---------------------------\n");
 | |
| 	print_testname("hardirq read-lock/lock-read");
 | |
| 	dotest(queued_read_lock_hardirq_RE_Er, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("hardirq lock-read/read-lock");
 | |
| 	dotest(queued_read_lock_hardirq_ER_rE, SUCCESS, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("hardirq inversion");
 | |
| 	dotest(queued_read_lock_hardirq_inversion, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("\n");
 | |
| }
 | |
| 
 | |
| static void fs_reclaim_correct_nesting(void)
 | |
| {
 | |
| 	fs_reclaim_acquire(GFP_KERNEL);
 | |
| 	might_alloc(GFP_NOFS);
 | |
| 	fs_reclaim_release(GFP_KERNEL);
 | |
| }
 | |
| 
 | |
| static void fs_reclaim_wrong_nesting(void)
 | |
| {
 | |
| 	fs_reclaim_acquire(GFP_KERNEL);
 | |
| 	might_alloc(GFP_KERNEL);
 | |
| 	fs_reclaim_release(GFP_KERNEL);
 | |
| }
 | |
| 
 | |
| static void fs_reclaim_protected_nesting(void)
 | |
| {
 | |
| 	unsigned int flags;
 | |
| 
 | |
| 	fs_reclaim_acquire(GFP_KERNEL);
 | |
| 	flags = memalloc_nofs_save();
 | |
| 	might_alloc(GFP_KERNEL);
 | |
| 	memalloc_nofs_restore(flags);
 | |
| 	fs_reclaim_release(GFP_KERNEL);
 | |
| }
 | |
| 
 | |
| static void fs_reclaim_tests(void)
 | |
| {
 | |
| 	printk("  --------------------\n");
 | |
| 	printk("  | fs_reclaim tests |\n");
 | |
| 	printk("  --------------------\n");
 | |
| 
 | |
| 	print_testname("correct nesting");
 | |
| 	dotest(fs_reclaim_correct_nesting, SUCCESS, 0);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("wrong nesting");
 | |
| 	dotest(fs_reclaim_wrong_nesting, FAILURE, 0);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("protected nesting");
 | |
| 	dotest(fs_reclaim_protected_nesting, SUCCESS, 0);
 | |
| 	pr_cont("\n");
 | |
| }
 | |
| 
 | |
| #define __guard(cleanup) __maybe_unused __attribute__((__cleanup__(cleanup)))
 | |
| 
 | |
| static void hardirq_exit(int *_)
 | |
| {
 | |
| 	HARDIRQ_EXIT();
 | |
| }
 | |
| 
 | |
| #define HARDIRQ_CONTEXT(name, ...)					\
 | |
| 	int hardirq_guard_##name __guard(hardirq_exit);			\
 | |
| 	HARDIRQ_ENTER();
 | |
| 
 | |
| #define NOTTHREADED_HARDIRQ_CONTEXT(name, ...)				\
 | |
| 	int notthreaded_hardirq_guard_##name __guard(hardirq_exit);	\
 | |
| 	local_irq_disable();						\
 | |
| 	__irq_enter();							\
 | |
| 	WARN_ON(!in_irq());
 | |
| 
 | |
| static void softirq_exit(int *_)
 | |
| {
 | |
| 	SOFTIRQ_EXIT();
 | |
| }
 | |
| 
 | |
| #define SOFTIRQ_CONTEXT(name, ...)				\
 | |
| 	int softirq_guard_##name __guard(softirq_exit);		\
 | |
| 	SOFTIRQ_ENTER();
 | |
| 
 | |
| static void rcu_exit(int *_)
 | |
| {
 | |
| 	rcu_read_unlock();
 | |
| }
 | |
| 
 | |
| #define RCU_CONTEXT(name, ...)					\
 | |
| 	int rcu_guard_##name __guard(rcu_exit);			\
 | |
| 	rcu_read_lock();
 | |
| 
 | |
| static void rcu_bh_exit(int *_)
 | |
| {
 | |
| 	rcu_read_unlock_bh();
 | |
| }
 | |
| 
 | |
| #define RCU_BH_CONTEXT(name, ...)				\
 | |
| 	int rcu_bh_guard_##name __guard(rcu_bh_exit);		\
 | |
| 	rcu_read_lock_bh();
 | |
| 
 | |
| static void rcu_sched_exit(int *_)
 | |
| {
 | |
| 	rcu_read_unlock_sched();
 | |
| }
 | |
| 
 | |
| #define RCU_SCHED_CONTEXT(name, ...)				\
 | |
| 	int rcu_sched_guard_##name __guard(rcu_sched_exit);	\
 | |
| 	rcu_read_lock_sched();
 | |
| 
 | |
| static void raw_spinlock_exit(raw_spinlock_t **lock)
 | |
| {
 | |
| 	raw_spin_unlock(*lock);
 | |
| }
 | |
| 
 | |
| #define RAW_SPINLOCK_CONTEXT(name, lock)						\
 | |
| 	raw_spinlock_t *raw_spinlock_guard_##name __guard(raw_spinlock_exit) = &(lock);	\
 | |
| 	raw_spin_lock(&(lock));
 | |
| 
 | |
| static void spinlock_exit(spinlock_t **lock)
 | |
| {
 | |
| 	spin_unlock(*lock);
 | |
| }
 | |
| 
 | |
| #define SPINLOCK_CONTEXT(name, lock)						\
 | |
| 	spinlock_t *spinlock_guard_##name __guard(spinlock_exit) = &(lock);	\
 | |
| 	spin_lock(&(lock));
 | |
| 
 | |
| static void mutex_exit(struct mutex **lock)
 | |
| {
 | |
| 	mutex_unlock(*lock);
 | |
| }
 | |
| 
 | |
| #define MUTEX_CONTEXT(name, lock)					\
 | |
| 	struct mutex *mutex_guard_##name __guard(mutex_exit) = &(lock);	\
 | |
| 	mutex_lock(&(lock));
 | |
| 
 | |
| #define GENERATE_2_CONTEXT_TESTCASE(outer, outer_lock, inner, inner_lock)	\
 | |
| 										\
 | |
| static void __maybe_unused inner##_in_##outer(void)				\
 | |
| {										\
 | |
| 	outer##_CONTEXT(_, outer_lock);						\
 | |
| 	{									\
 | |
| 		inner##_CONTEXT(_, inner_lock);					\
 | |
| 	}									\
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * wait contexts (considering PREEMPT_RT)
 | |
|  *
 | |
|  * o: inner is allowed in outer
 | |
|  * x: inner is disallowed in outer
 | |
|  *
 | |
|  *       \  inner |  RCU  | RAW_SPIN | SPIN | MUTEX
 | |
|  * outer  \       |       |          |      |
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * HARDIRQ        |   o   |    o     |  o   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * NOTTHREADED_IRQ|   o   |    o     |  x   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * SOFTIRQ        |   o   |    o     |  o   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * RCU            |   o   |    o     |  o   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * RCU_BH         |   o   |    o     |  o   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * RCU_SCHED      |   o   |    o     |  x   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * RAW_SPIN       |   o   |    o     |  x   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * SPIN           |   o   |    o     |  o   |  x
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  * MUTEX          |   o   |    o     |  o   |  o
 | |
|  * ---------------+-------+----------+------+-------
 | |
|  */
 | |
| 
 | |
| #define GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(inner, inner_lock)		\
 | |
| GENERATE_2_CONTEXT_TESTCASE(HARDIRQ, , inner, inner_lock)			\
 | |
| GENERATE_2_CONTEXT_TESTCASE(NOTTHREADED_HARDIRQ, , inner, inner_lock)		\
 | |
| GENERATE_2_CONTEXT_TESTCASE(SOFTIRQ, , inner, inner_lock)			\
 | |
| GENERATE_2_CONTEXT_TESTCASE(RCU, , inner, inner_lock)				\
 | |
| GENERATE_2_CONTEXT_TESTCASE(RCU_BH, , inner, inner_lock)			\
 | |
| GENERATE_2_CONTEXT_TESTCASE(RCU_SCHED, , inner, inner_lock)			\
 | |
| GENERATE_2_CONTEXT_TESTCASE(RAW_SPINLOCK, raw_lock_A, inner, inner_lock)	\
 | |
| GENERATE_2_CONTEXT_TESTCASE(SPINLOCK, lock_A, inner, inner_lock)		\
 | |
| GENERATE_2_CONTEXT_TESTCASE(MUTEX, mutex_A, inner, inner_lock)
 | |
| 
 | |
| GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(RCU, )
 | |
| GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(RAW_SPINLOCK, raw_lock_B)
 | |
| GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(SPINLOCK, lock_B)
 | |
| GENERATE_2_CONTEXT_TESTCASE_FOR_ALL_OUTER(MUTEX, mutex_B)
 | |
| 
 | |
| /* the outer context allows all kinds of preemption */
 | |
| #define DO_CONTEXT_TESTCASE_OUTER_PREEMPTIBLE(outer)			\
 | |
| 	dotest(RCU_in_##outer, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(RAW_SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN);	\
 | |
| 	dotest(SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN);		\
 | |
| 	dotest(MUTEX_in_##outer, SUCCESS, LOCKTYPE_MUTEX);		\
 | |
| 
 | |
| /*
 | |
|  * the outer context only allows the preemption introduced by spinlock_t (which
 | |
|  * is a sleepable lock for PREEMPT_RT)
 | |
|  */
 | |
| #define DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(outer)		\
 | |
| 	dotest(RCU_in_##outer, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(RAW_SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN);	\
 | |
| 	dotest(SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN);		\
 | |
| 	dotest(MUTEX_in_##outer, FAILURE, LOCKTYPE_MUTEX);		\
 | |
| 
 | |
| /* the outer doesn't allows any kind of preemption */
 | |
| #define DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(outer)			\
 | |
| 	dotest(RCU_in_##outer, SUCCESS, LOCKTYPE_RWLOCK);		\
 | |
| 	dotest(RAW_SPINLOCK_in_##outer, SUCCESS, LOCKTYPE_SPIN);	\
 | |
| 	dotest(SPINLOCK_in_##outer, FAILURE, LOCKTYPE_SPIN);		\
 | |
| 	dotest(MUTEX_in_##outer, FAILURE, LOCKTYPE_MUTEX);		\
 | |
| 
 | |
| static void wait_context_tests(void)
 | |
| {
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	printk("  | wait context tests |\n");
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	printk("                                 | rcu  | raw  | spin |mutex |\n");
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	print_testname("in hardirq context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(HARDIRQ);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in hardirq context (not threaded)");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(NOTTHREADED_HARDIRQ);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in softirq context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(SOFTIRQ);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in RCU context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(RCU);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in RCU-bh context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(RCU_BH);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in RCU-sched context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(RCU_SCHED);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in RAW_SPINLOCK context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_NOT_PREEMPTIBLE(RAW_SPINLOCK);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in SPINLOCK context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_LIMITED_PREEMPTIBLE(SPINLOCK);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("in MUTEX context");
 | |
| 	DO_CONTEXT_TESTCASE_OUTER_PREEMPTIBLE(MUTEX);
 | |
| 	pr_cont("\n");
 | |
| }
 | |
| 
 | |
| static void local_lock_2(void)
 | |
| {
 | |
| 	local_lock(&local_A);	/* IRQ-ON */
 | |
| 	local_unlock(&local_A);
 | |
| 
 | |
| 	HARDIRQ_ENTER();
 | |
| 	spin_lock(&lock_A);		/* IN-IRQ */
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_EXIT()
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_A);
 | |
| 	local_lock(&local_A);	/* IN-IRQ <-> IRQ-ON cycle, false */
 | |
| 	local_unlock(&local_A);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_ENABLE();
 | |
| }
 | |
| 
 | |
| static void local_lock_3A(void)
 | |
| {
 | |
| 	local_lock(&local_A);	/* IRQ-ON */
 | |
| 	spin_lock(&lock_B);		/* IRQ-ON */
 | |
| 	spin_unlock(&lock_B);
 | |
| 	local_unlock(&local_A);
 | |
| 
 | |
| 	HARDIRQ_ENTER();
 | |
| 	spin_lock(&lock_A);		/* IN-IRQ */
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_EXIT()
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_A);
 | |
| 	local_lock(&local_A);	/* IN-IRQ <-> IRQ-ON cycle only if we count local_lock(), false */
 | |
| 	local_unlock(&local_A);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_ENABLE();
 | |
| }
 | |
| 
 | |
| static void local_lock_3B(void)
 | |
| {
 | |
| 	local_lock(&local_A);	/* IRQ-ON */
 | |
| 	spin_lock(&lock_B);		/* IRQ-ON */
 | |
| 	spin_unlock(&lock_B);
 | |
| 	local_unlock(&local_A);
 | |
| 
 | |
| 	HARDIRQ_ENTER();
 | |
| 	spin_lock(&lock_A);		/* IN-IRQ */
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_EXIT()
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_A);
 | |
| 	local_lock(&local_A);	/* IN-IRQ <-> IRQ-ON cycle only if we count local_lock(), false */
 | |
| 	local_unlock(&local_A);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_ENABLE();
 | |
| 
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_A);
 | |
| 	spin_lock(&lock_B);		/* IN-IRQ <-> IRQ-ON cycle, true */
 | |
| 	spin_unlock(&lock_B);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_DISABLE();
 | |
| 
 | |
| }
 | |
| 
 | |
| static void local_lock_tests(void)
 | |
| {
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	printk("  | local_lock tests |\n");
 | |
| 	printk("  ---------------------\n");
 | |
| 
 | |
| 	print_testname("local_lock inversion  2");
 | |
| 	dotest(local_lock_2, SUCCESS, LOCKTYPE_LL);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("local_lock inversion 3A");
 | |
| 	dotest(local_lock_3A, SUCCESS, LOCKTYPE_LL);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("local_lock inversion 3B");
 | |
| 	dotest(local_lock_3B, FAILURE, LOCKTYPE_LL);
 | |
| 	pr_cont("\n");
 | |
| }
 | |
| 
 | |
| static void hardirq_deadlock_softirq_not_deadlock(void)
 | |
| {
 | |
| 	/* mutex_A is hardirq-unsafe and softirq-unsafe */
 | |
| 	/* mutex_A -> lock_C */
 | |
| 	mutex_lock(&mutex_A);
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_C);
 | |
| 	spin_unlock(&lock_C);
 | |
| 	HARDIRQ_ENABLE();
 | |
| 	mutex_unlock(&mutex_A);
 | |
| 
 | |
| 	/* lock_A is hardirq-safe */
 | |
| 	HARDIRQ_ENTER();
 | |
| 	spin_lock(&lock_A);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_EXIT();
 | |
| 
 | |
| 	/* lock_A -> lock_B */
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_A);
 | |
| 	spin_lock(&lock_B);
 | |
| 	spin_unlock(&lock_B);
 | |
| 	spin_unlock(&lock_A);
 | |
| 	HARDIRQ_ENABLE();
 | |
| 
 | |
| 	/* lock_B -> lock_C */
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_B);
 | |
| 	spin_lock(&lock_C);
 | |
| 	spin_unlock(&lock_C);
 | |
| 	spin_unlock(&lock_B);
 | |
| 	HARDIRQ_ENABLE();
 | |
| 
 | |
| 	/* lock_D is softirq-safe */
 | |
| 	SOFTIRQ_ENTER();
 | |
| 	spin_lock(&lock_D);
 | |
| 	spin_unlock(&lock_D);
 | |
| 	SOFTIRQ_EXIT();
 | |
| 
 | |
| 	/* And lock_D is hardirq-unsafe */
 | |
| 	SOFTIRQ_DISABLE();
 | |
| 	spin_lock(&lock_D);
 | |
| 	spin_unlock(&lock_D);
 | |
| 	SOFTIRQ_ENABLE();
 | |
| 
 | |
| 	/*
 | |
| 	 * mutex_A -> lock_C -> lock_D is softirq-unsafe -> softirq-safe, not
 | |
| 	 * deadlock.
 | |
| 	 *
 | |
| 	 * lock_A -> lock_B -> lock_C -> lock_D is hardirq-safe ->
 | |
| 	 * hardirq-unsafe, deadlock.
 | |
| 	 */
 | |
| 	HARDIRQ_DISABLE();
 | |
| 	spin_lock(&lock_C);
 | |
| 	spin_lock(&lock_D);
 | |
| 	spin_unlock(&lock_D);
 | |
| 	spin_unlock(&lock_C);
 | |
| 	HARDIRQ_ENABLE();
 | |
| }
 | |
| 
 | |
| void locking_selftest(void)
 | |
| {
 | |
| 	/*
 | |
| 	 * Got a locking failure before the selftest ran?
 | |
| 	 */
 | |
| 	if (!debug_locks) {
 | |
| 		printk("----------------------------------\n");
 | |
| 		printk("| Locking API testsuite disabled |\n");
 | |
| 		printk("----------------------------------\n");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * treats read_lock() as recursive read locks for testing purpose
 | |
| 	 */
 | |
| 	force_read_lock_recursive = 1;
 | |
| 
 | |
| 	/*
 | |
| 	 * Run the testsuite:
 | |
| 	 */
 | |
| 	printk("------------------------\n");
 | |
| 	printk("| Locking API testsuite:\n");
 | |
| 	printk("----------------------------------------------------------------------------\n");
 | |
| 	printk("                                 | spin |wlock |rlock |mutex | wsem | rsem |rtmutex\n");
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 
 | |
| 	init_shared_classes();
 | |
| 	lockdep_set_selftest_task(current);
 | |
| 
 | |
| 	DO_TESTCASE_6R("A-A deadlock", AA);
 | |
| 	DO_TESTCASE_6R("A-B-B-A deadlock", ABBA);
 | |
| 	DO_TESTCASE_6R("A-B-B-C-C-A deadlock", ABBCCA);
 | |
| 	DO_TESTCASE_6R("A-B-C-A-B-C deadlock", ABCABC);
 | |
| 	DO_TESTCASE_6R("A-B-B-C-C-D-D-A deadlock", ABBCCDDA);
 | |
| 	DO_TESTCASE_6R("A-B-C-D-B-D-D-A deadlock", ABCDBDDA);
 | |
| 	DO_TESTCASE_6R("A-B-C-D-B-C-D-A deadlock", ABCDBCDA);
 | |
| 	DO_TESTCASE_6("double unlock", double_unlock);
 | |
| 	DO_TESTCASE_6("initialize held", init_held);
 | |
| 
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	print_testname("recursive read-lock");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_AA1, SUCCESS, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rsem_AA1, FAILURE, LOCKTYPE_RWSEM);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("recursive read-lock #2");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_AA1B, SUCCESS, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rsem_AA1B, FAILURE, LOCKTYPE_RWSEM);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("mixed read-write-lock");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_AA2, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rsem_AA2, FAILURE, LOCKTYPE_RWSEM);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("mixed write-read-lock");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_AA3, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rsem_AA3, FAILURE, LOCKTYPE_RWSEM);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	print_testname("mixed read-lock/lock-write ABBA");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_ABBA1, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rwsem_ABBA1, FAILURE, LOCKTYPE_RWSEM);
 | |
| 
 | |
| 	print_testname("mixed read-lock/lock-read ABBA");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_ABBA2, SUCCESS, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rwsem_ABBA2, FAILURE, LOCKTYPE_RWSEM);
 | |
| 
 | |
| 	print_testname("mixed write-lock/lock-write ABBA");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_ABBA3, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rwsem_ABBA3, FAILURE, LOCKTYPE_RWSEM);
 | |
| 
 | |
| 	print_testname("chain cached mixed R-L/L-W ABBA");
 | |
| 	pr_cont("             |");
 | |
| 	dotest(rlock_chaincache_ABBA1, FAILURE, LOCKTYPE_RWLOCK);
 | |
| 
 | |
| 	DO_TESTCASE_6x1RRB("rlock W1R2/W2R3/W3R1", W1R2_W2R3_W3R1);
 | |
| 	DO_TESTCASE_6x1RRB("rlock W1W2/R2R3/W3R1", W1W2_R2R3_W3R1);
 | |
| 	DO_TESTCASE_6x1RR("rlock W1W2/R2R3/R3W1", W1W2_R2R3_R3W1);
 | |
| 	DO_TESTCASE_6x1RR("rlock W1R2/R2R3/W3W1", W1R2_R2R3_W3W1);
 | |
| 
 | |
| 	printk("  --------------------------------------------------------------------------\n");
 | |
| 	/*
 | |
| 	 * irq-context testcases:
 | |
| 	 */
 | |
| 	DO_TESTCASE_2x6("irqs-on + irq-safe-A", irqsafe1);
 | |
| 	NON_RT(DO_TESTCASE_2x3("sirq-safe-A => hirqs-on", irqsafe2A));
 | |
| 	DO_TESTCASE_2x6("safe-A + irqs-on", irqsafe2B);
 | |
| 	DO_TESTCASE_6x6("safe-A + unsafe-B #1", irqsafe3);
 | |
| 	DO_TESTCASE_6x6("safe-A + unsafe-B #2", irqsafe4);
 | |
| 	DO_TESTCASE_6x6RW("irq lock-inversion", irq_inversion);
 | |
| 
 | |
| 	DO_TESTCASE_6x2x2RW("irq read-recursion", irq_read_recursion);
 | |
| 	DO_TESTCASE_6x2x2RW("irq read-recursion #2", irq_read_recursion2);
 | |
| 	DO_TESTCASE_6x2x2RW("irq read-recursion #3", irq_read_recursion3);
 | |
| 
 | |
| 	ww_tests();
 | |
| 
 | |
| 	force_read_lock_recursive = 0;
 | |
| 	/*
 | |
| 	 * queued_read_lock() specific test cases can be put here
 | |
| 	 */
 | |
| 	if (IS_ENABLED(CONFIG_QUEUED_RWLOCKS))
 | |
| 		queued_read_lock_tests();
 | |
| 
 | |
| 	fs_reclaim_tests();
 | |
| 
 | |
| 	/* Wait context test cases that are specific for RAW_LOCK_NESTING */
 | |
| 	if (IS_ENABLED(CONFIG_PROVE_RAW_LOCK_NESTING))
 | |
| 		wait_context_tests();
 | |
| 
 | |
| 	local_lock_tests();
 | |
| 
 | |
| 	print_testname("hardirq_unsafe_softirq_safe");
 | |
| 	dotest(hardirq_deadlock_softirq_not_deadlock, FAILURE, LOCKTYPE_SPECIAL);
 | |
| 	pr_cont("\n");
 | |
| 
 | |
| 	if (unexpected_testcase_failures) {
 | |
| 		printk("-----------------------------------------------------------------\n");
 | |
| 		debug_locks = 0;
 | |
| 		printk("BUG: %3d unexpected failures (out of %3d) - debugging disabled! |\n",
 | |
| 			unexpected_testcase_failures, testcase_total);
 | |
| 		printk("-----------------------------------------------------------------\n");
 | |
| 	} else if (expected_testcase_failures && testcase_successes) {
 | |
| 		printk("--------------------------------------------------------\n");
 | |
| 		printk("%3d out of %3d testcases failed, as expected. |\n",
 | |
| 			expected_testcase_failures, testcase_total);
 | |
| 		printk("----------------------------------------------------\n");
 | |
| 		debug_locks = 1;
 | |
| 	} else if (expected_testcase_failures && !testcase_successes) {
 | |
| 		printk("--------------------------------------------------------\n");
 | |
| 		printk("All %3d testcases failed, as expected. |\n",
 | |
| 			expected_testcase_failures);
 | |
| 		printk("----------------------------------------\n");
 | |
| 		debug_locks = 1;
 | |
| 	} else {
 | |
| 		printk("-------------------------------------------------------\n");
 | |
| 		printk("Good, all %3d testcases passed! |\n",
 | |
| 			testcase_successes);
 | |
| 		printk("---------------------------------\n");
 | |
| 		debug_locks = 1;
 | |
| 	}
 | |
| 	lockdep_set_selftest_task(NULL);
 | |
| 	debug_locks_silent = 0;
 | |
| }
 |