mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	lockdep: Fix lockdep recursion
Steve reported that lockdep_assert*irq*(), when nested inside lockdep
itself, will trigger a false-positive.
One example is the stack-trace code, as called from inside lockdep,
triggering tracing, which in turn calls RCU, which then uses
lockdep_assert_irqs_disabled().
Fixes: a21ee6055c ("lockdep: Change hardirq{s_enabled,_context} to per-cpu variables")
Reported-by: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Ingo Molnar <mingo@kernel.org>
			
			
This commit is contained in:
		
							parent
							
								
									2bb8945bcc
								
							
						
					
					
						commit
						4d004099a6
					
				
					 2 changed files with 67 additions and 45 deletions
				
			
		| 
						 | 
					@ -534,6 +534,7 @@ do {									\
 | 
				
			||||||
 | 
					
 | 
				
			||||||
DECLARE_PER_CPU(int, hardirqs_enabled);
 | 
					DECLARE_PER_CPU(int, hardirqs_enabled);
 | 
				
			||||||
DECLARE_PER_CPU(int, hardirq_context);
 | 
					DECLARE_PER_CPU(int, hardirq_context);
 | 
				
			||||||
 | 
					DECLARE_PER_CPU(unsigned int, lockdep_recursion);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * The below lockdep_assert_*() macros use raw_cpu_read() to access the above
 | 
					 * The below lockdep_assert_*() macros use raw_cpu_read() to access the above
 | 
				
			||||||
| 
						 | 
					@ -543,25 +544,27 @@ DECLARE_PER_CPU(int, hardirq_context);
 | 
				
			||||||
 * read the value from our previous CPU.
 | 
					 * read the value from our previous CPU.
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#define __lockdep_enabled	(debug_locks && !raw_cpu_read(lockdep_recursion))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define lockdep_assert_irqs_enabled()					\
 | 
					#define lockdep_assert_irqs_enabled()					\
 | 
				
			||||||
do {									\
 | 
					do {									\
 | 
				
			||||||
	WARN_ON_ONCE(debug_locks && !raw_cpu_read(hardirqs_enabled));	\
 | 
						WARN_ON_ONCE(__lockdep_enabled && !raw_cpu_read(hardirqs_enabled)); \
 | 
				
			||||||
} while (0)
 | 
					} while (0)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define lockdep_assert_irqs_disabled()					\
 | 
					#define lockdep_assert_irqs_disabled()					\
 | 
				
			||||||
do {									\
 | 
					do {									\
 | 
				
			||||||
	WARN_ON_ONCE(debug_locks && raw_cpu_read(hardirqs_enabled));	\
 | 
						WARN_ON_ONCE(__lockdep_enabled && raw_cpu_read(hardirqs_enabled)); \
 | 
				
			||||||
} while (0)
 | 
					} while (0)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define lockdep_assert_in_irq()						\
 | 
					#define lockdep_assert_in_irq()						\
 | 
				
			||||||
do {									\
 | 
					do {									\
 | 
				
			||||||
	WARN_ON_ONCE(debug_locks && !raw_cpu_read(hardirq_context));	\
 | 
						WARN_ON_ONCE(__lockdep_enabled && !raw_cpu_read(hardirq_context)); \
 | 
				
			||||||
} while (0)
 | 
					} while (0)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define lockdep_assert_preemption_enabled()				\
 | 
					#define lockdep_assert_preemption_enabled()				\
 | 
				
			||||||
do {									\
 | 
					do {									\
 | 
				
			||||||
	WARN_ON_ONCE(IS_ENABLED(CONFIG_PREEMPT_COUNT)	&&		\
 | 
						WARN_ON_ONCE(IS_ENABLED(CONFIG_PREEMPT_COUNT)	&&		\
 | 
				
			||||||
		     debug_locks			&&		\
 | 
							     __lockdep_enabled			&&		\
 | 
				
			||||||
		     (preempt_count() != 0		||		\
 | 
							     (preempt_count() != 0		||		\
 | 
				
			||||||
		      !raw_cpu_read(hardirqs_enabled)));		\
 | 
							      !raw_cpu_read(hardirqs_enabled)));		\
 | 
				
			||||||
} while (0)
 | 
					} while (0)
 | 
				
			||||||
| 
						 | 
					@ -569,7 +572,7 @@ do {									\
 | 
				
			||||||
#define lockdep_assert_preemption_disabled()				\
 | 
					#define lockdep_assert_preemption_disabled()				\
 | 
				
			||||||
do {									\
 | 
					do {									\
 | 
				
			||||||
	WARN_ON_ONCE(IS_ENABLED(CONFIG_PREEMPT_COUNT)	&&		\
 | 
						WARN_ON_ONCE(IS_ENABLED(CONFIG_PREEMPT_COUNT)	&&		\
 | 
				
			||||||
		     debug_locks			&&		\
 | 
							     __lockdep_enabled			&&		\
 | 
				
			||||||
		     (preempt_count() == 0		&&		\
 | 
							     (preempt_count() == 0		&&		\
 | 
				
			||||||
		      raw_cpu_read(hardirqs_enabled)));			\
 | 
							      raw_cpu_read(hardirqs_enabled)));			\
 | 
				
			||||||
} while (0)
 | 
					} while (0)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -76,6 +76,23 @@ module_param(lock_stat, int, 0644);
 | 
				
			||||||
#define lock_stat 0
 | 
					#define lock_stat 0
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					DEFINE_PER_CPU(unsigned int, lockdep_recursion);
 | 
				
			||||||
 | 
					EXPORT_PER_CPU_SYMBOL_GPL(lockdep_recursion);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline bool lockdep_enabled(void)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						if (!debug_locks)
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (raw_cpu_read(lockdep_recursion))
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (current->lockdep_recursion)
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						return true;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * lockdep_lock: protects the lockdep graph, the hashes and the
 | 
					 * lockdep_lock: protects the lockdep graph, the hashes and the
 | 
				
			||||||
 *               class/list/hash allocators.
 | 
					 *               class/list/hash allocators.
 | 
				
			||||||
| 
						 | 
					@ -93,7 +110,7 @@ static inline void lockdep_lock(void)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	arch_spin_lock(&__lock);
 | 
						arch_spin_lock(&__lock);
 | 
				
			||||||
	__owner = current;
 | 
						__owner = current;
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						__this_cpu_inc(lockdep_recursion);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static inline void lockdep_unlock(void)
 | 
					static inline void lockdep_unlock(void)
 | 
				
			||||||
| 
						 | 
					@ -101,7 +118,7 @@ static inline void lockdep_unlock(void)
 | 
				
			||||||
	if (debug_locks && DEBUG_LOCKS_WARN_ON(__owner != current))
 | 
						if (debug_locks && DEBUG_LOCKS_WARN_ON(__owner != current))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion--;
 | 
						__this_cpu_dec(lockdep_recursion);
 | 
				
			||||||
	__owner = NULL;
 | 
						__owner = NULL;
 | 
				
			||||||
	arch_spin_unlock(&__lock);
 | 
						arch_spin_unlock(&__lock);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -393,10 +410,15 @@ void lockdep_init_task(struct task_struct *task)
 | 
				
			||||||
	task->lockdep_recursion = 0;
 | 
						task->lockdep_recursion = 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static __always_inline void lockdep_recursion_inc(void)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						__this_cpu_inc(lockdep_recursion);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static __always_inline void lockdep_recursion_finish(void)
 | 
					static __always_inline void lockdep_recursion_finish(void)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (WARN_ON_ONCE((--current->lockdep_recursion) & LOCKDEP_RECURSION_MASK))
 | 
						if (WARN_ON_ONCE(__this_cpu_dec_return(lockdep_recursion)))
 | 
				
			||||||
		current->lockdep_recursion = 0;
 | 
							__this_cpu_write(lockdep_recursion, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void lockdep_set_selftest_task(struct task_struct *task)
 | 
					void lockdep_set_selftest_task(struct task_struct *task)
 | 
				
			||||||
| 
						 | 
					@ -3659,7 +3681,7 @@ void lockdep_hardirqs_on_prepare(unsigned long ip)
 | 
				
			||||||
	if (unlikely(in_nmi()))
 | 
						if (unlikely(in_nmi()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion & LOCKDEP_RECURSION_MASK))
 | 
						if (unlikely(__this_cpu_read(lockdep_recursion)))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(lockdep_hardirqs_enabled())) {
 | 
						if (unlikely(lockdep_hardirqs_enabled())) {
 | 
				
			||||||
| 
						 | 
					@ -3695,7 +3717,7 @@ void lockdep_hardirqs_on_prepare(unsigned long ip)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->hardirq_chain_key = current->curr_chain_key;
 | 
						current->hardirq_chain_key = current->curr_chain_key;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	__trace_hardirqs_on_caller();
 | 
						__trace_hardirqs_on_caller();
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -3728,7 +3750,7 @@ void noinstr lockdep_hardirqs_on(unsigned long ip)
 | 
				
			||||||
		goto skip_checks;
 | 
							goto skip_checks;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion & LOCKDEP_RECURSION_MASK))
 | 
						if (unlikely(__this_cpu_read(lockdep_recursion)))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (lockdep_hardirqs_enabled()) {
 | 
						if (lockdep_hardirqs_enabled()) {
 | 
				
			||||||
| 
						 | 
					@ -3781,7 +3803,7 @@ void noinstr lockdep_hardirqs_off(unsigned long ip)
 | 
				
			||||||
	if (in_nmi()) {
 | 
						if (in_nmi()) {
 | 
				
			||||||
		if (!IS_ENABLED(CONFIG_TRACE_IRQFLAGS_NMI))
 | 
							if (!IS_ENABLED(CONFIG_TRACE_IRQFLAGS_NMI))
 | 
				
			||||||
			return;
 | 
								return;
 | 
				
			||||||
	} else if (current->lockdep_recursion & LOCKDEP_RECURSION_MASK)
 | 
						} else if (__this_cpu_read(lockdep_recursion))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
| 
						 | 
					@ -3814,7 +3836,7 @@ void lockdep_softirqs_on(unsigned long ip)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct irqtrace_events *trace = ¤t->irqtrace;
 | 
						struct irqtrace_events *trace = ¤t->irqtrace;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(!debug_locks || current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
| 
						 | 
					@ -3829,7 +3851,7 @@ void lockdep_softirqs_on(unsigned long ip)
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
	 * We'll do an OFF -> ON transition:
 | 
						 * We'll do an OFF -> ON transition:
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
| 
						 | 
					@ -3852,7 +3874,7 @@ void lockdep_softirqs_on(unsigned long ip)
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
void lockdep_softirqs_off(unsigned long ip)
 | 
					void lockdep_softirqs_off(unsigned long ip)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (unlikely(!debug_locks || current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
| 
						 | 
					@ -4233,11 +4255,11 @@ void lockdep_init_map_waits(struct lockdep_map *lock, const char *name,
 | 
				
			||||||
	if (subclass) {
 | 
						if (subclass) {
 | 
				
			||||||
		unsigned long flags;
 | 
							unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (DEBUG_LOCKS_WARN_ON(current->lockdep_recursion))
 | 
							if (DEBUG_LOCKS_WARN_ON(!lockdep_enabled()))
 | 
				
			||||||
			return;
 | 
								return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		raw_local_irq_save(flags);
 | 
							raw_local_irq_save(flags);
 | 
				
			||||||
		current->lockdep_recursion++;
 | 
							lockdep_recursion_inc();
 | 
				
			||||||
		register_lock_class(lock, subclass, 1);
 | 
							register_lock_class(lock, subclass, 1);
 | 
				
			||||||
		lockdep_recursion_finish();
 | 
							lockdep_recursion_finish();
 | 
				
			||||||
		raw_local_irq_restore(flags);
 | 
							raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					@ -4920,11 +4942,11 @@ void lock_set_class(struct lockdep_map *lock, const char *name,
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
	if (__lock_set_class(lock, name, key, subclass, ip))
 | 
						if (__lock_set_class(lock, name, key, subclass, ip))
 | 
				
			||||||
		check_chain_key(current);
 | 
							check_chain_key(current);
 | 
				
			||||||
| 
						 | 
					@ -4937,11 +4959,11 @@ void lock_downgrade(struct lockdep_map *lock, unsigned long ip)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
	if (__lock_downgrade(lock, ip))
 | 
						if (__lock_downgrade(lock, ip))
 | 
				
			||||||
		check_chain_key(current);
 | 
							check_chain_key(current);
 | 
				
			||||||
| 
						 | 
					@ -4979,7 +5001,7 @@ static void verify_lock_unused(struct lockdep_map *lock, struct held_lock *hlock
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static bool lockdep_nmi(void)
 | 
					static bool lockdep_nmi(void)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (current->lockdep_recursion & LOCKDEP_RECURSION_MASK)
 | 
						if (raw_cpu_read(lockdep_recursion))
 | 
				
			||||||
		return false;
 | 
							return false;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!in_nmi())
 | 
						if (!in_nmi())
 | 
				
			||||||
| 
						 | 
					@ -5000,7 +5022,10 @@ void lock_acquire(struct lockdep_map *lock, unsigned int subclass,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	trace_lock_acquire(lock, subclass, trylock, read, check, nest_lock, ip);
 | 
						trace_lock_acquire(lock, subclass, trylock, read, check, nest_lock, ip);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion)) {
 | 
						if (!debug_locks)
 | 
				
			||||||
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (unlikely(!lockdep_enabled())) {
 | 
				
			||||||
		/* XXX allow trylock from NMI ?!? */
 | 
							/* XXX allow trylock from NMI ?!? */
 | 
				
			||||||
		if (lockdep_nmi() && !trylock) {
 | 
							if (lockdep_nmi() && !trylock) {
 | 
				
			||||||
			struct held_lock hlock;
 | 
								struct held_lock hlock;
 | 
				
			||||||
| 
						 | 
					@ -5023,7 +5048,7 @@ void lock_acquire(struct lockdep_map *lock, unsigned int subclass,
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	__lock_acquire(lock, subclass, trylock, read, check,
 | 
						__lock_acquire(lock, subclass, trylock, read, check,
 | 
				
			||||||
		       irqs_disabled_flags(flags), nest_lock, ip, 0, 0);
 | 
							       irqs_disabled_flags(flags), nest_lock, ip, 0, 0);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
| 
						 | 
					@ -5037,13 +5062,13 @@ void lock_release(struct lockdep_map *lock, unsigned long ip)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	trace_lock_release(lock, ip);
 | 
						trace_lock_release(lock, ip);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	if (__lock_release(lock, ip))
 | 
						if (__lock_release(lock, ip))
 | 
				
			||||||
		check_chain_key(current);
 | 
							check_chain_key(current);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
| 
						 | 
					@ -5056,13 +5081,13 @@ noinstr int lock_is_held_type(const struct lockdep_map *lock, int read)
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
	int ret = 0;
 | 
						int ret = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return 1; /* avoid false negative lockdep_assert_held() */
 | 
							return 1; /* avoid false negative lockdep_assert_held() */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	ret = __lock_is_held(lock, read);
 | 
						ret = __lock_is_held(lock, read);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
	raw_local_irq_restore(flags);
 | 
						raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					@ -5077,13 +5102,13 @@ struct pin_cookie lock_pin_lock(struct lockdep_map *lock)
 | 
				
			||||||
	struct pin_cookie cookie = NIL_COOKIE;
 | 
						struct pin_cookie cookie = NIL_COOKIE;
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return cookie;
 | 
							return cookie;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	cookie = __lock_pin_lock(lock);
 | 
						cookie = __lock_pin_lock(lock);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
	raw_local_irq_restore(flags);
 | 
						raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					@ -5096,13 +5121,13 @@ void lock_repin_lock(struct lockdep_map *lock, struct pin_cookie cookie)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	__lock_repin_lock(lock, cookie);
 | 
						__lock_repin_lock(lock, cookie);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
	raw_local_irq_restore(flags);
 | 
						raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					@ -5113,13 +5138,13 @@ void lock_unpin_lock(struct lockdep_map *lock, struct pin_cookie cookie)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
						if (unlikely(!lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	__lock_unpin_lock(lock, cookie);
 | 
						__lock_unpin_lock(lock, cookie);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
	raw_local_irq_restore(flags);
 | 
						raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					@ -5249,15 +5274,12 @@ void lock_contended(struct lockdep_map *lock, unsigned long ip)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	trace_lock_acquired(lock, ip);
 | 
						trace_lock_acquired(lock, ip);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(!lock_stat || !debug_locks))
 | 
						if (unlikely(!lock_stat || !lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
					 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	__lock_contended(lock, ip);
 | 
						__lock_contended(lock, ip);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
	raw_local_irq_restore(flags);
 | 
						raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					@ -5270,15 +5292,12 @@ void lock_acquired(struct lockdep_map *lock, unsigned long ip)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	trace_lock_contended(lock, ip);
 | 
						trace_lock_contended(lock, ip);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(!lock_stat || !debug_locks))
 | 
						if (unlikely(!lock_stat || !lockdep_enabled()))
 | 
				
			||||||
		return;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (unlikely(current->lockdep_recursion))
 | 
					 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	raw_local_irq_save(flags);
 | 
						raw_local_irq_save(flags);
 | 
				
			||||||
	check_flags(flags);
 | 
						check_flags(flags);
 | 
				
			||||||
	current->lockdep_recursion++;
 | 
						lockdep_recursion_inc();
 | 
				
			||||||
	__lock_acquired(lock, ip);
 | 
						__lock_acquired(lock, ip);
 | 
				
			||||||
	lockdep_recursion_finish();
 | 
						lockdep_recursion_finish();
 | 
				
			||||||
	raw_local_irq_restore(flags);
 | 
						raw_local_irq_restore(flags);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue