mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	locking/lockdep: Move mark_lock() inside CONFIG_TRACE_IRQFLAGS && CONFIG_PROVE_LOCKING
The last cleanup patch triggered another issue, as now another function
should be moved into the same section:
 kernel/locking/lockdep.c:3580:12: error: 'mark_lock' defined but not used [-Werror=unused-function]
  static int mark_lock(struct task_struct *curr, struct held_lock *this,
Move mark_lock() into the same #ifdef section as its only caller, and
remove the now-unused mark_lock_irq() stub helper.
Signed-off-by: Arnd Bergmann <arnd@arndb.de>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Bart Van Assche <bvanassche@acm.org>
Cc: Frederic Weisbecker <frederic@kernel.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Waiman Long <longman@redhat.com>
Cc: Will Deacon <will.deacon@arm.com>
Cc: Yuyang Du <duyuyang@gmail.com>
Fixes: 0d2cc3b345 ("locking/lockdep: Move valid_state() inside CONFIG_TRACE_IRQFLAGS && CONFIG_PROVE_LOCKING")
Link: https://lkml.kernel.org/r/20190617124718.1232976-1-arnd@arndb.de
Signed-off-by: Ingo Molnar <mingo@kernel.org>
			
			
This commit is contained in:
		
							parent
							
								
									bf10c97adb
								
							
						
					
					
						commit
						886532aee3
					
				
					 1 changed files with 34 additions and 39 deletions
				
			
		| 
						 | 
				
			
			@ -437,13 +437,6 @@ static int verbose(struct lock_class *class)
 | 
			
		|||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 * Stack-trace: tightly packed array of stack backtrace
 | 
			
		||||
 * addresses. Protected by the graph_lock.
 | 
			
		||||
 */
 | 
			
		||||
unsigned long nr_stack_trace_entries;
 | 
			
		||||
static unsigned long stack_trace[MAX_STACK_TRACE_ENTRIES];
 | 
			
		||||
 | 
			
		||||
static void print_lockdep_off(const char *bug_msg)
 | 
			
		||||
{
 | 
			
		||||
	printk(KERN_DEBUG "%s\n", bug_msg);
 | 
			
		||||
| 
						 | 
				
			
			@ -453,6 +446,15 @@ static void print_lockdep_off(const char *bug_msg)
 | 
			
		|||
#endif
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
unsigned long nr_stack_trace_entries;
 | 
			
		||||
 | 
			
		||||
#if defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING)
 | 
			
		||||
/*
 | 
			
		||||
 * Stack-trace: tightly packed array of stack backtrace
 | 
			
		||||
 * addresses. Protected by the graph_lock.
 | 
			
		||||
 */
 | 
			
		||||
static unsigned long stack_trace[MAX_STACK_TRACE_ENTRIES];
 | 
			
		||||
 | 
			
		||||
static int save_trace(struct lock_trace *trace)
 | 
			
		||||
{
 | 
			
		||||
	unsigned long *entries = stack_trace + nr_stack_trace_entries;
 | 
			
		||||
| 
						 | 
				
			
			@ -475,6 +477,7 @@ static int save_trace(struct lock_trace *trace)
 | 
			
		|||
 | 
			
		||||
	return 1;
 | 
			
		||||
}
 | 
			
		||||
#endif
 | 
			
		||||
 | 
			
		||||
unsigned int nr_hardirq_chains;
 | 
			
		||||
unsigned int nr_softirq_chains;
 | 
			
		||||
| 
						 | 
				
			
			@ -488,6 +491,7 @@ unsigned int max_lockdep_depth;
 | 
			
		|||
DEFINE_PER_CPU(struct lockdep_stats, lockdep_stats);
 | 
			
		||||
#endif
 | 
			
		||||
 | 
			
		||||
#if defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING)
 | 
			
		||||
/*
 | 
			
		||||
 * Locking printouts:
 | 
			
		||||
 */
 | 
			
		||||
| 
						 | 
				
			
			@ -505,6 +509,7 @@ static const char *usage_str[] =
 | 
			
		|||
#undef LOCKDEP_STATE
 | 
			
		||||
	[LOCK_USED] = "INITIAL USE",
 | 
			
		||||
};
 | 
			
		||||
#endif
 | 
			
		||||
 | 
			
		||||
const char * __get_key_name(struct lockdep_subclass_key *key, char *str)
 | 
			
		||||
{
 | 
			
		||||
| 
						 | 
				
			
			@ -2964,12 +2969,10 @@ static void check_chain_key(struct task_struct *curr)
 | 
			
		|||
#endif
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#if defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING)
 | 
			
		||||
static int mark_lock(struct task_struct *curr, struct held_lock *this,
 | 
			
		||||
		     enum lock_usage_bit new_bit);
 | 
			
		||||
 | 
			
		||||
#if defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
static void print_usage_bug_scenario(struct held_lock *lock)
 | 
			
		||||
{
 | 
			
		||||
	struct lock_class *class = hlock_class(lock);
 | 
			
		||||
| 
						 | 
				
			
			@ -3545,35 +3548,6 @@ static int separate_irq_context(struct task_struct *curr,
 | 
			
		|||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#else /* defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING) */
 | 
			
		||||
 | 
			
		||||
static inline
 | 
			
		||||
int mark_lock_irq(struct task_struct *curr, struct held_lock *this,
 | 
			
		||||
		enum lock_usage_bit new_bit)
 | 
			
		||||
{
 | 
			
		||||
	WARN_ON(1); /* Impossible innit? when we don't have TRACE_IRQFLAG */
 | 
			
		||||
	return 1;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static inline int
 | 
			
		||||
mark_usage(struct task_struct *curr, struct held_lock *hlock, int check)
 | 
			
		||||
{
 | 
			
		||||
	return 1;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static inline unsigned int task_irq_context(struct task_struct *task)
 | 
			
		||||
{
 | 
			
		||||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static inline int separate_irq_context(struct task_struct *curr,
 | 
			
		||||
		struct held_lock *hlock)
 | 
			
		||||
{
 | 
			
		||||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#endif /* defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING) */
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 * Mark a lock with a usage bit, and validate the state transition:
 | 
			
		||||
 */
 | 
			
		||||
| 
						 | 
				
			
			@ -3634,6 +3608,27 @@ static int mark_lock(struct task_struct *curr, struct held_lock *this,
 | 
			
		|||
	return ret;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#else /* defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING) */
 | 
			
		||||
 | 
			
		||||
static inline int
 | 
			
		||||
mark_usage(struct task_struct *curr, struct held_lock *hlock, int check)
 | 
			
		||||
{
 | 
			
		||||
	return 1;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static inline unsigned int task_irq_context(struct task_struct *task)
 | 
			
		||||
{
 | 
			
		||||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static inline int separate_irq_context(struct task_struct *curr,
 | 
			
		||||
		struct held_lock *hlock)
 | 
			
		||||
{
 | 
			
		||||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#endif /* defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_PROVE_LOCKING) */
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 * Initialize a lock instance's lock-class mapping info:
 | 
			
		||||
 */
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue