mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	kmemleak: change some global variables to int
They don't have to be atomic_t, because they are simple boolean toggles. Signed-off-by: Li Zefan <lizefan@huawei.com> Acked-by: Catalin Marinas <catalin.marinas@arm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									5f3bf19aeb
								
							
						
					
					
						commit
						8910ae896c
					
				
					 1 changed files with 40 additions and 40 deletions
				
			
		| 
						 | 
					@ -192,15 +192,15 @@ static struct kmem_cache *object_cache;
 | 
				
			||||||
static struct kmem_cache *scan_area_cache;
 | 
					static struct kmem_cache *scan_area_cache;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* set if tracing memory operations is enabled */
 | 
					/* set if tracing memory operations is enabled */
 | 
				
			||||||
static atomic_t kmemleak_enabled = ATOMIC_INIT(0);
 | 
					static int kmemleak_enabled;
 | 
				
			||||||
/* set in the late_initcall if there were no errors */
 | 
					/* set in the late_initcall if there were no errors */
 | 
				
			||||||
static atomic_t kmemleak_initialized = ATOMIC_INIT(0);
 | 
					static int kmemleak_initialized;
 | 
				
			||||||
/* enables or disables early logging of the memory operations */
 | 
					/* enables or disables early logging of the memory operations */
 | 
				
			||||||
static atomic_t kmemleak_early_log = ATOMIC_INIT(1);
 | 
					static int kmemleak_early_log = 1;
 | 
				
			||||||
/* set if a kmemleak warning was issued */
 | 
					/* set if a kmemleak warning was issued */
 | 
				
			||||||
static atomic_t kmemleak_warning = ATOMIC_INIT(0);
 | 
					static int kmemleak_warning;
 | 
				
			||||||
/* set if a fatal kmemleak error has occurred */
 | 
					/* set if a fatal kmemleak error has occurred */
 | 
				
			||||||
static atomic_t kmemleak_error = ATOMIC_INIT(0);
 | 
					static int kmemleak_error;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* minimum and maximum address that may be valid pointers */
 | 
					/* minimum and maximum address that may be valid pointers */
 | 
				
			||||||
static unsigned long min_addr = ULONG_MAX;
 | 
					static unsigned long min_addr = ULONG_MAX;
 | 
				
			||||||
| 
						 | 
					@ -268,7 +268,7 @@ static void kmemleak_disable(void);
 | 
				
			||||||
#define kmemleak_warn(x...)	do {		\
 | 
					#define kmemleak_warn(x...)	do {		\
 | 
				
			||||||
	pr_warning(x);				\
 | 
						pr_warning(x);				\
 | 
				
			||||||
	dump_stack();				\
 | 
						dump_stack();				\
 | 
				
			||||||
	atomic_set(&kmemleak_warning, 1);	\
 | 
						kmemleak_warning = 1;			\
 | 
				
			||||||
} while (0)
 | 
					} while (0)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
| 
						 | 
					@ -806,7 +806,7 @@ static void __init log_early(int op_type, const void *ptr, size_t size,
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
	struct early_log *log;
 | 
						struct early_log *log;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_error)) {
 | 
						if (kmemleak_error) {
 | 
				
			||||||
		/* kmemleak stopped recording, just count the requests */
 | 
							/* kmemleak stopped recording, just count the requests */
 | 
				
			||||||
		crt_early_log++;
 | 
							crt_early_log++;
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
| 
						 | 
					@ -841,7 +841,7 @@ static void early_alloc(struct early_log *log)
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
	int i;
 | 
						int i;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!atomic_read(&kmemleak_enabled) || !log->ptr || IS_ERR(log->ptr))
 | 
						if (!kmemleak_enabled || !log->ptr || IS_ERR(log->ptr))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
| 
						 | 
					@ -894,9 +894,9 @@ void __ref kmemleak_alloc(const void *ptr, size_t size, int min_count,
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p, %zu, %d)\n", __func__, ptr, size, min_count);
 | 
						pr_debug("%s(0x%p, %zu, %d)\n", __func__, ptr, size, min_count);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		create_object((unsigned long)ptr, size, min_count, gfp);
 | 
							create_object((unsigned long)ptr, size, min_count, gfp);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_ALLOC, ptr, size, min_count);
 | 
							log_early(KMEMLEAK_ALLOC, ptr, size, min_count);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(kmemleak_alloc);
 | 
					EXPORT_SYMBOL_GPL(kmemleak_alloc);
 | 
				
			||||||
| 
						 | 
					@ -920,11 +920,11 @@ void __ref kmemleak_alloc_percpu(const void __percpu *ptr, size_t size)
 | 
				
			||||||
	 * Percpu allocations are only scanned and not reported as leaks
 | 
						 * Percpu allocations are only scanned and not reported as leaks
 | 
				
			||||||
	 * (min_count is set to 0).
 | 
						 * (min_count is set to 0).
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		for_each_possible_cpu(cpu)
 | 
							for_each_possible_cpu(cpu)
 | 
				
			||||||
			create_object((unsigned long)per_cpu_ptr(ptr, cpu),
 | 
								create_object((unsigned long)per_cpu_ptr(ptr, cpu),
 | 
				
			||||||
				      size, 0, GFP_KERNEL);
 | 
									      size, 0, GFP_KERNEL);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_ALLOC_PERCPU, ptr, size, 0);
 | 
							log_early(KMEMLEAK_ALLOC_PERCPU, ptr, size, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(kmemleak_alloc_percpu);
 | 
					EXPORT_SYMBOL_GPL(kmemleak_alloc_percpu);
 | 
				
			||||||
| 
						 | 
					@ -940,9 +940,9 @@ void __ref kmemleak_free(const void *ptr)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		delete_object_full((unsigned long)ptr);
 | 
							delete_object_full((unsigned long)ptr);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_FREE, ptr, 0, 0);
 | 
							log_early(KMEMLEAK_FREE, ptr, 0, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(kmemleak_free);
 | 
					EXPORT_SYMBOL_GPL(kmemleak_free);
 | 
				
			||||||
| 
						 | 
					@ -960,9 +960,9 @@ void __ref kmemleak_free_part(const void *ptr, size_t size)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		delete_object_part((unsigned long)ptr, size);
 | 
							delete_object_part((unsigned long)ptr, size);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_FREE_PART, ptr, size, 0);
 | 
							log_early(KMEMLEAK_FREE_PART, ptr, size, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(kmemleak_free_part);
 | 
					EXPORT_SYMBOL_GPL(kmemleak_free_part);
 | 
				
			||||||
| 
						 | 
					@ -980,11 +980,11 @@ void __ref kmemleak_free_percpu(const void __percpu *ptr)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		for_each_possible_cpu(cpu)
 | 
							for_each_possible_cpu(cpu)
 | 
				
			||||||
			delete_object_full((unsigned long)per_cpu_ptr(ptr,
 | 
								delete_object_full((unsigned long)per_cpu_ptr(ptr,
 | 
				
			||||||
								      cpu));
 | 
													      cpu));
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_FREE_PERCPU, ptr, 0, 0);
 | 
							log_early(KMEMLEAK_FREE_PERCPU, ptr, 0, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(kmemleak_free_percpu);
 | 
					EXPORT_SYMBOL_GPL(kmemleak_free_percpu);
 | 
				
			||||||
| 
						 | 
					@ -1000,9 +1000,9 @@ void __ref kmemleak_not_leak(const void *ptr)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		make_gray_object((unsigned long)ptr);
 | 
							make_gray_object((unsigned long)ptr);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_NOT_LEAK, ptr, 0, 0);
 | 
							log_early(KMEMLEAK_NOT_LEAK, ptr, 0, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(kmemleak_not_leak);
 | 
					EXPORT_SYMBOL(kmemleak_not_leak);
 | 
				
			||||||
| 
						 | 
					@ -1020,9 +1020,9 @@ void __ref kmemleak_ignore(const void *ptr)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		make_black_object((unsigned long)ptr);
 | 
							make_black_object((unsigned long)ptr);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_IGNORE, ptr, 0, 0);
 | 
							log_early(KMEMLEAK_IGNORE, ptr, 0, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(kmemleak_ignore);
 | 
					EXPORT_SYMBOL(kmemleak_ignore);
 | 
				
			||||||
| 
						 | 
					@ -1042,9 +1042,9 @@ void __ref kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && size && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && size && !IS_ERR(ptr))
 | 
				
			||||||
		add_scan_area((unsigned long)ptr, size, gfp);
 | 
							add_scan_area((unsigned long)ptr, size, gfp);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_SCAN_AREA, ptr, size, 0);
 | 
							log_early(KMEMLEAK_SCAN_AREA, ptr, size, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(kmemleak_scan_area);
 | 
					EXPORT_SYMBOL(kmemleak_scan_area);
 | 
				
			||||||
| 
						 | 
					@ -1062,9 +1062,9 @@ void __ref kmemleak_no_scan(const void *ptr)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
						pr_debug("%s(0x%p)\n", __func__, ptr);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr))
 | 
						if (kmemleak_enabled && ptr && !IS_ERR(ptr))
 | 
				
			||||||
		object_no_scan((unsigned long)ptr);
 | 
							object_no_scan((unsigned long)ptr);
 | 
				
			||||||
	else if (atomic_read(&kmemleak_early_log))
 | 
						else if (kmemleak_early_log)
 | 
				
			||||||
		log_early(KMEMLEAK_NO_SCAN, ptr, 0, 0);
 | 
							log_early(KMEMLEAK_NO_SCAN, ptr, 0, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(kmemleak_no_scan);
 | 
					EXPORT_SYMBOL(kmemleak_no_scan);
 | 
				
			||||||
| 
						 | 
					@ -1089,7 +1089,7 @@ static bool update_checksum(struct kmemleak_object *object)
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
static int scan_should_stop(void)
 | 
					static int scan_should_stop(void)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (!atomic_read(&kmemleak_enabled))
 | 
						if (!kmemleak_enabled)
 | 
				
			||||||
		return 1;
 | 
							return 1;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
| 
						 | 
					@ -1630,14 +1630,14 @@ static ssize_t kmemleak_write(struct file *file, const char __user *user_buf,
 | 
				
			||||||
		return ret;
 | 
							return ret;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (strncmp(buf, "clear", 5) == 0) {
 | 
						if (strncmp(buf, "clear", 5) == 0) {
 | 
				
			||||||
		if (atomic_read(&kmemleak_enabled))
 | 
							if (kmemleak_enabled)
 | 
				
			||||||
			kmemleak_clear();
 | 
								kmemleak_clear();
 | 
				
			||||||
		else
 | 
							else
 | 
				
			||||||
			__kmemleak_do_cleanup();
 | 
								__kmemleak_do_cleanup();
 | 
				
			||||||
		goto out;
 | 
							goto out;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!atomic_read(&kmemleak_enabled)) {
 | 
						if (!kmemleak_enabled) {
 | 
				
			||||||
		ret = -EBUSY;
 | 
							ret = -EBUSY;
 | 
				
			||||||
		goto out;
 | 
							goto out;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					@ -1726,14 +1726,14 @@ static DECLARE_WORK(cleanup_work, kmemleak_do_cleanup);
 | 
				
			||||||
static void kmemleak_disable(void)
 | 
					static void kmemleak_disable(void)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	/* atomically check whether it was already invoked */
 | 
						/* atomically check whether it was already invoked */
 | 
				
			||||||
	if (atomic_cmpxchg(&kmemleak_error, 0, 1))
 | 
						if (cmpxchg(&kmemleak_error, 0, 1))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* stop any memory operation tracing */
 | 
						/* stop any memory operation tracing */
 | 
				
			||||||
	atomic_set(&kmemleak_enabled, 0);
 | 
						kmemleak_enabled = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* check whether it is too early for a kernel thread */
 | 
						/* check whether it is too early for a kernel thread */
 | 
				
			||||||
	if (atomic_read(&kmemleak_initialized))
 | 
						if (kmemleak_initialized)
 | 
				
			||||||
		schedule_work(&cleanup_work);
 | 
							schedule_work(&cleanup_work);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	pr_info("Kernel memory leak detector disabled\n");
 | 
						pr_info("Kernel memory leak detector disabled\n");
 | 
				
			||||||
| 
						 | 
					@ -1775,9 +1775,10 @@ void __init kmemleak_init(void)
 | 
				
			||||||
	int i;
 | 
						int i;
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						kmemleak_early_log = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_DEBUG_KMEMLEAK_DEFAULT_OFF
 | 
					#ifdef CONFIG_DEBUG_KMEMLEAK_DEFAULT_OFF
 | 
				
			||||||
	if (!kmemleak_skip_disable) {
 | 
						if (!kmemleak_skip_disable) {
 | 
				
			||||||
		atomic_set(&kmemleak_early_log, 0);
 | 
					 | 
				
			||||||
		kmemleak_disable();
 | 
							kmemleak_disable();
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					@ -1795,12 +1796,11 @@ void __init kmemleak_init(void)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* the kernel is still in UP mode, so disabling the IRQs is enough */
 | 
						/* the kernel is still in UP mode, so disabling the IRQs is enough */
 | 
				
			||||||
	local_irq_save(flags);
 | 
						local_irq_save(flags);
 | 
				
			||||||
	atomic_set(&kmemleak_early_log, 0);
 | 
						if (kmemleak_error) {
 | 
				
			||||||
	if (atomic_read(&kmemleak_error)) {
 | 
					 | 
				
			||||||
		local_irq_restore(flags);
 | 
							local_irq_restore(flags);
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
	} else
 | 
						} else
 | 
				
			||||||
		atomic_set(&kmemleak_enabled, 1);
 | 
							kmemleak_enabled = 1;
 | 
				
			||||||
	local_irq_restore(flags);
 | 
						local_irq_restore(flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
| 
						 | 
					@ -1844,9 +1844,9 @@ void __init kmemleak_init(void)
 | 
				
			||||||
				      log->op_type);
 | 
									      log->op_type);
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (atomic_read(&kmemleak_warning)) {
 | 
							if (kmemleak_warning) {
 | 
				
			||||||
			print_log_trace(log);
 | 
								print_log_trace(log);
 | 
				
			||||||
			atomic_set(&kmemleak_warning, 0);
 | 
								kmemleak_warning = 0;
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -1858,9 +1858,9 @@ static int __init kmemleak_late_init(void)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct dentry *dentry;
 | 
						struct dentry *dentry;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	atomic_set(&kmemleak_initialized, 1);
 | 
						kmemleak_initialized = 1;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (atomic_read(&kmemleak_error)) {
 | 
						if (kmemleak_error) {
 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
		 * Some error occurred and kmemleak was disabled. There is a
 | 
							 * Some error occurred and kmemleak was disabled. There is a
 | 
				
			||||||
		 * small chance that kmemleak_disable() was called immediately
 | 
							 * small chance that kmemleak_disable() was called immediately
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue