mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Make sure usercopy hooks from linux/instrumented.h are invoked for copy_from_user_nmi(). This fixes KMSAN false positives reported when dumping opcodes for a stack trace. Link: https://lkml.kernel.org/r/20221102110611.1085175-2-glider@google.com Signed-off-by: Alexander Potapenko <glider@google.com> Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: Kees Cook <keescook@chromium.org> Cc: Borislav Petkov <bp@alien8.de> Cc: Dmitry Vyukov <dvyukov@google.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Marco Elver <elver@google.com> Cc: Masahiro Yamada <masahiroy@kernel.org> Cc: Nick Desaulniers <ndesaulniers@google.com> Cc: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
		
			
				
	
	
		
			55 lines
		
	
	
	
		
			1.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			55 lines
		
	
	
	
		
			1.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * User address space access functions.
 | 
						|
 *
 | 
						|
 *  For licencing details see kernel-base/COPYING
 | 
						|
 */
 | 
						|
 | 
						|
#include <linux/uaccess.h>
 | 
						|
#include <linux/export.h>
 | 
						|
#include <linux/instrumented.h>
 | 
						|
 | 
						|
#include <asm/tlbflush.h>
 | 
						|
 | 
						|
/**
 | 
						|
 * copy_from_user_nmi - NMI safe copy from user
 | 
						|
 * @to:		Pointer to the destination buffer
 | 
						|
 * @from:	Pointer to a user space address of the current task
 | 
						|
 * @n:		Number of bytes to copy
 | 
						|
 *
 | 
						|
 * Returns: The number of not copied bytes. 0 is success, i.e. all bytes copied
 | 
						|
 *
 | 
						|
 * Contrary to other copy_from_user() variants this function can be called
 | 
						|
 * from NMI context. Despite the name it is not restricted to be called
 | 
						|
 * from NMI context. It is safe to be called from any other context as
 | 
						|
 * well. It disables pagefaults across the copy which means a fault will
 | 
						|
 * abort the copy.
 | 
						|
 *
 | 
						|
 * For NMI context invocations this relies on the nested NMI work to allow
 | 
						|
 * atomic faults from the NMI path; the nested NMI paths are careful to
 | 
						|
 * preserve CR2.
 | 
						|
 */
 | 
						|
unsigned long
 | 
						|
copy_from_user_nmi(void *to, const void __user *from, unsigned long n)
 | 
						|
{
 | 
						|
	unsigned long ret;
 | 
						|
 | 
						|
	if (!__access_ok(from, n))
 | 
						|
		return n;
 | 
						|
 | 
						|
	if (!nmi_uaccess_okay())
 | 
						|
		return n;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Even though this function is typically called from NMI/IRQ context
 | 
						|
	 * disable pagefaults so that its behaviour is consistent even when
 | 
						|
	 * called from other contexts.
 | 
						|
	 */
 | 
						|
	pagefault_disable();
 | 
						|
	instrument_copy_from_user_before(to, from, n);
 | 
						|
	ret = raw_copy_from_user(to, from, n);
 | 
						|
	instrument_copy_from_user_after(to, from, n, ret);
 | 
						|
	pagefault_enable();
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
EXPORT_SYMBOL_GPL(copy_from_user_nmi);
 |