mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Having a stable wchan means the process must be blocked and for it to stay that way while performing stack unwinding. Suggested-by: Peter Zijlstra <peterz@infradead.org> Signed-off-by: Kees Cook <keescook@chromium.org> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Acked-by: Geert Uytterhoeven <geert@linux-m68k.org> Acked-by: Russell King (Oracle) <rmk+kernel@armlinux.org.uk> [arm] Tested-by: Mark Rutland <mark.rutland@arm.com> [arm64] Link: https://lkml.kernel.org/r/20211008111626.332092234@infradead.org
		
			
				
	
	
		
			274 lines
		
	
	
	
		
			7.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			274 lines
		
	
	
	
		
			7.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
// SPDX-License-Identifier: GPL-2.0-only
 | 
						|
/*
 | 
						|
 *	stacktrace.c : stacktracing APIs needed by rest of kernel
 | 
						|
 *			(wrappers over ARC dwarf based unwinder)
 | 
						|
 *
 | 
						|
 * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com)
 | 
						|
 *
 | 
						|
 *  vineetg: aug 2009
 | 
						|
 *  -Implemented CONFIG_STACKTRACE APIs, primarily save_stack_trace_tsk( )
 | 
						|
 *   for displaying task's kernel mode call stack in /proc/<pid>/stack
 | 
						|
 *  -Iterator based approach to have single copy of unwinding core and APIs
 | 
						|
 *   needing unwinding, implement the logic in iterator regarding:
 | 
						|
 *      = which frame onwards to start capture
 | 
						|
 *      = which frame to stop capturing (wchan)
 | 
						|
 *      = specifics of data structs where trace is saved(CONFIG_STACKTRACE etc)
 | 
						|
 *
 | 
						|
 *  vineetg: March 2009
 | 
						|
 *  -Implemented correct versions of thread_saved_pc() and __get_wchan()
 | 
						|
 *
 | 
						|
 *  rajeshwarr: 2008
 | 
						|
 *  -Initial implementation
 | 
						|
 */
 | 
						|
 | 
						|
#include <linux/ptrace.h>
 | 
						|
#include <linux/export.h>
 | 
						|
#include <linux/stacktrace.h>
 | 
						|
#include <linux/kallsyms.h>
 | 
						|
#include <linux/sched/debug.h>
 | 
						|
 | 
						|
#include <asm/arcregs.h>
 | 
						|
#include <asm/unwind.h>
 | 
						|
#include <asm/switch_to.h>
 | 
						|
 | 
						|
/*-------------------------------------------------------------------------
 | 
						|
 *              Unwinder Iterator
 | 
						|
 *-------------------------------------------------------------------------
 | 
						|
 */
 | 
						|
 | 
						|
#ifdef CONFIG_ARC_DW2_UNWIND
 | 
						|
 | 
						|
static int
 | 
						|
seed_unwind_frame_info(struct task_struct *tsk, struct pt_regs *regs,
 | 
						|
		       struct unwind_frame_info *frame_info)
 | 
						|
{
 | 
						|
	if (regs) {
 | 
						|
		/*
 | 
						|
		 * Asynchronous unwinding of intr/exception
 | 
						|
		 *  - Just uses the pt_regs passed
 | 
						|
		 */
 | 
						|
		frame_info->task = tsk;
 | 
						|
 | 
						|
		frame_info->regs.r27 = regs->fp;
 | 
						|
		frame_info->regs.r28 = regs->sp;
 | 
						|
		frame_info->regs.r31 = regs->blink;
 | 
						|
		frame_info->regs.r63 = regs->ret;
 | 
						|
		frame_info->call_frame = 0;
 | 
						|
	} else if (tsk == NULL || tsk == current) {
 | 
						|
		/*
 | 
						|
		 * synchronous unwinding (e.g. dump_stack)
 | 
						|
		 *  - uses current values of SP and friends
 | 
						|
		 */
 | 
						|
		unsigned long fp, sp, blink, ret;
 | 
						|
		frame_info->task = current;
 | 
						|
 | 
						|
		__asm__ __volatile__(
 | 
						|
			"mov %0,r27\n\t"
 | 
						|
			"mov %1,r28\n\t"
 | 
						|
			"mov %2,r31\n\t"
 | 
						|
			"mov %3,r63\n\t"
 | 
						|
			: "=r"(fp), "=r"(sp), "=r"(blink), "=r"(ret)
 | 
						|
		);
 | 
						|
 | 
						|
		frame_info->regs.r27 = fp;
 | 
						|
		frame_info->regs.r28 = sp;
 | 
						|
		frame_info->regs.r31 = blink;
 | 
						|
		frame_info->regs.r63 = ret;
 | 
						|
		frame_info->call_frame = 0;
 | 
						|
	} else {
 | 
						|
		/*
 | 
						|
		 * Asynchronous unwinding of a likely sleeping task
 | 
						|
		 *  - first ensure it is actually sleeping
 | 
						|
		 *  - if so, it will be in __switch_to, kernel mode SP of task
 | 
						|
		 *    is safe-kept and BLINK at a well known location in there
 | 
						|
		 */
 | 
						|
 | 
						|
		if (task_is_running(tsk))
 | 
						|
			return -1;
 | 
						|
 | 
						|
		frame_info->task = tsk;
 | 
						|
 | 
						|
		frame_info->regs.r27 = TSK_K_FP(tsk);
 | 
						|
		frame_info->regs.r28 = TSK_K_ESP(tsk);
 | 
						|
		frame_info->regs.r31 = TSK_K_BLINK(tsk);
 | 
						|
		frame_info->regs.r63 = (unsigned int)__switch_to;
 | 
						|
 | 
						|
		/* In the prologue of __switch_to, first FP is saved on stack
 | 
						|
		 * and then SP is copied to FP. Dwarf assumes cfa as FP based
 | 
						|
		 * but we didn't save FP. The value retrieved above is FP's
 | 
						|
		 * state in previous frame.
 | 
						|
		 * As a work around for this, we unwind from __switch_to start
 | 
						|
		 * and adjust SP accordingly. The other limitation is that
 | 
						|
		 * __switch_to macro is dwarf rules are not generated for inline
 | 
						|
		 * assembly code
 | 
						|
		 */
 | 
						|
		frame_info->regs.r27 = 0;
 | 
						|
		frame_info->regs.r28 += 60;
 | 
						|
		frame_info->call_frame = 0;
 | 
						|
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
#endif
 | 
						|
 | 
						|
notrace noinline unsigned int
 | 
						|
arc_unwind_core(struct task_struct *tsk, struct pt_regs *regs,
 | 
						|
		int (*consumer_fn) (unsigned int, void *), void *arg)
 | 
						|
{
 | 
						|
#ifdef CONFIG_ARC_DW2_UNWIND
 | 
						|
	int ret = 0, cnt = 0;
 | 
						|
	unsigned int address;
 | 
						|
	struct unwind_frame_info frame_info;
 | 
						|
 | 
						|
	if (seed_unwind_frame_info(tsk, regs, &frame_info))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	while (1) {
 | 
						|
		address = UNW_PC(&frame_info);
 | 
						|
 | 
						|
		if (!address || !__kernel_text_address(address))
 | 
						|
			break;
 | 
						|
 | 
						|
		if (consumer_fn(address, arg) == -1)
 | 
						|
			break;
 | 
						|
 | 
						|
		ret = arc_unwind(&frame_info);
 | 
						|
		if (ret)
 | 
						|
			break;
 | 
						|
 | 
						|
		frame_info.regs.r63 = frame_info.regs.r31;
 | 
						|
 | 
						|
		if (cnt++ > 128) {
 | 
						|
			printk("unwinder looping too long, aborting !\n");
 | 
						|
			return 0;
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return address;		/* return the last address it saw */
 | 
						|
#else
 | 
						|
	/* On ARC, only Dward based unwinder works. fp based backtracing is
 | 
						|
	 * not possible (-fno-omit-frame-pointer) because of the way function
 | 
						|
	 * prologue is setup (callee regs saved and then fp set and not other
 | 
						|
	 * way around
 | 
						|
	 */
 | 
						|
	pr_warn_once("CONFIG_ARC_DW2_UNWIND needs to be enabled\n");
 | 
						|
	return 0;
 | 
						|
 | 
						|
#endif
 | 
						|
}
 | 
						|
 | 
						|
/*-------------------------------------------------------------------------
 | 
						|
 * callbacks called by unwinder iterator to implement kernel APIs
 | 
						|
 *
 | 
						|
 * The callback can return -1 to force the iterator to stop, which by default
 | 
						|
 * keeps going till the bottom-most frame.
 | 
						|
 *-------------------------------------------------------------------------
 | 
						|
 */
 | 
						|
 | 
						|
/* Call-back which plugs into unwinding core to dump the stack in
 | 
						|
 * case of panic/OOPs/BUG etc
 | 
						|
 */
 | 
						|
static int __print_sym(unsigned int address, void *arg)
 | 
						|
{
 | 
						|
	const char *loglvl = arg;
 | 
						|
 | 
						|
	printk("%s  %pS\n", loglvl, (void *)address);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
#ifdef CONFIG_STACKTRACE
 | 
						|
 | 
						|
/* Call-back which plugs into unwinding core to capture the
 | 
						|
 * traces needed by kernel on /proc/<pid>/stack
 | 
						|
 */
 | 
						|
static int __collect_all(unsigned int address, void *arg)
 | 
						|
{
 | 
						|
	struct stack_trace *trace = arg;
 | 
						|
 | 
						|
	if (trace->skip > 0)
 | 
						|
		trace->skip--;
 | 
						|
	else
 | 
						|
		trace->entries[trace->nr_entries++] = address;
 | 
						|
 | 
						|
	if (trace->nr_entries >= trace->max_entries)
 | 
						|
		return -1;
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int __collect_all_but_sched(unsigned int address, void *arg)
 | 
						|
{
 | 
						|
	struct stack_trace *trace = arg;
 | 
						|
 | 
						|
	if (in_sched_functions(address))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	if (trace->skip > 0)
 | 
						|
		trace->skip--;
 | 
						|
	else
 | 
						|
		trace->entries[trace->nr_entries++] = address;
 | 
						|
 | 
						|
	if (trace->nr_entries >= trace->max_entries)
 | 
						|
		return -1;
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
#endif
 | 
						|
 | 
						|
static int __get_first_nonsched(unsigned int address, void *unused)
 | 
						|
{
 | 
						|
	if (in_sched_functions(address))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	return -1;
 | 
						|
}
 | 
						|
 | 
						|
/*-------------------------------------------------------------------------
 | 
						|
 *              APIs expected by various kernel sub-systems
 | 
						|
 *-------------------------------------------------------------------------
 | 
						|
 */
 | 
						|
 | 
						|
noinline void show_stacktrace(struct task_struct *tsk, struct pt_regs *regs,
 | 
						|
			      const char *loglvl)
 | 
						|
{
 | 
						|
	printk("%s\nStack Trace:\n", loglvl);
 | 
						|
	arc_unwind_core(tsk, regs, __print_sym, (void *)loglvl);
 | 
						|
}
 | 
						|
EXPORT_SYMBOL(show_stacktrace);
 | 
						|
 | 
						|
/* Expected by sched Code */
 | 
						|
void show_stack(struct task_struct *tsk, unsigned long *sp, const char *loglvl)
 | 
						|
{
 | 
						|
	show_stacktrace(tsk, NULL, loglvl);
 | 
						|
}
 | 
						|
 | 
						|
/* Another API expected by schedular, shows up in "ps" as Wait Channel
 | 
						|
 * Of course just returning schedule( ) would be pointless so unwind until
 | 
						|
 * the function is not in schedular code
 | 
						|
 */
 | 
						|
unsigned int __get_wchan(struct task_struct *tsk)
 | 
						|
{
 | 
						|
	return arc_unwind_core(tsk, NULL, __get_first_nonsched, NULL);
 | 
						|
}
 | 
						|
 | 
						|
#ifdef CONFIG_STACKTRACE
 | 
						|
 | 
						|
/*
 | 
						|
 * API required by CONFIG_STACKTRACE, CONFIG_LATENCYTOP.
 | 
						|
 * A typical use is when /proc/<pid>/stack is queried by userland
 | 
						|
 */
 | 
						|
void save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace)
 | 
						|
{
 | 
						|
	/* Assumes @tsk is sleeping so unwinds from __switch_to */
 | 
						|
	arc_unwind_core(tsk, NULL, __collect_all_but_sched, trace);
 | 
						|
}
 | 
						|
 | 
						|
void save_stack_trace(struct stack_trace *trace)
 | 
						|
{
 | 
						|
	/* Pass NULL for task so it unwinds the current call frame */
 | 
						|
	arc_unwind_core(NULL, NULL, __collect_all, trace);
 | 
						|
}
 | 
						|
EXPORT_SYMBOL_GPL(save_stack_trace);
 | 
						|
#endif
 |