mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 16:48:26 +02:00 
			
		
		
		
	 dcd5c1662d
			
		
	
	
		dcd5c1662d
		
	
	
	
	
		
			
			perf_arch_fetch_caller_regs() is exported for the overriden x86 version, but not for the generic weak version. As a general rule, weak functions should not have their symbol exported in the same file they are defined. So let's export it on trace_event_perf.c as it is used by trace events only. This fixes: ERROR: ".perf_arch_fetch_caller_regs" [fs/xfs/xfs.ko] undefined! ERROR: ".perf_arch_fetch_caller_regs" [arch/powerpc/platforms/cell/spufs/spufs.ko] undefined! -v2: And also only build it if trace events are enabled. -v3: Fix changelog mistake Reported-by: Stephen Rothwell <sfr@canb.auug.org.au> Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com> Cc: Paul Mackerras <paulus@samba.org> LKML-Reference: <1268697902-9518-1-git-send-regression-fweisbec@gmail.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
		
			
				
	
	
		
			168 lines
		
	
	
	
		
			3.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			168 lines
		
	
	
	
		
			3.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * trace event based perf event profiling/tracing
 | |
|  *
 | |
|  * Copyright (C) 2009 Red Hat Inc, Peter Zijlstra <pzijlstr@redhat.com>
 | |
|  * Copyright (C) 2009-2010 Frederic Weisbecker <fweisbec@gmail.com>
 | |
|  */
 | |
| 
 | |
| #include <linux/module.h>
 | |
| #include <linux/kprobes.h>
 | |
| #include "trace.h"
 | |
| 
 | |
| DEFINE_PER_CPU(struct pt_regs, perf_trace_regs);
 | |
| EXPORT_PER_CPU_SYMBOL_GPL(perf_trace_regs);
 | |
| 
 | |
| EXPORT_SYMBOL_GPL(perf_arch_fetch_caller_regs);
 | |
| 
 | |
| static char *perf_trace_buf;
 | |
| static char *perf_trace_buf_nmi;
 | |
| 
 | |
| typedef typeof(char [PERF_MAX_TRACE_SIZE]) perf_trace_t ;
 | |
| 
 | |
| /* Count the events in use (per event id, not per instance) */
 | |
| static int	total_ref_count;
 | |
| 
 | |
| static int perf_trace_event_enable(struct ftrace_event_call *event)
 | |
| {
 | |
| 	char *buf;
 | |
| 	int ret = -ENOMEM;
 | |
| 
 | |
| 	if (event->perf_refcount++ > 0)
 | |
| 		return 0;
 | |
| 
 | |
| 	if (!total_ref_count) {
 | |
| 		buf = (char *)alloc_percpu(perf_trace_t);
 | |
| 		if (!buf)
 | |
| 			goto fail_buf;
 | |
| 
 | |
| 		rcu_assign_pointer(perf_trace_buf, buf);
 | |
| 
 | |
| 		buf = (char *)alloc_percpu(perf_trace_t);
 | |
| 		if (!buf)
 | |
| 			goto fail_buf_nmi;
 | |
| 
 | |
| 		rcu_assign_pointer(perf_trace_buf_nmi, buf);
 | |
| 	}
 | |
| 
 | |
| 	ret = event->perf_event_enable(event);
 | |
| 	if (!ret) {
 | |
| 		total_ref_count++;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| fail_buf_nmi:
 | |
| 	if (!total_ref_count) {
 | |
| 		free_percpu(perf_trace_buf_nmi);
 | |
| 		free_percpu(perf_trace_buf);
 | |
| 		perf_trace_buf_nmi = NULL;
 | |
| 		perf_trace_buf = NULL;
 | |
| 	}
 | |
| fail_buf:
 | |
| 	event->perf_refcount--;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int perf_trace_enable(int event_id)
 | |
| {
 | |
| 	struct ftrace_event_call *event;
 | |
| 	int ret = -EINVAL;
 | |
| 
 | |
| 	mutex_lock(&event_mutex);
 | |
| 	list_for_each_entry(event, &ftrace_events, list) {
 | |
| 		if (event->id == event_id && event->perf_event_enable &&
 | |
| 		    try_module_get(event->mod)) {
 | |
| 			ret = perf_trace_event_enable(event);
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	mutex_unlock(&event_mutex);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void perf_trace_event_disable(struct ftrace_event_call *event)
 | |
| {
 | |
| 	char *buf, *nmi_buf;
 | |
| 
 | |
| 	if (--event->perf_refcount > 0)
 | |
| 		return;
 | |
| 
 | |
| 	event->perf_event_disable(event);
 | |
| 
 | |
| 	if (!--total_ref_count) {
 | |
| 		buf = perf_trace_buf;
 | |
| 		rcu_assign_pointer(perf_trace_buf, NULL);
 | |
| 
 | |
| 		nmi_buf = perf_trace_buf_nmi;
 | |
| 		rcu_assign_pointer(perf_trace_buf_nmi, NULL);
 | |
| 
 | |
| 		/*
 | |
| 		 * Ensure every events in profiling have finished before
 | |
| 		 * releasing the buffers
 | |
| 		 */
 | |
| 		synchronize_sched();
 | |
| 
 | |
| 		free_percpu(buf);
 | |
| 		free_percpu(nmi_buf);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void perf_trace_disable(int event_id)
 | |
| {
 | |
| 	struct ftrace_event_call *event;
 | |
| 
 | |
| 	mutex_lock(&event_mutex);
 | |
| 	list_for_each_entry(event, &ftrace_events, list) {
 | |
| 		if (event->id == event_id) {
 | |
| 			perf_trace_event_disable(event);
 | |
| 			module_put(event->mod);
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	mutex_unlock(&event_mutex);
 | |
| }
 | |
| 
 | |
| __kprobes void *perf_trace_buf_prepare(int size, unsigned short type,
 | |
| 				       int *rctxp, unsigned long *irq_flags)
 | |
| {
 | |
| 	struct trace_entry *entry;
 | |
| 	char *trace_buf, *raw_data;
 | |
| 	int pc, cpu;
 | |
| 
 | |
| 	pc = preempt_count();
 | |
| 
 | |
| 	/* Protect the per cpu buffer, begin the rcu read side */
 | |
| 	local_irq_save(*irq_flags);
 | |
| 
 | |
| 	*rctxp = perf_swevent_get_recursion_context();
 | |
| 	if (*rctxp < 0)
 | |
| 		goto err_recursion;
 | |
| 
 | |
| 	cpu = smp_processor_id();
 | |
| 
 | |
| 	if (in_nmi())
 | |
| 		trace_buf = rcu_dereference(perf_trace_buf_nmi);
 | |
| 	else
 | |
| 		trace_buf = rcu_dereference(perf_trace_buf);
 | |
| 
 | |
| 	if (!trace_buf)
 | |
| 		goto err;
 | |
| 
 | |
| 	raw_data = per_cpu_ptr(trace_buf, cpu);
 | |
| 
 | |
| 	/* zero the dead bytes from align to not leak stack to user */
 | |
| 	*(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL;
 | |
| 
 | |
| 	entry = (struct trace_entry *)raw_data;
 | |
| 	tracing_generic_entry_update(entry, *irq_flags, pc);
 | |
| 	entry->type = type;
 | |
| 
 | |
| 	return raw_data;
 | |
| err:
 | |
| 	perf_swevent_put_recursion_context(*rctxp);
 | |
| err_recursion:
 | |
| 	local_irq_restore(*irq_flags);
 | |
| 	return NULL;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(perf_trace_buf_prepare);
 |