mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	The kernel perf event creation path shouldn't use find_task_by_vpid() because a vpid exists in a specific namespace. find_task_by_vpid() uses current's pid namespace which isn't always the correct namespace to use for the vpid in all the places perf_event_create_kernel_counter() (and thus find_get_context()) is called. The goal is to clean up pid namespace handling and prevent bugs like: https://bugzilla.kernel.org/show_bug.cgi?id=17281 Instead of using pids switch find_get_context() to use task struct pointers directly. The syscall is responsible for resolving the pid to a task struct. This moves the pid namespace resolution into the syscall much like every other syscall that takes pid parameters. Signed-off-by: Matt Helsley <matthltc@us.ibm.com> Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Robin Green <greenrd@greenrd.org> Cc: Prasad <prasad@linux.vnet.ibm.com> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: Steven Rostedt <rostedt@goodmis.org> Cc: Will Deacon <will.deacon@arm.com> Cc: Mahesh Salgaonkar <mahesh@linux.vnet.ibm.com> LKML-Reference: <a134e5e392ab0204961fd1a62c84a222bf5874a9.1284407763.git.matthltc@us.ibm.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
		
			
				
	
	
		
			426 lines
		
	
	
	
		
			9.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			426 lines
		
	
	
	
		
			9.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/**
 | 
						|
 * @file common.c
 | 
						|
 *
 | 
						|
 * @remark Copyright 2004 Oprofile Authors
 | 
						|
 * @remark Copyright 2010 ARM Ltd.
 | 
						|
 * @remark Read the file COPYING
 | 
						|
 *
 | 
						|
 * @author Zwane Mwaikambo
 | 
						|
 * @author Will Deacon [move to perf]
 | 
						|
 */
 | 
						|
 | 
						|
#include <linux/cpumask.h>
 | 
						|
#include <linux/err.h>
 | 
						|
#include <linux/errno.h>
 | 
						|
#include <linux/init.h>
 | 
						|
#include <linux/mutex.h>
 | 
						|
#include <linux/oprofile.h>
 | 
						|
#include <linux/perf_event.h>
 | 
						|
#include <linux/platform_device.h>
 | 
						|
#include <linux/slab.h>
 | 
						|
#include <asm/stacktrace.h>
 | 
						|
#include <linux/uaccess.h>
 | 
						|
 | 
						|
#include <asm/perf_event.h>
 | 
						|
#include <asm/ptrace.h>
 | 
						|
 | 
						|
#ifdef CONFIG_HW_PERF_EVENTS
 | 
						|
/*
 | 
						|
 * Per performance monitor configuration as set via oprofilefs.
 | 
						|
 */
 | 
						|
struct op_counter_config {
 | 
						|
	unsigned long count;
 | 
						|
	unsigned long enabled;
 | 
						|
	unsigned long event;
 | 
						|
	unsigned long unit_mask;
 | 
						|
	unsigned long kernel;
 | 
						|
	unsigned long user;
 | 
						|
	struct perf_event_attr attr;
 | 
						|
};
 | 
						|
 | 
						|
static int op_arm_enabled;
 | 
						|
static DEFINE_MUTEX(op_arm_mutex);
 | 
						|
 | 
						|
static struct op_counter_config *counter_config;
 | 
						|
static struct perf_event **perf_events[nr_cpumask_bits];
 | 
						|
static int perf_num_counters;
 | 
						|
 | 
						|
/*
 | 
						|
 * Overflow callback for oprofile.
 | 
						|
 */
 | 
						|
static void op_overflow_handler(struct perf_event *event, int unused,
 | 
						|
			struct perf_sample_data *data, struct pt_regs *regs)
 | 
						|
{
 | 
						|
	int id;
 | 
						|
	u32 cpu = smp_processor_id();
 | 
						|
 | 
						|
	for (id = 0; id < perf_num_counters; ++id)
 | 
						|
		if (perf_events[cpu][id] == event)
 | 
						|
			break;
 | 
						|
 | 
						|
	if (id != perf_num_counters)
 | 
						|
		oprofile_add_sample(regs, id);
 | 
						|
	else
 | 
						|
		pr_warning("oprofile: ignoring spurious overflow "
 | 
						|
				"on cpu %u\n", cpu);
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Called by op_arm_setup to create perf attributes to mirror the oprofile
 | 
						|
 * settings in counter_config. Attributes are created as `pinned' events and
 | 
						|
 * so are permanently scheduled on the PMU.
 | 
						|
 */
 | 
						|
static void op_perf_setup(void)
 | 
						|
{
 | 
						|
	int i;
 | 
						|
	u32 size = sizeof(struct perf_event_attr);
 | 
						|
	struct perf_event_attr *attr;
 | 
						|
 | 
						|
	for (i = 0; i < perf_num_counters; ++i) {
 | 
						|
		attr = &counter_config[i].attr;
 | 
						|
		memset(attr, 0, size);
 | 
						|
		attr->type		= PERF_TYPE_RAW;
 | 
						|
		attr->size		= size;
 | 
						|
		attr->config		= counter_config[i].event;
 | 
						|
		attr->sample_period	= counter_config[i].count;
 | 
						|
		attr->pinned		= 1;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static int op_create_counter(int cpu, int event)
 | 
						|
{
 | 
						|
	int ret = 0;
 | 
						|
	struct perf_event *pevent;
 | 
						|
 | 
						|
	if (!counter_config[event].enabled || (perf_events[cpu][event] != NULL))
 | 
						|
		return ret;
 | 
						|
 | 
						|
	pevent = perf_event_create_kernel_counter(&counter_config[event].attr,
 | 
						|
						  cpu, NULL,
 | 
						|
						  op_overflow_handler);
 | 
						|
 | 
						|
	if (IS_ERR(pevent)) {
 | 
						|
		ret = PTR_ERR(pevent);
 | 
						|
	} else if (pevent->state != PERF_EVENT_STATE_ACTIVE) {
 | 
						|
		pr_warning("oprofile: failed to enable event %d "
 | 
						|
				"on CPU %d\n", event, cpu);
 | 
						|
		ret = -EBUSY;
 | 
						|
	} else {
 | 
						|
		perf_events[cpu][event] = pevent;
 | 
						|
	}
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static void op_destroy_counter(int cpu, int event)
 | 
						|
{
 | 
						|
	struct perf_event *pevent = perf_events[cpu][event];
 | 
						|
 | 
						|
	if (pevent) {
 | 
						|
		perf_event_release_kernel(pevent);
 | 
						|
		perf_events[cpu][event] = NULL;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Called by op_arm_start to create active perf events based on the
 | 
						|
 * perviously configured attributes.
 | 
						|
 */
 | 
						|
static int op_perf_start(void)
 | 
						|
{
 | 
						|
	int cpu, event, ret = 0;
 | 
						|
 | 
						|
	for_each_online_cpu(cpu) {
 | 
						|
		for (event = 0; event < perf_num_counters; ++event) {
 | 
						|
			ret = op_create_counter(cpu, event);
 | 
						|
			if (ret)
 | 
						|
				goto out;
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
out:
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Called by op_arm_stop at the end of a profiling run.
 | 
						|
 */
 | 
						|
static void op_perf_stop(void)
 | 
						|
{
 | 
						|
	int cpu, event;
 | 
						|
 | 
						|
	for_each_online_cpu(cpu)
 | 
						|
		for (event = 0; event < perf_num_counters; ++event)
 | 
						|
			op_destroy_counter(cpu, event);
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
static char *op_name_from_perf_id(enum arm_perf_pmu_ids id)
 | 
						|
{
 | 
						|
	switch (id) {
 | 
						|
	case ARM_PERF_PMU_ID_XSCALE1:
 | 
						|
		return "arm/xscale1";
 | 
						|
	case ARM_PERF_PMU_ID_XSCALE2:
 | 
						|
		return "arm/xscale2";
 | 
						|
	case ARM_PERF_PMU_ID_V6:
 | 
						|
		return "arm/armv6";
 | 
						|
	case ARM_PERF_PMU_ID_V6MP:
 | 
						|
		return "arm/mpcore";
 | 
						|
	case ARM_PERF_PMU_ID_CA8:
 | 
						|
		return "arm/armv7";
 | 
						|
	case ARM_PERF_PMU_ID_CA9:
 | 
						|
		return "arm/armv7-ca9";
 | 
						|
	default:
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static int op_arm_create_files(struct super_block *sb, struct dentry *root)
 | 
						|
{
 | 
						|
	unsigned int i;
 | 
						|
 | 
						|
	for (i = 0; i < perf_num_counters; i++) {
 | 
						|
		struct dentry *dir;
 | 
						|
		char buf[4];
 | 
						|
 | 
						|
		snprintf(buf, sizeof buf, "%d", i);
 | 
						|
		dir = oprofilefs_mkdir(sb, root, buf);
 | 
						|
		oprofilefs_create_ulong(sb, dir, "enabled", &counter_config[i].enabled);
 | 
						|
		oprofilefs_create_ulong(sb, dir, "event", &counter_config[i].event);
 | 
						|
		oprofilefs_create_ulong(sb, dir, "count", &counter_config[i].count);
 | 
						|
		oprofilefs_create_ulong(sb, dir, "unit_mask", &counter_config[i].unit_mask);
 | 
						|
		oprofilefs_create_ulong(sb, dir, "kernel", &counter_config[i].kernel);
 | 
						|
		oprofilefs_create_ulong(sb, dir, "user", &counter_config[i].user);
 | 
						|
	}
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int op_arm_setup(void)
 | 
						|
{
 | 
						|
	spin_lock(&oprofilefs_lock);
 | 
						|
	op_perf_setup();
 | 
						|
	spin_unlock(&oprofilefs_lock);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int op_arm_start(void)
 | 
						|
{
 | 
						|
	int ret = -EBUSY;
 | 
						|
 | 
						|
	mutex_lock(&op_arm_mutex);
 | 
						|
	if (!op_arm_enabled) {
 | 
						|
		ret = 0;
 | 
						|
		op_perf_start();
 | 
						|
		op_arm_enabled = 1;
 | 
						|
	}
 | 
						|
	mutex_unlock(&op_arm_mutex);
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static void op_arm_stop(void)
 | 
						|
{
 | 
						|
	mutex_lock(&op_arm_mutex);
 | 
						|
	if (op_arm_enabled)
 | 
						|
		op_perf_stop();
 | 
						|
	op_arm_enabled = 0;
 | 
						|
	mutex_unlock(&op_arm_mutex);
 | 
						|
}
 | 
						|
 | 
						|
#ifdef CONFIG_PM
 | 
						|
static int op_arm_suspend(struct platform_device *dev, pm_message_t state)
 | 
						|
{
 | 
						|
	mutex_lock(&op_arm_mutex);
 | 
						|
	if (op_arm_enabled)
 | 
						|
		op_perf_stop();
 | 
						|
	mutex_unlock(&op_arm_mutex);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int op_arm_resume(struct platform_device *dev)
 | 
						|
{
 | 
						|
	mutex_lock(&op_arm_mutex);
 | 
						|
	if (op_arm_enabled && op_perf_start())
 | 
						|
		op_arm_enabled = 0;
 | 
						|
	mutex_unlock(&op_arm_mutex);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct platform_driver oprofile_driver = {
 | 
						|
	.driver		= {
 | 
						|
		.name		= "arm-oprofile",
 | 
						|
	},
 | 
						|
	.resume		= op_arm_resume,
 | 
						|
	.suspend	= op_arm_suspend,
 | 
						|
};
 | 
						|
 | 
						|
static struct platform_device *oprofile_pdev;
 | 
						|
 | 
						|
static int __init init_driverfs(void)
 | 
						|
{
 | 
						|
	int ret;
 | 
						|
 | 
						|
	ret = platform_driver_register(&oprofile_driver);
 | 
						|
	if (ret)
 | 
						|
		goto out;
 | 
						|
 | 
						|
	oprofile_pdev =	platform_device_register_simple(
 | 
						|
				oprofile_driver.driver.name, 0, NULL, 0);
 | 
						|
	if (IS_ERR(oprofile_pdev)) {
 | 
						|
		ret = PTR_ERR(oprofile_pdev);
 | 
						|
		platform_driver_unregister(&oprofile_driver);
 | 
						|
	}
 | 
						|
 | 
						|
out:
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static void  exit_driverfs(void)
 | 
						|
{
 | 
						|
	platform_device_unregister(oprofile_pdev);
 | 
						|
	platform_driver_unregister(&oprofile_driver);
 | 
						|
}
 | 
						|
#else
 | 
						|
static int __init init_driverfs(void) { return 0; }
 | 
						|
#define exit_driverfs() do { } while (0)
 | 
						|
#endif /* CONFIG_PM */
 | 
						|
 | 
						|
static int report_trace(struct stackframe *frame, void *d)
 | 
						|
{
 | 
						|
	unsigned int *depth = d;
 | 
						|
 | 
						|
	if (*depth) {
 | 
						|
		oprofile_add_trace(frame->pc);
 | 
						|
		(*depth)--;
 | 
						|
	}
 | 
						|
 | 
						|
	return *depth == 0;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * The registers we're interested in are at the end of the variable
 | 
						|
 * length saved register structure. The fp points at the end of this
 | 
						|
 * structure so the address of this struct is:
 | 
						|
 * (struct frame_tail *)(xxx->fp)-1
 | 
						|
 */
 | 
						|
struct frame_tail {
 | 
						|
	struct frame_tail *fp;
 | 
						|
	unsigned long sp;
 | 
						|
	unsigned long lr;
 | 
						|
} __attribute__((packed));
 | 
						|
 | 
						|
static struct frame_tail* user_backtrace(struct frame_tail *tail)
 | 
						|
{
 | 
						|
	struct frame_tail buftail[2];
 | 
						|
 | 
						|
	/* Also check accessibility of one struct frame_tail beyond */
 | 
						|
	if (!access_ok(VERIFY_READ, tail, sizeof(buftail)))
 | 
						|
		return NULL;
 | 
						|
	if (__copy_from_user_inatomic(buftail, tail, sizeof(buftail)))
 | 
						|
		return NULL;
 | 
						|
 | 
						|
	oprofile_add_trace(buftail[0].lr);
 | 
						|
 | 
						|
	/* frame pointers should strictly progress back up the stack
 | 
						|
	 * (towards higher addresses) */
 | 
						|
	if (tail >= buftail[0].fp)
 | 
						|
		return NULL;
 | 
						|
 | 
						|
	return buftail[0].fp-1;
 | 
						|
}
 | 
						|
 | 
						|
static void arm_backtrace(struct pt_regs * const regs, unsigned int depth)
 | 
						|
{
 | 
						|
	struct frame_tail *tail = ((struct frame_tail *) regs->ARM_fp) - 1;
 | 
						|
 | 
						|
	if (!user_mode(regs)) {
 | 
						|
		struct stackframe frame;
 | 
						|
		frame.fp = regs->ARM_fp;
 | 
						|
		frame.sp = regs->ARM_sp;
 | 
						|
		frame.lr = regs->ARM_lr;
 | 
						|
		frame.pc = regs->ARM_pc;
 | 
						|
		walk_stackframe(&frame, report_trace, &depth);
 | 
						|
		return;
 | 
						|
	}
 | 
						|
 | 
						|
	while (depth-- && tail && !((unsigned long) tail & 3))
 | 
						|
		tail = user_backtrace(tail);
 | 
						|
}
 | 
						|
 | 
						|
int __init oprofile_arch_init(struct oprofile_operations *ops)
 | 
						|
{
 | 
						|
	int cpu, ret = 0;
 | 
						|
 | 
						|
	perf_num_counters = armpmu_get_max_events();
 | 
						|
 | 
						|
	counter_config = kcalloc(perf_num_counters,
 | 
						|
			sizeof(struct op_counter_config), GFP_KERNEL);
 | 
						|
 | 
						|
	if (!counter_config) {
 | 
						|
		pr_info("oprofile: failed to allocate %d "
 | 
						|
				"counters\n", perf_num_counters);
 | 
						|
		return -ENOMEM;
 | 
						|
	}
 | 
						|
 | 
						|
	ret = init_driverfs();
 | 
						|
	if (ret) {
 | 
						|
		kfree(counter_config);
 | 
						|
		return ret;
 | 
						|
	}
 | 
						|
 | 
						|
	for_each_possible_cpu(cpu) {
 | 
						|
		perf_events[cpu] = kcalloc(perf_num_counters,
 | 
						|
				sizeof(struct perf_event *), GFP_KERNEL);
 | 
						|
		if (!perf_events[cpu]) {
 | 
						|
			pr_info("oprofile: failed to allocate %d perf events "
 | 
						|
					"for cpu %d\n", perf_num_counters, cpu);
 | 
						|
			while (--cpu >= 0)
 | 
						|
				kfree(perf_events[cpu]);
 | 
						|
			return -ENOMEM;
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	ops->backtrace		= arm_backtrace;
 | 
						|
	ops->create_files	= op_arm_create_files;
 | 
						|
	ops->setup		= op_arm_setup;
 | 
						|
	ops->start		= op_arm_start;
 | 
						|
	ops->stop		= op_arm_stop;
 | 
						|
	ops->shutdown		= op_arm_stop;
 | 
						|
	ops->cpu_type		= op_name_from_perf_id(armpmu_get_pmu_id());
 | 
						|
 | 
						|
	if (!ops->cpu_type)
 | 
						|
		ret = -ENODEV;
 | 
						|
	else
 | 
						|
		pr_info("oprofile: using %s\n", ops->cpu_type);
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
void oprofile_arch_exit(void)
 | 
						|
{
 | 
						|
	int cpu, id;
 | 
						|
	struct perf_event *event;
 | 
						|
 | 
						|
	if (*perf_events) {
 | 
						|
		exit_driverfs();
 | 
						|
		for_each_possible_cpu(cpu) {
 | 
						|
			for (id = 0; id < perf_num_counters; ++id) {
 | 
						|
				event = perf_events[cpu][id];
 | 
						|
				if (event != NULL)
 | 
						|
					perf_event_release_kernel(event);
 | 
						|
			}
 | 
						|
			kfree(perf_events[cpu]);
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	if (counter_config)
 | 
						|
		kfree(counter_config);
 | 
						|
}
 | 
						|
#else
 | 
						|
int __init oprofile_arch_init(struct oprofile_operations *ops)
 | 
						|
{
 | 
						|
	pr_info("oprofile: hardware counters not available\n");
 | 
						|
	return -ENODEV;
 | 
						|
}
 | 
						|
void oprofile_arch_exit(void) {}
 | 
						|
#endif /* CONFIG_HW_PERF_EVENTS */
 |