forked from mirrors/linux
		
	 c83212d79b
			
		
	
	
		c83212d79b
		
	
	
	
	
		
			
			In sdei_device_freeze(), the input parameter of cpuhp_remove_state() is
passed as 'sdei_entry_point' by mistake. Change it to 'sdei_hp_state'.
Fixes: d2c48b2387 ("firmware: arm_sdei: Fix sleep from invalid context BUG")
Signed-off-by: Xiongfeng Wang <wangxiongfeng2@huawei.com>
Reviewed-by: James Morse <james.morse@arm.com>
Link: https://lore.kernel.org/r/20241016084740.183353-1-wangxiongfeng2@huawei.com
Signed-off-by: Will Deacon <will@kernel.org>
		
	
			
		
			
				
	
	
		
			1116 lines
		
	
	
	
		
			24 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1116 lines
		
	
	
	
		
			24 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| // Copyright (C) 2017 Arm Ltd.
 | |
| #define pr_fmt(fmt) "sdei: " fmt
 | |
| 
 | |
| #include <acpi/ghes.h>
 | |
| #include <linux/acpi.h>
 | |
| #include <linux/arm_sdei.h>
 | |
| #include <linux/arm-smccc.h>
 | |
| #include <linux/atomic.h>
 | |
| #include <linux/bitops.h>
 | |
| #include <linux/compiler.h>
 | |
| #include <linux/cpuhotplug.h>
 | |
| #include <linux/cpu.h>
 | |
| #include <linux/cpu_pm.h>
 | |
| #include <linux/errno.h>
 | |
| #include <linux/hardirq.h>
 | |
| #include <linux/kernel.h>
 | |
| #include <linux/kprobes.h>
 | |
| #include <linux/kvm_host.h>
 | |
| #include <linux/list.h>
 | |
| #include <linux/mutex.h>
 | |
| #include <linux/notifier.h>
 | |
| #include <linux/of.h>
 | |
| #include <linux/of_platform.h>
 | |
| #include <linux/percpu.h>
 | |
| #include <linux/platform_device.h>
 | |
| #include <linux/pm.h>
 | |
| #include <linux/ptrace.h>
 | |
| #include <linux/preempt.h>
 | |
| #include <linux/reboot.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/smp.h>
 | |
| #include <linux/spinlock.h>
 | |
| 
 | |
| /*
 | |
|  * The call to use to reach the firmware.
 | |
|  */
 | |
| static asmlinkage void (*sdei_firmware_call)(unsigned long function_id,
 | |
| 		      unsigned long arg0, unsigned long arg1,
 | |
| 		      unsigned long arg2, unsigned long arg3,
 | |
| 		      unsigned long arg4, struct arm_smccc_res *res);
 | |
| 
 | |
| /* entry point from firmware to arch asm code */
 | |
| static unsigned long sdei_entry_point;
 | |
| 
 | |
| static int sdei_hp_state;
 | |
| 
 | |
| struct sdei_event {
 | |
| 	/* These three are protected by the sdei_list_lock */
 | |
| 	struct list_head	list;
 | |
| 	bool			reregister;
 | |
| 	bool			reenable;
 | |
| 
 | |
| 	u32			event_num;
 | |
| 	u8			type;
 | |
| 	u8			priority;
 | |
| 
 | |
| 	/* This pointer is handed to firmware as the event argument. */
 | |
| 	union {
 | |
| 		/* Shared events */
 | |
| 		struct sdei_registered_event *registered;
 | |
| 
 | |
| 		/* CPU private events */
 | |
| 		struct sdei_registered_event __percpu *private_registered;
 | |
| 	};
 | |
| };
 | |
| 
 | |
| /* Take the mutex for any API call or modification. Take the mutex first. */
 | |
| static DEFINE_MUTEX(sdei_events_lock);
 | |
| 
 | |
| /* and then hold this when modifying the list */
 | |
| static DEFINE_SPINLOCK(sdei_list_lock);
 | |
| static LIST_HEAD(sdei_list);
 | |
| 
 | |
| /* Private events are registered/enabled via IPI passing one of these */
 | |
| struct sdei_crosscall_args {
 | |
| 	struct sdei_event *event;
 | |
| 	atomic_t errors;
 | |
| 	int first_error;
 | |
| };
 | |
| 
 | |
| #define CROSSCALL_INIT(arg, event)		\
 | |
| 	do {					\
 | |
| 		arg.event = event;		\
 | |
| 		arg.first_error = 0;		\
 | |
| 		atomic_set(&arg.errors, 0);	\
 | |
| 	} while (0)
 | |
| 
 | |
| static inline int sdei_do_local_call(smp_call_func_t fn,
 | |
| 				     struct sdei_event *event)
 | |
| {
 | |
| 	struct sdei_crosscall_args arg;
 | |
| 
 | |
| 	CROSSCALL_INIT(arg, event);
 | |
| 	fn(&arg);
 | |
| 
 | |
| 	return arg.first_error;
 | |
| }
 | |
| 
 | |
| static inline int sdei_do_cross_call(smp_call_func_t fn,
 | |
| 				     struct sdei_event *event)
 | |
| {
 | |
| 	struct sdei_crosscall_args arg;
 | |
| 
 | |
| 	CROSSCALL_INIT(arg, event);
 | |
| 	on_each_cpu(fn, &arg, true);
 | |
| 
 | |
| 	return arg.first_error;
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| sdei_cross_call_return(struct sdei_crosscall_args *arg, int err)
 | |
| {
 | |
| 	if (err && (atomic_inc_return(&arg->errors) == 1))
 | |
| 		arg->first_error = err;
 | |
| }
 | |
| 
 | |
| static int sdei_to_linux_errno(unsigned long sdei_err)
 | |
| {
 | |
| 	switch (sdei_err) {
 | |
| 	case SDEI_NOT_SUPPORTED:
 | |
| 		return -EOPNOTSUPP;
 | |
| 	case SDEI_INVALID_PARAMETERS:
 | |
| 		return -EINVAL;
 | |
| 	case SDEI_DENIED:
 | |
| 		return -EPERM;
 | |
| 	case SDEI_PENDING:
 | |
| 		return -EINPROGRESS;
 | |
| 	case SDEI_OUT_OF_RESOURCE:
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int invoke_sdei_fn(unsigned long function_id, unsigned long arg0,
 | |
| 			  unsigned long arg1, unsigned long arg2,
 | |
| 			  unsigned long arg3, unsigned long arg4,
 | |
| 			  u64 *result)
 | |
| {
 | |
| 	int err;
 | |
| 	struct arm_smccc_res res;
 | |
| 
 | |
| 	if (sdei_firmware_call) {
 | |
| 		sdei_firmware_call(function_id, arg0, arg1, arg2, arg3, arg4,
 | |
| 				   &res);
 | |
| 		err = sdei_to_linux_errno(res.a0);
 | |
| 	} else {
 | |
| 		/*
 | |
| 		 * !sdei_firmware_call means we failed to probe or called
 | |
| 		 * sdei_mark_interface_broken(). -EIO is not an error returned
 | |
| 		 * by sdei_to_linux_errno() and is used to suppress messages
 | |
| 		 * from this driver.
 | |
| 		 */
 | |
| 		err = -EIO;
 | |
| 		res.a0 = SDEI_NOT_SUPPORTED;
 | |
| 	}
 | |
| 
 | |
| 	if (result)
 | |
| 		*result = res.a0;
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| NOKPROBE_SYMBOL(invoke_sdei_fn);
 | |
| 
 | |
| static struct sdei_event *sdei_event_find(u32 event_num)
 | |
| {
 | |
| 	struct sdei_event *e, *found = NULL;
 | |
| 
 | |
| 	lockdep_assert_held(&sdei_events_lock);
 | |
| 
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	list_for_each_entry(e, &sdei_list, list) {
 | |
| 		if (e->event_num == event_num) {
 | |
| 			found = e;
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 
 | |
| 	return found;
 | |
| }
 | |
| 
 | |
| int sdei_api_event_context(u32 query, u64 *result)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_CONTEXT, query, 0, 0, 0, 0,
 | |
| 			      result);
 | |
| }
 | |
| NOKPROBE_SYMBOL(sdei_api_event_context);
 | |
| 
 | |
| static int sdei_api_event_get_info(u32 event, u32 info, u64 *result)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_GET_INFO, event, info, 0,
 | |
| 			      0, 0, result);
 | |
| }
 | |
| 
 | |
| static struct sdei_event *sdei_event_create(u32 event_num,
 | |
| 					    sdei_event_callback *cb,
 | |
| 					    void *cb_arg)
 | |
| {
 | |
| 	int err;
 | |
| 	u64 result;
 | |
| 	struct sdei_event *event;
 | |
| 	struct sdei_registered_event *reg;
 | |
| 
 | |
| 	lockdep_assert_held(&sdei_events_lock);
 | |
| 
 | |
| 	event = kzalloc(sizeof(*event), GFP_KERNEL);
 | |
| 	if (!event) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto fail;
 | |
| 	}
 | |
| 
 | |
| 	INIT_LIST_HEAD(&event->list);
 | |
| 	event->event_num = event_num;
 | |
| 
 | |
| 	err = sdei_api_event_get_info(event_num, SDEI_EVENT_INFO_EV_PRIORITY,
 | |
| 				      &result);
 | |
| 	if (err)
 | |
| 		goto fail;
 | |
| 	event->priority = result;
 | |
| 
 | |
| 	err = sdei_api_event_get_info(event_num, SDEI_EVENT_INFO_EV_TYPE,
 | |
| 				      &result);
 | |
| 	if (err)
 | |
| 		goto fail;
 | |
| 	event->type = result;
 | |
| 
 | |
| 	if (event->type == SDEI_EVENT_TYPE_SHARED) {
 | |
| 		reg = kzalloc(sizeof(*reg), GFP_KERNEL);
 | |
| 		if (!reg) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto fail;
 | |
| 		}
 | |
| 
 | |
| 		reg->event_num = event->event_num;
 | |
| 		reg->priority = event->priority;
 | |
| 
 | |
| 		reg->callback = cb;
 | |
| 		reg->callback_arg = cb_arg;
 | |
| 		event->registered = reg;
 | |
| 	} else {
 | |
| 		int cpu;
 | |
| 		struct sdei_registered_event __percpu *regs;
 | |
| 
 | |
| 		regs = alloc_percpu(struct sdei_registered_event);
 | |
| 		if (!regs) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto fail;
 | |
| 		}
 | |
| 
 | |
| 		for_each_possible_cpu(cpu) {
 | |
| 			reg = per_cpu_ptr(regs, cpu);
 | |
| 
 | |
| 			reg->event_num = event->event_num;
 | |
| 			reg->priority = event->priority;
 | |
| 			reg->callback = cb;
 | |
| 			reg->callback_arg = cb_arg;
 | |
| 		}
 | |
| 
 | |
| 		event->private_registered = regs;
 | |
| 	}
 | |
| 
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	list_add(&event->list, &sdei_list);
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 
 | |
| 	return event;
 | |
| 
 | |
| fail:
 | |
| 	kfree(event);
 | |
| 	return ERR_PTR(err);
 | |
| }
 | |
| 
 | |
| static void sdei_event_destroy_llocked(struct sdei_event *event)
 | |
| {
 | |
| 	lockdep_assert_held(&sdei_events_lock);
 | |
| 	lockdep_assert_held(&sdei_list_lock);
 | |
| 
 | |
| 	list_del(&event->list);
 | |
| 
 | |
| 	if (event->type == SDEI_EVENT_TYPE_SHARED)
 | |
| 		kfree(event->registered);
 | |
| 	else
 | |
| 		free_percpu(event->private_registered);
 | |
| 
 | |
| 	kfree(event);
 | |
| }
 | |
| 
 | |
| static void sdei_event_destroy(struct sdei_event *event)
 | |
| {
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	sdei_event_destroy_llocked(event);
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| }
 | |
| 
 | |
| static int sdei_api_get_version(u64 *version)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_VERSION, 0, 0, 0, 0, 0, version);
 | |
| }
 | |
| 
 | |
| int sdei_mask_local_cpu(void)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	err = invoke_sdei_fn(SDEI_1_0_FN_SDEI_PE_MASK, 0, 0, 0, 0, 0, NULL);
 | |
| 	if (err && err != -EIO) {
 | |
| 		pr_warn_once("failed to mask CPU[%u]: %d\n",
 | |
| 			      smp_processor_id(), err);
 | |
| 		return err;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void _ipi_mask_cpu(void *ignored)
 | |
| {
 | |
| 	WARN_ON_ONCE(preemptible());
 | |
| 	sdei_mask_local_cpu();
 | |
| }
 | |
| 
 | |
| int sdei_unmask_local_cpu(void)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	err = invoke_sdei_fn(SDEI_1_0_FN_SDEI_PE_UNMASK, 0, 0, 0, 0, 0, NULL);
 | |
| 	if (err && err != -EIO) {
 | |
| 		pr_warn_once("failed to unmask CPU[%u]: %d\n",
 | |
| 			     smp_processor_id(), err);
 | |
| 		return err;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void _ipi_unmask_cpu(void *ignored)
 | |
| {
 | |
| 	WARN_ON_ONCE(preemptible());
 | |
| 	sdei_unmask_local_cpu();
 | |
| }
 | |
| 
 | |
| static void _ipi_private_reset(void *ignored)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	WARN_ON_ONCE(preemptible());
 | |
| 
 | |
| 	err = invoke_sdei_fn(SDEI_1_0_FN_SDEI_PRIVATE_RESET, 0, 0, 0, 0, 0,
 | |
| 			     NULL);
 | |
| 	if (err && err != -EIO)
 | |
| 		pr_warn_once("failed to reset CPU[%u]: %d\n",
 | |
| 			     smp_processor_id(), err);
 | |
| }
 | |
| 
 | |
| static int sdei_api_shared_reset(void)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_SHARED_RESET, 0, 0, 0, 0, 0,
 | |
| 			      NULL);
 | |
| }
 | |
| 
 | |
| static void sdei_mark_interface_broken(void)
 | |
| {
 | |
| 	pr_err("disabling SDEI firmware interface\n");
 | |
| 	on_each_cpu(&_ipi_mask_cpu, NULL, true);
 | |
| 	sdei_firmware_call = NULL;
 | |
| }
 | |
| 
 | |
| static int sdei_platform_reset(void)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	on_each_cpu(&_ipi_private_reset, NULL, true);
 | |
| 	err = sdei_api_shared_reset();
 | |
| 	if (err) {
 | |
| 		pr_err("Failed to reset platform: %d\n", err);
 | |
| 		sdei_mark_interface_broken();
 | |
| 	}
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_api_event_enable(u32 event_num)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_ENABLE, event_num, 0, 0, 0,
 | |
| 			      0, NULL);
 | |
| }
 | |
| 
 | |
| /* Called directly by the hotplug callbacks */
 | |
| static void _local_event_enable(void *data)
 | |
| {
 | |
| 	int err;
 | |
| 	struct sdei_crosscall_args *arg = data;
 | |
| 
 | |
| 	err = sdei_api_event_enable(arg->event->event_num);
 | |
| 
 | |
| 	sdei_cross_call_return(arg, err);
 | |
| }
 | |
| 
 | |
| int sdei_event_enable(u32 event_num)
 | |
| {
 | |
| 	int err = -EINVAL;
 | |
| 	struct sdei_event *event;
 | |
| 
 | |
| 	mutex_lock(&sdei_events_lock);
 | |
| 	event = sdei_event_find(event_num);
 | |
| 	if (!event) {
 | |
| 		mutex_unlock(&sdei_events_lock);
 | |
| 		return -ENOENT;
 | |
| 	}
 | |
| 
 | |
| 
 | |
| 	cpus_read_lock();
 | |
| 	if (event->type == SDEI_EVENT_TYPE_SHARED)
 | |
| 		err = sdei_api_event_enable(event->event_num);
 | |
| 	else
 | |
| 		err = sdei_do_cross_call(_local_event_enable, event);
 | |
| 
 | |
| 	if (!err) {
 | |
| 		spin_lock(&sdei_list_lock);
 | |
| 		event->reenable = true;
 | |
| 		spin_unlock(&sdei_list_lock);
 | |
| 	}
 | |
| 	cpus_read_unlock();
 | |
| 	mutex_unlock(&sdei_events_lock);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_api_event_disable(u32 event_num)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_DISABLE, event_num, 0, 0,
 | |
| 			      0, 0, NULL);
 | |
| }
 | |
| 
 | |
| static void _ipi_event_disable(void *data)
 | |
| {
 | |
| 	int err;
 | |
| 	struct sdei_crosscall_args *arg = data;
 | |
| 
 | |
| 	err = sdei_api_event_disable(arg->event->event_num);
 | |
| 
 | |
| 	sdei_cross_call_return(arg, err);
 | |
| }
 | |
| 
 | |
| int sdei_event_disable(u32 event_num)
 | |
| {
 | |
| 	int err = -EINVAL;
 | |
| 	struct sdei_event *event;
 | |
| 
 | |
| 	mutex_lock(&sdei_events_lock);
 | |
| 	event = sdei_event_find(event_num);
 | |
| 	if (!event) {
 | |
| 		mutex_unlock(&sdei_events_lock);
 | |
| 		return -ENOENT;
 | |
| 	}
 | |
| 
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	event->reenable = false;
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 
 | |
| 	if (event->type == SDEI_EVENT_TYPE_SHARED)
 | |
| 		err = sdei_api_event_disable(event->event_num);
 | |
| 	else
 | |
| 		err = sdei_do_cross_call(_ipi_event_disable, event);
 | |
| 	mutex_unlock(&sdei_events_lock);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_api_event_unregister(u32 event_num)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_UNREGISTER, event_num, 0,
 | |
| 			      0, 0, 0, NULL);
 | |
| }
 | |
| 
 | |
| /* Called directly by the hotplug callbacks */
 | |
| static void _local_event_unregister(void *data)
 | |
| {
 | |
| 	int err;
 | |
| 	struct sdei_crosscall_args *arg = data;
 | |
| 
 | |
| 	err = sdei_api_event_unregister(arg->event->event_num);
 | |
| 
 | |
| 	sdei_cross_call_return(arg, err);
 | |
| }
 | |
| 
 | |
| int sdei_event_unregister(u32 event_num)
 | |
| {
 | |
| 	int err;
 | |
| 	struct sdei_event *event;
 | |
| 
 | |
| 	WARN_ON(in_nmi());
 | |
| 
 | |
| 	mutex_lock(&sdei_events_lock);
 | |
| 	event = sdei_event_find(event_num);
 | |
| 	if (!event) {
 | |
| 		pr_warn("Event %u not registered\n", event_num);
 | |
| 		err = -ENOENT;
 | |
| 		goto unlock;
 | |
| 	}
 | |
| 
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	event->reregister = false;
 | |
| 	event->reenable = false;
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 
 | |
| 	if (event->type == SDEI_EVENT_TYPE_SHARED)
 | |
| 		err = sdei_api_event_unregister(event->event_num);
 | |
| 	else
 | |
| 		err = sdei_do_cross_call(_local_event_unregister, event);
 | |
| 
 | |
| 	if (err)
 | |
| 		goto unlock;
 | |
| 
 | |
| 	sdei_event_destroy(event);
 | |
| unlock:
 | |
| 	mutex_unlock(&sdei_events_lock);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * unregister events, but don't destroy them as they are re-registered by
 | |
|  * sdei_reregister_shared().
 | |
|  */
 | |
| static int sdei_unregister_shared(void)
 | |
| {
 | |
| 	int err = 0;
 | |
| 	struct sdei_event *event;
 | |
| 
 | |
| 	mutex_lock(&sdei_events_lock);
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	list_for_each_entry(event, &sdei_list, list) {
 | |
| 		if (event->type != SDEI_EVENT_TYPE_SHARED)
 | |
| 			continue;
 | |
| 
 | |
| 		err = sdei_api_event_unregister(event->event_num);
 | |
| 		if (err)
 | |
| 			break;
 | |
| 	}
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 	mutex_unlock(&sdei_events_lock);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_api_event_register(u32 event_num, unsigned long entry_point,
 | |
| 				   void *arg, u64 flags, u64 affinity)
 | |
| {
 | |
| 	return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_REGISTER, event_num,
 | |
| 			      (unsigned long)entry_point, (unsigned long)arg,
 | |
| 			      flags, affinity, NULL);
 | |
| }
 | |
| 
 | |
| /* Called directly by the hotplug callbacks */
 | |
| static void _local_event_register(void *data)
 | |
| {
 | |
| 	int err;
 | |
| 	struct sdei_registered_event *reg;
 | |
| 	struct sdei_crosscall_args *arg = data;
 | |
| 
 | |
| 	reg = per_cpu_ptr(arg->event->private_registered, smp_processor_id());
 | |
| 	err = sdei_api_event_register(arg->event->event_num, sdei_entry_point,
 | |
| 				      reg, 0, 0);
 | |
| 
 | |
| 	sdei_cross_call_return(arg, err);
 | |
| }
 | |
| 
 | |
| int sdei_event_register(u32 event_num, sdei_event_callback *cb, void *arg)
 | |
| {
 | |
| 	int err;
 | |
| 	struct sdei_event *event;
 | |
| 
 | |
| 	WARN_ON(in_nmi());
 | |
| 
 | |
| 	mutex_lock(&sdei_events_lock);
 | |
| 	if (sdei_event_find(event_num)) {
 | |
| 		pr_warn("Event %u already registered\n", event_num);
 | |
| 		err = -EBUSY;
 | |
| 		goto unlock;
 | |
| 	}
 | |
| 
 | |
| 	event = sdei_event_create(event_num, cb, arg);
 | |
| 	if (IS_ERR(event)) {
 | |
| 		err = PTR_ERR(event);
 | |
| 		pr_warn("Failed to create event %u: %d\n", event_num, err);
 | |
| 		goto unlock;
 | |
| 	}
 | |
| 
 | |
| 	cpus_read_lock();
 | |
| 	if (event->type == SDEI_EVENT_TYPE_SHARED) {
 | |
| 		err = sdei_api_event_register(event->event_num,
 | |
| 					      sdei_entry_point,
 | |
| 					      event->registered,
 | |
| 					      SDEI_EVENT_REGISTER_RM_ANY, 0);
 | |
| 	} else {
 | |
| 		err = sdei_do_cross_call(_local_event_register, event);
 | |
| 		if (err)
 | |
| 			sdei_do_cross_call(_local_event_unregister, event);
 | |
| 	}
 | |
| 
 | |
| 	if (err) {
 | |
| 		sdei_event_destroy(event);
 | |
| 		pr_warn("Failed to register event %u: %d\n", event_num, err);
 | |
| 		goto cpu_unlock;
 | |
| 	}
 | |
| 
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	event->reregister = true;
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| cpu_unlock:
 | |
| 	cpus_read_unlock();
 | |
| unlock:
 | |
| 	mutex_unlock(&sdei_events_lock);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_reregister_shared(void)
 | |
| {
 | |
| 	int err = 0;
 | |
| 	struct sdei_event *event;
 | |
| 
 | |
| 	mutex_lock(&sdei_events_lock);
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	list_for_each_entry(event, &sdei_list, list) {
 | |
| 		if (event->type != SDEI_EVENT_TYPE_SHARED)
 | |
| 			continue;
 | |
| 
 | |
| 		if (event->reregister) {
 | |
| 			err = sdei_api_event_register(event->event_num,
 | |
| 					sdei_entry_point, event->registered,
 | |
| 					SDEI_EVENT_REGISTER_RM_ANY, 0);
 | |
| 			if (err) {
 | |
| 				pr_err("Failed to re-register event %u\n",
 | |
| 				       event->event_num);
 | |
| 				sdei_event_destroy_llocked(event);
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		if (event->reenable) {
 | |
| 			err = sdei_api_event_enable(event->event_num);
 | |
| 			if (err) {
 | |
| 				pr_err("Failed to re-enable event %u\n",
 | |
| 				       event->event_num);
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 	mutex_unlock(&sdei_events_lock);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_cpuhp_down(unsigned int cpu)
 | |
| {
 | |
| 	struct sdei_event *event;
 | |
| 	int err;
 | |
| 
 | |
| 	/* un-register private events */
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	list_for_each_entry(event, &sdei_list, list) {
 | |
| 		if (event->type == SDEI_EVENT_TYPE_SHARED)
 | |
| 			continue;
 | |
| 
 | |
| 		err = sdei_do_local_call(_local_event_unregister, event);
 | |
| 		if (err) {
 | |
| 			pr_err("Failed to unregister event %u: %d\n",
 | |
| 			       event->event_num, err);
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 
 | |
| 	return sdei_mask_local_cpu();
 | |
| }
 | |
| 
 | |
| static int sdei_cpuhp_up(unsigned int cpu)
 | |
| {
 | |
| 	struct sdei_event *event;
 | |
| 	int err;
 | |
| 
 | |
| 	/* re-register/enable private events */
 | |
| 	spin_lock(&sdei_list_lock);
 | |
| 	list_for_each_entry(event, &sdei_list, list) {
 | |
| 		if (event->type == SDEI_EVENT_TYPE_SHARED)
 | |
| 			continue;
 | |
| 
 | |
| 		if (event->reregister) {
 | |
| 			err = sdei_do_local_call(_local_event_register, event);
 | |
| 			if (err) {
 | |
| 				pr_err("Failed to re-register event %u: %d\n",
 | |
| 				       event->event_num, err);
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		if (event->reenable) {
 | |
| 			err = sdei_do_local_call(_local_event_enable, event);
 | |
| 			if (err) {
 | |
| 				pr_err("Failed to re-enable event %u: %d\n",
 | |
| 				       event->event_num, err);
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(&sdei_list_lock);
 | |
| 
 | |
| 	return sdei_unmask_local_cpu();
 | |
| }
 | |
| 
 | |
| /* When entering idle, mask/unmask events for this cpu */
 | |
| static int sdei_pm_notifier(struct notifier_block *nb, unsigned long action,
 | |
| 			    void *data)
 | |
| {
 | |
| 	int rv;
 | |
| 
 | |
| 	WARN_ON_ONCE(preemptible());
 | |
| 
 | |
| 	switch (action) {
 | |
| 	case CPU_PM_ENTER:
 | |
| 		rv = sdei_mask_local_cpu();
 | |
| 		break;
 | |
| 	case CPU_PM_EXIT:
 | |
| 	case CPU_PM_ENTER_FAILED:
 | |
| 		rv = sdei_unmask_local_cpu();
 | |
| 		break;
 | |
| 	default:
 | |
| 		return NOTIFY_DONE;
 | |
| 	}
 | |
| 
 | |
| 	if (rv)
 | |
| 		return notifier_from_errno(rv);
 | |
| 
 | |
| 	return NOTIFY_OK;
 | |
| }
 | |
| 
 | |
| static struct notifier_block sdei_pm_nb = {
 | |
| 	.notifier_call = sdei_pm_notifier,
 | |
| };
 | |
| 
 | |
| static int sdei_device_suspend(struct device *dev)
 | |
| {
 | |
| 	on_each_cpu(_ipi_mask_cpu, NULL, true);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int sdei_device_resume(struct device *dev)
 | |
| {
 | |
| 	on_each_cpu(_ipi_unmask_cpu, NULL, true);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * We need all events to be reregistered when we resume from hibernate.
 | |
|  *
 | |
|  * The sequence is freeze->thaw. Reboot. freeze->restore. We unregister
 | |
|  * events during freeze, then re-register and re-enable them during thaw
 | |
|  * and restore.
 | |
|  */
 | |
| static int sdei_device_freeze(struct device *dev)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	/* unregister private events */
 | |
| 	cpuhp_remove_state(sdei_hp_state);
 | |
| 
 | |
| 	err = sdei_unregister_shared();
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int sdei_device_thaw(struct device *dev)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	/* re-register shared events */
 | |
| 	err = sdei_reregister_shared();
 | |
| 	if (err) {
 | |
| 		pr_warn("Failed to re-register shared events...\n");
 | |
| 		sdei_mark_interface_broken();
 | |
| 		return err;
 | |
| 	}
 | |
| 
 | |
| 	err = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "SDEI",
 | |
| 				&sdei_cpuhp_up, &sdei_cpuhp_down);
 | |
| 	if (err < 0) {
 | |
| 		pr_warn("Failed to re-register CPU hotplug notifier...\n");
 | |
| 		return err;
 | |
| 	}
 | |
| 
 | |
| 	sdei_hp_state = err;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int sdei_device_restore(struct device *dev)
 | |
| {
 | |
| 	int err;
 | |
| 
 | |
| 	err = sdei_platform_reset();
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	return sdei_device_thaw(dev);
 | |
| }
 | |
| 
 | |
| static const struct dev_pm_ops sdei_pm_ops = {
 | |
| 	.suspend = sdei_device_suspend,
 | |
| 	.resume = sdei_device_resume,
 | |
| 	.freeze = sdei_device_freeze,
 | |
| 	.thaw = sdei_device_thaw,
 | |
| 	.restore = sdei_device_restore,
 | |
| };
 | |
| 
 | |
| /*
 | |
|  * Mask all CPUs and unregister all events on panic, reboot or kexec.
 | |
|  */
 | |
| static int sdei_reboot_notifier(struct notifier_block *nb, unsigned long action,
 | |
| 				void *data)
 | |
| {
 | |
| 	/*
 | |
| 	 * We are going to reset the interface, after this there is no point
 | |
| 	 * doing work when we take CPUs offline.
 | |
| 	 */
 | |
| 	cpuhp_remove_state(sdei_hp_state);
 | |
| 
 | |
| 	sdei_platform_reset();
 | |
| 
 | |
| 	return NOTIFY_OK;
 | |
| }
 | |
| 
 | |
| static struct notifier_block sdei_reboot_nb = {
 | |
| 	.notifier_call = sdei_reboot_notifier,
 | |
| };
 | |
| 
 | |
| static void sdei_smccc_smc(unsigned long function_id,
 | |
| 			   unsigned long arg0, unsigned long arg1,
 | |
| 			   unsigned long arg2, unsigned long arg3,
 | |
| 			   unsigned long arg4, struct arm_smccc_res *res)
 | |
| {
 | |
| 	arm_smccc_smc(function_id, arg0, arg1, arg2, arg3, arg4, 0, 0, res);
 | |
| }
 | |
| NOKPROBE_SYMBOL(sdei_smccc_smc);
 | |
| 
 | |
| static void sdei_smccc_hvc(unsigned long function_id,
 | |
| 			   unsigned long arg0, unsigned long arg1,
 | |
| 			   unsigned long arg2, unsigned long arg3,
 | |
| 			   unsigned long arg4, struct arm_smccc_res *res)
 | |
| {
 | |
| 	arm_smccc_hvc(function_id, arg0, arg1, arg2, arg3, arg4, 0, 0, res);
 | |
| }
 | |
| NOKPROBE_SYMBOL(sdei_smccc_hvc);
 | |
| 
 | |
| int sdei_register_ghes(struct ghes *ghes, sdei_event_callback *normal_cb,
 | |
| 		       sdei_event_callback *critical_cb)
 | |
| {
 | |
| 	int err;
 | |
| 	u64 result;
 | |
| 	u32 event_num;
 | |
| 	sdei_event_callback *cb;
 | |
| 
 | |
| 	if (!IS_ENABLED(CONFIG_ACPI_APEI_GHES))
 | |
| 		return -EOPNOTSUPP;
 | |
| 
 | |
| 	event_num = ghes->generic->notify.vector;
 | |
| 	if (event_num == 0) {
 | |
| 		/*
 | |
| 		 * Event 0 is reserved by the specification for
 | |
| 		 * SDEI_EVENT_SIGNAL.
 | |
| 		 */
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	err = sdei_api_event_get_info(event_num, SDEI_EVENT_INFO_EV_PRIORITY,
 | |
| 				      &result);
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	if (result == SDEI_EVENT_PRIORITY_CRITICAL)
 | |
| 		cb = critical_cb;
 | |
| 	else
 | |
| 		cb = normal_cb;
 | |
| 
 | |
| 	err = sdei_event_register(event_num, cb, ghes);
 | |
| 	if (!err)
 | |
| 		err = sdei_event_enable(event_num);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| int sdei_unregister_ghes(struct ghes *ghes)
 | |
| {
 | |
| 	int i;
 | |
| 	int err;
 | |
| 	u32 event_num = ghes->generic->notify.vector;
 | |
| 
 | |
| 	might_sleep();
 | |
| 
 | |
| 	if (!IS_ENABLED(CONFIG_ACPI_APEI_GHES))
 | |
| 		return -EOPNOTSUPP;
 | |
| 
 | |
| 	/*
 | |
| 	 * The event may be running on another CPU. Disable it
 | |
| 	 * to stop new events, then try to unregister a few times.
 | |
| 	 */
 | |
| 	err = sdei_event_disable(event_num);
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	for (i = 0; i < 3; i++) {
 | |
| 		err = sdei_event_unregister(event_num);
 | |
| 		if (err != -EINPROGRESS)
 | |
| 			break;
 | |
| 
 | |
| 		schedule();
 | |
| 	}
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int sdei_get_conduit(struct platform_device *pdev)
 | |
| {
 | |
| 	const char *method;
 | |
| 	struct device_node *np = pdev->dev.of_node;
 | |
| 
 | |
| 	sdei_firmware_call = NULL;
 | |
| 	if (np) {
 | |
| 		if (of_property_read_string(np, "method", &method)) {
 | |
| 			pr_warn("missing \"method\" property\n");
 | |
| 			return SMCCC_CONDUIT_NONE;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp("hvc", method)) {
 | |
| 			sdei_firmware_call = &sdei_smccc_hvc;
 | |
| 			return SMCCC_CONDUIT_HVC;
 | |
| 		} else if (!strcmp("smc", method)) {
 | |
| 			sdei_firmware_call = &sdei_smccc_smc;
 | |
| 			return SMCCC_CONDUIT_SMC;
 | |
| 		}
 | |
| 
 | |
| 		pr_warn("invalid \"method\" property: %s\n", method);
 | |
| 	} else if (!acpi_disabled) {
 | |
| 		if (acpi_psci_use_hvc()) {
 | |
| 			sdei_firmware_call = &sdei_smccc_hvc;
 | |
| 			return SMCCC_CONDUIT_HVC;
 | |
| 		} else {
 | |
| 			sdei_firmware_call = &sdei_smccc_smc;
 | |
| 			return SMCCC_CONDUIT_SMC;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return SMCCC_CONDUIT_NONE;
 | |
| }
 | |
| 
 | |
| static int sdei_probe(struct platform_device *pdev)
 | |
| {
 | |
| 	int err;
 | |
| 	u64 ver = 0;
 | |
| 	int conduit;
 | |
| 
 | |
| 	conduit = sdei_get_conduit(pdev);
 | |
| 	if (!sdei_firmware_call)
 | |
| 		return 0;
 | |
| 
 | |
| 	err = sdei_api_get_version(&ver);
 | |
| 	if (err) {
 | |
| 		pr_err("Failed to get SDEI version: %d\n", err);
 | |
| 		sdei_mark_interface_broken();
 | |
| 		return err;
 | |
| 	}
 | |
| 
 | |
| 	pr_info("SDEIv%d.%d (0x%x) detected in firmware.\n",
 | |
| 		(int)SDEI_VERSION_MAJOR(ver), (int)SDEI_VERSION_MINOR(ver),
 | |
| 		(int)SDEI_VERSION_VENDOR(ver));
 | |
| 
 | |
| 	if (SDEI_VERSION_MAJOR(ver) != 1) {
 | |
| 		pr_warn("Conflicting SDEI version detected.\n");
 | |
| 		sdei_mark_interface_broken();
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	err = sdei_platform_reset();
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	sdei_entry_point = sdei_arch_get_entry_point(conduit);
 | |
| 	if (!sdei_entry_point) {
 | |
| 		/* Not supported due to hardware or boot configuration */
 | |
| 		sdei_mark_interface_broken();
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	err = cpu_pm_register_notifier(&sdei_pm_nb);
 | |
| 	if (err) {
 | |
| 		pr_warn("Failed to register CPU PM notifier...\n");
 | |
| 		goto error;
 | |
| 	}
 | |
| 
 | |
| 	err = register_reboot_notifier(&sdei_reboot_nb);
 | |
| 	if (err) {
 | |
| 		pr_warn("Failed to register reboot notifier...\n");
 | |
| 		goto remove_cpupm;
 | |
| 	}
 | |
| 
 | |
| 	err = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "SDEI",
 | |
| 				&sdei_cpuhp_up, &sdei_cpuhp_down);
 | |
| 	if (err < 0) {
 | |
| 		pr_warn("Failed to register CPU hotplug notifier...\n");
 | |
| 		goto remove_reboot;
 | |
| 	}
 | |
| 
 | |
| 	sdei_hp_state = err;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| remove_reboot:
 | |
| 	unregister_reboot_notifier(&sdei_reboot_nb);
 | |
| 
 | |
| remove_cpupm:
 | |
| 	cpu_pm_unregister_notifier(&sdei_pm_nb);
 | |
| 
 | |
| error:
 | |
| 	sdei_mark_interface_broken();
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static const struct of_device_id sdei_of_match[] = {
 | |
| 	{ .compatible = "arm,sdei-1.0" },
 | |
| 	{}
 | |
| };
 | |
| 
 | |
| static struct platform_driver sdei_driver = {
 | |
| 	.driver		= {
 | |
| 		.name			= "sdei",
 | |
| 		.pm			= &sdei_pm_ops,
 | |
| 		.of_match_table		= sdei_of_match,
 | |
| 	},
 | |
| 	.probe		= sdei_probe,
 | |
| };
 | |
| 
 | |
| static bool __init sdei_present_acpi(void)
 | |
| {
 | |
| 	acpi_status status;
 | |
| 	struct acpi_table_header *sdei_table_header;
 | |
| 
 | |
| 	if (acpi_disabled)
 | |
| 		return false;
 | |
| 
 | |
| 	status = acpi_get_table(ACPI_SIG_SDEI, 0, &sdei_table_header);
 | |
| 	if (ACPI_FAILURE(status) && status != AE_NOT_FOUND) {
 | |
| 		const char *msg = acpi_format_exception(status);
 | |
| 
 | |
| 		pr_info("Failed to get ACPI:SDEI table, %s\n", msg);
 | |
| 	}
 | |
| 	if (ACPI_FAILURE(status))
 | |
| 		return false;
 | |
| 
 | |
| 	acpi_put_table(sdei_table_header);
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| void __init sdei_init(void)
 | |
| {
 | |
| 	struct platform_device *pdev;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = platform_driver_register(&sdei_driver);
 | |
| 	if (ret || !sdei_present_acpi())
 | |
| 		return;
 | |
| 
 | |
| 	pdev = platform_device_register_simple(sdei_driver.driver.name,
 | |
| 					       0, NULL, 0);
 | |
| 	if (IS_ERR(pdev)) {
 | |
| 		ret = PTR_ERR(pdev);
 | |
| 		platform_driver_unregister(&sdei_driver);
 | |
| 		pr_info("Failed to register ACPI:SDEI platform device %d\n",
 | |
| 			ret);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| int sdei_event_handler(struct pt_regs *regs,
 | |
| 		       struct sdei_registered_event *arg)
 | |
| {
 | |
| 	int err;
 | |
| 	u32 event_num = arg->event_num;
 | |
| 
 | |
| 	err = arg->callback(event_num, regs, arg->callback_arg);
 | |
| 	if (err)
 | |
| 		pr_err_ratelimited("event %u on CPU %u failed with error: %d\n",
 | |
| 				   event_num, smp_processor_id(), err);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| NOKPROBE_SYMBOL(sdei_event_handler);
 | |
| 
 | |
| void sdei_handler_abort(void)
 | |
| {
 | |
| 	/*
 | |
| 	 * If the crash happened in an SDEI event handler then we need to
 | |
| 	 * finish the handler with the firmware so that we can have working
 | |
| 	 * interrupts in the crash kernel.
 | |
| 	 */
 | |
| 	if (__this_cpu_read(sdei_active_critical_event)) {
 | |
| 	        pr_warn("still in SDEI critical event context, attempting to finish handler.\n");
 | |
| 	        __sdei_handler_abort();
 | |
| 	        __this_cpu_write(sdei_active_critical_event, NULL);
 | |
| 	}
 | |
| 	if (__this_cpu_read(sdei_active_normal_event)) {
 | |
| 	        pr_warn("still in SDEI normal event context, attempting to finish handler.\n");
 | |
| 	        __sdei_handler_abort();
 | |
| 	        __this_cpu_write(sdei_active_normal_event, NULL);
 | |
| 	}
 | |
| }
 |