forked from mirrors/linux
		
	 0f613bfa82
			
		
	
	
		0f613bfa82
		
	
	
	
	
		
			
			Now that we have raw_atomic*_<op>() definitions, there's no need to use arch_atomic*_<op>() definitions outside of the low-level atomic definitions. Move treewide users of arch_atomic*_<op>() over to the equivalent raw_atomic*_<op>(). There should be no functional change as a result of this patch. Signed-off-by: Mark Rutland <mark.rutland@arm.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Reviewed-by: Kees Cook <keescook@chromium.org> Link: https://lore.kernel.org/r/20230605070124.3741859-19-mark.rutland@arm.com
		
			
				
	
	
		
			166 lines
		
	
	
	
		
			4.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			166 lines
		
	
	
	
		
			4.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* SPDX-License-Identifier: GPL-2.0 */
 | |
| #ifndef _LINUX_CONTEXT_TRACKING_H
 | |
| #define _LINUX_CONTEXT_TRACKING_H
 | |
| 
 | |
| #include <linux/sched.h>
 | |
| #include <linux/vtime.h>
 | |
| #include <linux/context_tracking_state.h>
 | |
| #include <linux/instrumentation.h>
 | |
| 
 | |
| #include <asm/ptrace.h>
 | |
| 
 | |
| 
 | |
| #ifdef CONFIG_CONTEXT_TRACKING_USER
 | |
| extern void ct_cpu_track_user(int cpu);
 | |
| 
 | |
| /* Called with interrupts disabled.  */
 | |
| extern void __ct_user_enter(enum ctx_state state);
 | |
| extern void __ct_user_exit(enum ctx_state state);
 | |
| 
 | |
| extern void ct_user_enter(enum ctx_state state);
 | |
| extern void ct_user_exit(enum ctx_state state);
 | |
| 
 | |
| extern void user_enter_callable(void);
 | |
| extern void user_exit_callable(void);
 | |
| 
 | |
| static inline void user_enter(void)
 | |
| {
 | |
| 	if (context_tracking_enabled())
 | |
| 		ct_user_enter(CONTEXT_USER);
 | |
| 
 | |
| }
 | |
| static inline void user_exit(void)
 | |
| {
 | |
| 	if (context_tracking_enabled())
 | |
| 		ct_user_exit(CONTEXT_USER);
 | |
| }
 | |
| 
 | |
| /* Called with interrupts disabled.  */
 | |
| static __always_inline void user_enter_irqoff(void)
 | |
| {
 | |
| 	if (context_tracking_enabled())
 | |
| 		__ct_user_enter(CONTEXT_USER);
 | |
| 
 | |
| }
 | |
| static __always_inline void user_exit_irqoff(void)
 | |
| {
 | |
| 	if (context_tracking_enabled())
 | |
| 		__ct_user_exit(CONTEXT_USER);
 | |
| }
 | |
| 
 | |
| static inline enum ctx_state exception_enter(void)
 | |
| {
 | |
| 	enum ctx_state prev_ctx;
 | |
| 
 | |
| 	if (IS_ENABLED(CONFIG_HAVE_CONTEXT_TRACKING_USER_OFFSTACK) ||
 | |
| 	    !context_tracking_enabled())
 | |
| 		return 0;
 | |
| 
 | |
| 	prev_ctx = __ct_state();
 | |
| 	if (prev_ctx != CONTEXT_KERNEL)
 | |
| 		ct_user_exit(prev_ctx);
 | |
| 
 | |
| 	return prev_ctx;
 | |
| }
 | |
| 
 | |
| static inline void exception_exit(enum ctx_state prev_ctx)
 | |
| {
 | |
| 	if (!IS_ENABLED(CONFIG_HAVE_CONTEXT_TRACKING_USER_OFFSTACK) &&
 | |
| 	    context_tracking_enabled()) {
 | |
| 		if (prev_ctx != CONTEXT_KERNEL)
 | |
| 			ct_user_enter(prev_ctx);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static __always_inline bool context_tracking_guest_enter(void)
 | |
| {
 | |
| 	if (context_tracking_enabled())
 | |
| 		__ct_user_enter(CONTEXT_GUEST);
 | |
| 
 | |
| 	return context_tracking_enabled_this_cpu();
 | |
| }
 | |
| 
 | |
| static __always_inline void context_tracking_guest_exit(void)
 | |
| {
 | |
| 	if (context_tracking_enabled())
 | |
| 		__ct_user_exit(CONTEXT_GUEST);
 | |
| }
 | |
| 
 | |
| #define CT_WARN_ON(cond) WARN_ON(context_tracking_enabled() && (cond))
 | |
| 
 | |
| #else
 | |
| static inline void user_enter(void) { }
 | |
| static inline void user_exit(void) { }
 | |
| static inline void user_enter_irqoff(void) { }
 | |
| static inline void user_exit_irqoff(void) { }
 | |
| static inline int exception_enter(void) { return 0; }
 | |
| static inline void exception_exit(enum ctx_state prev_ctx) { }
 | |
| static inline int ct_state(void) { return -1; }
 | |
| static inline int __ct_state(void) { return -1; }
 | |
| static __always_inline bool context_tracking_guest_enter(void) { return false; }
 | |
| static __always_inline void context_tracking_guest_exit(void) { }
 | |
| #define CT_WARN_ON(cond) do { } while (0)
 | |
| #endif /* !CONFIG_CONTEXT_TRACKING_USER */
 | |
| 
 | |
| #ifdef CONFIG_CONTEXT_TRACKING_USER_FORCE
 | |
| extern void context_tracking_init(void);
 | |
| #else
 | |
| static inline void context_tracking_init(void) { }
 | |
| #endif /* CONFIG_CONTEXT_TRACKING_USER_FORCE */
 | |
| 
 | |
| #ifdef CONFIG_CONTEXT_TRACKING_IDLE
 | |
| extern void ct_idle_enter(void);
 | |
| extern void ct_idle_exit(void);
 | |
| 
 | |
| /*
 | |
|  * Is the current CPU in an extended quiescent state?
 | |
|  *
 | |
|  * No ordering, as we are sampling CPU-local information.
 | |
|  */
 | |
| static __always_inline bool rcu_dynticks_curr_cpu_in_eqs(void)
 | |
| {
 | |
| 	return !(raw_atomic_read(this_cpu_ptr(&context_tracking.state)) & RCU_DYNTICKS_IDX);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Increment the current CPU's context_tracking structure's ->state field
 | |
|  * with ordering.  Return the new value.
 | |
|  */
 | |
| static __always_inline unsigned long ct_state_inc(int incby)
 | |
| {
 | |
| 	return raw_atomic_add_return(incby, this_cpu_ptr(&context_tracking.state));
 | |
| }
 | |
| 
 | |
| static __always_inline bool warn_rcu_enter(void)
 | |
| {
 | |
| 	bool ret = false;
 | |
| 
 | |
| 	/*
 | |
| 	 * Horrible hack to shut up recursive RCU isn't watching fail since
 | |
| 	 * lots of the actual reporting also relies on RCU.
 | |
| 	 */
 | |
| 	preempt_disable_notrace();
 | |
| 	if (rcu_dynticks_curr_cpu_in_eqs()) {
 | |
| 		ret = true;
 | |
| 		ct_state_inc(RCU_DYNTICKS_IDX);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static __always_inline void warn_rcu_exit(bool rcu)
 | |
| {
 | |
| 	if (rcu)
 | |
| 		ct_state_inc(RCU_DYNTICKS_IDX);
 | |
| 	preempt_enable_notrace();
 | |
| }
 | |
| 
 | |
| #else
 | |
| static inline void ct_idle_enter(void) { }
 | |
| static inline void ct_idle_exit(void) { }
 | |
| 
 | |
| static __always_inline bool warn_rcu_enter(void) { return false; }
 | |
| static __always_inline void warn_rcu_exit(bool rcu) { }
 | |
| #endif /* !CONFIG_CONTEXT_TRACKING_IDLE */
 | |
| 
 | |
| #endif
 |