mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 00:28:52 +02:00 
			
		
		
		
	
		
			
				
	
	
		
			1857 lines
		
	
	
	
		
			56 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1857 lines
		
	
	
	
		
			56 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-only
 | |
| // Copyright (C) 2022 Linutronix GmbH, John Ogness
 | |
| // Copyright (C) 2022 Intel, Thomas Gleixner
 | |
| 
 | |
| #include <linux/atomic.h>
 | |
| #include <linux/bug.h>
 | |
| #include <linux/console.h>
 | |
| #include <linux/delay.h>
 | |
| #include <linux/errno.h>
 | |
| #include <linux/export.h>
 | |
| #include <linux/init.h>
 | |
| #include <linux/irqflags.h>
 | |
| #include <linux/kthread.h>
 | |
| #include <linux/minmax.h>
 | |
| #include <linux/percpu.h>
 | |
| #include <linux/preempt.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/smp.h>
 | |
| #include <linux/stddef.h>
 | |
| #include <linux/string.h>
 | |
| #include <linux/types.h>
 | |
| #include "internal.h"
 | |
| #include "printk_ringbuffer.h"
 | |
| /*
 | |
|  * Printk console printing implementation for consoles which does not depend
 | |
|  * on the legacy style console_lock mechanism.
 | |
|  *
 | |
|  * The state of the console is maintained in the "nbcon_state" atomic
 | |
|  * variable.
 | |
|  *
 | |
|  * The console is locked when:
 | |
|  *
 | |
|  *   - The 'prio' field contains the priority of the context that owns the
 | |
|  *     console. Only higher priority contexts are allowed to take over the
 | |
|  *     lock. A value of 0 (NBCON_PRIO_NONE) means the console is not locked.
 | |
|  *
 | |
|  *   - The 'cpu' field denotes on which CPU the console is locked. It is used
 | |
|  *     to prevent busy waiting on the same CPU. Also it informs the lock owner
 | |
|  *     that it has lost the lock in a more complex scenario when the lock was
 | |
|  *     taken over by a higher priority context, released, and taken on another
 | |
|  *     CPU with the same priority as the interrupted owner.
 | |
|  *
 | |
|  * The acquire mechanism uses a few more fields:
 | |
|  *
 | |
|  *   - The 'req_prio' field is used by the handover approach to make the
 | |
|  *     current owner aware that there is a context with a higher priority
 | |
|  *     waiting for the friendly handover.
 | |
|  *
 | |
|  *   - The 'unsafe' field allows to take over the console in a safe way in the
 | |
|  *     middle of emitting a message. The field is set only when accessing some
 | |
|  *     shared resources or when the console device is manipulated. It can be
 | |
|  *     cleared, for example, after emitting one character when the console
 | |
|  *     device is in a consistent state.
 | |
|  *
 | |
|  *   - The 'unsafe_takeover' field is set when a hostile takeover took the
 | |
|  *     console in an unsafe state. The console will stay in the unsafe state
 | |
|  *     until re-initialized.
 | |
|  *
 | |
|  * The acquire mechanism uses three approaches:
 | |
|  *
 | |
|  *   1) Direct acquire when the console is not owned or is owned by a lower
 | |
|  *      priority context and is in a safe state.
 | |
|  *
 | |
|  *   2) Friendly handover mechanism uses a request/grant handshake. It is used
 | |
|  *      when the current owner has lower priority and the console is in an
 | |
|  *      unsafe state.
 | |
|  *
 | |
|  *      The requesting context:
 | |
|  *
 | |
|  *        a) Sets its priority into the 'req_prio' field.
 | |
|  *
 | |
|  *        b) Waits (with a timeout) for the owning context to unlock the
 | |
|  *           console.
 | |
|  *
 | |
|  *        c) Takes the lock and clears the 'req_prio' field.
 | |
|  *
 | |
|  *      The owning context:
 | |
|  *
 | |
|  *        a) Observes the 'req_prio' field set on exit from the unsafe
 | |
|  *           console state.
 | |
|  *
 | |
|  *        b) Gives up console ownership by clearing the 'prio' field.
 | |
|  *
 | |
|  *   3) Unsafe hostile takeover allows to take over the lock even when the
 | |
|  *      console is an unsafe state. It is used only in panic() by the final
 | |
|  *      attempt to flush consoles in a try and hope mode.
 | |
|  *
 | |
|  *      Note that separate record buffers are used in panic(). As a result,
 | |
|  *      the messages can be read and formatted without any risk even after
 | |
|  *      using the hostile takeover in unsafe state.
 | |
|  *
 | |
|  * The release function simply clears the 'prio' field.
 | |
|  *
 | |
|  * All operations on @console::nbcon_state are atomic cmpxchg based to
 | |
|  * handle concurrency.
 | |
|  *
 | |
|  * The acquire/release functions implement only minimal policies:
 | |
|  *
 | |
|  *   - Preference for higher priority contexts.
 | |
|  *   - Protection of the panic CPU.
 | |
|  *
 | |
|  * All other policy decisions must be made at the call sites:
 | |
|  *
 | |
|  *   - What is marked as an unsafe section.
 | |
|  *   - Whether to spin-wait if there is already an owner and the console is
 | |
|  *     in an unsafe state.
 | |
|  *   - Whether to attempt an unsafe hostile takeover.
 | |
|  *
 | |
|  * The design allows to implement the well known:
 | |
|  *
 | |
|  *     acquire()
 | |
|  *     output_one_printk_record()
 | |
|  *     release()
 | |
|  *
 | |
|  * The output of one printk record might be interrupted with a higher priority
 | |
|  * context. The new owner is supposed to reprint the entire interrupted record
 | |
|  * from scratch.
 | |
|  */
 | |
| 
 | |
| /**
 | |
|  * nbcon_state_set - Helper function to set the console state
 | |
|  * @con:	Console to update
 | |
|  * @new:	The new state to write
 | |
|  *
 | |
|  * Only to be used when the console is not yet or no longer visible in the
 | |
|  * system. Otherwise use nbcon_state_try_cmpxchg().
 | |
|  */
 | |
| static inline void nbcon_state_set(struct console *con, struct nbcon_state *new)
 | |
| {
 | |
| 	atomic_set(&ACCESS_PRIVATE(con, nbcon_state), new->atom);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_state_read - Helper function to read the console state
 | |
|  * @con:	Console to read
 | |
|  * @state:	The state to store the result
 | |
|  */
 | |
| static inline void nbcon_state_read(struct console *con, struct nbcon_state *state)
 | |
| {
 | |
| 	state->atom = atomic_read(&ACCESS_PRIVATE(con, nbcon_state));
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_state_try_cmpxchg() - Helper function for atomic_try_cmpxchg() on console state
 | |
|  * @con:	Console to update
 | |
|  * @cur:	Old/expected state
 | |
|  * @new:	New state
 | |
|  *
 | |
|  * Return: True on success. False on fail and @cur is updated.
 | |
|  */
 | |
| static inline bool nbcon_state_try_cmpxchg(struct console *con, struct nbcon_state *cur,
 | |
| 					   struct nbcon_state *new)
 | |
| {
 | |
| 	return atomic_try_cmpxchg(&ACCESS_PRIVATE(con, nbcon_state), &cur->atom, new->atom);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_seq_read - Read the current console sequence
 | |
|  * @con:	Console to read the sequence of
 | |
|  *
 | |
|  * Return:	Sequence number of the next record to print on @con.
 | |
|  */
 | |
| u64 nbcon_seq_read(struct console *con)
 | |
| {
 | |
| 	unsigned long nbcon_seq = atomic_long_read(&ACCESS_PRIVATE(con, nbcon_seq));
 | |
| 
 | |
| 	return __ulseq_to_u64seq(prb, nbcon_seq);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_seq_force - Force console sequence to a specific value
 | |
|  * @con:	Console to work on
 | |
|  * @seq:	Sequence number value to set
 | |
|  *
 | |
|  * Only to be used during init (before registration) or in extreme situations
 | |
|  * (such as panic with CONSOLE_REPLAY_ALL).
 | |
|  */
 | |
| void nbcon_seq_force(struct console *con, u64 seq)
 | |
| {
 | |
| 	/*
 | |
| 	 * If the specified record no longer exists, the oldest available record
 | |
| 	 * is chosen. This is especially important on 32bit systems because only
 | |
| 	 * the lower 32 bits of the sequence number are stored. The upper 32 bits
 | |
| 	 * are derived from the sequence numbers available in the ringbuffer.
 | |
| 	 */
 | |
| 	u64 valid_seq = max_t(u64, seq, prb_first_valid_seq(prb));
 | |
| 
 | |
| 	atomic_long_set(&ACCESS_PRIVATE(con, nbcon_seq), __u64seq_to_ulseq(valid_seq));
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_seq_try_update - Try to update the console sequence number
 | |
|  * @ctxt:	Pointer to an acquire context that contains
 | |
|  *		all information about the acquire mode
 | |
|  * @new_seq:	The new sequence number to set
 | |
|  *
 | |
|  * @ctxt->seq is updated to the new value of @con::nbcon_seq (expanded to
 | |
|  * the 64bit value). This could be a different value than @new_seq if
 | |
|  * nbcon_seq_force() was used or the current context no longer owns the
 | |
|  * console. In the later case, it will stop printing anyway.
 | |
|  */
 | |
| static void nbcon_seq_try_update(struct nbcon_context *ctxt, u64 new_seq)
 | |
| {
 | |
| 	unsigned long nbcon_seq = __u64seq_to_ulseq(ctxt->seq);
 | |
| 	struct console *con = ctxt->console;
 | |
| 
 | |
| 	if (atomic_long_try_cmpxchg(&ACCESS_PRIVATE(con, nbcon_seq), &nbcon_seq,
 | |
| 				    __u64seq_to_ulseq(new_seq))) {
 | |
| 		ctxt->seq = new_seq;
 | |
| 	} else {
 | |
| 		ctxt->seq = nbcon_seq_read(con);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_try_acquire_direct - Try to acquire directly
 | |
|  * @ctxt:		The context of the caller
 | |
|  * @cur:		The current console state
 | |
|  * @is_reacquire:	This acquire is a reacquire
 | |
|  *
 | |
|  * Acquire the console when it is released. Also acquire the console when
 | |
|  * the current owner has a lower priority and the console is in a safe state.
 | |
|  *
 | |
|  * Return:	0 on success. Otherwise, an error code on failure. Also @cur
 | |
|  *		is updated to the latest state when failed to modify it.
 | |
|  *
 | |
|  * Errors:
 | |
|  *
 | |
|  *	-EPERM:		A panic is in progress and this is neither the panic
 | |
|  *			CPU nor is this a reacquire. Or the current owner or
 | |
|  *			waiter has the same or higher priority. No acquire
 | |
|  *			method can be successful in these cases.
 | |
|  *
 | |
|  *	-EBUSY:		The current owner has a lower priority but the console
 | |
|  *			in an unsafe state. The caller should try using
 | |
|  *			the handover acquire method.
 | |
|  */
 | |
| static int nbcon_context_try_acquire_direct(struct nbcon_context *ctxt,
 | |
| 					    struct nbcon_state *cur, bool is_reacquire)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state new;
 | |
| 
 | |
| 	do {
 | |
| 		/*
 | |
| 		 * Panic does not imply that the console is owned. However,
 | |
| 		 * since all non-panic CPUs are stopped during panic(), it
 | |
| 		 * is safer to have them avoid gaining console ownership.
 | |
| 		 *
 | |
| 		 * If this acquire is a reacquire (and an unsafe takeover
 | |
| 		 * has not previously occurred) then it is allowed to attempt
 | |
| 		 * a direct acquire in panic. This gives console drivers an
 | |
| 		 * opportunity to perform any necessary cleanup if they were
 | |
| 		 * interrupted by the panic CPU while printing.
 | |
| 		 */
 | |
| 		if (other_cpu_in_panic() &&
 | |
| 		    (!is_reacquire || cur->unsafe_takeover)) {
 | |
| 			return -EPERM;
 | |
| 		}
 | |
| 
 | |
| 		if (ctxt->prio <= cur->prio || ctxt->prio <= cur->req_prio)
 | |
| 			return -EPERM;
 | |
| 
 | |
| 		if (cur->unsafe)
 | |
| 			return -EBUSY;
 | |
| 
 | |
| 		/*
 | |
| 		 * The console should never be safe for a direct acquire
 | |
| 		 * if an unsafe hostile takeover has ever happened.
 | |
| 		 */
 | |
| 		WARN_ON_ONCE(cur->unsafe_takeover);
 | |
| 
 | |
| 		new.atom = cur->atom;
 | |
| 		new.prio	= ctxt->prio;
 | |
| 		new.req_prio	= NBCON_PRIO_NONE;
 | |
| 		new.unsafe	= cur->unsafe_takeover;
 | |
| 		new.cpu		= cpu;
 | |
| 
 | |
| 	} while (!nbcon_state_try_cmpxchg(con, cur, &new));
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static bool nbcon_waiter_matches(struct nbcon_state *cur, int expected_prio)
 | |
| {
 | |
| 	/*
 | |
| 	 * The request context is well defined by the @req_prio because:
 | |
| 	 *
 | |
| 	 * - Only a context with a priority higher than the owner can become
 | |
| 	 *   a waiter.
 | |
| 	 * - Only a context with a priority higher than the waiter can
 | |
| 	 *   directly take over the request.
 | |
| 	 * - There are only three priorities.
 | |
| 	 * - Only one CPU is allowed to request PANIC priority.
 | |
| 	 * - Lower priorities are ignored during panic() until reboot.
 | |
| 	 *
 | |
| 	 * As a result, the following scenario is *not* possible:
 | |
| 	 *
 | |
| 	 * 1. This context is currently a waiter.
 | |
| 	 * 2. Another context with a higher priority than this context
 | |
| 	 *    directly takes ownership.
 | |
| 	 * 3. The higher priority context releases the ownership.
 | |
| 	 * 4. Another lower priority context takes the ownership.
 | |
| 	 * 5. Another context with the same priority as this context
 | |
| 	 *    creates a request and starts waiting.
 | |
| 	 *
 | |
| 	 * Event #1 implies this context is EMERGENCY.
 | |
| 	 * Event #2 implies the new context is PANIC.
 | |
| 	 * Event #3 occurs when panic() has flushed the console.
 | |
| 	 * Event #4 occurs when a non-panic CPU reacquires.
 | |
| 	 * Event #5 is not possible due to the other_cpu_in_panic() check
 | |
| 	 *          in nbcon_context_try_acquire_handover().
 | |
| 	 */
 | |
| 
 | |
| 	return (cur->req_prio == expected_prio);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_try_acquire_requested - Try to acquire after having
 | |
|  *					 requested a handover
 | |
|  * @ctxt:	The context of the caller
 | |
|  * @cur:	The current console state
 | |
|  *
 | |
|  * This is a helper function for nbcon_context_try_acquire_handover().
 | |
|  * It is called when the console is in an unsafe state. The current
 | |
|  * owner will release the console on exit from the unsafe region.
 | |
|  *
 | |
|  * Return:	0 on success and @cur is updated to the new console state.
 | |
|  *		Otherwise an error code on failure.
 | |
|  *
 | |
|  * Errors:
 | |
|  *
 | |
|  *	-EPERM:		A panic is in progress and this is not the panic CPU
 | |
|  *			or this context is no longer the waiter.
 | |
|  *
 | |
|  *	-EBUSY:		The console is still locked. The caller should
 | |
|  *			continue waiting.
 | |
|  *
 | |
|  * Note: The caller must still remove the request when an error has occurred
 | |
|  *       except when this context is no longer the waiter.
 | |
|  */
 | |
| static int nbcon_context_try_acquire_requested(struct nbcon_context *ctxt,
 | |
| 					       struct nbcon_state *cur)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state new;
 | |
| 
 | |
| 	/* Note that the caller must still remove the request! */
 | |
| 	if (other_cpu_in_panic())
 | |
| 		return -EPERM;
 | |
| 
 | |
| 	/*
 | |
| 	 * Note that the waiter will also change if there was an unsafe
 | |
| 	 * hostile takeover.
 | |
| 	 */
 | |
| 	if (!nbcon_waiter_matches(cur, ctxt->prio))
 | |
| 		return -EPERM;
 | |
| 
 | |
| 	/* If still locked, caller should continue waiting. */
 | |
| 	if (cur->prio != NBCON_PRIO_NONE)
 | |
| 		return -EBUSY;
 | |
| 
 | |
| 	/*
 | |
| 	 * The previous owner should have never released ownership
 | |
| 	 * in an unsafe region.
 | |
| 	 */
 | |
| 	WARN_ON_ONCE(cur->unsafe);
 | |
| 
 | |
| 	new.atom = cur->atom;
 | |
| 	new.prio	= ctxt->prio;
 | |
| 	new.req_prio	= NBCON_PRIO_NONE;
 | |
| 	new.unsafe	= cur->unsafe_takeover;
 | |
| 	new.cpu		= cpu;
 | |
| 
 | |
| 	if (!nbcon_state_try_cmpxchg(con, cur, &new)) {
 | |
| 		/*
 | |
| 		 * The acquire could fail only when it has been taken
 | |
| 		 * over by a higher priority context.
 | |
| 		 */
 | |
| 		WARN_ON_ONCE(nbcon_waiter_matches(cur, ctxt->prio));
 | |
| 		return -EPERM;
 | |
| 	}
 | |
| 
 | |
| 	/* Handover success. This context now owns the console. */
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_try_acquire_handover - Try to acquire via handover
 | |
|  * @ctxt:	The context of the caller
 | |
|  * @cur:	The current console state
 | |
|  *
 | |
|  * The function must be called only when the context has higher priority
 | |
|  * than the current owner and the console is in an unsafe state.
 | |
|  * It is the case when nbcon_context_try_acquire_direct() returns -EBUSY.
 | |
|  *
 | |
|  * The function sets "req_prio" field to make the current owner aware of
 | |
|  * the request. Then it waits until the current owner releases the console,
 | |
|  * or an even higher context takes over the request, or timeout expires.
 | |
|  *
 | |
|  * The current owner checks the "req_prio" field on exit from the unsafe
 | |
|  * region and releases the console. It does not touch the "req_prio" field
 | |
|  * so that the console stays reserved for the waiter.
 | |
|  *
 | |
|  * Return:	0 on success. Otherwise, an error code on failure. Also @cur
 | |
|  *		is updated to the latest state when failed to modify it.
 | |
|  *
 | |
|  * Errors:
 | |
|  *
 | |
|  *	-EPERM:		A panic is in progress and this is not the panic CPU.
 | |
|  *			Or a higher priority context has taken over the
 | |
|  *			console or the handover request.
 | |
|  *
 | |
|  *	-EBUSY:		The current owner is on the same CPU so that the hand
 | |
|  *			shake could not work. Or the current owner is not
 | |
|  *			willing to wait (zero timeout). Or the console does
 | |
|  *			not enter the safe state before timeout passed. The
 | |
|  *			caller might still use the unsafe hostile takeover
 | |
|  *			when allowed.
 | |
|  *
 | |
|  *	-EAGAIN:	@cur has changed when creating the handover request.
 | |
|  *			The caller should retry with direct acquire.
 | |
|  */
 | |
| static int nbcon_context_try_acquire_handover(struct nbcon_context *ctxt,
 | |
| 					      struct nbcon_state *cur)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state new;
 | |
| 	int timeout;
 | |
| 	int request_err = -EBUSY;
 | |
| 
 | |
| 	/*
 | |
| 	 * Check that the handover is called when the direct acquire failed
 | |
| 	 * with -EBUSY.
 | |
| 	 */
 | |
| 	WARN_ON_ONCE(ctxt->prio <= cur->prio || ctxt->prio <= cur->req_prio);
 | |
| 	WARN_ON_ONCE(!cur->unsafe);
 | |
| 
 | |
| 	/*
 | |
| 	 * Panic does not imply that the console is owned. However, it
 | |
| 	 * is critical that non-panic CPUs during panic are unable to
 | |
| 	 * wait for a handover in order to satisfy the assumptions of
 | |
| 	 * nbcon_waiter_matches(). In particular, the assumption that
 | |
| 	 * lower priorities are ignored during panic.
 | |
| 	 */
 | |
| 	if (other_cpu_in_panic())
 | |
| 		return -EPERM;
 | |
| 
 | |
| 	/* Handover is not possible on the same CPU. */
 | |
| 	if (cur->cpu == cpu)
 | |
| 		return -EBUSY;
 | |
| 
 | |
| 	/*
 | |
| 	 * Console stays unsafe after an unsafe takeover until re-initialized.
 | |
| 	 * Waiting is not going to help in this case.
 | |
| 	 */
 | |
| 	if (cur->unsafe_takeover)
 | |
| 		return -EBUSY;
 | |
| 
 | |
| 	/* Is the caller willing to wait? */
 | |
| 	if (ctxt->spinwait_max_us == 0)
 | |
| 		return -EBUSY;
 | |
| 
 | |
| 	/*
 | |
| 	 * Setup a request for the handover. The caller should try to acquire
 | |
| 	 * the console directly when the current state has been modified.
 | |
| 	 */
 | |
| 	new.atom = cur->atom;
 | |
| 	new.req_prio = ctxt->prio;
 | |
| 	if (!nbcon_state_try_cmpxchg(con, cur, &new))
 | |
| 		return -EAGAIN;
 | |
| 
 | |
| 	cur->atom = new.atom;
 | |
| 
 | |
| 	/* Wait until there is no owner and then acquire the console. */
 | |
| 	for (timeout = ctxt->spinwait_max_us; timeout >= 0; timeout--) {
 | |
| 		/* On successful acquire, this request is cleared. */
 | |
| 		request_err = nbcon_context_try_acquire_requested(ctxt, cur);
 | |
| 		if (!request_err)
 | |
| 			return 0;
 | |
| 
 | |
| 		/*
 | |
| 		 * If the acquire should be aborted, it must be ensured
 | |
| 		 * that the request is removed before returning to caller.
 | |
| 		 */
 | |
| 		if (request_err == -EPERM)
 | |
| 			break;
 | |
| 
 | |
| 		udelay(1);
 | |
| 
 | |
| 		/* Re-read the state because some time has passed. */
 | |
| 		nbcon_state_read(con, cur);
 | |
| 	}
 | |
| 
 | |
| 	/* Timed out or aborted. Carefully remove handover request. */
 | |
| 	do {
 | |
| 		/*
 | |
| 		 * No need to remove request if there is a new waiter. This
 | |
| 		 * can only happen if a higher priority context has taken over
 | |
| 		 * the console or the handover request.
 | |
| 		 */
 | |
| 		if (!nbcon_waiter_matches(cur, ctxt->prio))
 | |
| 			return -EPERM;
 | |
| 
 | |
| 		/* Unset request for handover. */
 | |
| 		new.atom = cur->atom;
 | |
| 		new.req_prio = NBCON_PRIO_NONE;
 | |
| 		if (nbcon_state_try_cmpxchg(con, cur, &new)) {
 | |
| 			/*
 | |
| 			 * Request successfully unset. Report failure of
 | |
| 			 * acquiring via handover.
 | |
| 			 */
 | |
| 			cur->atom = new.atom;
 | |
| 			return request_err;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * Unable to remove request. Try to acquire in case
 | |
| 		 * the owner has released the lock.
 | |
| 		 */
 | |
| 	} while (nbcon_context_try_acquire_requested(ctxt, cur));
 | |
| 
 | |
| 	/* Lucky timing. The acquire succeeded while removing the request. */
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_try_acquire_hostile - Acquire via unsafe hostile takeover
 | |
|  * @ctxt:	The context of the caller
 | |
|  * @cur:	The current console state
 | |
|  *
 | |
|  * Acquire the console even in the unsafe state.
 | |
|  *
 | |
|  * It can be permitted by setting the 'allow_unsafe_takeover' field only
 | |
|  * by the final attempt to flush messages in panic().
 | |
|  *
 | |
|  * Return:	0 on success. -EPERM when not allowed by the context.
 | |
|  */
 | |
| static int nbcon_context_try_acquire_hostile(struct nbcon_context *ctxt,
 | |
| 					     struct nbcon_state *cur)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state new;
 | |
| 
 | |
| 	if (!ctxt->allow_unsafe_takeover)
 | |
| 		return -EPERM;
 | |
| 
 | |
| 	/* Ensure caller is allowed to perform unsafe hostile takeovers. */
 | |
| 	if (WARN_ON_ONCE(ctxt->prio != NBCON_PRIO_PANIC))
 | |
| 		return -EPERM;
 | |
| 
 | |
| 	/*
 | |
| 	 * Check that try_acquire_direct() and try_acquire_handover() returned
 | |
| 	 * -EBUSY in the right situation.
 | |
| 	 */
 | |
| 	WARN_ON_ONCE(ctxt->prio <= cur->prio || ctxt->prio <= cur->req_prio);
 | |
| 	WARN_ON_ONCE(cur->unsafe != true);
 | |
| 
 | |
| 	do {
 | |
| 		new.atom = cur->atom;
 | |
| 		new.cpu			= cpu;
 | |
| 		new.prio		= ctxt->prio;
 | |
| 		new.unsafe		|= cur->unsafe_takeover;
 | |
| 		new.unsafe_takeover	|= cur->unsafe;
 | |
| 
 | |
| 	} while (!nbcon_state_try_cmpxchg(con, cur, &new));
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static struct printk_buffers panic_nbcon_pbufs;
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_try_acquire - Try to acquire nbcon console
 | |
|  * @ctxt:		The context of the caller
 | |
|  * @is_reacquire:	This acquire is a reacquire
 | |
|  *
 | |
|  * Context:	Under @ctxt->con->device_lock() or local_irq_save().
 | |
|  * Return:	True if the console was acquired. False otherwise.
 | |
|  *
 | |
|  * If the caller allowed an unsafe hostile takeover, on success the
 | |
|  * caller should check the current console state to see if it is
 | |
|  * in an unsafe state. Otherwise, on success the caller may assume
 | |
|  * the console is not in an unsafe state.
 | |
|  */
 | |
| static bool nbcon_context_try_acquire(struct nbcon_context *ctxt, bool is_reacquire)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state cur;
 | |
| 	int err;
 | |
| 
 | |
| 	nbcon_state_read(con, &cur);
 | |
| try_again:
 | |
| 	err = nbcon_context_try_acquire_direct(ctxt, &cur, is_reacquire);
 | |
| 	if (err != -EBUSY)
 | |
| 		goto out;
 | |
| 
 | |
| 	err = nbcon_context_try_acquire_handover(ctxt, &cur);
 | |
| 	if (err == -EAGAIN)
 | |
| 		goto try_again;
 | |
| 	if (err != -EBUSY)
 | |
| 		goto out;
 | |
| 
 | |
| 	err = nbcon_context_try_acquire_hostile(ctxt, &cur);
 | |
| out:
 | |
| 	if (err)
 | |
| 		return false;
 | |
| 
 | |
| 	/* Acquire succeeded. */
 | |
| 
 | |
| 	/* Assign the appropriate buffer for this context. */
 | |
| 	if (atomic_read(&panic_cpu) == cpu)
 | |
| 		ctxt->pbufs = &panic_nbcon_pbufs;
 | |
| 	else
 | |
| 		ctxt->pbufs = con->pbufs;
 | |
| 
 | |
| 	/* Set the record sequence for this context to print. */
 | |
| 	ctxt->seq = nbcon_seq_read(ctxt->console);
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| static bool nbcon_owner_matches(struct nbcon_state *cur, int expected_cpu,
 | |
| 				int expected_prio)
 | |
| {
 | |
| 	/*
 | |
| 	 * A similar function, nbcon_waiter_matches(), only deals with
 | |
| 	 * EMERGENCY and PANIC priorities. However, this function must also
 | |
| 	 * deal with the NORMAL priority, which requires additional checks
 | |
| 	 * and constraints.
 | |
| 	 *
 | |
| 	 * For the case where preemption and interrupts are disabled, it is
 | |
| 	 * enough to also verify that the owning CPU has not changed.
 | |
| 	 *
 | |
| 	 * For the case where preemption or interrupts are enabled, an
 | |
| 	 * external synchronization method *must* be used. In particular,
 | |
| 	 * the driver-specific locking mechanism used in device_lock()
 | |
| 	 * (including disabling migration) should be used. It prevents
 | |
| 	 * scenarios such as:
 | |
| 	 *
 | |
| 	 * 1. [Task A] owns a context with NBCON_PRIO_NORMAL on [CPU X] and
 | |
| 	 *    is scheduled out.
 | |
| 	 * 2. Another context takes over the lock with NBCON_PRIO_EMERGENCY
 | |
| 	 *    and releases it.
 | |
| 	 * 3. [Task B] acquires a context with NBCON_PRIO_NORMAL on [CPU X]
 | |
| 	 *    and is scheduled out.
 | |
| 	 * 4. [Task A] gets running on [CPU X] and sees that the console is
 | |
| 	 *    still owned by a task on [CPU X] with NBON_PRIO_NORMAL. Thus
 | |
| 	 *    [Task A] thinks it is the owner when it is not.
 | |
| 	 */
 | |
| 
 | |
| 	if (cur->prio != expected_prio)
 | |
| 		return false;
 | |
| 
 | |
| 	if (cur->cpu != expected_cpu)
 | |
| 		return false;
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_release - Release the console
 | |
|  * @ctxt:	The nbcon context from nbcon_context_try_acquire()
 | |
|  */
 | |
| static void nbcon_context_release(struct nbcon_context *ctxt)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state cur;
 | |
| 	struct nbcon_state new;
 | |
| 
 | |
| 	nbcon_state_read(con, &cur);
 | |
| 
 | |
| 	do {
 | |
| 		if (!nbcon_owner_matches(&cur, cpu, ctxt->prio))
 | |
| 			break;
 | |
| 
 | |
| 		new.atom = cur.atom;
 | |
| 		new.prio = NBCON_PRIO_NONE;
 | |
| 
 | |
| 		/*
 | |
| 		 * If @unsafe_takeover is set, it is kept set so that
 | |
| 		 * the state remains permanently unsafe.
 | |
| 		 */
 | |
| 		new.unsafe |= cur.unsafe_takeover;
 | |
| 
 | |
| 	} while (!nbcon_state_try_cmpxchg(con, &cur, &new));
 | |
| 
 | |
| 	ctxt->pbufs = NULL;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_context_can_proceed - Check whether ownership can proceed
 | |
|  * @ctxt:	The nbcon context from nbcon_context_try_acquire()
 | |
|  * @cur:	The current console state
 | |
|  *
 | |
|  * Return:	True if this context still owns the console. False if
 | |
|  *		ownership was handed over or taken.
 | |
|  *
 | |
|  * Must be invoked when entering the unsafe state to make sure that it still
 | |
|  * owns the lock. Also must be invoked when exiting the unsafe context
 | |
|  * to eventually free the lock for a higher priority context which asked
 | |
|  * for the friendly handover.
 | |
|  *
 | |
|  * It can be called inside an unsafe section when the console is just
 | |
|  * temporary in safe state instead of exiting and entering the unsafe
 | |
|  * state.
 | |
|  *
 | |
|  * Also it can be called in the safe context before doing an expensive
 | |
|  * safe operation. It does not make sense to do the operation when
 | |
|  * a higher priority context took the lock.
 | |
|  *
 | |
|  * When this function returns false then the calling context no longer owns
 | |
|  * the console and is no longer allowed to go forward. In this case it must
 | |
|  * back out immediately and carefully. The buffer content is also no longer
 | |
|  * trusted since it no longer belongs to the calling context.
 | |
|  */
 | |
| static bool nbcon_context_can_proceed(struct nbcon_context *ctxt, struct nbcon_state *cur)
 | |
| {
 | |
| 	unsigned int cpu = smp_processor_id();
 | |
| 
 | |
| 	/* Make sure this context still owns the console. */
 | |
| 	if (!nbcon_owner_matches(cur, cpu, ctxt->prio))
 | |
| 		return false;
 | |
| 
 | |
| 	/* The console owner can proceed if there is no waiter. */
 | |
| 	if (cur->req_prio == NBCON_PRIO_NONE)
 | |
| 		return true;
 | |
| 
 | |
| 	/*
 | |
| 	 * A console owner within an unsafe region is always allowed to
 | |
| 	 * proceed, even if there are waiters. It can perform a handover
 | |
| 	 * when exiting the unsafe region. Otherwise the waiter will
 | |
| 	 * need to perform an unsafe hostile takeover.
 | |
| 	 */
 | |
| 	if (cur->unsafe)
 | |
| 		return true;
 | |
| 
 | |
| 	/* Waiters always have higher priorities than owners. */
 | |
| 	WARN_ON_ONCE(cur->req_prio <= cur->prio);
 | |
| 
 | |
| 	/*
 | |
| 	 * Having a safe point for take over and eventually a few
 | |
| 	 * duplicated characters or a full line is way better than a
 | |
| 	 * hostile takeover. Post processing can take care of the garbage.
 | |
| 	 * Release and hand over.
 | |
| 	 */
 | |
| 	nbcon_context_release(ctxt);
 | |
| 
 | |
| 	/*
 | |
| 	 * It is not clear whether the waiter really took over ownership. The
 | |
| 	 * outermost callsite must make the final decision whether console
 | |
| 	 * ownership is needed for it to proceed. If yes, it must reacquire
 | |
| 	 * ownership (possibly hostile) before carefully proceeding.
 | |
| 	 *
 | |
| 	 * The calling context no longer owns the console so go back all the
 | |
| 	 * way instead of trying to implement reacquire heuristics in tons of
 | |
| 	 * places.
 | |
| 	 */
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_can_proceed - Check whether ownership can proceed
 | |
|  * @wctxt:	The write context that was handed to the write function
 | |
|  *
 | |
|  * Return:	True if this context still owns the console. False if
 | |
|  *		ownership was handed over or taken.
 | |
|  *
 | |
|  * It is used in nbcon_enter_unsafe() to make sure that it still owns the
 | |
|  * lock. Also it is used in nbcon_exit_unsafe() to eventually free the lock
 | |
|  * for a higher priority context which asked for the friendly handover.
 | |
|  *
 | |
|  * It can be called inside an unsafe section when the console is just
 | |
|  * temporary in safe state instead of exiting and entering the unsafe state.
 | |
|  *
 | |
|  * Also it can be called in the safe context before doing an expensive safe
 | |
|  * operation. It does not make sense to do the operation when a higher
 | |
|  * priority context took the lock.
 | |
|  *
 | |
|  * When this function returns false then the calling context no longer owns
 | |
|  * the console and is no longer allowed to go forward. In this case it must
 | |
|  * back out immediately and carefully. The buffer content is also no longer
 | |
|  * trusted since it no longer belongs to the calling context.
 | |
|  */
 | |
| bool nbcon_can_proceed(struct nbcon_write_context *wctxt)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state cur;
 | |
| 
 | |
| 	nbcon_state_read(con, &cur);
 | |
| 
 | |
| 	return nbcon_context_can_proceed(ctxt, &cur);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(nbcon_can_proceed);
 | |
| 
 | |
| #define nbcon_context_enter_unsafe(c)	__nbcon_context_update_unsafe(c, true)
 | |
| #define nbcon_context_exit_unsafe(c)	__nbcon_context_update_unsafe(c, false)
 | |
| 
 | |
| /**
 | |
|  * __nbcon_context_update_unsafe - Update the unsafe bit in @con->nbcon_state
 | |
|  * @ctxt:	The nbcon context from nbcon_context_try_acquire()
 | |
|  * @unsafe:	The new value for the unsafe bit
 | |
|  *
 | |
|  * Return:	True if the unsafe state was updated and this context still
 | |
|  *		owns the console. Otherwise false if ownership was handed
 | |
|  *		over or taken.
 | |
|  *
 | |
|  * This function allows console owners to modify the unsafe status of the
 | |
|  * console.
 | |
|  *
 | |
|  * When this function returns false then the calling context no longer owns
 | |
|  * the console and is no longer allowed to go forward. In this case it must
 | |
|  * back out immediately and carefully. The buffer content is also no longer
 | |
|  * trusted since it no longer belongs to the calling context.
 | |
|  *
 | |
|  * Internal helper to avoid duplicated code.
 | |
|  */
 | |
| static bool __nbcon_context_update_unsafe(struct nbcon_context *ctxt, bool unsafe)
 | |
| {
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state cur;
 | |
| 	struct nbcon_state new;
 | |
| 
 | |
| 	nbcon_state_read(con, &cur);
 | |
| 
 | |
| 	do {
 | |
| 		/*
 | |
| 		 * The unsafe bit must not be cleared if an
 | |
| 		 * unsafe hostile takeover has occurred.
 | |
| 		 */
 | |
| 		if (!unsafe && cur.unsafe_takeover)
 | |
| 			goto out;
 | |
| 
 | |
| 		if (!nbcon_context_can_proceed(ctxt, &cur))
 | |
| 			return false;
 | |
| 
 | |
| 		new.atom = cur.atom;
 | |
| 		new.unsafe = unsafe;
 | |
| 	} while (!nbcon_state_try_cmpxchg(con, &cur, &new));
 | |
| 
 | |
| 	cur.atom = new.atom;
 | |
| out:
 | |
| 	return nbcon_context_can_proceed(ctxt, &cur);
 | |
| }
 | |
| 
 | |
| static void nbcon_write_context_set_buf(struct nbcon_write_context *wctxt,
 | |
| 					char *buf, unsigned int len)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 	struct console *con = ctxt->console;
 | |
| 	struct nbcon_state cur;
 | |
| 
 | |
| 	wctxt->outbuf = buf;
 | |
| 	wctxt->len = len;
 | |
| 	nbcon_state_read(con, &cur);
 | |
| 	wctxt->unsafe_takeover = cur.unsafe_takeover;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_enter_unsafe - Enter an unsafe region in the driver
 | |
|  * @wctxt:	The write context that was handed to the write function
 | |
|  *
 | |
|  * Return:	True if this context still owns the console. False if
 | |
|  *		ownership was handed over or taken.
 | |
|  *
 | |
|  * When this function returns false then the calling context no longer owns
 | |
|  * the console and is no longer allowed to go forward. In this case it must
 | |
|  * back out immediately and carefully. The buffer content is also no longer
 | |
|  * trusted since it no longer belongs to the calling context.
 | |
|  */
 | |
| bool nbcon_enter_unsafe(struct nbcon_write_context *wctxt)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 	bool is_owner;
 | |
| 
 | |
| 	is_owner = nbcon_context_enter_unsafe(ctxt);
 | |
| 	if (!is_owner)
 | |
| 		nbcon_write_context_set_buf(wctxt, NULL, 0);
 | |
| 	return is_owner;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(nbcon_enter_unsafe);
 | |
| 
 | |
| /**
 | |
|  * nbcon_exit_unsafe - Exit an unsafe region in the driver
 | |
|  * @wctxt:	The write context that was handed to the write function
 | |
|  *
 | |
|  * Return:	True if this context still owns the console. False if
 | |
|  *		ownership was handed over or taken.
 | |
|  *
 | |
|  * When this function returns false then the calling context no longer owns
 | |
|  * the console and is no longer allowed to go forward. In this case it must
 | |
|  * back out immediately and carefully. The buffer content is also no longer
 | |
|  * trusted since it no longer belongs to the calling context.
 | |
|  */
 | |
| bool nbcon_exit_unsafe(struct nbcon_write_context *wctxt)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 	bool ret;
 | |
| 
 | |
| 	ret = nbcon_context_exit_unsafe(ctxt);
 | |
| 	if (!ret)
 | |
| 		nbcon_write_context_set_buf(wctxt, NULL, 0);
 | |
| 	return ret;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(nbcon_exit_unsafe);
 | |
| 
 | |
| /**
 | |
|  * nbcon_reacquire_nobuf - Reacquire a console after losing ownership
 | |
|  *				while printing
 | |
|  * @wctxt:	The write context that was handed to the write callback
 | |
|  *
 | |
|  * Since ownership can be lost at any time due to handover or takeover, a
 | |
|  * printing context _must_ be prepared to back out immediately and
 | |
|  * carefully. However, there are scenarios where the printing context must
 | |
|  * reacquire ownership in order to finalize or revert hardware changes.
 | |
|  *
 | |
|  * This function allows a printing context to reacquire ownership using the
 | |
|  * same priority as its previous ownership.
 | |
|  *
 | |
|  * Note that after a successful reacquire the printing context will have no
 | |
|  * output buffer because that has been lost. This function cannot be used to
 | |
|  * resume printing.
 | |
|  */
 | |
| void nbcon_reacquire_nobuf(struct nbcon_write_context *wctxt)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 
 | |
| 	while (!nbcon_context_try_acquire(ctxt, true))
 | |
| 		cpu_relax();
 | |
| 
 | |
| 	nbcon_write_context_set_buf(wctxt, NULL, 0);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(nbcon_reacquire_nobuf);
 | |
| 
 | |
| /**
 | |
|  * nbcon_emit_next_record - Emit a record in the acquired context
 | |
|  * @wctxt:	The write context that will be handed to the write function
 | |
|  * @use_atomic:	True if the write_atomic() callback is to be used
 | |
|  *
 | |
|  * Return:	True if this context still owns the console. False if
 | |
|  *		ownership was handed over or taken.
 | |
|  *
 | |
|  * When this function returns false then the calling context no longer owns
 | |
|  * the console and is no longer allowed to go forward. In this case it must
 | |
|  * back out immediately and carefully. The buffer content is also no longer
 | |
|  * trusted since it no longer belongs to the calling context. If the caller
 | |
|  * wants to do more it must reacquire the console first.
 | |
|  *
 | |
|  * When true is returned, @wctxt->ctxt.backlog indicates whether there are
 | |
|  * still records pending in the ringbuffer,
 | |
|  */
 | |
| static bool nbcon_emit_next_record(struct nbcon_write_context *wctxt, bool use_atomic)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 	struct console *con = ctxt->console;
 | |
| 	bool is_extended = console_srcu_read_flags(con) & CON_EXTENDED;
 | |
| 	struct printk_message pmsg = {
 | |
| 		.pbufs = ctxt->pbufs,
 | |
| 	};
 | |
| 	unsigned long con_dropped;
 | |
| 	struct nbcon_state cur;
 | |
| 	unsigned long dropped;
 | |
| 	unsigned long ulseq;
 | |
| 
 | |
| 	/*
 | |
| 	 * This function should never be called for consoles that have not
 | |
| 	 * implemented the necessary callback for writing: i.e. legacy
 | |
| 	 * consoles and, when atomic, nbcon consoles with no write_atomic().
 | |
| 	 * Handle it as if ownership was lost and try to continue.
 | |
| 	 *
 | |
| 	 * Note that for nbcon consoles the write_thread() callback is
 | |
| 	 * mandatory and was already checked in nbcon_alloc().
 | |
| 	 */
 | |
| 	if (WARN_ON_ONCE((use_atomic && !con->write_atomic) ||
 | |
| 			 !(console_srcu_read_flags(con) & CON_NBCON))) {
 | |
| 		nbcon_context_release(ctxt);
 | |
| 		return false;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * The printk buffers are filled within an unsafe section. This
 | |
| 	 * prevents NBCON_PRIO_NORMAL and NBCON_PRIO_EMERGENCY from
 | |
| 	 * clobbering each other.
 | |
| 	 */
 | |
| 
 | |
| 	if (!nbcon_context_enter_unsafe(ctxt))
 | |
| 		return false;
 | |
| 
 | |
| 	ctxt->backlog = printk_get_next_message(&pmsg, ctxt->seq, is_extended, true);
 | |
| 	if (!ctxt->backlog)
 | |
| 		return nbcon_context_exit_unsafe(ctxt);
 | |
| 
 | |
| 	/*
 | |
| 	 * @con->dropped is not protected in case of an unsafe hostile
 | |
| 	 * takeover. In that situation the update can be racy so
 | |
| 	 * annotate it accordingly.
 | |
| 	 */
 | |
| 	con_dropped = data_race(READ_ONCE(con->dropped));
 | |
| 
 | |
| 	dropped = con_dropped + pmsg.dropped;
 | |
| 	if (dropped && !is_extended)
 | |
| 		console_prepend_dropped(&pmsg, dropped);
 | |
| 
 | |
| 	/*
 | |
| 	 * If the previous owner was assigned the same record, this context
 | |
| 	 * has taken over ownership and is replaying the record. Prepend a
 | |
| 	 * message to let the user know the record is replayed.
 | |
| 	 */
 | |
| 	ulseq = atomic_long_read(&ACCESS_PRIVATE(con, nbcon_prev_seq));
 | |
| 	if (__ulseq_to_u64seq(prb, ulseq) == pmsg.seq) {
 | |
| 		console_prepend_replay(&pmsg);
 | |
| 	} else {
 | |
| 		/*
 | |
| 		 * Ensure this context is still the owner before trying to
 | |
| 		 * update @nbcon_prev_seq. Otherwise the value in @ulseq may
 | |
| 		 * not be from the previous owner and instead be some later
 | |
| 		 * value from the context that took over ownership.
 | |
| 		 */
 | |
| 		nbcon_state_read(con, &cur);
 | |
| 		if (!nbcon_context_can_proceed(ctxt, &cur))
 | |
| 			return false;
 | |
| 
 | |
| 		atomic_long_try_cmpxchg(&ACCESS_PRIVATE(con, nbcon_prev_seq), &ulseq,
 | |
| 					__u64seq_to_ulseq(pmsg.seq));
 | |
| 	}
 | |
| 
 | |
| 	if (!nbcon_context_exit_unsafe(ctxt))
 | |
| 		return false;
 | |
| 
 | |
| 	/* For skipped records just update seq/dropped in @con. */
 | |
| 	if (pmsg.outbuf_len == 0)
 | |
| 		goto update_con;
 | |
| 
 | |
| 	/* Initialize the write context for driver callbacks. */
 | |
| 	nbcon_write_context_set_buf(wctxt, &pmsg.pbufs->outbuf[0], pmsg.outbuf_len);
 | |
| 
 | |
| 	if (use_atomic)
 | |
| 		con->write_atomic(con, wctxt);
 | |
| 	else
 | |
| 		con->write_thread(con, wctxt);
 | |
| 
 | |
| 	if (!wctxt->outbuf) {
 | |
| 		/*
 | |
| 		 * Ownership was lost and reacquired by the driver. Handle it
 | |
| 		 * as if ownership was lost.
 | |
| 		 */
 | |
| 		nbcon_context_release(ctxt);
 | |
| 		return false;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Ownership may have been lost but _not_ reacquired by the driver.
 | |
| 	 * This case is detected and handled when entering unsafe to update
 | |
| 	 * dropped/seq values.
 | |
| 	 */
 | |
| 
 | |
| 	/*
 | |
| 	 * Since any dropped message was successfully output, reset the
 | |
| 	 * dropped count for the console.
 | |
| 	 */
 | |
| 	dropped = 0;
 | |
| update_con:
 | |
| 	/*
 | |
| 	 * The dropped count and the sequence number are updated within an
 | |
| 	 * unsafe section. This limits update races to the panic context and
 | |
| 	 * allows the panic context to win.
 | |
| 	 */
 | |
| 
 | |
| 	if (!nbcon_context_enter_unsafe(ctxt))
 | |
| 		return false;
 | |
| 
 | |
| 	if (dropped != con_dropped) {
 | |
| 		/* Counterpart to the READ_ONCE() above. */
 | |
| 		WRITE_ONCE(con->dropped, dropped);
 | |
| 	}
 | |
| 
 | |
| 	nbcon_seq_try_update(ctxt, pmsg.seq + 1);
 | |
| 
 | |
| 	return nbcon_context_exit_unsafe(ctxt);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * nbcon_emit_one - Print one record for an nbcon console using the
 | |
|  *			specified callback
 | |
|  * @wctxt:	An initialized write context struct to use for this context
 | |
|  * @use_atomic:	True if the write_atomic() callback is to be used
 | |
|  *
 | |
|  * Return:	True, when a record has been printed and there are still
 | |
|  *		pending records. The caller might want to continue flushing.
 | |
|  *
 | |
|  *		False, when there is no pending record, or when the console
 | |
|  *		context cannot be acquired, or the ownership has been lost.
 | |
|  *		The caller should give up. Either the job is done, cannot be
 | |
|  *		done, or will be handled by the owning context.
 | |
|  *
 | |
|  * This is an internal helper to handle the locking of the console before
 | |
|  * calling nbcon_emit_next_record().
 | |
|  */
 | |
| static bool nbcon_emit_one(struct nbcon_write_context *wctxt, bool use_atomic)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
 | |
| 	struct console *con = ctxt->console;
 | |
| 	unsigned long flags;
 | |
| 	bool ret = false;
 | |
| 
 | |
| 	if (!use_atomic) {
 | |
| 		con->device_lock(con, &flags);
 | |
| 
 | |
| 		/*
 | |
| 		 * Ensure this stays on the CPU to make handover and
 | |
| 		 * takeover possible.
 | |
| 		 */
 | |
| 		cant_migrate();
 | |
| 	}
 | |
| 
 | |
| 	if (!nbcon_context_try_acquire(ctxt, false))
 | |
| 		goto out;
 | |
| 
 | |
| 	/*
 | |
| 	 * nbcon_emit_next_record() returns false when the console was
 | |
| 	 * handed over or taken over. In both cases the context is no
 | |
| 	 * longer valid.
 | |
| 	 *
 | |
| 	 * The higher priority printing context takes over responsibility
 | |
| 	 * to print the pending records.
 | |
| 	 */
 | |
| 	if (!nbcon_emit_next_record(wctxt, use_atomic))
 | |
| 		goto out;
 | |
| 
 | |
| 	nbcon_context_release(ctxt);
 | |
| 
 | |
| 	ret = ctxt->backlog;
 | |
| out:
 | |
| 	if (!use_atomic)
 | |
| 		con->device_unlock(con, flags);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_kthread_should_wakeup - Check whether a printer thread should wakeup
 | |
|  * @con:	Console to operate on
 | |
|  * @ctxt:	The nbcon context from nbcon_context_try_acquire()
 | |
|  *
 | |
|  * Return:	True if the thread should shutdown or if the console is
 | |
|  *		allowed to print and a record is available. False otherwise.
 | |
|  *
 | |
|  * After the thread wakes up, it must first check if it should shutdown before
 | |
|  * attempting any printing.
 | |
|  */
 | |
| static bool nbcon_kthread_should_wakeup(struct console *con, struct nbcon_context *ctxt)
 | |
| {
 | |
| 	bool ret = false;
 | |
| 	short flags;
 | |
| 	int cookie;
 | |
| 
 | |
| 	if (kthread_should_stop())
 | |
| 		return true;
 | |
| 
 | |
| 	cookie = console_srcu_read_lock();
 | |
| 
 | |
| 	flags = console_srcu_read_flags(con);
 | |
| 	if (console_is_usable(con, flags, false)) {
 | |
| 		/* Bring the sequence in @ctxt up to date */
 | |
| 		ctxt->seq = nbcon_seq_read(con);
 | |
| 
 | |
| 		ret = prb_read_valid(prb, ctxt->seq, NULL);
 | |
| 	}
 | |
| 
 | |
| 	console_srcu_read_unlock(cookie);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_kthread_func - The printer thread function
 | |
|  * @__console:	Console to operate on
 | |
|  *
 | |
|  * Return:	0
 | |
|  */
 | |
| static int nbcon_kthread_func(void *__console)
 | |
| {
 | |
| 	struct console *con = __console;
 | |
| 	struct nbcon_write_context wctxt = {
 | |
| 		.ctxt.console	= con,
 | |
| 		.ctxt.prio	= NBCON_PRIO_NORMAL,
 | |
| 	};
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(&wctxt, ctxt);
 | |
| 	short con_flags;
 | |
| 	bool backlog;
 | |
| 	int cookie;
 | |
| 
 | |
| wait_for_event:
 | |
| 	/*
 | |
| 	 * Guarantee this task is visible on the rcuwait before
 | |
| 	 * checking the wake condition.
 | |
| 	 *
 | |
| 	 * The full memory barrier within set_current_state() of
 | |
| 	 * ___rcuwait_wait_event() pairs with the full memory
 | |
| 	 * barrier within rcuwait_has_sleeper().
 | |
| 	 *
 | |
| 	 * This pairs with rcuwait_has_sleeper:A and nbcon_kthread_wake:A.
 | |
| 	 */
 | |
| 	rcuwait_wait_event(&con->rcuwait,
 | |
| 			   nbcon_kthread_should_wakeup(con, ctxt),
 | |
| 			   TASK_INTERRUPTIBLE); /* LMM(nbcon_kthread_func:A) */
 | |
| 
 | |
| 	do {
 | |
| 		if (kthread_should_stop())
 | |
| 			return 0;
 | |
| 
 | |
| 		backlog = false;
 | |
| 
 | |
| 		/*
 | |
| 		 * Keep the srcu read lock around the entire operation so that
 | |
| 		 * synchronize_srcu() can guarantee that the kthread stopped
 | |
| 		 * or suspended printing.
 | |
| 		 */
 | |
| 		cookie = console_srcu_read_lock();
 | |
| 
 | |
| 		con_flags = console_srcu_read_flags(con);
 | |
| 
 | |
| 		if (console_is_usable(con, con_flags, false))
 | |
| 			backlog = nbcon_emit_one(&wctxt, false);
 | |
| 
 | |
| 		console_srcu_read_unlock(cookie);
 | |
| 
 | |
| 		cond_resched();
 | |
| 
 | |
| 	} while (backlog);
 | |
| 
 | |
| 	goto wait_for_event;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_irq_work - irq work to wake console printer thread
 | |
|  * @irq_work:	The irq work to operate on
 | |
|  */
 | |
| static void nbcon_irq_work(struct irq_work *irq_work)
 | |
| {
 | |
| 	struct console *con = container_of(irq_work, struct console, irq_work);
 | |
| 
 | |
| 	nbcon_kthread_wake(con);
 | |
| }
 | |
| 
 | |
| static inline bool rcuwait_has_sleeper(struct rcuwait *w)
 | |
| {
 | |
| 	/*
 | |
| 	 * Guarantee any new records can be seen by tasks preparing to wait
 | |
| 	 * before this context checks if the rcuwait is empty.
 | |
| 	 *
 | |
| 	 * This full memory barrier pairs with the full memory barrier within
 | |
| 	 * set_current_state() of ___rcuwait_wait_event(), which is called
 | |
| 	 * after prepare_to_rcuwait() adds the waiter but before it has
 | |
| 	 * checked the wait condition.
 | |
| 	 *
 | |
| 	 * This pairs with nbcon_kthread_func:A.
 | |
| 	 */
 | |
| 	smp_mb(); /* LMM(rcuwait_has_sleeper:A) */
 | |
| 	return rcuwait_active(w);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_kthreads_wake - Wake up printing threads using irq_work
 | |
|  */
 | |
| void nbcon_kthreads_wake(void)
 | |
| {
 | |
| 	struct console *con;
 | |
| 	int cookie;
 | |
| 
 | |
| 	if (!printk_kthreads_running)
 | |
| 		return;
 | |
| 
 | |
| 	cookie = console_srcu_read_lock();
 | |
| 	for_each_console_srcu(con) {
 | |
| 		if (!(console_srcu_read_flags(con) & CON_NBCON))
 | |
| 			continue;
 | |
| 
 | |
| 		/*
 | |
| 		 * Only schedule irq_work if the printing thread is
 | |
| 		 * actively waiting. If not waiting, the thread will
 | |
| 		 * notice by itself that it has work to do.
 | |
| 		 */
 | |
| 		if (rcuwait_has_sleeper(&con->rcuwait))
 | |
| 			irq_work_queue(&con->irq_work);
 | |
| 	}
 | |
| 	console_srcu_read_unlock(cookie);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * nbcon_kthread_stop - Stop a console printer thread
 | |
|  * @con:	Console to operate on
 | |
|  */
 | |
| void nbcon_kthread_stop(struct console *con)
 | |
| {
 | |
| 	lockdep_assert_console_list_lock_held();
 | |
| 
 | |
| 	if (!con->kthread)
 | |
| 		return;
 | |
| 
 | |
| 	kthread_stop(con->kthread);
 | |
| 	con->kthread = NULL;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_kthread_create - Create a console printer thread
 | |
|  * @con:	Console to operate on
 | |
|  *
 | |
|  * Return:	True if the kthread was started or already exists.
 | |
|  *		Otherwise false and @con must not be registered.
 | |
|  *
 | |
|  * This function is called when it will be expected that nbcon consoles are
 | |
|  * flushed using the kthread. The messages printed with NBCON_PRIO_NORMAL
 | |
|  * will be no longer flushed by the legacy loop. This is why failure must
 | |
|  * be fatal for console registration.
 | |
|  *
 | |
|  * If @con was already registered and this function fails, @con must be
 | |
|  * unregistered before the global state variable @printk_kthreads_running
 | |
|  * can be set.
 | |
|  */
 | |
| bool nbcon_kthread_create(struct console *con)
 | |
| {
 | |
| 	struct task_struct *kt;
 | |
| 
 | |
| 	lockdep_assert_console_list_lock_held();
 | |
| 
 | |
| 	if (con->kthread)
 | |
| 		return true;
 | |
| 
 | |
| 	kt = kthread_run(nbcon_kthread_func, con, "pr/%s%d", con->name, con->index);
 | |
| 	if (WARN_ON(IS_ERR(kt))) {
 | |
| 		con_printk(KERN_ERR, con, "failed to start printing thread\n");
 | |
| 		return false;
 | |
| 	}
 | |
| 
 | |
| 	con->kthread = kt;
 | |
| 
 | |
| 	/*
 | |
| 	 * It is important that console printing threads are scheduled
 | |
| 	 * shortly after a printk call and with generous runtime budgets.
 | |
| 	 */
 | |
| 	sched_set_normal(con->kthread, -20);
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| /* Track the nbcon emergency nesting per CPU. */
 | |
| static DEFINE_PER_CPU(unsigned int, nbcon_pcpu_emergency_nesting);
 | |
| static unsigned int early_nbcon_pcpu_emergency_nesting __initdata;
 | |
| 
 | |
| /**
 | |
|  * nbcon_get_cpu_emergency_nesting - Get the per CPU emergency nesting pointer
 | |
|  *
 | |
|  * Context:	For reading, any context. For writing, any context which could
 | |
|  *		not be migrated to another CPU.
 | |
|  * Return:	Either a pointer to the per CPU emergency nesting counter of
 | |
|  *		the current CPU or to the init data during early boot.
 | |
|  *
 | |
|  * The function is safe for reading per-CPU variables in any context because
 | |
|  * preemption is disabled if the current CPU is in the emergency state. See
 | |
|  * also nbcon_cpu_emergency_enter().
 | |
|  */
 | |
| static __ref unsigned int *nbcon_get_cpu_emergency_nesting(void)
 | |
| {
 | |
| 	/*
 | |
| 	 * The value of __printk_percpu_data_ready gets set in normal
 | |
| 	 * context and before SMP initialization. As a result it could
 | |
| 	 * never change while inside an nbcon emergency section.
 | |
| 	 */
 | |
| 	if (!printk_percpu_data_ready())
 | |
| 		return &early_nbcon_pcpu_emergency_nesting;
 | |
| 
 | |
| 	return raw_cpu_ptr(&nbcon_pcpu_emergency_nesting);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_get_default_prio - The appropriate nbcon priority to use for nbcon
 | |
|  *				printing on the current CPU
 | |
|  *
 | |
|  * Context:	Any context.
 | |
|  * Return:	The nbcon_prio to use for acquiring an nbcon console in this
 | |
|  *		context for printing.
 | |
|  *
 | |
|  * The function is safe for reading per-CPU data in any context because
 | |
|  * preemption is disabled if the current CPU is in the emergency or panic
 | |
|  * state.
 | |
|  */
 | |
| enum nbcon_prio nbcon_get_default_prio(void)
 | |
| {
 | |
| 	unsigned int *cpu_emergency_nesting;
 | |
| 
 | |
| 	if (this_cpu_in_panic())
 | |
| 		return NBCON_PRIO_PANIC;
 | |
| 
 | |
| 	cpu_emergency_nesting = nbcon_get_cpu_emergency_nesting();
 | |
| 	if (*cpu_emergency_nesting)
 | |
| 		return NBCON_PRIO_EMERGENCY;
 | |
| 
 | |
| 	return NBCON_PRIO_NORMAL;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_legacy_emit_next_record - Print one record for an nbcon console
 | |
|  *					in legacy contexts
 | |
|  * @con:	The console to print on
 | |
|  * @handover:	Will be set to true if a printk waiter has taken over the
 | |
|  *		console_lock, in which case the caller is no longer holding
 | |
|  *		both the console_lock and the SRCU read lock. Otherwise it
 | |
|  *		is set to false.
 | |
|  * @cookie:	The cookie from the SRCU read lock.
 | |
|  * @use_atomic: Set true when called in an atomic or unknown context.
 | |
|  *		It affects which nbcon callback will be used: write_atomic()
 | |
|  *		or write_thread().
 | |
|  *
 | |
|  *		When false, the write_thread() callback is used and would be
 | |
|  *		called in a preemtible context unless disabled by the
 | |
|  *		device_lock. The legacy handover is not allowed in this mode.
 | |
|  *
 | |
|  * Context:	Any context except NMI.
 | |
|  * Return:	True, when a record has been printed and there are still
 | |
|  *		pending records. The caller might want to continue flushing.
 | |
|  *
 | |
|  *		False, when there is no pending record, or when the console
 | |
|  *		context cannot be acquired, or the ownership has been lost.
 | |
|  *		The caller should give up. Either the job is done, cannot be
 | |
|  *		done, or will be handled by the owning context.
 | |
|  *
 | |
|  * This function is meant to be called by console_flush_all() to print records
 | |
|  * on nbcon consoles from legacy context (printing via console unlocking).
 | |
|  * Essentially it is the nbcon version of console_emit_next_record().
 | |
|  */
 | |
| bool nbcon_legacy_emit_next_record(struct console *con, bool *handover,
 | |
| 				   int cookie, bool use_atomic)
 | |
| {
 | |
| 	struct nbcon_write_context wctxt = { };
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(&wctxt, ctxt);
 | |
| 	unsigned long flags;
 | |
| 	bool progress;
 | |
| 
 | |
| 	ctxt->console	= con;
 | |
| 	ctxt->prio	= nbcon_get_default_prio();
 | |
| 
 | |
| 	if (use_atomic) {
 | |
| 		/*
 | |
| 		 * In an atomic or unknown context, use the same procedure as
 | |
| 		 * in console_emit_next_record(). It allows to handover.
 | |
| 		 */
 | |
| 		printk_safe_enter_irqsave(flags);
 | |
| 		console_lock_spinning_enable();
 | |
| 		stop_critical_timings();
 | |
| 	}
 | |
| 
 | |
| 	progress = nbcon_emit_one(&wctxt, use_atomic);
 | |
| 
 | |
| 	if (use_atomic) {
 | |
| 		start_critical_timings();
 | |
| 		*handover = console_lock_spinning_disable_and_check(cookie);
 | |
| 		printk_safe_exit_irqrestore(flags);
 | |
| 	} else {
 | |
| 		/* Non-atomic does not perform legacy spinning handovers. */
 | |
| 		*handover = false;
 | |
| 	}
 | |
| 
 | |
| 	return progress;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * __nbcon_atomic_flush_pending_con - Flush specified nbcon console using its
 | |
|  *					write_atomic() callback
 | |
|  * @con:			The nbcon console to flush
 | |
|  * @stop_seq:			Flush up until this record
 | |
|  * @allow_unsafe_takeover:	True, to allow unsafe hostile takeovers
 | |
|  *
 | |
|  * Return:	0 if @con was flushed up to @stop_seq Otherwise, error code on
 | |
|  *		failure.
 | |
|  *
 | |
|  * Errors:
 | |
|  *
 | |
|  *	-EPERM:		Unable to acquire console ownership.
 | |
|  *
 | |
|  *	-EAGAIN:	Another context took over ownership while printing.
 | |
|  *
 | |
|  *	-ENOENT:	A record before @stop_seq is not available.
 | |
|  *
 | |
|  * If flushing up to @stop_seq was not successful, it only makes sense for the
 | |
|  * caller to try again when -EAGAIN was returned. When -EPERM is returned,
 | |
|  * this context is not allowed to acquire the console. When -ENOENT is
 | |
|  * returned, it cannot be expected that the unfinalized record will become
 | |
|  * available.
 | |
|  */
 | |
| static int __nbcon_atomic_flush_pending_con(struct console *con, u64 stop_seq,
 | |
| 					    bool allow_unsafe_takeover)
 | |
| {
 | |
| 	struct nbcon_write_context wctxt = { };
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(&wctxt, ctxt);
 | |
| 	int err = 0;
 | |
| 
 | |
| 	ctxt->console			= con;
 | |
| 	ctxt->spinwait_max_us		= 2000;
 | |
| 	ctxt->prio			= nbcon_get_default_prio();
 | |
| 	ctxt->allow_unsafe_takeover	= allow_unsafe_takeover;
 | |
| 
 | |
| 	if (!nbcon_context_try_acquire(ctxt, false))
 | |
| 		return -EPERM;
 | |
| 
 | |
| 	while (nbcon_seq_read(con) < stop_seq) {
 | |
| 		/*
 | |
| 		 * nbcon_emit_next_record() returns false when the console was
 | |
| 		 * handed over or taken over. In both cases the context is no
 | |
| 		 * longer valid.
 | |
| 		 */
 | |
| 		if (!nbcon_emit_next_record(&wctxt, true))
 | |
| 			return -EAGAIN;
 | |
| 
 | |
| 		if (!ctxt->backlog) {
 | |
| 			/* Are there reserved but not yet finalized records? */
 | |
| 			if (nbcon_seq_read(con) < stop_seq)
 | |
| 				err = -ENOENT;
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	nbcon_context_release(ctxt);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_atomic_flush_pending_con - Flush specified nbcon console using its
 | |
|  *					write_atomic() callback
 | |
|  * @con:			The nbcon console to flush
 | |
|  * @stop_seq:			Flush up until this record
 | |
|  * @allow_unsafe_takeover:	True, to allow unsafe hostile takeovers
 | |
|  *
 | |
|  * This will stop flushing before @stop_seq if another context has ownership.
 | |
|  * That context is then responsible for the flushing. Likewise, if new records
 | |
|  * are added while this context was flushing and there is no other context
 | |
|  * to handle the printing, this context must also flush those records.
 | |
|  */
 | |
| static void nbcon_atomic_flush_pending_con(struct console *con, u64 stop_seq,
 | |
| 					   bool allow_unsafe_takeover)
 | |
| {
 | |
| 	struct console_flush_type ft;
 | |
| 	unsigned long flags;
 | |
| 	int err;
 | |
| 
 | |
| again:
 | |
| 	/*
 | |
| 	 * Atomic flushing does not use console driver synchronization (i.e.
 | |
| 	 * it does not hold the port lock for uart consoles). Therefore IRQs
 | |
| 	 * must be disabled to avoid being interrupted and then calling into
 | |
| 	 * a driver that will deadlock trying to acquire console ownership.
 | |
| 	 */
 | |
| 	local_irq_save(flags);
 | |
| 
 | |
| 	err = __nbcon_atomic_flush_pending_con(con, stop_seq, allow_unsafe_takeover);
 | |
| 
 | |
| 	local_irq_restore(flags);
 | |
| 
 | |
| 	/*
 | |
| 	 * If there was a new owner (-EPERM, -EAGAIN), that context is
 | |
| 	 * responsible for completing.
 | |
| 	 *
 | |
| 	 * Do not wait for records not yet finalized (-ENOENT) to avoid a
 | |
| 	 * possible deadlock. They will either get flushed by the writer or
 | |
| 	 * eventually skipped on panic CPU.
 | |
| 	 */
 | |
| 	if (err)
 | |
| 		return;
 | |
| 
 | |
| 	/*
 | |
| 	 * If flushing was successful but more records are available, this
 | |
| 	 * context must flush those remaining records if the printer thread
 | |
| 	 * is not available do it.
 | |
| 	 */
 | |
| 	printk_get_console_flush_type(&ft);
 | |
| 	if (!ft.nbcon_offload &&
 | |
| 	    prb_read_valid(prb, nbcon_seq_read(con), NULL)) {
 | |
| 		stop_seq = prb_next_reserve_seq(prb);
 | |
| 		goto again;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * __nbcon_atomic_flush_pending - Flush all nbcon consoles using their
 | |
|  *					write_atomic() callback
 | |
|  * @stop_seq:			Flush up until this record
 | |
|  * @allow_unsafe_takeover:	True, to allow unsafe hostile takeovers
 | |
|  */
 | |
| static void __nbcon_atomic_flush_pending(u64 stop_seq, bool allow_unsafe_takeover)
 | |
| {
 | |
| 	struct console *con;
 | |
| 	int cookie;
 | |
| 
 | |
| 	cookie = console_srcu_read_lock();
 | |
| 	for_each_console_srcu(con) {
 | |
| 		short flags = console_srcu_read_flags(con);
 | |
| 
 | |
| 		if (!(flags & CON_NBCON))
 | |
| 			continue;
 | |
| 
 | |
| 		if (!console_is_usable(con, flags, true))
 | |
| 			continue;
 | |
| 
 | |
| 		if (nbcon_seq_read(con) >= stop_seq)
 | |
| 			continue;
 | |
| 
 | |
| 		nbcon_atomic_flush_pending_con(con, stop_seq, allow_unsafe_takeover);
 | |
| 	}
 | |
| 	console_srcu_read_unlock(cookie);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_atomic_flush_pending - Flush all nbcon consoles using their
 | |
|  *				write_atomic() callback
 | |
|  *
 | |
|  * Flush the backlog up through the currently newest record. Any new
 | |
|  * records added while flushing will not be flushed if there is another
 | |
|  * context available to handle the flushing. This is to avoid one CPU
 | |
|  * printing unbounded because other CPUs continue to add records.
 | |
|  */
 | |
| void nbcon_atomic_flush_pending(void)
 | |
| {
 | |
| 	__nbcon_atomic_flush_pending(prb_next_reserve_seq(prb), false);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_atomic_flush_unsafe - Flush all nbcon consoles using their
 | |
|  *	write_atomic() callback and allowing unsafe hostile takeovers
 | |
|  *
 | |
|  * Flush the backlog up through the currently newest record. Unsafe hostile
 | |
|  * takeovers will be performed, if necessary.
 | |
|  */
 | |
| void nbcon_atomic_flush_unsafe(void)
 | |
| {
 | |
| 	__nbcon_atomic_flush_pending(prb_next_reserve_seq(prb), true);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_cpu_emergency_enter - Enter an emergency section where printk()
 | |
|  *				messages for that CPU are flushed directly
 | |
|  *
 | |
|  * Context:	Any context. Disables preemption.
 | |
|  *
 | |
|  * When within an emergency section, printk() calls will attempt to flush any
 | |
|  * pending messages in the ringbuffer.
 | |
|  */
 | |
| void nbcon_cpu_emergency_enter(void)
 | |
| {
 | |
| 	unsigned int *cpu_emergency_nesting;
 | |
| 
 | |
| 	preempt_disable();
 | |
| 
 | |
| 	cpu_emergency_nesting = nbcon_get_cpu_emergency_nesting();
 | |
| 	(*cpu_emergency_nesting)++;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_cpu_emergency_exit - Exit an emergency section
 | |
|  *
 | |
|  * Context:	Within an emergency section. Enables preemption.
 | |
|  */
 | |
| void nbcon_cpu_emergency_exit(void)
 | |
| {
 | |
| 	unsigned int *cpu_emergency_nesting;
 | |
| 
 | |
| 	cpu_emergency_nesting = nbcon_get_cpu_emergency_nesting();
 | |
| 
 | |
| 	if (!WARN_ON_ONCE(*cpu_emergency_nesting == 0))
 | |
| 		(*cpu_emergency_nesting)--;
 | |
| 
 | |
| 	preempt_enable();
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_alloc - Allocate and init the nbcon console specific data
 | |
|  * @con:	Console to initialize
 | |
|  *
 | |
|  * Return:	True if the console was fully allocated and initialized.
 | |
|  *		Otherwise @con must not be registered.
 | |
|  *
 | |
|  * When allocation and init was successful, the console must be properly
 | |
|  * freed using nbcon_free() once it is no longer needed.
 | |
|  */
 | |
| bool nbcon_alloc(struct console *con)
 | |
| {
 | |
| 	struct nbcon_state state = { };
 | |
| 
 | |
| 	/* Synchronize the kthread start. */
 | |
| 	lockdep_assert_console_list_lock_held();
 | |
| 
 | |
| 	/* The write_thread() callback is mandatory. */
 | |
| 	if (WARN_ON(!con->write_thread))
 | |
| 		return false;
 | |
| 
 | |
| 	rcuwait_init(&con->rcuwait);
 | |
| 	init_irq_work(&con->irq_work, nbcon_irq_work);
 | |
| 	atomic_long_set(&ACCESS_PRIVATE(con, nbcon_prev_seq), -1UL);
 | |
| 	nbcon_state_set(con, &state);
 | |
| 
 | |
| 	/*
 | |
| 	 * Initialize @nbcon_seq to the highest possible sequence number so
 | |
| 	 * that practically speaking it will have nothing to print until a
 | |
| 	 * desired initial sequence number has been set via nbcon_seq_force().
 | |
| 	 */
 | |
| 	atomic_long_set(&ACCESS_PRIVATE(con, nbcon_seq), ULSEQ_MAX(prb));
 | |
| 
 | |
| 	if (con->flags & CON_BOOT) {
 | |
| 		/*
 | |
| 		 * Boot console printing is synchronized with legacy console
 | |
| 		 * printing, so boot consoles can share the same global printk
 | |
| 		 * buffers.
 | |
| 		 */
 | |
| 		con->pbufs = &printk_shared_pbufs;
 | |
| 	} else {
 | |
| 		con->pbufs = kmalloc(sizeof(*con->pbufs), GFP_KERNEL);
 | |
| 		if (!con->pbufs) {
 | |
| 			con_printk(KERN_ERR, con, "failed to allocate printing buffer\n");
 | |
| 			return false;
 | |
| 		}
 | |
| 
 | |
| 		if (printk_kthreads_ready && !have_boot_console) {
 | |
| 			if (!nbcon_kthread_create(con)) {
 | |
| 				kfree(con->pbufs);
 | |
| 				con->pbufs = NULL;
 | |
| 				return false;
 | |
| 			}
 | |
| 
 | |
| 			/* Might be the first kthread. */
 | |
| 			printk_kthreads_running = true;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_free - Free and cleanup the nbcon console specific data
 | |
|  * @con:	Console to free/cleanup nbcon data
 | |
|  *
 | |
|  * Important: @have_nbcon_console must be updated before calling
 | |
|  *	this function. In particular, it can be set only when there
 | |
|  *	is still another nbcon console registered.
 | |
|  */
 | |
| void nbcon_free(struct console *con)
 | |
| {
 | |
| 	struct nbcon_state state = { };
 | |
| 
 | |
| 	/* Synchronize the kthread stop. */
 | |
| 	lockdep_assert_console_list_lock_held();
 | |
| 
 | |
| 	if (printk_kthreads_running) {
 | |
| 		nbcon_kthread_stop(con);
 | |
| 
 | |
| 		/* Might be the last nbcon console.
 | |
| 		 *
 | |
| 		 * Do not rely on printk_kthreads_check_locked(). It is not
 | |
| 		 * called in some code paths, see nbcon_free() callers.
 | |
| 		 */
 | |
| 		if (!have_nbcon_console)
 | |
| 			printk_kthreads_running = false;
 | |
| 	}
 | |
| 
 | |
| 	nbcon_state_set(con, &state);
 | |
| 
 | |
| 	/* Boot consoles share global printk buffers. */
 | |
| 	if (!(con->flags & CON_BOOT))
 | |
| 		kfree(con->pbufs);
 | |
| 
 | |
| 	con->pbufs = NULL;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nbcon_device_try_acquire - Try to acquire nbcon console and enter unsafe
 | |
|  *				section
 | |
|  * @con:	The nbcon console to acquire
 | |
|  *
 | |
|  * Context:	Under the locking mechanism implemented in
 | |
|  *		@con->device_lock() including disabling migration.
 | |
|  * Return:	True if the console was acquired. False otherwise.
 | |
|  *
 | |
|  * Console drivers will usually use their own internal synchronization
 | |
|  * mechasism to synchronize between console printing and non-printing
 | |
|  * activities (such as setting baud rates). However, nbcon console drivers
 | |
|  * supporting atomic consoles may also want to mark unsafe sections when
 | |
|  * performing non-printing activities in order to synchronize against their
 | |
|  * atomic_write() callback.
 | |
|  *
 | |
|  * This function acquires the nbcon console using priority NBCON_PRIO_NORMAL
 | |
|  * and marks it unsafe for handover/takeover.
 | |
|  */
 | |
| bool nbcon_device_try_acquire(struct console *con)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(con, nbcon_device_ctxt);
 | |
| 
 | |
| 	cant_migrate();
 | |
| 
 | |
| 	memset(ctxt, 0, sizeof(*ctxt));
 | |
| 	ctxt->console	= con;
 | |
| 	ctxt->prio	= NBCON_PRIO_NORMAL;
 | |
| 
 | |
| 	if (!nbcon_context_try_acquire(ctxt, false))
 | |
| 		return false;
 | |
| 
 | |
| 	if (!nbcon_context_enter_unsafe(ctxt))
 | |
| 		return false;
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(nbcon_device_try_acquire);
 | |
| 
 | |
| /**
 | |
|  * nbcon_device_release - Exit unsafe section and release the nbcon console
 | |
|  * @con:	The nbcon console acquired in nbcon_device_try_acquire()
 | |
|  */
 | |
| void nbcon_device_release(struct console *con)
 | |
| {
 | |
| 	struct nbcon_context *ctxt = &ACCESS_PRIVATE(con, nbcon_device_ctxt);
 | |
| 	struct console_flush_type ft;
 | |
| 	int cookie;
 | |
| 
 | |
| 	if (!nbcon_context_exit_unsafe(ctxt))
 | |
| 		return;
 | |
| 
 | |
| 	nbcon_context_release(ctxt);
 | |
| 
 | |
| 	/*
 | |
| 	 * This context must flush any new records added while the console
 | |
| 	 * was locked if the printer thread is not available to do it. The
 | |
| 	 * console_srcu_read_lock must be taken to ensure the console is
 | |
| 	 * usable throughout flushing.
 | |
| 	 */
 | |
| 	cookie = console_srcu_read_lock();
 | |
| 	printk_get_console_flush_type(&ft);
 | |
| 	if (console_is_usable(con, console_srcu_read_flags(con), true) &&
 | |
| 	    !ft.nbcon_offload &&
 | |
| 	    prb_read_valid(prb, nbcon_seq_read(con), NULL)) {
 | |
| 		/*
 | |
| 		 * If nbcon_atomic flushing is not available, fallback to
 | |
| 		 * using the legacy loop.
 | |
| 		 */
 | |
| 		if (ft.nbcon_atomic) {
 | |
| 			__nbcon_atomic_flush_pending_con(con, prb_next_reserve_seq(prb), false);
 | |
| 		} else if (ft.legacy_direct) {
 | |
| 			if (console_trylock())
 | |
| 				console_unlock();
 | |
| 		} else if (ft.legacy_offload) {
 | |
| 			printk_trigger_flush();
 | |
| 		}
 | |
| 	}
 | |
| 	console_srcu_read_unlock(cookie);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(nbcon_device_release);
 | 
