mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Impact: change existing irq_chip API Not much point with gentle transition here: the struct irq_chip's setaffinity method signature needs to change. Fortunately, not widely used code, but hits a few architectures. Note: In irq_select_affinity() I save a temporary in by mangling irq_desc[irq].affinity directly. Ingo, does this break anything? (Folded in fix from KOSAKI Motohiro) Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> Signed-off-by: Mike Travis <travis@sgi.com> Reviewed-by: Grant Grundler <grundler@parisc-linux.org> Acked-by: Ingo Molnar <mingo@redhat.com> Cc: ralf@linux-mips.org Cc: grundler@parisc-linux.org Cc: jeremy@xensource.com Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
		
			
				
	
	
		
			64 lines
		
	
	
	
		
			1.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			64 lines
		
	
	
	
		
			1.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
 | 
						|
#include <linux/irq.h>
 | 
						|
 | 
						|
void move_masked_irq(int irq)
 | 
						|
{
 | 
						|
	struct irq_desc *desc = irq_to_desc(irq);
 | 
						|
 | 
						|
	if (likely(!(desc->status & IRQ_MOVE_PENDING)))
 | 
						|
		return;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
 | 
						|
	 */
 | 
						|
	if (CHECK_IRQ_PER_CPU(desc->status)) {
 | 
						|
		WARN_ON(1);
 | 
						|
		return;
 | 
						|
	}
 | 
						|
 | 
						|
	desc->status &= ~IRQ_MOVE_PENDING;
 | 
						|
 | 
						|
	if (unlikely(cpumask_empty(&desc->pending_mask)))
 | 
						|
		return;
 | 
						|
 | 
						|
	if (!desc->chip->set_affinity)
 | 
						|
		return;
 | 
						|
 | 
						|
	assert_spin_locked(&desc->lock);
 | 
						|
 | 
						|
	/*
 | 
						|
	 * If there was a valid mask to work with, please
 | 
						|
	 * do the disable, re-program, enable sequence.
 | 
						|
	 * This is *not* particularly important for level triggered
 | 
						|
	 * but in a edge trigger case, we might be setting rte
 | 
						|
	 * when an active trigger is comming in. This could
 | 
						|
	 * cause some ioapics to mal-function.
 | 
						|
	 * Being paranoid i guess!
 | 
						|
	 *
 | 
						|
	 * For correct operation this depends on the caller
 | 
						|
	 * masking the irqs.
 | 
						|
	 */
 | 
						|
	if (likely(cpumask_any_and(&desc->pending_mask, cpu_online_mask)
 | 
						|
		   < nr_cpu_ids)) {
 | 
						|
		cpumask_and(&desc->affinity,
 | 
						|
			    &desc->pending_mask, cpu_online_mask);
 | 
						|
		desc->chip->set_affinity(irq, &desc->affinity);
 | 
						|
	}
 | 
						|
	cpumask_clear(&desc->pending_mask);
 | 
						|
}
 | 
						|
 | 
						|
void move_native_irq(int irq)
 | 
						|
{
 | 
						|
	struct irq_desc *desc = irq_to_desc(irq);
 | 
						|
 | 
						|
	if (likely(!(desc->status & IRQ_MOVE_PENDING)))
 | 
						|
		return;
 | 
						|
 | 
						|
	if (unlikely(desc->status & IRQ_DISABLED))
 | 
						|
		return;
 | 
						|
 | 
						|
	desc->chip->mask(irq);
 | 
						|
	move_masked_irq(irq);
 | 
						|
	desc->chip->unmask(irq);
 | 
						|
}
 | 
						|
 |