forked from mirrors/linux
		
	x86: Use generic idle loop
Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Rusty Russell <rusty@rustcorp.com.au> Cc: Paul McKenney <paulmck@linux.vnet.ibm.com> Cc: Peter Zijlstra <peterz@infradead.org> Reviewed-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> Cc: Magnus Damm <magnus.damm@gmail.com> Link: http://lkml.kernel.org/r/20130321215235.486594473@linutronix.de Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: x86@kernel.org
This commit is contained in:
		
							parent
							
								
									aba92c9e2c
								
							
						
					
					
						commit
						7d1a941731
					
				
					 4 changed files with 31 additions and 81 deletions
				
			
		|  | @ -97,6 +97,7 @@ config X86 | ||||||
| 	select GENERIC_IOMAP | 	select GENERIC_IOMAP | ||||||
| 	select DCACHE_WORD_ACCESS | 	select DCACHE_WORD_ACCESS | ||||||
| 	select GENERIC_SMP_IDLE_THREAD | 	select GENERIC_SMP_IDLE_THREAD | ||||||
|  | 	select GENERIC_IDLE_LOOP | ||||||
| 	select ARCH_WANT_IPC_PARSE_VERSION if X86_32 | 	select ARCH_WANT_IPC_PARSE_VERSION if X86_32 | ||||||
| 	select HAVE_ARCH_SECCOMP_FILTER | 	select HAVE_ARCH_SECCOMP_FILTER | ||||||
| 	select BUILDTIME_EXTABLE_SORT | 	select BUILDTIME_EXTABLE_SORT | ||||||
|  |  | ||||||
|  | @ -301,13 +301,7 @@ void exit_idle(void) | ||||||
| } | } | ||||||
| #endif | #endif | ||||||
| 
 | 
 | ||||||
| /*
 | void arch_cpu_idle_prepare(void) | ||||||
|  * The idle thread. There's no useful work to be |  | ||||||
|  * done, so just try to conserve power and have a |  | ||||||
|  * low exit latency (ie sit in a loop waiting for |  | ||||||
|  * somebody to say that they'd like to reschedule) |  | ||||||
|  */ |  | ||||||
| void cpu_idle(void) |  | ||||||
| { | { | ||||||
| 	/*
 | 	/*
 | ||||||
| 	 * If we're the non-boot CPU, nothing set the stack canary up | 	 * If we're the non-boot CPU, nothing set the stack canary up | ||||||
|  | @ -317,71 +311,40 @@ void cpu_idle(void) | ||||||
| 	 * canaries already on the stack wont ever trigger). | 	 * canaries already on the stack wont ever trigger). | ||||||
| 	 */ | 	 */ | ||||||
| 	boot_init_stack_canary(); | 	boot_init_stack_canary(); | ||||||
| 	current_thread_info()->status |= TS_POLLING; | } | ||||||
| 
 | 
 | ||||||
| 	while (1) { | void arch_cpu_idle_enter(void) | ||||||
| 		tick_nohz_idle_enter(); | { | ||||||
| 
 |  | ||||||
| 		while (!need_resched()) { |  | ||||||
| 			rmb(); |  | ||||||
| 
 |  | ||||||
| 			if (cpu_is_offline(smp_processor_id())) |  | ||||||
| 				play_dead(); |  | ||||||
| 
 |  | ||||||
| 			/*
 |  | ||||||
| 			 * Idle routines should keep interrupts disabled |  | ||||||
| 			 * from here on, until they go to idle. |  | ||||||
| 			 * Otherwise, idle callbacks can misfire. |  | ||||||
| 			 */ |  | ||||||
| 	local_touch_nmi(); | 	local_touch_nmi(); | ||||||
| 			local_irq_disable(); |  | ||||||
| 
 |  | ||||||
| 	enter_idle(); | 	enter_idle(); | ||||||
|  | } | ||||||
| 
 | 
 | ||||||
| 			/* Don't trace irqs off for idle */ | void arch_cpu_idle_exit(void) | ||||||
| 			stop_critical_timings(); | { | ||||||
| 
 |  | ||||||
| 			/* enter_idle() needs rcu for notifiers */ |  | ||||||
| 			rcu_idle_enter(); |  | ||||||
| 
 |  | ||||||
| 			if (cpuidle_idle_call()) |  | ||||||
| 				x86_idle(); |  | ||||||
| 
 |  | ||||||
| 			rcu_idle_exit(); |  | ||||||
| 			start_critical_timings(); |  | ||||||
| 
 |  | ||||||
| 			/* In many cases the interrupt that ended idle
 |  | ||||||
| 			   has already called exit_idle. But some idle |  | ||||||
| 			   loops can be woken up without interrupt. */ |  | ||||||
| 	__exit_idle(); | 	__exit_idle(); | ||||||
| 		} | } | ||||||
| 
 | 
 | ||||||
| 		tick_nohz_idle_exit(); | void arch_cpu_idle_dead(void) | ||||||
| 		preempt_enable_no_resched(); | { | ||||||
| 		schedule(); | 	play_dead(); | ||||||
| 		preempt_disable(); |  | ||||||
| 	} |  | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /*
 | /*
 | ||||||
|  * We use this if we don't have any better |  * Called from the generic idle code. | ||||||
|  * idle routine.. |  */ | ||||||
|  | void arch_cpu_idle(void) | ||||||
|  | { | ||||||
|  | 	if (cpuidle_idle_call()) | ||||||
|  | 		x86_idle(); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | /*
 | ||||||
|  |  * We use this if we don't have any better idle routine.. | ||||||
|  */ |  */ | ||||||
| void default_idle(void) | void default_idle(void) | ||||||
| { | { | ||||||
| 	trace_cpu_idle_rcuidle(1, smp_processor_id()); | 	trace_cpu_idle_rcuidle(1, smp_processor_id()); | ||||||
| 	current_thread_info()->status &= ~TS_POLLING; | 	safe_halt(); | ||||||
| 	/*
 |  | ||||||
| 	 * TS_POLLING-cleared state must be visible before we |  | ||||||
| 	 * test NEED_RESCHED: |  | ||||||
| 	 */ |  | ||||||
| 	smp_mb(); |  | ||||||
| 
 |  | ||||||
| 	if (!need_resched()) |  | ||||||
| 		safe_halt();	/* enables interrupts racelessly */ |  | ||||||
| 	else |  | ||||||
| 		local_irq_enable(); |  | ||||||
| 	current_thread_info()->status |= TS_POLLING; |  | ||||||
| 	trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id()); | 	trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id()); | ||||||
| } | } | ||||||
| #ifdef CONFIG_APM_MODULE | #ifdef CONFIG_APM_MODULE | ||||||
|  | @ -411,20 +374,6 @@ void stop_this_cpu(void *dummy) | ||||||
| 		halt(); | 		halt(); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /*
 |  | ||||||
|  * On SMP it's slightly faster (but much more power-consuming!) |  | ||||||
|  * to poll the ->work.need_resched flag instead of waiting for the |  | ||||||
|  * cross-CPU IPI to arrive. Use this option with caution. |  | ||||||
|  */ |  | ||||||
| static void poll_idle(void) |  | ||||||
| { |  | ||||||
| 	trace_cpu_idle_rcuidle(0, smp_processor_id()); |  | ||||||
| 	local_irq_enable(); |  | ||||||
| 	while (!need_resched()) |  | ||||||
| 		cpu_relax(); |  | ||||||
| 	trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id()); |  | ||||||
| } |  | ||||||
| 
 |  | ||||||
| bool amd_e400_c1e_detected; | bool amd_e400_c1e_detected; | ||||||
| EXPORT_SYMBOL(amd_e400_c1e_detected); | EXPORT_SYMBOL(amd_e400_c1e_detected); | ||||||
| 
 | 
 | ||||||
|  | @ -489,10 +438,10 @@ static void amd_e400_idle(void) | ||||||
| void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c) | void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c) | ||||||
| { | { | ||||||
| #ifdef CONFIG_SMP | #ifdef CONFIG_SMP | ||||||
| 	if (x86_idle == poll_idle && smp_num_siblings > 1) | 	if (boot_option_idle_override == IDLE_POLL && smp_num_siblings > 1) | ||||||
| 		pr_warn_once("WARNING: polling idle and HT enabled, performance may degrade\n"); | 		pr_warn_once("WARNING: polling idle and HT enabled, performance may degrade\n"); | ||||||
| #endif | #endif | ||||||
| 	if (x86_idle) | 	if (x86_idle || boot_option_idle_override == IDLE_POLL) | ||||||
| 		return; | 		return; | ||||||
| 
 | 
 | ||||||
| 	if (cpu_has_amd_erratum(amd_erratum_400)) { | 	if (cpu_has_amd_erratum(amd_erratum_400)) { | ||||||
|  | @ -517,8 +466,8 @@ static int __init idle_setup(char *str) | ||||||
| 
 | 
 | ||||||
| 	if (!strcmp(str, "poll")) { | 	if (!strcmp(str, "poll")) { | ||||||
| 		pr_info("using polling idle threads\n"); | 		pr_info("using polling idle threads\n"); | ||||||
| 		x86_idle = poll_idle; |  | ||||||
| 		boot_option_idle_override = IDLE_POLL; | 		boot_option_idle_override = IDLE_POLL; | ||||||
|  | 		cpu_idle_poll_ctrl(true); | ||||||
| 	} else if (!strcmp(str, "halt")) { | 	} else if (!strcmp(str, "halt")) { | ||||||
| 		/*
 | 		/*
 | ||||||
| 		 * When the boot option of idle=halt is added, halt is | 		 * When the boot option of idle=halt is added, halt is | ||||||
|  |  | ||||||
|  | @ -284,7 +284,7 @@ notrace static void __cpuinit start_secondary(void *unused) | ||||||
| 	x86_cpuinit.setup_percpu_clockev(); | 	x86_cpuinit.setup_percpu_clockev(); | ||||||
| 
 | 
 | ||||||
| 	wmb(); | 	wmb(); | ||||||
| 	cpu_idle(); | 	cpu_startup_entry(CPUHP_ONLINE); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| void __init smp_store_boot_cpu_info(void) | void __init smp_store_boot_cpu_info(void) | ||||||
|  |  | ||||||
|  | @ -95,7 +95,7 @@ static void __cpuinit cpu_bringup(void) | ||||||
| static void __cpuinit cpu_bringup_and_idle(void) | static void __cpuinit cpu_bringup_and_idle(void) | ||||||
| { | { | ||||||
| 	cpu_bringup(); | 	cpu_bringup(); | ||||||
| 	cpu_idle(); | 	cpu_startup_entry(CPUHP_ONLINE); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int xen_smp_intr_init(unsigned int cpu) | static int xen_smp_intr_init(unsigned int cpu) | ||||||
|  |  | ||||||
		Loading…
	
		Reference in a new issue
	
	 Thomas Gleixner
						Thomas Gleixner