|  | /* | 
|  | * The idle loop for all SuperH platforms. | 
|  | * | 
|  | *  Copyright (C) 2002 - 2009  Paul Mundt | 
|  | * | 
|  | * This file is subject to the terms and conditions of the GNU General Public | 
|  | * License.  See the file "COPYING" in the main directory of this archive | 
|  | * for more details. | 
|  | */ | 
|  | #include <linux/module.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/mm.h> | 
|  | #include <linux/pm.h> | 
|  | #include <linux/tick.h> | 
|  | #include <linux/preempt.h> | 
|  | #include <linux/thread_info.h> | 
|  | #include <linux/irqflags.h> | 
|  | #include <linux/smp.h> | 
|  | #include <linux/cpuidle.h> | 
|  | #include <linux/atomic.h> | 
|  | #include <asm/pgalloc.h> | 
|  | #include <asm/smp.h> | 
|  | #include <asm/bl_bit.h> | 
|  |  | 
|  | void (*pm_idle)(void); | 
|  |  | 
|  | static int hlt_counter; | 
|  |  | 
|  | static int __init nohlt_setup(char *__unused) | 
|  | { | 
|  | hlt_counter = 1; | 
|  | return 1; | 
|  | } | 
|  | __setup("nohlt", nohlt_setup); | 
|  |  | 
|  | static int __init hlt_setup(char *__unused) | 
|  | { | 
|  | hlt_counter = 0; | 
|  | return 1; | 
|  | } | 
|  | __setup("hlt", hlt_setup); | 
|  |  | 
|  | static inline int hlt_works(void) | 
|  | { | 
|  | return !hlt_counter; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * On SMP it's slightly faster (but much more power-consuming!) | 
|  | * to poll the ->work.need_resched flag instead of waiting for the | 
|  | * cross-CPU IPI to arrive. Use this option with caution. | 
|  | */ | 
|  | static void poll_idle(void) | 
|  | { | 
|  | local_irq_enable(); | 
|  | while (!need_resched()) | 
|  | cpu_relax(); | 
|  | } | 
|  |  | 
|  | void default_idle(void) | 
|  | { | 
|  | if (hlt_works()) { | 
|  | clear_thread_flag(TIF_POLLING_NRFLAG); | 
|  | smp_mb__after_clear_bit(); | 
|  |  | 
|  | set_bl_bit(); | 
|  | if (!need_resched()) { | 
|  | local_irq_enable(); | 
|  | cpu_sleep(); | 
|  | } else | 
|  | local_irq_enable(); | 
|  |  | 
|  | set_thread_flag(TIF_POLLING_NRFLAG); | 
|  | clear_bl_bit(); | 
|  | } else | 
|  | poll_idle(); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * The idle thread. There's no useful work to be done, so just try to conserve | 
|  | * power and have a low exit latency (ie sit in a loop waiting for somebody to | 
|  | * say that they'd like to reschedule) | 
|  | */ | 
|  | void cpu_idle(void) | 
|  | { | 
|  | unsigned int cpu = smp_processor_id(); | 
|  |  | 
|  | set_thread_flag(TIF_POLLING_NRFLAG); | 
|  |  | 
|  | /* endless idle loop with no priority at all */ | 
|  | while (1) { | 
|  | tick_nohz_idle_enter(); | 
|  | rcu_idle_enter(); | 
|  |  | 
|  | while (!need_resched()) { | 
|  | check_pgt_cache(); | 
|  | rmb(); | 
|  |  | 
|  | if (cpu_is_offline(cpu)) | 
|  | play_dead(); | 
|  |  | 
|  | local_irq_disable(); | 
|  | /* Don't trace irqs off for idle */ | 
|  | stop_critical_timings(); | 
|  | if (cpuidle_idle_call()) | 
|  | pm_idle(); | 
|  | /* | 
|  | * Sanity check to ensure that pm_idle() returns | 
|  | * with IRQs enabled | 
|  | */ | 
|  | WARN_ON(irqs_disabled()); | 
|  | start_critical_timings(); | 
|  | } | 
|  |  | 
|  | rcu_idle_exit(); | 
|  | tick_nohz_idle_exit(); | 
|  | schedule_preempt_disabled(); | 
|  | } | 
|  | } | 
|  |  | 
|  | void __init select_idle_routine(void) | 
|  | { | 
|  | /* | 
|  | * If a platform has set its own idle routine, leave it alone. | 
|  | */ | 
|  | if (pm_idle) | 
|  | return; | 
|  |  | 
|  | if (hlt_works()) | 
|  | pm_idle = default_idle; | 
|  | else | 
|  | pm_idle = poll_idle; | 
|  | } | 
|  |  | 
|  | void stop_this_cpu(void *unused) | 
|  | { | 
|  | local_irq_disable(); | 
|  | set_cpu_online(smp_processor_id(), false); | 
|  |  | 
|  | for (;;) | 
|  | cpu_sleep(); | 
|  | } |