Loading arch/arm/kernel/process.c +4 −9 Original line number Diff line number Diff line Loading @@ -172,14 +172,9 @@ static void default_idle(void) local_irq_enable(); } void (*pm_idle)(void) = default_idle; EXPORT_SYMBOL(pm_idle); /* * The idle thread, has rather strange semantics for calling pm_idle, * but this is what x86 does and we need to do the same, so that * things like cpuidle get called in the same way. The only difference * is that we always respect 'hlt_counter' to prevent low power idle. * The idle thread. * We always respect 'hlt_counter' to prevent low power idle. */ void cpu_idle(void) { Loading Loading @@ -210,10 +205,10 @@ void cpu_idle(void) } else if (!need_resched()) { stop_critical_timings(); if (cpuidle_idle_call()) pm_idle(); default_idle(); start_critical_timings(); /* * pm_idle functions must always * default_idle functions must always * return with IRQs enabled. */ WARN_ON(irqs_disabled()); Loading arch/arm64/kernel/process.c +4 −9 Original line number Diff line number Diff line Loading @@ -97,14 +97,9 @@ static void default_idle(void) local_irq_enable(); } void (*pm_idle)(void) = default_idle; EXPORT_SYMBOL_GPL(pm_idle); /* * The idle thread, has rather strange semantics for calling pm_idle, * but this is what x86 does and we need to do the same, so that * things like cpuidle get called in the same way. The only difference * is that we always respect 'hlt_counter' to prevent low power idle. * The idle thread. * We always respect 'hlt_counter' to prevent low power idle. */ void cpu_idle(void) { Loading @@ -122,10 +117,10 @@ void cpu_idle(void) local_irq_disable(); if (!need_resched()) { stop_critical_timings(); pm_idle(); default_idle(); start_critical_timings(); /* * pm_idle functions should always return * default_idle functions should always return * with IRQs enabled. */ WARN_ON(irqs_disabled()); Loading arch/blackfin/kernel/process.c +0 −7 Original line number Diff line number Diff line Loading @@ -39,12 +39,6 @@ int nr_l1stack_tasks; void *l1_stack_base; unsigned long l1_stack_len; /* * Powermanagement idle function, if any.. */ void (*pm_idle)(void) = NULL; EXPORT_SYMBOL(pm_idle); void (*pm_power_off)(void) = NULL; EXPORT_SYMBOL(pm_power_off); Loading Loading @@ -81,7 +75,6 @@ void cpu_idle(void) { /* endless idle loop with no priority at all */ while (1) { void (*idle)(void) = pm_idle; #ifdef CONFIG_HOTPLUG_CPU if (cpu_is_offline(smp_processor_id())) Loading arch/cris/kernel/process.c +1 −10 Original line number Diff line number Diff line Loading @@ -54,11 +54,6 @@ void enable_hlt(void) EXPORT_SYMBOL(enable_hlt); /* * The following aren't currently used. */ void (*pm_idle)(void); extern void default_idle(void); void (*pm_power_off)(void); Loading @@ -77,16 +72,12 @@ void cpu_idle (void) while (1) { rcu_idle_enter(); while (!need_resched()) { void (*idle)(void); /* * Mark this as an RCU critical section so that * synchronize_kernel() in the unload path waits * for our completion. */ idle = pm_idle; if (!idle) idle = default_idle; idle(); default_idle(); } rcu_idle_exit(); schedule_preempt_disabled(); Loading arch/ia64/kernel/process.c +0 −3 Original line number Diff line number Diff line Loading @@ -57,8 +57,6 @@ void (*ia64_mark_idle)(int); unsigned long boot_option_idle_override = IDLE_NO_OVERRIDE; EXPORT_SYMBOL(boot_option_idle_override); void (*pm_idle) (void); EXPORT_SYMBOL(pm_idle); void (*pm_power_off) (void); EXPORT_SYMBOL(pm_power_off); Loading Loading @@ -301,7 +299,6 @@ cpu_idle (void) if (mark_idle) (*mark_idle)(1); idle = pm_idle; if (!idle) idle = default_idle; (*idle)(); Loading Loading
arch/arm/kernel/process.c +4 −9 Original line number Diff line number Diff line Loading @@ -172,14 +172,9 @@ static void default_idle(void) local_irq_enable(); } void (*pm_idle)(void) = default_idle; EXPORT_SYMBOL(pm_idle); /* * The idle thread, has rather strange semantics for calling pm_idle, * but this is what x86 does and we need to do the same, so that * things like cpuidle get called in the same way. The only difference * is that we always respect 'hlt_counter' to prevent low power idle. * The idle thread. * We always respect 'hlt_counter' to prevent low power idle. */ void cpu_idle(void) { Loading Loading @@ -210,10 +205,10 @@ void cpu_idle(void) } else if (!need_resched()) { stop_critical_timings(); if (cpuidle_idle_call()) pm_idle(); default_idle(); start_critical_timings(); /* * pm_idle functions must always * default_idle functions must always * return with IRQs enabled. */ WARN_ON(irqs_disabled()); Loading
arch/arm64/kernel/process.c +4 −9 Original line number Diff line number Diff line Loading @@ -97,14 +97,9 @@ static void default_idle(void) local_irq_enable(); } void (*pm_idle)(void) = default_idle; EXPORT_SYMBOL_GPL(pm_idle); /* * The idle thread, has rather strange semantics for calling pm_idle, * but this is what x86 does and we need to do the same, so that * things like cpuidle get called in the same way. The only difference * is that we always respect 'hlt_counter' to prevent low power idle. * The idle thread. * We always respect 'hlt_counter' to prevent low power idle. */ void cpu_idle(void) { Loading @@ -122,10 +117,10 @@ void cpu_idle(void) local_irq_disable(); if (!need_resched()) { stop_critical_timings(); pm_idle(); default_idle(); start_critical_timings(); /* * pm_idle functions should always return * default_idle functions should always return * with IRQs enabled. */ WARN_ON(irqs_disabled()); Loading
arch/blackfin/kernel/process.c +0 −7 Original line number Diff line number Diff line Loading @@ -39,12 +39,6 @@ int nr_l1stack_tasks; void *l1_stack_base; unsigned long l1_stack_len; /* * Powermanagement idle function, if any.. */ void (*pm_idle)(void) = NULL; EXPORT_SYMBOL(pm_idle); void (*pm_power_off)(void) = NULL; EXPORT_SYMBOL(pm_power_off); Loading Loading @@ -81,7 +75,6 @@ void cpu_idle(void) { /* endless idle loop with no priority at all */ while (1) { void (*idle)(void) = pm_idle; #ifdef CONFIG_HOTPLUG_CPU if (cpu_is_offline(smp_processor_id())) Loading
arch/cris/kernel/process.c +1 −10 Original line number Diff line number Diff line Loading @@ -54,11 +54,6 @@ void enable_hlt(void) EXPORT_SYMBOL(enable_hlt); /* * The following aren't currently used. */ void (*pm_idle)(void); extern void default_idle(void); void (*pm_power_off)(void); Loading @@ -77,16 +72,12 @@ void cpu_idle (void) while (1) { rcu_idle_enter(); while (!need_resched()) { void (*idle)(void); /* * Mark this as an RCU critical section so that * synchronize_kernel() in the unload path waits * for our completion. */ idle = pm_idle; if (!idle) idle = default_idle; idle(); default_idle(); } rcu_idle_exit(); schedule_preempt_disabled(); Loading
arch/ia64/kernel/process.c +0 −3 Original line number Diff line number Diff line Loading @@ -57,8 +57,6 @@ void (*ia64_mark_idle)(int); unsigned long boot_option_idle_override = IDLE_NO_OVERRIDE; EXPORT_SYMBOL(boot_option_idle_override); void (*pm_idle) (void); EXPORT_SYMBOL(pm_idle); void (*pm_power_off) (void); EXPORT_SYMBOL(pm_power_off); Loading Loading @@ -301,7 +299,6 @@ cpu_idle (void) if (mark_idle) (*mark_idle)(1); idle = pm_idle; if (!idle) idle = default_idle; (*idle)(); Loading