Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 76fac077 authored by Alok Kataria's avatar Alok Kataria Committed by H. Peter Anvin
Browse files

x86, kexec: Make sure to stop all CPUs before exiting the kernel



x86 smp_ops now has a new op, stop_other_cpus which takes a parameter
"wait" this allows the caller to specify if it wants to stop until all
the cpus have processed the stop IPI.  This is required specifically
for the kexec case where we should wait for all the cpus to be stopped
before starting the new kernel.  We now wait for the cpus to stop in
all cases except for panic/kdump where we expect things to be broken
and we are doing our best to make things work anyway.

This patch fixes a legitimate regression, which was introduced during
2.6.30, by commit id 4ef702c1.

Signed-off-by: default avatarAlok N Kataria <akataria@vmware.com>
LKML-Reference: <1286833028.1372.20.camel@ank32.eng.vmware.com>
Cc: Eric W. Biederman <ebiederm@xmission.com>
Cc: Jeremy Fitzhardinge <jeremy@xensource.com>
Cc: <stable@kernel.org> v2.6.30-36
Signed-off-by: default avatarH. Peter Anvin <hpa@linux.intel.com>
parent 03f1a17c
Loading
Loading
Loading
Loading
+7 −2
Original line number Diff line number Diff line
@@ -50,7 +50,7 @@ struct smp_ops {
	void (*smp_prepare_cpus)(unsigned max_cpus);
	void (*smp_cpus_done)(unsigned max_cpus);

	void (*smp_send_stop)(void);
	void (*stop_other_cpus)(int wait);
	void (*smp_send_reschedule)(int cpu);

	int (*cpu_up)(unsigned cpu);
@@ -73,7 +73,12 @@ extern struct smp_ops smp_ops;

static inline void smp_send_stop(void)
{
	smp_ops.smp_send_stop();
	smp_ops.stop_other_cpus(0);
}

static inline void stop_other_cpus(void)
{
	smp_ops.stop_other_cpus(1);
}

static inline void smp_prepare_boot_cpu(void)
+1 −1
Original line number Diff line number Diff line
@@ -641,7 +641,7 @@ void native_machine_shutdown(void)
	/* O.K Now that I'm on the appropriate processor,
	 * stop all of the others.
	 */
	smp_send_stop();
	stop_other_cpus();
#endif

	lapic_shutdown();
+9 −6
Original line number Diff line number Diff line
@@ -159,10 +159,10 @@ asmlinkage void smp_reboot_interrupt(void)
	irq_exit();
}

static void native_smp_send_stop(void)
static void native_stop_other_cpus(int wait)
{
	unsigned long flags;
	unsigned long wait;
	unsigned long timeout;

	if (reboot_force)
		return;
@@ -179,9 +179,12 @@ static void native_smp_send_stop(void)
	if (num_online_cpus() > 1) {
		apic->send_IPI_allbutself(REBOOT_VECTOR);

		/* Don't wait longer than a second */
		wait = USEC_PER_SEC;
		while (num_online_cpus() > 1 && wait--)
		/*
		 * Don't wait longer than a second if the caller
		 * didn't ask us to wait.
		 */
		timeout = USEC_PER_SEC;
		while (num_online_cpus() > 1 && (wait || timeout--))
			udelay(1);
	}

@@ -227,7 +230,7 @@ struct smp_ops smp_ops = {
	.smp_prepare_cpus	= native_smp_prepare_cpus,
	.smp_cpus_done		= native_smp_cpus_done,

	.smp_send_stop		= native_smp_send_stop,
	.stop_other_cpus	= native_stop_other_cpus,
	.smp_send_reschedule	= native_smp_send_reschedule,

	.cpu_up			= native_cpu_up,
+1 −1
Original line number Diff line number Diff line
@@ -1018,7 +1018,7 @@ static void xen_reboot(int reason)
	struct sched_shutdown r = { .reason = reason };

#ifdef CONFIG_SMP
	smp_send_stop();
	stop_other_cpus();
#endif

	if (HYPERVISOR_sched_op(SCHEDOP_shutdown, &r))
+3 −3
Original line number Diff line number Diff line
@@ -400,9 +400,9 @@ static void stop_self(void *v)
	BUG();
}

static void xen_smp_send_stop(void)
static void xen_stop_other_cpus(int wait)
{
	smp_call_function(stop_self, NULL, 0);
	smp_call_function(stop_self, NULL, wait);
}

static void xen_smp_send_reschedule(int cpu)
@@ -470,7 +470,7 @@ static const struct smp_ops xen_smp_ops __initdata = {
	.cpu_disable = xen_cpu_disable,
	.play_dead = xen_play_dead,

	.smp_send_stop = xen_smp_send_stop,
	.stop_other_cpus = xen_stop_other_cpus,
	.smp_send_reschedule = xen_smp_send_reschedule,

	.send_call_func_ipi = xen_smp_send_call_function_ipi,