x86, mce: cleanup mce_start()

Simplify interface of mce_start():

-       no_way_out = mce_start(no_way_out, &order);
+       order = mce_start(&no_way_out);

Now Monarch and Subjects share same exit(return) in usual path.

Signed-off-by: Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
Signed-off-by: H. Peter Anvin <hpa@zytor.com>
This commit is contained in:
Hidetoshi Seto
2009-06-15 18:18:43 +09:00
committed by H. Peter Anvin
parent 33edbf02a9
commit 7fb06fc967

View File

@@ -691,23 +691,21 @@ static atomic_t global_nwo;
* in the entry order. * in the entry order.
* TBD double check parallel CPU hotunplug * TBD double check parallel CPU hotunplug
*/ */
static int mce_start(int no_way_out, int *order) static int mce_start(int *no_way_out)
{ {
int nwo; int order;
int cpus = num_online_cpus(); int cpus = num_online_cpus();
u64 timeout = (u64)monarch_timeout * NSEC_PER_USEC; u64 timeout = (u64)monarch_timeout * NSEC_PER_USEC;
if (!timeout) { if (!timeout)
*order = -1; return -1;
return no_way_out;
}
atomic_add(no_way_out, &global_nwo); atomic_add(*no_way_out, &global_nwo);
/* /*
* global_nwo should be updated before mce_callin * global_nwo should be updated before mce_callin
*/ */
smp_wmb(); smp_wmb();
*order = atomic_add_return(1, &mce_callin); order = atomic_add_return(1, &mce_callin);
/* /*
* Wait for everyone. * Wait for everyone.
@@ -715,8 +713,7 @@ static int mce_start(int no_way_out, int *order)
while (atomic_read(&mce_callin) != cpus) { while (atomic_read(&mce_callin) != cpus) {
if (mce_timed_out(&timeout)) { if (mce_timed_out(&timeout)) {
atomic_set(&global_nwo, 0); atomic_set(&global_nwo, 0);
*order = -1; return -1;
return no_way_out;
} }
ndelay(SPINUNIT); ndelay(SPINUNIT);
} }
@@ -725,34 +722,34 @@ static int mce_start(int no_way_out, int *order)
* mce_callin should be read before global_nwo * mce_callin should be read before global_nwo
*/ */
smp_rmb(); smp_rmb();
if (order == 1) {
/*
* Monarch: Starts executing now, the others wait.
*/
atomic_set(&mce_executing, 1);
} else {
/*
* Subject: Now start the scanning loop one by one in
* the original callin order.
* This way when there are any shared banks it will be
* only seen by one CPU before cleared, avoiding duplicates.
*/
while (atomic_read(&mce_executing) < order) {
if (mce_timed_out(&timeout)) {
atomic_set(&global_nwo, 0);
return -1;
}
ndelay(SPINUNIT);
}
}
/* /*
* Cache the global no_way_out state. * Cache the global no_way_out state.
*/ */
nwo = atomic_read(&global_nwo); *no_way_out = atomic_read(&global_nwo);
/* return order;
* Monarch starts executing now, the others wait.
*/
if (*order == 1) {
atomic_set(&mce_executing, 1);
return nwo;
}
/*
* Now start the scanning loop one by one
* in the original callin order.
* This way when there are any shared banks it will
* be only seen by one CPU before cleared, avoiding duplicates.
*/
while (atomic_read(&mce_executing) < *order) {
if (mce_timed_out(&timeout)) {
atomic_set(&global_nwo, 0);
*order = -1;
return no_way_out;
}
ndelay(SPINUNIT);
}
return nwo;
} }
/* /*
@@ -871,8 +868,7 @@ void do_machine_check(struct pt_regs *regs, long error_code)
* Establish sequential order between the CPUs entering the machine * Establish sequential order between the CPUs entering the machine
* check handler. * check handler.
*/ */
int order = -1; int order;
/* /*
* If no_way_out gets set, there is no safe way to recover from this * If no_way_out gets set, there is no safe way to recover from this
* MCE. If tolerant is cranked up, we'll try anyway. * MCE. If tolerant is cranked up, we'll try anyway.
@@ -917,7 +913,7 @@ void do_machine_check(struct pt_regs *regs, long error_code)
* This way we don't report duplicated events on shared banks * This way we don't report duplicated events on shared banks
* because the first one to see it will clear it. * because the first one to see it will clear it.
*/ */
no_way_out = mce_start(no_way_out, &order); order = mce_start(&no_way_out);
for (i = 0; i < banks; i++) { for (i = 0; i < banks; i++) {
__clear_bit(i, toclear); __clear_bit(i, toclear);
if (!bank[i]) if (!bank[i])