提交 723aae25 编写于 作者: M Milton Miller 提交者: Linus Torvalds

smp_call_function_many: handle concurrent clearing of mask

Mike Galbraith reported finding a lockup ("perma-spin bug") where the
cpumask passed to smp_call_function_many was cleared by other cpu(s)
while a cpu was preparing its call_data block, resulting in no cpu to
clear the last ref and unlock the block.

Having cpus clear their bit asynchronously could be useful on a mask of
cpus that might have a translation context, or cpus that need a push to
complete an rcu window.

Instead of adding a BUG_ON and requiring yet another cpumask copy, just
detect the race and handle it.

Note: arch_send_call_function_ipi_mask must still handle an empty
cpumask because the data block is globally visible before the that arch
callback is made.  And (obviously) there are no guarantees to which cpus
are notified if the mask is changed during the call; only cpus that were
online and had their mask bit set during the whole call are guaranteed
to be called.
Reported-by: NMike Galbraith <efault@gmx.de>
Reported-by: NJan Beulich <JBeulich@novell.com>
Acked-by: NJan Beulich <jbeulich@novell.com>
Cc: stable@kernel.org
Signed-off-by: NMilton Miller <miltonm@bga.com>
Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
上级 45a57919
...@@ -450,7 +450,7 @@ void smp_call_function_many(const struct cpumask *mask, ...@@ -450,7 +450,7 @@ void smp_call_function_many(const struct cpumask *mask,
{ {
struct call_function_data *data; struct call_function_data *data;
unsigned long flags; unsigned long flags;
int cpu, next_cpu, this_cpu = smp_processor_id(); int refs, cpu, next_cpu, this_cpu = smp_processor_id();
/* /*
* Can deadlock when called with interrupts disabled. * Can deadlock when called with interrupts disabled.
...@@ -461,7 +461,7 @@ void smp_call_function_many(const struct cpumask *mask, ...@@ -461,7 +461,7 @@ void smp_call_function_many(const struct cpumask *mask,
WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled() WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled()
&& !oops_in_progress && !early_boot_irqs_disabled); && !oops_in_progress && !early_boot_irqs_disabled);
/* So, what's a CPU they want? Ignoring this one. */ /* Try to fastpath. So, what's a CPU they want? Ignoring this one. */
cpu = cpumask_first_and(mask, cpu_online_mask); cpu = cpumask_first_and(mask, cpu_online_mask);
if (cpu == this_cpu) if (cpu == this_cpu)
cpu = cpumask_next_and(cpu, mask, cpu_online_mask); cpu = cpumask_next_and(cpu, mask, cpu_online_mask);
...@@ -519,6 +519,13 @@ void smp_call_function_many(const struct cpumask *mask, ...@@ -519,6 +519,13 @@ void smp_call_function_many(const struct cpumask *mask,
/* We rely on the "and" being processed before the store */ /* We rely on the "and" being processed before the store */
cpumask_and(data->cpumask, mask, cpu_online_mask); cpumask_and(data->cpumask, mask, cpu_online_mask);
cpumask_clear_cpu(this_cpu, data->cpumask); cpumask_clear_cpu(this_cpu, data->cpumask);
refs = cpumask_weight(data->cpumask);
/* Some callers race with other cpus changing the passed mask */
if (unlikely(!refs)) {
csd_unlock(&data->csd);
return;
}
raw_spin_lock_irqsave(&call_function.lock, flags); raw_spin_lock_irqsave(&call_function.lock, flags);
/* /*
...@@ -532,7 +539,7 @@ void smp_call_function_many(const struct cpumask *mask, ...@@ -532,7 +539,7 @@ void smp_call_function_many(const struct cpumask *mask,
* to the cpumask before this write to refs, which indicates * to the cpumask before this write to refs, which indicates
* data is on the list and is ready to be processed. * data is on the list and is ready to be processed.
*/ */
atomic_set(&data->refs, cpumask_weight(data->cpumask)); atomic_set(&data->refs, refs);
raw_spin_unlock_irqrestore(&call_function.lock, flags); raw_spin_unlock_irqrestore(&call_function.lock, flags);
/* /*
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册