提交 3da1c84c 编写于 作者: O Oleg Nesterov 提交者: Linus Torvalds

workqueues: make get_online_cpus() useable for work->func()

workqueue_cpu_callback(CPU_DEAD) flushes cwq->thread under
cpu_maps_update_begin().  This means that the multithreaded workqueues
can't use get_online_cpus() due to the possible deadlock, very bad and
very old problem.

Introduce the new state, CPU_POST_DEAD, which is called after
cpu_hotplug_done() but before cpu_maps_update_done().

Change workqueue_cpu_callback() to use CPU_POST_DEAD instead of CPU_DEAD.
This means that create/destroy functions can't rely on get_online_cpus()
any longer and should take cpu_add_remove_lock instead.

[akpm@linux-foundation.org: fix CONFIG_SMP=n]
Signed-off-by: NOleg Nesterov <oleg@tv-sign.ru>
Acked-by: NGautham R Shenoy <ego@in.ibm.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Max Krasnyansky <maxk@qualcomm.com>
Cc: Paul Jackson <pj@sgi.com>
Cc: Paul Menage <menage@google.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Ingo Molnar <mingo@elte.hu>
Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
上级 8616a89a
...@@ -69,10 +69,11 @@ static inline void unregister_cpu_notifier(struct notifier_block *nb) ...@@ -69,10 +69,11 @@ static inline void unregister_cpu_notifier(struct notifier_block *nb)
#endif #endif
int cpu_up(unsigned int cpu); int cpu_up(unsigned int cpu);
extern void cpu_hotplug_init(void); extern void cpu_hotplug_init(void);
extern void cpu_maps_update_begin(void);
extern void cpu_maps_update_done(void);
#else #else /* CONFIG_SMP */
static inline int register_cpu_notifier(struct notifier_block *nb) static inline int register_cpu_notifier(struct notifier_block *nb)
{ {
...@@ -87,10 +88,16 @@ static inline void cpu_hotplug_init(void) ...@@ -87,10 +88,16 @@ static inline void cpu_hotplug_init(void)
{ {
} }
static inline void cpu_maps_update_begin(void)
{
}
static inline void cpu_maps_update_done(void)
{
}
#endif /* CONFIG_SMP */ #endif /* CONFIG_SMP */
extern struct sysdev_class cpu_sysdev_class; extern struct sysdev_class cpu_sysdev_class;
extern void cpu_maps_update_begin(void);
extern void cpu_maps_update_done(void);
#ifdef CONFIG_HOTPLUG_CPU #ifdef CONFIG_HOTPLUG_CPU
/* Stop CPUs going up and down. */ /* Stop CPUs going up and down. */
......
...@@ -214,6 +214,8 @@ static inline int notifier_to_errno(int ret) ...@@ -214,6 +214,8 @@ static inline int notifier_to_errno(int ret)
#define CPU_DEAD 0x0007 /* CPU (unsigned)v dead */ #define CPU_DEAD 0x0007 /* CPU (unsigned)v dead */
#define CPU_DYING 0x0008 /* CPU (unsigned)v not running any task, #define CPU_DYING 0x0008 /* CPU (unsigned)v not running any task,
* not handling interrupts, soon dead */ * not handling interrupts, soon dead */
#define CPU_POST_DEAD 0x0009 /* CPU (unsigned)v dead, cpu_hotplug
* lock is dropped */
/* Used for CPU hotplug events occuring while tasks are frozen due to a suspend /* Used for CPU hotplug events occuring while tasks are frozen due to a suspend
* operation in progress * operation in progress
......
...@@ -285,6 +285,11 @@ static int __ref _cpu_down(unsigned int cpu, int tasks_frozen) ...@@ -285,6 +285,11 @@ static int __ref _cpu_down(unsigned int cpu, int tasks_frozen)
set_cpus_allowed_ptr(current, &old_allowed); set_cpus_allowed_ptr(current, &old_allowed);
out_release: out_release:
cpu_hotplug_done(); cpu_hotplug_done();
if (!err) {
if (raw_notifier_call_chain(&cpu_chain, CPU_POST_DEAD | mod,
hcpu) == NOTIFY_BAD)
BUG();
}
return err; return err;
} }
......
...@@ -828,7 +828,7 @@ struct workqueue_struct *__create_workqueue_key(const char *name, ...@@ -828,7 +828,7 @@ struct workqueue_struct *__create_workqueue_key(const char *name,
err = create_workqueue_thread(cwq, singlethread_cpu); err = create_workqueue_thread(cwq, singlethread_cpu);
start_workqueue_thread(cwq, -1); start_workqueue_thread(cwq, -1);
} else { } else {
get_online_cpus(); cpu_maps_update_begin();
spin_lock(&workqueue_lock); spin_lock(&workqueue_lock);
list_add(&wq->list, &workqueues); list_add(&wq->list, &workqueues);
spin_unlock(&workqueue_lock); spin_unlock(&workqueue_lock);
...@@ -840,7 +840,7 @@ struct workqueue_struct *__create_workqueue_key(const char *name, ...@@ -840,7 +840,7 @@ struct workqueue_struct *__create_workqueue_key(const char *name,
err = create_workqueue_thread(cwq, cpu); err = create_workqueue_thread(cwq, cpu);
start_workqueue_thread(cwq, cpu); start_workqueue_thread(cwq, cpu);
} }
put_online_cpus(); cpu_maps_update_done();
} }
if (err) { if (err) {
...@@ -854,8 +854,8 @@ EXPORT_SYMBOL_GPL(__create_workqueue_key); ...@@ -854,8 +854,8 @@ EXPORT_SYMBOL_GPL(__create_workqueue_key);
static void cleanup_workqueue_thread(struct cpu_workqueue_struct *cwq) static void cleanup_workqueue_thread(struct cpu_workqueue_struct *cwq)
{ {
/* /*
* Our caller is either destroy_workqueue() or CPU_DEAD, * Our caller is either destroy_workqueue() or CPU_POST_DEAD,
* get_online_cpus() protects cwq->thread. * cpu_add_remove_lock protects cwq->thread.
*/ */
if (cwq->thread == NULL) if (cwq->thread == NULL)
return; return;
...@@ -865,7 +865,7 @@ static void cleanup_workqueue_thread(struct cpu_workqueue_struct *cwq) ...@@ -865,7 +865,7 @@ static void cleanup_workqueue_thread(struct cpu_workqueue_struct *cwq)
flush_cpu_workqueue(cwq); flush_cpu_workqueue(cwq);
/* /*
* If the caller is CPU_DEAD and cwq->worklist was not empty, * If the caller is CPU_POST_DEAD and cwq->worklist was not empty,
* a concurrent flush_workqueue() can insert a barrier after us. * a concurrent flush_workqueue() can insert a barrier after us.
* However, in that case run_workqueue() won't return and check * However, in that case run_workqueue() won't return and check
* kthread_should_stop() until it flushes all work_struct's. * kthread_should_stop() until it flushes all work_struct's.
...@@ -889,14 +889,14 @@ void destroy_workqueue(struct workqueue_struct *wq) ...@@ -889,14 +889,14 @@ void destroy_workqueue(struct workqueue_struct *wq)
const cpumask_t *cpu_map = wq_cpu_map(wq); const cpumask_t *cpu_map = wq_cpu_map(wq);
int cpu; int cpu;
get_online_cpus(); cpu_maps_update_begin();
spin_lock(&workqueue_lock); spin_lock(&workqueue_lock);
list_del(&wq->list); list_del(&wq->list);
spin_unlock(&workqueue_lock); spin_unlock(&workqueue_lock);
for_each_cpu_mask_nr(cpu, *cpu_map) for_each_cpu_mask_nr(cpu, *cpu_map)
cleanup_workqueue_thread(per_cpu_ptr(wq->cpu_wq, cpu)); cleanup_workqueue_thread(per_cpu_ptr(wq->cpu_wq, cpu));
put_online_cpus(); cpu_maps_update_done();
free_percpu(wq->cpu_wq); free_percpu(wq->cpu_wq);
kfree(wq); kfree(wq);
...@@ -935,7 +935,7 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb, ...@@ -935,7 +935,7 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb,
case CPU_UP_CANCELED: case CPU_UP_CANCELED:
start_workqueue_thread(cwq, -1); start_workqueue_thread(cwq, -1);
case CPU_DEAD: case CPU_POST_DEAD:
cleanup_workqueue_thread(cwq); cleanup_workqueue_thread(cwq);
break; break;
} }
...@@ -943,7 +943,7 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb, ...@@ -943,7 +943,7 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb,
switch (action) { switch (action) {
case CPU_UP_CANCELED: case CPU_UP_CANCELED:
case CPU_DEAD: case CPU_POST_DEAD:
cpu_clear(cpu, cpu_populated_map); cpu_clear(cpu, cpu_populated_map);
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册