提交 da531125 编写于 作者: A Andi Kleen 提交者: Linus Torvalds

[PATCH] x86_64: Fix race in cpu_local_* on preemptible kernels

When a process changes CPUs while doing the non atomic cpu_local_*
operations it might operate on the local_t of a different CPUs.

Fix that by disabling preemption.

Pointed out by Christopher Lameter
Signed-off-by: NAndi Kleen <ak@suse.de>
Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
上级 75bd665c
...@@ -55,12 +55,26 @@ static __inline__ void local_sub(long i, local_t *v) ...@@ -55,12 +55,26 @@ static __inline__ void local_sub(long i, local_t *v)
* much more efficient than these naive implementations. Note they take * much more efficient than these naive implementations. Note they take
* a variable, not an address. * a variable, not an address.
*/ */
#define cpu_local_read(v) local_read(&__get_cpu_var(v))
#define cpu_local_set(v, i) local_set(&__get_cpu_var(v), (i)) /* Need to disable preemption for the cpu local counters otherwise we could
#define cpu_local_inc(v) local_inc(&__get_cpu_var(v)) still access a variable of a previous CPU in a non atomic way. */
#define cpu_local_dec(v) local_dec(&__get_cpu_var(v)) #define cpu_local_wrap_v(v) \
#define cpu_local_add(i, v) local_add((i), &__get_cpu_var(v)) ({ local_t res__; \
#define cpu_local_sub(i, v) local_sub((i), &__get_cpu_var(v)) preempt_disable(); \
res__ = (v); \
preempt_enable(); \
res__; })
#define cpu_local_wrap(v) \
({ preempt_disable(); \
v; \
preempt_enable(); }) \
#define cpu_local_read(v) cpu_local_wrap_v(local_read(&__get_cpu_var(v)))
#define cpu_local_set(v, i) cpu_local_wrap(local_set(&__get_cpu_var(v), (i)))
#define cpu_local_inc(v) cpu_local_wrap(local_inc(&__get_cpu_var(v)))
#define cpu_local_dec(v) cpu_local_wrap(local_dec(&__get_cpu_var(v)))
#define cpu_local_add(i, v) cpu_local_wrap(local_add((i), &__get_cpu_var(v)))
#define cpu_local_sub(i, v) cpu_local_wrap(local_sub((i), &__get_cpu_var(v)))
#define __cpu_local_inc(v) cpu_local_inc(v) #define __cpu_local_inc(v) cpu_local_inc(v)
#define __cpu_local_dec(v) cpu_local_dec(v) #define __cpu_local_dec(v) cpu_local_dec(v)
......
...@@ -59,12 +59,26 @@ static inline void local_sub(long i, local_t *v) ...@@ -59,12 +59,26 @@ static inline void local_sub(long i, local_t *v)
* This could be done better if we moved the per cpu data directly * This could be done better if we moved the per cpu data directly
* after GS. * after GS.
*/ */
#define cpu_local_read(v) local_read(&__get_cpu_var(v))
#define cpu_local_set(v, i) local_set(&__get_cpu_var(v), (i)) /* Need to disable preemption for the cpu local counters otherwise we could
#define cpu_local_inc(v) local_inc(&__get_cpu_var(v)) still access a variable of a previous CPU in a non atomic way. */
#define cpu_local_dec(v) local_dec(&__get_cpu_var(v)) #define cpu_local_wrap_v(v) \
#define cpu_local_add(i, v) local_add((i), &__get_cpu_var(v)) ({ local_t res__; \
#define cpu_local_sub(i, v) local_sub((i), &__get_cpu_var(v)) preempt_disable(); \
res__ = (v); \
preempt_enable(); \
res__; })
#define cpu_local_wrap(v) \
({ preempt_disable(); \
v; \
preempt_enable(); }) \
#define cpu_local_read(v) cpu_local_wrap_v(local_read(&__get_cpu_var(v)))
#define cpu_local_set(v, i) cpu_local_wrap(local_set(&__get_cpu_var(v), (i)))
#define cpu_local_inc(v) cpu_local_wrap(local_inc(&__get_cpu_var(v)))
#define cpu_local_dec(v) cpu_local_wrap(local_dec(&__get_cpu_var(v)))
#define cpu_local_add(i, v) cpu_local_wrap(local_add((i), &__get_cpu_var(v)))
#define cpu_local_sub(i, v) cpu_local_wrap(local_sub((i), &__get_cpu_var(v)))
#define __cpu_local_inc(v) cpu_local_inc(v) #define __cpu_local_inc(v) cpu_local_inc(v)
#define __cpu_local_dec(v) cpu_local_dec(v) #define __cpu_local_dec(v) cpu_local_dec(v)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册