提交 41945f6c 编写于 作者: P Peter Zijlstra 提交者: Ingo Molnar

perf: Avoid RCU vs preemption assumptions

The per-pmu per-cpu context patch converted things from
get_cpu_var() to this_cpu_ptr(), but that only works if
rcu_read_lock() actually disables preemption, and since
there is no such guarantee, we need to fix that.

Use the newly introduced {get,put}_cpu_ptr().
Signed-off-by: NPeter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Tejun Heo <tj@kernel.org>
LKML-Reference: <20100917093009.308453028@chello.nl>
Signed-off-by: NIngo Molnar <mingo@elte.hu>
上级 8b8e2ec1
...@@ -3836,18 +3836,20 @@ static void perf_event_task_event(struct perf_task_event *task_event) ...@@ -3836,18 +3836,20 @@ static void perf_event_task_event(struct perf_task_event *task_event)
rcu_read_lock(); rcu_read_lock();
list_for_each_entry_rcu(pmu, &pmus, entry) { list_for_each_entry_rcu(pmu, &pmus, entry) {
cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); cpuctx = get_cpu_ptr(pmu->pmu_cpu_context);
perf_event_task_ctx(&cpuctx->ctx, task_event); perf_event_task_ctx(&cpuctx->ctx, task_event);
ctx = task_event->task_ctx; ctx = task_event->task_ctx;
if (!ctx) { if (!ctx) {
ctxn = pmu->task_ctx_nr; ctxn = pmu->task_ctx_nr;
if (ctxn < 0) if (ctxn < 0)
continue; goto next;
ctx = rcu_dereference(current->perf_event_ctxp[ctxn]); ctx = rcu_dereference(current->perf_event_ctxp[ctxn]);
} }
if (ctx) if (ctx)
perf_event_task_ctx(ctx, task_event); perf_event_task_ctx(ctx, task_event);
next:
put_cpu_ptr(pmu->pmu_cpu_context);
} }
rcu_read_unlock(); rcu_read_unlock();
} }
...@@ -3969,16 +3971,18 @@ static void perf_event_comm_event(struct perf_comm_event *comm_event) ...@@ -3969,16 +3971,18 @@ static void perf_event_comm_event(struct perf_comm_event *comm_event)
rcu_read_lock(); rcu_read_lock();
list_for_each_entry_rcu(pmu, &pmus, entry) { list_for_each_entry_rcu(pmu, &pmus, entry) {
cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); cpuctx = get_cpu_ptr(pmu->pmu_cpu_context);
perf_event_comm_ctx(&cpuctx->ctx, comm_event); perf_event_comm_ctx(&cpuctx->ctx, comm_event);
ctxn = pmu->task_ctx_nr; ctxn = pmu->task_ctx_nr;
if (ctxn < 0) if (ctxn < 0)
continue; goto next;
ctx = rcu_dereference(current->perf_event_ctxp[ctxn]); ctx = rcu_dereference(current->perf_event_ctxp[ctxn]);
if (ctx) if (ctx)
perf_event_comm_ctx(ctx, comm_event); perf_event_comm_ctx(ctx, comm_event);
next:
put_cpu_ptr(pmu->pmu_cpu_context);
} }
rcu_read_unlock(); rcu_read_unlock();
} }
...@@ -4152,19 +4156,21 @@ static void perf_event_mmap_event(struct perf_mmap_event *mmap_event) ...@@ -4152,19 +4156,21 @@ static void perf_event_mmap_event(struct perf_mmap_event *mmap_event)
rcu_read_lock(); rcu_read_lock();
list_for_each_entry_rcu(pmu, &pmus, entry) { list_for_each_entry_rcu(pmu, &pmus, entry) {
cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); cpuctx = get_cpu_ptr(pmu->pmu_cpu_context);
perf_event_mmap_ctx(&cpuctx->ctx, mmap_event, perf_event_mmap_ctx(&cpuctx->ctx, mmap_event,
vma->vm_flags & VM_EXEC); vma->vm_flags & VM_EXEC);
ctxn = pmu->task_ctx_nr; ctxn = pmu->task_ctx_nr;
if (ctxn < 0) if (ctxn < 0)
continue; goto next;
ctx = rcu_dereference(current->perf_event_ctxp[ctxn]); ctx = rcu_dereference(current->perf_event_ctxp[ctxn]);
if (ctx) { if (ctx) {
perf_event_mmap_ctx(ctx, mmap_event, perf_event_mmap_ctx(ctx, mmap_event,
vma->vm_flags & VM_EXEC); vma->vm_flags & VM_EXEC);
} }
next:
put_cpu_ptr(pmu->pmu_cpu_context);
} }
rcu_read_unlock(); rcu_read_unlock();
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册