提交 5622f295 编写于 作者: M Markus Metzger 提交者: Ingo Molnar

x86, perf_counter, bts: Optimize BTS overflow handling

Draining the BTS buffer on a buffer overflow interrupt takes too
long resulting in a kernel lockup when tracing the kernel.

Restructure perf_counter sampling into sample creation and sample
output.

Prepare a single reference sample for BTS sampling and update the
from and to address fields when draining the BTS buffer. Drain the
entire BTS buffer between a single perf_output_begin() /
perf_output_end() pair.
Signed-off-by: NMarkus Metzger <markus.t.metzger@intel.com>
Signed-off-by: NPeter Zijlstra <a.p.zijlstra@chello.nl>
LKML-Reference: <20090915130023.A16204@sedona.ch.intel.com>
Signed-off-by: NIngo Molnar <mingo@elte.hu>
上级 4b77a729
...@@ -36,10 +36,10 @@ static u64 perf_counter_mask __read_mostly; ...@@ -36,10 +36,10 @@ static u64 perf_counter_mask __read_mostly;
#define BTS_RECORD_SIZE 24 #define BTS_RECORD_SIZE 24
/* The size of a per-cpu BTS buffer in bytes: */ /* The size of a per-cpu BTS buffer in bytes: */
#define BTS_BUFFER_SIZE (BTS_RECORD_SIZE * 1024) #define BTS_BUFFER_SIZE (BTS_RECORD_SIZE * 2048)
/* The BTS overflow threshold in bytes from the end of the buffer: */ /* The BTS overflow threshold in bytes from the end of the buffer: */
#define BTS_OVFL_TH (BTS_RECORD_SIZE * 64) #define BTS_OVFL_TH (BTS_RECORD_SIZE * 128)
/* /*
...@@ -1488,8 +1488,7 @@ void perf_counter_print_debug(void) ...@@ -1488,8 +1488,7 @@ void perf_counter_print_debug(void)
local_irq_restore(flags); local_irq_restore(flags);
} }
static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc, static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc)
struct perf_sample_data *data)
{ {
struct debug_store *ds = cpuc->ds; struct debug_store *ds = cpuc->ds;
struct bts_record { struct bts_record {
...@@ -1498,8 +1497,11 @@ static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc, ...@@ -1498,8 +1497,11 @@ static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc,
u64 flags; u64 flags;
}; };
struct perf_counter *counter = cpuc->counters[X86_PMC_IDX_FIXED_BTS]; struct perf_counter *counter = cpuc->counters[X86_PMC_IDX_FIXED_BTS];
unsigned long orig_ip = data->regs->ip;
struct bts_record *at, *top; struct bts_record *at, *top;
struct perf_output_handle handle;
struct perf_event_header header;
struct perf_sample_data data;
struct pt_regs regs;
if (!counter) if (!counter)
return; return;
...@@ -1510,19 +1512,38 @@ static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc, ...@@ -1510,19 +1512,38 @@ static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc,
at = (struct bts_record *)(unsigned long)ds->bts_buffer_base; at = (struct bts_record *)(unsigned long)ds->bts_buffer_base;
top = (struct bts_record *)(unsigned long)ds->bts_index; top = (struct bts_record *)(unsigned long)ds->bts_index;
if (top <= at)
return;
ds->bts_index = ds->bts_buffer_base; ds->bts_index = ds->bts_buffer_base;
data.period = counter->hw.last_period;
data.addr = 0;
regs.ip = 0;
/*
* Prepare a generic sample, i.e. fill in the invariant fields.
* We will overwrite the from and to address before we output
* the sample.
*/
perf_prepare_sample(&header, &data, counter, &regs);
if (perf_output_begin(&handle, counter,
header.size * (top - at), 1, 1))
return;
for (; at < top; at++) { for (; at < top; at++) {
data->regs->ip = at->from; data.ip = at->from;
data->addr = at->to; data.addr = at->to;
perf_counter_output(counter, 1, data); perf_output_sample(&handle, &header, &data, counter);
} }
data->regs->ip = orig_ip; perf_output_end(&handle);
data->addr = 0;
/* There's new data available. */ /* There's new data available. */
counter->hw.interrupts++;
counter->pending_kill = POLL_IN; counter->pending_kill = POLL_IN;
} }
...@@ -1552,13 +1573,9 @@ static void x86_pmu_disable(struct perf_counter *counter) ...@@ -1552,13 +1573,9 @@ static void x86_pmu_disable(struct perf_counter *counter)
x86_perf_counter_update(counter, hwc, idx); x86_perf_counter_update(counter, hwc, idx);
/* Drain the remaining BTS records. */ /* Drain the remaining BTS records. */
if (unlikely(idx == X86_PMC_IDX_FIXED_BTS)) { if (unlikely(idx == X86_PMC_IDX_FIXED_BTS))
struct perf_sample_data data; intel_pmu_drain_bts_buffer(cpuc);
struct pt_regs regs;
data.regs = &regs;
intel_pmu_drain_bts_buffer(cpuc, &data);
}
cpuc->counters[idx] = NULL; cpuc->counters[idx] = NULL;
clear_bit(idx, cpuc->used_mask); clear_bit(idx, cpuc->used_mask);
...@@ -1619,7 +1636,6 @@ static int p6_pmu_handle_irq(struct pt_regs *regs) ...@@ -1619,7 +1636,6 @@ static int p6_pmu_handle_irq(struct pt_regs *regs)
int idx, handled = 0; int idx, handled = 0;
u64 val; u64 val;
data.regs = regs;
data.addr = 0; data.addr = 0;
cpuc = &__get_cpu_var(cpu_hw_counters); cpuc = &__get_cpu_var(cpu_hw_counters);
...@@ -1644,7 +1660,7 @@ static int p6_pmu_handle_irq(struct pt_regs *regs) ...@@ -1644,7 +1660,7 @@ static int p6_pmu_handle_irq(struct pt_regs *regs)
if (!x86_perf_counter_set_period(counter, hwc, idx)) if (!x86_perf_counter_set_period(counter, hwc, idx))
continue; continue;
if (perf_counter_overflow(counter, 1, &data)) if (perf_counter_overflow(counter, 1, &data, regs))
p6_pmu_disable_counter(hwc, idx); p6_pmu_disable_counter(hwc, idx);
} }
...@@ -1665,13 +1681,12 @@ static int intel_pmu_handle_irq(struct pt_regs *regs) ...@@ -1665,13 +1681,12 @@ static int intel_pmu_handle_irq(struct pt_regs *regs)
int bit, loops; int bit, loops;
u64 ack, status; u64 ack, status;
data.regs = regs;
data.addr = 0; data.addr = 0;
cpuc = &__get_cpu_var(cpu_hw_counters); cpuc = &__get_cpu_var(cpu_hw_counters);
perf_disable(); perf_disable();
intel_pmu_drain_bts_buffer(cpuc, &data); intel_pmu_drain_bts_buffer(cpuc);
status = intel_pmu_get_status(); status = intel_pmu_get_status();
if (!status) { if (!status) {
perf_enable(); perf_enable();
...@@ -1702,7 +1717,7 @@ static int intel_pmu_handle_irq(struct pt_regs *regs) ...@@ -1702,7 +1717,7 @@ static int intel_pmu_handle_irq(struct pt_regs *regs)
data.period = counter->hw.last_period; data.period = counter->hw.last_period;
if (perf_counter_overflow(counter, 1, &data)) if (perf_counter_overflow(counter, 1, &data, regs))
intel_pmu_disable_counter(&counter->hw, bit); intel_pmu_disable_counter(&counter->hw, bit);
} }
...@@ -1729,7 +1744,6 @@ static int amd_pmu_handle_irq(struct pt_regs *regs) ...@@ -1729,7 +1744,6 @@ static int amd_pmu_handle_irq(struct pt_regs *regs)
int idx, handled = 0; int idx, handled = 0;
u64 val; u64 val;
data.regs = regs;
data.addr = 0; data.addr = 0;
cpuc = &__get_cpu_var(cpu_hw_counters); cpuc = &__get_cpu_var(cpu_hw_counters);
...@@ -1754,7 +1768,7 @@ static int amd_pmu_handle_irq(struct pt_regs *regs) ...@@ -1754,7 +1768,7 @@ static int amd_pmu_handle_irq(struct pt_regs *regs)
if (!x86_perf_counter_set_period(counter, hwc, idx)) if (!x86_perf_counter_set_period(counter, hwc, idx))
continue; continue;
if (perf_counter_overflow(counter, 1, &data)) if (perf_counter_overflow(counter, 1, &data, regs))
amd_pmu_disable_counter(hwc, idx); amd_pmu_disable_counter(hwc, idx);
} }
......
...@@ -691,6 +691,17 @@ struct perf_cpu_context { ...@@ -691,6 +691,17 @@ struct perf_cpu_context {
int recursion[4]; int recursion[4];
}; };
struct perf_output_handle {
struct perf_counter *counter;
struct perf_mmap_data *data;
unsigned long head;
unsigned long offset;
int nmi;
int sample;
int locked;
unsigned long flags;
};
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_COUNTERS
/* /*
...@@ -722,16 +733,38 @@ extern int hw_perf_group_sched_in(struct perf_counter *group_leader, ...@@ -722,16 +733,38 @@ extern int hw_perf_group_sched_in(struct perf_counter *group_leader,
extern void perf_counter_update_userpage(struct perf_counter *counter); extern void perf_counter_update_userpage(struct perf_counter *counter);
struct perf_sample_data { struct perf_sample_data {
struct pt_regs *regs; u64 type;
u64 ip;
struct {
u32 pid;
u32 tid;
} tid_entry;
u64 time;
u64 addr; u64 addr;
u64 id;
u64 stream_id;
struct {
u32 cpu;
u32 reserved;
} cpu_entry;
u64 period; u64 period;
struct perf_callchain_entry *callchain;
struct perf_raw_record *raw; struct perf_raw_record *raw;
}; };
extern void perf_output_sample(struct perf_output_handle *handle,
struct perf_event_header *header,
struct perf_sample_data *data,
struct perf_counter *counter);
extern void perf_prepare_sample(struct perf_event_header *header,
struct perf_sample_data *data,
struct perf_counter *counter,
struct pt_regs *regs);
extern int perf_counter_overflow(struct perf_counter *counter, int nmi, extern int perf_counter_overflow(struct perf_counter *counter, int nmi,
struct perf_sample_data *data); struct perf_sample_data *data,
extern void perf_counter_output(struct perf_counter *counter, int nmi, struct pt_regs *regs);
struct perf_sample_data *data);
/* /*
* Return 1 for a software counter, 0 for a hardware counter * Return 1 for a software counter, 0 for a hardware counter
...@@ -781,6 +814,12 @@ extern void perf_tpcounter_event(int event_id, u64 addr, u64 count, ...@@ -781,6 +814,12 @@ extern void perf_tpcounter_event(int event_id, u64 addr, u64 count,
#define perf_instruction_pointer(regs) instruction_pointer(regs) #define perf_instruction_pointer(regs) instruction_pointer(regs)
#endif #endif
extern int perf_output_begin(struct perf_output_handle *handle,
struct perf_counter *counter, unsigned int size,
int nmi, int sample);
extern void perf_output_end(struct perf_output_handle *handle);
extern void perf_output_copy(struct perf_output_handle *handle,
const void *buf, unsigned int len);
#else #else
static inline void static inline void
perf_counter_task_sched_in(struct task_struct *task, int cpu) { } perf_counter_task_sched_in(struct task_struct *task, int cpu) { }
...@@ -807,7 +846,28 @@ static inline void perf_counter_mmap(struct vm_area_struct *vma) { } ...@@ -807,7 +846,28 @@ static inline void perf_counter_mmap(struct vm_area_struct *vma) { }
static inline void perf_counter_comm(struct task_struct *tsk) { } static inline void perf_counter_comm(struct task_struct *tsk) { }
static inline void perf_counter_fork(struct task_struct *tsk) { } static inline void perf_counter_fork(struct task_struct *tsk) { }
static inline void perf_counter_init(void) { } static inline void perf_counter_init(void) { }
static inline int
perf_output_begin(struct perf_output_handle *handle, struct perf_counter *c,
unsigned int size, int nmi, int sample) { }
static inline void perf_output_end(struct perf_output_handle *handle) { }
static inline void
perf_output_copy(struct perf_output_handle *handle,
const void *buf, unsigned int len) { }
static inline void
perf_output_sample(struct perf_output_handle *handle,
struct perf_event_header *header,
struct perf_sample_data *data,
struct perf_counter *counter) { }
static inline void
perf_prepare_sample(struct perf_event_header *header,
struct perf_sample_data *data,
struct perf_counter *counter,
struct pt_regs *regs) { }
#endif #endif
#define perf_output_put(handle, x) \
perf_output_copy((handle), &(x), sizeof(x))
#endif /* __KERNEL__ */ #endif /* __KERNEL__ */
#endif /* _LINUX_PERF_COUNTER_H */ #endif /* _LINUX_PERF_COUNTER_H */
...@@ -2512,18 +2512,6 @@ __weak struct perf_callchain_entry *perf_callchain(struct pt_regs *regs) ...@@ -2512,18 +2512,6 @@ __weak struct perf_callchain_entry *perf_callchain(struct pt_regs *regs)
/* /*
* Output * Output
*/ */
struct perf_output_handle {
struct perf_counter *counter;
struct perf_mmap_data *data;
unsigned long head;
unsigned long offset;
int nmi;
int sample;
int locked;
unsigned long flags;
};
static bool perf_output_space(struct perf_mmap_data *data, unsigned long tail, static bool perf_output_space(struct perf_mmap_data *data, unsigned long tail,
unsigned long offset, unsigned long head) unsigned long offset, unsigned long head)
{ {
...@@ -2633,8 +2621,8 @@ static void perf_output_unlock(struct perf_output_handle *handle) ...@@ -2633,8 +2621,8 @@ static void perf_output_unlock(struct perf_output_handle *handle)
local_irq_restore(handle->flags); local_irq_restore(handle->flags);
} }
static void perf_output_copy(struct perf_output_handle *handle, void perf_output_copy(struct perf_output_handle *handle,
const void *buf, unsigned int len) const void *buf, unsigned int len)
{ {
unsigned int pages_mask; unsigned int pages_mask;
unsigned int offset; unsigned int offset;
...@@ -2669,12 +2657,9 @@ static void perf_output_copy(struct perf_output_handle *handle, ...@@ -2669,12 +2657,9 @@ static void perf_output_copy(struct perf_output_handle *handle,
WARN_ON_ONCE(((long)(handle->head - handle->offset)) < 0); WARN_ON_ONCE(((long)(handle->head - handle->offset)) < 0);
} }
#define perf_output_put(handle, x) \ int perf_output_begin(struct perf_output_handle *handle,
perf_output_copy((handle), &(x), sizeof(x)) struct perf_counter *counter, unsigned int size,
int nmi, int sample)
static int perf_output_begin(struct perf_output_handle *handle,
struct perf_counter *counter, unsigned int size,
int nmi, int sample)
{ {
struct perf_counter *output_counter; struct perf_counter *output_counter;
struct perf_mmap_data *data; struct perf_mmap_data *data;
...@@ -2756,7 +2741,7 @@ static int perf_output_begin(struct perf_output_handle *handle, ...@@ -2756,7 +2741,7 @@ static int perf_output_begin(struct perf_output_handle *handle,
return -ENOSPC; return -ENOSPC;
} }
static void perf_output_end(struct perf_output_handle *handle) void perf_output_end(struct perf_output_handle *handle)
{ {
struct perf_counter *counter = handle->counter; struct perf_counter *counter = handle->counter;
struct perf_mmap_data *data = handle->data; struct perf_mmap_data *data = handle->data;
...@@ -2870,82 +2855,151 @@ static void perf_output_read(struct perf_output_handle *handle, ...@@ -2870,82 +2855,151 @@ static void perf_output_read(struct perf_output_handle *handle,
perf_output_read_one(handle, counter); perf_output_read_one(handle, counter);
} }
void perf_counter_output(struct perf_counter *counter, int nmi, void perf_output_sample(struct perf_output_handle *handle,
struct perf_sample_data *data) struct perf_event_header *header,
struct perf_sample_data *data,
struct perf_counter *counter)
{
u64 sample_type = data->type;
perf_output_put(handle, *header);
if (sample_type & PERF_SAMPLE_IP)
perf_output_put(handle, data->ip);
if (sample_type & PERF_SAMPLE_TID)
perf_output_put(handle, data->tid_entry);
if (sample_type & PERF_SAMPLE_TIME)
perf_output_put(handle, data->time);
if (sample_type & PERF_SAMPLE_ADDR)
perf_output_put(handle, data->addr);
if (sample_type & PERF_SAMPLE_ID)
perf_output_put(handle, data->id);
if (sample_type & PERF_SAMPLE_STREAM_ID)
perf_output_put(handle, data->stream_id);
if (sample_type & PERF_SAMPLE_CPU)
perf_output_put(handle, data->cpu_entry);
if (sample_type & PERF_SAMPLE_PERIOD)
perf_output_put(handle, data->period);
if (sample_type & PERF_SAMPLE_READ)
perf_output_read(handle, counter);
if (sample_type & PERF_SAMPLE_CALLCHAIN) {
if (data->callchain) {
int size = 1;
if (data->callchain)
size += data->callchain->nr;
size *= sizeof(u64);
perf_output_copy(handle, data->callchain, size);
} else {
u64 nr = 0;
perf_output_put(handle, nr);
}
}
if (sample_type & PERF_SAMPLE_RAW) {
if (data->raw) {
perf_output_put(handle, data->raw->size);
perf_output_copy(handle, data->raw->data,
data->raw->size);
} else {
struct {
u32 size;
u32 data;
} raw = {
.size = sizeof(u32),
.data = 0,
};
perf_output_put(handle, raw);
}
}
}
void perf_prepare_sample(struct perf_event_header *header,
struct perf_sample_data *data,
struct perf_counter *counter,
struct pt_regs *regs)
{ {
int ret;
u64 sample_type = counter->attr.sample_type; u64 sample_type = counter->attr.sample_type;
struct perf_output_handle handle;
struct perf_event_header header;
u64 ip;
struct {
u32 pid, tid;
} tid_entry;
struct perf_callchain_entry *callchain = NULL;
int callchain_size = 0;
u64 time;
struct {
u32 cpu, reserved;
} cpu_entry;
header.type = PERF_EVENT_SAMPLE; data->type = sample_type;
header.size = sizeof(header);
header.misc = 0; header->type = PERF_EVENT_SAMPLE;
header.misc |= perf_misc_flags(data->regs); header->size = sizeof(*header);
header->misc = 0;
header->misc |= perf_misc_flags(regs);
if (sample_type & PERF_SAMPLE_IP) { if (sample_type & PERF_SAMPLE_IP) {
ip = perf_instruction_pointer(data->regs); data->ip = perf_instruction_pointer(regs);
header.size += sizeof(ip);
header->size += sizeof(data->ip);
} }
if (sample_type & PERF_SAMPLE_TID) { if (sample_type & PERF_SAMPLE_TID) {
/* namespace issues */ /* namespace issues */
tid_entry.pid = perf_counter_pid(counter, current); data->tid_entry.pid = perf_counter_pid(counter, current);
tid_entry.tid = perf_counter_tid(counter, current); data->tid_entry.tid = perf_counter_tid(counter, current);
header.size += sizeof(tid_entry); header->size += sizeof(data->tid_entry);
} }
if (sample_type & PERF_SAMPLE_TIME) { if (sample_type & PERF_SAMPLE_TIME) {
/* /*
* Maybe do better on x86 and provide cpu_clock_nmi() * Maybe do better on x86 and provide cpu_clock_nmi()
*/ */
time = sched_clock(); data->time = sched_clock();
header.size += sizeof(u64); header->size += sizeof(data->time);
} }
if (sample_type & PERF_SAMPLE_ADDR) if (sample_type & PERF_SAMPLE_ADDR)
header.size += sizeof(u64); header->size += sizeof(data->addr);
if (sample_type & PERF_SAMPLE_ID) if (sample_type & PERF_SAMPLE_ID) {
header.size += sizeof(u64); data->id = primary_counter_id(counter);
if (sample_type & PERF_SAMPLE_STREAM_ID) header->size += sizeof(data->id);
header.size += sizeof(u64); }
if (sample_type & PERF_SAMPLE_STREAM_ID) {
data->stream_id = counter->id;
header->size += sizeof(data->stream_id);
}
if (sample_type & PERF_SAMPLE_CPU) { if (sample_type & PERF_SAMPLE_CPU) {
header.size += sizeof(cpu_entry); data->cpu_entry.cpu = raw_smp_processor_id();
data->cpu_entry.reserved = 0;
cpu_entry.cpu = raw_smp_processor_id(); header->size += sizeof(data->cpu_entry);
cpu_entry.reserved = 0;
} }
if (sample_type & PERF_SAMPLE_PERIOD) if (sample_type & PERF_SAMPLE_PERIOD)
header.size += sizeof(u64); header->size += sizeof(data->period);
if (sample_type & PERF_SAMPLE_READ) if (sample_type & PERF_SAMPLE_READ)
header.size += perf_counter_read_size(counter); header->size += perf_counter_read_size(counter);
if (sample_type & PERF_SAMPLE_CALLCHAIN) { if (sample_type & PERF_SAMPLE_CALLCHAIN) {
callchain = perf_callchain(data->regs); int size = 1;
if (callchain) { data->callchain = perf_callchain(regs);
callchain_size = (1 + callchain->nr) * sizeof(u64);
header.size += callchain_size; if (data->callchain)
} else size += data->callchain->nr;
header.size += sizeof(u64);
header->size += size * sizeof(u64);
} }
if (sample_type & PERF_SAMPLE_RAW) { if (sample_type & PERF_SAMPLE_RAW) {
...@@ -2957,69 +3011,23 @@ void perf_counter_output(struct perf_counter *counter, int nmi, ...@@ -2957,69 +3011,23 @@ void perf_counter_output(struct perf_counter *counter, int nmi,
size += sizeof(u32); size += sizeof(u32);
WARN_ON_ONCE(size & (sizeof(u64)-1)); WARN_ON_ONCE(size & (sizeof(u64)-1));
header.size += size; header->size += size;
}
ret = perf_output_begin(&handle, counter, header.size, nmi, 1);
if (ret)
return;
perf_output_put(&handle, header);
if (sample_type & PERF_SAMPLE_IP)
perf_output_put(&handle, ip);
if (sample_type & PERF_SAMPLE_TID)
perf_output_put(&handle, tid_entry);
if (sample_type & PERF_SAMPLE_TIME)
perf_output_put(&handle, time);
if (sample_type & PERF_SAMPLE_ADDR)
perf_output_put(&handle, data->addr);
if (sample_type & PERF_SAMPLE_ID) {
u64 id = primary_counter_id(counter);
perf_output_put(&handle, id);
} }
}
if (sample_type & PERF_SAMPLE_STREAM_ID) static void perf_counter_output(struct perf_counter *counter, int nmi,
perf_output_put(&handle, counter->id); struct perf_sample_data *data,
struct pt_regs *regs)
if (sample_type & PERF_SAMPLE_CPU) {
perf_output_put(&handle, cpu_entry); struct perf_output_handle handle;
struct perf_event_header header;
if (sample_type & PERF_SAMPLE_PERIOD)
perf_output_put(&handle, data->period);
if (sample_type & PERF_SAMPLE_READ) perf_prepare_sample(&header, data, counter, regs);
perf_output_read(&handle, counter);
if (sample_type & PERF_SAMPLE_CALLCHAIN) { if (perf_output_begin(&handle, counter, header.size, nmi, 1))
if (callchain) return;
perf_output_copy(&handle, callchain, callchain_size);
else {
u64 nr = 0;
perf_output_put(&handle, nr);
}
}
if (sample_type & PERF_SAMPLE_RAW) { perf_output_sample(&handle, &header, data, counter);
if (data->raw) {
perf_output_put(&handle, data->raw->size);
perf_output_copy(&handle, data->raw->data, data->raw->size);
} else {
struct {
u32 size;
u32 data;
} raw = {
.size = sizeof(u32),
.data = 0,
};
perf_output_put(&handle, raw);
}
}
perf_output_end(&handle); perf_output_end(&handle);
} }
...@@ -3501,7 +3509,8 @@ static void perf_log_throttle(struct perf_counter *counter, int enable) ...@@ -3501,7 +3509,8 @@ static void perf_log_throttle(struct perf_counter *counter, int enable)
*/ */
static int __perf_counter_overflow(struct perf_counter *counter, int nmi, static int __perf_counter_overflow(struct perf_counter *counter, int nmi,
int throttle, struct perf_sample_data *data) int throttle, struct perf_sample_data *data,
struct pt_regs *regs)
{ {
int events = atomic_read(&counter->event_limit); int events = atomic_read(&counter->event_limit);
struct hw_perf_counter *hwc = &counter->hw; struct hw_perf_counter *hwc = &counter->hw;
...@@ -3557,14 +3566,15 @@ static int __perf_counter_overflow(struct perf_counter *counter, int nmi, ...@@ -3557,14 +3566,15 @@ static int __perf_counter_overflow(struct perf_counter *counter, int nmi,
perf_counter_disable(counter); perf_counter_disable(counter);
} }
perf_counter_output(counter, nmi, data); perf_counter_output(counter, nmi, data, regs);
return ret; return ret;
} }
int perf_counter_overflow(struct perf_counter *counter, int nmi, int perf_counter_overflow(struct perf_counter *counter, int nmi,
struct perf_sample_data *data) struct perf_sample_data *data,
struct pt_regs *regs)
{ {
return __perf_counter_overflow(counter, nmi, 1, data); return __perf_counter_overflow(counter, nmi, 1, data, regs);
} }
/* /*
...@@ -3602,7 +3612,8 @@ static u64 perf_swcounter_set_period(struct perf_counter *counter) ...@@ -3602,7 +3612,8 @@ static u64 perf_swcounter_set_period(struct perf_counter *counter)
} }
static void perf_swcounter_overflow(struct perf_counter *counter, static void perf_swcounter_overflow(struct perf_counter *counter,
int nmi, struct perf_sample_data *data) int nmi, struct perf_sample_data *data,
struct pt_regs *regs)
{ {
struct hw_perf_counter *hwc = &counter->hw; struct hw_perf_counter *hwc = &counter->hw;
int throttle = 0; int throttle = 0;
...@@ -3615,7 +3626,8 @@ static void perf_swcounter_overflow(struct perf_counter *counter, ...@@ -3615,7 +3626,8 @@ static void perf_swcounter_overflow(struct perf_counter *counter,
return; return;
for (; overflow; overflow--) { for (; overflow; overflow--) {
if (__perf_counter_overflow(counter, nmi, throttle, data)) { if (__perf_counter_overflow(counter, nmi, throttle,
data, regs)) {
/* /*
* We inhibit the overflow from happening when * We inhibit the overflow from happening when
* hwc->interrupts == MAX_INTERRUPTS. * hwc->interrupts == MAX_INTERRUPTS.
...@@ -3634,7 +3646,8 @@ static void perf_swcounter_unthrottle(struct perf_counter *counter) ...@@ -3634,7 +3646,8 @@ static void perf_swcounter_unthrottle(struct perf_counter *counter)
} }
static void perf_swcounter_add(struct perf_counter *counter, u64 nr, static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
int nmi, struct perf_sample_data *data) int nmi, struct perf_sample_data *data,
struct pt_regs *regs)
{ {
struct hw_perf_counter *hwc = &counter->hw; struct hw_perf_counter *hwc = &counter->hw;
...@@ -3643,11 +3656,11 @@ static void perf_swcounter_add(struct perf_counter *counter, u64 nr, ...@@ -3643,11 +3656,11 @@ static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
if (!hwc->sample_period) if (!hwc->sample_period)
return; return;
if (!data->regs) if (!regs)
return; return;
if (!atomic64_add_negative(nr, &hwc->period_left)) if (!atomic64_add_negative(nr, &hwc->period_left))
perf_swcounter_overflow(counter, nmi, data); perf_swcounter_overflow(counter, nmi, data, regs);
} }
static int perf_swcounter_is_counting(struct perf_counter *counter) static int perf_swcounter_is_counting(struct perf_counter *counter)
...@@ -3706,7 +3719,8 @@ static int perf_swcounter_match(struct perf_counter *counter, ...@@ -3706,7 +3719,8 @@ static int perf_swcounter_match(struct perf_counter *counter,
static void perf_swcounter_ctx_event(struct perf_counter_context *ctx, static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
enum perf_type_id type, enum perf_type_id type,
u32 event, u64 nr, int nmi, u32 event, u64 nr, int nmi,
struct perf_sample_data *data) struct perf_sample_data *data,
struct pt_regs *regs)
{ {
struct perf_counter *counter; struct perf_counter *counter;
...@@ -3715,8 +3729,8 @@ static void perf_swcounter_ctx_event(struct perf_counter_context *ctx, ...@@ -3715,8 +3729,8 @@ static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
rcu_read_lock(); rcu_read_lock();
list_for_each_entry_rcu(counter, &ctx->event_list, event_entry) { list_for_each_entry_rcu(counter, &ctx->event_list, event_entry) {
if (perf_swcounter_match(counter, type, event, data->regs)) if (perf_swcounter_match(counter, type, event, regs))
perf_swcounter_add(counter, nr, nmi, data); perf_swcounter_add(counter, nr, nmi, data, regs);
} }
rcu_read_unlock(); rcu_read_unlock();
} }
...@@ -3737,7 +3751,8 @@ static int *perf_swcounter_recursion_context(struct perf_cpu_context *cpuctx) ...@@ -3737,7 +3751,8 @@ static int *perf_swcounter_recursion_context(struct perf_cpu_context *cpuctx)
static void do_perf_swcounter_event(enum perf_type_id type, u32 event, static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
u64 nr, int nmi, u64 nr, int nmi,
struct perf_sample_data *data) struct perf_sample_data *data,
struct pt_regs *regs)
{ {
struct perf_cpu_context *cpuctx = &get_cpu_var(perf_cpu_context); struct perf_cpu_context *cpuctx = &get_cpu_var(perf_cpu_context);
int *recursion = perf_swcounter_recursion_context(cpuctx); int *recursion = perf_swcounter_recursion_context(cpuctx);
...@@ -3750,7 +3765,7 @@ static void do_perf_swcounter_event(enum perf_type_id type, u32 event, ...@@ -3750,7 +3765,7 @@ static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
barrier(); barrier();
perf_swcounter_ctx_event(&cpuctx->ctx, type, event, perf_swcounter_ctx_event(&cpuctx->ctx, type, event,
nr, nmi, data); nr, nmi, data, regs);
rcu_read_lock(); rcu_read_lock();
/* /*
* doesn't really matter which of the child contexts the * doesn't really matter which of the child contexts the
...@@ -3758,7 +3773,7 @@ static void do_perf_swcounter_event(enum perf_type_id type, u32 event, ...@@ -3758,7 +3773,7 @@ static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
*/ */
ctx = rcu_dereference(current->perf_counter_ctxp); ctx = rcu_dereference(current->perf_counter_ctxp);
if (ctx) if (ctx)
perf_swcounter_ctx_event(ctx, type, event, nr, nmi, data); perf_swcounter_ctx_event(ctx, type, event, nr, nmi, data, regs);
rcu_read_unlock(); rcu_read_unlock();
barrier(); barrier();
...@@ -3772,11 +3787,11 @@ void __perf_swcounter_event(u32 event, u64 nr, int nmi, ...@@ -3772,11 +3787,11 @@ void __perf_swcounter_event(u32 event, u64 nr, int nmi,
struct pt_regs *regs, u64 addr) struct pt_regs *regs, u64 addr)
{ {
struct perf_sample_data data = { struct perf_sample_data data = {
.regs = regs,
.addr = addr, .addr = addr,
}; };
do_perf_swcounter_event(PERF_TYPE_SOFTWARE, event, nr, nmi, &data); do_perf_swcounter_event(PERF_TYPE_SOFTWARE, event, nr, nmi,
&data, regs);
} }
static void perf_swcounter_read(struct perf_counter *counter) static void perf_swcounter_read(struct perf_counter *counter)
...@@ -3813,6 +3828,7 @@ static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer) ...@@ -3813,6 +3828,7 @@ static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer)
{ {
enum hrtimer_restart ret = HRTIMER_RESTART; enum hrtimer_restart ret = HRTIMER_RESTART;
struct perf_sample_data data; struct perf_sample_data data;
struct pt_regs *regs;
struct perf_counter *counter; struct perf_counter *counter;
u64 period; u64 period;
...@@ -3820,17 +3836,17 @@ static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer) ...@@ -3820,17 +3836,17 @@ static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer)
counter->pmu->read(counter); counter->pmu->read(counter);
data.addr = 0; data.addr = 0;
data.regs = get_irq_regs(); regs = get_irq_regs();
/* /*
* In case we exclude kernel IPs or are somehow not in interrupt * In case we exclude kernel IPs or are somehow not in interrupt
* context, provide the next best thing, the user IP. * context, provide the next best thing, the user IP.
*/ */
if ((counter->attr.exclude_kernel || !data.regs) && if ((counter->attr.exclude_kernel || !regs) &&
!counter->attr.exclude_user) !counter->attr.exclude_user)
data.regs = task_pt_regs(current); regs = task_pt_regs(current);
if (data.regs) { if (regs) {
if (perf_counter_overflow(counter, 0, &data)) if (perf_counter_overflow(counter, 0, &data, regs))
ret = HRTIMER_NORESTART; ret = HRTIMER_NORESTART;
} }
...@@ -3966,15 +3982,17 @@ void perf_tpcounter_event(int event_id, u64 addr, u64 count, void *record, ...@@ -3966,15 +3982,17 @@ void perf_tpcounter_event(int event_id, u64 addr, u64 count, void *record,
}; };
struct perf_sample_data data = { struct perf_sample_data data = {
.regs = get_irq_regs(),
.addr = addr, .addr = addr,
.raw = &raw, .raw = &raw,
}; };
if (!data.regs) struct pt_regs *regs = get_irq_regs();
data.regs = task_pt_regs(current);
if (!regs)
regs = task_pt_regs(current);
do_perf_swcounter_event(PERF_TYPE_TRACEPOINT, event_id, count, 1, &data); do_perf_swcounter_event(PERF_TYPE_TRACEPOINT, event_id, count, 1,
&data, regs);
} }
EXPORT_SYMBOL_GPL(perf_tpcounter_event); EXPORT_SYMBOL_GPL(perf_tpcounter_event);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册