提交 82b89778 编写于 作者: A Adrian Hunter 提交者: Ingo Molnar

perf: Differentiate exec() and non-exec() comm events

perf tools like 'perf report' can aggregate samples by comm strings,
which generally works.  However, there are other potential use-cases.
For example, to pair up 'calls' with 'returns' accurately (from branch
events like Intel BTS) it is necessary to identify whether the process
has exec'd.  Although a comm event is generated when an 'exec' happens
it is also generated whenever the comm string is changed on a whim
(e.g. by prctl PR_SET_NAME).  This patch adds a flag to the comm event
to differentiate one case from the other.

In order to determine whether the kernel supports the new flag, a
selection bit named 'exec' is added to struct perf_event_attr.  The
bit does nothing but will cause perf_event_open() to fail if the bit
is set on kernels that do not have it defined.
Signed-off-by: NAdrian Hunter <adrian.hunter@intel.com>
Signed-off-by: NPeter Zijlstra <peterz@infradead.org>
Link: http://lkml.kernel.org/r/537D9EBE.7030806@intel.com
Cc: Paul Mackerras <paulus@samba.org>
Cc: Dave Jones <davej@redhat.com>
Cc: Arnaldo Carvalho de Melo <acme@kernel.org>
Cc: David Ahern <dsahern@gmail.com>
Cc: Jiri Olsa <jolsa@redhat.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: linux-fsdevel@vger.kernel.org
Cc: linux-kernel@vger.kernel.org
Signed-off-by: NIngo Molnar <mingo@kernel.org>
上级 ec000109
...@@ -1046,13 +1046,13 @@ EXPORT_SYMBOL_GPL(get_task_comm); ...@@ -1046,13 +1046,13 @@ EXPORT_SYMBOL_GPL(get_task_comm);
* so that a new one can be started * so that a new one can be started
*/ */
void set_task_comm(struct task_struct *tsk, const char *buf) void __set_task_comm(struct task_struct *tsk, const char *buf, bool exec)
{ {
task_lock(tsk); task_lock(tsk);
trace_task_rename(tsk, buf); trace_task_rename(tsk, buf);
strlcpy(tsk->comm, buf, sizeof(tsk->comm)); strlcpy(tsk->comm, buf, sizeof(tsk->comm));
task_unlock(tsk); task_unlock(tsk);
perf_event_comm(tsk); perf_event_comm(tsk, exec);
} }
int flush_old_exec(struct linux_binprm * bprm) int flush_old_exec(struct linux_binprm * bprm)
...@@ -1111,7 +1111,7 @@ void setup_new_exec(struct linux_binprm * bprm) ...@@ -1111,7 +1111,7 @@ void setup_new_exec(struct linux_binprm * bprm)
set_dumpable(current->mm, suid_dumpable); set_dumpable(current->mm, suid_dumpable);
perf_event_exec(); perf_event_exec();
set_task_comm(current, kbasename(bprm->filename)); __set_task_comm(current, kbasename(bprm->filename), true);
/* Set the new mm task size. We have to do that late because it may /* Set the new mm task size. We have to do that late because it may
* depend on TIF_32BIT which is only updated in flush_thread() on * depend on TIF_32BIT which is only updated in flush_thread() on
......
...@@ -707,7 +707,7 @@ extern int perf_register_guest_info_callbacks(struct perf_guest_info_callbacks * ...@@ -707,7 +707,7 @@ extern int perf_register_guest_info_callbacks(struct perf_guest_info_callbacks *
extern int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *callbacks); extern int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *callbacks);
extern void perf_event_exec(void); extern void perf_event_exec(void);
extern void perf_event_comm(struct task_struct *tsk); extern void perf_event_comm(struct task_struct *tsk, bool exec);
extern void perf_event_fork(struct task_struct *tsk); extern void perf_event_fork(struct task_struct *tsk);
/* Callchains */ /* Callchains */
...@@ -815,7 +815,7 @@ static inline int perf_unregister_guest_info_callbacks ...@@ -815,7 +815,7 @@ static inline int perf_unregister_guest_info_callbacks
static inline void perf_event_mmap(struct vm_area_struct *vma) { } static inline void perf_event_mmap(struct vm_area_struct *vma) { }
static inline void perf_event_exec(void) { } static inline void perf_event_exec(void) { }
static inline void perf_event_comm(struct task_struct *tsk) { } static inline void perf_event_comm(struct task_struct *tsk, bool exec) { }
static inline void perf_event_fork(struct task_struct *tsk) { } static inline void perf_event_fork(struct task_struct *tsk) { }
static inline void perf_event_init(void) { } static inline void perf_event_init(void) { }
static inline int perf_swevent_get_recursion_context(void) { return -1; } static inline int perf_swevent_get_recursion_context(void) { return -1; }
......
...@@ -2379,7 +2379,11 @@ extern long do_fork(unsigned long, unsigned long, unsigned long, int __user *, i ...@@ -2379,7 +2379,11 @@ extern long do_fork(unsigned long, unsigned long, unsigned long, int __user *, i
struct task_struct *fork_idle(int); struct task_struct *fork_idle(int);
extern pid_t kernel_thread(int (*fn)(void *), void *arg, unsigned long flags); extern pid_t kernel_thread(int (*fn)(void *), void *arg, unsigned long flags);
extern void set_task_comm(struct task_struct *tsk, const char *from); extern void __set_task_comm(struct task_struct *tsk, const char *from, bool exec);
static inline void set_task_comm(struct task_struct *tsk, const char *from)
{
__set_task_comm(tsk, from, false);
}
extern char *get_task_comm(char *to, struct task_struct *tsk); extern char *get_task_comm(char *to, struct task_struct *tsk);
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
......
...@@ -302,8 +302,8 @@ struct perf_event_attr { ...@@ -302,8 +302,8 @@ struct perf_event_attr {
exclude_callchain_kernel : 1, /* exclude kernel callchains */ exclude_callchain_kernel : 1, /* exclude kernel callchains */
exclude_callchain_user : 1, /* exclude user callchains */ exclude_callchain_user : 1, /* exclude user callchains */
mmap2 : 1, /* include mmap with inode data */ mmap2 : 1, /* include mmap with inode data */
comm_exec : 1, /* flag comm events that are due to an exec */
__reserved_1 : 40; __reserved_1 : 39;
union { union {
__u32 wakeup_events; /* wakeup every n events */ __u32 wakeup_events; /* wakeup every n events */
...@@ -502,7 +502,12 @@ struct perf_event_mmap_page { ...@@ -502,7 +502,12 @@ struct perf_event_mmap_page {
#define PERF_RECORD_MISC_GUEST_KERNEL (4 << 0) #define PERF_RECORD_MISC_GUEST_KERNEL (4 << 0)
#define PERF_RECORD_MISC_GUEST_USER (5 << 0) #define PERF_RECORD_MISC_GUEST_USER (5 << 0)
/*
* PERF_RECORD_MISC_MMAP_DATA and PERF_RECORD_MISC_COMM_EXEC are used on
* different events so can reuse the same bit position.
*/
#define PERF_RECORD_MISC_MMAP_DATA (1 << 13) #define PERF_RECORD_MISC_MMAP_DATA (1 << 13)
#define PERF_RECORD_MISC_COMM_EXEC (1 << 13)
/* /*
* Indicates that the content of PERF_SAMPLE_IP points to * Indicates that the content of PERF_SAMPLE_IP points to
* the actual instruction that triggered the event. See also * the actual instruction that triggered the event. See also
......
...@@ -5090,7 +5090,7 @@ static void perf_event_comm_event(struct perf_comm_event *comm_event) ...@@ -5090,7 +5090,7 @@ static void perf_event_comm_event(struct perf_comm_event *comm_event)
NULL); NULL);
} }
void perf_event_comm(struct task_struct *task) void perf_event_comm(struct task_struct *task, bool exec)
{ {
struct perf_comm_event comm_event; struct perf_comm_event comm_event;
...@@ -5104,7 +5104,7 @@ void perf_event_comm(struct task_struct *task) ...@@ -5104,7 +5104,7 @@ void perf_event_comm(struct task_struct *task)
.event_id = { .event_id = {
.header = { .header = {
.type = PERF_RECORD_COMM, .type = PERF_RECORD_COMM,
.misc = 0, .misc = exec ? PERF_RECORD_MISC_COMM_EXEC : 0,
/* .size */ /* .size */
}, },
/* .pid */ /* .pid */
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册