event.c 33.7 KB
Newer Older
1
#include <linux/types.h>
2
#include <sys/mman.h>
3 4
#include "event.h"
#include "debug.h"
5
#include "hist.h"
6
#include "machine.h"
7
#include "sort.h"
8
#include "string.h"
9
#include "strlist.h"
10
#include "thread.h"
11
#include "thread_map.h"
12
#include "symbol/kallsyms.h"
13 14
#include "asm/bug.h"
#include "stat.h"
15

16
static const char *perf_event__names[] = {
17 18
	[0]					= "TOTAL",
	[PERF_RECORD_MMAP]			= "MMAP",
19
	[PERF_RECORD_MMAP2]			= "MMAP2",
20 21 22 23 24 25 26 27
	[PERF_RECORD_LOST]			= "LOST",
	[PERF_RECORD_COMM]			= "COMM",
	[PERF_RECORD_EXIT]			= "EXIT",
	[PERF_RECORD_THROTTLE]			= "THROTTLE",
	[PERF_RECORD_UNTHROTTLE]		= "UNTHROTTLE",
	[PERF_RECORD_FORK]			= "FORK",
	[PERF_RECORD_READ]			= "READ",
	[PERF_RECORD_SAMPLE]			= "SAMPLE",
28
	[PERF_RECORD_AUX]			= "AUX",
29
	[PERF_RECORD_ITRACE_START]		= "ITRACE_START",
30
	[PERF_RECORD_LOST_SAMPLES]		= "LOST_SAMPLES",
31 32
	[PERF_RECORD_SWITCH]			= "SWITCH",
	[PERF_RECORD_SWITCH_CPU_WIDE]		= "SWITCH_CPU_WIDE",
33 34 35 36 37
	[PERF_RECORD_HEADER_ATTR]		= "ATTR",
	[PERF_RECORD_HEADER_EVENT_TYPE]		= "EVENT_TYPE",
	[PERF_RECORD_HEADER_TRACING_DATA]	= "TRACING_DATA",
	[PERF_RECORD_HEADER_BUILD_ID]		= "BUILD_ID",
	[PERF_RECORD_FINISHED_ROUND]		= "FINISHED_ROUND",
A
Adrian Hunter 已提交
38
	[PERF_RECORD_ID_INDEX]			= "ID_INDEX",
39 40
	[PERF_RECORD_AUXTRACE_INFO]		= "AUXTRACE_INFO",
	[PERF_RECORD_AUXTRACE]			= "AUXTRACE",
41
	[PERF_RECORD_AUXTRACE_ERROR]		= "AUXTRACE_ERROR",
42
	[PERF_RECORD_THREAD_MAP]		= "THREAD_MAP",
43
	[PERF_RECORD_CPU_MAP]			= "CPU_MAP",
44
	[PERF_RECORD_STAT_CONFIG]		= "STAT_CONFIG",
45 46
};

47
const char *perf_event__name(unsigned int id)
48
{
49
	if (id >= ARRAY_SIZE(perf_event__names))
50
		return "INVALID";
51
	if (!perf_event__names[id])
52
		return "UNKNOWN";
53
	return perf_event__names[id];
54 55
}

56
static struct perf_sample synth_sample = {
57 58 59 60 61 62 63 64
	.pid	   = -1,
	.tid	   = -1,
	.time	   = -1,
	.stream_id = -1,
	.cpu	   = -1,
	.period	   = 1,
};

65 66
/*
 * Assumes that the first 4095 bytes of /proc/pid/stat contains
67
 * the comm, tgid and ppid.
68
 */
69 70
static int perf_event__get_comm_ids(pid_t pid, char *comm, size_t len,
				    pid_t *tgid, pid_t *ppid)
71 72
{
	char filename[PATH_MAX];
73 74
	char bf[4096];
	int fd;
75 76
	size_t size = 0;
	ssize_t n;
77 78 79 80
	char *nl, *name, *tgids, *ppids;

	*tgid = -1;
	*ppid = -1;
81 82 83

	snprintf(filename, sizeof(filename), "/proc/%d/status", pid);

84 85
	fd = open(filename, O_RDONLY);
	if (fd < 0) {
86
		pr_debug("couldn't open %s\n", filename);
87
		return -1;
88 89
	}

90 91 92
	n = read(fd, bf, sizeof(bf) - 1);
	close(fd);
	if (n <= 0) {
93
		pr_warning("Couldn't get COMM, tigd and ppid for pid %d\n",
94 95
			   pid);
		return -1;
96
	}
97
	bf[n] = '\0';
98

99 100
	name = strstr(bf, "Name:");
	tgids = strstr(bf, "Tgid:");
101
	ppids = strstr(bf, "PPid:");
102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123

	if (name) {
		name += 5;  /* strlen("Name:") */

		while (*name && isspace(*name))
			++name;

		nl = strchr(name, '\n');
		if (nl)
			*nl = '\0';

		size = strlen(name);
		if (size >= len)
			size = len - 1;
		memcpy(comm, name, size);
		comm[size] = '\0';
	} else {
		pr_debug("Name: string not found for pid %d\n", pid);
	}

	if (tgids) {
		tgids += 5;  /* strlen("Tgid:") */
124
		*tgid = atoi(tgids);
125 126 127
	} else {
		pr_debug("Tgid: string not found for pid %d\n", pid);
	}
128

129 130 131 132 133 134 135 136
	if (ppids) {
		ppids += 5;  /* strlen("PPid:") */
		*ppid = atoi(ppids);
	} else {
		pr_debug("PPid: string not found for pid %d\n", pid);
	}

	return 0;
137 138
}

139 140 141
static int perf_event__prepare_comm(union perf_event *event, pid_t pid,
				    struct machine *machine,
				    pid_t *tgid, pid_t *ppid)
142 143
{
	size_t size;
144 145

	*ppid = -1;
146 147 148

	memset(&event->comm, 0, sizeof(event->comm));

149 150 151 152 153 154 155 156 157
	if (machine__is_host(machine)) {
		if (perf_event__get_comm_ids(pid, event->comm.comm,
					     sizeof(event->comm.comm),
					     tgid, ppid) != 0) {
			return -1;
		}
	} else {
		*tgid = machine->pid;
	}
158

159 160
	if (*tgid < 0)
		return -1;
161

162
	event->comm.pid = *tgid;
163
	event->comm.header.type = PERF_RECORD_COMM;
164 165

	size = strlen(event->comm.comm) + 1;
166
	size = PERF_ALIGN(size, sizeof(u64));
167
	memset(event->comm.comm + size, 0, machine->id_hdr_size);
168 169
	event->comm.header.size = (sizeof(event->comm) -
				(sizeof(event->comm.comm) - size) +
170
				machine->id_hdr_size);
171
	event->comm.tid = pid;
172 173

	return 0;
174 175
}

176
pid_t perf_event__synthesize_comm(struct perf_tool *tool,
177 178 179 180
					 union perf_event *event, pid_t pid,
					 perf_event__handler_t process,
					 struct machine *machine)
{
181
	pid_t tgid, ppid;
182

183 184
	if (perf_event__prepare_comm(event, pid, machine, &tgid, &ppid) != 0)
		return -1;
185

186 187
	if (process(tool, event, &synth_sample, machine) != 0)
		return -1;
188

189
	return tgid;
190 191
}

192
static int perf_event__synthesize_fork(struct perf_tool *tool,
193 194 195
				       union perf_event *event,
				       pid_t pid, pid_t tgid, pid_t ppid,
				       perf_event__handler_t process,
196 197 198 199
				       struct machine *machine)
{
	memset(&event->fork, 0, sizeof(event->fork) + machine->id_hdr_size);

200 201 202 203 204 205 206 207 208 209 210 211
	/*
	 * for main thread set parent to ppid from status file. For other
	 * threads set parent pid to main thread. ie., assume main thread
	 * spawns all threads in a process
	*/
	if (tgid == pid) {
		event->fork.ppid = ppid;
		event->fork.ptid = ppid;
	} else {
		event->fork.ppid = tgid;
		event->fork.ptid = tgid;
	}
212 213 214 215 216 217 218 219 220 221 222 223
	event->fork.pid  = tgid;
	event->fork.tid  = pid;
	event->fork.header.type = PERF_RECORD_FORK;

	event->fork.header.size = (sizeof(event->fork) + machine->id_hdr_size);

	if (process(tool, event, &synth_sample, machine) != 0)
		return -1;

	return 0;
}

224 225 226 227 228
int perf_event__synthesize_mmap_events(struct perf_tool *tool,
				       union perf_event *event,
				       pid_t pid, pid_t tgid,
				       perf_event__handler_t process,
				       struct machine *machine,
229 230
				       bool mmap_data,
				       unsigned int proc_map_timeout)
231 232 233
{
	char filename[PATH_MAX];
	FILE *fp;
234 235
	unsigned long long t;
	bool truncation = false;
236
	unsigned long long timeout = proc_map_timeout * 1000000ULL;
237
	int rc = 0;
238

239 240 241
	if (machine__is_default_guest(machine))
		return 0;

242 243
	snprintf(filename, sizeof(filename), "%s/proc/%d/maps",
		 machine->root_dir, pid);
244 245 246 247 248 249 250 251 252 253

	fp = fopen(filename, "r");
	if (fp == NULL) {
		/*
		 * We raced with a task exiting - just return:
		 */
		pr_debug("couldn't open %s\n", filename);
		return -1;
	}

254
	event->header.type = PERF_RECORD_MMAP2;
255
	t = rdclock();
256

257
	while (1) {
258 259 260 261
		char bf[BUFSIZ];
		char prot[5];
		char execname[PATH_MAX];
		char anonstr[] = "//anon";
262
		unsigned int ino;
263
		size_t size;
264
		ssize_t n;
265

266 267 268
		if (fgets(bf, sizeof(bf), fp) == NULL)
			break;

269 270 271 272 273
		if ((rdclock() - t) > timeout) {
			pr_warning("Reading %s time out. "
				   "You may want to increase "
				   "the time limit by --proc-map-timeout\n",
				   filename);
274 275 276 277
			truncation = true;
			goto out;
		}

278 279 280
		/* ensure null termination since stack will be reused. */
		strcpy(execname, "");

281
		/* 00400000-0040c000 r-xp 00000000 fd:01 41038  /bin/cat */
282 283 284 285 286 287
		n = sscanf(bf, "%"PRIx64"-%"PRIx64" %s %"PRIx64" %x:%x %u %s\n",
		       &event->mmap2.start, &event->mmap2.len, prot,
		       &event->mmap2.pgoff, &event->mmap2.maj,
		       &event->mmap2.min,
		       &ino, execname);

288 289 290
		/*
 		 * Anon maps don't have the execname.
 		 */
291
		if (n < 7)
292
			continue;
293 294 295

		event->mmap2.ino = (u64)ino;

296 297 298
		/*
		 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
		 */
299 300 301 302
		if (machine__is_host(machine))
			event->header.misc = PERF_RECORD_MISC_USER;
		else
			event->header.misc = PERF_RECORD_MISC_GUEST_USER;
303

304 305 306 307 308 309 310 311 312 313 314 315 316 317 318
		/* map protection and flags bits */
		event->mmap2.prot = 0;
		event->mmap2.flags = 0;
		if (prot[0] == 'r')
			event->mmap2.prot |= PROT_READ;
		if (prot[1] == 'w')
			event->mmap2.prot |= PROT_WRITE;
		if (prot[2] == 'x')
			event->mmap2.prot |= PROT_EXEC;

		if (prot[3] == 's')
			event->mmap2.flags |= MAP_SHARED;
		else
			event->mmap2.flags |= MAP_PRIVATE;

319 320 321 322 323 324
		if (prot[2] != 'x') {
			if (!mmap_data || prot[0] != 'r')
				continue;

			event->header.misc |= PERF_RECORD_MISC_MMAP_DATA;
		}
325

326 327 328 329
out:
		if (truncation)
			event->header.misc |= PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT;

330 331 332 333
		if (!strcmp(execname, ""))
			strcpy(execname, anonstr);

		size = strlen(execname) + 1;
334
		memcpy(event->mmap2.filename, execname, size);
335
		size = PERF_ALIGN(size, sizeof(u64));
336 337 338 339 340 341 342
		event->mmap2.len -= event->mmap.start;
		event->mmap2.header.size = (sizeof(event->mmap2) -
					(sizeof(event->mmap2.filename) - size));
		memset(event->mmap2.filename + size, 0, machine->id_hdr_size);
		event->mmap2.header.size += machine->id_hdr_size;
		event->mmap2.pid = tgid;
		event->mmap2.tid = pid;
343 344 345 346

		if (process(tool, event, &synth_sample, machine) != 0) {
			rc = -1;
			break;
347
		}
348 349 350

		if (truncation)
			break;
351 352 353
	}

	fclose(fp);
354
	return rc;
355 356
}

357
int perf_event__synthesize_modules(struct perf_tool *tool,
358
				   perf_event__handler_t process,
359
				   struct machine *machine)
360
{
361
	int rc = 0;
362
	struct map *pos;
363
	struct map_groups *kmaps = &machine->kmaps;
364
	struct maps *maps = &kmaps->maps[MAP__FUNCTION];
365
	union perf_event *event = zalloc((sizeof(event->mmap) +
366
					  machine->id_hdr_size));
367 368 369 370 371 372 373
	if (event == NULL) {
		pr_debug("Not enough memory synthesizing mmap event "
			 "for kernel modules\n");
		return -1;
	}

	event->header.type = PERF_RECORD_MMAP;
374

375 376 377 378
	/*
	 * kernel uses 0 for user space maps, see kernel/perf_event.c
	 * __perf_event_mmap
	 */
379
	if (machine__is_host(machine))
380
		event->header.misc = PERF_RECORD_MISC_KERNEL;
381
	else
382
		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
383

384
	for (pos = maps__first(maps); pos; pos = map__next(pos)) {
385 386
		size_t size;

387
		if (__map__is_kernel(pos))
388 389
			continue;

390
		size = PERF_ALIGN(pos->dso->long_name_len + 1, sizeof(u64));
391 392 393
		event->mmap.header.type = PERF_RECORD_MMAP;
		event->mmap.header.size = (sizeof(event->mmap) -
				        (sizeof(event->mmap.filename) - size));
394 395
		memset(event->mmap.filename + size, 0, machine->id_hdr_size);
		event->mmap.header.size += machine->id_hdr_size;
396 397 398 399 400
		event->mmap.start = pos->start;
		event->mmap.len   = pos->end - pos->start;
		event->mmap.pid   = machine->pid;

		memcpy(event->mmap.filename, pos->dso->long_name,
401
		       pos->dso->long_name_len + 1);
402 403 404 405
		if (process(tool, event, &synth_sample, machine) != 0) {
			rc = -1;
			break;
		}
406 407
	}

408
	free(event);
409
	return rc;
410 411
}

412 413
static int __event__synthesize_thread(union perf_event *comm_event,
				      union perf_event *mmap_event,
414
				      union perf_event *fork_event,
415 416
				      pid_t pid, int full,
					  perf_event__handler_t process,
417
				      struct perf_tool *tool,
418 419 420
				      struct machine *machine,
				      bool mmap_data,
				      unsigned int proc_map_timeout)
421
{
422 423 424
	char filename[PATH_MAX];
	DIR *tasks;
	struct dirent dirent, *next;
425
	pid_t tgid, ppid;
426
	int rc = 0;
427 428 429 430 431 432 433 434 435 436

	/* special case: only send one comm event using passed in pid */
	if (!full) {
		tgid = perf_event__synthesize_comm(tool, comm_event, pid,
						   process, machine);

		if (tgid == -1)
			return -1;

		return perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
437 438
							  process, machine, mmap_data,
							  proc_map_timeout);
439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460
	}

	if (machine__is_default_guest(machine))
		return 0;

	snprintf(filename, sizeof(filename), "%s/proc/%d/task",
		 machine->root_dir, pid);

	tasks = opendir(filename);
	if (tasks == NULL) {
		pr_debug("couldn't open %s\n", filename);
		return 0;
	}

	while (!readdir_r(tasks, &dirent, &next) && next) {
		char *end;
		pid_t _pid;

		_pid = strtol(dirent.d_name, &end, 10);
		if (*end)
			continue;

461
		rc = -1;
462 463
		if (perf_event__prepare_comm(comm_event, _pid, machine,
					     &tgid, &ppid) != 0)
464
			break;
465

466
		if (perf_event__synthesize_fork(tool, fork_event, _pid, tgid,
467
						ppid, process, machine) < 0)
468
			break;
469 470 471 472
		/*
		 * Send the prepared comm event
		 */
		if (process(tool, comm_event, &synth_sample, machine) != 0)
473
			break;
474

475
		rc = 0;
476 477 478
		if (_pid == pid) {
			/* process the parent's maps too */
			rc = perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
479
						process, machine, mmap_data, proc_map_timeout);
480 481
			if (rc)
				break;
482
		}
483 484 485
	}

	closedir(tasks);
486
	return rc;
487 488
}

489
int perf_event__synthesize_thread_map(struct perf_tool *tool,
490
				      struct thread_map *threads,
491
				      perf_event__handler_t process,
492
				      struct machine *machine,
493 494
				      bool mmap_data,
				      unsigned int proc_map_timeout)
495
{
496
	union perf_event *comm_event, *mmap_event, *fork_event;
497
	int err = -1, thread, j;
498

499
	comm_event = malloc(sizeof(comm_event->comm) + machine->id_hdr_size);
500 501 502
	if (comm_event == NULL)
		goto out;

503
	mmap_event = malloc(sizeof(mmap_event->mmap) + machine->id_hdr_size);
504 505 506
	if (mmap_event == NULL)
		goto out_free_comm;

507 508 509 510
	fork_event = malloc(sizeof(fork_event->fork) + machine->id_hdr_size);
	if (fork_event == NULL)
		goto out_free_mmap;

511 512 513
	err = 0;
	for (thread = 0; thread < threads->nr; ++thread) {
		if (__event__synthesize_thread(comm_event, mmap_event,
514
					       fork_event,
515
					       thread_map__pid(threads, thread), 0,
516
					       process, tool, machine,
517
					       mmap_data, proc_map_timeout)) {
518 519 520
			err = -1;
			break;
		}
521 522 523 524 525

		/*
		 * comm.pid is set to thread group id by
		 * perf_event__synthesize_comm
		 */
526
		if ((int) comm_event->comm.pid != thread_map__pid(threads, thread)) {
527 528 529 530
			bool need_leader = true;

			/* is thread group leader in thread_map? */
			for (j = 0; j < threads->nr; ++j) {
531
				if ((int) comm_event->comm.pid == thread_map__pid(threads, j)) {
532 533 534 535 536 537 538
					need_leader = false;
					break;
				}
			}

			/* if not, generate events for it */
			if (need_leader &&
539
			    __event__synthesize_thread(comm_event, mmap_event,
540
						       fork_event,
541 542
						       comm_event->comm.pid, 0,
						       process, tool, machine,
543
						       mmap_data, proc_map_timeout)) {
544 545 546 547
				err = -1;
				break;
			}
		}
548
	}
549 550
	free(fork_event);
out_free_mmap:
551 552 553 554 555 556 557
	free(mmap_event);
out_free_comm:
	free(comm_event);
out:
	return err;
}

558
int perf_event__synthesize_threads(struct perf_tool *tool,
559
				   perf_event__handler_t process,
560 561 562
				   struct machine *machine,
				   bool mmap_data,
				   unsigned int proc_map_timeout)
563 564
{
	DIR *proc;
565
	char proc_path[PATH_MAX];
566
	struct dirent dirent, *next;
567
	union perf_event *comm_event, *mmap_event, *fork_event;
568 569
	int err = -1;

570 571 572
	if (machine__is_default_guest(machine))
		return 0;

573
	comm_event = malloc(sizeof(comm_event->comm) + machine->id_hdr_size);
574 575 576
	if (comm_event == NULL)
		goto out;

577
	mmap_event = malloc(sizeof(mmap_event->mmap) + machine->id_hdr_size);
578 579
	if (mmap_event == NULL)
		goto out_free_comm;
580

581 582 583 584
	fork_event = malloc(sizeof(fork_event->fork) + machine->id_hdr_size);
	if (fork_event == NULL)
		goto out_free_mmap;

585 586 587
	snprintf(proc_path, sizeof(proc_path), "%s/proc", machine->root_dir);
	proc = opendir(proc_path);

588
	if (proc == NULL)
589
		goto out_free_fork;
590 591 592 593 594 595 596

	while (!readdir_r(proc, &dirent, &next) && next) {
		char *end;
		pid_t pid = strtol(dirent.d_name, &end, 10);

		if (*end) /* only interested in proper numerical dirents */
			continue;
597 598 599 600
		/*
 		 * We may race with exiting thread, so don't stop just because
 		 * one thread couldn't be synthesized.
 		 */
601
		__event__synthesize_thread(comm_event, mmap_event, fork_event, pid,
602 603
					   1, process, tool, machine, mmap_data,
					   proc_map_timeout);
604 605
	}

606
	err = 0;
607
	closedir(proc);
608 609
out_free_fork:
	free(fork_event);
610 611 612 613 614 615
out_free_mmap:
	free(mmap_event);
out_free_comm:
	free(comm_event);
out:
	return err;
616
}
617

618 619 620 621 622
struct process_symbol_args {
	const char *name;
	u64	   start;
};

623
static int find_symbol_cb(void *arg, const char *name, char type,
624
			  u64 start)
625 626 627
{
	struct process_symbol_args *args = arg;

628 629 630 631 632 633
	/*
	 * Must be a function or at least an alias, as in PARISC64, where "_text" is
	 * an 'A' to the same address as "_stext".
	 */
	if (!(symbol_type__is_a(type, MAP__FUNCTION) ||
	      type == 'A') || strcmp(name, args->name))
634 635 636 637 638 639
		return 0;

	args->start = start;
	return 1;
}

640 641 642 643 644 645 646 647 648 649 650
u64 kallsyms__get_function_start(const char *kallsyms_filename,
				 const char *symbol_name)
{
	struct process_symbol_args args = { .name = symbol_name, };

	if (kallsyms__parse(kallsyms_filename, &args, find_symbol_cb) <= 0)
		return 0;

	return args.start;
}

651
int perf_event__synthesize_kernel_mmap(struct perf_tool *tool,
652
				       perf_event__handler_t process,
653
				       struct machine *machine)
654 655
{
	size_t size;
656
	const char *mmap_name;
657
	char name_buff[PATH_MAX];
658
	struct map *map = machine__kernel_map(machine);
659
	struct kmap *kmap;
660
	int err;
661 662
	union perf_event *event;

663
	if (map == NULL)
664 665
		return -1;

666 667 668 669 670
	/*
	 * We should get this from /sys/kernel/sections/.text, but till that is
	 * available use this, and after it is use this as a fallback for older
	 * kernels.
	 */
671
	event = zalloc((sizeof(event->mmap) + machine->id_hdr_size));
672 673 674 675 676
	if (event == NULL) {
		pr_debug("Not enough memory synthesizing mmap event "
			 "for kernel modules\n");
		return -1;
	}
677

678
	mmap_name = machine__mmap_name(machine, name_buff, sizeof(name_buff));
679
	if (machine__is_host(machine)) {
680 681 682 683
		/*
		 * kernel uses PERF_RECORD_MISC_USER for user space maps,
		 * see kernel/perf_event.c __perf_event_mmap
		 */
684
		event->header.misc = PERF_RECORD_MISC_KERNEL;
685
	} else {
686
		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
687
	}
688

689
	kmap = map__kmap(map);
690
	size = snprintf(event->mmap.filename, sizeof(event->mmap.filename),
691
			"%s%s", mmap_name, kmap->ref_reloc_sym->name) + 1;
692
	size = PERF_ALIGN(size, sizeof(u64));
693 694
	event->mmap.header.type = PERF_RECORD_MMAP;
	event->mmap.header.size = (sizeof(event->mmap) -
695
			(sizeof(event->mmap.filename) - size) + machine->id_hdr_size);
696
	event->mmap.pgoff = kmap->ref_reloc_sym->addr;
697 698 699 700
	event->mmap.start = map->start;
	event->mmap.len   = map->end - event->mmap.start;
	event->mmap.pid   = machine->pid;

701
	err = process(tool, event, &synth_sample, machine);
702 703 704
	free(event);

	return err;
705 706
}

707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742
int perf_event__synthesize_thread_map2(struct perf_tool *tool,
				      struct thread_map *threads,
				      perf_event__handler_t process,
				      struct machine *machine)
{
	union perf_event *event;
	int i, err, size;

	size  = sizeof(event->thread_map);
	size +=	threads->nr * sizeof(event->thread_map.entries[0]);

	event = zalloc(size);
	if (!event)
		return -ENOMEM;

	event->header.type = PERF_RECORD_THREAD_MAP;
	event->header.size = size;
	event->thread_map.nr = threads->nr;

	for (i = 0; i < threads->nr; i++) {
		struct thread_map_event_entry *entry = &event->thread_map.entries[i];
		char *comm = thread_map__comm(threads, i);

		if (!comm)
			comm = (char *) "";

		entry->pid = thread_map__pid(threads, i);
		strncpy((char *) &entry->comm, comm, sizeof(entry->comm));
	}

	err = process(tool, event, NULL, machine);

	free(event);
	return err;
}

743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873
static void synthesize_cpus(struct cpu_map_entries *cpus,
			    struct cpu_map *map)
{
	int i;

	cpus->nr = map->nr;

	for (i = 0; i < map->nr; i++)
		cpus->cpu[i] = map->map[i];
}

static void synthesize_mask(struct cpu_map_mask *mask,
			    struct cpu_map *map, int max)
{
	int i;

	mask->nr = BITS_TO_LONGS(max);
	mask->long_size = sizeof(long);

	for (i = 0; i < map->nr; i++)
		set_bit(map->map[i], mask->mask);
}

static size_t cpus_size(struct cpu_map *map)
{
	return sizeof(struct cpu_map_entries) + map->nr * sizeof(u16);
}

static size_t mask_size(struct cpu_map *map, int *max)
{
	int i;

	*max = 0;

	for (i = 0; i < map->nr; i++) {
		/* bit possition of the cpu is + 1 */
		int bit = map->map[i] + 1;

		if (bit > *max)
			*max = bit;
	}

	return sizeof(struct cpu_map_mask) + BITS_TO_LONGS(*max) * sizeof(long);
}

void *cpu_map_data__alloc(struct cpu_map *map, size_t *size, u16 *type, int *max)
{
	size_t size_cpus, size_mask;
	bool is_dummy = cpu_map__empty(map);

	/*
	 * Both array and mask data have variable size based
	 * on the number of cpus and their actual values.
	 * The size of the 'struct cpu_map_data' is:
	 *
	 *   array = size of 'struct cpu_map_entries' +
	 *           number of cpus * sizeof(u64)
	 *
	 *   mask  = size of 'struct cpu_map_mask' +
	 *           maximum cpu bit converted to size of longs
	 *
	 * and finaly + the size of 'struct cpu_map_data'.
	 */
	size_cpus = cpus_size(map);
	size_mask = mask_size(map, max);

	if (is_dummy || (size_cpus < size_mask)) {
		*size += size_cpus;
		*type  = PERF_CPU_MAP__CPUS;
	} else {
		*size += size_mask;
		*type  = PERF_CPU_MAP__MASK;
	}

	*size += sizeof(struct cpu_map_data);
	return zalloc(*size);
}

void cpu_map_data__synthesize(struct cpu_map_data *data, struct cpu_map *map,
			      u16 type, int max)
{
	data->type = type;

	switch (type) {
	case PERF_CPU_MAP__CPUS:
		synthesize_cpus((struct cpu_map_entries *) data->data, map);
		break;
	case PERF_CPU_MAP__MASK:
		synthesize_mask((struct cpu_map_mask *) data->data, map, max);
	default:
		break;
	};
}

static struct cpu_map_event* cpu_map_event__new(struct cpu_map *map)
{
	size_t size = sizeof(struct cpu_map_event);
	struct cpu_map_event *event;
	int max;
	u16 type;

	event = cpu_map_data__alloc(map, &size, &type, &max);
	if (!event)
		return NULL;

	event->header.type = PERF_RECORD_CPU_MAP;
	event->header.size = size;
	event->data.type   = type;

	cpu_map_data__synthesize(&event->data, map, type, max);
	return event;
}

int perf_event__synthesize_cpu_map(struct perf_tool *tool,
				   struct cpu_map *map,
				   perf_event__handler_t process,
				   struct machine *machine)
{
	struct cpu_map_event *event;
	int err;

	event = cpu_map_event__new(map);
	if (!event)
		return -ENOMEM;

	err = process(tool, (union perf_event *) event, NULL, machine);

	free(event);
	return err;
}

874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911
int perf_event__synthesize_stat_config(struct perf_tool *tool,
				       struct perf_stat_config *config,
				       perf_event__handler_t process,
				       struct machine *machine)
{
	struct stat_config_event *event;
	int size, i = 0, err;

	size  = sizeof(*event);
	size += (PERF_STAT_CONFIG_TERM__MAX * sizeof(event->data[0]));

	event = zalloc(size);
	if (!event)
		return -ENOMEM;

	event->header.type = PERF_RECORD_STAT_CONFIG;
	event->header.size = size;
	event->nr          = PERF_STAT_CONFIG_TERM__MAX;

#define ADD(__term, __val)					\
	event->data[i].tag = PERF_STAT_CONFIG_TERM__##__term;	\
	event->data[i].val = __val;				\
	i++;

	ADD(AGGR_MODE,	config->aggr_mode)
	ADD(INTERVAL,	config->interval)
	ADD(SCALE,	config->scale)

	WARN_ONCE(i != PERF_STAT_CONFIG_TERM__MAX,
		  "stat config terms unbalanced\n");
#undef ADD

	err = process(tool, (union perf_event *) event, NULL, machine);

	free(event);
	return err;
}

912 913
size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp)
{
914 915 916 917 918 919 920
	const char *s;

	if (event->header.misc & PERF_RECORD_MISC_COMM_EXEC)
		s = " exec";
	else
		s = "";

921
	return fprintf(fp, "%s: %s:%d/%d\n", s, event->comm.comm, event->comm.pid, event->comm.tid);
922 923
}

924
int perf_event__process_comm(struct perf_tool *tool __maybe_unused,
925
			     union perf_event *event,
926
			     struct perf_sample *sample,
927
			     struct machine *machine)
928
{
929
	return machine__process_comm_event(machine, event, sample);
930 931
}

932
int perf_event__process_lost(struct perf_tool *tool __maybe_unused,
933
			     union perf_event *event,
934
			     struct perf_sample *sample,
935
			     struct machine *machine)
936
{
937
	return machine__process_lost_event(machine, event, sample);
938
}
939

940 941 942 943 944 945 946 947
int perf_event__process_aux(struct perf_tool *tool __maybe_unused,
			    union perf_event *event,
			    struct perf_sample *sample __maybe_unused,
			    struct machine *machine)
{
	return machine__process_aux_event(machine, event);
}

948 949 950 951 952 953 954 955
int perf_event__process_itrace_start(struct perf_tool *tool __maybe_unused,
				     union perf_event *event,
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine)
{
	return machine__process_itrace_start_event(machine, event);
}

956 957 958 959 960 961 962 963
int perf_event__process_lost_samples(struct perf_tool *tool __maybe_unused,
				     union perf_event *event,
				     struct perf_sample *sample,
				     struct machine *machine)
{
	return machine__process_lost_samples_event(machine, event, sample);
}

964 965 966 967 968 969 970 971
int perf_event__process_switch(struct perf_tool *tool __maybe_unused,
			       union perf_event *event,
			       struct perf_sample *sample __maybe_unused,
			       struct machine *machine)
{
	return machine__process_switch_event(machine, event);
}

972 973
size_t perf_event__fprintf_mmap(union perf_event *event, FILE *fp)
{
974
	return fprintf(fp, " %d/%d: [%#" PRIx64 "(%#" PRIx64 ") @ %#" PRIx64 "]: %c %s\n",
975
		       event->mmap.pid, event->mmap.tid, event->mmap.start,
976 977 978
		       event->mmap.len, event->mmap.pgoff,
		       (event->header.misc & PERF_RECORD_MISC_MMAP_DATA) ? 'r' : 'x',
		       event->mmap.filename);
979 980
}

981 982 983
size_t perf_event__fprintf_mmap2(union perf_event *event, FILE *fp)
{
	return fprintf(fp, " %d/%d: [%#" PRIx64 "(%#" PRIx64 ") @ %#" PRIx64
984
			   " %02x:%02x %"PRIu64" %"PRIu64"]: %c%c%c%c %s\n",
985 986 987 988
		       event->mmap2.pid, event->mmap2.tid, event->mmap2.start,
		       event->mmap2.len, event->mmap2.pgoff, event->mmap2.maj,
		       event->mmap2.min, event->mmap2.ino,
		       event->mmap2.ino_generation,
989 990 991 992
		       (event->mmap2.prot & PROT_READ) ? 'r' : '-',
		       (event->mmap2.prot & PROT_WRITE) ? 'w' : '-',
		       (event->mmap2.prot & PROT_EXEC) ? 'x' : '-',
		       (event->mmap2.flags & MAP_SHARED) ? 's' : 'p',
993 994 995
		       event->mmap2.filename);
}

996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011
size_t perf_event__fprintf_thread_map(union perf_event *event, FILE *fp)
{
	struct thread_map *threads = thread_map__new_event(&event->thread_map);
	size_t ret;

	ret = fprintf(fp, " nr: ");

	if (threads)
		ret += thread_map__fprintf(threads, fp);
	else
		ret += fprintf(fp, "failed to get threads from event\n");

	thread_map__put(threads);
	return ret;
}

1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027
size_t perf_event__fprintf_cpu_map(union perf_event *event, FILE *fp)
{
	struct cpu_map *cpus = cpu_map__new_data(&event->cpu_map.data);
	size_t ret;

	ret = fprintf(fp, " nr: ");

	if (cpus)
		ret += cpu_map__fprintf(cpus, fp);
	else
		ret += fprintf(fp, "failed to get cpumap from event\n");

	cpu_map__put(cpus);
	return ret;
}

1028
int perf_event__process_mmap(struct perf_tool *tool __maybe_unused,
1029
			     union perf_event *event,
1030
			     struct perf_sample *sample,
1031
			     struct machine *machine)
1032
{
1033
	return machine__process_mmap_event(machine, event, sample);
1034 1035
}

1036 1037
int perf_event__process_mmap2(struct perf_tool *tool __maybe_unused,
			     union perf_event *event,
1038
			     struct perf_sample *sample,
1039 1040
			     struct machine *machine)
{
1041
	return machine__process_mmap2_event(machine, event, sample);
1042 1043
}

1044 1045 1046 1047 1048 1049 1050
size_t perf_event__fprintf_task(union perf_event *event, FILE *fp)
{
	return fprintf(fp, "(%d:%d):(%d:%d)\n",
		       event->fork.pid, event->fork.tid,
		       event->fork.ppid, event->fork.ptid);
}

1051
int perf_event__process_fork(struct perf_tool *tool __maybe_unused,
1052
			     union perf_event *event,
1053
			     struct perf_sample *sample,
1054
			     struct machine *machine)
1055
{
1056
	return machine__process_fork_event(machine, event, sample);
1057
}
1058

1059 1060
int perf_event__process_exit(struct perf_tool *tool __maybe_unused,
			     union perf_event *event,
1061
			     struct perf_sample *sample,
1062 1063
			     struct machine *machine)
{
1064
	return machine__process_exit_event(machine, event, sample);
1065 1066
}

1067 1068 1069 1070 1071 1072 1073 1074 1075
size_t perf_event__fprintf_aux(union perf_event *event, FILE *fp)
{
	return fprintf(fp, " offset: %#"PRIx64" size: %#"PRIx64" flags: %#"PRIx64" [%s%s]\n",
		       event->aux.aux_offset, event->aux.aux_size,
		       event->aux.flags,
		       event->aux.flags & PERF_AUX_FLAG_TRUNCATED ? "T" : "",
		       event->aux.flags & PERF_AUX_FLAG_OVERWRITE ? "O" : "");
}

1076 1077 1078 1079 1080 1081
size_t perf_event__fprintf_itrace_start(union perf_event *event, FILE *fp)
{
	return fprintf(fp, " pid: %u tid: %u\n",
		       event->itrace_start.pid, event->itrace_start.tid);
}

1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095
size_t perf_event__fprintf_switch(union perf_event *event, FILE *fp)
{
	bool out = event->header.misc & PERF_RECORD_MISC_SWITCH_OUT;
	const char *in_out = out ? "OUT" : "IN ";

	if (event->header.type == PERF_RECORD_SWITCH)
		return fprintf(fp, " %s\n", in_out);

	return fprintf(fp, " %s  %s pid/tid: %5u/%-5u\n",
		       in_out, out ? "next" : "prev",
		       event->context_switch.next_prev_pid,
		       event->context_switch.next_prev_tid);
}

1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111
size_t perf_event__fprintf(union perf_event *event, FILE *fp)
{
	size_t ret = fprintf(fp, "PERF_RECORD_%s",
			     perf_event__name(event->header.type));

	switch (event->header.type) {
	case PERF_RECORD_COMM:
		ret += perf_event__fprintf_comm(event, fp);
		break;
	case PERF_RECORD_FORK:
	case PERF_RECORD_EXIT:
		ret += perf_event__fprintf_task(event, fp);
		break;
	case PERF_RECORD_MMAP:
		ret += perf_event__fprintf_mmap(event, fp);
		break;
1112 1113 1114
	case PERF_RECORD_MMAP2:
		ret += perf_event__fprintf_mmap2(event, fp);
		break;
1115 1116 1117
	case PERF_RECORD_AUX:
		ret += perf_event__fprintf_aux(event, fp);
		break;
1118 1119 1120
	case PERF_RECORD_ITRACE_START:
		ret += perf_event__fprintf_itrace_start(event, fp);
		break;
1121 1122 1123 1124
	case PERF_RECORD_SWITCH:
	case PERF_RECORD_SWITCH_CPU_WIDE:
		ret += perf_event__fprintf_switch(event, fp);
		break;
1125 1126 1127 1128 1129 1130 1131
	default:
		ret += fprintf(fp, "\n");
	}

	return ret;
}

1132 1133
int perf_event__process(struct perf_tool *tool __maybe_unused,
			union perf_event *event,
1134
			struct perf_sample *sample,
1135
			struct machine *machine)
1136
{
1137
	return machine__process_event(machine, event, sample);
1138 1139
}

1140
void thread__find_addr_map(struct thread *thread, u8 cpumode,
1141
			   enum map_type type, u64 addr,
1142
			   struct addr_location *al)
1143
{
1144
	struct map_groups *mg = thread->mg;
1145
	struct machine *machine = mg->machine;
1146
	bool load_map = false;
1147

1148
	al->machine = machine;
1149
	al->thread = thread;
1150
	al->addr = addr;
1151
	al->cpumode = cpumode;
1152
	al->filtered = 0;
1153

1154 1155 1156 1157 1158
	if (machine == NULL) {
		al->map = NULL;
		return;
	}

1159
	if (cpumode == PERF_RECORD_MISC_KERNEL && perf_host) {
1160
		al->level = 'k';
1161
		mg = &machine->kmaps;
1162
		load_map = true;
1163
	} else if (cpumode == PERF_RECORD_MISC_USER && perf_host) {
1164
		al->level = '.';
1165 1166
	} else if (cpumode == PERF_RECORD_MISC_GUEST_KERNEL && perf_guest) {
		al->level = 'g';
1167
		mg = &machine->kmaps;
1168
		load_map = true;
1169 1170
	} else if (cpumode == PERF_RECORD_MISC_GUEST_USER && perf_guest) {
		al->level = 'u';
1171
	} else {
1172
		al->level = 'H';
1173
		al->map = NULL;
1174 1175 1176 1177

		if ((cpumode == PERF_RECORD_MISC_GUEST_USER ||
			cpumode == PERF_RECORD_MISC_GUEST_KERNEL) &&
			!perf_guest)
1178
			al->filtered |= (1 << HIST_FILTER__GUEST);
1179 1180 1181
		if ((cpumode == PERF_RECORD_MISC_USER ||
			cpumode == PERF_RECORD_MISC_KERNEL) &&
			!perf_host)
1182
			al->filtered |= (1 << HIST_FILTER__HOST);
1183

1184 1185 1186
		return;
	}
try_again:
1187
	al->map = map_groups__find(mg, type, al->addr);
1188 1189 1190 1191 1192 1193 1194 1195 1196 1197
	if (al->map == NULL) {
		/*
		 * If this is outside of all known maps, and is a negative
		 * address, try to look it up in the kernel dso, as it might be
		 * a vsyscall or vdso (which executes in user-mode).
		 *
		 * XXX This is nasty, we should have a symbol list in the
		 * "[vdso]" dso, but for now lets use the old trick of looking
		 * in the whole kernel symbol list.
		 */
1198 1199 1200
		if (cpumode == PERF_RECORD_MISC_USER && machine &&
		    mg != &machine->kmaps &&
		    machine__kernel_ip(machine, al->addr)) {
1201
			mg = &machine->kmaps;
1202
			load_map = true;
1203 1204
			goto try_again;
		}
1205 1206 1207 1208 1209 1210
	} else {
		/*
		 * Kernel maps might be changed when loading symbols so loading
		 * must be done prior to using kernel maps.
		 */
		if (load_map)
1211
			map__load(al->map, machine->symbol_filter);
1212
		al->addr = al->map->map_ip(al->map, al->addr);
1213
	}
1214 1215
}

1216
void thread__find_addr_location(struct thread *thread,
1217
				u8 cpumode, enum map_type type, u64 addr,
1218
				struct addr_location *al)
1219
{
1220
	thread__find_addr_map(thread, cpumode, type, addr, al);
1221
	if (al->map != NULL)
1222
		al->sym = map__find_symbol(al->map, al->addr,
1223
					   thread->mg->machine->symbol_filter);
1224 1225
	else
		al->sym = NULL;
1226 1227
}

1228 1229 1230 1231
/*
 * Callers need to drop the reference to al->thread, obtained in
 * machine__findnew_thread()
 */
1232
int perf_event__preprocess_sample(const union perf_event *event,
1233
				  struct machine *machine,
1234
				  struct addr_location *al,
1235
				  struct perf_sample *sample)
1236
{
1237
	u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
1238
	struct thread *thread = machine__findnew_thread(machine, sample->pid,
1239
							sample->tid);
1240

1241 1242 1243
	if (thread == NULL)
		return -1;

1244
	dump_printf(" ... thread: %s:%d\n", thread__comm_str(thread), thread->tid);
1245
	/*
1246
	 * Have we already created the kernel maps for this machine?
1247 1248 1249 1250 1251 1252
	 *
	 * This should have happened earlier, when we processed the kernel MMAP
	 * events, but for older perf.data files there was no such thing, so do
	 * it now.
	 */
	if (cpumode == PERF_RECORD_MISC_KERNEL &&
1253
	    machine__kernel_map(machine) == NULL)
1254
		machine__create_kernel_maps(machine);
1255

1256
	thread__find_addr_map(thread, cpumode, MAP__FUNCTION, sample->ip, al);
1257 1258 1259
	dump_printf(" ...... dso: %s\n",
		    al->map ? al->map->dso->long_name :
			al->level == 'H' ? "[hypervisor]" : "<not found>");
1260 1261 1262 1263

	if (thread__is_filtered(thread))
		al->filtered |= (1 << HIST_FILTER__THREAD);

1264
	al->sym = NULL;
1265
	al->cpu = sample->cpu;
1266 1267 1268 1269 1270 1271 1272 1273
	al->socket = -1;

	if (al->cpu >= 0) {
		struct perf_env *env = machine->env;

		if (env && env->cpu)
			al->socket = env->cpu[al->cpu].socket_id;
	}
1274 1275

	if (al->map) {
1276 1277
		struct dso *dso = al->map->dso;

1278
		if (symbol_conf.dso_list &&
1279 1280 1281 1282
		    (!dso || !(strlist__has_entry(symbol_conf.dso_list,
						  dso->short_name) ||
			       (dso->short_name != dso->long_name &&
				strlist__has_entry(symbol_conf.dso_list,
1283 1284 1285
						   dso->long_name))))) {
			al->filtered |= (1 << HIST_FILTER__DSO);
		}
1286

1287 1288
		al->sym = map__find_symbol(al->map, al->addr,
					   machine->symbol_filter);
1289
	}
1290

1291 1292
	if (symbol_conf.sym_list &&
		(!al->sym || !strlist__has_entry(symbol_conf.sym_list,
1293 1294 1295
						al->sym->name))) {
		al->filtered |= (1 << HIST_FILTER__SYMBOL);
	}
1296 1297

	return 0;
1298
}
1299

1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310
/*
 * The preprocess_sample method will return with reference counts for the
 * in it, when done using (and perhaps getting ref counts if needing to
 * keep a pointer to one of those entries) it must be paired with
 * addr_location__put(), so that the refcounts can be decremented.
 */
void addr_location__put(struct addr_location *al)
{
	thread__zput(al->thread);
}

1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338
bool is_bts_event(struct perf_event_attr *attr)
{
	return attr->type == PERF_TYPE_HARDWARE &&
	       (attr->config & PERF_COUNT_HW_BRANCH_INSTRUCTIONS) &&
	       attr->sample_period == 1;
}

bool sample_addr_correlates_sym(struct perf_event_attr *attr)
{
	if (attr->type == PERF_TYPE_SOFTWARE &&
	    (attr->config == PERF_COUNT_SW_PAGE_FAULTS ||
	     attr->config == PERF_COUNT_SW_PAGE_FAULTS_MIN ||
	     attr->config == PERF_COUNT_SW_PAGE_FAULTS_MAJ))
		return true;

	if (is_bts_event(attr))
		return true;

	return false;
}

void perf_event__preprocess_sample_addr(union perf_event *event,
					struct perf_sample *sample,
					struct thread *thread,
					struct addr_location *al)
{
	u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;

1339
	thread__find_addr_map(thread, cpumode, MAP__FUNCTION, sample->addr, al);
1340
	if (!al->map)
1341
		thread__find_addr_map(thread, cpumode, MAP__VARIABLE,
1342 1343 1344 1345 1346 1347 1348 1349
				      sample->addr, al);

	al->cpu = sample->cpu;
	al->sym = NULL;

	if (al->map)
		al->sym = map__find_symbol(al->map, al->addr, NULL);
}