event.c 34.2 KB
Newer Older
1
#include <linux/types.h>
2
#include <sys/mman.h>
3 4
#include "event.h"
#include "debug.h"
5
#include "hist.h"
6
#include "machine.h"
7
#include "sort.h"
8
#include "string.h"
9
#include "strlist.h"
10
#include "thread.h"
11
#include "thread_map.h"
12
#include "symbol/kallsyms.h"
13 14
#include "asm/bug.h"
#include "stat.h"
15

16
static const char *perf_event__names[] = {
17 18
	[0]					= "TOTAL",
	[PERF_RECORD_MMAP]			= "MMAP",
19
	[PERF_RECORD_MMAP2]			= "MMAP2",
20 21 22 23 24 25 26 27
	[PERF_RECORD_LOST]			= "LOST",
	[PERF_RECORD_COMM]			= "COMM",
	[PERF_RECORD_EXIT]			= "EXIT",
	[PERF_RECORD_THROTTLE]			= "THROTTLE",
	[PERF_RECORD_UNTHROTTLE]		= "UNTHROTTLE",
	[PERF_RECORD_FORK]			= "FORK",
	[PERF_RECORD_READ]			= "READ",
	[PERF_RECORD_SAMPLE]			= "SAMPLE",
28
	[PERF_RECORD_AUX]			= "AUX",
29
	[PERF_RECORD_ITRACE_START]		= "ITRACE_START",
30
	[PERF_RECORD_LOST_SAMPLES]		= "LOST_SAMPLES",
31 32
	[PERF_RECORD_SWITCH]			= "SWITCH",
	[PERF_RECORD_SWITCH_CPU_WIDE]		= "SWITCH_CPU_WIDE",
33 34 35 36 37
	[PERF_RECORD_HEADER_ATTR]		= "ATTR",
	[PERF_RECORD_HEADER_EVENT_TYPE]		= "EVENT_TYPE",
	[PERF_RECORD_HEADER_TRACING_DATA]	= "TRACING_DATA",
	[PERF_RECORD_HEADER_BUILD_ID]		= "BUILD_ID",
	[PERF_RECORD_FINISHED_ROUND]		= "FINISHED_ROUND",
A
Adrian Hunter 已提交
38
	[PERF_RECORD_ID_INDEX]			= "ID_INDEX",
39 40
	[PERF_RECORD_AUXTRACE_INFO]		= "AUXTRACE_INFO",
	[PERF_RECORD_AUXTRACE]			= "AUXTRACE",
41
	[PERF_RECORD_AUXTRACE_ERROR]		= "AUXTRACE_ERROR",
42
	[PERF_RECORD_THREAD_MAP]		= "THREAD_MAP",
43
	[PERF_RECORD_CPU_MAP]			= "CPU_MAP",
44
	[PERF_RECORD_STAT_CONFIG]		= "STAT_CONFIG",
45 46
};

47
const char *perf_event__name(unsigned int id)
48
{
49
	if (id >= ARRAY_SIZE(perf_event__names))
50
		return "INVALID";
51
	if (!perf_event__names[id])
52
		return "UNKNOWN";
53
	return perf_event__names[id];
54 55
}

56
static struct perf_sample synth_sample = {
57 58 59 60 61 62 63 64
	.pid	   = -1,
	.tid	   = -1,
	.time	   = -1,
	.stream_id = -1,
	.cpu	   = -1,
	.period	   = 1,
};

65 66
/*
 * Assumes that the first 4095 bytes of /proc/pid/stat contains
67
 * the comm, tgid and ppid.
68
 */
69 70
static int perf_event__get_comm_ids(pid_t pid, char *comm, size_t len,
				    pid_t *tgid, pid_t *ppid)
71 72
{
	char filename[PATH_MAX];
73 74
	char bf[4096];
	int fd;
75 76
	size_t size = 0;
	ssize_t n;
77 78 79 80
	char *nl, *name, *tgids, *ppids;

	*tgid = -1;
	*ppid = -1;
81 82 83

	snprintf(filename, sizeof(filename), "/proc/%d/status", pid);

84 85
	fd = open(filename, O_RDONLY);
	if (fd < 0) {
86
		pr_debug("couldn't open %s\n", filename);
87
		return -1;
88 89
	}

90 91 92
	n = read(fd, bf, sizeof(bf) - 1);
	close(fd);
	if (n <= 0) {
93
		pr_warning("Couldn't get COMM, tigd and ppid for pid %d\n",
94 95
			   pid);
		return -1;
96
	}
97
	bf[n] = '\0';
98

99 100
	name = strstr(bf, "Name:");
	tgids = strstr(bf, "Tgid:");
101
	ppids = strstr(bf, "PPid:");
102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123

	if (name) {
		name += 5;  /* strlen("Name:") */

		while (*name && isspace(*name))
			++name;

		nl = strchr(name, '\n');
		if (nl)
			*nl = '\0';

		size = strlen(name);
		if (size >= len)
			size = len - 1;
		memcpy(comm, name, size);
		comm[size] = '\0';
	} else {
		pr_debug("Name: string not found for pid %d\n", pid);
	}

	if (tgids) {
		tgids += 5;  /* strlen("Tgid:") */
124
		*tgid = atoi(tgids);
125 126 127
	} else {
		pr_debug("Tgid: string not found for pid %d\n", pid);
	}
128

129 130 131 132 133 134 135 136
	if (ppids) {
		ppids += 5;  /* strlen("PPid:") */
		*ppid = atoi(ppids);
	} else {
		pr_debug("PPid: string not found for pid %d\n", pid);
	}

	return 0;
137 138
}

139 140 141
static int perf_event__prepare_comm(union perf_event *event, pid_t pid,
				    struct machine *machine,
				    pid_t *tgid, pid_t *ppid)
142 143
{
	size_t size;
144 145

	*ppid = -1;
146 147 148

	memset(&event->comm, 0, sizeof(event->comm));

149 150 151 152 153 154 155 156 157
	if (machine__is_host(machine)) {
		if (perf_event__get_comm_ids(pid, event->comm.comm,
					     sizeof(event->comm.comm),
					     tgid, ppid) != 0) {
			return -1;
		}
	} else {
		*tgid = machine->pid;
	}
158

159 160
	if (*tgid < 0)
		return -1;
161

162
	event->comm.pid = *tgid;
163
	event->comm.header.type = PERF_RECORD_COMM;
164 165

	size = strlen(event->comm.comm) + 1;
166
	size = PERF_ALIGN(size, sizeof(u64));
167
	memset(event->comm.comm + size, 0, machine->id_hdr_size);
168 169
	event->comm.header.size = (sizeof(event->comm) -
				(sizeof(event->comm.comm) - size) +
170
				machine->id_hdr_size);
171
	event->comm.tid = pid;
172 173

	return 0;
174 175
}

176
pid_t perf_event__synthesize_comm(struct perf_tool *tool,
177 178 179 180
					 union perf_event *event, pid_t pid,
					 perf_event__handler_t process,
					 struct machine *machine)
{
181
	pid_t tgid, ppid;
182

183 184
	if (perf_event__prepare_comm(event, pid, machine, &tgid, &ppid) != 0)
		return -1;
185

186 187
	if (process(tool, event, &synth_sample, machine) != 0)
		return -1;
188

189
	return tgid;
190 191
}

192
static int perf_event__synthesize_fork(struct perf_tool *tool,
193 194 195
				       union perf_event *event,
				       pid_t pid, pid_t tgid, pid_t ppid,
				       perf_event__handler_t process,
196 197 198 199
				       struct machine *machine)
{
	memset(&event->fork, 0, sizeof(event->fork) + machine->id_hdr_size);

200 201 202 203 204 205 206 207 208 209 210 211
	/*
	 * for main thread set parent to ppid from status file. For other
	 * threads set parent pid to main thread. ie., assume main thread
	 * spawns all threads in a process
	*/
	if (tgid == pid) {
		event->fork.ppid = ppid;
		event->fork.ptid = ppid;
	} else {
		event->fork.ppid = tgid;
		event->fork.ptid = tgid;
	}
212 213 214 215 216 217 218 219 220 221 222 223
	event->fork.pid  = tgid;
	event->fork.tid  = pid;
	event->fork.header.type = PERF_RECORD_FORK;

	event->fork.header.size = (sizeof(event->fork) + machine->id_hdr_size);

	if (process(tool, event, &synth_sample, machine) != 0)
		return -1;

	return 0;
}

224 225 226 227 228
int perf_event__synthesize_mmap_events(struct perf_tool *tool,
				       union perf_event *event,
				       pid_t pid, pid_t tgid,
				       perf_event__handler_t process,
				       struct machine *machine,
229 230
				       bool mmap_data,
				       unsigned int proc_map_timeout)
231 232 233
{
	char filename[PATH_MAX];
	FILE *fp;
234 235
	unsigned long long t;
	bool truncation = false;
236
	unsigned long long timeout = proc_map_timeout * 1000000ULL;
237
	int rc = 0;
238

239 240 241
	if (machine__is_default_guest(machine))
		return 0;

242 243
	snprintf(filename, sizeof(filename), "%s/proc/%d/maps",
		 machine->root_dir, pid);
244 245 246 247 248 249 250 251 252 253

	fp = fopen(filename, "r");
	if (fp == NULL) {
		/*
		 * We raced with a task exiting - just return:
		 */
		pr_debug("couldn't open %s\n", filename);
		return -1;
	}

254
	event->header.type = PERF_RECORD_MMAP2;
255
	t = rdclock();
256

257
	while (1) {
258 259 260 261
		char bf[BUFSIZ];
		char prot[5];
		char execname[PATH_MAX];
		char anonstr[] = "//anon";
262
		unsigned int ino;
263
		size_t size;
264
		ssize_t n;
265

266 267 268
		if (fgets(bf, sizeof(bf), fp) == NULL)
			break;

269 270 271 272 273
		if ((rdclock() - t) > timeout) {
			pr_warning("Reading %s time out. "
				   "You may want to increase "
				   "the time limit by --proc-map-timeout\n",
				   filename);
274 275 276 277
			truncation = true;
			goto out;
		}

278 279 280
		/* ensure null termination since stack will be reused. */
		strcpy(execname, "");

281
		/* 00400000-0040c000 r-xp 00000000 fd:01 41038  /bin/cat */
282 283 284 285 286 287
		n = sscanf(bf, "%"PRIx64"-%"PRIx64" %s %"PRIx64" %x:%x %u %s\n",
		       &event->mmap2.start, &event->mmap2.len, prot,
		       &event->mmap2.pgoff, &event->mmap2.maj,
		       &event->mmap2.min,
		       &ino, execname);

288 289 290
		/*
 		 * Anon maps don't have the execname.
 		 */
291
		if (n < 7)
292
			continue;
293 294 295

		event->mmap2.ino = (u64)ino;

296 297 298
		/*
		 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
		 */
299 300 301 302
		if (machine__is_host(machine))
			event->header.misc = PERF_RECORD_MISC_USER;
		else
			event->header.misc = PERF_RECORD_MISC_GUEST_USER;
303

304 305 306 307 308 309 310 311 312 313 314 315 316 317 318
		/* map protection and flags bits */
		event->mmap2.prot = 0;
		event->mmap2.flags = 0;
		if (prot[0] == 'r')
			event->mmap2.prot |= PROT_READ;
		if (prot[1] == 'w')
			event->mmap2.prot |= PROT_WRITE;
		if (prot[2] == 'x')
			event->mmap2.prot |= PROT_EXEC;

		if (prot[3] == 's')
			event->mmap2.flags |= MAP_SHARED;
		else
			event->mmap2.flags |= MAP_PRIVATE;

319 320 321 322 323 324
		if (prot[2] != 'x') {
			if (!mmap_data || prot[0] != 'r')
				continue;

			event->header.misc |= PERF_RECORD_MISC_MMAP_DATA;
		}
325

326 327 328 329
out:
		if (truncation)
			event->header.misc |= PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT;

330 331 332 333
		if (!strcmp(execname, ""))
			strcpy(execname, anonstr);

		size = strlen(execname) + 1;
334
		memcpy(event->mmap2.filename, execname, size);
335
		size = PERF_ALIGN(size, sizeof(u64));
336 337 338 339 340 341 342
		event->mmap2.len -= event->mmap.start;
		event->mmap2.header.size = (sizeof(event->mmap2) -
					(sizeof(event->mmap2.filename) - size));
		memset(event->mmap2.filename + size, 0, machine->id_hdr_size);
		event->mmap2.header.size += machine->id_hdr_size;
		event->mmap2.pid = tgid;
		event->mmap2.tid = pid;
343 344 345 346

		if (process(tool, event, &synth_sample, machine) != 0) {
			rc = -1;
			break;
347
		}
348 349 350

		if (truncation)
			break;
351 352 353
	}

	fclose(fp);
354
	return rc;
355 356
}

357
int perf_event__synthesize_modules(struct perf_tool *tool,
358
				   perf_event__handler_t process,
359
				   struct machine *machine)
360
{
361
	int rc = 0;
362
	struct map *pos;
363
	struct map_groups *kmaps = &machine->kmaps;
364
	struct maps *maps = &kmaps->maps[MAP__FUNCTION];
365
	union perf_event *event = zalloc((sizeof(event->mmap) +
366
					  machine->id_hdr_size));
367 368 369 370 371 372 373
	if (event == NULL) {
		pr_debug("Not enough memory synthesizing mmap event "
			 "for kernel modules\n");
		return -1;
	}

	event->header.type = PERF_RECORD_MMAP;
374

375 376 377 378
	/*
	 * kernel uses 0 for user space maps, see kernel/perf_event.c
	 * __perf_event_mmap
	 */
379
	if (machine__is_host(machine))
380
		event->header.misc = PERF_RECORD_MISC_KERNEL;
381
	else
382
		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
383

384
	for (pos = maps__first(maps); pos; pos = map__next(pos)) {
385 386
		size_t size;

387
		if (__map__is_kernel(pos))
388 389
			continue;

390
		size = PERF_ALIGN(pos->dso->long_name_len + 1, sizeof(u64));
391 392 393
		event->mmap.header.type = PERF_RECORD_MMAP;
		event->mmap.header.size = (sizeof(event->mmap) -
				        (sizeof(event->mmap.filename) - size));
394 395
		memset(event->mmap.filename + size, 0, machine->id_hdr_size);
		event->mmap.header.size += machine->id_hdr_size;
396 397 398 399 400
		event->mmap.start = pos->start;
		event->mmap.len   = pos->end - pos->start;
		event->mmap.pid   = machine->pid;

		memcpy(event->mmap.filename, pos->dso->long_name,
401
		       pos->dso->long_name_len + 1);
402 403 404 405
		if (process(tool, event, &synth_sample, machine) != 0) {
			rc = -1;
			break;
		}
406 407
	}

408
	free(event);
409
	return rc;
410 411
}

412 413
static int __event__synthesize_thread(union perf_event *comm_event,
				      union perf_event *mmap_event,
414
				      union perf_event *fork_event,
415 416
				      pid_t pid, int full,
					  perf_event__handler_t process,
417
				      struct perf_tool *tool,
418 419 420
				      struct machine *machine,
				      bool mmap_data,
				      unsigned int proc_map_timeout)
421
{
422 423 424
	char filename[PATH_MAX];
	DIR *tasks;
	struct dirent dirent, *next;
425
	pid_t tgid, ppid;
426
	int rc = 0;
427 428 429 430 431 432 433 434 435 436

	/* special case: only send one comm event using passed in pid */
	if (!full) {
		tgid = perf_event__synthesize_comm(tool, comm_event, pid,
						   process, machine);

		if (tgid == -1)
			return -1;

		return perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
437 438
							  process, machine, mmap_data,
							  proc_map_timeout);
439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460
	}

	if (machine__is_default_guest(machine))
		return 0;

	snprintf(filename, sizeof(filename), "%s/proc/%d/task",
		 machine->root_dir, pid);

	tasks = opendir(filename);
	if (tasks == NULL) {
		pr_debug("couldn't open %s\n", filename);
		return 0;
	}

	while (!readdir_r(tasks, &dirent, &next) && next) {
		char *end;
		pid_t _pid;

		_pid = strtol(dirent.d_name, &end, 10);
		if (*end)
			continue;

461
		rc = -1;
462 463
		if (perf_event__prepare_comm(comm_event, _pid, machine,
					     &tgid, &ppid) != 0)
464
			break;
465

466
		if (perf_event__synthesize_fork(tool, fork_event, _pid, tgid,
467
						ppid, process, machine) < 0)
468
			break;
469 470 471 472
		/*
		 * Send the prepared comm event
		 */
		if (process(tool, comm_event, &synth_sample, machine) != 0)
473
			break;
474

475
		rc = 0;
476 477 478
		if (_pid == pid) {
			/* process the parent's maps too */
			rc = perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
479
						process, machine, mmap_data, proc_map_timeout);
480 481
			if (rc)
				break;
482
		}
483 484 485
	}

	closedir(tasks);
486
	return rc;
487 488
}

489
int perf_event__synthesize_thread_map(struct perf_tool *tool,
490
				      struct thread_map *threads,
491
				      perf_event__handler_t process,
492
				      struct machine *machine,
493 494
				      bool mmap_data,
				      unsigned int proc_map_timeout)
495
{
496
	union perf_event *comm_event, *mmap_event, *fork_event;
497
	int err = -1, thread, j;
498

499
	comm_event = malloc(sizeof(comm_event->comm) + machine->id_hdr_size);
500 501 502
	if (comm_event == NULL)
		goto out;

503
	mmap_event = malloc(sizeof(mmap_event->mmap) + machine->id_hdr_size);
504 505 506
	if (mmap_event == NULL)
		goto out_free_comm;

507 508 509 510
	fork_event = malloc(sizeof(fork_event->fork) + machine->id_hdr_size);
	if (fork_event == NULL)
		goto out_free_mmap;

511 512 513
	err = 0;
	for (thread = 0; thread < threads->nr; ++thread) {
		if (__event__synthesize_thread(comm_event, mmap_event,
514
					       fork_event,
515
					       thread_map__pid(threads, thread), 0,
516
					       process, tool, machine,
517
					       mmap_data, proc_map_timeout)) {
518 519 520
			err = -1;
			break;
		}
521 522 523 524 525

		/*
		 * comm.pid is set to thread group id by
		 * perf_event__synthesize_comm
		 */
526
		if ((int) comm_event->comm.pid != thread_map__pid(threads, thread)) {
527 528 529 530
			bool need_leader = true;

			/* is thread group leader in thread_map? */
			for (j = 0; j < threads->nr; ++j) {
531
				if ((int) comm_event->comm.pid == thread_map__pid(threads, j)) {
532 533 534 535 536 537 538
					need_leader = false;
					break;
				}
			}

			/* if not, generate events for it */
			if (need_leader &&
539
			    __event__synthesize_thread(comm_event, mmap_event,
540
						       fork_event,
541 542
						       comm_event->comm.pid, 0,
						       process, tool, machine,
543
						       mmap_data, proc_map_timeout)) {
544 545 546 547
				err = -1;
				break;
			}
		}
548
	}
549 550
	free(fork_event);
out_free_mmap:
551 552 553 554 555 556 557
	free(mmap_event);
out_free_comm:
	free(comm_event);
out:
	return err;
}

558
int perf_event__synthesize_threads(struct perf_tool *tool,
559
				   perf_event__handler_t process,
560 561 562
				   struct machine *machine,
				   bool mmap_data,
				   unsigned int proc_map_timeout)
563 564
{
	DIR *proc;
565
	char proc_path[PATH_MAX];
566
	struct dirent dirent, *next;
567
	union perf_event *comm_event, *mmap_event, *fork_event;
568 569
	int err = -1;

570 571 572
	if (machine__is_default_guest(machine))
		return 0;

573
	comm_event = malloc(sizeof(comm_event->comm) + machine->id_hdr_size);
574 575 576
	if (comm_event == NULL)
		goto out;

577
	mmap_event = malloc(sizeof(mmap_event->mmap) + machine->id_hdr_size);
578 579
	if (mmap_event == NULL)
		goto out_free_comm;
580

581 582 583 584
	fork_event = malloc(sizeof(fork_event->fork) + machine->id_hdr_size);
	if (fork_event == NULL)
		goto out_free_mmap;

585 586 587
	snprintf(proc_path, sizeof(proc_path), "%s/proc", machine->root_dir);
	proc = opendir(proc_path);

588
	if (proc == NULL)
589
		goto out_free_fork;
590 591 592 593 594 595 596

	while (!readdir_r(proc, &dirent, &next) && next) {
		char *end;
		pid_t pid = strtol(dirent.d_name, &end, 10);

		if (*end) /* only interested in proper numerical dirents */
			continue;
597 598 599 600
		/*
 		 * We may race with exiting thread, so don't stop just because
 		 * one thread couldn't be synthesized.
 		 */
601
		__event__synthesize_thread(comm_event, mmap_event, fork_event, pid,
602 603
					   1, process, tool, machine, mmap_data,
					   proc_map_timeout);
604 605
	}

606
	err = 0;
607
	closedir(proc);
608 609
out_free_fork:
	free(fork_event);
610 611 612 613 614 615
out_free_mmap:
	free(mmap_event);
out_free_comm:
	free(comm_event);
out:
	return err;
616
}
617

618 619 620 621 622
struct process_symbol_args {
	const char *name;
	u64	   start;
};

623
static int find_symbol_cb(void *arg, const char *name, char type,
624
			  u64 start)
625 626 627
{
	struct process_symbol_args *args = arg;

628 629 630 631 632 633
	/*
	 * Must be a function or at least an alias, as in PARISC64, where "_text" is
	 * an 'A' to the same address as "_stext".
	 */
	if (!(symbol_type__is_a(type, MAP__FUNCTION) ||
	      type == 'A') || strcmp(name, args->name))
634 635 636 637 638 639
		return 0;

	args->start = start;
	return 1;
}

640 641 642 643 644 645 646 647 648 649 650
u64 kallsyms__get_function_start(const char *kallsyms_filename,
				 const char *symbol_name)
{
	struct process_symbol_args args = { .name = symbol_name, };

	if (kallsyms__parse(kallsyms_filename, &args, find_symbol_cb) <= 0)
		return 0;

	return args.start;
}

651
int perf_event__synthesize_kernel_mmap(struct perf_tool *tool,
652
				       perf_event__handler_t process,
653
				       struct machine *machine)
654 655
{
	size_t size;
656
	const char *mmap_name;
657
	char name_buff[PATH_MAX];
658
	struct map *map = machine__kernel_map(machine);
659
	struct kmap *kmap;
660
	int err;
661 662
	union perf_event *event;

663
	if (map == NULL)
664 665
		return -1;

666 667 668 669 670
	/*
	 * We should get this from /sys/kernel/sections/.text, but till that is
	 * available use this, and after it is use this as a fallback for older
	 * kernels.
	 */
671
	event = zalloc((sizeof(event->mmap) + machine->id_hdr_size));
672 673 674 675 676
	if (event == NULL) {
		pr_debug("Not enough memory synthesizing mmap event "
			 "for kernel modules\n");
		return -1;
	}
677

678
	mmap_name = machine__mmap_name(machine, name_buff, sizeof(name_buff));
679
	if (machine__is_host(machine)) {
680 681 682 683
		/*
		 * kernel uses PERF_RECORD_MISC_USER for user space maps,
		 * see kernel/perf_event.c __perf_event_mmap
		 */
684
		event->header.misc = PERF_RECORD_MISC_KERNEL;
685
	} else {
686
		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
687
	}
688

689
	kmap = map__kmap(map);
690
	size = snprintf(event->mmap.filename, sizeof(event->mmap.filename),
691
			"%s%s", mmap_name, kmap->ref_reloc_sym->name) + 1;
692
	size = PERF_ALIGN(size, sizeof(u64));
693 694
	event->mmap.header.type = PERF_RECORD_MMAP;
	event->mmap.header.size = (sizeof(event->mmap) -
695
			(sizeof(event->mmap.filename) - size) + machine->id_hdr_size);
696
	event->mmap.pgoff = kmap->ref_reloc_sym->addr;
697 698 699 700
	event->mmap.start = map->start;
	event->mmap.len   = map->end - event->mmap.start;
	event->mmap.pid   = machine->pid;

701
	err = process(tool, event, &synth_sample, machine);
702 703 704
	free(event);

	return err;
705 706
}

707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742
int perf_event__synthesize_thread_map2(struct perf_tool *tool,
				      struct thread_map *threads,
				      perf_event__handler_t process,
				      struct machine *machine)
{
	union perf_event *event;
	int i, err, size;

	size  = sizeof(event->thread_map);
	size +=	threads->nr * sizeof(event->thread_map.entries[0]);

	event = zalloc(size);
	if (!event)
		return -ENOMEM;

	event->header.type = PERF_RECORD_THREAD_MAP;
	event->header.size = size;
	event->thread_map.nr = threads->nr;

	for (i = 0; i < threads->nr; i++) {
		struct thread_map_event_entry *entry = &event->thread_map.entries[i];
		char *comm = thread_map__comm(threads, i);

		if (!comm)
			comm = (char *) "";

		entry->pid = thread_map__pid(threads, i);
		strncpy((char *) &entry->comm, comm, sizeof(entry->comm));
	}

	err = process(tool, event, NULL, machine);

	free(event);
	return err;
}

743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873
static void synthesize_cpus(struct cpu_map_entries *cpus,
			    struct cpu_map *map)
{
	int i;

	cpus->nr = map->nr;

	for (i = 0; i < map->nr; i++)
		cpus->cpu[i] = map->map[i];
}

static void synthesize_mask(struct cpu_map_mask *mask,
			    struct cpu_map *map, int max)
{
	int i;

	mask->nr = BITS_TO_LONGS(max);
	mask->long_size = sizeof(long);

	for (i = 0; i < map->nr; i++)
		set_bit(map->map[i], mask->mask);
}

static size_t cpus_size(struct cpu_map *map)
{
	return sizeof(struct cpu_map_entries) + map->nr * sizeof(u16);
}

static size_t mask_size(struct cpu_map *map, int *max)
{
	int i;

	*max = 0;

	for (i = 0; i < map->nr; i++) {
		/* bit possition of the cpu is + 1 */
		int bit = map->map[i] + 1;

		if (bit > *max)
			*max = bit;
	}

	return sizeof(struct cpu_map_mask) + BITS_TO_LONGS(*max) * sizeof(long);
}

void *cpu_map_data__alloc(struct cpu_map *map, size_t *size, u16 *type, int *max)
{
	size_t size_cpus, size_mask;
	bool is_dummy = cpu_map__empty(map);

	/*
	 * Both array and mask data have variable size based
	 * on the number of cpus and their actual values.
	 * The size of the 'struct cpu_map_data' is:
	 *
	 *   array = size of 'struct cpu_map_entries' +
	 *           number of cpus * sizeof(u64)
	 *
	 *   mask  = size of 'struct cpu_map_mask' +
	 *           maximum cpu bit converted to size of longs
	 *
	 * and finaly + the size of 'struct cpu_map_data'.
	 */
	size_cpus = cpus_size(map);
	size_mask = mask_size(map, max);

	if (is_dummy || (size_cpus < size_mask)) {
		*size += size_cpus;
		*type  = PERF_CPU_MAP__CPUS;
	} else {
		*size += size_mask;
		*type  = PERF_CPU_MAP__MASK;
	}

	*size += sizeof(struct cpu_map_data);
	return zalloc(*size);
}

void cpu_map_data__synthesize(struct cpu_map_data *data, struct cpu_map *map,
			      u16 type, int max)
{
	data->type = type;

	switch (type) {
	case PERF_CPU_MAP__CPUS:
		synthesize_cpus((struct cpu_map_entries *) data->data, map);
		break;
	case PERF_CPU_MAP__MASK:
		synthesize_mask((struct cpu_map_mask *) data->data, map, max);
	default:
		break;
	};
}

static struct cpu_map_event* cpu_map_event__new(struct cpu_map *map)
{
	size_t size = sizeof(struct cpu_map_event);
	struct cpu_map_event *event;
	int max;
	u16 type;

	event = cpu_map_data__alloc(map, &size, &type, &max);
	if (!event)
		return NULL;

	event->header.type = PERF_RECORD_CPU_MAP;
	event->header.size = size;
	event->data.type   = type;

	cpu_map_data__synthesize(&event->data, map, type, max);
	return event;
}

int perf_event__synthesize_cpu_map(struct perf_tool *tool,
				   struct cpu_map *map,
				   perf_event__handler_t process,
				   struct machine *machine)
{
	struct cpu_map_event *event;
	int err;

	event = cpu_map_event__new(map);
	if (!event)
		return -ENOMEM;

	err = process(tool, (union perf_event *) event, NULL, machine);

	free(event);
	return err;
}

874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911
int perf_event__synthesize_stat_config(struct perf_tool *tool,
				       struct perf_stat_config *config,
				       perf_event__handler_t process,
				       struct machine *machine)
{
	struct stat_config_event *event;
	int size, i = 0, err;

	size  = sizeof(*event);
	size += (PERF_STAT_CONFIG_TERM__MAX * sizeof(event->data[0]));

	event = zalloc(size);
	if (!event)
		return -ENOMEM;

	event->header.type = PERF_RECORD_STAT_CONFIG;
	event->header.size = size;
	event->nr          = PERF_STAT_CONFIG_TERM__MAX;

#define ADD(__term, __val)					\
	event->data[i].tag = PERF_STAT_CONFIG_TERM__##__term;	\
	event->data[i].val = __val;				\
	i++;

	ADD(AGGR_MODE,	config->aggr_mode)
	ADD(INTERVAL,	config->interval)
	ADD(SCALE,	config->scale)

	WARN_ONCE(i != PERF_STAT_CONFIG_TERM__MAX,
		  "stat config terms unbalanced\n");
#undef ADD

	err = process(tool, (union perf_event *) event, NULL, machine);

	free(event);
	return err;
}

912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935
void perf_event__read_stat_config(struct perf_stat_config *config,
				  struct stat_config_event *event)
{
	unsigned i;

	for (i = 0; i < event->nr; i++) {

		switch (event->data[i].tag) {
#define CASE(__term, __val)					\
		case PERF_STAT_CONFIG_TERM__##__term:		\
			config->__val = event->data[i].val;	\
			break;

		CASE(AGGR_MODE, aggr_mode)
		CASE(SCALE,     scale)
		CASE(INTERVAL,  interval)
#undef CASE
		default:
			pr_warning("unknown stat config term %" PRIu64 "\n",
				   event->data[i].tag);
		}
	}
}

936 937
size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp)
{
938 939 940 941 942 943 944
	const char *s;

	if (event->header.misc & PERF_RECORD_MISC_COMM_EXEC)
		s = " exec";
	else
		s = "";

945
	return fprintf(fp, "%s: %s:%d/%d\n", s, event->comm.comm, event->comm.pid, event->comm.tid);
946 947
}

948
int perf_event__process_comm(struct perf_tool *tool __maybe_unused,
949
			     union perf_event *event,
950
			     struct perf_sample *sample,
951
			     struct machine *machine)
952
{
953
	return machine__process_comm_event(machine, event, sample);
954 955
}

956
int perf_event__process_lost(struct perf_tool *tool __maybe_unused,
957
			     union perf_event *event,
958
			     struct perf_sample *sample,
959
			     struct machine *machine)
960
{
961
	return machine__process_lost_event(machine, event, sample);
962
}
963

964 965 966 967 968 969 970 971
int perf_event__process_aux(struct perf_tool *tool __maybe_unused,
			    union perf_event *event,
			    struct perf_sample *sample __maybe_unused,
			    struct machine *machine)
{
	return machine__process_aux_event(machine, event);
}

972 973 974 975 976 977 978 979
int perf_event__process_itrace_start(struct perf_tool *tool __maybe_unused,
				     union perf_event *event,
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine)
{
	return machine__process_itrace_start_event(machine, event);
}

980 981 982 983 984 985 986 987
int perf_event__process_lost_samples(struct perf_tool *tool __maybe_unused,
				     union perf_event *event,
				     struct perf_sample *sample,
				     struct machine *machine)
{
	return machine__process_lost_samples_event(machine, event, sample);
}

988 989 990 991 992 993 994 995
int perf_event__process_switch(struct perf_tool *tool __maybe_unused,
			       union perf_event *event,
			       struct perf_sample *sample __maybe_unused,
			       struct machine *machine)
{
	return machine__process_switch_event(machine, event);
}

996 997
size_t perf_event__fprintf_mmap(union perf_event *event, FILE *fp)
{
998
	return fprintf(fp, " %d/%d: [%#" PRIx64 "(%#" PRIx64 ") @ %#" PRIx64 "]: %c %s\n",
999
		       event->mmap.pid, event->mmap.tid, event->mmap.start,
1000 1001 1002
		       event->mmap.len, event->mmap.pgoff,
		       (event->header.misc & PERF_RECORD_MISC_MMAP_DATA) ? 'r' : 'x',
		       event->mmap.filename);
1003 1004
}

1005 1006 1007
size_t perf_event__fprintf_mmap2(union perf_event *event, FILE *fp)
{
	return fprintf(fp, " %d/%d: [%#" PRIx64 "(%#" PRIx64 ") @ %#" PRIx64
1008
			   " %02x:%02x %"PRIu64" %"PRIu64"]: %c%c%c%c %s\n",
1009 1010 1011 1012
		       event->mmap2.pid, event->mmap2.tid, event->mmap2.start,
		       event->mmap2.len, event->mmap2.pgoff, event->mmap2.maj,
		       event->mmap2.min, event->mmap2.ino,
		       event->mmap2.ino_generation,
1013 1014 1015 1016
		       (event->mmap2.prot & PROT_READ) ? 'r' : '-',
		       (event->mmap2.prot & PROT_WRITE) ? 'w' : '-',
		       (event->mmap2.prot & PROT_EXEC) ? 'x' : '-',
		       (event->mmap2.flags & MAP_SHARED) ? 's' : 'p',
1017 1018 1019
		       event->mmap2.filename);
}

1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035
size_t perf_event__fprintf_thread_map(union perf_event *event, FILE *fp)
{
	struct thread_map *threads = thread_map__new_event(&event->thread_map);
	size_t ret;

	ret = fprintf(fp, " nr: ");

	if (threads)
		ret += thread_map__fprintf(threads, fp);
	else
		ret += fprintf(fp, "failed to get threads from event\n");

	thread_map__put(threads);
	return ret;
}

1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051
size_t perf_event__fprintf_cpu_map(union perf_event *event, FILE *fp)
{
	struct cpu_map *cpus = cpu_map__new_data(&event->cpu_map.data);
	size_t ret;

	ret = fprintf(fp, " nr: ");

	if (cpus)
		ret += cpu_map__fprintf(cpus, fp);
	else
		ret += fprintf(fp, "failed to get cpumap from event\n");

	cpu_map__put(cpus);
	return ret;
}

1052
int perf_event__process_mmap(struct perf_tool *tool __maybe_unused,
1053
			     union perf_event *event,
1054
			     struct perf_sample *sample,
1055
			     struct machine *machine)
1056
{
1057
	return machine__process_mmap_event(machine, event, sample);
1058 1059
}

1060 1061
int perf_event__process_mmap2(struct perf_tool *tool __maybe_unused,
			     union perf_event *event,
1062
			     struct perf_sample *sample,
1063 1064
			     struct machine *machine)
{
1065
	return machine__process_mmap2_event(machine, event, sample);
1066 1067
}

1068 1069 1070 1071 1072 1073 1074
size_t perf_event__fprintf_task(union perf_event *event, FILE *fp)
{
	return fprintf(fp, "(%d:%d):(%d:%d)\n",
		       event->fork.pid, event->fork.tid,
		       event->fork.ppid, event->fork.ptid);
}

1075
int perf_event__process_fork(struct perf_tool *tool __maybe_unused,
1076
			     union perf_event *event,
1077
			     struct perf_sample *sample,
1078
			     struct machine *machine)
1079
{
1080
	return machine__process_fork_event(machine, event, sample);
1081
}
1082

1083 1084
int perf_event__process_exit(struct perf_tool *tool __maybe_unused,
			     union perf_event *event,
1085
			     struct perf_sample *sample,
1086 1087
			     struct machine *machine)
{
1088
	return machine__process_exit_event(machine, event, sample);
1089 1090
}

1091 1092 1093 1094 1095 1096 1097 1098 1099
size_t perf_event__fprintf_aux(union perf_event *event, FILE *fp)
{
	return fprintf(fp, " offset: %#"PRIx64" size: %#"PRIx64" flags: %#"PRIx64" [%s%s]\n",
		       event->aux.aux_offset, event->aux.aux_size,
		       event->aux.flags,
		       event->aux.flags & PERF_AUX_FLAG_TRUNCATED ? "T" : "",
		       event->aux.flags & PERF_AUX_FLAG_OVERWRITE ? "O" : "");
}

1100 1101 1102 1103 1104 1105
size_t perf_event__fprintf_itrace_start(union perf_event *event, FILE *fp)
{
	return fprintf(fp, " pid: %u tid: %u\n",
		       event->itrace_start.pid, event->itrace_start.tid);
}

1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119
size_t perf_event__fprintf_switch(union perf_event *event, FILE *fp)
{
	bool out = event->header.misc & PERF_RECORD_MISC_SWITCH_OUT;
	const char *in_out = out ? "OUT" : "IN ";

	if (event->header.type == PERF_RECORD_SWITCH)
		return fprintf(fp, " %s\n", in_out);

	return fprintf(fp, " %s  %s pid/tid: %5u/%-5u\n",
		       in_out, out ? "next" : "prev",
		       event->context_switch.next_prev_pid,
		       event->context_switch.next_prev_tid);
}

1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135
size_t perf_event__fprintf(union perf_event *event, FILE *fp)
{
	size_t ret = fprintf(fp, "PERF_RECORD_%s",
			     perf_event__name(event->header.type));

	switch (event->header.type) {
	case PERF_RECORD_COMM:
		ret += perf_event__fprintf_comm(event, fp);
		break;
	case PERF_RECORD_FORK:
	case PERF_RECORD_EXIT:
		ret += perf_event__fprintf_task(event, fp);
		break;
	case PERF_RECORD_MMAP:
		ret += perf_event__fprintf_mmap(event, fp);
		break;
1136 1137 1138
	case PERF_RECORD_MMAP2:
		ret += perf_event__fprintf_mmap2(event, fp);
		break;
1139 1140 1141
	case PERF_RECORD_AUX:
		ret += perf_event__fprintf_aux(event, fp);
		break;
1142 1143 1144
	case PERF_RECORD_ITRACE_START:
		ret += perf_event__fprintf_itrace_start(event, fp);
		break;
1145 1146 1147 1148
	case PERF_RECORD_SWITCH:
	case PERF_RECORD_SWITCH_CPU_WIDE:
		ret += perf_event__fprintf_switch(event, fp);
		break;
1149 1150 1151 1152 1153 1154 1155
	default:
		ret += fprintf(fp, "\n");
	}

	return ret;
}

1156 1157
int perf_event__process(struct perf_tool *tool __maybe_unused,
			union perf_event *event,
1158
			struct perf_sample *sample,
1159
			struct machine *machine)
1160
{
1161
	return machine__process_event(machine, event, sample);
1162 1163
}

1164
void thread__find_addr_map(struct thread *thread, u8 cpumode,
1165
			   enum map_type type, u64 addr,
1166
			   struct addr_location *al)
1167
{
1168
	struct map_groups *mg = thread->mg;
1169
	struct machine *machine = mg->machine;
1170
	bool load_map = false;
1171

1172
	al->machine = machine;
1173
	al->thread = thread;
1174
	al->addr = addr;
1175
	al->cpumode = cpumode;
1176
	al->filtered = 0;
1177

1178 1179 1180 1181 1182
	if (machine == NULL) {
		al->map = NULL;
		return;
	}

1183
	if (cpumode == PERF_RECORD_MISC_KERNEL && perf_host) {
1184
		al->level = 'k';
1185
		mg = &machine->kmaps;
1186
		load_map = true;
1187
	} else if (cpumode == PERF_RECORD_MISC_USER && perf_host) {
1188
		al->level = '.';
1189 1190
	} else if (cpumode == PERF_RECORD_MISC_GUEST_KERNEL && perf_guest) {
		al->level = 'g';
1191
		mg = &machine->kmaps;
1192
		load_map = true;
1193 1194
	} else if (cpumode == PERF_RECORD_MISC_GUEST_USER && perf_guest) {
		al->level = 'u';
1195
	} else {
1196
		al->level = 'H';
1197
		al->map = NULL;
1198 1199 1200 1201

		if ((cpumode == PERF_RECORD_MISC_GUEST_USER ||
			cpumode == PERF_RECORD_MISC_GUEST_KERNEL) &&
			!perf_guest)
1202
			al->filtered |= (1 << HIST_FILTER__GUEST);
1203 1204 1205
		if ((cpumode == PERF_RECORD_MISC_USER ||
			cpumode == PERF_RECORD_MISC_KERNEL) &&
			!perf_host)
1206
			al->filtered |= (1 << HIST_FILTER__HOST);
1207

1208 1209 1210
		return;
	}
try_again:
1211
	al->map = map_groups__find(mg, type, al->addr);
1212 1213 1214 1215 1216 1217 1218 1219 1220 1221
	if (al->map == NULL) {
		/*
		 * If this is outside of all known maps, and is a negative
		 * address, try to look it up in the kernel dso, as it might be
		 * a vsyscall or vdso (which executes in user-mode).
		 *
		 * XXX This is nasty, we should have a symbol list in the
		 * "[vdso]" dso, but for now lets use the old trick of looking
		 * in the whole kernel symbol list.
		 */
1222 1223 1224
		if (cpumode == PERF_RECORD_MISC_USER && machine &&
		    mg != &machine->kmaps &&
		    machine__kernel_ip(machine, al->addr)) {
1225
			mg = &machine->kmaps;
1226
			load_map = true;
1227 1228
			goto try_again;
		}
1229 1230 1231 1232 1233 1234
	} else {
		/*
		 * Kernel maps might be changed when loading symbols so loading
		 * must be done prior to using kernel maps.
		 */
		if (load_map)
1235
			map__load(al->map, machine->symbol_filter);
1236
		al->addr = al->map->map_ip(al->map, al->addr);
1237
	}
1238 1239
}

1240
void thread__find_addr_location(struct thread *thread,
1241
				u8 cpumode, enum map_type type, u64 addr,
1242
				struct addr_location *al)
1243
{
1244
	thread__find_addr_map(thread, cpumode, type, addr, al);
1245
	if (al->map != NULL)
1246
		al->sym = map__find_symbol(al->map, al->addr,
1247
					   thread->mg->machine->symbol_filter);
1248 1249
	else
		al->sym = NULL;
1250 1251
}

1252 1253 1254 1255
/*
 * Callers need to drop the reference to al->thread, obtained in
 * machine__findnew_thread()
 */
1256
int perf_event__preprocess_sample(const union perf_event *event,
1257
				  struct machine *machine,
1258
				  struct addr_location *al,
1259
				  struct perf_sample *sample)
1260
{
1261
	u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
1262
	struct thread *thread = machine__findnew_thread(machine, sample->pid,
1263
							sample->tid);
1264

1265 1266 1267
	if (thread == NULL)
		return -1;

1268
	dump_printf(" ... thread: %s:%d\n", thread__comm_str(thread), thread->tid);
1269
	/*
1270
	 * Have we already created the kernel maps for this machine?
1271 1272 1273 1274 1275 1276
	 *
	 * This should have happened earlier, when we processed the kernel MMAP
	 * events, but for older perf.data files there was no such thing, so do
	 * it now.
	 */
	if (cpumode == PERF_RECORD_MISC_KERNEL &&
1277
	    machine__kernel_map(machine) == NULL)
1278
		machine__create_kernel_maps(machine);
1279

1280
	thread__find_addr_map(thread, cpumode, MAP__FUNCTION, sample->ip, al);
1281 1282 1283
	dump_printf(" ...... dso: %s\n",
		    al->map ? al->map->dso->long_name :
			al->level == 'H' ? "[hypervisor]" : "<not found>");
1284 1285 1286 1287

	if (thread__is_filtered(thread))
		al->filtered |= (1 << HIST_FILTER__THREAD);

1288
	al->sym = NULL;
1289
	al->cpu = sample->cpu;
1290 1291 1292 1293 1294 1295 1296 1297
	al->socket = -1;

	if (al->cpu >= 0) {
		struct perf_env *env = machine->env;

		if (env && env->cpu)
			al->socket = env->cpu[al->cpu].socket_id;
	}
1298 1299

	if (al->map) {
1300 1301
		struct dso *dso = al->map->dso;

1302
		if (symbol_conf.dso_list &&
1303 1304 1305 1306
		    (!dso || !(strlist__has_entry(symbol_conf.dso_list,
						  dso->short_name) ||
			       (dso->short_name != dso->long_name &&
				strlist__has_entry(symbol_conf.dso_list,
1307 1308 1309
						   dso->long_name))))) {
			al->filtered |= (1 << HIST_FILTER__DSO);
		}
1310

1311 1312
		al->sym = map__find_symbol(al->map, al->addr,
					   machine->symbol_filter);
1313
	}
1314

1315 1316
	if (symbol_conf.sym_list &&
		(!al->sym || !strlist__has_entry(symbol_conf.sym_list,
1317 1318 1319
						al->sym->name))) {
		al->filtered |= (1 << HIST_FILTER__SYMBOL);
	}
1320 1321

	return 0;
1322
}
1323

1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334
/*
 * The preprocess_sample method will return with reference counts for the
 * in it, when done using (and perhaps getting ref counts if needing to
 * keep a pointer to one of those entries) it must be paired with
 * addr_location__put(), so that the refcounts can be decremented.
 */
void addr_location__put(struct addr_location *al)
{
	thread__zput(al->thread);
}

1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362
bool is_bts_event(struct perf_event_attr *attr)
{
	return attr->type == PERF_TYPE_HARDWARE &&
	       (attr->config & PERF_COUNT_HW_BRANCH_INSTRUCTIONS) &&
	       attr->sample_period == 1;
}

bool sample_addr_correlates_sym(struct perf_event_attr *attr)
{
	if (attr->type == PERF_TYPE_SOFTWARE &&
	    (attr->config == PERF_COUNT_SW_PAGE_FAULTS ||
	     attr->config == PERF_COUNT_SW_PAGE_FAULTS_MIN ||
	     attr->config == PERF_COUNT_SW_PAGE_FAULTS_MAJ))
		return true;

	if (is_bts_event(attr))
		return true;

	return false;
}

void perf_event__preprocess_sample_addr(union perf_event *event,
					struct perf_sample *sample,
					struct thread *thread,
					struct addr_location *al)
{
	u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;

1363
	thread__find_addr_map(thread, cpumode, MAP__FUNCTION, sample->addr, al);
1364
	if (!al->map)
1365
		thread__find_addr_map(thread, cpumode, MAP__VARIABLE,
1366 1367 1368 1369 1370 1371 1372 1373
				      sample->addr, al);

	al->cpu = sample->cpu;
	al->sym = NULL;

	if (al->map)
		al->sym = map__find_symbol(al->map, al->addr, NULL);
}