session.c 42.7 KB
Newer Older
1 2
#define _FILE_OFFSET_BITS 64

3 4
#include <linux/kernel.h>

5
#include <byteswap.h>
6 7
#include <unistd.h>
#include <sys/types.h>
8
#include <sys/mman.h>
9

10 11
#include "evlist.h"
#include "evsel.h"
12
#include "session.h"
13
#include "tool.h"
14
#include "sort.h"
15
#include "util.h"
16
#include "cpumap.h"
17
#include "event-parse.h"
18

19 20 21 22 23 24 25 26 27 28 29 30 31
int perf_session__parse_sample(struct perf_session *session,
			       const union perf_event *event,
			       struct perf_sample *sample)
{
	struct perf_evsel *first;
	first = list_entry(session->evlist->entries.next, struct perf_evsel, node);

	return perf_event__parse_sample(event, session->sample_type,
					first->sample_size,
					session->sample_id_all, sample,
					session->header.needs_swap);
}

32 33 34 35
static int perf_session__open(struct perf_session *self, bool force)
{
	struct stat input_stat;

36 37 38 39
	if (!strcmp(self->filename, "-")) {
		self->fd_pipe = true;
		self->fd = STDIN_FILENO;

40
		if (perf_session__read_header(self, self->fd) < 0)
41
			pr_err("incompatible file format (rerun with -v to learn more)");
42 43 44 45

		return 0;
	}

46
	self->fd = open(self->filename, O_RDONLY);
47
	if (self->fd < 0) {
48 49 50 51
		int err = errno;

		pr_err("failed to open %s: %s", self->filename, strerror(err));
		if (err == ENOENT && !strcmp(self->filename, "perf.data"))
52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71
			pr_err("  (try 'perf record' first)");
		pr_err("\n");
		return -errno;
	}

	if (fstat(self->fd, &input_stat) < 0)
		goto out_close;

	if (!force && input_stat.st_uid && (input_stat.st_uid != geteuid())) {
		pr_err("file %s not owned by current user or root\n",
		       self->filename);
		goto out_close;
	}

	if (!input_stat.st_size) {
		pr_info("zero-sized file (%s), nothing to do!\n",
			self->filename);
		goto out_close;
	}

72
	if (perf_session__read_header(self, self->fd) < 0) {
73
		pr_err("incompatible file format (rerun with -v to learn more)");
74 75 76
		goto out_close;
	}

77 78 79 80 81 82 83 84 85 86
	if (!perf_evlist__valid_sample_type(self->evlist)) {
		pr_err("non matching sample_type");
		goto out_close;
	}

	if (!perf_evlist__valid_sample_id_all(self->evlist)) {
		pr_err("non matching sample_id_all");
		goto out_close;
	}

87 88 89 90 91 92 93 94 95
	self->size = input_stat.st_size;
	return 0;

out_close:
	close(self->fd);
	self->fd = -1;
	return -1;
}

96 97
void perf_session__update_sample_type(struct perf_session *self)
{
98 99
	self->sample_type = perf_evlist__sample_type(self->evlist);
	self->sample_id_all = perf_evlist__sample_id_all(self->evlist);
100
	self->id_hdr_size = perf_evlist__id_hdr_size(self->evlist);
101
	self->host_machine.id_hdr_size = self->id_hdr_size;
102
	machines__set_id_hdr_size(&self->machines, self->id_hdr_size);
103 104
}

105 106
int perf_session__create_kernel_maps(struct perf_session *self)
{
107
	int ret = machine__create_kernel_maps(&self->host_machine);
108 109

	if (ret >= 0)
110
		ret = machines__create_guest_kernel_maps(&self->machines);
111 112 113
	return ret;
}

114 115 116 117 118 119
static void perf_session__destroy_kernel_maps(struct perf_session *self)
{
	machine__destroy_kernel_maps(&self->host_machine);
	machines__destroy_guest_kernel_maps(&self->machines);
}

120 121
struct perf_session *perf_session__new(const char *filename, int mode,
				       bool force, bool repipe,
122
				       struct perf_tool *tool)
123
{
124 125 126 127 128 129 130 131 132 133 134 135 136
	struct perf_session *self;
	struct stat st;
	size_t len;

	if (!filename || !strlen(filename)) {
		if (!fstat(STDIN_FILENO, &st) && S_ISFIFO(st.st_mode))
			filename = "-";
		else
			filename = "perf.data";
	}

	len = strlen(filename);
	self = zalloc(sizeof(*self) + len);
137 138 139 140 141

	if (self == NULL)
		goto out;

	memcpy(self->filename, filename, len);
142 143 144 145 146 147 148 149 150
	/*
	 * On 64bit we can mmap the data file in one go. No need for tiny mmap
	 * slices. On 32bit we use 32MB.
	 */
#if BITS_PER_LONG == 64
	self->mmap_window = ULLONG_MAX;
#else
	self->mmap_window = 32 * 1024 * 1024ULL;
#endif
151
	self->machines = RB_ROOT;
T
Tom Zanussi 已提交
152
	self->repipe = repipe;
153
	INIT_LIST_HEAD(&self->ordered_samples.samples);
154
	INIT_LIST_HEAD(&self->ordered_samples.sample_cache);
155
	INIT_LIST_HEAD(&self->ordered_samples.to_free);
156
	machine__init(&self->host_machine, "", HOST_KERNEL_ID);
157
	hists__init(&self->hists);
158

159 160 161
	if (mode == O_RDONLY) {
		if (perf_session__open(self, force) < 0)
			goto out_delete;
162
		perf_session__update_sample_type(self);
163 164 165
	} else if (mode == O_WRONLY) {
		/*
		 * In O_RDONLY mode this will be performed when reading the
166
		 * kernel MMAP event, in perf_event__process_mmap().
167 168 169 170
		 */
		if (perf_session__create_kernel_maps(self) < 0)
			goto out_delete;
	}
171

172 173
	if (tool && tool->ordering_requires_timestamps &&
	    tool->ordered_samples && !self->sample_id_all) {
174
		dump_printf("WARNING: No sample_id_all support, falling back to unordered processing\n");
175
		tool->ordered_samples = false;
176 177
	}

178 179
out:
	return self;
180 181 182
out_delete:
	perf_session__delete(self);
	return NULL;
183 184
}

185
static void machine__delete_dead_threads(struct machine *machine)
186 187 188
{
	struct thread *n, *t;

189
	list_for_each_entry_safe(t, n, &machine->dead_threads, node) {
190 191 192 193 194
		list_del(&t->node);
		thread__delete(t);
	}
}

195 196 197 198 199 200
static void perf_session__delete_dead_threads(struct perf_session *session)
{
	machine__delete_dead_threads(&session->host_machine);
}

static void machine__delete_threads(struct machine *self)
201 202 203 204 205 206 207 208 209 210 211 212
{
	struct rb_node *nd = rb_first(&self->threads);

	while (nd) {
		struct thread *t = rb_entry(nd, struct thread, rb_node);

		rb_erase(&t->rb_node, &self->threads);
		nd = rb_next(nd);
		thread__delete(t);
	}
}

213 214 215 216 217
static void perf_session__delete_threads(struct perf_session *session)
{
	machine__delete_threads(&session->host_machine);
}

218 219
void perf_session__delete(struct perf_session *self)
{
220
	perf_session__destroy_kernel_maps(self);
221 222 223
	perf_session__delete_dead_threads(self);
	perf_session__delete_threads(self);
	machine__exit(&self->host_machine);
224 225 226
	close(self->fd);
	free(self);
}
227

228
void machine__remove_thread(struct machine *self, struct thread *th)
229
{
230
	self->last_match = NULL;
231 232 233 234 235 236 237 238
	rb_erase(&th->rb_node, &self->threads);
	/*
	 * We may have references to this thread, for instance in some hist_entry
	 * instances, so just move them to a separate list.
	 */
	list_add_tail(&th->node, &self->dead_threads);
}

239 240 241 242 243 244 245 246
static bool symbol__match_parent_regex(struct symbol *sym)
{
	if (sym->name && !regexec(&parent_regex, sym->name, 0, NULL, 0))
		return 1;

	return 0;
}

247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280
static const u8 cpumodes[] = {
	PERF_RECORD_MISC_USER,
	PERF_RECORD_MISC_KERNEL,
	PERF_RECORD_MISC_GUEST_USER,
	PERF_RECORD_MISC_GUEST_KERNEL
};
#define NCPUMODES (sizeof(cpumodes)/sizeof(u8))

static void ip__resolve_ams(struct machine *self, struct thread *thread,
			    struct addr_map_symbol *ams,
			    u64 ip)
{
	struct addr_location al;
	size_t i;
	u8 m;

	memset(&al, 0, sizeof(al));

	for (i = 0; i < NCPUMODES; i++) {
		m = cpumodes[i];
		/*
		 * We cannot use the header.misc hint to determine whether a
		 * branch stack address is user, kernel, guest, hypervisor.
		 * Branches may straddle the kernel/user/hypervisor boundaries.
		 * Thus, we have to try consecutively until we find a match
		 * or else, the symbol is unknown
		 */
		thread__find_addr_location(thread, self, m, MAP__FUNCTION,
				ip, &al, NULL);
		if (al.sym)
			goto found;
	}
found:
	ams->addr = ip;
281
	ams->al_addr = al.addr;
282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304
	ams->sym = al.sym;
	ams->map = al.map;
}

struct branch_info *machine__resolve_bstack(struct machine *self,
					    struct thread *thr,
					    struct branch_stack *bs)
{
	struct branch_info *bi;
	unsigned int i;

	bi = calloc(bs->nr, sizeof(struct branch_info));
	if (!bi)
		return NULL;

	for (i = 0; i < bs->nr; i++) {
		ip__resolve_ams(self, thr, &bi[i].to, bs->entries[i].to);
		ip__resolve_ams(self, thr, &bi[i].from, bs->entries[i].from);
		bi[i].flags = bs->entries[i].flags;
	}
	return bi;
}

305
int machine__resolve_callchain(struct machine *self,
306 307 308
			       struct thread *thread,
			       struct ip_callchain *chain,
			       struct symbol **parent)
309 310 311
{
	u8 cpumode = PERF_RECORD_MISC_USER;
	unsigned int i;
312
	int err;
313

314
	callchain_cursor_reset(&callchain_cursor);
315

316 317 318 319 320
	if (chain->nr > PERF_MAX_STACK_DEPTH) {
		pr_warning("corrupted callchain. skipping...\n");
		return 0;
	}

321
	for (i = 0; i < chain->nr; i++) {
322
		u64 ip;
323 324
		struct addr_location al;

325 326 327 328 329
		if (callchain_param.order == ORDER_CALLEE)
			ip = chain->ips[i];
		else
			ip = chain->ips[chain->nr - i - 1];

330 331 332 333 334 335 336 337 338
		if (ip >= PERF_CONTEXT_MAX) {
			switch (ip) {
			case PERF_CONTEXT_HV:
				cpumode = PERF_RECORD_MISC_HYPERVISOR;	break;
			case PERF_CONTEXT_KERNEL:
				cpumode = PERF_RECORD_MISC_KERNEL;	break;
			case PERF_CONTEXT_USER:
				cpumode = PERF_RECORD_MISC_USER;	break;
			default:
339 340 341 342 343 344 345 346
				pr_debug("invalid callchain context: "
					 "%"PRId64"\n", (s64) ip);
				/*
				 * It seems the callchain is corrupted.
				 * Discard all.
				 */
				callchain_cursor_reset(&callchain_cursor);
				return 0;
347 348 349 350
			}
			continue;
		}

351
		al.filtered = false;
352
		thread__find_addr_location(thread, self, cpumode,
353
					   MAP__FUNCTION, ip, &al, NULL);
354 355 356 357
		if (al.sym != NULL) {
			if (sort__has_parent && !*parent &&
			    symbol__match_parent_regex(al.sym))
				*parent = al.sym;
358
			if (!symbol_conf.use_callchain)
359 360
				break;
		}
361

362
		err = callchain_cursor_append(&callchain_cursor,
363 364 365
					      ip, al.map, al.sym);
		if (err)
			return err;
366 367
	}

368
	return 0;
369
}
370

371 372 373 374 375 376 377
static int process_event_synth_tracing_data_stub(union perf_event *event __used,
						 struct perf_session *session __used)
{
	dump_printf(": unhandled!\n");
	return 0;
}

378 379 380 381 382 383 384
static int process_event_synth_attr_stub(union perf_event *event __used,
					 struct perf_evlist **pevlist __used)
{
	dump_printf(": unhandled!\n");
	return 0;
}

385
static int process_event_sample_stub(struct perf_tool *tool __used,
386
				     union perf_event *event __used,
387 388
				     struct perf_sample *sample __used,
				     struct perf_evsel *evsel __used,
389
				     struct machine *machine __used)
390 391 392 393 394
{
	dump_printf(": unhandled!\n");
	return 0;
}

395
static int process_event_stub(struct perf_tool *tool __used,
396
			      union perf_event *event __used,
397
			      struct perf_sample *sample __used,
398
			      struct machine *machine __used)
399 400 401 402 403
{
	dump_printf(": unhandled!\n");
	return 0;
}

404
static int process_finished_round_stub(struct perf_tool *tool __used,
405
				       union perf_event *event __used,
406 407 408 409 410 411
				       struct perf_session *perf_session __used)
{
	dump_printf(": unhandled!\n");
	return 0;
}

412
static int process_event_type_stub(struct perf_tool *tool __used,
413
				   union perf_event *event __used)
414 415 416 417 418
{
	dump_printf(": unhandled!\n");
	return 0;
}

419
static int process_finished_round(struct perf_tool *tool,
420 421
				  union perf_event *event,
				  struct perf_session *session);
422

423
static void perf_tool__fill_defaults(struct perf_tool *tool)
424
{
425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453
	if (tool->sample == NULL)
		tool->sample = process_event_sample_stub;
	if (tool->mmap == NULL)
		tool->mmap = process_event_stub;
	if (tool->comm == NULL)
		tool->comm = process_event_stub;
	if (tool->fork == NULL)
		tool->fork = process_event_stub;
	if (tool->exit == NULL)
		tool->exit = process_event_stub;
	if (tool->lost == NULL)
		tool->lost = perf_event__process_lost;
	if (tool->read == NULL)
		tool->read = process_event_sample_stub;
	if (tool->throttle == NULL)
		tool->throttle = process_event_stub;
	if (tool->unthrottle == NULL)
		tool->unthrottle = process_event_stub;
	if (tool->attr == NULL)
		tool->attr = process_event_synth_attr_stub;
	if (tool->event_type == NULL)
		tool->event_type = process_event_type_stub;
	if (tool->tracing_data == NULL)
		tool->tracing_data = process_event_synth_tracing_data_stub;
	if (tool->build_id == NULL)
		tool->build_id = process_finished_round_stub;
	if (tool->finished_round == NULL) {
		if (tool->ordered_samples)
			tool->finished_round = process_finished_round;
454
		else
455
			tool->finished_round = process_finished_round_stub;
456
	}
457
}
458 459 460 461 462 463 464 465 466 467
 
void mem_bswap_32(void *src, int byte_size)
{
	u32 *m = src;
	while (byte_size > 0) {
		*m = bswap_32(*m);
		byte_size -= sizeof(u32);
		++m;
	}
}
468

469 470 471 472 473 474 475 476 477 478 479
void mem_bswap_64(void *src, int byte_size)
{
	u64 *m = src;

	while (byte_size > 0) {
		*m = bswap_64(*m);
		byte_size -= sizeof(u64);
		++m;
	}
}

480 481 482 483 484 485 486 487 488 489 490
static void swap_sample_id_all(union perf_event *event, void *data)
{
	void *end = (void *) event + event->header.size;
	int size = end - data;

	BUG_ON(size % sizeof(u64));
	mem_bswap_64(data, size);
}

static void perf_event__all64_swap(union perf_event *event,
				   bool sample_id_all __used)
491
{
492 493
	struct perf_event_header *hdr = &event->header;
	mem_bswap_64(hdr + 1, event->header.size - sizeof(*hdr));
494 495
}

496
static void perf_event__comm_swap(union perf_event *event, bool sample_id_all)
497
{
498 499
	event->comm.pid = bswap_32(event->comm.pid);
	event->comm.tid = bswap_32(event->comm.tid);
500 501 502 503 504 505 506

	if (sample_id_all) {
		void *data = &event->comm.comm;

		data += ALIGN(strlen(data) + 1, sizeof(u64));
		swap_sample_id_all(event, data);
	}
507 508
}

509 510
static void perf_event__mmap_swap(union perf_event *event,
				  bool sample_id_all)
511
{
512 513 514 515 516
	event->mmap.pid	  = bswap_32(event->mmap.pid);
	event->mmap.tid	  = bswap_32(event->mmap.tid);
	event->mmap.start = bswap_64(event->mmap.start);
	event->mmap.len	  = bswap_64(event->mmap.len);
	event->mmap.pgoff = bswap_64(event->mmap.pgoff);
517 518 519 520 521 522 523

	if (sample_id_all) {
		void *data = &event->mmap.filename;

		data += ALIGN(strlen(data) + 1, sizeof(u64));
		swap_sample_id_all(event, data);
	}
524 525
}

526
static void perf_event__task_swap(union perf_event *event, bool sample_id_all)
527
{
528 529 530 531 532
	event->fork.pid	 = bswap_32(event->fork.pid);
	event->fork.tid	 = bswap_32(event->fork.tid);
	event->fork.ppid = bswap_32(event->fork.ppid);
	event->fork.ptid = bswap_32(event->fork.ptid);
	event->fork.time = bswap_64(event->fork.time);
533 534 535

	if (sample_id_all)
		swap_sample_id_all(event, &event->fork + 1);
536 537
}

538
static void perf_event__read_swap(union perf_event *event, bool sample_id_all)
539
{
540 541 542 543 544 545
	event->read.pid		 = bswap_32(event->read.pid);
	event->read.tid		 = bswap_32(event->read.tid);
	event->read.value	 = bswap_64(event->read.value);
	event->read.time_enabled = bswap_64(event->read.time_enabled);
	event->read.time_running = bswap_64(event->read.time_running);
	event->read.id		 = bswap_64(event->read.id);
546 547 548

	if (sample_id_all)
		swap_sample_id_all(event, &event->read + 1);
549 550
}

551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582
static u8 revbyte(u8 b)
{
	int rev = (b >> 4) | ((b & 0xf) << 4);
	rev = ((rev & 0xcc) >> 2) | ((rev & 0x33) << 2);
	rev = ((rev & 0xaa) >> 1) | ((rev & 0x55) << 1);
	return (u8) rev;
}

/*
 * XXX this is hack in attempt to carry flags bitfield
 * throught endian village. ABI says:
 *
 * Bit-fields are allocated from right to left (least to most significant)
 * on little-endian implementations and from left to right (most to least
 * significant) on big-endian implementations.
 *
 * The above seems to be byte specific, so we need to reverse each
 * byte of the bitfield. 'Internet' also says this might be implementation
 * specific and we probably need proper fix and carry perf_event_attr
 * bitfield flags in separate data file FEAT_ section. Thought this seems
 * to work for now.
 */
static void swap_bitfield(u8 *p, unsigned len)
{
	unsigned i;

	for (i = 0; i < len; i++) {
		*p = revbyte(*p);
		p++;
	}
}

583 584 585 586 587 588 589 590 591 592 593 594 595
/* exported for swapping attributes in file header */
void perf_event__attr_swap(struct perf_event_attr *attr)
{
	attr->type		= bswap_32(attr->type);
	attr->size		= bswap_32(attr->size);
	attr->config		= bswap_64(attr->config);
	attr->sample_period	= bswap_64(attr->sample_period);
	attr->sample_type	= bswap_64(attr->sample_type);
	attr->read_format	= bswap_64(attr->read_format);
	attr->wakeup_events	= bswap_32(attr->wakeup_events);
	attr->bp_type		= bswap_32(attr->bp_type);
	attr->bp_addr		= bswap_64(attr->bp_addr);
	attr->bp_len		= bswap_64(attr->bp_len);
596 597

	swap_bitfield((u8 *) (&attr->read_format + 1), sizeof(u64));
598 599
}

600 601
static void perf_event__hdr_attr_swap(union perf_event *event,
				      bool sample_id_all __used)
602 603 604
{
	size_t size;

605
	perf_event__attr_swap(&event->attr.attr);
606

607 608 609
	size = event->header.size;
	size -= (void *)&event->attr.id - (void *)event;
	mem_bswap_64(event->attr.id, size);
610 611
}

612 613
static void perf_event__event_type_swap(union perf_event *event,
					bool sample_id_all __used)
614
{
615 616
	event->event_type.event_type.event_id =
		bswap_64(event->event_type.event_type.event_id);
617 618
}

619 620
static void perf_event__tracing_data_swap(union perf_event *event,
					  bool sample_id_all __used)
621
{
622
	event->tracing_data.size = bswap_32(event->tracing_data.size);
623 624
}

625 626
typedef void (*perf_event__swap_op)(union perf_event *event,
				    bool sample_id_all);
627

628 629 630 631 632 633 634 635
static perf_event__swap_op perf_event__swap_ops[] = {
	[PERF_RECORD_MMAP]		  = perf_event__mmap_swap,
	[PERF_RECORD_COMM]		  = perf_event__comm_swap,
	[PERF_RECORD_FORK]		  = perf_event__task_swap,
	[PERF_RECORD_EXIT]		  = perf_event__task_swap,
	[PERF_RECORD_LOST]		  = perf_event__all64_swap,
	[PERF_RECORD_READ]		  = perf_event__read_swap,
	[PERF_RECORD_SAMPLE]		  = perf_event__all64_swap,
636
	[PERF_RECORD_HEADER_ATTR]	  = perf_event__hdr_attr_swap,
637 638 639 640
	[PERF_RECORD_HEADER_EVENT_TYPE]	  = perf_event__event_type_swap,
	[PERF_RECORD_HEADER_TRACING_DATA] = perf_event__tracing_data_swap,
	[PERF_RECORD_HEADER_BUILD_ID]	  = NULL,
	[PERF_RECORD_HEADER_MAX]	  = NULL,
641 642
};

643 644
struct sample_queue {
	u64			timestamp;
645
	u64			file_offset;
646
	union perf_event	*event;
647 648 649
	struct list_head	list;
};

650 651 652 653
static void perf_session_free_sample_buffers(struct perf_session *session)
{
	struct ordered_samples *os = &session->ordered_samples;

654
	while (!list_empty(&os->to_free)) {
655 656
		struct sample_queue *sq;

657
		sq = list_entry(os->to_free.next, struct sample_queue, list);
658 659 660 661 662
		list_del(&sq->list);
		free(sq);
	}
}

663
static int perf_session_deliver_event(struct perf_session *session,
664
				      union perf_event *event,
665
				      struct perf_sample *sample,
666
				      struct perf_tool *tool,
667
				      u64 file_offset);
668

669
static void flush_sample_queue(struct perf_session *s,
670
			       struct perf_tool *tool)
671
{
672 673
	struct ordered_samples *os = &s->ordered_samples;
	struct list_head *head = &os->samples;
674
	struct sample_queue *tmp, *iter;
675
	struct perf_sample sample;
676 677
	u64 limit = os->next_flush;
	u64 last_ts = os->last_sample ? os->last_sample->timestamp : 0ULL;
678
	unsigned idx = 0, progress_next = os->nr_samples / 16;
679
	int ret;
680

681
	if (!tool->ordered_samples || !limit)
682 683 684 685
		return;

	list_for_each_entry_safe(iter, tmp, head, list) {
		if (iter->timestamp > limit)
686
			break;
687

688 689 690 691
		ret = perf_session__parse_sample(s, iter->event, &sample);
		if (ret)
			pr_err("Can't parse sample, err = %d\n", ret);
		else
692
			perf_session_deliver_event(s, iter->event, &sample, tool,
693
						   iter->file_offset);
694

695
		os->last_flush = iter->timestamp;
696
		list_del(&iter->list);
697
		list_add(&iter->list, &os->sample_cache);
698 699 700 701 702
		if (++idx >= progress_next) {
			progress_next += os->nr_samples / 16;
			ui_progress__update(idx, os->nr_samples,
					    "Processing time ordered events...");
		}
703
	}
704 705 706 707 708 709 710

	if (list_empty(head)) {
		os->last_sample = NULL;
	} else if (last_ts <= limit) {
		os->last_sample =
			list_entry(head->prev, struct sample_queue, list);
	}
711 712

	os->nr_samples = 0;
713 714
}

715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753
/*
 * When perf record finishes a pass on every buffers, it records this pseudo
 * event.
 * We record the max timestamp t found in the pass n.
 * Assuming these timestamps are monotonic across cpus, we know that if
 * a buffer still has events with timestamps below t, they will be all
 * available and then read in the pass n + 1.
 * Hence when we start to read the pass n + 2, we can safely flush every
 * events with timestamps below t.
 *
 *    ============ PASS n =================
 *       CPU 0         |   CPU 1
 *                     |
 *    cnt1 timestamps  |   cnt2 timestamps
 *          1          |         2
 *          2          |         3
 *          -          |         4  <--- max recorded
 *
 *    ============ PASS n + 1 ==============
 *       CPU 0         |   CPU 1
 *                     |
 *    cnt1 timestamps  |   cnt2 timestamps
 *          3          |         5
 *          4          |         6
 *          5          |         7 <---- max recorded
 *
 *      Flush every events below timestamp 4
 *
 *    ============ PASS n + 2 ==============
 *       CPU 0         |   CPU 1
 *                     |
 *    cnt1 timestamps  |   cnt2 timestamps
 *          6          |         8
 *          7          |         9
 *          -          |         10
 *
 *      Flush every events below timestamp 7
 *      etc...
 */
754
static int process_finished_round(struct perf_tool *tool,
755 756
				  union perf_event *event __used,
				  struct perf_session *session)
757
{
758
	flush_sample_queue(session, tool);
759 760 761 762 763
	session->ordered_samples.next_flush = session->ordered_samples.max_timestamp;

	return 0;
}

764
/* The queue is ordered by time */
765
static void __queue_event(struct sample_queue *new, struct perf_session *s)
766
{
767 768 769 770
	struct ordered_samples *os = &s->ordered_samples;
	struct sample_queue *sample = os->last_sample;
	u64 timestamp = new->timestamp;
	struct list_head *p;
771

772
	++os->nr_samples;
773
	os->last_sample = new;
774

775 776 777
	if (!sample) {
		list_add(&new->list, &os->samples);
		os->max_timestamp = timestamp;
778 779 780 781
		return;
	}

	/*
782 783 784
	 * last_sample might point to some random place in the list as it's
	 * the last queued event. We expect that the new event is close to
	 * this.
785
	 */
786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807
	if (sample->timestamp <= timestamp) {
		while (sample->timestamp <= timestamp) {
			p = sample->list.next;
			if (p == &os->samples) {
				list_add_tail(&new->list, &os->samples);
				os->max_timestamp = timestamp;
				return;
			}
			sample = list_entry(p, struct sample_queue, list);
		}
		list_add_tail(&new->list, &sample->list);
	} else {
		while (sample->timestamp > timestamp) {
			p = sample->list.prev;
			if (p == &os->samples) {
				list_add(&new->list, &os->samples);
				return;
			}
			sample = list_entry(p, struct sample_queue, list);
		}
		list_add(&new->list, &sample->list);
	}
808 809
}

810 811
#define MAX_SAMPLE_BUFFER	(64 * 1024 / sizeof(struct sample_queue))

812
static int perf_session_queue_event(struct perf_session *s, union perf_event *event,
813
				    struct perf_sample *sample, u64 file_offset)
814
{
815 816
	struct ordered_samples *os = &s->ordered_samples;
	struct list_head *sc = &os->sample_cache;
817
	u64 timestamp = sample->time;
818 819
	struct sample_queue *new;

820
	if (!timestamp || timestamp == ~0ULL)
821 822
		return -ETIME;

823 824 825 826 827
	if (timestamp < s->ordered_samples.last_flush) {
		printf("Warning: Timestamp below last timeslice flush\n");
		return -EINVAL;
	}

828 829 830
	if (!list_empty(sc)) {
		new = list_entry(sc->next, struct sample_queue, list);
		list_del(&new->list);
831 832 833 834
	} else if (os->sample_buffer) {
		new = os->sample_buffer + os->sample_buffer_idx;
		if (++os->sample_buffer_idx == MAX_SAMPLE_BUFFER)
			os->sample_buffer = NULL;
835
	} else {
836 837
		os->sample_buffer = malloc(MAX_SAMPLE_BUFFER * sizeof(*new));
		if (!os->sample_buffer)
838
			return -ENOMEM;
839 840 841
		list_add(&os->sample_buffer->list, &os->to_free);
		os->sample_buffer_idx = 2;
		new = os->sample_buffer + 1;
842
	}
843 844

	new->timestamp = timestamp;
845
	new->file_offset = file_offset;
846
	new->event = event;
847

848
	__queue_event(new, s);
849 850 851

	return 0;
}
852

853
static void callchain__printf(struct perf_sample *sample)
854 855
{
	unsigned int i;
856

857
	printf("... chain: nr:%" PRIu64 "\n", sample->callchain->nr);
858 859

	for (i = 0; i < sample->callchain->nr; i++)
860 861
		printf("..... %2d: %016" PRIx64 "\n",
		       i, sample->callchain->ips[i]);
862 863
}

864 865 866 867 868 869 870 871 872 873 874 875
static void branch_stack__printf(struct perf_sample *sample)
{
	uint64_t i;

	printf("... branch stack: nr:%" PRIu64 "\n", sample->branch_stack->nr);

	for (i = 0; i < sample->branch_stack->nr; i++)
		printf("..... %2"PRIu64": %016" PRIx64 " -> %016" PRIx64 "\n",
			i, sample->branch_stack->entries[i].from,
			sample->branch_stack->entries[i].to);
}

876
static void perf_session__print_tstamp(struct perf_session *session,
877
				       union perf_event *event,
878
				       struct perf_sample *sample)
879 880 881 882 883 884 885 886 887 888 889
{
	if (event->header.type != PERF_RECORD_SAMPLE &&
	    !session->sample_id_all) {
		fputs("-1 -1 ", stdout);
		return;
	}

	if ((session->sample_type & PERF_SAMPLE_CPU))
		printf("%u ", sample->cpu);

	if (session->sample_type & PERF_SAMPLE_TIME)
890
		printf("%" PRIu64 " ", sample->time);
891 892
}

893
static void dump_event(struct perf_session *session, union perf_event *event,
894
		       u64 file_offset, struct perf_sample *sample)
895 896 897 898
{
	if (!dump_trace)
		return;

899 900
	printf("\n%#" PRIx64 " [%#x]: event: %d\n",
	       file_offset, event->header.size, event->header.type);
901 902 903 904 905 906

	trace_event(event);

	if (sample)
		perf_session__print_tstamp(session, event, sample);

907
	printf("%#" PRIx64 " [%#x]: PERF_RECORD_%s", file_offset,
908
	       event->header.size, perf_event__name(event->header.type));
909 910
}

911
static void dump_sample(struct perf_session *session, union perf_event *event,
912
			struct perf_sample *sample)
913
{
914 915 916
	if (!dump_trace)
		return;

917
	printf("(IP, %d): %d/%d: %#" PRIx64 " period: %" PRIu64 " addr: %#" PRIx64 "\n",
918
	       event->header.misc, sample->pid, sample->tid, sample->ip,
919
	       sample->period, sample->addr);
920 921

	if (session->sample_type & PERF_SAMPLE_CALLCHAIN)
922
		callchain__printf(sample);
923 924 925

	if (session->sample_type & PERF_SAMPLE_BRANCH_STACK)
		branch_stack__printf(sample);
926 927
}

928 929 930 931 932 933
static struct machine *
	perf_session__find_machine_for_cpumode(struct perf_session *session,
					       union perf_event *event)
{
	const u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;

934 935 936
	if (perf_guest &&
	    ((cpumode == PERF_RECORD_MISC_GUEST_KERNEL) ||
	     (cpumode == PERF_RECORD_MISC_GUEST_USER))) {
937 938 939 940 941 942 943
		u32 pid;

		if (event->header.type == PERF_RECORD_MMAP)
			pid = event->mmap.pid;
		else
			pid = event->ip.pid;

944
		return perf_session__findnew_machine(session, pid);
945
	}
946 947 948 949

	return perf_session__find_host_machine(session);
}

950
static int perf_session_deliver_event(struct perf_session *session,
951
				      union perf_event *event,
952
				      struct perf_sample *sample,
953
				      struct perf_tool *tool,
954
				      u64 file_offset)
955
{
956
	struct perf_evsel *evsel;
957
	struct machine *machine;
958

959 960
	dump_event(session, event, file_offset, sample);

961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978
	evsel = perf_evlist__id2evsel(session->evlist, sample->id);
	if (evsel != NULL && event->header.type != PERF_RECORD_SAMPLE) {
		/*
		 * XXX We're leaving PERF_RECORD_SAMPLE unnacounted here
		 * because the tools right now may apply filters, discarding
		 * some of the samples. For consistency, in the future we
		 * should have something like nr_filtered_samples and remove
		 * the sample->period from total_sample_period, etc, KISS for
		 * now tho.
		 *
		 * Also testing against NULL allows us to handle files without
		 * attr.sample_id_all and/or without PERF_SAMPLE_ID. In the
		 * future probably it'll be a good idea to restrict event
		 * processing via perf_session to files with both set.
		 */
		hists__inc_nr_events(&evsel->hists, event->header.type);
	}

979 980
	machine = perf_session__find_machine_for_cpumode(session, event);

981 982
	switch (event->header.type) {
	case PERF_RECORD_SAMPLE:
983
		dump_sample(session, event, sample);
984 985
		if (evsel == NULL) {
			++session->hists.stats.nr_unknown_id;
986
			return 0;
987
		}
988 989
		if (machine == NULL) {
			++session->hists.stats.nr_unprocessable_samples;
990
			return 0;
991
		}
992
		return tool->sample(tool, event, sample, evsel, machine);
993
	case PERF_RECORD_MMAP:
994
		return tool->mmap(tool, event, sample, machine);
995
	case PERF_RECORD_COMM:
996
		return tool->comm(tool, event, sample, machine);
997
	case PERF_RECORD_FORK:
998
		return tool->fork(tool, event, sample, machine);
999
	case PERF_RECORD_EXIT:
1000
		return tool->exit(tool, event, sample, machine);
1001
	case PERF_RECORD_LOST:
1002
		if (tool->lost == perf_event__process_lost)
1003
			session->hists.stats.total_lost += event->lost.lost;
1004
		return tool->lost(tool, event, sample, machine);
1005
	case PERF_RECORD_READ:
1006
		return tool->read(tool, event, sample, evsel, machine);
1007
	case PERF_RECORD_THROTTLE:
1008
		return tool->throttle(tool, event, sample, machine);
1009
	case PERF_RECORD_UNTHROTTLE:
1010
		return tool->unthrottle(tool, event, sample, machine);
1011 1012 1013 1014 1015 1016
	default:
		++session->hists.stats.nr_unknown_events;
		return -1;
	}
}

1017
static int perf_session__preprocess_sample(struct perf_session *session,
1018
					   union perf_event *event, struct perf_sample *sample)
1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032
{
	if (event->header.type != PERF_RECORD_SAMPLE ||
	    !(session->sample_type & PERF_SAMPLE_CALLCHAIN))
		return 0;

	if (!ip_callchain__valid(sample->callchain, event)) {
		pr_debug("call-chain problem with event, skipping it.\n");
		++session->hists.stats.nr_invalid_chains;
		session->hists.stats.total_invalid_chains += sample->period;
		return -EINVAL;
	}
	return 0;
}

1033
static int perf_session__process_user_event(struct perf_session *session, union perf_event *event,
1034
					    struct perf_tool *tool, u64 file_offset)
1035
{
1036 1037
	int err;

1038
	dump_event(session, event, file_offset, NULL);
1039

1040
	/* These events are processed right away */
1041
	switch (event->header.type) {
1042
	case PERF_RECORD_HEADER_ATTR:
1043
		err = tool->attr(event, &session->evlist);
1044 1045 1046
		if (err == 0)
			perf_session__update_sample_type(session);
		return err;
1047
	case PERF_RECORD_HEADER_EVENT_TYPE:
1048
		return tool->event_type(tool, event);
1049 1050
	case PERF_RECORD_HEADER_TRACING_DATA:
		/* setup for reading amidst mmap */
1051
		lseek(session->fd, file_offset, SEEK_SET);
1052
		return tool->tracing_data(event, session);
1053
	case PERF_RECORD_HEADER_BUILD_ID:
1054
		return tool->build_id(tool, event, session);
1055
	case PERF_RECORD_FINISHED_ROUND:
1056
		return tool->finished_round(tool, event, session);
1057
	default:
1058
		return -EINVAL;
1059
	}
1060 1061
}

1062 1063 1064 1065 1066 1067 1068 1069 1070
static void event_swap(union perf_event *event, bool sample_id_all)
{
	perf_event__swap_op swap;

	swap = perf_event__swap_ops[event->header.type];
	if (swap)
		swap(event, sample_id_all);
}

1071
static int perf_session__process_event(struct perf_session *session,
1072
				       union perf_event *event,
1073
				       struct perf_tool *tool,
1074 1075
				       u64 file_offset)
{
1076
	struct perf_sample sample;
1077 1078
	int ret;

1079 1080
	if (session->header.needs_swap)
		event_swap(event, session->sample_id_all);
1081 1082 1083 1084 1085 1086 1087

	if (event->header.type >= PERF_RECORD_HEADER_MAX)
		return -EINVAL;

	hists__inc_nr_events(&session->hists, event->header.type);

	if (event->header.type >= PERF_RECORD_USER_TYPE_START)
1088
		return perf_session__process_user_event(session, event, tool, file_offset);
1089

1090 1091 1092
	/*
	 * For all kernel events we get the sample data
	 */
1093 1094 1095
	ret = perf_session__parse_sample(session, event, &sample);
	if (ret)
		return ret;
1096 1097 1098 1099 1100

	/* Preprocess sample records - precheck callchains */
	if (perf_session__preprocess_sample(session, event, &sample))
		return 0;

1101
	if (tool->ordered_samples) {
1102 1103
		ret = perf_session_queue_event(session, event, &sample,
					       file_offset);
1104 1105 1106 1107
		if (ret != -ETIME)
			return ret;
	}

1108
	return perf_session_deliver_event(session, event, &sample, tool,
1109
					  file_offset);
1110 1111
}

1112 1113 1114 1115 1116 1117 1118
void perf_event_header__bswap(struct perf_event_header *self)
{
	self->type = bswap_32(self->type);
	self->misc = bswap_16(self->misc);
	self->size = bswap_16(self->size);
}

1119 1120 1121 1122 1123
struct thread *perf_session__findnew(struct perf_session *session, pid_t pid)
{
	return machine__findnew_thread(&session->host_machine, pid);
}

1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135
static struct thread *perf_session__register_idle_thread(struct perf_session *self)
{
	struct thread *thread = perf_session__findnew(self, 0);

	if (thread == NULL || thread__set_comm(thread, "swapper")) {
		pr_err("problem inserting idle task.\n");
		thread = NULL;
	}

	return thread;
}

1136
static void perf_session__warn_about_errors(const struct perf_session *session,
1137
					    const struct perf_tool *tool)
1138
{
1139
	if (tool->lost == perf_event__process_lost &&
1140 1141 1142 1143 1144
	    session->hists.stats.nr_events[PERF_RECORD_LOST] != 0) {
		ui__warning("Processed %d events and lost %d chunks!\n\n"
			    "Check IO/CPU overload!\n\n",
			    session->hists.stats.nr_events[0],
			    session->hists.stats.nr_events[PERF_RECORD_LOST]);
1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155
	}

	if (session->hists.stats.nr_unknown_events != 0) {
		ui__warning("Found %u unknown events!\n\n"
			    "Is this an older tool processing a perf.data "
			    "file generated by a more recent tool?\n\n"
			    "If that is not the case, consider "
			    "reporting to linux-kernel@vger.kernel.org.\n\n",
			    session->hists.stats.nr_unknown_events);
	}

1156 1157 1158 1159 1160
	if (session->hists.stats.nr_unknown_id != 0) {
		ui__warning("%u samples with id not present in the header\n",
			    session->hists.stats.nr_unknown_id);
	}

1161 1162 1163 1164 1165 1166 1167
 	if (session->hists.stats.nr_invalid_chains != 0) {
 		ui__warning("Found invalid callchains!\n\n"
 			    "%u out of %u events were discarded for this reason.\n\n"
 			    "Consider reporting to linux-kernel@vger.kernel.org.\n\n",
 			    session->hists.stats.nr_invalid_chains,
 			    session->hists.stats.nr_events[PERF_RECORD_SAMPLE]);
 	}
1168 1169 1170 1171 1172 1173

	if (session->hists.stats.nr_unprocessable_samples != 0) {
		ui__warning("%u unprocessable samples recorded.\n"
			    "Do you have a KVM guest running and not using 'perf kvm'?\n",
			    session->hists.stats.nr_unprocessable_samples);
	}
1174 1175
}

1176 1177 1178 1179
#define session_done()	(*(volatile int *)(&session_done))
volatile int session_done;

static int __perf_session__process_pipe_events(struct perf_session *self,
1180
					       struct perf_tool *tool)
1181
{
1182 1183 1184
	union perf_event *event;
	uint32_t size, cur_size = 0;
	void *buf = NULL;
1185 1186 1187 1188 1189
	int skip = 0;
	u64 head;
	int err;
	void *p;

1190
	perf_tool__fill_defaults(tool);
1191 1192

	head = 0;
1193 1194 1195 1196 1197
	cur_size = sizeof(union perf_event);

	buf = malloc(cur_size);
	if (!buf)
		return -errno;
1198
more:
1199 1200
	event = buf;
	err = readn(self->fd, event, sizeof(struct perf_event_header));
1201 1202 1203 1204 1205 1206 1207 1208 1209
	if (err <= 0) {
		if (err == 0)
			goto done;

		pr_err("failed to read event header\n");
		goto out_err;
	}

	if (self->header.needs_swap)
1210
		perf_event_header__bswap(&event->header);
1211

1212
	size = event->header.size;
1213 1214 1215
	if (size == 0)
		size = 8;

1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226
	if (size > cur_size) {
		void *new = realloc(buf, size);
		if (!new) {
			pr_err("failed to allocate memory to read event\n");
			goto out_err;
		}
		buf = new;
		cur_size = size;
		event = buf;
	}
	p = event;
1227 1228
	p += sizeof(struct perf_event_header);

1229
	if (size - sizeof(struct perf_event_header)) {
1230
		err = readn(self->fd, p, size - sizeof(struct perf_event_header));
1231 1232 1233 1234 1235
		if (err <= 0) {
			if (err == 0) {
				pr_err("unexpected end of event stream\n");
				goto done;
			}
1236

1237 1238 1239
			pr_err("failed to read event data\n");
			goto out_err;
		}
1240 1241
	}

1242
	if ((skip = perf_session__process_event(self, event, tool, head)) < 0) {
1243
		pr_err("%#" PRIx64 " [%#x]: failed to process type: %d\n",
1244
		       head, event->header.size, event->header.type);
1245 1246
		err = -EINVAL;
		goto out_err;
1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258
	}

	head += size;

	if (skip > 0)
		head += skip;

	if (!session_done())
		goto more;
done:
	err = 0;
out_err:
1259
	free(buf);
1260
	perf_session__warn_about_errors(self, tool);
1261
	perf_session_free_sample_buffers(self);
1262 1263 1264
	return err;
}

1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288
static union perf_event *
fetch_mmaped_event(struct perf_session *session,
		   u64 head, size_t mmap_size, char *buf)
{
	union perf_event *event;

	/*
	 * Ensure we have enough space remaining to read
	 * the size of the event in the headers.
	 */
	if (head + sizeof(event->header) > mmap_size)
		return NULL;

	event = (union perf_event *)(buf + head);

	if (session->header.needs_swap)
		perf_event_header__bswap(&event->header);

	if (head + event->header.size > mmap_size)
		return NULL;

	return event;
}

1289
int __perf_session__process_events(struct perf_session *session,
1290
				   u64 data_offset, u64 data_size,
1291
				   u64 file_size, struct perf_tool *tool)
1292
{
1293
	u64 head, page_offset, file_offset, file_pos, progress_next;
1294
	int err, mmap_prot, mmap_flags, map_idx = 0;
1295
	size_t	page_size, mmap_size;
1296
	char *buf, *mmaps[8];
1297
	union perf_event *event;
1298
	uint32_t size;
1299

1300
	perf_tool__fill_defaults(tool);
1301

1302
	page_size = sysconf(_SC_PAGESIZE);
1303

1304 1305 1306
	page_offset = page_size * (data_offset / page_size);
	file_offset = page_offset;
	head = data_offset - page_offset;
1307

1308 1309 1310
	if (data_offset + data_size < file_size)
		file_size = data_offset + data_size;

1311 1312 1313 1314 1315 1316
	progress_next = file_size / 16;

	mmap_size = session->mmap_window;
	if (mmap_size > file_size)
		mmap_size = file_size;

1317 1318
	memset(mmaps, 0, sizeof(mmaps));

1319 1320 1321
	mmap_prot  = PROT_READ;
	mmap_flags = MAP_SHARED;

1322
	if (session->header.needs_swap) {
1323 1324 1325
		mmap_prot  |= PROT_WRITE;
		mmap_flags = MAP_PRIVATE;
	}
1326
remap:
1327 1328
	buf = mmap(NULL, mmap_size, mmap_prot, mmap_flags, session->fd,
		   file_offset);
1329 1330 1331 1332 1333
	if (buf == MAP_FAILED) {
		pr_err("failed to mmap file\n");
		err = -errno;
		goto out_err;
	}
1334 1335
	mmaps[map_idx] = buf;
	map_idx = (map_idx + 1) & (ARRAY_SIZE(mmaps) - 1);
1336
	file_pos = file_offset + head;
1337 1338

more:
1339 1340
	event = fetch_mmaped_event(session, head, mmap_size, buf);
	if (!event) {
1341 1342 1343 1344
		if (mmaps[map_idx]) {
			munmap(mmaps[map_idx], mmap_size);
			mmaps[map_idx] = NULL;
		}
1345

1346 1347 1348
		page_offset = page_size * (head / page_size);
		file_offset += page_offset;
		head -= page_offset;
1349 1350 1351 1352 1353
		goto remap;
	}

	size = event->header.size;

1354
	if (size == 0 ||
1355
	    perf_session__process_event(session, event, tool, file_pos) < 0) {
1356 1357 1358 1359 1360
		pr_err("%#" PRIx64 " [%#x]: failed to process type: %d\n",
		       file_offset + head, event->header.size,
		       event->header.type);
		err = -EINVAL;
		goto out_err;
1361 1362 1363
	}

	head += size;
1364
	file_pos += size;
1365

1366 1367
	if (file_pos >= progress_next) {
		progress_next += file_size / 16;
1368 1369
		ui_progress__update(file_pos, file_size,
				    "Processing events...");
1370 1371
	}

1372
	if (file_pos < file_size)
1373
		goto more;
1374

1375
	err = 0;
1376
	/* do the final flush for ordered samples */
1377
	session->ordered_samples.next_flush = ULLONG_MAX;
1378
	flush_sample_queue(session, tool);
1379
out_err:
1380
	perf_session__warn_about_errors(session, tool);
1381
	perf_session_free_sample_buffers(session);
1382 1383
	return err;
}
1384

1385
int perf_session__process_events(struct perf_session *self,
1386
				 struct perf_tool *tool)
1387 1388 1389 1390 1391 1392
{
	int err;

	if (perf_session__register_idle_thread(self) == NULL)
		return -ENOMEM;

1393 1394 1395 1396
	if (!self->fd_pipe)
		err = __perf_session__process_events(self,
						     self->header.data_offset,
						     self->header.data_size,
1397
						     self->size, tool);
1398
	else
1399
		err = __perf_session__process_pipe_events(self, tool);
1400

1401 1402 1403
	return err;
}

1404
bool perf_session__has_traces(struct perf_session *self, const char *msg)
1405 1406
{
	if (!(self->sample_type & PERF_SAMPLE_RAW)) {
1407 1408
		pr_err("No trace sample to read. Did you call 'perf %s'?\n", msg);
		return false;
1409 1410
	}

1411
	return true;
1412
}
1413

1414 1415
int maps__set_kallsyms_ref_reloc_sym(struct map **maps,
				     const char *symbol_name, u64 addr)
1416 1417
{
	char *bracket;
1418
	enum map_type i;
1419 1420 1421 1422 1423
	struct ref_reloc_sym *ref;

	ref = zalloc(sizeof(struct ref_reloc_sym));
	if (ref == NULL)
		return -ENOMEM;
1424

1425 1426 1427
	ref->name = strdup(symbol_name);
	if (ref->name == NULL) {
		free(ref);
1428
		return -ENOMEM;
1429
	}
1430

1431
	bracket = strchr(ref->name, ']');
1432 1433 1434
	if (bracket)
		*bracket = '\0';

1435
	ref->addr = addr;
1436 1437

	for (i = 0; i < MAP__NR_TYPES; ++i) {
1438 1439
		struct kmap *kmap = map__kmap(maps[i]);
		kmap->ref_reloc_sym = ref;
1440 1441
	}

1442 1443
	return 0;
}
1444 1445 1446 1447 1448 1449 1450

size_t perf_session__fprintf_dsos(struct perf_session *self, FILE *fp)
{
	return __dsos__fprintf(&self->host_machine.kernel_dsos, fp) +
	       __dsos__fprintf(&self->host_machine.user_dsos, fp) +
	       machines__fprintf_dsos(&self->machines, fp);
}
1451 1452 1453 1454 1455 1456 1457

size_t perf_session__fprintf_dsos_buildid(struct perf_session *self, FILE *fp,
					  bool with_hits)
{
	size_t ret = machine__fprintf_dsos_buildid(&self->host_machine, fp, with_hits);
	return ret + machines__fprintf_dsos_buildid(&self->machines, fp, with_hits);
}
1458 1459 1460 1461 1462 1463 1464 1465 1466

size_t perf_session__fprintf_nr_events(struct perf_session *session, FILE *fp)
{
	struct perf_evsel *pos;
	size_t ret = fprintf(fp, "Aggregated stats:\n");

	ret += hists__fprintf_nr_events(&session->hists, fp);

	list_for_each_entry(pos, &session->evlist->entries, node) {
1467
		ret += fprintf(fp, "%s stats:\n", perf_evsel__name(pos));
1468 1469 1470 1471 1472
		ret += hists__fprintf_nr_events(&pos->hists, fp);
	}

	return ret;
}
1473

1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494
size_t perf_session__fprintf(struct perf_session *session, FILE *fp)
{
	/*
	 * FIXME: Here we have to actually print all the machines in this
	 * session, not just the host...
	 */
	return machine__fprintf(&session->host_machine, fp);
}

void perf_session__remove_thread(struct perf_session *session,
				 struct thread *th)
{
	/*
	 * FIXME: This one makes no sense, we need to remove the thread from
	 * the machine it belongs to, perf_session can have many machines, so
	 * doing it always on ->host_machine is wrong.  Fix when auditing all
	 * the 'perf kvm' code.
	 */
	machine__remove_thread(&session->host_machine, th);
}

1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506
struct perf_evsel *perf_session__find_first_evtype(struct perf_session *session,
					      unsigned int type)
{
	struct perf_evsel *pos;

	list_for_each_entry(pos, &session->evlist->entries, node) {
		if (pos->attr.type == type)
			return pos;
	}
	return NULL;
}

1507
void perf_event__print_ip(union perf_event *event, struct perf_sample *sample,
1508 1509
			  struct machine *machine, int print_sym,
			  int print_dso, int print_symoffset)
1510 1511 1512 1513
{
	struct addr_location al;
	struct callchain_cursor_node *node;

1514
	if (perf_event__preprocess_sample(event, machine, &al, sample,
1515 1516 1517 1518 1519 1520 1521 1522
					  NULL) < 0) {
		error("problem processing %d event, skipping it.\n",
			event->header.type);
		return;
	}

	if (symbol_conf.use_callchain && sample->callchain) {

1523
		if (machine__resolve_callchain(machine, al.thread,
1524 1525 1526 1527 1528
						sample->callchain, NULL) != 0) {
			if (verbose)
				error("Failed to resolve callchain. Skipping\n");
			return;
		}
1529
		callchain_cursor_commit(&callchain_cursor);
1530 1531

		while (1) {
1532
			node = callchain_cursor_current(&callchain_cursor);
1533 1534 1535
			if (!node)
				break;

1536 1537
			printf("\t%16" PRIx64, node->ip);
			if (print_sym) {
1538 1539
				printf(" ");
				symbol__fprintf_symname(node->sym, stdout);
1540 1541
			}
			if (print_dso) {
1542
				printf(" (");
1543
				map__fprintf_dsoname(node->map, stdout);
1544
				printf(")");
1545 1546
			}
			printf("\n");
1547

1548
			callchain_cursor_advance(&callchain_cursor);
1549 1550 1551
		}

	} else {
1552
		printf("%16" PRIx64, sample->ip);
1553
		if (print_sym) {
1554
			printf(" ");
1555 1556 1557 1558 1559
			if (print_symoffset)
				symbol__fprintf_symname_offs(al.sym, &al,
							     stdout);
			else
				symbol__fprintf_symname(al.sym, stdout);
1560 1561 1562
		}

		if (print_dso) {
1563 1564 1565
			printf(" (");
			map__fprintf_dsoname(al.map, stdout);
			printf(")");
1566
		}
1567 1568
	}
}
1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590

int perf_session__cpu_bitmap(struct perf_session *session,
			     const char *cpu_list, unsigned long *cpu_bitmap)
{
	int i;
	struct cpu_map *map;

	for (i = 0; i < PERF_TYPE_MAX; ++i) {
		struct perf_evsel *evsel;

		evsel = perf_session__find_first_evtype(session, i);
		if (!evsel)
			continue;

		if (!(evsel->attr.sample_type & PERF_SAMPLE_CPU)) {
			pr_err("File does not contain CPU events. "
			       "Remove -c option to proceed.\n");
			return -1;
		}
	}

	map = cpu_map__new(cpu_list);
1591 1592 1593 1594
	if (map == NULL) {
		pr_err("Invalid cpu_list\n");
		return -1;
	}
1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609

	for (i = 0; i < map->nr; i++) {
		int cpu = map->map[i];

		if (cpu >= MAX_NR_CPUS) {
			pr_err("Requested CPU %d too large. "
			       "Consider raising MAX_NR_CPUS\n", cpu);
			return -1;
		}

		set_bit(cpu, cpu_bitmap);
	}

	return 0;
}
1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628

void perf_session__fprintf_info(struct perf_session *session, FILE *fp,
				bool full)
{
	struct stat st;
	int ret;

	if (session == NULL || fp == NULL)
		return;

	ret = fstat(session->fd, &st);
	if (ret == -1)
		return;

	fprintf(fp, "# ========\n");
	fprintf(fp, "# captured on: %s", ctime(&st.st_ctime));
	perf_header__fprintf_info(session, fp, full);
	fprintf(fp, "# ========\n#\n");
}
1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683


int __perf_session__set_tracepoints_handlers(struct perf_session *session,
					     const struct perf_evsel_str_handler *assocs,
					     size_t nr_assocs)
{
	struct perf_evlist *evlist = session->evlist;
	struct event_format *format;
	struct perf_evsel *evsel;
	char *tracepoint, *name;
	size_t i;
	int err;

	for (i = 0; i < nr_assocs; i++) {
		err = -ENOMEM;
		tracepoint = strdup(assocs[i].name);
		if (tracepoint == NULL)
			goto out;

		err = -ENOENT;
		name = strchr(tracepoint, ':');
		if (name == NULL)
			goto out_free;

		*name++ = '\0';
		format = pevent_find_event_by_name(session->pevent,
						   tracepoint, name);
		if (format == NULL) {
			/*
			 * Adding a handler for an event not in the session,
			 * just ignore it.
			 */
			goto next;
		}

		evsel = perf_evlist__find_tracepoint_by_id(evlist, format->id);
		if (evsel == NULL)
			goto next;

		err = -EEXIST;
		if (evsel->handler.func != NULL)
			goto out_free;
		evsel->handler.func = assocs[i].handler;
next:
		free(tracepoint);
	}

	err = 0;
out:
	return err;

out_free:
	free(tracepoint);
	goto out;
}