builtin-record.c 24.4 KB
Newer Older
I
Ingo Molnar 已提交
1
/*
2 3 4 5 6
 * builtin-record.c
 *
 * Builtin record command: Record the profile of a workload
 * (or a CPU, or a PID) into the perf.data output file - for
 * later analysis via perf report.
I
Ingo Molnar 已提交
7
 */
8
#include "builtin.h"
9 10 11

#include "perf.h"

12
#include "util/build-id.h"
13
#include "util/util.h"
14
#include "util/parse-options.h"
15
#include "util/parse-events.h"
16

17
#include "util/header.h"
18
#include "util/event.h"
19
#include "util/evlist.h"
20
#include "util/evsel.h"
21
#include "util/debug.h"
22
#include "util/session.h"
23
#include "util/tool.h"
24
#include "util/symbol.h"
25
#include "util/cpumap.h"
26
#include "util/thread_map.h"
27
#include "util/data.h"
28

29
#include <unistd.h>
30
#include <sched.h>
31
#include <sys/mman.h>
32

33
#ifndef HAVE_ON_EXIT_SUPPORT
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64
#ifndef ATEXIT_MAX
#define ATEXIT_MAX 32
#endif
static int __on_exit_count = 0;
typedef void (*on_exit_func_t) (int, void *);
static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
static void *__on_exit_args[ATEXIT_MAX];
static int __exitcode = 0;
static void __handle_on_exit_funcs(void);
static int on_exit(on_exit_func_t function, void *arg);
#define exit(x) (exit)(__exitcode = (x))

static int on_exit(on_exit_func_t function, void *arg)
{
	if (__on_exit_count == ATEXIT_MAX)
		return -ENOMEM;
	else if (__on_exit_count == 0)
		atexit(__handle_on_exit_funcs);
	__on_exit_funcs[__on_exit_count] = function;
	__on_exit_args[__on_exit_count++] = arg;
	return 0;
}

static void __handle_on_exit_funcs(void)
{
	int i;
	for (i = 0; i < __on_exit_count; i++)
		__on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
}
#endif

65
struct record {
66
	struct perf_tool	tool;
67
	struct record_opts	opts;
68
	u64			bytes_written;
69
	struct perf_data_file	file;
70 71 72 73 74 75 76
	struct perf_evlist	*evlist;
	struct perf_session	*session;
	const char		*progname;
	int			realtime_prio;
	bool			no_buildid;
	bool			no_buildid_cache;
	long			samples;
77
};
78

79
static int record__write(struct record *rec, void *bf, size_t size)
80
{
81
	if (perf_data_file__write(rec->session->file, bf, size) < 0) {
82 83
		pr_err("failed to write perf data, error: %m\n");
		return -1;
84
	}
85

86
	rec->bytes_written += size;
87
	return 0;
88 89
}

90
static int process_synthesized_event(struct perf_tool *tool,
91
				     union perf_event *event,
92 93
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine __maybe_unused)
94
{
95 96
	struct record *rec = container_of(tool, struct record, tool);
	return record__write(rec, event, event->header.size);
97 98
}

99
static int record__mmap_read(struct record *rec, struct perf_mmap *md)
100
{
101
	unsigned int head = perf_mmap__read_head(md);
102
	unsigned int old = md->prev;
J
Jiri Olsa 已提交
103
	unsigned char *data = md->base + page_size;
104 105
	unsigned long size;
	void *buf;
106
	int rc = 0;
107

108
	if (old == head)
109
		return 0;
110

111
	rec->samples++;
112 113 114 115 116 117 118

	size = head - old;

	if ((old & md->mask) + size != (head & md->mask)) {
		buf = &data[old & md->mask];
		size = md->mask + 1 - (old & md->mask);
		old += size;
119

120
		if (record__write(rec, buf, size) < 0) {
121 122 123
			rc = -1;
			goto out;
		}
124 125 126 127 128
	}

	buf = &data[old & md->mask];
	size = head - old;
	old += size;
129

130
	if (record__write(rec, buf, size) < 0) {
131 132 133
		rc = -1;
		goto out;
	}
134 135

	md->prev = old;
136
	perf_mmap__write_tail(md, old);
137 138 139

out:
	return rc;
140 141 142
}

static volatile int done = 0;
143
static volatile int signr = -1;
144
static volatile int child_finished = 0;
145

146
static void sig_handler(int sig)
147
{
148 149 150
	if (sig == SIGCHLD)
		child_finished = 1;

151
	done = 1;
152 153 154
	signr = sig;
}

155
static void record__sig_exit(int exit_status __maybe_unused, void *arg)
156
{
157
	struct record *rec = arg;
158 159
	int status;

160
	if (rec->evlist->workload.pid > 0) {
161
		if (!child_finished)
162
			kill(rec->evlist->workload.pid, SIGTERM);
163 164 165

		wait(&status);
		if (WIFSIGNALED(status))
166
			psignal(WTERMSIG(status), rec->progname);
167
	}
168

169
	if (signr == -1 || signr == SIGUSR1)
170 171 172
		return;

	signal(signr, SIG_DFL);
173 174
}

175
static int record__open(struct record *rec)
176
{
177
	char msg[512];
178
	struct perf_evsel *pos;
179 180
	struct perf_evlist *evlist = rec->evlist;
	struct perf_session *session = rec->session;
181
	struct record_opts *opts = &rec->opts;
182
	int rc = 0;
183

184
	perf_evlist__config(evlist, opts);
185

186 187
	list_for_each_entry(pos, &evlist->entries, node) {
try_again:
188
		if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
189
			if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
190
				if (verbose)
191
					ui__warning("%s\n", msg);
192 193
				goto try_again;
			}
194

195 196 197 198
			rc = -errno;
			perf_evsel__open_strerror(pos, &opts->target,
						  errno, msg, sizeof(msg));
			ui__error("%s\n", msg);
199
			goto out;
L
Li Zefan 已提交
200 201
		}
	}
202

203
	if (perf_evlist__apply_filters(evlist)) {
204 205
		error("failed to set filter with %d (%s)\n", errno,
			strerror(errno));
206 207
		rc = -1;
		goto out;
208 209
	}

210
	if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
211 212 213 214 215
		if (errno == EPERM) {
			pr_err("Permission error mapping pages.\n"
			       "Consider increasing "
			       "/proc/sys/kernel/perf_event_mlock_kb,\n"
			       "or try again with a smaller value of -m/--mmap_pages.\n"
216
			       "(current value: %u)\n", opts->mmap_pages);
217 218 219 220 221 222
			rc = -errno;
		} else {
			pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
			rc = -errno;
		}
		goto out;
223
	}
224

225
	session->evlist = evlist;
226
	perf_session__set_id_hdr_size(session);
227 228
out:
	return rc;
229 230
}

231
static int process_buildids(struct record *rec)
232
{
233 234
	struct perf_data_file *file  = &rec->file;
	struct perf_session *session = rec->session;
235
	u64 start = session->header.data_offset;
236

237
	u64 size = lseek(file->fd, 0, SEEK_CUR);
238 239 240
	if (size == 0)
		return 0;

241 242
	return __perf_session__process_events(session, start,
					      size - start,
243 244 245
					      size, &build_id__mark_dso_hit_ops);
}

246
static void record__exit(int status, void *arg)
247
{
248
	struct record *rec = arg;
249
	struct perf_data_file *file = &rec->file;
250

251 252 253
	if (status != 0)
		return;

254
	if (!file->is_pipe) {
255 256 257 258 259
		rec->session->header.data_size += rec->bytes_written;

		if (!rec->no_buildid)
			process_buildids(rec);
		perf_session__write_header(rec->session, rec->evlist,
260
					   file->fd, true);
261 262
		perf_session__delete(rec->session);
		perf_evlist__delete(rec->evlist);
263
		symbol__exit();
264
	}
265 266
}

267
static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
268 269
{
	int err;
270
	struct perf_tool *tool = data;
271 272 273 274 275 276 277 278
	/*
	 *As for guest kernel when processing subcommand record&report,
	 *we arrange module mmap prior to guest kernel mmap and trigger
	 *a preload dso because default guest module symbols are loaded
	 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
	 *method is used to avoid symbol missing when the first addr is
	 *in module instead of in guest kernel.
	 */
279
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
280
					     machine);
281 282
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
283
		       " relocation symbol.\n", machine->pid);
284 285 286 287 288

	/*
	 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
	 * have no _text sometimes.
	 */
289
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
290
						 machine, "_text");
291
	if (err < 0)
292
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
293
							 machine, "_stext");
294 295
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
296
		       " relocation symbol.\n", machine->pid);
297 298
}

299 300 301 302 303
static struct perf_event_header finished_round_event = {
	.size = sizeof(struct perf_event_header),
	.type = PERF_RECORD_FINISHED_ROUND,
};

304
static int record__mmap_read_all(struct record *rec)
305
{
306
	int i;
307
	int rc = 0;
308

309
	for (i = 0; i < rec->evlist->nr_mmaps; i++) {
310
		if (rec->evlist->mmap[i].base) {
311
			if (record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
312 313 314 315
				rc = -1;
				goto out;
			}
		}
316 317
	}

318
	if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
319
		rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
320 321 322

out:
	return rc;
323 324
}

325
static void record__init_features(struct record *rec)
326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343
{
	struct perf_evlist *evsel_list = rec->evlist;
	struct perf_session *session = rec->session;
	int feat;

	for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
		perf_header__set_feat(&session->header, feat);

	if (rec->no_buildid)
		perf_header__clear_feat(&session->header, HEADER_BUILD_ID);

	if (!have_tracepoints(&evsel_list->entries))
		perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);

	if (!rec->opts.branch_stack)
		perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
}

344
static int __cmd_record(struct record *rec, int argc, const char **argv)
345
{
346
	int err;
347
	unsigned long waking = 0;
348
	const bool forks = argc > 0;
349
	struct machine *machine;
350
	struct perf_tool *tool = &rec->tool;
351
	struct record_opts *opts = &rec->opts;
352
	struct perf_evlist *evsel_list = rec->evlist;
353
	struct perf_data_file *file = &rec->file;
354
	struct perf_session *session;
355
	bool disabled = false;
356

357
	rec->progname = argv[0];
358

359
	on_exit(record__sig_exit, rec);
360 361
	signal(SIGCHLD, sig_handler);
	signal(SIGINT, sig_handler);
362
	signal(SIGUSR1, sig_handler);
363
	signal(SIGTERM, sig_handler);
364

365
	session = perf_session__new(file, false, NULL);
366
	if (session == NULL) {
367 368 369 370
		pr_err("Not enough memory for reading perf file header\n");
		return -1;
	}

371 372
	rec->session = session;

373
	record__init_features(rec);
374

375
	if (forks) {
376
		err = perf_evlist__prepare_workload(evsel_list, &opts->target,
377
						    argv, file->is_pipe,
378
						    true);
379 380 381
		if (err < 0) {
			pr_err("Couldn't run the workload!\n");
			goto out_delete_session;
382 383 384
		}
	}

385
	if (record__open(rec) != 0) {
386 387 388
		err = -1;
		goto out_delete_session;
	}
389

390 391 392
	if (!evsel_list->nr_groups)
		perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);

393
	/*
394
	 * perf_session__delete(session) will be called at record__exit()
395
	 */
396
	on_exit(record__exit, rec);
397

398 399
	if (file->is_pipe) {
		err = perf_header__write_pipe(file->fd);
400
		if (err < 0)
401
			goto out_delete_session;
402
	} else {
403
		err = perf_session__write_header(session, evsel_list,
404
						 file->fd, false);
405
		if (err < 0)
406
			goto out_delete_session;
407 408
	}

409
	if (!rec->no_buildid
410
	    && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
411
		pr_err("Couldn't generate buildids. "
412
		       "Use --no-buildid to profile anyway.\n");
413 414
		err = -1;
		goto out_delete_session;
415 416
	}

417
	machine = &session->machines.host;
418

419
	if (file->is_pipe) {
420
		err = perf_event__synthesize_attrs(tool, session,
421
						   process_synthesized_event);
422 423
		if (err < 0) {
			pr_err("Couldn't synthesize attrs.\n");
424
			goto out_delete_session;
425
		}
426

427
		if (have_tracepoints(&evsel_list->entries)) {
428 429 430 431 432 433 434 435
			/*
			 * FIXME err <= 0 here actually means that
			 * there were no tracepoints so its not really
			 * an error, just that we don't need to
			 * synthesize anything.  We really have to
			 * return this more properly and also
			 * propagate errors that now are calling die()
			 */
436
			err = perf_event__synthesize_tracing_data(tool, file->fd, evsel_list,
437
								  process_synthesized_event);
438 439
			if (err <= 0) {
				pr_err("Couldn't record tracing data.\n");
440
				goto out_delete_session;
441
			}
442
			rec->bytes_written += err;
443
		}
444 445
	}

446
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
447
						 machine, "_text");
448
	if (err < 0)
449
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
450
							 machine, "_stext");
451 452 453 454
	if (err < 0)
		pr_err("Couldn't record kernel reference relocation symbol\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/kallsyms permission or run as root.\n");
455

456
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
457
					     machine);
458 459 460 461 462
	if (err < 0)
		pr_err("Couldn't record kernel module information.\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/modules permission or run as root.\n");

463
	if (perf_guest) {
464 465
		machines__process_guests(&session->machines,
					 perf_event__synthesize_guest_os, tool);
466
	}
467

468 469
	err = __machine__synthesize_threads(machine, tool, &opts->target, evsel_list->threads,
					    process_synthesized_event, opts->sample_address);
470 471 472
	if (err != 0)
		goto out_delete_session;

473
	if (rec->realtime_prio) {
474 475
		struct sched_param param;

476
		param.sched_priority = rec->realtime_prio;
477
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
478
			pr_err("Could not set realtime priority.\n");
479 480
			err = -1;
			goto out_delete_session;
481 482 483
		}
	}

484 485 486 487 488
	/*
	 * When perf is starting the traced process, all the events
	 * (apart from group members) have enable_on_exec=1 set,
	 * so don't spoil it by prematurely enabling them.
	 */
489
	if (!target__none(&opts->target))
490
		perf_evlist__enable(evsel_list);
491

492 493 494
	/*
	 * Let the child rip
	 */
495
	if (forks)
496
		perf_evlist__start_workload(evsel_list);
497

498
	for (;;) {
499
		int hits = rec->samples;
500

501
		if (record__mmap_read_all(rec) < 0) {
502 503 504
			err = -1;
			goto out_delete_session;
		}
505

506
		if (hits == rec->samples) {
507 508
			if (done)
				break;
509
			err = poll(evsel_list->pollfd, evsel_list->nr_fds, -1);
510 511 512
			waking++;
		}

513 514 515 516 517
		/*
		 * When perf is starting the traced process, at the end events
		 * die with the process and we wait for that. Thus no need to
		 * disable events in this case.
		 */
518
		if (done && !disabled && !target__none(&opts->target)) {
519
			perf_evlist__disable(evsel_list);
520 521
			disabled = true;
		}
522 523
	}

524
	if (quiet || signr == SIGUSR1)
525 526
		return 0;

527 528
	fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);

529 530 531 532
	/*
	 * Approximate RIP event size: 24 bytes.
	 */
	fprintf(stderr,
533
		"[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
534
		(double)rec->bytes_written / 1024.0 / 1024.0,
535
		file->path,
536
		rec->bytes_written / 24);
537

538
	return 0;
539 540 541 542

out_delete_session:
	perf_session__delete(session);
	return err;
543
}
544

545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562
#define BRANCH_OPT(n, m) \
	{ .name = n, .mode = (m) }

#define BRANCH_END { .name = NULL }

struct branch_mode {
	const char *name;
	int mode;
};

static const struct branch_mode branch_modes[] = {
	BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
	BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
	BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
	BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
	BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
	BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
	BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
563 564 565
	BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
	BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
	BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
566 567 568 569
	BRANCH_END
};

static int
570
parse_branch_stack(const struct option *opt, const char *str, int unset)
571 572 573 574 575 576 577 578
{
#define ONLY_PLM \
	(PERF_SAMPLE_BRANCH_USER	|\
	 PERF_SAMPLE_BRANCH_KERNEL	|\
	 PERF_SAMPLE_BRANCH_HV)

	uint64_t *mode = (uint64_t *)opt->value;
	const struct branch_mode *br;
579
	char *s, *os = NULL, *p;
580 581
	int ret = -1;

582 583
	if (unset)
		return 0;
584

585 586 587 588
	/*
	 * cannot set it twice, -b + --branch-filter for instance
	 */
	if (*mode)
589 590
		return -1;

591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611
	/* str may be NULL in case no arg is passed to -b */
	if (str) {
		/* because str is read-only */
		s = os = strdup(str);
		if (!s)
			return -1;

		for (;;) {
			p = strchr(s, ',');
			if (p)
				*p = '\0';

			for (br = branch_modes; br->name; br++) {
				if (!strcasecmp(s, br->name))
					break;
			}
			if (!br->name) {
				ui__warning("unknown branch filter %s,"
					    " check man page\n", s);
				goto error;
			}
612

613
			*mode |= br->mode;
614

615 616
			if (!p)
				break;
617

618 619
			s = p + 1;
		}
620 621 622
	}
	ret = 0;

623
	/* default to any branch */
624
	if ((*mode & ~ONLY_PLM) == 0) {
625
		*mode = PERF_SAMPLE_BRANCH_ANY;
626 627 628 629 630 631
	}
error:
	free(os);
	return ret;
}

632
#ifdef HAVE_LIBUNWIND_SUPPORT
633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657
static int get_stack_size(char *str, unsigned long *_size)
{
	char *endptr;
	unsigned long size;
	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));

	size = strtoul(str, &endptr, 0);

	do {
		if (*endptr)
			break;

		size = round_up(size, sizeof(u64));
		if (!size || size > max_size)
			break;

		*_size = size;
		return 0;

	} while (0);

	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
	       max_size, str);
	return -1;
}
658
#endif /* HAVE_LIBUNWIND_SUPPORT */
659

660
int record_parse_callchain(const char *arg, struct record_opts *opts)
661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679
{
	char *tok, *name, *saveptr = NULL;
	char *buf;
	int ret = -1;

	/* We need buffer that we know we can write to. */
	buf = malloc(strlen(arg) + 1);
	if (!buf)
		return -ENOMEM;

	strcpy(buf, arg);

	tok = strtok_r((char *)buf, ",", &saveptr);
	name = tok ? : (char *)buf;

	do {
		/* Framepointer style */
		if (!strncmp(name, "fp", sizeof("fp"))) {
			if (!strtok_r(NULL, ",", &saveptr)) {
680
				opts->call_graph = CALLCHAIN_FP;
681 682 683 684 685 686
				ret = 0;
			} else
				pr_err("callchain: No more arguments "
				       "needed for -g fp\n");
			break;

687
#ifdef HAVE_LIBUNWIND_SUPPORT
688 689
		/* Dwarf style */
		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
690 691
			const unsigned long default_stack_dump_size = 8192;

692
			ret = 0;
693 694
			opts->call_graph = CALLCHAIN_DWARF;
			opts->stack_dump_size = default_stack_dump_size;
695 696 697 698 699 700

			tok = strtok_r(NULL, ",", &saveptr);
			if (tok) {
				unsigned long size = 0;

				ret = get_stack_size(tok, &size);
701
				opts->stack_dump_size = size;
702
			}
703
#endif /* HAVE_LIBUNWIND_SUPPORT */
704
		} else {
J
Jiri Olsa 已提交
705
			pr_err("callchain: Unknown --call-graph option "
706 707 708 709 710 711 712
			       "value: %s\n", arg);
			break;
		}

	} while (0);

	free(buf);
J
Jiri Olsa 已提交
713 714 715
	return ret;
}

716
static void callchain_debug(struct record_opts *opts)
J
Jiri Olsa 已提交
717 718
{
	pr_debug("callchain: type %d\n", opts->call_graph);
719

J
Jiri Olsa 已提交
720 721 722 723 724 725 726 727 728
	if (opts->call_graph == CALLCHAIN_DWARF)
		pr_debug("callchain: stack dump size %d\n",
			 opts->stack_dump_size);
}

int record_parse_callchain_opt(const struct option *opt,
			       const char *arg,
			       int unset)
{
729
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
730 731 732 733 734 735 736 737 738 739
	int ret;

	/* --no-call-graph */
	if (unset) {
		opts->call_graph = CALLCHAIN_NONE;
		pr_debug("callchain: disabled\n");
		return 0;
	}

	ret = record_parse_callchain(arg, opts);
740
	if (!ret)
J
Jiri Olsa 已提交
741
		callchain_debug(opts);
742 743 744 745

	return ret;
}

J
Jiri Olsa 已提交
746 747 748 749
int record_callchain_opt(const struct option *opt,
			 const char *arg __maybe_unused,
			 int unset __maybe_unused)
{
750
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
751 752 753 754 755 756 757 758

	if (opts->call_graph == CALLCHAIN_NONE)
		opts->call_graph = CALLCHAIN_FP;

	callchain_debug(opts);
	return 0;
}

759
static const char * const record_usage[] = {
760 761
	"perf record [<options>] [<command>]",
	"perf record [<options>] -- <command> [<options>]",
762 763 764
	NULL
};

765
/*
766 767
 * XXX Ideally would be local to cmd_record() and passed to a record__new
 * because we need to have access to it in record__exit, that is called
768 769 770 771 772 773 774
 * after cmd_record() exits, but since record_options need to be accessible to
 * builtin-script, leave it here.
 *
 * At least we don't ouch it in all the other functions here directly.
 *
 * Just say no to tons of global variables, sigh.
 */
775
static struct record record = {
776 777 778 779
	.opts = {
		.mmap_pages	     = UINT_MAX,
		.user_freq	     = UINT_MAX,
		.user_interval	     = ULLONG_MAX,
780
		.freq		     = 4000,
N
Namhyung Kim 已提交
781 782
		.target		     = {
			.uses_mmap   = true,
783
			.default_per_cpu = true,
N
Namhyung Kim 已提交
784
		},
785 786
	},
};
787

J
Jiri Olsa 已提交
788
#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
789

790
#ifdef HAVE_LIBUNWIND_SUPPORT
J
Jiri Olsa 已提交
791
const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
792
#else
J
Jiri Olsa 已提交
793
const char record_callchain_help[] = CALLCHAIN_HELP "fp";
794 795
#endif

796 797 798
/*
 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
 * with it and switch to use the library functions in perf_evlist that came
799
 * from builtin-record.c, i.e. use record_opts,
800 801 802
 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
 * using pipes, etc.
 */
803
const struct option record_options[] = {
804
	OPT_CALLBACK('e', "event", &record.evlist, "event",
805
		     "event selector. use 'perf list' to list available events",
806
		     parse_events_option),
807
	OPT_CALLBACK(0, "filter", &record.evlist, "filter",
L
Li Zefan 已提交
808
		     "event filter", parse_filter),
809
	OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
810
		    "record events on existing process id"),
811
	OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
812
		    "record events on existing thread id"),
813
	OPT_INTEGER('r', "realtime", &record.realtime_prio,
814
		    "collect data with this RT SCHED_FIFO priority"),
815
	OPT_BOOLEAN('D', "no-delay", &record.opts.no_delay,
816
		    "collect data without buffering"),
817
	OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
818
		    "collect raw sample records from all opened counters"),
819
	OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
820
			    "system-wide collection from all CPUs"),
821
	OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
822
		    "list of cpus to monitor"),
823
	OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
824
	OPT_STRING('o', "output", &record.file.path, "file",
I
Ingo Molnar 已提交
825
		    "output file name"),
826 827 828
	OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
			&record.opts.no_inherit_set,
			"child tasks do not inherit counters"),
829
	OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
830 831 832
	OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
		     "number of mmap data pages",
		     perf_evlist__parse_mmap_pages),
833
	OPT_BOOLEAN(0, "group", &record.opts.group,
834
		    "put the counters into a counter group"),
J
Jiri Olsa 已提交
835 836 837 838 839 840
	OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
			   NULL, "enables call-graph recording" ,
			   &record_callchain_opt),
	OPT_CALLBACK(0, "call-graph", &record.opts,
		     "mode[,dump_size]", record_callchain_help,
		     &record_parse_callchain_opt),
841
	OPT_INCR('v', "verbose", &verbose,
842
		    "be more verbose (show counter open errors, etc)"),
843
	OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
844
	OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
845
		    "per thread counts"),
846
	OPT_BOOLEAN('d', "data", &record.opts.sample_address,
847
		    "Sample addresses"),
848
	OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
849
	OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
850
	OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
851
		    "don't sample"),
852
	OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
853
		    "do not update the buildid cache"),
854
	OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
855
		    "do not collect buildids in perf.data"),
856
	OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
S
Stephane Eranian 已提交
857 858
		     "monitor event in cgroup name only",
		     parse_cgroups),
859 860
	OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
		   "user to profile"),
861 862 863 864 865 866 867

	OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
		     "branch any", "sample any taken branches",
		     parse_branch_stack),

	OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
		     "branch filter mask", "branch stack filter modes",
868
		     parse_branch_stack),
869 870
	OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
		    "sample by weight (on special events only)"),
871 872
	OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
		    "sample transaction flags (special events only)"),
873 874
	OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
		    "use per-thread mmaps"),
875 876 877
	OPT_END()
};

878
int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
879
{
880
	int err = -ENOMEM;
881
	struct perf_evlist *evsel_list;
882
	struct record *rec = &record;
883
	char errbuf[BUFSIZ];
884

885
	evsel_list = perf_evlist__new();
886 887 888
	if (evsel_list == NULL)
		return -ENOMEM;

889 890
	rec->evlist = evsel_list;

891
	argc = parse_options(argc, argv, record_options, record_usage,
892
			    PARSE_OPT_STOP_AT_NON_OPTION);
893
	if (!argc && target__none(&rec->opts.target))
894
		usage_with_options(record_usage, record_options);
895

896
	if (nr_cgroups && !rec->opts.target.system_wide) {
897 898
		ui__error("cgroup monitoring only available in"
			  " system-wide mode\n");
S
Stephane Eranian 已提交
899 900 901
		usage_with_options(record_usage, record_options);
	}

902
	symbol__init();
903

904
	if (symbol_conf.kptr_restrict)
905 906 907 908 909 910 911 912
		pr_warning(
"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
"check /proc/sys/kernel/kptr_restrict.\n\n"
"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
"file is not found in the buildid cache or in the vmlinux path.\n\n"
"Samples in kernel modules won't be resolved at all.\n\n"
"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
"even with a suitable vmlinux or kallsyms file.\n\n");
913

914
	if (rec->no_buildid_cache || rec->no_buildid)
915
		disable_buildid_cache();
916

917 918
	if (evsel_list->nr_entries == 0 &&
	    perf_evlist__add_default(evsel_list) < 0) {
919 920
		pr_err("Not enough memory for event selector list\n");
		goto out_symbol_exit;
921
	}
922

923 924 925
	if (rec->opts.target.tid && !rec->opts.no_inherit_set)
		rec->opts.no_inherit = true;

926
	err = target__validate(&rec->opts.target);
927
	if (err) {
928
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
929 930 931
		ui__warning("%s", errbuf);
	}

932
	err = target__parse_uid(&rec->opts.target);
933 934
	if (err) {
		int saved_errno = errno;
935

936
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
937
		ui__error("%s", errbuf);
938 939

		err = -saved_errno;
940
		goto out_symbol_exit;
941
	}
942

943
	err = -ENOMEM;
944
	if (perf_evlist__create_maps(evsel_list, &rec->opts.target) < 0)
945
		usage_with_options(record_usage, record_options);
946

947
	if (record_opts__config(&rec->opts)) {
948
		err = -EINVAL;
949
		goto out_free_fd;
950 951
	}

952
	err = __cmd_record(&record, argc, argv);
953 954 955

	perf_evlist__munmap(evsel_list);
	perf_evlist__close(evsel_list);
956
out_free_fd:
957
	perf_evlist__delete_maps(evsel_list);
958 959
out_symbol_exit:
	symbol__exit();
960
	return err;
961
}