builtin-record.c 24.9 KB
Newer Older
I
Ingo Molnar 已提交
1
/*
2 3 4 5 6
 * builtin-record.c
 *
 * Builtin record command: Record the profile of a workload
 * (or a CPU, or a PID) into the perf.data output file - for
 * later analysis via perf report.
I
Ingo Molnar 已提交
7
 */
8
#include "builtin.h"
9 10 11

#include "perf.h"

12
#include "util/build-id.h"
13
#include "util/util.h"
14
#include "util/parse-options.h"
15
#include "util/parse-events.h"
16

17
#include "util/header.h"
18
#include "util/event.h"
19
#include "util/evlist.h"
20
#include "util/evsel.h"
21
#include "util/debug.h"
22
#include "util/session.h"
23
#include "util/tool.h"
24
#include "util/symbol.h"
25
#include "util/cpumap.h"
26
#include "util/thread_map.h"
27
#include "util/data.h"
28

29
#include <unistd.h>
30
#include <sched.h>
31
#include <sys/mman.h>
32

33
#ifndef HAVE_ON_EXIT_SUPPORT
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64
#ifndef ATEXIT_MAX
#define ATEXIT_MAX 32
#endif
static int __on_exit_count = 0;
typedef void (*on_exit_func_t) (int, void *);
static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
static void *__on_exit_args[ATEXIT_MAX];
static int __exitcode = 0;
static void __handle_on_exit_funcs(void);
static int on_exit(on_exit_func_t function, void *arg);
#define exit(x) (exit)(__exitcode = (x))

static int on_exit(on_exit_func_t function, void *arg)
{
	if (__on_exit_count == ATEXIT_MAX)
		return -ENOMEM;
	else if (__on_exit_count == 0)
		atexit(__handle_on_exit_funcs);
	__on_exit_funcs[__on_exit_count] = function;
	__on_exit_args[__on_exit_count++] = arg;
	return 0;
}

static void __handle_on_exit_funcs(void)
{
	int i;
	for (i = 0; i < __on_exit_count; i++)
		__on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
}
#endif

65
struct record {
66
	struct perf_tool	tool;
67
	struct record_opts	opts;
68
	u64			bytes_written;
69
	struct perf_data_file	file;
70 71 72 73 74 75 76
	struct perf_evlist	*evlist;
	struct perf_session	*session;
	const char		*progname;
	int			realtime_prio;
	bool			no_buildid;
	bool			no_buildid_cache;
	long			samples;
77
};
78

79
static int record__write(struct record *rec, void *bf, size_t size)
80
{
81
	if (perf_data_file__write(rec->session->file, bf, size) < 0) {
82 83
		pr_err("failed to write perf data, error: %m\n");
		return -1;
84
	}
85

86
	rec->bytes_written += size;
87
	return 0;
88 89
}

90
static int process_synthesized_event(struct perf_tool *tool,
91
				     union perf_event *event,
92 93
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine __maybe_unused)
94
{
95 96
	struct record *rec = container_of(tool, struct record, tool);
	return record__write(rec, event, event->header.size);
97 98
}

99
static int record__mmap_read(struct record *rec, struct perf_mmap *md)
100
{
101
	unsigned int head = perf_mmap__read_head(md);
102
	unsigned int old = md->prev;
J
Jiri Olsa 已提交
103
	unsigned char *data = md->base + page_size;
104 105
	unsigned long size;
	void *buf;
106
	int rc = 0;
107

108
	if (old == head)
109
		return 0;
110

111
	rec->samples++;
112 113 114 115 116 117 118

	size = head - old;

	if ((old & md->mask) + size != (head & md->mask)) {
		buf = &data[old & md->mask];
		size = md->mask + 1 - (old & md->mask);
		old += size;
119

120
		if (record__write(rec, buf, size) < 0) {
121 122 123
			rc = -1;
			goto out;
		}
124 125 126 127 128
	}

	buf = &data[old & md->mask];
	size = head - old;
	old += size;
129

130
	if (record__write(rec, buf, size) < 0) {
131 132 133
		rc = -1;
		goto out;
	}
134 135

	md->prev = old;
136
	perf_mmap__write_tail(md, old);
137 138 139

out:
	return rc;
140 141 142
}

static volatile int done = 0;
143
static volatile int signr = -1;
144
static volatile int child_finished = 0;
145

146
static void sig_handler(int sig)
147
{
148 149 150
	if (sig == SIGCHLD)
		child_finished = 1;

151
	done = 1;
152 153 154
	signr = sig;
}

155
static void record__sig_exit(int exit_status __maybe_unused, void *arg)
156
{
157
	struct record *rec = arg;
158 159
	int status;

160
	if (rec->evlist->workload.pid > 0) {
161
		if (!child_finished)
162
			kill(rec->evlist->workload.pid, SIGTERM);
163 164 165

		wait(&status);
		if (WIFSIGNALED(status))
166
			psignal(WTERMSIG(status), rec->progname);
167
	}
168

169
	if (signr == -1 || signr == SIGUSR1)
170 171 172
		return;

	signal(signr, SIG_DFL);
173 174
}

175
static int record__open(struct record *rec)
176
{
177
	char msg[512];
178
	struct perf_evsel *pos;
179 180
	struct perf_evlist *evlist = rec->evlist;
	struct perf_session *session = rec->session;
181
	struct record_opts *opts = &rec->opts;
182
	int rc = 0;
183

184
	perf_evlist__config(evlist, opts);
185

186 187
	list_for_each_entry(pos, &evlist->entries, node) {
try_again:
188
		if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
189
			if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
190
				if (verbose)
191
					ui__warning("%s\n", msg);
192 193
				goto try_again;
			}
194

195 196 197 198
			rc = -errno;
			perf_evsel__open_strerror(pos, &opts->target,
						  errno, msg, sizeof(msg));
			ui__error("%s\n", msg);
199
			goto out;
L
Li Zefan 已提交
200 201
		}
	}
202

203
	if (perf_evlist__apply_filters(evlist)) {
204 205
		error("failed to set filter with %d (%s)\n", errno,
			strerror(errno));
206 207
		rc = -1;
		goto out;
208 209
	}

210
	if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
211 212 213 214 215
		if (errno == EPERM) {
			pr_err("Permission error mapping pages.\n"
			       "Consider increasing "
			       "/proc/sys/kernel/perf_event_mlock_kb,\n"
			       "or try again with a smaller value of -m/--mmap_pages.\n"
216
			       "(current value: %u)\n", opts->mmap_pages);
217 218 219 220 221 222
			rc = -errno;
		} else {
			pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
			rc = -errno;
		}
		goto out;
223
	}
224

225
	session->evlist = evlist;
226
	perf_session__set_id_hdr_size(session);
227 228
out:
	return rc;
229 230
}

231
static int process_buildids(struct record *rec)
232
{
233 234
	struct perf_data_file *file  = &rec->file;
	struct perf_session *session = rec->session;
235
	u64 start = session->header.data_offset;
236

237
	u64 size = lseek(file->fd, 0, SEEK_CUR);
238 239 240
	if (size == 0)
		return 0;

241 242
	return __perf_session__process_events(session, start,
					      size - start,
243 244 245
					      size, &build_id__mark_dso_hit_ops);
}

246
static void record__exit(int status, void *arg)
247
{
248
	struct record *rec = arg;
249
	struct perf_data_file *file = &rec->file;
250

251 252 253
	if (status != 0)
		return;

254
	if (!file->is_pipe) {
255 256 257 258 259
		rec->session->header.data_size += rec->bytes_written;

		if (!rec->no_buildid)
			process_buildids(rec);
		perf_session__write_header(rec->session, rec->evlist,
260
					   file->fd, true);
261 262
		perf_session__delete(rec->session);
		perf_evlist__delete(rec->evlist);
263
		symbol__exit();
264
	}
265 266
}

267
static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
268 269
{
	int err;
270
	struct perf_tool *tool = data;
271 272 273 274 275 276 277 278
	/*
	 *As for guest kernel when processing subcommand record&report,
	 *we arrange module mmap prior to guest kernel mmap and trigger
	 *a preload dso because default guest module symbols are loaded
	 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
	 *method is used to avoid symbol missing when the first addr is
	 *in module instead of in guest kernel.
	 */
279
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
280
					     machine);
281 282
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
283
		       " relocation symbol.\n", machine->pid);
284 285 286 287 288

	/*
	 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
	 * have no _text sometimes.
	 */
289
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
290
						 machine, "_text");
291
	if (err < 0)
292
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
293
							 machine, "_stext");
294 295
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
296
		       " relocation symbol.\n", machine->pid);
297 298
}

299 300 301 302 303
static struct perf_event_header finished_round_event = {
	.size = sizeof(struct perf_event_header),
	.type = PERF_RECORD_FINISHED_ROUND,
};

304
static int record__mmap_read_all(struct record *rec)
305
{
306
	int i;
307
	int rc = 0;
308

309
	for (i = 0; i < rec->evlist->nr_mmaps; i++) {
310
		if (rec->evlist->mmap[i].base) {
311
			if (record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
312 313 314 315
				rc = -1;
				goto out;
			}
		}
316 317
	}

318
	if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
319
		rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
320 321 322

out:
	return rc;
323 324
}

325
static void record__init_features(struct record *rec)
326 327 328 329 330 331 332 333 334 335
{
	struct perf_session *session = rec->session;
	int feat;

	for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
		perf_header__set_feat(&session->header, feat);

	if (rec->no_buildid)
		perf_header__clear_feat(&session->header, HEADER_BUILD_ID);

336
	if (!have_tracepoints(&rec->evlist->entries))
337 338 339 340 341 342
		perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);

	if (!rec->opts.branch_stack)
		perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
}

343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358
static volatile int workload_exec_errno;

/*
 * perf_evlist__prepare_workload will send a SIGUSR1
 * if the fork fails, since we asked by setting its
 * want_signal to true.
 */
static void workload_exec_failed_signal(int signo, siginfo_t *info,
					void *ucontext __maybe_unused)
{
	workload_exec_errno = info->si_value.sival_int;
	done = 1;
	signr = signo;
	child_finished = 1;
}

359
static int __cmd_record(struct record *rec, int argc, const char **argv)
360
{
361
	int err;
362
	unsigned long waking = 0;
363
	const bool forks = argc > 0;
364
	struct machine *machine;
365
	struct perf_tool *tool = &rec->tool;
366
	struct record_opts *opts = &rec->opts;
367
	struct perf_data_file *file = &rec->file;
368
	struct perf_session *session;
369
	bool disabled = false;
370

371
	rec->progname = argv[0];
372

373
	on_exit(record__sig_exit, rec);
374 375
	signal(SIGCHLD, sig_handler);
	signal(SIGINT, sig_handler);
376
	signal(SIGTERM, sig_handler);
377

378
	session = perf_session__new(file, false, NULL);
379
	if (session == NULL) {
380 381 382 383
		pr_err("Not enough memory for reading perf file header\n");
		return -1;
	}

384 385
	rec->session = session;

386
	record__init_features(rec);
387

388
	if (forks) {
389
		err = perf_evlist__prepare_workload(rec->evlist, &opts->target,
390
						    argv, file->is_pipe,
391
						    workload_exec_failed_signal);
392 393 394
		if (err < 0) {
			pr_err("Couldn't run the workload!\n");
			goto out_delete_session;
395 396 397
		}
	}

398
	if (record__open(rec) != 0) {
399 400 401
		err = -1;
		goto out_delete_session;
	}
402

403
	if (!rec->evlist->nr_groups)
404 405
		perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);

406
	/*
407
	 * perf_session__delete(session) will be called at record__exit()
408
	 */
409
	on_exit(record__exit, rec);
410

411 412
	if (file->is_pipe) {
		err = perf_header__write_pipe(file->fd);
413
		if (err < 0)
414
			goto out_delete_session;
415
	} else {
416
		err = perf_session__write_header(session, rec->evlist,
417
						 file->fd, false);
418
		if (err < 0)
419
			goto out_delete_session;
420 421
	}

422
	if (!rec->no_buildid
423
	    && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
424
		pr_err("Couldn't generate buildids. "
425
		       "Use --no-buildid to profile anyway.\n");
426 427
		err = -1;
		goto out_delete_session;
428 429
	}

430
	machine = &session->machines.host;
431

432
	if (file->is_pipe) {
433
		err = perf_event__synthesize_attrs(tool, session,
434
						   process_synthesized_event);
435 436
		if (err < 0) {
			pr_err("Couldn't synthesize attrs.\n");
437
			goto out_delete_session;
438
		}
439

440
		if (have_tracepoints(&rec->evlist->entries)) {
441 442 443 444 445 446 447 448
			/*
			 * FIXME err <= 0 here actually means that
			 * there were no tracepoints so its not really
			 * an error, just that we don't need to
			 * synthesize anything.  We really have to
			 * return this more properly and also
			 * propagate errors that now are calling die()
			 */
449
			err = perf_event__synthesize_tracing_data(tool, file->fd, rec->evlist,
450
								  process_synthesized_event);
451 452
			if (err <= 0) {
				pr_err("Couldn't record tracing data.\n");
453
				goto out_delete_session;
454
			}
455
			rec->bytes_written += err;
456
		}
457 458
	}

459
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
460
						 machine, "_text");
461
	if (err < 0)
462
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
463
							 machine, "_stext");
464 465 466 467
	if (err < 0)
		pr_err("Couldn't record kernel reference relocation symbol\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/kallsyms permission or run as root.\n");
468

469
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
470
					     machine);
471 472 473 474 475
	if (err < 0)
		pr_err("Couldn't record kernel module information.\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/modules permission or run as root.\n");

476
	if (perf_guest) {
477 478
		machines__process_guests(&session->machines,
					 perf_event__synthesize_guest_os, tool);
479
	}
480

481
	err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
482
					    process_synthesized_event, opts->sample_address);
483 484 485
	if (err != 0)
		goto out_delete_session;

486
	if (rec->realtime_prio) {
487 488
		struct sched_param param;

489
		param.sched_priority = rec->realtime_prio;
490
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
491
			pr_err("Could not set realtime priority.\n");
492 493
			err = -1;
			goto out_delete_session;
494 495 496
		}
	}

497 498 499 500 501
	/*
	 * When perf is starting the traced process, all the events
	 * (apart from group members) have enable_on_exec=1 set,
	 * so don't spoil it by prematurely enabling them.
	 */
502
	if (!target__none(&opts->target))
503
		perf_evlist__enable(rec->evlist);
504

505 506 507
	/*
	 * Let the child rip
	 */
508
	if (forks)
509
		perf_evlist__start_workload(rec->evlist);
510

511
	for (;;) {
512
		int hits = rec->samples;
513

514
		if (record__mmap_read_all(rec) < 0) {
515 516 517
			err = -1;
			goto out_delete_session;
		}
518

519
		if (hits == rec->samples) {
520 521
			if (done)
				break;
522
			err = poll(rec->evlist->pollfd, rec->evlist->nr_fds, -1);
523 524 525
			waking++;
		}

526 527 528 529 530
		/*
		 * When perf is starting the traced process, at the end events
		 * die with the process and we wait for that. Thus no need to
		 * disable events in this case.
		 */
531
		if (done && !disabled && !target__none(&opts->target)) {
532
			perf_evlist__disable(rec->evlist);
533 534
			disabled = true;
		}
535 536
	}

537 538 539 540 541 542 543 544
	if (forks && workload_exec_errno) {
		char msg[512];
		const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
		pr_err("Workload failed: %s\n", emsg);
		err = -1;
		goto out_delete_session;
	}

545
	if (quiet || signr == SIGUSR1)
546 547
		return 0;

548 549
	fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);

550 551 552 553
	/*
	 * Approximate RIP event size: 24 bytes.
	 */
	fprintf(stderr,
554
		"[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
555
		(double)rec->bytes_written / 1024.0 / 1024.0,
556
		file->path,
557
		rec->bytes_written / 24);
558

559
	return 0;
560 561 562 563

out_delete_session:
	perf_session__delete(session);
	return err;
564
}
565

566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583
#define BRANCH_OPT(n, m) \
	{ .name = n, .mode = (m) }

#define BRANCH_END { .name = NULL }

struct branch_mode {
	const char *name;
	int mode;
};

static const struct branch_mode branch_modes[] = {
	BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
	BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
	BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
	BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
	BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
	BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
	BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
584 585 586
	BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
	BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
	BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
587 588 589 590
	BRANCH_END
};

static int
591
parse_branch_stack(const struct option *opt, const char *str, int unset)
592 593 594 595 596 597 598 599
{
#define ONLY_PLM \
	(PERF_SAMPLE_BRANCH_USER	|\
	 PERF_SAMPLE_BRANCH_KERNEL	|\
	 PERF_SAMPLE_BRANCH_HV)

	uint64_t *mode = (uint64_t *)opt->value;
	const struct branch_mode *br;
600
	char *s, *os = NULL, *p;
601 602
	int ret = -1;

603 604
	if (unset)
		return 0;
605

606 607 608 609
	/*
	 * cannot set it twice, -b + --branch-filter for instance
	 */
	if (*mode)
610 611
		return -1;

612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632
	/* str may be NULL in case no arg is passed to -b */
	if (str) {
		/* because str is read-only */
		s = os = strdup(str);
		if (!s)
			return -1;

		for (;;) {
			p = strchr(s, ',');
			if (p)
				*p = '\0';

			for (br = branch_modes; br->name; br++) {
				if (!strcasecmp(s, br->name))
					break;
			}
			if (!br->name) {
				ui__warning("unknown branch filter %s,"
					    " check man page\n", s);
				goto error;
			}
633

634
			*mode |= br->mode;
635

636 637
			if (!p)
				break;
638

639 640
			s = p + 1;
		}
641 642 643
	}
	ret = 0;

644
	/* default to any branch */
645
	if ((*mode & ~ONLY_PLM) == 0) {
646
		*mode = PERF_SAMPLE_BRANCH_ANY;
647 648 649 650 651 652
	}
error:
	free(os);
	return ret;
}

653
#ifdef HAVE_LIBUNWIND_SUPPORT
654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678
static int get_stack_size(char *str, unsigned long *_size)
{
	char *endptr;
	unsigned long size;
	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));

	size = strtoul(str, &endptr, 0);

	do {
		if (*endptr)
			break;

		size = round_up(size, sizeof(u64));
		if (!size || size > max_size)
			break;

		*_size = size;
		return 0;

	} while (0);

	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
	       max_size, str);
	return -1;
}
679
#endif /* HAVE_LIBUNWIND_SUPPORT */
680

681
int record_parse_callchain(const char *arg, struct record_opts *opts)
682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700
{
	char *tok, *name, *saveptr = NULL;
	char *buf;
	int ret = -1;

	/* We need buffer that we know we can write to. */
	buf = malloc(strlen(arg) + 1);
	if (!buf)
		return -ENOMEM;

	strcpy(buf, arg);

	tok = strtok_r((char *)buf, ",", &saveptr);
	name = tok ? : (char *)buf;

	do {
		/* Framepointer style */
		if (!strncmp(name, "fp", sizeof("fp"))) {
			if (!strtok_r(NULL, ",", &saveptr)) {
701
				opts->call_graph = CALLCHAIN_FP;
702 703 704 705 706 707
				ret = 0;
			} else
				pr_err("callchain: No more arguments "
				       "needed for -g fp\n");
			break;

708
#ifdef HAVE_LIBUNWIND_SUPPORT
709 710
		/* Dwarf style */
		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
711 712
			const unsigned long default_stack_dump_size = 8192;

713
			ret = 0;
714 715
			opts->call_graph = CALLCHAIN_DWARF;
			opts->stack_dump_size = default_stack_dump_size;
716 717 718 719 720 721

			tok = strtok_r(NULL, ",", &saveptr);
			if (tok) {
				unsigned long size = 0;

				ret = get_stack_size(tok, &size);
722
				opts->stack_dump_size = size;
723
			}
724
#endif /* HAVE_LIBUNWIND_SUPPORT */
725
		} else {
J
Jiri Olsa 已提交
726
			pr_err("callchain: Unknown --call-graph option "
727 728 729 730 731 732 733
			       "value: %s\n", arg);
			break;
		}

	} while (0);

	free(buf);
J
Jiri Olsa 已提交
734 735 736
	return ret;
}

737
static void callchain_debug(struct record_opts *opts)
J
Jiri Olsa 已提交
738 739
{
	pr_debug("callchain: type %d\n", opts->call_graph);
740

J
Jiri Olsa 已提交
741 742 743 744 745 746 747 748 749
	if (opts->call_graph == CALLCHAIN_DWARF)
		pr_debug("callchain: stack dump size %d\n",
			 opts->stack_dump_size);
}

int record_parse_callchain_opt(const struct option *opt,
			       const char *arg,
			       int unset)
{
750
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
751 752 753 754 755 756 757 758 759 760
	int ret;

	/* --no-call-graph */
	if (unset) {
		opts->call_graph = CALLCHAIN_NONE;
		pr_debug("callchain: disabled\n");
		return 0;
	}

	ret = record_parse_callchain(arg, opts);
761
	if (!ret)
J
Jiri Olsa 已提交
762
		callchain_debug(opts);
763 764 765 766

	return ret;
}

J
Jiri Olsa 已提交
767 768 769 770
int record_callchain_opt(const struct option *opt,
			 const char *arg __maybe_unused,
			 int unset __maybe_unused)
{
771
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
772 773 774 775 776 777 778 779

	if (opts->call_graph == CALLCHAIN_NONE)
		opts->call_graph = CALLCHAIN_FP;

	callchain_debug(opts);
	return 0;
}

780
static const char * const record_usage[] = {
781 782
	"perf record [<options>] [<command>]",
	"perf record [<options>] -- <command> [<options>]",
783 784 785
	NULL
};

786
/*
787 788
 * XXX Ideally would be local to cmd_record() and passed to a record__new
 * because we need to have access to it in record__exit, that is called
789 790 791 792 793 794 795
 * after cmd_record() exits, but since record_options need to be accessible to
 * builtin-script, leave it here.
 *
 * At least we don't ouch it in all the other functions here directly.
 *
 * Just say no to tons of global variables, sigh.
 */
796
static struct record record = {
797 798 799 800
	.opts = {
		.mmap_pages	     = UINT_MAX,
		.user_freq	     = UINT_MAX,
		.user_interval	     = ULLONG_MAX,
801
		.freq		     = 4000,
N
Namhyung Kim 已提交
802 803
		.target		     = {
			.uses_mmap   = true,
804
			.default_per_cpu = true,
N
Namhyung Kim 已提交
805
		},
806 807
	},
};
808

J
Jiri Olsa 已提交
809
#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
810

811
#ifdef HAVE_LIBUNWIND_SUPPORT
J
Jiri Olsa 已提交
812
const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
813
#else
J
Jiri Olsa 已提交
814
const char record_callchain_help[] = CALLCHAIN_HELP "fp";
815 816
#endif

817 818 819
/*
 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
 * with it and switch to use the library functions in perf_evlist that came
820
 * from builtin-record.c, i.e. use record_opts,
821 822 823
 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
 * using pipes, etc.
 */
824
const struct option record_options[] = {
825
	OPT_CALLBACK('e', "event", &record.evlist, "event",
826
		     "event selector. use 'perf list' to list available events",
827
		     parse_events_option),
828
	OPT_CALLBACK(0, "filter", &record.evlist, "filter",
L
Li Zefan 已提交
829
		     "event filter", parse_filter),
830
	OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
831
		    "record events on existing process id"),
832
	OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
833
		    "record events on existing thread id"),
834
	OPT_INTEGER('r', "realtime", &record.realtime_prio,
835
		    "collect data with this RT SCHED_FIFO priority"),
836
	OPT_BOOLEAN('D', "no-delay", &record.opts.no_delay,
837
		    "collect data without buffering"),
838
	OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
839
		    "collect raw sample records from all opened counters"),
840
	OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
841
			    "system-wide collection from all CPUs"),
842
	OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
843
		    "list of cpus to monitor"),
844
	OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
845
	OPT_STRING('o', "output", &record.file.path, "file",
I
Ingo Molnar 已提交
846
		    "output file name"),
847 848 849
	OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
			&record.opts.no_inherit_set,
			"child tasks do not inherit counters"),
850
	OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
851 852 853
	OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
		     "number of mmap data pages",
		     perf_evlist__parse_mmap_pages),
854
	OPT_BOOLEAN(0, "group", &record.opts.group,
855
		    "put the counters into a counter group"),
J
Jiri Olsa 已提交
856 857 858 859 860 861
	OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
			   NULL, "enables call-graph recording" ,
			   &record_callchain_opt),
	OPT_CALLBACK(0, "call-graph", &record.opts,
		     "mode[,dump_size]", record_callchain_help,
		     &record_parse_callchain_opt),
862
	OPT_INCR('v', "verbose", &verbose,
863
		    "be more verbose (show counter open errors, etc)"),
864
	OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
865
	OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
866
		    "per thread counts"),
867
	OPT_BOOLEAN('d', "data", &record.opts.sample_address,
868
		    "Sample addresses"),
869
	OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
870
	OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
871
	OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
872
		    "don't sample"),
873
	OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
874
		    "do not update the buildid cache"),
875
	OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
876
		    "do not collect buildids in perf.data"),
877
	OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
S
Stephane Eranian 已提交
878 879
		     "monitor event in cgroup name only",
		     parse_cgroups),
880 881
	OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
		   "user to profile"),
882 883 884 885 886 887 888

	OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
		     "branch any", "sample any taken branches",
		     parse_branch_stack),

	OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
		     "branch filter mask", "branch stack filter modes",
889
		     parse_branch_stack),
890 891
	OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
		    "sample by weight (on special events only)"),
892 893
	OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
		    "sample transaction flags (special events only)"),
894 895
	OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
		    "use per-thread mmaps"),
896 897 898
	OPT_END()
};

899
int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
900
{
901
	int err = -ENOMEM;
902
	struct record *rec = &record;
903
	char errbuf[BUFSIZ];
904

905 906
	rec->evlist = perf_evlist__new();
	if (rec->evlist == NULL)
907 908
		return -ENOMEM;

909
	argc = parse_options(argc, argv, record_options, record_usage,
910
			    PARSE_OPT_STOP_AT_NON_OPTION);
911
	if (!argc && target__none(&rec->opts.target))
912
		usage_with_options(record_usage, record_options);
913

914
	if (nr_cgroups && !rec->opts.target.system_wide) {
915 916
		ui__error("cgroup monitoring only available in"
			  " system-wide mode\n");
S
Stephane Eranian 已提交
917 918 919
		usage_with_options(record_usage, record_options);
	}

920
	symbol__init();
921

922
	if (symbol_conf.kptr_restrict)
923 924 925 926 927 928 929 930
		pr_warning(
"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
"check /proc/sys/kernel/kptr_restrict.\n\n"
"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
"file is not found in the buildid cache or in the vmlinux path.\n\n"
"Samples in kernel modules won't be resolved at all.\n\n"
"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
"even with a suitable vmlinux or kallsyms file.\n\n");
931

932
	if (rec->no_buildid_cache || rec->no_buildid)
933
		disable_buildid_cache();
934

935 936
	if (rec->evlist->nr_entries == 0 &&
	    perf_evlist__add_default(rec->evlist) < 0) {
937 938
		pr_err("Not enough memory for event selector list\n");
		goto out_symbol_exit;
939
	}
940

941 942 943
	if (rec->opts.target.tid && !rec->opts.no_inherit_set)
		rec->opts.no_inherit = true;

944
	err = target__validate(&rec->opts.target);
945
	if (err) {
946
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
947 948 949
		ui__warning("%s", errbuf);
	}

950
	err = target__parse_uid(&rec->opts.target);
951 952
	if (err) {
		int saved_errno = errno;
953

954
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
955
		ui__error("%s", errbuf);
956 957

		err = -saved_errno;
958
		goto out_symbol_exit;
959
	}
960

961
	err = -ENOMEM;
962
	if (perf_evlist__create_maps(rec->evlist, &rec->opts.target) < 0)
963
		usage_with_options(record_usage, record_options);
964

965
	if (record_opts__config(&rec->opts)) {
966
		err = -EINVAL;
967
		goto out_free_fd;
968 969
	}

970
	err = __cmd_record(&record, argc, argv);
971

972 973
	perf_evlist__munmap(rec->evlist);
	perf_evlist__close(rec->evlist);
974
out_free_fd:
975
	perf_evlist__delete_maps(rec->evlist);
976 977
out_symbol_exit:
	symbol__exit();
978
	return err;
979
}